shivank21 commited on
Commit
89cf807
·
verified ·
1 Parent(s): 31ad2f1

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151643,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 3584,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 18944,
12
+ "max_position_embeddings": 131072,
13
+ "max_window_layers": 28,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 28,
16
+ "num_hidden_layers": 28,
17
+ "num_key_value_heads": 4,
18
+ "rms_norm_eps": 1e-06,
19
+ "rope_scaling": null,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": 131072,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.51.3",
25
+ "use_cache": false,
26
+ "use_sliding_window": false,
27
+ "vocab_size": 152064
28
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.51.3"
6
+ }
model-00001-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:468b83871f021e2a25abf6d9240ee30f38c371fdda0322d25f79bd3c7e1d5317
3
+ size 4976687216
model-00002-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7978e2c2e92b5e394a5ebb2cd6d8a6c63a83cdfad18592a43b79d4e5f69f68ea
3
+ size 4778622352
model-00003-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff260157b787a34bb38c73acf7a2c23f75b9e5f79cab4ca5424557cbac007b6c
3
+ size 4932743960
model-00004-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:970418097234401db7e8dc8f9d85de9dc16d3225785cdbd2083dfb26d351d586
3
+ size 4932743992
model-00005-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd75650bf34dd8932540917df356ba83c2a487b87d563b9e1c21bbc5e8cf2404
3
+ size 4998852296
model-00006-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cf75dcb0930fb6ce7a4de61bba1cb612e41241fd3be47a071956764ca047e4e
3
+ size 3662865184
model-00007-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a585f22246c30e93cf7e2fe800e7932f560a581c6bb01f461593191413484f37
3
+ size 2179989632
model.safetensors.index.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 30462466048
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00007-of-00007.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00007.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00007.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
13
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00007.safetensors",
14
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
15
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
16
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00007.safetensors",
17
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
18
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00007.safetensors",
19
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
20
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00007.safetensors",
21
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
22
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
23
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
24
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
25
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00007.safetensors",
26
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
27
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
28
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00007.safetensors",
29
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
30
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00007.safetensors",
31
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
32
+ "model.layers.10.input_layernorm.weight": "model-00003-of-00007.safetensors",
33
+ "model.layers.10.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
34
+ "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
35
+ "model.layers.10.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
36
+ "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
37
+ "model.layers.10.self_attn.k_proj.bias": "model-00003-of-00007.safetensors",
38
+ "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
39
+ "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
40
+ "model.layers.10.self_attn.q_proj.bias": "model-00003-of-00007.safetensors",
41
+ "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
42
+ "model.layers.10.self_attn.v_proj.bias": "model-00003-of-00007.safetensors",
43
+ "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
44
+ "model.layers.11.input_layernorm.weight": "model-00003-of-00007.safetensors",
45
+ "model.layers.11.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
46
+ "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
47
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
48
+ "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
49
+ "model.layers.11.self_attn.k_proj.bias": "model-00003-of-00007.safetensors",
50
+ "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
51
+ "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
52
+ "model.layers.11.self_attn.q_proj.bias": "model-00003-of-00007.safetensors",
53
+ "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
54
+ "model.layers.11.self_attn.v_proj.bias": "model-00003-of-00007.safetensors",
55
+ "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
56
+ "model.layers.12.input_layernorm.weight": "model-00003-of-00007.safetensors",
57
+ "model.layers.12.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
58
+ "model.layers.12.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
59
+ "model.layers.12.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
60
+ "model.layers.12.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
61
+ "model.layers.12.self_attn.k_proj.bias": "model-00003-of-00007.safetensors",
62
+ "model.layers.12.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
63
+ "model.layers.12.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
64
+ "model.layers.12.self_attn.q_proj.bias": "model-00003-of-00007.safetensors",
65
+ "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
66
+ "model.layers.12.self_attn.v_proj.bias": "model-00003-of-00007.safetensors",
67
+ "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
68
+ "model.layers.13.input_layernorm.weight": "model-00004-of-00007.safetensors",
69
+ "model.layers.13.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
70
+ "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
71
+ "model.layers.13.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
72
+ "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
73
+ "model.layers.13.self_attn.k_proj.bias": "model-00003-of-00007.safetensors",
74
+ "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
75
+ "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
76
+ "model.layers.13.self_attn.q_proj.bias": "model-00003-of-00007.safetensors",
77
+ "model.layers.13.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
78
+ "model.layers.13.self_attn.v_proj.bias": "model-00003-of-00007.safetensors",
79
+ "model.layers.13.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
80
+ "model.layers.14.input_layernorm.weight": "model-00004-of-00007.safetensors",
81
+ "model.layers.14.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
82
+ "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
83
+ "model.layers.14.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
84
+ "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
85
+ "model.layers.14.self_attn.k_proj.bias": "model-00004-of-00007.safetensors",
86
+ "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
87
+ "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
88
+ "model.layers.14.self_attn.q_proj.bias": "model-00004-of-00007.safetensors",
89
+ "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
90
+ "model.layers.14.self_attn.v_proj.bias": "model-00004-of-00007.safetensors",
91
+ "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
92
+ "model.layers.15.input_layernorm.weight": "model-00004-of-00007.safetensors",
93
+ "model.layers.15.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
94
+ "model.layers.15.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
95
+ "model.layers.15.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
96
+ "model.layers.15.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
97
+ "model.layers.15.self_attn.k_proj.bias": "model-00004-of-00007.safetensors",
98
+ "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
99
+ "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
100
+ "model.layers.15.self_attn.q_proj.bias": "model-00004-of-00007.safetensors",
101
+ "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
102
+ "model.layers.15.self_attn.v_proj.bias": "model-00004-of-00007.safetensors",
103
+ "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
104
+ "model.layers.16.input_layernorm.weight": "model-00004-of-00007.safetensors",
105
+ "model.layers.16.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
106
+ "model.layers.16.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
107
+ "model.layers.16.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
108
+ "model.layers.16.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
109
+ "model.layers.16.self_attn.k_proj.bias": "model-00004-of-00007.safetensors",
110
+ "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
111
+ "model.layers.16.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
112
+ "model.layers.16.self_attn.q_proj.bias": "model-00004-of-00007.safetensors",
113
+ "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
114
+ "model.layers.16.self_attn.v_proj.bias": "model-00004-of-00007.safetensors",
115
+ "model.layers.16.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
116
+ "model.layers.17.input_layernorm.weight": "model-00004-of-00007.safetensors",
117
+ "model.layers.17.mlp.down_proj.weight": "model-00004-of-00007.safetensors",
118
+ "model.layers.17.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
119
+ "model.layers.17.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
120
+ "model.layers.17.post_attention_layernorm.weight": "model-00004-of-00007.safetensors",
121
+ "model.layers.17.self_attn.k_proj.bias": "model-00004-of-00007.safetensors",
122
+ "model.layers.17.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
123
+ "model.layers.17.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
124
+ "model.layers.17.self_attn.q_proj.bias": "model-00004-of-00007.safetensors",
125
+ "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
126
+ "model.layers.17.self_attn.v_proj.bias": "model-00004-of-00007.safetensors",
127
+ "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
128
+ "model.layers.18.input_layernorm.weight": "model-00005-of-00007.safetensors",
129
+ "model.layers.18.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
130
+ "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00007.safetensors",
131
+ "model.layers.18.mlp.up_proj.weight": "model-00004-of-00007.safetensors",
132
+ "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
133
+ "model.layers.18.self_attn.k_proj.bias": "model-00004-of-00007.safetensors",
134
+ "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00007.safetensors",
135
+ "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00007.safetensors",
136
+ "model.layers.18.self_attn.q_proj.bias": "model-00004-of-00007.safetensors",
137
+ "model.layers.18.self_attn.q_proj.weight": "model-00004-of-00007.safetensors",
138
+ "model.layers.18.self_attn.v_proj.bias": "model-00004-of-00007.safetensors",
139
+ "model.layers.18.self_attn.v_proj.weight": "model-00004-of-00007.safetensors",
140
+ "model.layers.19.input_layernorm.weight": "model-00005-of-00007.safetensors",
141
+ "model.layers.19.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
142
+ "model.layers.19.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
143
+ "model.layers.19.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
144
+ "model.layers.19.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
145
+ "model.layers.19.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
146
+ "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
147
+ "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
148
+ "model.layers.19.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
149
+ "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
150
+ "model.layers.19.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
151
+ "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
152
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00007.safetensors",
153
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00007.safetensors",
154
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00007.safetensors",
155
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00007.safetensors",
156
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00007.safetensors",
157
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00007.safetensors",
158
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00007.safetensors",
159
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00007.safetensors",
160
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00007.safetensors",
161
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00007.safetensors",
162
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00007.safetensors",
163
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00007.safetensors",
164
+ "model.layers.20.input_layernorm.weight": "model-00005-of-00007.safetensors",
165
+ "model.layers.20.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
166
+ "model.layers.20.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
167
+ "model.layers.20.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
168
+ "model.layers.20.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
169
+ "model.layers.20.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
170
+ "model.layers.20.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
171
+ "model.layers.20.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
172
+ "model.layers.20.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
173
+ "model.layers.20.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
174
+ "model.layers.20.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
175
+ "model.layers.20.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
176
+ "model.layers.21.input_layernorm.weight": "model-00005-of-00007.safetensors",
177
+ "model.layers.21.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
178
+ "model.layers.21.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
179
+ "model.layers.21.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
180
+ "model.layers.21.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
181
+ "model.layers.21.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
182
+ "model.layers.21.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
183
+ "model.layers.21.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
184
+ "model.layers.21.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
185
+ "model.layers.21.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
186
+ "model.layers.21.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
187
+ "model.layers.21.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
188
+ "model.layers.22.input_layernorm.weight": "model-00005-of-00007.safetensors",
189
+ "model.layers.22.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
190
+ "model.layers.22.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
191
+ "model.layers.22.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
192
+ "model.layers.22.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
193
+ "model.layers.22.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
194
+ "model.layers.22.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
195
+ "model.layers.22.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
196
+ "model.layers.22.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
197
+ "model.layers.22.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
198
+ "model.layers.22.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
199
+ "model.layers.22.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
200
+ "model.layers.23.input_layernorm.weight": "model-00005-of-00007.safetensors",
201
+ "model.layers.23.mlp.down_proj.weight": "model-00005-of-00007.safetensors",
202
+ "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00007.safetensors",
203
+ "model.layers.23.mlp.up_proj.weight": "model-00005-of-00007.safetensors",
204
+ "model.layers.23.post_attention_layernorm.weight": "model-00005-of-00007.safetensors",
205
+ "model.layers.23.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
206
+ "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
207
+ "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00007.safetensors",
208
+ "model.layers.23.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
209
+ "model.layers.23.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
210
+ "model.layers.23.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
211
+ "model.layers.23.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
212
+ "model.layers.24.input_layernorm.weight": "model-00006-of-00007.safetensors",
213
+ "model.layers.24.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
214
+ "model.layers.24.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
215
+ "model.layers.24.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
216
+ "model.layers.24.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
217
+ "model.layers.24.self_attn.k_proj.bias": "model-00005-of-00007.safetensors",
218
+ "model.layers.24.self_attn.k_proj.weight": "model-00005-of-00007.safetensors",
219
+ "model.layers.24.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
220
+ "model.layers.24.self_attn.q_proj.bias": "model-00005-of-00007.safetensors",
221
+ "model.layers.24.self_attn.q_proj.weight": "model-00005-of-00007.safetensors",
222
+ "model.layers.24.self_attn.v_proj.bias": "model-00005-of-00007.safetensors",
223
+ "model.layers.24.self_attn.v_proj.weight": "model-00005-of-00007.safetensors",
224
+ "model.layers.25.input_layernorm.weight": "model-00006-of-00007.safetensors",
225
+ "model.layers.25.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
226
+ "model.layers.25.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
227
+ "model.layers.25.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
228
+ "model.layers.25.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
229
+ "model.layers.25.self_attn.k_proj.bias": "model-00006-of-00007.safetensors",
230
+ "model.layers.25.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
231
+ "model.layers.25.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
232
+ "model.layers.25.self_attn.q_proj.bias": "model-00006-of-00007.safetensors",
233
+ "model.layers.25.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
234
+ "model.layers.25.self_attn.v_proj.bias": "model-00006-of-00007.safetensors",
235
+ "model.layers.25.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
236
+ "model.layers.26.input_layernorm.weight": "model-00006-of-00007.safetensors",
237
+ "model.layers.26.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
238
+ "model.layers.26.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
239
+ "model.layers.26.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
240
+ "model.layers.26.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
241
+ "model.layers.26.self_attn.k_proj.bias": "model-00006-of-00007.safetensors",
242
+ "model.layers.26.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
243
+ "model.layers.26.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
244
+ "model.layers.26.self_attn.q_proj.bias": "model-00006-of-00007.safetensors",
245
+ "model.layers.26.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
246
+ "model.layers.26.self_attn.v_proj.bias": "model-00006-of-00007.safetensors",
247
+ "model.layers.26.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
248
+ "model.layers.27.input_layernorm.weight": "model-00006-of-00007.safetensors",
249
+ "model.layers.27.mlp.down_proj.weight": "model-00006-of-00007.safetensors",
250
+ "model.layers.27.mlp.gate_proj.weight": "model-00006-of-00007.safetensors",
251
+ "model.layers.27.mlp.up_proj.weight": "model-00006-of-00007.safetensors",
252
+ "model.layers.27.post_attention_layernorm.weight": "model-00006-of-00007.safetensors",
253
+ "model.layers.27.self_attn.k_proj.bias": "model-00006-of-00007.safetensors",
254
+ "model.layers.27.self_attn.k_proj.weight": "model-00006-of-00007.safetensors",
255
+ "model.layers.27.self_attn.o_proj.weight": "model-00006-of-00007.safetensors",
256
+ "model.layers.27.self_attn.q_proj.bias": "model-00006-of-00007.safetensors",
257
+ "model.layers.27.self_attn.q_proj.weight": "model-00006-of-00007.safetensors",
258
+ "model.layers.27.self_attn.v_proj.bias": "model-00006-of-00007.safetensors",
259
+ "model.layers.27.self_attn.v_proj.weight": "model-00006-of-00007.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00007.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
265
+ "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
266
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
267
+ "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
268
+ "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
270
+ "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
271
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
272
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00007.safetensors",
273
+ "model.layers.4.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
274
+ "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
275
+ "model.layers.4.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
276
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
277
+ "model.layers.4.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
278
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
279
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
280
+ "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
281
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
282
+ "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
283
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
284
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00007.safetensors",
285
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
286
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
287
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
288
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
289
+ "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
290
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
291
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
292
+ "model.layers.5.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
293
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
294
+ "model.layers.5.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
295
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
296
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00007.safetensors",
297
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
298
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
299
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
300
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
301
+ "model.layers.6.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
302
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
303
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
304
+ "model.layers.6.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
305
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
306
+ "model.layers.6.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
307
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
308
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00007.safetensors",
309
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00007.safetensors",
310
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00007.safetensors",
311
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00007.safetensors",
312
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00007.safetensors",
313
+ "model.layers.7.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
314
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
315
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
316
+ "model.layers.7.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
317
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
318
+ "model.layers.7.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
319
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
320
+ "model.layers.8.input_layernorm.weight": "model-00003-of-00007.safetensors",
321
+ "model.layers.8.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
322
+ "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
323
+ "model.layers.8.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
324
+ "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
325
+ "model.layers.8.self_attn.k_proj.bias": "model-00002-of-00007.safetensors",
326
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00007.safetensors",
327
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00007.safetensors",
328
+ "model.layers.8.self_attn.q_proj.bias": "model-00002-of-00007.safetensors",
329
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00007.safetensors",
330
+ "model.layers.8.self_attn.v_proj.bias": "model-00002-of-00007.safetensors",
331
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00007.safetensors",
332
+ "model.layers.9.input_layernorm.weight": "model-00003-of-00007.safetensors",
333
+ "model.layers.9.mlp.down_proj.weight": "model-00003-of-00007.safetensors",
334
+ "model.layers.9.mlp.gate_proj.weight": "model-00003-of-00007.safetensors",
335
+ "model.layers.9.mlp.up_proj.weight": "model-00003-of-00007.safetensors",
336
+ "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00007.safetensors",
337
+ "model.layers.9.self_attn.k_proj.bias": "model-00003-of-00007.safetensors",
338
+ "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00007.safetensors",
339
+ "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00007.safetensors",
340
+ "model.layers.9.self_attn.q_proj.bias": "model-00003-of-00007.safetensors",
341
+ "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00007.safetensors",
342
+ "model.layers.9.self_attn.v_proj.bias": "model-00003-of-00007.safetensors",
343
+ "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00007.safetensors",
344
+ "model.norm.weight": "model-00006-of-00007.safetensors"
345
+ }
346
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|endoftext|>",
37
+ "errors": "replace",
38
+ "extra_special_tokens": {},
39
+ "model_max_length": 32768,
40
+ "pad_token": "<|endoftext|>",
41
+ "split_special_tokens": false,
42
+ "tokenizer_class": "Qwen2Tokenizer",
43
+ "unk_token": null
44
+ }
trainer_state.json ADDED
@@ -0,0 +1,2031 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 7500,
3
+ "best_metric": 0.7491397857666016,
4
+ "best_model_checkpoint": "./results/checkpoint-7500",
5
+ "epoch": 4.997752808988764,
6
+ "eval_steps": 250,
7
+ "global_step": 9455,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.026437541308658295,
14
+ "grad_norm": 1.9868909120559692,
15
+ "learning_rate": 0.00034625958983852136,
16
+ "loss": 1.3782,
17
+ "mean_token_accuracy": 0.7697586753964424,
18
+ "num_tokens": 1638400.0,
19
+ "step": 50
20
+ },
21
+ {
22
+ "epoch": 0.05287508261731659,
23
+ "grad_norm": 0.9880499839782715,
24
+ "learning_rate": 0.000407611186724682,
25
+ "loss": 1.2966,
26
+ "mean_token_accuracy": 0.7892405907809734,
27
+ "num_tokens": 3276800.0,
28
+ "step": 100
29
+ },
30
+ {
31
+ "epoch": 0.07931262392597488,
32
+ "grad_norm": 0.6774707436561584,
33
+ "learning_rate": 0.0004434995702624468,
34
+ "loss": 0.939,
35
+ "mean_token_accuracy": 0.8182881197333336,
36
+ "num_tokens": 4915200.0,
37
+ "step": 150
38
+ },
39
+ {
40
+ "epoch": 0.10575016523463318,
41
+ "grad_norm": 0.84303218126297,
42
+ "learning_rate": 0.0004689627836108426,
43
+ "loss": 0.9239,
44
+ "mean_token_accuracy": 0.8214302301406861,
45
+ "num_tokens": 6553600.0,
46
+ "step": 200
47
+ },
48
+ {
49
+ "epoch": 0.13218770654329148,
50
+ "grad_norm": 2.8259122371673584,
51
+ "learning_rate": 0.0004887135863147016,
52
+ "loss": 1.0103,
53
+ "step": 250
54
+ },
55
+ {
56
+ "epoch": 0.13218770654329148,
57
+ "eval_loss": 1.3388922214508057,
58
+ "eval_mean_token_accuracy": 0.767919923740008,
59
+ "eval_num_tokens": 8192000.0,
60
+ "eval_runtime": 1597.0105,
61
+ "eval_samples_per_second": 4.737,
62
+ "eval_steps_per_second": 0.592,
63
+ "step": 250
64
+ },
65
+ {
66
+ "epoch": 0.15862524785194976,
67
+ "grad_norm": 6.46172571182251,
68
+ "learning_rate": 0.0004991822047759241,
69
+ "loss": 1.6239,
70
+ "mean_token_accuracy": 0.767223238646984,
71
+ "num_tokens": 9830400.0,
72
+ "step": 300
73
+ },
74
+ {
75
+ "epoch": 0.18506278916060806,
76
+ "grad_norm": 17.279727935791016,
77
+ "learning_rate": 0.0004964562206956711,
78
+ "loss": 1.3753,
79
+ "mean_token_accuracy": 0.7646566477417945,
80
+ "num_tokens": 11468800.0,
81
+ "step": 350
82
+ },
83
+ {
84
+ "epoch": 0.21150033046926636,
85
+ "grad_norm": 9.562877655029297,
86
+ "learning_rate": 0.0004937302366154182,
87
+ "loss": 1.7307,
88
+ "mean_token_accuracy": 0.7092528110742569,
89
+ "num_tokens": 13107200.0,
90
+ "step": 400
91
+ },
92
+ {
93
+ "epoch": 0.23793787177792466,
94
+ "grad_norm": 57.86962890625,
95
+ "learning_rate": 0.0004910042525351653,
96
+ "loss": 3.8636,
97
+ "mean_token_accuracy": 0.44049181263893844,
98
+ "num_tokens": 14745600.0,
99
+ "step": 450
100
+ },
101
+ {
102
+ "epoch": 0.26437541308658297,
103
+ "grad_norm": 2.1134138107299805,
104
+ "learning_rate": 0.00048827826845491225,
105
+ "loss": 3.3381,
106
+ "step": 500
107
+ },
108
+ {
109
+ "epoch": 0.26437541308658297,
110
+ "eval_loss": 3.0370047092437744,
111
+ "eval_mean_token_accuracy": 0.5421812577177052,
112
+ "eval_num_tokens": 16384000.0,
113
+ "eval_runtime": 1593.4634,
114
+ "eval_samples_per_second": 4.748,
115
+ "eval_steps_per_second": 0.594,
116
+ "step": 500
117
+ },
118
+ {
119
+ "epoch": 0.29081295439524124,
120
+ "grad_norm": 2.805110454559326,
121
+ "learning_rate": 0.0004855522843746593,
122
+ "loss": 2.8023,
123
+ "mean_token_accuracy": 0.5313697456568479,
124
+ "num_tokens": 18022400.0,
125
+ "step": 550
126
+ },
127
+ {
128
+ "epoch": 0.3172504957038995,
129
+ "grad_norm": 2.2337939739227295,
130
+ "learning_rate": 0.00048282630029440626,
131
+ "loss": 2.3643,
132
+ "mean_token_accuracy": 0.6257539093494415,
133
+ "num_tokens": 19660800.0,
134
+ "step": 600
135
+ },
136
+ {
137
+ "epoch": 0.34368803701255785,
138
+ "grad_norm": 1.3661304712295532,
139
+ "learning_rate": 0.00048010031621415335,
140
+ "loss": 2.2117,
141
+ "mean_token_accuracy": 0.6438269788026809,
142
+ "num_tokens": 21299200.0,
143
+ "step": 650
144
+ },
145
+ {
146
+ "epoch": 0.3701255783212161,
147
+ "grad_norm": 1.319533109664917,
148
+ "learning_rate": 0.0004773743321339004,
149
+ "loss": 2.0377,
150
+ "mean_token_accuracy": 0.6656103357672691,
151
+ "num_tokens": 22937600.0,
152
+ "step": 700
153
+ },
154
+ {
155
+ "epoch": 0.3965631196298744,
156
+ "grad_norm": 0.9889560341835022,
157
+ "learning_rate": 0.00047464834805364736,
158
+ "loss": 1.9612,
159
+ "step": 750
160
+ },
161
+ {
162
+ "epoch": 0.3965631196298744,
163
+ "eval_loss": 1.9166280031204224,
164
+ "eval_mean_token_accuracy": 0.6814079303570076,
165
+ "eval_num_tokens": 24576000.0,
166
+ "eval_runtime": 1597.3511,
167
+ "eval_samples_per_second": 4.736,
168
+ "eval_steps_per_second": 0.592,
169
+ "step": 750
170
+ },
171
+ {
172
+ "epoch": 0.4230006609385327,
173
+ "grad_norm": 0.9154905676841736,
174
+ "learning_rate": 0.0004719223639733944,
175
+ "loss": 1.895,
176
+ "mean_token_accuracy": 0.6779982282221317,
177
+ "num_tokens": 26214400.0,
178
+ "step": 800
179
+ },
180
+ {
181
+ "epoch": 0.449438202247191,
182
+ "grad_norm": 1.1073395013809204,
183
+ "learning_rate": 0.0004691963798931415,
184
+ "loss": 1.8532,
185
+ "mean_token_accuracy": 0.6910386118292808,
186
+ "num_tokens": 27852800.0,
187
+ "step": 850
188
+ },
189
+ {
190
+ "epoch": 0.47587574355584933,
191
+ "grad_norm": 0.7696494460105896,
192
+ "learning_rate": 0.00046647039581288846,
193
+ "loss": 1.7828,
194
+ "mean_token_accuracy": 0.7011858496069908,
195
+ "num_tokens": 29491200.0,
196
+ "step": 900
197
+ },
198
+ {
199
+ "epoch": 0.5023132848645075,
200
+ "grad_norm": 0.8735769987106323,
201
+ "learning_rate": 0.0004637444117326355,
202
+ "loss": 1.7661,
203
+ "mean_token_accuracy": 0.7030816239118576,
204
+ "num_tokens": 31129600.0,
205
+ "step": 950
206
+ },
207
+ {
208
+ "epoch": 0.5287508261731659,
209
+ "grad_norm": 0.8184662461280823,
210
+ "learning_rate": 0.0004610184276523825,
211
+ "loss": 1.7303,
212
+ "step": 1000
213
+ },
214
+ {
215
+ "epoch": 0.5287508261731659,
216
+ "eval_loss": 1.732823133468628,
217
+ "eval_mean_token_accuracy": 0.7076210416774669,
218
+ "eval_num_tokens": 32768000.0,
219
+ "eval_runtime": 1598.294,
220
+ "eval_samples_per_second": 4.733,
221
+ "eval_steps_per_second": 0.592,
222
+ "step": 1000
223
+ },
224
+ {
225
+ "epoch": 0.5551883674818242,
226
+ "grad_norm": 0.805102527141571,
227
+ "learning_rate": 0.00045829244357212956,
228
+ "loss": 1.7074,
229
+ "mean_token_accuracy": 0.7088553883135319,
230
+ "num_tokens": 34406400.0,
231
+ "step": 1050
232
+ },
233
+ {
234
+ "epoch": 0.5816259087904825,
235
+ "grad_norm": 1.1714200973510742,
236
+ "learning_rate": 0.0004555664594918766,
237
+ "loss": 1.6578,
238
+ "mean_token_accuracy": 0.7183681574463844,
239
+ "num_tokens": 36044800.0,
240
+ "step": 1100
241
+ },
242
+ {
243
+ "epoch": 0.6080634500991408,
244
+ "grad_norm": 0.801713228225708,
245
+ "learning_rate": 0.0004528404754116236,
246
+ "loss": 1.639,
247
+ "mean_token_accuracy": 0.7187829902768135,
248
+ "num_tokens": 37683200.0,
249
+ "step": 1150
250
+ },
251
+ {
252
+ "epoch": 0.634500991407799,
253
+ "grad_norm": 0.776907205581665,
254
+ "learning_rate": 0.0004501144913313706,
255
+ "loss": 1.6155,
256
+ "mean_token_accuracy": 0.7220100191235542,
257
+ "num_tokens": 39321600.0,
258
+ "step": 1200
259
+ },
260
+ {
261
+ "epoch": 0.6609385327164574,
262
+ "grad_norm": 0.5754767656326294,
263
+ "learning_rate": 0.0004473885072511177,
264
+ "loss": 1.5987,
265
+ "step": 1250
266
+ },
267
+ {
268
+ "epoch": 0.6609385327164574,
269
+ "eval_loss": 1.5740511417388916,
270
+ "eval_mean_token_accuracy": 0.7283713231001041,
271
+ "eval_num_tokens": 40960000.0,
272
+ "eval_runtime": 1599.1655,
273
+ "eval_samples_per_second": 4.731,
274
+ "eval_steps_per_second": 0.592,
275
+ "step": 1250
276
+ },
277
+ {
278
+ "epoch": 0.6873760740251157,
279
+ "grad_norm": 0.5739009976387024,
280
+ "learning_rate": 0.0004446625231708647,
281
+ "loss": 1.5626,
282
+ "mean_token_accuracy": 0.7280584080517292,
283
+ "num_tokens": 42598400.0,
284
+ "step": 1300
285
+ },
286
+ {
287
+ "epoch": 0.713813615333774,
288
+ "grad_norm": 0.56184983253479,
289
+ "learning_rate": 0.0004419365390906117,
290
+ "loss": 1.5383,
291
+ "mean_token_accuracy": 0.7336229240894317,
292
+ "num_tokens": 44236800.0,
293
+ "step": 1350
294
+ },
295
+ {
296
+ "epoch": 0.7402511566424322,
297
+ "grad_norm": 0.7078151106834412,
298
+ "learning_rate": 0.00043921055501035873,
299
+ "loss": 1.4998,
300
+ "mean_token_accuracy": 0.7408009549975395,
301
+ "num_tokens": 45875200.0,
302
+ "step": 1400
303
+ },
304
+ {
305
+ "epoch": 0.7666886979510905,
306
+ "grad_norm": 0.6344922184944153,
307
+ "learning_rate": 0.0004364845709301058,
308
+ "loss": 1.4751,
309
+ "mean_token_accuracy": 0.74332783639431,
310
+ "num_tokens": 47513600.0,
311
+ "step": 1450
312
+ },
313
+ {
314
+ "epoch": 0.7931262392597488,
315
+ "grad_norm": 0.4986041486263275,
316
+ "learning_rate": 0.0004337585868498528,
317
+ "loss": 1.4716,
318
+ "step": 1500
319
+ },
320
+ {
321
+ "epoch": 0.7931262392597488,
322
+ "eval_loss": 1.4890165328979492,
323
+ "eval_mean_token_accuracy": 0.7397930833174361,
324
+ "eval_num_tokens": 49152000.0,
325
+ "eval_runtime": 1599.8275,
326
+ "eval_samples_per_second": 4.729,
327
+ "eval_steps_per_second": 0.591,
328
+ "step": 1500
329
+ },
330
+ {
331
+ "epoch": 0.8195637805684072,
332
+ "grad_norm": 0.508725643157959,
333
+ "learning_rate": 0.00043103260276959983,
334
+ "loss": 1.4734,
335
+ "mean_token_accuracy": 0.7423943056166172,
336
+ "num_tokens": 50790400.0,
337
+ "step": 1550
338
+ },
339
+ {
340
+ "epoch": 0.8460013218770654,
341
+ "grad_norm": 0.5139680504798889,
342
+ "learning_rate": 0.0004283066186893468,
343
+ "loss": 1.4513,
344
+ "mean_token_accuracy": 0.7446151030063629,
345
+ "num_tokens": 52428800.0,
346
+ "step": 1600
347
+ },
348
+ {
349
+ "epoch": 0.8724388631857237,
350
+ "grad_norm": 0.5360570549964905,
351
+ "learning_rate": 0.0004255806346090939,
352
+ "loss": 1.4587,
353
+ "mean_token_accuracy": 0.7414125129580498,
354
+ "num_tokens": 54067200.0,
355
+ "step": 1650
356
+ },
357
+ {
358
+ "epoch": 0.898876404494382,
359
+ "grad_norm": 0.46601545810699463,
360
+ "learning_rate": 0.00042285465052884093,
361
+ "loss": 1.4468,
362
+ "mean_token_accuracy": 0.7438508039712906,
363
+ "num_tokens": 55705600.0,
364
+ "step": 1700
365
+ },
366
+ {
367
+ "epoch": 0.9253139458030403,
368
+ "grad_norm": 0.4491994380950928,
369
+ "learning_rate": 0.0004201286664485879,
370
+ "loss": 1.4234,
371
+ "step": 1750
372
+ },
373
+ {
374
+ "epoch": 0.9253139458030403,
375
+ "eval_loss": 1.405325174331665,
376
+ "eval_mean_token_accuracy": 0.7498895639975025,
377
+ "eval_num_tokens": 57344000.0,
378
+ "eval_runtime": 1600.1713,
379
+ "eval_samples_per_second": 4.728,
380
+ "eval_steps_per_second": 0.591,
381
+ "step": 1750
382
+ },
383
+ {
384
+ "epoch": 0.9517514871116987,
385
+ "grad_norm": 0.40710222721099854,
386
+ "learning_rate": 0.00041740268236833495,
387
+ "loss": 1.3835,
388
+ "mean_token_accuracy": 0.7491015987098217,
389
+ "num_tokens": 58982400.0,
390
+ "step": 1800
391
+ },
392
+ {
393
+ "epoch": 0.9781890284203569,
394
+ "grad_norm": 0.4693559408187866,
395
+ "learning_rate": 0.00041467669828808203,
396
+ "loss": 1.3498,
397
+ "mean_token_accuracy": 0.7580449655652046,
398
+ "num_tokens": 60620800.0,
399
+ "step": 1850
400
+ },
401
+ {
402
+ "epoch": 1.0042300066093852,
403
+ "grad_norm": 0.4662095606327057,
404
+ "learning_rate": 0.00041200523388943414,
405
+ "loss": 1.3744,
406
+ "mean_token_accuracy": 0.7523588169044649,
407
+ "num_tokens": 62234624.0,
408
+ "step": 1900
409
+ },
410
+ {
411
+ "epoch": 1.0306675479180436,
412
+ "grad_norm": 0.39532390236854553,
413
+ "learning_rate": 0.0004092792498091811,
414
+ "loss": 1.3057,
415
+ "mean_token_accuracy": 0.7592387574911118,
416
+ "num_tokens": 63873024.0,
417
+ "step": 1950
418
+ },
419
+ {
420
+ "epoch": 1.057105089226702,
421
+ "grad_norm": 0.4154648780822754,
422
+ "learning_rate": 0.00040655326572892816,
423
+ "loss": 1.3248,
424
+ "step": 2000
425
+ },
426
+ {
427
+ "epoch": 1.057105089226702,
428
+ "eval_loss": 1.339290976524353,
429
+ "eval_mean_token_accuracy": 0.7579027240422513,
430
+ "eval_num_tokens": 65511424.0,
431
+ "eval_runtime": 1599.5937,
432
+ "eval_samples_per_second": 4.729,
433
+ "eval_steps_per_second": 0.591,
434
+ "step": 2000
435
+ },
436
+ {
437
+ "epoch": 1.0835426305353602,
438
+ "grad_norm": 0.3895817697048187,
439
+ "learning_rate": 0.00040382728164867513,
440
+ "loss": 1.2956,
441
+ "mean_token_accuracy": 0.7590417274832726,
442
+ "num_tokens": 67149824.0,
443
+ "step": 2050
444
+ },
445
+ {
446
+ "epoch": 1.1099801718440185,
447
+ "grad_norm": 0.39260634779930115,
448
+ "learning_rate": 0.0004011012975684222,
449
+ "loss": 1.3223,
450
+ "mean_token_accuracy": 0.756996577680111,
451
+ "num_tokens": 68788224.0,
452
+ "step": 2100
453
+ },
454
+ {
455
+ "epoch": 1.1364177131526767,
456
+ "grad_norm": 0.3638737201690674,
457
+ "learning_rate": 0.00039837531348816925,
458
+ "loss": 1.268,
459
+ "mean_token_accuracy": 0.7646328181028366,
460
+ "num_tokens": 70426624.0,
461
+ "step": 2150
462
+ },
463
+ {
464
+ "epoch": 1.162855254461335,
465
+ "grad_norm": 0.3186447322368622,
466
+ "learning_rate": 0.00039564932940791623,
467
+ "loss": 1.2705,
468
+ "mean_token_accuracy": 0.7647727259993553,
469
+ "num_tokens": 72065024.0,
470
+ "step": 2200
471
+ },
472
+ {
473
+ "epoch": 1.1892927957699935,
474
+ "grad_norm": 0.37439003586769104,
475
+ "learning_rate": 0.00039292334532766327,
476
+ "loss": 1.2631,
477
+ "step": 2250
478
+ },
479
+ {
480
+ "epoch": 1.1892927957699935,
481
+ "eval_loss": 1.278181791305542,
482
+ "eval_mean_token_accuracy": 0.7655498584531578,
483
+ "eval_num_tokens": 73703424.0,
484
+ "eval_runtime": 1599.9443,
485
+ "eval_samples_per_second": 4.728,
486
+ "eval_steps_per_second": 0.591,
487
+ "step": 2250
488
+ },
489
+ {
490
+ "epoch": 1.2157303370786516,
491
+ "grad_norm": 0.36414834856987,
492
+ "learning_rate": 0.00039019736124741035,
493
+ "loss": 1.2556,
494
+ "mean_token_accuracy": 0.7657642959058285,
495
+ "num_tokens": 75341824.0,
496
+ "step": 2300
497
+ },
498
+ {
499
+ "epoch": 1.24216787838731,
500
+ "grad_norm": 0.38630911707878113,
501
+ "learning_rate": 0.00038747137716715733,
502
+ "loss": 1.2453,
503
+ "mean_token_accuracy": 0.7686192587018013,
504
+ "num_tokens": 76980224.0,
505
+ "step": 2350
506
+ },
507
+ {
508
+ "epoch": 1.2686054196959682,
509
+ "grad_norm": 0.34793412685394287,
510
+ "learning_rate": 0.00038474539308690437,
511
+ "loss": 1.2113,
512
+ "mean_token_accuracy": 0.7736504143476486,
513
+ "num_tokens": 78618624.0,
514
+ "step": 2400
515
+ },
516
+ {
517
+ "epoch": 1.2950429610046266,
518
+ "grad_norm": 0.3544578552246094,
519
+ "learning_rate": 0.0003820194090066514,
520
+ "loss": 1.1924,
521
+ "mean_token_accuracy": 0.7755889534950257,
522
+ "num_tokens": 80257024.0,
523
+ "step": 2450
524
+ },
525
+ {
526
+ "epoch": 1.321480502313285,
527
+ "grad_norm": 0.30794623494148254,
528
+ "learning_rate": 0.00037929342492639843,
529
+ "loss": 1.1748,
530
+ "step": 2500
531
+ },
532
+ {
533
+ "epoch": 1.321480502313285,
534
+ "eval_loss": 1.198885440826416,
535
+ "eval_mean_token_accuracy": 0.776488389197666,
536
+ "eval_num_tokens": 81895424.0,
537
+ "eval_runtime": 1599.7215,
538
+ "eval_samples_per_second": 4.729,
539
+ "eval_steps_per_second": 0.591,
540
+ "step": 2500
541
+ },
542
+ {
543
+ "epoch": 1.3479180436219431,
544
+ "grad_norm": 0.2978927493095398,
545
+ "learning_rate": 0.00037656744084614547,
546
+ "loss": 1.1624,
547
+ "mean_token_accuracy": 0.7798365721106529,
548
+ "num_tokens": 83533824.0,
549
+ "step": 2550
550
+ },
551
+ {
552
+ "epoch": 1.3743555849306015,
553
+ "grad_norm": 0.3153753876686096,
554
+ "learning_rate": 0.0003738414567658925,
555
+ "loss": 1.1462,
556
+ "mean_token_accuracy": 0.7827309390902519,
557
+ "num_tokens": 85172224.0,
558
+ "step": 2600
559
+ },
560
+ {
561
+ "epoch": 1.4007931262392597,
562
+ "grad_norm": 0.31813791394233704,
563
+ "learning_rate": 0.0003711154726856395,
564
+ "loss": 1.1274,
565
+ "mean_token_accuracy": 0.7860245615243912,
566
+ "num_tokens": 86810624.0,
567
+ "step": 2650
568
+ },
569
+ {
570
+ "epoch": 1.427230667547918,
571
+ "grad_norm": 0.30844855308532715,
572
+ "learning_rate": 0.00036838948860538656,
573
+ "loss": 1.118,
574
+ "mean_token_accuracy": 0.7876050838828087,
575
+ "num_tokens": 88449024.0,
576
+ "step": 2700
577
+ },
578
+ {
579
+ "epoch": 1.4536682088565764,
580
+ "grad_norm": 0.3054572343826294,
581
+ "learning_rate": 0.0003656635045251336,
582
+ "loss": 1.1336,
583
+ "step": 2750
584
+ },
585
+ {
586
+ "epoch": 1.4536682088565764,
587
+ "eval_loss": 1.1190927028656006,
588
+ "eval_mean_token_accuracy": 0.7886134508926319,
589
+ "eval_num_tokens": 90087424.0,
590
+ "eval_runtime": 1599.6779,
591
+ "eval_samples_per_second": 4.729,
592
+ "eval_steps_per_second": 0.591,
593
+ "step": 2750
594
+ },
595
+ {
596
+ "epoch": 1.4801057501652346,
597
+ "grad_norm": 0.28417208790779114,
598
+ "learning_rate": 0.0003629375204448806,
599
+ "loss": 1.1039,
600
+ "mean_token_accuracy": 0.7864858260750771,
601
+ "num_tokens": 91725824.0,
602
+ "step": 2800
603
+ },
604
+ {
605
+ "epoch": 1.5065432914738928,
606
+ "grad_norm": 0.307099312543869,
607
+ "learning_rate": 0.0003602115363646276,
608
+ "loss": 1.0909,
609
+ "mean_token_accuracy": 0.7900452110171318,
610
+ "num_tokens": 93364224.0,
611
+ "step": 2850
612
+ },
613
+ {
614
+ "epoch": 1.5329808327825512,
615
+ "grad_norm": 0.30008023977279663,
616
+ "learning_rate": 0.0003574855522843747,
617
+ "loss": 1.0824,
618
+ "mean_token_accuracy": 0.7928054749965667,
619
+ "num_tokens": 95002624.0,
620
+ "step": 2900
621
+ },
622
+ {
623
+ "epoch": 1.5594183740912095,
624
+ "grad_norm": 0.27622541785240173,
625
+ "learning_rate": 0.0003547595682041217,
626
+ "loss": 1.055,
627
+ "mean_token_accuracy": 0.7961241453886032,
628
+ "num_tokens": 96641024.0,
629
+ "step": 2950
630
+ },
631
+ {
632
+ "epoch": 1.585855915399868,
633
+ "grad_norm": 0.2670520544052124,
634
+ "learning_rate": 0.0003520335841238687,
635
+ "loss": 1.0466,
636
+ "step": 3000
637
+ },
638
+ {
639
+ "epoch": 1.585855915399868,
640
+ "eval_loss": 1.0528658628463745,
641
+ "eval_mean_token_accuracy": 0.7982593539149262,
642
+ "eval_num_tokens": 98279424.0,
643
+ "eval_runtime": 1599.5737,
644
+ "eval_samples_per_second": 4.729,
645
+ "eval_steps_per_second": 0.591,
646
+ "step": 3000
647
+ },
648
+ {
649
+ "epoch": 1.612293456708526,
650
+ "grad_norm": 0.26690635085105896,
651
+ "learning_rate": 0.00034930760004361574,
652
+ "loss": 1.0354,
653
+ "mean_token_accuracy": 0.7988346171379089,
654
+ "num_tokens": 99917824.0,
655
+ "step": 3050
656
+ },
657
+ {
658
+ "epoch": 1.6387309980171842,
659
+ "grad_norm": 0.27989307045936584,
660
+ "learning_rate": 0.0003465816159633628,
661
+ "loss": 1.0225,
662
+ "mean_token_accuracy": 0.8015902996063232,
663
+ "num_tokens": 101556224.0,
664
+ "step": 3100
665
+ },
666
+ {
667
+ "epoch": 1.6651685393258426,
668
+ "grad_norm": 0.21368129551410675,
669
+ "learning_rate": 0.0003438556318831098,
670
+ "loss": 1.0197,
671
+ "mean_token_accuracy": 0.8019238775968551,
672
+ "num_tokens": 103194624.0,
673
+ "step": 3150
674
+ },
675
+ {
676
+ "epoch": 1.691606080634501,
677
+ "grad_norm": 0.288343220949173,
678
+ "learning_rate": 0.00034112964780285684,
679
+ "loss": 1.0174,
680
+ "mean_token_accuracy": 0.8012603887915611,
681
+ "num_tokens": 104833024.0,
682
+ "step": 3200
683
+ },
684
+ {
685
+ "epoch": 1.7180436219431594,
686
+ "grad_norm": 0.245047464966774,
687
+ "learning_rate": 0.0003384036637226039,
688
+ "loss": 0.9922,
689
+ "step": 3250
690
+ },
691
+ {
692
+ "epoch": 1.7180436219431594,
693
+ "eval_loss": 1.0065803527832031,
694
+ "eval_mean_token_accuracy": 0.8048020523773943,
695
+ "eval_num_tokens": 106471424.0,
696
+ "eval_runtime": 1599.1563,
697
+ "eval_samples_per_second": 4.731,
698
+ "eval_steps_per_second": 0.592,
699
+ "step": 3250
700
+ },
701
+ {
702
+ "epoch": 1.7444811632518176,
703
+ "grad_norm": 0.23827126622200012,
704
+ "learning_rate": 0.0003356776796423509,
705
+ "loss": 0.9838,
706
+ "mean_token_accuracy": 0.8066547532379628,
707
+ "num_tokens": 108109824.0,
708
+ "step": 3300
709
+ },
710
+ {
711
+ "epoch": 1.7709187045604757,
712
+ "grad_norm": 0.22703391313552856,
713
+ "learning_rate": 0.00033295169556209794,
714
+ "loss": 0.9587,
715
+ "mean_token_accuracy": 0.8113178130984307,
716
+ "num_tokens": 109748224.0,
717
+ "step": 3350
718
+ },
719
+ {
720
+ "epoch": 1.7973562458691341,
721
+ "grad_norm": 0.25331422686576843,
722
+ "learning_rate": 0.0003302257114818449,
723
+ "loss": 0.9697,
724
+ "mean_token_accuracy": 0.8093206259608269,
725
+ "num_tokens": 111386624.0,
726
+ "step": 3400
727
+ },
728
+ {
729
+ "epoch": 1.8237937871777925,
730
+ "grad_norm": 0.264260470867157,
731
+ "learning_rate": 0.000327499727401592,
732
+ "loss": 0.956,
733
+ "mean_token_accuracy": 0.8123435971140861,
734
+ "num_tokens": 113025024.0,
735
+ "step": 3450
736
+ },
737
+ {
738
+ "epoch": 1.8502313284864509,
739
+ "grad_norm": 0.2458537220954895,
740
+ "learning_rate": 0.00032477374332133904,
741
+ "loss": 0.9539,
742
+ "step": 3500
743
+ },
744
+ {
745
+ "epoch": 1.8502313284864509,
746
+ "eval_loss": 0.9670175909996033,
747
+ "eval_mean_token_accuracy": 0.8104942284729214,
748
+ "eval_num_tokens": 114663424.0,
749
+ "eval_runtime": 1599.1718,
750
+ "eval_samples_per_second": 4.731,
751
+ "eval_steps_per_second": 0.592,
752
+ "step": 3500
753
+ },
754
+ {
755
+ "epoch": 1.876668869795109,
756
+ "grad_norm": 0.20451125502586365,
757
+ "learning_rate": 0.000322047759241086,
758
+ "loss": 0.9479,
759
+ "mean_token_accuracy": 0.8118679732084274,
760
+ "num_tokens": 116301824.0,
761
+ "step": 3550
762
+ },
763
+ {
764
+ "epoch": 1.9031064111037672,
765
+ "grad_norm": 0.22584660351276398,
766
+ "learning_rate": 0.00031932177516083305,
767
+ "loss": 0.9688,
768
+ "mean_token_accuracy": 0.8094049346446991,
769
+ "num_tokens": 117940224.0,
770
+ "step": 3600
771
+ },
772
+ {
773
+ "epoch": 1.9295439524124256,
774
+ "grad_norm": 0.2119428962469101,
775
+ "learning_rate": 0.00031659579108058014,
776
+ "loss": 0.9229,
777
+ "mean_token_accuracy": 0.816270771920681,
778
+ "num_tokens": 119578624.0,
779
+ "step": 3650
780
+ },
781
+ {
782
+ "epoch": 1.955981493721084,
783
+ "grad_norm": 0.2210853099822998,
784
+ "learning_rate": 0.0003138698070003271,
785
+ "loss": 0.9341,
786
+ "mean_token_accuracy": 0.814192325770855,
787
+ "num_tokens": 121217024.0,
788
+ "step": 3700
789
+ },
790
+ {
791
+ "epoch": 1.9824190350297424,
792
+ "grad_norm": 0.1966710090637207,
793
+ "learning_rate": 0.00031114382292007415,
794
+ "loss": 0.9283,
795
+ "step": 3750
796
+ },
797
+ {
798
+ "epoch": 1.9824190350297424,
799
+ "eval_loss": 0.9337447881698608,
800
+ "eval_mean_token_accuracy": 0.8153352634725308,
801
+ "eval_num_tokens": 122855424.0,
802
+ "eval_runtime": 1599.5384,
803
+ "eval_samples_per_second": 4.729,
804
+ "eval_steps_per_second": 0.591,
805
+ "step": 3750
806
+ },
807
+ {
808
+ "epoch": 2.0084600132187704,
809
+ "grad_norm": 0.20168109238147736,
810
+ "learning_rate": 0.00030847235852142626,
811
+ "loss": 0.8927,
812
+ "mean_token_accuracy": 0.8173428005175266,
813
+ "num_tokens": 124469248.0,
814
+ "step": 3800
815
+ },
816
+ {
817
+ "epoch": 2.034897554527429,
818
+ "grad_norm": 0.1905670017004013,
819
+ "learning_rate": 0.00030574637444117324,
820
+ "loss": 0.8427,
821
+ "mean_token_accuracy": 0.8238172018527985,
822
+ "num_tokens": 126107648.0,
823
+ "step": 3850
824
+ },
825
+ {
826
+ "epoch": 2.061335095836087,
827
+ "grad_norm": 0.19004780054092407,
828
+ "learning_rate": 0.0003030203903609203,
829
+ "loss": 0.8536,
830
+ "mean_token_accuracy": 0.8214613863825798,
831
+ "num_tokens": 127746048.0,
832
+ "step": 3900
833
+ },
834
+ {
835
+ "epoch": 2.0877726371447456,
836
+ "grad_norm": 0.21092021465301514,
837
+ "learning_rate": 0.00030029440628066736,
838
+ "loss": 0.8347,
839
+ "mean_token_accuracy": 0.8248136582970619,
840
+ "num_tokens": 129384448.0,
841
+ "step": 3950
842
+ },
843
+ {
844
+ "epoch": 2.114210178453404,
845
+ "grad_norm": 0.2002408355474472,
846
+ "learning_rate": 0.00029756842220041434,
847
+ "loss": 0.8409,
848
+ "step": 4000
849
+ },
850
+ {
851
+ "epoch": 2.114210178453404,
852
+ "eval_loss": 0.913899838924408,
853
+ "eval_mean_token_accuracy": 0.8182373826271888,
854
+ "eval_num_tokens": 131022848.0,
855
+ "eval_runtime": 1599.607,
856
+ "eval_samples_per_second": 4.729,
857
+ "eval_steps_per_second": 0.591,
858
+ "step": 4000
859
+ },
860
+ {
861
+ "epoch": 2.140647719762062,
862
+ "grad_norm": 0.22307777404785156,
863
+ "learning_rate": 0.0002948424381201614,
864
+ "loss": 0.8428,
865
+ "mean_token_accuracy": 0.8242137080430985,
866
+ "num_tokens": 132661248.0,
867
+ "step": 4050
868
+ },
869
+ {
870
+ "epoch": 2.1670852610707203,
871
+ "grad_norm": 0.1873617023229599,
872
+ "learning_rate": 0.0002921164540399084,
873
+ "loss": 0.8439,
874
+ "mean_token_accuracy": 0.8233803743124009,
875
+ "num_tokens": 134299648.0,
876
+ "step": 4100
877
+ },
878
+ {
879
+ "epoch": 2.1935228023793787,
880
+ "grad_norm": 0.1888233870267868,
881
+ "learning_rate": 0.00028939046995965544,
882
+ "loss": 0.8406,
883
+ "mean_token_accuracy": 0.8241153433918953,
884
+ "num_tokens": 135938048.0,
885
+ "step": 4150
886
+ },
887
+ {
888
+ "epoch": 2.219960343688037,
889
+ "grad_norm": 0.19996315240859985,
890
+ "learning_rate": 0.00028666448587940247,
891
+ "loss": 0.8337,
892
+ "mean_token_accuracy": 0.8248002156615257,
893
+ "num_tokens": 137576448.0,
894
+ "step": 4200
895
+ },
896
+ {
897
+ "epoch": 2.2463978849966955,
898
+ "grad_norm": 0.21117758750915527,
899
+ "learning_rate": 0.0002839385017991495,
900
+ "loss": 0.8411,
901
+ "step": 4250
902
+ },
903
+ {
904
+ "epoch": 2.2463978849966955,
905
+ "eval_loss": 0.893865704536438,
906
+ "eval_mean_token_accuracy": 0.8210062818093733,
907
+ "eval_num_tokens": 139214848.0,
908
+ "eval_runtime": 1599.858,
909
+ "eval_samples_per_second": 4.729,
910
+ "eval_steps_per_second": 0.591,
911
+ "step": 4250
912
+ },
913
+ {
914
+ "epoch": 2.2728354263053534,
915
+ "grad_norm": 0.20331983268260956,
916
+ "learning_rate": 0.00028121251771889654,
917
+ "loss": 0.8389,
918
+ "mean_token_accuracy": 0.824597994685173,
919
+ "num_tokens": 140853248.0,
920
+ "step": 4300
921
+ },
922
+ {
923
+ "epoch": 2.299272967614012,
924
+ "grad_norm": 0.19736993312835693,
925
+ "learning_rate": 0.00027848653363864357,
926
+ "loss": 0.8168,
927
+ "mean_token_accuracy": 0.8279356023669243,
928
+ "num_tokens": 142491648.0,
929
+ "step": 4350
930
+ },
931
+ {
932
+ "epoch": 2.32571050892267,
933
+ "grad_norm": 0.1942383050918579,
934
+ "learning_rate": 0.0002757605495583906,
935
+ "loss": 0.8158,
936
+ "mean_token_accuracy": 0.8288569149374961,
937
+ "num_tokens": 144130048.0,
938
+ "step": 4400
939
+ },
940
+ {
941
+ "epoch": 2.3521480502313286,
942
+ "grad_norm": 0.18327121436595917,
943
+ "learning_rate": 0.0002730345654781376,
944
+ "loss": 0.8097,
945
+ "mean_token_accuracy": 0.8300702553987503,
946
+ "num_tokens": 145768448.0,
947
+ "step": 4450
948
+ },
949
+ {
950
+ "epoch": 2.378585591539987,
951
+ "grad_norm": 0.17920152842998505,
952
+ "learning_rate": 0.00027030858139788467,
953
+ "loss": 0.8017,
954
+ "step": 4500
955
+ },
956
+ {
957
+ "epoch": 2.378585591539987,
958
+ "eval_loss": 0.874257504940033,
959
+ "eval_mean_token_accuracy": 0.823767999828996,
960
+ "eval_num_tokens": 147406848.0,
961
+ "eval_runtime": 1599.5025,
962
+ "eval_samples_per_second": 4.73,
963
+ "eval_steps_per_second": 0.591,
964
+ "step": 4500
965
+ },
966
+ {
967
+ "epoch": 2.405023132848645,
968
+ "grad_norm": 0.18811027705669403,
969
+ "learning_rate": 0.0002675825973176317,
970
+ "loss": 0.8215,
971
+ "mean_token_accuracy": 0.8293267333507538,
972
+ "num_tokens": 149045248.0,
973
+ "step": 4550
974
+ },
975
+ {
976
+ "epoch": 2.4314606741573033,
977
+ "grad_norm": 0.20340368151664734,
978
+ "learning_rate": 0.0002648566132373787,
979
+ "loss": 0.8249,
980
+ "mean_token_accuracy": 0.8268548348546028,
981
+ "num_tokens": 150683648.0,
982
+ "step": 4600
983
+ },
984
+ {
985
+ "epoch": 2.4578982154659617,
986
+ "grad_norm": 0.18492697179317474,
987
+ "learning_rate": 0.0002621306291571257,
988
+ "loss": 0.7914,
989
+ "mean_token_accuracy": 0.832571476995945,
990
+ "num_tokens": 152322048.0,
991
+ "step": 4650
992
+ },
993
+ {
994
+ "epoch": 2.48433575677462,
995
+ "grad_norm": 0.19855117797851562,
996
+ "learning_rate": 0.0002594046450768728,
997
+ "loss": 0.8077,
998
+ "mean_token_accuracy": 0.8298674210906029,
999
+ "num_tokens": 153960448.0,
1000
+ "step": 4700
1001
+ },
1002
+ {
1003
+ "epoch": 2.5107732980832784,
1004
+ "grad_norm": 0.1997339129447937,
1005
+ "learning_rate": 0.0002566786609966198,
1006
+ "loss": 0.809,
1007
+ "step": 4750
1008
+ },
1009
+ {
1010
+ "epoch": 2.5107732980832784,
1011
+ "eval_loss": 0.8553281426429749,
1012
+ "eval_mean_token_accuracy": 0.8265610535729511,
1013
+ "eval_num_tokens": 155598848.0,
1014
+ "eval_runtime": 1599.9059,
1015
+ "eval_samples_per_second": 4.728,
1016
+ "eval_steps_per_second": 0.591,
1017
+ "step": 4750
1018
+ },
1019
+ {
1020
+ "epoch": 2.5372108393919364,
1021
+ "grad_norm": 0.19008329510688782,
1022
+ "learning_rate": 0.0002539526769163668,
1023
+ "loss": 0.797,
1024
+ "mean_token_accuracy": 0.8298222103714943,
1025
+ "num_tokens": 157237248.0,
1026
+ "step": 4800
1027
+ },
1028
+ {
1029
+ "epoch": 2.5636483807005948,
1030
+ "grad_norm": 0.18476171791553497,
1031
+ "learning_rate": 0.00025122669283611385,
1032
+ "loss": 0.7987,
1033
+ "mean_token_accuracy": 0.8304337722063064,
1034
+ "num_tokens": 158875648.0,
1035
+ "step": 4850
1036
+ },
1037
+ {
1038
+ "epoch": 2.590085922009253,
1039
+ "grad_norm": 0.18693213164806366,
1040
+ "learning_rate": 0.0002485007087558609,
1041
+ "loss": 0.8042,
1042
+ "mean_token_accuracy": 0.8297446221113205,
1043
+ "num_tokens": 160514048.0,
1044
+ "step": 4900
1045
+ },
1046
+ {
1047
+ "epoch": 2.6165234633179115,
1048
+ "grad_norm": 0.19470660388469696,
1049
+ "learning_rate": 0.0002457747246756079,
1050
+ "loss": 0.8024,
1051
+ "mean_token_accuracy": 0.8308174461126328,
1052
+ "num_tokens": 162152448.0,
1053
+ "step": 4950
1054
+ },
1055
+ {
1056
+ "epoch": 2.64296100462657,
1057
+ "grad_norm": 0.23168876767158508,
1058
+ "learning_rate": 0.00024304874059535492,
1059
+ "loss": 0.7903,
1060
+ "step": 5000
1061
+ },
1062
+ {
1063
+ "epoch": 2.64296100462657,
1064
+ "eval_loss": 0.8376234769821167,
1065
+ "eval_mean_token_accuracy": 0.828871109394896,
1066
+ "eval_num_tokens": 163790848.0,
1067
+ "eval_runtime": 1600.0988,
1068
+ "eval_samples_per_second": 4.728,
1069
+ "eval_steps_per_second": 0.591,
1070
+ "step": 5000
1071
+ },
1072
+ {
1073
+ "epoch": 2.669398545935228,
1074
+ "grad_norm": 0.15908803045749664,
1075
+ "learning_rate": 0.00024032275651510195,
1076
+ "loss": 0.7967,
1077
+ "mean_token_accuracy": 0.8314005956053734,
1078
+ "num_tokens": 165429248.0,
1079
+ "step": 5050
1080
+ },
1081
+ {
1082
+ "epoch": 2.6958360872438862,
1083
+ "grad_norm": 0.1805862933397293,
1084
+ "learning_rate": 0.000237596772434849,
1085
+ "loss": 0.7774,
1086
+ "mean_token_accuracy": 0.8344085997343064,
1087
+ "num_tokens": 167067648.0,
1088
+ "step": 5100
1089
+ },
1090
+ {
1091
+ "epoch": 2.7222736285525446,
1092
+ "grad_norm": 0.17997150123119354,
1093
+ "learning_rate": 0.00023487078835459602,
1094
+ "loss": 0.7851,
1095
+ "mean_token_accuracy": 0.8325213807821273,
1096
+ "num_tokens": 168706048.0,
1097
+ "step": 5150
1098
+ },
1099
+ {
1100
+ "epoch": 2.748711169861203,
1101
+ "grad_norm": 0.18113110959529877,
1102
+ "learning_rate": 0.00023214480427434303,
1103
+ "loss": 0.776,
1104
+ "mean_token_accuracy": 0.8346639758348465,
1105
+ "num_tokens": 170344448.0,
1106
+ "step": 5200
1107
+ },
1108
+ {
1109
+ "epoch": 2.7751487111698614,
1110
+ "grad_norm": 0.18302254378795624,
1111
+ "learning_rate": 0.00022941882019409009,
1112
+ "loss": 0.7854,
1113
+ "step": 5250
1114
+ },
1115
+ {
1116
+ "epoch": 2.7751487111698614,
1117
+ "eval_loss": 0.8233165144920349,
1118
+ "eval_mean_token_accuracy": 0.830954508725987,
1119
+ "eval_num_tokens": 171982848.0,
1120
+ "eval_runtime": 1599.9718,
1121
+ "eval_samples_per_second": 4.728,
1122
+ "eval_steps_per_second": 0.591,
1123
+ "step": 5250
1124
+ },
1125
+ {
1126
+ "epoch": 2.8015862524785193,
1127
+ "grad_norm": 0.1922728568315506,
1128
+ "learning_rate": 0.0002266928361138371,
1129
+ "loss": 0.7936,
1130
+ "mean_token_accuracy": 0.8322769993543625,
1131
+ "num_tokens": 173621248.0,
1132
+ "step": 5300
1133
+ },
1134
+ {
1135
+ "epoch": 2.8280237937871777,
1136
+ "grad_norm": 0.1617008000612259,
1137
+ "learning_rate": 0.00022396685203358413,
1138
+ "loss": 0.7738,
1139
+ "mean_token_accuracy": 0.8344037118554115,
1140
+ "num_tokens": 175259648.0,
1141
+ "step": 5350
1142
+ },
1143
+ {
1144
+ "epoch": 2.854461335095836,
1145
+ "grad_norm": 0.17171062529087067,
1146
+ "learning_rate": 0.00022124086795333116,
1147
+ "loss": 0.7697,
1148
+ "mean_token_accuracy": 0.8351166906952858,
1149
+ "num_tokens": 176898048.0,
1150
+ "step": 5400
1151
+ },
1152
+ {
1153
+ "epoch": 2.8808988764044945,
1154
+ "grad_norm": 0.1803775280714035,
1155
+ "learning_rate": 0.0002185148838730782,
1156
+ "loss": 0.7735,
1157
+ "mean_token_accuracy": 0.8350091609358787,
1158
+ "num_tokens": 178536448.0,
1159
+ "step": 5450
1160
+ },
1161
+ {
1162
+ "epoch": 2.907336417713153,
1163
+ "grad_norm": 0.17305733263492584,
1164
+ "learning_rate": 0.0002157888997928252,
1165
+ "loss": 0.7716,
1166
+ "step": 5500
1167
+ },
1168
+ {
1169
+ "epoch": 2.907336417713153,
1170
+ "eval_loss": 0.8076795339584351,
1171
+ "eval_mean_token_accuracy": 0.8331229730841977,
1172
+ "eval_num_tokens": 180174848.0,
1173
+ "eval_runtime": 1600.6859,
1174
+ "eval_samples_per_second": 4.726,
1175
+ "eval_steps_per_second": 0.591,
1176
+ "step": 5500
1177
+ },
1178
+ {
1179
+ "epoch": 2.933773959021811,
1180
+ "grad_norm": 0.17064611613750458,
1181
+ "learning_rate": 0.00021306291571257226,
1182
+ "loss": 0.7713,
1183
+ "mean_token_accuracy": 0.8356136959791184,
1184
+ "num_tokens": 181813248.0,
1185
+ "step": 5550
1186
+ },
1187
+ {
1188
+ "epoch": 2.960211500330469,
1189
+ "grad_norm": 0.18137440085411072,
1190
+ "learning_rate": 0.00021033693163231926,
1191
+ "loss": 0.7667,
1192
+ "mean_token_accuracy": 0.8351374611258506,
1193
+ "num_tokens": 183451648.0,
1194
+ "step": 5600
1195
+ },
1196
+ {
1197
+ "epoch": 2.9866490416391276,
1198
+ "grad_norm": 0.17405763268470764,
1199
+ "learning_rate": 0.0002076109475520663,
1200
+ "loss": 0.7495,
1201
+ "mean_token_accuracy": 0.8385416662693024,
1202
+ "num_tokens": 185090048.0,
1203
+ "step": 5650
1204
+ },
1205
+ {
1206
+ "epoch": 3.012690019828156,
1207
+ "grad_norm": 0.17279721796512604,
1208
+ "learning_rate": 0.0002049394831534184,
1209
+ "loss": 0.7159,
1210
+ "mean_token_accuracy": 0.8417613173499325,
1211
+ "num_tokens": 186703872.0,
1212
+ "step": 5700
1213
+ },
1214
+ {
1215
+ "epoch": 3.0391275611368145,
1216
+ "grad_norm": 0.19387085735797882,
1217
+ "learning_rate": 0.0002022134990731654,
1218
+ "loss": 0.666,
1219
+ "step": 5750
1220
+ },
1221
+ {
1222
+ "epoch": 3.0391275611368145,
1223
+ "eval_loss": 0.8053749799728394,
1224
+ "eval_mean_token_accuracy": 0.8340540434416959,
1225
+ "eval_num_tokens": 188342272.0,
1226
+ "eval_runtime": 1600.205,
1227
+ "eval_samples_per_second": 4.728,
1228
+ "eval_steps_per_second": 0.591,
1229
+ "step": 5750
1230
+ },
1231
+ {
1232
+ "epoch": 3.0655651024454724,
1233
+ "grad_norm": 0.18193645775318146,
1234
+ "learning_rate": 0.00019948751499291245,
1235
+ "loss": 0.6644,
1236
+ "mean_token_accuracy": 0.8480684906244278,
1237
+ "num_tokens": 189980672.0,
1238
+ "step": 5800
1239
+ },
1240
+ {
1241
+ "epoch": 3.092002643754131,
1242
+ "grad_norm": 0.16633963584899902,
1243
+ "learning_rate": 0.00019676153091265948,
1244
+ "loss": 0.6691,
1245
+ "mean_token_accuracy": 0.847120603621006,
1246
+ "num_tokens": 191619072.0,
1247
+ "step": 5850
1248
+ },
1249
+ {
1250
+ "epoch": 3.118440185062789,
1251
+ "grad_norm": 0.17585037648677826,
1252
+ "learning_rate": 0.0001940355468324065,
1253
+ "loss": 0.6636,
1254
+ "mean_token_accuracy": 0.84809934258461,
1255
+ "num_tokens": 193257472.0,
1256
+ "step": 5900
1257
+ },
1258
+ {
1259
+ "epoch": 3.1448777263714476,
1260
+ "grad_norm": 0.1676415503025055,
1261
+ "learning_rate": 0.00019130956275215352,
1262
+ "loss": 0.6672,
1263
+ "mean_token_accuracy": 0.8475995865464211,
1264
+ "num_tokens": 194895872.0,
1265
+ "step": 5950
1266
+ },
1267
+ {
1268
+ "epoch": 3.1713152676801055,
1269
+ "grad_norm": 0.18070462346076965,
1270
+ "learning_rate": 0.00018858357867190058,
1271
+ "loss": 0.6627,
1272
+ "step": 6000
1273
+ },
1274
+ {
1275
+ "epoch": 3.1713152676801055,
1276
+ "eval_loss": 0.801948070526123,
1277
+ "eval_mean_token_accuracy": 0.8345137661909704,
1278
+ "eval_num_tokens": 196534272.0,
1279
+ "eval_runtime": 1599.8066,
1280
+ "eval_samples_per_second": 4.729,
1281
+ "eval_steps_per_second": 0.591,
1282
+ "step": 6000
1283
+ },
1284
+ {
1285
+ "epoch": 3.197752808988764,
1286
+ "grad_norm": 0.16841137409210205,
1287
+ "learning_rate": 0.00018585759459164758,
1288
+ "loss": 0.6569,
1289
+ "mean_token_accuracy": 0.8492378443479538,
1290
+ "num_tokens": 198172672.0,
1291
+ "step": 6050
1292
+ },
1293
+ {
1294
+ "epoch": 3.2241903502974223,
1295
+ "grad_norm": 0.18084491789340973,
1296
+ "learning_rate": 0.00018313161051139462,
1297
+ "loss": 0.6678,
1298
+ "mean_token_accuracy": 0.8477779817581177,
1299
+ "num_tokens": 199811072.0,
1300
+ "step": 6100
1301
+ },
1302
+ {
1303
+ "epoch": 3.2506278916060807,
1304
+ "grad_norm": 0.17532089352607727,
1305
+ "learning_rate": 0.00018040562643114165,
1306
+ "loss": 0.6693,
1307
+ "mean_token_accuracy": 0.8475476580858231,
1308
+ "num_tokens": 201449472.0,
1309
+ "step": 6150
1310
+ },
1311
+ {
1312
+ "epoch": 3.277065432914739,
1313
+ "grad_norm": 0.17762629687786102,
1314
+ "learning_rate": 0.00017767964235088868,
1315
+ "loss": 0.6568,
1316
+ "mean_token_accuracy": 0.8500018376111984,
1317
+ "num_tokens": 203087872.0,
1318
+ "step": 6200
1319
+ },
1320
+ {
1321
+ "epoch": 3.303502974223397,
1322
+ "grad_norm": 0.17803572118282318,
1323
+ "learning_rate": 0.0001749536582706357,
1324
+ "loss": 0.6664,
1325
+ "step": 6250
1326
+ },
1327
+ {
1328
+ "epoch": 3.303502974223397,
1329
+ "eval_loss": 0.7924287915229797,
1330
+ "eval_mean_token_accuracy": 0.8360363316838384,
1331
+ "eval_num_tokens": 204726272.0,
1332
+ "eval_runtime": 1600.5314,
1333
+ "eval_samples_per_second": 4.727,
1334
+ "eval_steps_per_second": 0.591,
1335
+ "step": 6250
1336
+ },
1337
+ {
1338
+ "epoch": 3.3299405155320554,
1339
+ "grad_norm": 0.1736496537923813,
1340
+ "learning_rate": 0.00017222767419038275,
1341
+ "loss": 0.6626,
1342
+ "mean_token_accuracy": 0.8480022014677524,
1343
+ "num_tokens": 206364672.0,
1344
+ "step": 6300
1345
+ },
1346
+ {
1347
+ "epoch": 3.3563780568407138,
1348
+ "grad_norm": 0.1790972799062729,
1349
+ "learning_rate": 0.00016950169011012976,
1350
+ "loss": 0.666,
1351
+ "mean_token_accuracy": 0.8478036442399025,
1352
+ "num_tokens": 208003072.0,
1353
+ "step": 6350
1354
+ },
1355
+ {
1356
+ "epoch": 3.382815598149372,
1357
+ "grad_norm": 0.17161910235881805,
1358
+ "learning_rate": 0.0001667757060298768,
1359
+ "loss": 0.6635,
1360
+ "mean_token_accuracy": 0.8481677681207657,
1361
+ "num_tokens": 209641472.0,
1362
+ "step": 6400
1363
+ },
1364
+ {
1365
+ "epoch": 3.4092531394580305,
1366
+ "grad_norm": 0.17608526349067688,
1367
+ "learning_rate": 0.00016404972194962382,
1368
+ "loss": 0.6483,
1369
+ "mean_token_accuracy": 0.8513996881246567,
1370
+ "num_tokens": 211279872.0,
1371
+ "step": 6450
1372
+ },
1373
+ {
1374
+ "epoch": 3.4356906807666885,
1375
+ "grad_norm": 0.17622597515583038,
1376
+ "learning_rate": 0.00016132373786937086,
1377
+ "loss": 0.6562,
1378
+ "step": 6500
1379
+ },
1380
+ {
1381
+ "epoch": 3.4356906807666885,
1382
+ "eval_loss": 0.7829640507698059,
1383
+ "eval_mean_token_accuracy": 0.8375042610768284,
1384
+ "eval_num_tokens": 212918272.0,
1385
+ "eval_runtime": 1600.7277,
1386
+ "eval_samples_per_second": 4.726,
1387
+ "eval_steps_per_second": 0.591,
1388
+ "step": 6500
1389
+ },
1390
+ {
1391
+ "epoch": 3.462128222075347,
1392
+ "grad_norm": 0.18006405234336853,
1393
+ "learning_rate": 0.00015859775378911786,
1394
+ "loss": 0.6498,
1395
+ "mean_token_accuracy": 0.8504380528628827,
1396
+ "num_tokens": 214556672.0,
1397
+ "step": 6550
1398
+ },
1399
+ {
1400
+ "epoch": 3.4885657633840053,
1401
+ "grad_norm": 0.16343793272972107,
1402
+ "learning_rate": 0.0001558717697088649,
1403
+ "loss": 0.6519,
1404
+ "mean_token_accuracy": 0.850884655714035,
1405
+ "num_tokens": 216195072.0,
1406
+ "step": 6600
1407
+ },
1408
+ {
1409
+ "epoch": 3.5150033046926636,
1410
+ "grad_norm": 0.16798467934131622,
1411
+ "learning_rate": 0.00015314578562861193,
1412
+ "loss": 0.6648,
1413
+ "mean_token_accuracy": 0.8490127098560333,
1414
+ "num_tokens": 217833472.0,
1415
+ "step": 6650
1416
+ },
1417
+ {
1418
+ "epoch": 3.541440846001322,
1419
+ "grad_norm": 0.15794213116168976,
1420
+ "learning_rate": 0.00015041980154835896,
1421
+ "loss": 0.6471,
1422
+ "mean_token_accuracy": 0.8517173796892166,
1423
+ "num_tokens": 219471872.0,
1424
+ "step": 6700
1425
+ },
1426
+ {
1427
+ "epoch": 3.56787838730998,
1428
+ "grad_norm": 0.1636921763420105,
1429
+ "learning_rate": 0.00014769381746810597,
1430
+ "loss": 0.6424,
1431
+ "step": 6750
1432
+ },
1433
+ {
1434
+ "epoch": 3.56787838730998,
1435
+ "eval_loss": 0.773522138595581,
1436
+ "eval_mean_token_accuracy": 0.8390711046928583,
1437
+ "eval_num_tokens": 221110272.0,
1438
+ "eval_runtime": 1600.7216,
1439
+ "eval_samples_per_second": 4.726,
1440
+ "eval_steps_per_second": 0.591,
1441
+ "step": 6750
1442
+ },
1443
+ {
1444
+ "epoch": 3.5943159286186384,
1445
+ "grad_norm": 0.15980064868927002,
1446
+ "learning_rate": 0.00014496783338785303,
1447
+ "loss": 0.6571,
1448
+ "mean_token_accuracy": 0.851312015503645,
1449
+ "num_tokens": 222748672.0,
1450
+ "step": 6800
1451
+ },
1452
+ {
1453
+ "epoch": 3.6207534699272967,
1454
+ "grad_norm": 0.1708955615758896,
1455
+ "learning_rate": 0.00014224184930760003,
1456
+ "loss": 0.6484,
1457
+ "mean_token_accuracy": 0.8513654717803001,
1458
+ "num_tokens": 224387072.0,
1459
+ "step": 6850
1460
+ },
1461
+ {
1462
+ "epoch": 3.647191011235955,
1463
+ "grad_norm": 0.16906002163887024,
1464
+ "learning_rate": 0.00013951586522734707,
1465
+ "loss": 0.6517,
1466
+ "mean_token_accuracy": 0.8500537672638893,
1467
+ "num_tokens": 226025472.0,
1468
+ "step": 6900
1469
+ },
1470
+ {
1471
+ "epoch": 3.6736285525446135,
1472
+ "grad_norm": 0.16365185379981995,
1473
+ "learning_rate": 0.0001367898811470941,
1474
+ "loss": 0.6372,
1475
+ "mean_token_accuracy": 0.8536284250020981,
1476
+ "num_tokens": 227663872.0,
1477
+ "step": 6950
1478
+ },
1479
+ {
1480
+ "epoch": 3.7000660938532715,
1481
+ "grad_norm": 0.17780087888240814,
1482
+ "learning_rate": 0.00013406389706684113,
1483
+ "loss": 0.6501,
1484
+ "step": 7000
1485
+ },
1486
+ {
1487
+ "epoch": 3.7000660938532715,
1488
+ "eval_loss": 0.7657620906829834,
1489
+ "eval_mean_token_accuracy": 0.8401046276848614,
1490
+ "eval_num_tokens": 229302272.0,
1491
+ "eval_runtime": 1600.1467,
1492
+ "eval_samples_per_second": 4.728,
1493
+ "eval_steps_per_second": 0.591,
1494
+ "step": 7000
1495
+ },
1496
+ {
1497
+ "epoch": 3.72650363516193,
1498
+ "grad_norm": 0.17722897231578827,
1499
+ "learning_rate": 0.00013133791298658814,
1500
+ "loss": 0.6527,
1501
+ "mean_token_accuracy": 0.8508571648597717,
1502
+ "num_tokens": 230940672.0,
1503
+ "step": 7050
1504
+ },
1505
+ {
1506
+ "epoch": 3.7529411764705882,
1507
+ "grad_norm": 0.16244906187057495,
1508
+ "learning_rate": 0.0001286119289063352,
1509
+ "loss": 0.6356,
1510
+ "mean_token_accuracy": 0.8537634432315826,
1511
+ "num_tokens": 232579072.0,
1512
+ "step": 7100
1513
+ },
1514
+ {
1515
+ "epoch": 3.7793787177792466,
1516
+ "grad_norm": 0.15864387154579163,
1517
+ "learning_rate": 0.0001258859448260822,
1518
+ "loss": 0.6452,
1519
+ "mean_token_accuracy": 0.8518102434277535,
1520
+ "num_tokens": 234217472.0,
1521
+ "step": 7150
1522
+ },
1523
+ {
1524
+ "epoch": 3.805816259087905,
1525
+ "grad_norm": 0.16620229184627533,
1526
+ "learning_rate": 0.00012315996074582924,
1527
+ "loss": 0.6418,
1528
+ "mean_token_accuracy": 0.8521817001700401,
1529
+ "num_tokens": 235855872.0,
1530
+ "step": 7200
1531
+ },
1532
+ {
1533
+ "epoch": 3.832253800396563,
1534
+ "grad_norm": 0.1765565574169159,
1535
+ "learning_rate": 0.00012043397666557627,
1536
+ "loss": 0.6387,
1537
+ "step": 7250
1538
+ },
1539
+ {
1540
+ "epoch": 3.832253800396563,
1541
+ "eval_loss": 0.7578161358833313,
1542
+ "eval_mean_token_accuracy": 0.8413670561404359,
1543
+ "eval_num_tokens": 237494272.0,
1544
+ "eval_runtime": 1602.8152,
1545
+ "eval_samples_per_second": 4.72,
1546
+ "eval_steps_per_second": 0.59,
1547
+ "step": 7250
1548
+ },
1549
+ {
1550
+ "epoch": 3.8586913417052213,
1551
+ "grad_norm": 0.15968503057956696,
1552
+ "learning_rate": 0.0001177079925853233,
1553
+ "loss": 0.6365,
1554
+ "mean_token_accuracy": 0.8533528861403465,
1555
+ "num_tokens": 239132672.0,
1556
+ "step": 7300
1557
+ },
1558
+ {
1559
+ "epoch": 3.8851288830138797,
1560
+ "grad_norm": 0.15743543207645416,
1561
+ "learning_rate": 0.00011498200850507034,
1562
+ "loss": 0.6486,
1563
+ "mean_token_accuracy": 0.8513813573122024,
1564
+ "num_tokens": 240771072.0,
1565
+ "step": 7350
1566
+ },
1567
+ {
1568
+ "epoch": 3.911566424322538,
1569
+ "grad_norm": 0.18122394382953644,
1570
+ "learning_rate": 0.00011225602442481736,
1571
+ "loss": 0.6384,
1572
+ "mean_token_accuracy": 0.8533547213673591,
1573
+ "num_tokens": 242409472.0,
1574
+ "step": 7400
1575
+ },
1576
+ {
1577
+ "epoch": 3.9380039656311965,
1578
+ "grad_norm": 0.15892641246318817,
1579
+ "learning_rate": 0.00010953004034456439,
1580
+ "loss": 0.6338,
1581
+ "mean_token_accuracy": 0.8538844108581543,
1582
+ "num_tokens": 244047872.0,
1583
+ "step": 7450
1584
+ },
1585
+ {
1586
+ "epoch": 3.9644415069398544,
1587
+ "grad_norm": 0.16563069820404053,
1588
+ "learning_rate": 0.00010680405626431142,
1589
+ "loss": 0.6256,
1590
+ "step": 7500
1591
+ },
1592
+ {
1593
+ "epoch": 3.9644415069398544,
1594
+ "eval_loss": 0.7491397857666016,
1595
+ "eval_mean_token_accuracy": 0.8425506683535102,
1596
+ "eval_num_tokens": 245686272.0,
1597
+ "eval_runtime": 1603.9187,
1598
+ "eval_samples_per_second": 4.717,
1599
+ "eval_steps_per_second": 0.59,
1600
+ "step": 7500
1601
+ },
1602
+ {
1603
+ "epoch": 3.990879048248513,
1604
+ "grad_norm": 0.1561686098575592,
1605
+ "learning_rate": 0.00010407807218405844,
1606
+ "loss": 0.6398,
1607
+ "mean_token_accuracy": 0.8541101579368114,
1608
+ "num_tokens": 247324672.0,
1609
+ "step": 7550
1610
+ },
1611
+ {
1612
+ "epoch": 4.016920026437541,
1613
+ "grad_norm": 0.17185606062412262,
1614
+ "learning_rate": 0.00010135208810380548,
1615
+ "loss": 0.5504,
1616
+ "mean_token_accuracy": 0.8682107241625713,
1617
+ "num_tokens": 248938496.0,
1618
+ "step": 7600
1619
+ },
1620
+ {
1621
+ "epoch": 4.0433575677462,
1622
+ "grad_norm": 0.17470529675483704,
1623
+ "learning_rate": 9.86261040235525e-05,
1624
+ "loss": 0.5029,
1625
+ "mean_token_accuracy": 0.8748790314793586,
1626
+ "num_tokens": 250576896.0,
1627
+ "step": 7650
1628
+ },
1629
+ {
1630
+ "epoch": 4.069795109054858,
1631
+ "grad_norm": 0.1801612824201584,
1632
+ "learning_rate": 9.590011994329953e-05,
1633
+ "loss": 0.5043,
1634
+ "mean_token_accuracy": 0.8748985821008682,
1635
+ "num_tokens": 252215296.0,
1636
+ "step": 7700
1637
+ },
1638
+ {
1639
+ "epoch": 4.0962326503635165,
1640
+ "grad_norm": 0.16825653612613678,
1641
+ "learning_rate": 9.317413586304656e-05,
1642
+ "loss": 0.4967,
1643
+ "step": 7750
1644
+ },
1645
+ {
1646
+ "epoch": 4.0962326503635165,
1647
+ "eval_loss": 0.7883051037788391,
1648
+ "eval_mean_token_accuracy": 0.8408105385983973,
1649
+ "eval_num_tokens": 253853696.0,
1650
+ "eval_runtime": 1607.5856,
1651
+ "eval_samples_per_second": 4.706,
1652
+ "eval_steps_per_second": 0.588,
1653
+ "step": 7750
1654
+ },
1655
+ {
1656
+ "epoch": 4.122670191672174,
1657
+ "grad_norm": 0.17985741794109344,
1658
+ "learning_rate": 9.044815178279358e-05,
1659
+ "loss": 0.5031,
1660
+ "mean_token_accuracy": 0.875472262352705,
1661
+ "num_tokens": 255492096.0,
1662
+ "step": 7800
1663
+ },
1664
+ {
1665
+ "epoch": 4.149107732980832,
1666
+ "grad_norm": 0.17613214254379272,
1667
+ "learning_rate": 8.772216770254061e-05,
1668
+ "loss": 0.4969,
1669
+ "mean_token_accuracy": 0.8762671053409576,
1670
+ "num_tokens": 257130496.0,
1671
+ "step": 7850
1672
+ },
1673
+ {
1674
+ "epoch": 4.175545274289491,
1675
+ "grad_norm": 0.17405198514461517,
1676
+ "learning_rate": 8.499618362228765e-05,
1677
+ "loss": 0.5095,
1678
+ "mean_token_accuracy": 0.8734744620323182,
1679
+ "num_tokens": 258768896.0,
1680
+ "step": 7900
1681
+ },
1682
+ {
1683
+ "epoch": 4.201982815598149,
1684
+ "grad_norm": 0.17185764014720917,
1685
+ "learning_rate": 8.227019954203467e-05,
1686
+ "loss": 0.5074,
1687
+ "mean_token_accuracy": 0.8739729967713356,
1688
+ "num_tokens": 260407296.0,
1689
+ "step": 7950
1690
+ },
1691
+ {
1692
+ "epoch": 4.228420356906808,
1693
+ "grad_norm": 0.17758677899837494,
1694
+ "learning_rate": 7.95442154617817e-05,
1695
+ "loss": 0.5085,
1696
+ "step": 8000
1697
+ },
1698
+ {
1699
+ "epoch": 4.228420356906808,
1700
+ "eval_loss": 0.7870664000511169,
1701
+ "eval_mean_token_accuracy": 0.8414596145929292,
1702
+ "eval_num_tokens": 262045696.0,
1703
+ "eval_runtime": 1607.3849,
1704
+ "eval_samples_per_second": 4.706,
1705
+ "eval_steps_per_second": 0.589,
1706
+ "step": 8000
1707
+ },
1708
+ {
1709
+ "epoch": 4.254857898215466,
1710
+ "grad_norm": 0.16629241406917572,
1711
+ "learning_rate": 7.681823138152873e-05,
1712
+ "loss": 0.5032,
1713
+ "mean_token_accuracy": 0.8741639178991317,
1714
+ "num_tokens": 263684096.0,
1715
+ "step": 8050
1716
+ },
1717
+ {
1718
+ "epoch": 4.281295439524124,
1719
+ "grad_norm": 0.173508420586586,
1720
+ "learning_rate": 7.409224730127575e-05,
1721
+ "loss": 0.4909,
1722
+ "mean_token_accuracy": 0.8775629255175591,
1723
+ "num_tokens": 265322496.0,
1724
+ "step": 8100
1725
+ },
1726
+ {
1727
+ "epoch": 4.307732980832783,
1728
+ "grad_norm": 0.1713671237230301,
1729
+ "learning_rate": 7.136626322102279e-05,
1730
+ "loss": 0.4923,
1731
+ "mean_token_accuracy": 0.8772788345813751,
1732
+ "num_tokens": 266960896.0,
1733
+ "step": 8150
1734
+ },
1735
+ {
1736
+ "epoch": 4.334170522141441,
1737
+ "grad_norm": 0.17122632265090942,
1738
+ "learning_rate": 6.864027914076983e-05,
1739
+ "loss": 0.5,
1740
+ "mean_token_accuracy": 0.8755180832743644,
1741
+ "num_tokens": 268599296.0,
1742
+ "step": 8200
1743
+ },
1744
+ {
1745
+ "epoch": 4.360608063450099,
1746
+ "grad_norm": 0.17359545826911926,
1747
+ "learning_rate": 6.591429506051685e-05,
1748
+ "loss": 0.4943,
1749
+ "step": 8250
1750
+ },
1751
+ {
1752
+ "epoch": 4.360608063450099,
1753
+ "eval_loss": 0.7823996543884277,
1754
+ "eval_mean_token_accuracy": 0.8421699439370355,
1755
+ "eval_num_tokens": 270237696.0,
1756
+ "eval_runtime": 1607.7567,
1757
+ "eval_samples_per_second": 4.705,
1758
+ "eval_steps_per_second": 0.588,
1759
+ "step": 8250
1760
+ },
1761
+ {
1762
+ "epoch": 4.387045604758757,
1763
+ "grad_norm": 0.17702388763427734,
1764
+ "learning_rate": 6.318831098026388e-05,
1765
+ "loss": 0.4904,
1766
+ "mean_token_accuracy": 0.8775449013710022,
1767
+ "num_tokens": 271876096.0,
1768
+ "step": 8300
1769
+ },
1770
+ {
1771
+ "epoch": 4.413483146067415,
1772
+ "grad_norm": 0.18663644790649414,
1773
+ "learning_rate": 6.0462326900010904e-05,
1774
+ "loss": 0.4959,
1775
+ "mean_token_accuracy": 0.8762383911013604,
1776
+ "num_tokens": 273514496.0,
1777
+ "step": 8350
1778
+ },
1779
+ {
1780
+ "epoch": 4.439920687376074,
1781
+ "grad_norm": 0.1880512684583664,
1782
+ "learning_rate": 5.773634281975793e-05,
1783
+ "loss": 0.4931,
1784
+ "mean_token_accuracy": 0.8767839661240577,
1785
+ "num_tokens": 275152896.0,
1786
+ "step": 8400
1787
+ },
1788
+ {
1789
+ "epoch": 4.466358228684732,
1790
+ "grad_norm": 0.18527589738368988,
1791
+ "learning_rate": 5.5010358739504963e-05,
1792
+ "loss": 0.4877,
1793
+ "mean_token_accuracy": 0.87819525629282,
1794
+ "num_tokens": 276791296.0,
1795
+ "step": 8450
1796
+ },
1797
+ {
1798
+ "epoch": 4.492795769993391,
1799
+ "grad_norm": 0.19010977447032928,
1800
+ "learning_rate": 5.228437465925199e-05,
1801
+ "loss": 0.4894,
1802
+ "step": 8500
1803
+ },
1804
+ {
1805
+ "epoch": 4.492795769993391,
1806
+ "eval_loss": 0.7803131341934204,
1807
+ "eval_mean_token_accuracy": 0.8430041650637008,
1808
+ "eval_num_tokens": 278429696.0,
1809
+ "eval_runtime": 1610.9854,
1810
+ "eval_samples_per_second": 4.696,
1811
+ "eval_steps_per_second": 0.587,
1812
+ "step": 8500
1813
+ },
1814
+ {
1815
+ "epoch": 4.519233311302049,
1816
+ "grad_norm": 0.17016442120075226,
1817
+ "learning_rate": 4.9558390578999016e-05,
1818
+ "loss": 0.4847,
1819
+ "mean_token_accuracy": 0.8786284182965756,
1820
+ "num_tokens": 280068096.0,
1821
+ "step": 8550
1822
+ },
1823
+ {
1824
+ "epoch": 4.545670852610707,
1825
+ "grad_norm": 0.1719425618648529,
1826
+ "learning_rate": 4.683240649874604e-05,
1827
+ "loss": 0.4875,
1828
+ "mean_token_accuracy": 0.8785123375058174,
1829
+ "num_tokens": 281706496.0,
1830
+ "step": 8600
1831
+ },
1832
+ {
1833
+ "epoch": 4.572108393919366,
1834
+ "grad_norm": 0.17816464602947235,
1835
+ "learning_rate": 4.4106422418493076e-05,
1836
+ "loss": 0.4863,
1837
+ "mean_token_accuracy": 0.8782337459921837,
1838
+ "num_tokens": 283344896.0,
1839
+ "step": 8650
1840
+ },
1841
+ {
1842
+ "epoch": 4.598545935228024,
1843
+ "grad_norm": 0.1728549599647522,
1844
+ "learning_rate": 4.138043833824011e-05,
1845
+ "loss": 0.4879,
1846
+ "mean_token_accuracy": 0.8787457209825515,
1847
+ "num_tokens": 284983296.0,
1848
+ "step": 8700
1849
+ },
1850
+ {
1851
+ "epoch": 4.624983476536682,
1852
+ "grad_norm": 0.18577666580677032,
1853
+ "learning_rate": 3.8654454257987135e-05,
1854
+ "loss": 0.4914,
1855
+ "step": 8750
1856
+ },
1857
+ {
1858
+ "epoch": 4.624983476536682,
1859
+ "eval_loss": 0.7784421443939209,
1860
+ "eval_mean_token_accuracy": 0.8436387255000262,
1861
+ "eval_num_tokens": 286621696.0,
1862
+ "eval_runtime": 1611.4393,
1863
+ "eval_samples_per_second": 4.695,
1864
+ "eval_steps_per_second": 0.587,
1865
+ "step": 8750
1866
+ },
1867
+ {
1868
+ "epoch": 4.65142101784534,
1869
+ "grad_norm": 0.16825436055660248,
1870
+ "learning_rate": 3.592847017773417e-05,
1871
+ "loss": 0.4756,
1872
+ "mean_token_accuracy": 0.8792506690323353,
1873
+ "num_tokens": 288260096.0,
1874
+ "step": 8800
1875
+ },
1876
+ {
1877
+ "epoch": 4.677858559153998,
1878
+ "grad_norm": 0.18510740995407104,
1879
+ "learning_rate": 3.3202486097481194e-05,
1880
+ "loss": 0.4788,
1881
+ "mean_token_accuracy": 0.8801001918315887,
1882
+ "num_tokens": 289898496.0,
1883
+ "step": 8850
1884
+ },
1885
+ {
1886
+ "epoch": 4.704296100462657,
1887
+ "grad_norm": 0.18907974660396576,
1888
+ "learning_rate": 3.0476502017228217e-05,
1889
+ "loss": 0.4837,
1890
+ "mean_token_accuracy": 0.8794446450471878,
1891
+ "num_tokens": 291536896.0,
1892
+ "step": 8900
1893
+ },
1894
+ {
1895
+ "epoch": 4.730733641771315,
1896
+ "grad_norm": 0.1798245906829834,
1897
+ "learning_rate": 2.775051793697525e-05,
1898
+ "loss": 0.4883,
1899
+ "mean_token_accuracy": 0.8778897827863693,
1900
+ "num_tokens": 293175296.0,
1901
+ "step": 8950
1902
+ },
1903
+ {
1904
+ "epoch": 4.757171183079974,
1905
+ "grad_norm": 0.17980748414993286,
1906
+ "learning_rate": 2.502453385672228e-05,
1907
+ "loss": 0.475,
1908
+ "step": 9000
1909
+ },
1910
+ {
1911
+ "epoch": 4.757171183079974,
1912
+ "eval_loss": 0.7753015756607056,
1913
+ "eval_mean_token_accuracy": 0.8443130426754659,
1914
+ "eval_num_tokens": 294813696.0,
1915
+ "eval_runtime": 1611.452,
1916
+ "eval_samples_per_second": 4.695,
1917
+ "eval_steps_per_second": 0.587,
1918
+ "step": 9000
1919
+ },
1920
+ {
1921
+ "epoch": 4.783608724388632,
1922
+ "grad_norm": 0.17731408774852753,
1923
+ "learning_rate": 2.2298549776469306e-05,
1924
+ "loss": 0.4657,
1925
+ "mean_token_accuracy": 0.8821294555068016,
1926
+ "num_tokens": 296452096.0,
1927
+ "step": 9050
1928
+ },
1929
+ {
1930
+ "epoch": 4.81004626569729,
1931
+ "grad_norm": 0.19258248805999756,
1932
+ "learning_rate": 1.9572565696216336e-05,
1933
+ "loss": 0.4779,
1934
+ "mean_token_accuracy": 0.8807239699363708,
1935
+ "num_tokens": 298090496.0,
1936
+ "step": 9100
1937
+ },
1938
+ {
1939
+ "epoch": 4.836483807005949,
1940
+ "grad_norm": 0.17705880105495453,
1941
+ "learning_rate": 1.6846581615963362e-05,
1942
+ "loss": 0.476,
1943
+ "mean_token_accuracy": 0.8808369943499565,
1944
+ "num_tokens": 299728896.0,
1945
+ "step": 9150
1946
+ },
1947
+ {
1948
+ "epoch": 4.8629213483146065,
1949
+ "grad_norm": 0.1794816255569458,
1950
+ "learning_rate": 1.4120597535710392e-05,
1951
+ "loss": 0.4742,
1952
+ "mean_token_accuracy": 0.8813196429610253,
1953
+ "num_tokens": 301367296.0,
1954
+ "step": 9200
1955
+ },
1956
+ {
1957
+ "epoch": 4.889358889623265,
1958
+ "grad_norm": 0.17823387682437897,
1959
+ "learning_rate": 1.139461345545742e-05,
1960
+ "loss": 0.4719,
1961
+ "step": 9250
1962
+ },
1963
+ {
1964
+ "epoch": 4.889358889623265,
1965
+ "eval_loss": 0.7754274010658264,
1966
+ "eval_mean_token_accuracy": 0.844791491931387,
1967
+ "eval_num_tokens": 303005696.0,
1968
+ "eval_runtime": 1610.8654,
1969
+ "eval_samples_per_second": 4.696,
1970
+ "eval_steps_per_second": 0.587,
1971
+ "step": 9250
1972
+ },
1973
+ {
1974
+ "epoch": 4.915796430931923,
1975
+ "grad_norm": 0.16834519803524017,
1976
+ "learning_rate": 8.668629375204448e-06,
1977
+ "loss": 0.4653,
1978
+ "mean_token_accuracy": 0.8821077673137188,
1979
+ "num_tokens": 304644096.0,
1980
+ "step": 9300
1981
+ },
1982
+ {
1983
+ "epoch": 4.942233972240581,
1984
+ "grad_norm": 0.17272663116455078,
1985
+ "learning_rate": 5.942645294951477e-06,
1986
+ "loss": 0.4783,
1987
+ "mean_token_accuracy": 0.8806390488147735,
1988
+ "num_tokens": 306282496.0,
1989
+ "step": 9350
1990
+ },
1991
+ {
1992
+ "epoch": 4.96867151354924,
1993
+ "grad_norm": 0.17334023118019104,
1994
+ "learning_rate": 3.2166612146985063e-06,
1995
+ "loss": 0.4794,
1996
+ "mean_token_accuracy": 0.8807239702343941,
1997
+ "num_tokens": 307920896.0,
1998
+ "step": 9400
1999
+ },
2000
+ {
2001
+ "epoch": 4.995109054857898,
2002
+ "grad_norm": 0.17255398631095886,
2003
+ "learning_rate": 4.906771344455349e-07,
2004
+ "loss": 0.4793,
2005
+ "mean_token_accuracy": 0.8803439608216286,
2006
+ "num_tokens": 309559296.0,
2007
+ "step": 9450
2008
+ }
2009
+ ],
2010
+ "logging_steps": 50,
2011
+ "max_steps": 9455,
2012
+ "num_input_tokens_seen": 0,
2013
+ "num_train_epochs": 5,
2014
+ "save_steps": 500,
2015
+ "stateful_callbacks": {
2016
+ "TrainerControl": {
2017
+ "args": {
2018
+ "should_epoch_stop": false,
2019
+ "should_evaluate": false,
2020
+ "should_log": false,
2021
+ "should_save": true,
2022
+ "should_training_stop": true
2023
+ },
2024
+ "attributes": {}
2025
+ }
2026
+ },
2027
+ "total_flos": 619390244487168.0,
2028
+ "train_batch_size": 1,
2029
+ "trial_name": null,
2030
+ "trial_params": null
2031
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a1b022ea3a1df7cc77be31885f0fdddcaa9dc83bf07e50c0a759385eeb4f06d
3
+ size 8120