starlineventures commited on
Commit
a92485f
·
verified ·
1 Parent(s): 392dcd2

starlineventures/chatgpt

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: microsoft/phi-2
3
  library_name: peft
4
  license: mit
5
  tags:
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # outputs
18
 
19
- This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
1
  ---
2
+ base_model: microsoft/Phi-3.5-mini-instruct
3
  library_name: peft
4
  license: mit
5
  tags:
 
16
 
17
  # outputs
18
 
19
+ This model is a fine-tuned version of [microsoft/Phi-3.5-mini-instruct](https://huggingface.co/microsoft/Phi-3.5-mini-instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
adapter_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
- "base_model_class": "PhiForCausalLM",
5
- "parent_library": "transformers.models.phi.modeling_phi"
6
  },
7
- "base_model_name_or_path": "microsoft/phi-2",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "gate_proj",
27
- "q_proj",
28
  "k_proj",
29
- "down_proj",
30
  "up_proj",
 
31
  "o_proj",
32
- "v_proj"
 
 
33
  ],
34
  "task_type": null,
35
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
+ "base_model_class": "Phi3ForCausalLM",
5
+ "parent_library": "transformers_modules.microsoft.Phi-3.5-mini-instruct.3145e03a9fd4cdd7cd953c34d9bbf7ad606122ca.modeling_phi3"
6
  },
7
+ "base_model_name_or_path": "microsoft/Phi-3.5-mini-instruct",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "k_proj",
 
27
  "up_proj",
28
+ "q_proj",
29
  "o_proj",
30
+ "v_proj",
31
+ "gate_proj",
32
+ "down_proj"
33
  ],
34
  "task_type": null,
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:537817d8a957663d34acfd9678d3fcdb85a1092df7428631b9b5efebce235b80
3
- size 94422368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fff7b051d12ef77bb0c869afde32d845678781cf3b5494203b66572011f998b2
3
+ size 35668592
added_tokens.json CHANGED
@@ -1,40 +1,13 @@
1
  {
2
- "\t\t": 50294,
3
- "\t\t\t": 50293,
4
- "\t\t\t\t": 50292,
5
- "\t\t\t\t\t": 50291,
6
- "\t\t\t\t\t\t": 50290,
7
- "\t\t\t\t\t\t\t": 50289,
8
- "\t\t\t\t\t\t\t\t": 50288,
9
- "\t\t\t\t\t\t\t\t\t": 50287,
10
- " ": 50286,
11
- " ": 50285,
12
- " ": 50284,
13
- " ": 50283,
14
- " ": 50282,
15
- " ": 50281,
16
- " ": 50280,
17
- " ": 50279,
18
- " ": 50278,
19
- " ": 50277,
20
- " ": 50276,
21
- " ": 50275,
22
- " ": 50274,
23
- " ": 50273,
24
- " ": 50272,
25
- " ": 50271,
26
- " ": 50270,
27
- " ": 50269,
28
- " ": 50268,
29
- " ": 50267,
30
- " ": 50266,
31
- " ": 50265,
32
- " ": 50264,
33
- " ": 50263,
34
- " ": 50262,
35
- " ": 50261,
36
- " ": 50260,
37
- " ": 50259,
38
- " ": 50258,
39
- " ": 50257
40
  }
 
1
  {
2
+ "<|assistant|>": 32001,
3
+ "<|endoftext|>": 32000,
4
+ "<|end|>": 32007,
5
+ "<|placeholder1|>": 32002,
6
+ "<|placeholder2|>": 32003,
7
+ "<|placeholder3|>": 32004,
8
+ "<|placeholder4|>": 32005,
9
+ "<|placeholder5|>": 32008,
10
+ "<|placeholder6|>": 32009,
11
+ "<|system|>": 32006,
12
+ "<|user|>": 32010
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  }
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.4417308241713281,
5
- "train_runtime": 112.6516,
6
- "train_samples_per_second": 5.406,
7
- "train_steps_per_second": 1.811
8
  }
 
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.4158199636959562,
5
+ "train_runtime": 99.4711,
6
+ "train_samples_per_second": 6.122,
7
+ "train_steps_per_second": 2.051
8
  }
special_tokens_map.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "bos_token": {
3
- "content": "<|endoftext|>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
@@ -15,7 +15,7 @@
15
  },
16
  "pad_token": "<|endoftext|>",
17
  "unk_token": {
18
- "content": "<|endoftext|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
 
1
  {
2
  "bos_token": {
3
+ "content": "<s>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
 
15
  },
16
  "pad_token": "<|endoftext|>",
17
  "unk_token": {
18
+ "content": "<unk>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json CHANGED
@@ -1,327 +1,132 @@
1
  {
2
  "add_bos_token": false,
3
- "add_prefix_space": false,
 
4
  "added_tokens_decoder": {
5
- "50256": {
6
- "content": "<|endoftext|>",
7
  "lstrip": false,
8
  "normalized": false,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
12
  },
13
- "50257": {
14
- "content": " ",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": false
20
- },
21
- "50258": {
22
- "content": " ",
23
- "lstrip": false,
24
- "normalized": true,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": false
28
- },
29
- "50259": {
30
- "content": " ",
31
- "lstrip": false,
32
- "normalized": true,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": false
36
- },
37
- "50260": {
38
- "content": " ",
39
- "lstrip": false,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": false
44
- },
45
- "50261": {
46
- "content": " ",
47
- "lstrip": false,
48
- "normalized": true,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": false
52
- },
53
- "50262": {
54
- "content": " ",
55
- "lstrip": false,
56
- "normalized": true,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": false
60
- },
61
- "50263": {
62
- "content": " ",
63
- "lstrip": false,
64
- "normalized": true,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": false
68
- },
69
- "50264": {
70
- "content": " ",
71
- "lstrip": false,
72
- "normalized": true,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": false
76
- },
77
- "50265": {
78
- "content": " ",
79
- "lstrip": false,
80
- "normalized": true,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": false
84
- },
85
- "50266": {
86
- "content": " ",
87
- "lstrip": false,
88
- "normalized": true,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": false
92
- },
93
- "50267": {
94
- "content": " ",
95
- "lstrip": false,
96
- "normalized": true,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": false
100
- },
101
- "50268": {
102
- "content": " ",
103
- "lstrip": false,
104
- "normalized": true,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": false
108
- },
109
- "50269": {
110
- "content": " ",
111
- "lstrip": false,
112
- "normalized": true,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": false
116
- },
117
- "50270": {
118
- "content": " ",
119
  "lstrip": false,
120
- "normalized": true,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": false
124
- },
125
- "50271": {
126
- "content": " ",
127
- "lstrip": false,
128
- "normalized": true,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": false
132
- },
133
- "50272": {
134
- "content": " ",
135
- "lstrip": false,
136
- "normalized": true,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "50273": {
142
- "content": " ",
143
- "lstrip": false,
144
- "normalized": true,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "50274": {
150
- "content": " ",
151
- "lstrip": false,
152
- "normalized": true,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "50275": {
158
- "content": " ",
159
- "lstrip": false,
160
- "normalized": true,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
- },
165
- "50276": {
166
- "content": " ",
167
- "lstrip": false,
168
- "normalized": true,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "50277": {
174
- "content": " ",
175
- "lstrip": false,
176
- "normalized": true,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
- },
181
- "50278": {
182
- "content": " ",
183
- "lstrip": false,
184
- "normalized": true,
185
- "rstrip": false,
186
- "single_word": false,
187
- "special": false
188
- },
189
- "50279": {
190
- "content": " ",
191
- "lstrip": false,
192
- "normalized": true,
193
- "rstrip": false,
194
- "single_word": false,
195
- "special": false
196
- },
197
- "50280": {
198
- "content": " ",
199
- "lstrip": false,
200
- "normalized": true,
201
- "rstrip": false,
202
- "single_word": false,
203
- "special": false
204
- },
205
- "50281": {
206
- "content": " ",
207
- "lstrip": false,
208
- "normalized": true,
209
- "rstrip": false,
210
- "single_word": false,
211
- "special": false
212
- },
213
- "50282": {
214
- "content": " ",
215
- "lstrip": false,
216
- "normalized": true,
217
  "rstrip": false,
218
  "single_word": false,
219
- "special": false
220
  },
221
- "50283": {
222
- "content": " ",
223
  "lstrip": false,
224
- "normalized": true,
225
- "rstrip": false,
226
  "single_word": false,
227
  "special": false
228
  },
229
- "50284": {
230
- "content": " ",
231
  "lstrip": false,
232
- "normalized": true,
233
  "rstrip": false,
234
  "single_word": false,
235
- "special": false
236
  },
237
- "50285": {
238
- "content": " ",
239
  "lstrip": false,
240
- "normalized": true,
241
- "rstrip": false,
242
  "single_word": false,
243
- "special": false
244
  },
245
- "50286": {
246
- "content": " ",
247
  "lstrip": false,
248
- "normalized": true,
249
- "rstrip": false,
250
  "single_word": false,
251
- "special": false
252
  },
253
- "50287": {
254
- "content": "\t\t\t\t\t\t\t\t\t",
255
  "lstrip": false,
256
- "normalized": true,
257
- "rstrip": false,
258
  "single_word": false,
259
- "special": false
260
  },
261
- "50288": {
262
- "content": "\t\t\t\t\t\t\t\t",
263
  "lstrip": false,
264
- "normalized": true,
265
- "rstrip": false,
266
  "single_word": false,
267
- "special": false
268
  },
269
- "50289": {
270
- "content": "\t\t\t\t\t\t\t",
271
  "lstrip": false,
272
- "normalized": true,
273
- "rstrip": false,
274
  "single_word": false,
275
- "special": false
276
  },
277
- "50290": {
278
- "content": "\t\t\t\t\t\t",
279
  "lstrip": false,
280
- "normalized": true,
281
- "rstrip": false,
282
  "single_word": false,
283
- "special": false
284
  },
285
- "50291": {
286
- "content": "\t\t\t\t\t",
287
  "lstrip": false,
288
- "normalized": true,
289
- "rstrip": false,
290
  "single_word": false,
291
- "special": false
292
  },
293
- "50292": {
294
- "content": "\t\t\t\t",
295
  "lstrip": false,
296
- "normalized": true,
297
- "rstrip": false,
298
  "single_word": false,
299
- "special": false
300
  },
301
- "50293": {
302
- "content": "\t\t\t",
303
  "lstrip": false,
304
- "normalized": true,
305
- "rstrip": false,
306
  "single_word": false,
307
- "special": false
308
  },
309
- "50294": {
310
- "content": "\t\t",
311
  "lstrip": false,
312
- "normalized": true,
313
- "rstrip": false,
314
  "single_word": false,
315
- "special": false
316
  }
317
  },
318
- "bos_token": "<|endoftext|>",
319
- "clean_up_tokenization_spaces": true,
 
320
  "eos_token": "<|endoftext|>",
321
- "errors": "replace",
322
  "model_max_length": 2048,
323
  "pad_token": "<|endoftext|>",
324
- "return_token_type_ids": false,
325
- "tokenizer_class": "CodeGenTokenizer",
326
- "unk_token": "<|endoftext|>"
 
 
 
327
  }
 
1
  {
2
  "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
11
  "single_word": false,
12
  "special": true
13
  },
14
+ "1": {
15
+ "content": "<s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "lstrip": false,
17
+ "normalized": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  "rstrip": false,
19
  "single_word": false,
20
+ "special": true
21
  },
22
+ "2": {
23
+ "content": "</s>",
24
  "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": true,
27
  "single_word": false,
28
  "special": false
29
  },
30
+ "32000": {
31
+ "content": "<|endoftext|>",
32
  "lstrip": false,
33
+ "normalized": false,
34
  "rstrip": false,
35
  "single_word": false,
36
+ "special": true
37
  },
38
+ "32001": {
39
+ "content": "<|assistant|>",
40
  "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": true,
43
  "single_word": false,
44
+ "special": true
45
  },
46
+ "32002": {
47
+ "content": "<|placeholder1|>",
48
  "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": true,
51
  "single_word": false,
52
+ "special": true
53
  },
54
+ "32003": {
55
+ "content": "<|placeholder2|>",
56
  "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": true,
59
  "single_word": false,
60
+ "special": true
61
  },
62
+ "32004": {
63
+ "content": "<|placeholder3|>",
64
  "lstrip": false,
65
+ "normalized": false,
66
+ "rstrip": true,
67
  "single_word": false,
68
+ "special": true
69
  },
70
+ "32005": {
71
+ "content": "<|placeholder4|>",
72
  "lstrip": false,
73
+ "normalized": false,
74
+ "rstrip": true,
75
  "single_word": false,
76
+ "special": true
77
  },
78
+ "32006": {
79
+ "content": "<|system|>",
80
  "lstrip": false,
81
+ "normalized": false,
82
+ "rstrip": true,
83
  "single_word": false,
84
+ "special": true
85
  },
86
+ "32007": {
87
+ "content": "<|end|>",
88
  "lstrip": false,
89
+ "normalized": false,
90
+ "rstrip": true,
91
  "single_word": false,
92
+ "special": true
93
  },
94
+ "32008": {
95
+ "content": "<|placeholder5|>",
96
  "lstrip": false,
97
+ "normalized": false,
98
+ "rstrip": true,
99
  "single_word": false,
100
+ "special": true
101
  },
102
+ "32009": {
103
+ "content": "<|placeholder6|>",
104
  "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": true,
107
  "single_word": false,
108
+ "special": true
109
  },
110
+ "32010": {
111
+ "content": "<|user|>",
112
  "lstrip": false,
113
+ "normalized": false,
114
+ "rstrip": true,
115
  "single_word": false,
116
+ "special": true
117
  }
118
  },
119
+ "bos_token": "<s>",
120
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}",
121
+ "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
+ "legacy": false,
124
  "model_max_length": 2048,
125
  "pad_token": "<|endoftext|>",
126
+ "padding_side": "right",
127
+ "sp_model_kwargs": {},
128
+ "spaces_between_special_tokens": false,
129
+ "tokenizer_class": "LlamaTokenizer",
130
+ "unk_token": "<unk>",
131
+ "use_default_system_prompt": false
132
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.4417308241713281,
5
- "train_runtime": 112.6516,
6
- "train_samples_per_second": 5.406,
7
- "train_steps_per_second": 1.811
8
  }
 
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.4158199636959562,
5
+ "train_runtime": 99.4711,
6
+ "train_samples_per_second": 6.122,
7
+ "train_steps_per_second": 2.051
8
  }
trainer_state.json CHANGED
@@ -10,152 +10,152 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.14705882352941177,
13
- "grad_norm": 5.014659881591797,
14
  "learning_rate": 9.509803921568627e-05,
15
- "loss": 0.8047,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.29411764705882354,
20
- "grad_norm": 3.9647512435913086,
21
  "learning_rate": 9.019607843137255e-05,
22
- "loss": 0.572,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.4411764705882353,
27
- "grad_norm": 3.050414800643921,
28
  "learning_rate": 8.529411764705883e-05,
29
- "loss": 0.5933,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.5882352941176471,
34
- "grad_norm": 1.585654377937317,
35
  "learning_rate": 8.039215686274511e-05,
36
- "loss": 0.4563,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.7352941176470589,
41
- "grad_norm": 1.262968897819519,
42
  "learning_rate": 7.549019607843137e-05,
43
- "loss": 0.4342,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.8823529411764706,
48
- "grad_norm": 0.7713326215744019,
49
  "learning_rate": 7.058823529411765e-05,
50
- "loss": 0.4967,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 1.0294117647058822,
55
- "grad_norm": 1.4584881067276,
56
  "learning_rate": 6.568627450980392e-05,
57
- "loss": 0.5211,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 1.1764705882352942,
62
- "grad_norm": 2.0033822059631348,
63
  "learning_rate": 6.078431372549019e-05,
64
- "loss": 0.3788,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.3235294117647058,
69
- "grad_norm": 2.14152455329895,
70
  "learning_rate": 5.588235294117647e-05,
71
- "loss": 0.4405,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 1.4705882352941178,
76
- "grad_norm": 2.8985249996185303,
77
  "learning_rate": 5.0980392156862745e-05,
78
- "loss": 0.3855,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 1.6176470588235294,
83
- "grad_norm": 2.2722222805023193,
84
  "learning_rate": 4.607843137254902e-05,
85
- "loss": 0.3668,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 1.7647058823529411,
90
- "grad_norm": 1.7401117086410522,
91
  "learning_rate": 4.11764705882353e-05,
92
- "loss": 0.4555,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 1.9117647058823528,
97
- "grad_norm": 2.3890228271484375,
98
  "learning_rate": 3.627450980392157e-05,
99
- "loss": 0.4121,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 2.0588235294117645,
104
- "grad_norm": 2.051992654800415,
105
  "learning_rate": 3.137254901960784e-05,
106
- "loss": 0.3758,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 2.2058823529411766,
111
- "grad_norm": 2.088096857070923,
112
  "learning_rate": 2.647058823529412e-05,
113
- "loss": 0.3738,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 2.3529411764705883,
118
- "grad_norm": 1.384030818939209,
119
  "learning_rate": 2.1568627450980395e-05,
120
- "loss": 0.393,
121
  "step": 160
122
  },
123
  {
124
  "epoch": 2.5,
125
- "grad_norm": 1.6448745727539062,
126
  "learning_rate": 1.6666666666666667e-05,
127
- "loss": 0.2952,
128
  "step": 170
129
  },
130
  {
131
  "epoch": 2.6470588235294117,
132
- "grad_norm": 1.227141261100769,
133
  "learning_rate": 1.1764705882352942e-05,
134
- "loss": 0.3821,
135
  "step": 180
136
  },
137
  {
138
  "epoch": 2.7941176470588234,
139
- "grad_norm": 2.20298433303833,
140
  "learning_rate": 6.862745098039216e-06,
141
- "loss": 0.3722,
142
  "step": 190
143
  },
144
  {
145
  "epoch": 2.9411764705882355,
146
- "grad_norm": 1.9124171733856201,
147
  "learning_rate": 1.96078431372549e-06,
148
- "loss": 0.36,
149
  "step": 200
150
  },
151
  {
152
  "epoch": 3.0,
153
  "step": 204,
154
  "total_flos": 0.0,
155
- "train_loss": 0.4417308241713281,
156
- "train_runtime": 112.6516,
157
- "train_samples_per_second": 5.406,
158
- "train_steps_per_second": 1.811
159
  }
160
  ],
161
  "logging_steps": 10,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.14705882352941177,
13
+ "grad_norm": 1.1240489482879639,
14
  "learning_rate": 9.509803921568627e-05,
15
+ "loss": 3.006,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.29411764705882354,
20
+ "grad_norm": 0.9972021579742432,
21
  "learning_rate": 9.019607843137255e-05,
22
+ "loss": 0.3771,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.4411764705882353,
27
+ "grad_norm": 1.274720549583435,
28
  "learning_rate": 8.529411764705883e-05,
29
+ "loss": 0.4497,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 0.5882352941176471,
34
+ "grad_norm": 0.8991940021514893,
35
  "learning_rate": 8.039215686274511e-05,
36
+ "loss": 0.3536,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 0.7352941176470589,
41
+ "grad_norm": 0.9119272828102112,
42
  "learning_rate": 7.549019607843137e-05,
43
+ "loss": 0.3353,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 0.8823529411764706,
48
+ "grad_norm": 0.6662648320198059,
49
  "learning_rate": 7.058823529411765e-05,
50
+ "loss": 0.4034,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 1.0294117647058822,
55
+ "grad_norm": 0.7458919882774353,
56
  "learning_rate": 6.568627450980392e-05,
57
+ "loss": 0.4318,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 1.1764705882352942,
62
+ "grad_norm": 0.937239944934845,
63
  "learning_rate": 6.078431372549019e-05,
64
+ "loss": 0.2594,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 1.3235294117647058,
69
+ "grad_norm": 0.8029321432113647,
70
  "learning_rate": 5.588235294117647e-05,
71
+ "loss": 0.2861,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 1.4705882352941178,
76
+ "grad_norm": 0.7472988367080688,
77
  "learning_rate": 5.0980392156862745e-05,
78
+ "loss": 0.2578,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 1.6176470588235294,
83
+ "grad_norm": 0.6276091933250427,
84
  "learning_rate": 4.607843137254902e-05,
85
+ "loss": 0.2554,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 1.7647058823529411,
90
+ "grad_norm": 0.551144540309906,
91
  "learning_rate": 4.11764705882353e-05,
92
+ "loss": 0.3111,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 1.9117647058823528,
97
+ "grad_norm": 1.0049179792404175,
98
  "learning_rate": 3.627450980392157e-05,
99
+ "loss": 0.2822,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 2.0588235294117645,
104
+ "grad_norm": 0.6848428845405579,
105
  "learning_rate": 3.137254901960784e-05,
106
+ "loss": 0.2257,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 2.2058823529411766,
111
+ "grad_norm": 0.7736006379127502,
112
  "learning_rate": 2.647058823529412e-05,
113
+ "loss": 0.2035,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 2.3529411764705883,
118
+ "grad_norm": 0.640957236289978,
119
  "learning_rate": 2.1568627450980395e-05,
120
+ "loss": 0.2181,
121
  "step": 160
122
  },
123
  {
124
  "epoch": 2.5,
125
+ "grad_norm": 1.5639573335647583,
126
  "learning_rate": 1.6666666666666667e-05,
127
+ "loss": 0.1645,
128
  "step": 170
129
  },
130
  {
131
  "epoch": 2.6470588235294117,
132
+ "grad_norm": 0.870328962802887,
133
  "learning_rate": 1.1764705882352942e-05,
134
+ "loss": 0.1894,
135
  "step": 180
136
  },
137
  {
138
  "epoch": 2.7941176470588234,
139
+ "grad_norm": 1.0381242036819458,
140
  "learning_rate": 6.862745098039216e-06,
141
+ "loss": 0.2089,
142
  "step": 190
143
  },
144
  {
145
  "epoch": 2.9411764705882355,
146
+ "grad_norm": 1.1926450729370117,
147
  "learning_rate": 1.96078431372549e-06,
148
+ "loss": 0.1855,
149
  "step": 200
150
  },
151
  {
152
  "epoch": 3.0,
153
  "step": 204,
154
  "total_flos": 0.0,
155
+ "train_loss": 0.4158199636959562,
156
+ "train_runtime": 99.4711,
157
+ "train_samples_per_second": 6.122,
158
+ "train_steps_per_second": 2.051
159
  }
160
  ],
161
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a4e8f2d3250958257fd416d205e9505bf49a86acfd66fa7e8bdd8b8358c882d
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2952133924800416bdb80493b0f16baf261b3a81b8d0a0bcd5bfedc4aed9112f
3
  size 5432