Yucheng Yin commited on
Commit
65f6508
·
1 Parent(s): 6b57efb

add model weights

Browse files
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ metrics:
5
+ - accuracy
6
+ model-index:
7
+ - name: test-mlm-bert-base-uncased-netflow-3M-mixed
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # test-mlm-bert-base-uncased-netflow-3M-mixed
15
+
16
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 0.7583
19
+ - Accuracy: 0.7759
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 5e-05
39
+ - train_batch_size: 32
40
+ - eval_batch_size: 32
41
+ - seed: 42
42
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
+ - lr_scheduler_type: linear
44
+ - num_epochs: 3.0
45
+
46
+ ### Training results
47
+
48
+
49
+
50
+ ### Framework versions
51
+
52
+ - Transformers 4.31.0.dev0
53
+ - Pytorch 2.0.1+cu117
54
+ - Datasets 2.13.0
55
+ - Tokenizers 0.13.3
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.7758689395616559,
4
+ "eval_loss": 0.7582568526268005,
5
+ "eval_runtime": 224.7846,
6
+ "eval_samples": 57799,
7
+ "eval_samples_per_second": 257.131,
8
+ "eval_steps_per_second": 8.039,
9
+ "perplexity": 2.134552136676338,
10
+ "train_loss": 0.3894348529865434,
11
+ "train_runtime": 39614.9645,
12
+ "train_samples": 1098180,
13
+ "train_samples_per_second": 83.164,
14
+ "train_steps_per_second": 2.599
15
+ }
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 3072,
12
+ "layer_norm_eps": 1e-12,
13
+ "max_position_embeddings": 512,
14
+ "model_type": "bert",
15
+ "num_attention_heads": 12,
16
+ "num_hidden_layers": 12,
17
+ "pad_token_id": 0,
18
+ "position_embedding_type": "absolute",
19
+ "torch_dtype": "float32",
20
+ "transformers_version": "4.31.0.dev0",
21
+ "type_vocab_size": 2,
22
+ "use_cache": true,
23
+ "vocab_size": 30522
24
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.7758689395616559,
4
+ "eval_loss": 0.7582568526268005,
5
+ "eval_runtime": 224.7846,
6
+ "eval_samples": 57799,
7
+ "eval_samples_per_second": 257.131,
8
+ "eval_steps_per_second": 8.039,
9
+ "perplexity": 2.134552136676338
10
+ }
generation_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "pad_token_id": 0,
4
+ "transformers_version": "4.31.0.dev0"
5
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b77e608defdc59e62add6ecd1258f2865f70c56bade1774d07abe73257c7c1e
3
+ size 438130617
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": true,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 512,
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.3894348529865434,
4
+ "train_runtime": 39614.9645,
5
+ "train_samples": 1098180,
6
+ "train_samples_per_second": 83.164,
7
+ "train_steps_per_second": 2.599
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,637 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "global_step": 102957,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "learning_rate": 4.9514360364035474e-05,
13
+ "loss": 1.3581,
14
+ "step": 1000
15
+ },
16
+ {
17
+ "epoch": 0.06,
18
+ "learning_rate": 4.9028720728070946e-05,
19
+ "loss": 0.6756,
20
+ "step": 2000
21
+ },
22
+ {
23
+ "epoch": 0.09,
24
+ "learning_rate": 4.854308109210642e-05,
25
+ "loss": 0.5975,
26
+ "step": 3000
27
+ },
28
+ {
29
+ "epoch": 0.12,
30
+ "learning_rate": 4.805744145614189e-05,
31
+ "loss": 0.5572,
32
+ "step": 4000
33
+ },
34
+ {
35
+ "epoch": 0.15,
36
+ "learning_rate": 4.757180182017736e-05,
37
+ "loss": 0.5289,
38
+ "step": 5000
39
+ },
40
+ {
41
+ "epoch": 0.17,
42
+ "learning_rate": 4.7086162184212826e-05,
43
+ "loss": 0.5127,
44
+ "step": 6000
45
+ },
46
+ {
47
+ "epoch": 0.2,
48
+ "learning_rate": 4.66005225482483e-05,
49
+ "loss": 0.4999,
50
+ "step": 7000
51
+ },
52
+ {
53
+ "epoch": 0.23,
54
+ "learning_rate": 4.611488291228377e-05,
55
+ "loss": 0.4843,
56
+ "step": 8000
57
+ },
58
+ {
59
+ "epoch": 0.26,
60
+ "learning_rate": 4.562924327631924e-05,
61
+ "loss": 0.4701,
62
+ "step": 9000
63
+ },
64
+ {
65
+ "epoch": 0.29,
66
+ "learning_rate": 4.5143603640354713e-05,
67
+ "loss": 0.4619,
68
+ "step": 10000
69
+ },
70
+ {
71
+ "epoch": 0.32,
72
+ "learning_rate": 4.4657964004390185e-05,
73
+ "loss": 0.4581,
74
+ "step": 11000
75
+ },
76
+ {
77
+ "epoch": 0.35,
78
+ "learning_rate": 4.417232436842566e-05,
79
+ "loss": 0.4538,
80
+ "step": 12000
81
+ },
82
+ {
83
+ "epoch": 0.38,
84
+ "learning_rate": 4.368668473246112e-05,
85
+ "loss": 0.4452,
86
+ "step": 13000
87
+ },
88
+ {
89
+ "epoch": 0.41,
90
+ "learning_rate": 4.3201045096496594e-05,
91
+ "loss": 0.4426,
92
+ "step": 14000
93
+ },
94
+ {
95
+ "epoch": 0.44,
96
+ "learning_rate": 4.271540546053207e-05,
97
+ "loss": 0.4381,
98
+ "step": 15000
99
+ },
100
+ {
101
+ "epoch": 0.47,
102
+ "learning_rate": 4.2229765824567544e-05,
103
+ "loss": 0.4302,
104
+ "step": 16000
105
+ },
106
+ {
107
+ "epoch": 0.5,
108
+ "learning_rate": 4.174412618860301e-05,
109
+ "loss": 0.4299,
110
+ "step": 17000
111
+ },
112
+ {
113
+ "epoch": 0.52,
114
+ "learning_rate": 4.125848655263848e-05,
115
+ "loss": 0.4218,
116
+ "step": 18000
117
+ },
118
+ {
119
+ "epoch": 0.55,
120
+ "learning_rate": 4.077284691667395e-05,
121
+ "loss": 0.4219,
122
+ "step": 19000
123
+ },
124
+ {
125
+ "epoch": 0.58,
126
+ "learning_rate": 4.0287207280709424e-05,
127
+ "loss": 0.4178,
128
+ "step": 20000
129
+ },
130
+ {
131
+ "epoch": 0.61,
132
+ "learning_rate": 3.9801567644744896e-05,
133
+ "loss": 0.4169,
134
+ "step": 21000
135
+ },
136
+ {
137
+ "epoch": 0.64,
138
+ "learning_rate": 3.931592800878037e-05,
139
+ "loss": 0.4161,
140
+ "step": 22000
141
+ },
142
+ {
143
+ "epoch": 0.67,
144
+ "learning_rate": 3.883028837281584e-05,
145
+ "loss": 0.409,
146
+ "step": 23000
147
+ },
148
+ {
149
+ "epoch": 0.7,
150
+ "learning_rate": 3.8344648736851305e-05,
151
+ "loss": 0.4057,
152
+ "step": 24000
153
+ },
154
+ {
155
+ "epoch": 0.73,
156
+ "learning_rate": 3.7859009100886777e-05,
157
+ "loss": 0.4045,
158
+ "step": 25000
159
+ },
160
+ {
161
+ "epoch": 0.76,
162
+ "learning_rate": 3.7373369464922255e-05,
163
+ "loss": 0.4049,
164
+ "step": 26000
165
+ },
166
+ {
167
+ "epoch": 0.79,
168
+ "learning_rate": 3.688772982895772e-05,
169
+ "loss": 0.399,
170
+ "step": 27000
171
+ },
172
+ {
173
+ "epoch": 0.82,
174
+ "learning_rate": 3.640209019299319e-05,
175
+ "loss": 0.3971,
176
+ "step": 28000
177
+ },
178
+ {
179
+ "epoch": 0.85,
180
+ "learning_rate": 3.5916450557028664e-05,
181
+ "loss": 0.3951,
182
+ "step": 29000
183
+ },
184
+ {
185
+ "epoch": 0.87,
186
+ "learning_rate": 3.5430810921064135e-05,
187
+ "loss": 0.3963,
188
+ "step": 30000
189
+ },
190
+ {
191
+ "epoch": 0.9,
192
+ "learning_rate": 3.49451712850996e-05,
193
+ "loss": 0.3933,
194
+ "step": 31000
195
+ },
196
+ {
197
+ "epoch": 0.93,
198
+ "learning_rate": 3.445953164913508e-05,
199
+ "loss": 0.3924,
200
+ "step": 32000
201
+ },
202
+ {
203
+ "epoch": 0.96,
204
+ "learning_rate": 3.397389201317055e-05,
205
+ "loss": 0.3889,
206
+ "step": 33000
207
+ },
208
+ {
209
+ "epoch": 0.99,
210
+ "learning_rate": 3.348825237720602e-05,
211
+ "loss": 0.3868,
212
+ "step": 34000
213
+ },
214
+ {
215
+ "epoch": 1.02,
216
+ "learning_rate": 3.300261274124149e-05,
217
+ "loss": 0.3844,
218
+ "step": 35000
219
+ },
220
+ {
221
+ "epoch": 1.05,
222
+ "learning_rate": 3.251697310527696e-05,
223
+ "loss": 0.3802,
224
+ "step": 36000
225
+ },
226
+ {
227
+ "epoch": 1.08,
228
+ "learning_rate": 3.203133346931244e-05,
229
+ "loss": 0.3825,
230
+ "step": 37000
231
+ },
232
+ {
233
+ "epoch": 1.11,
234
+ "learning_rate": 3.15456938333479e-05,
235
+ "loss": 0.3802,
236
+ "step": 38000
237
+ },
238
+ {
239
+ "epoch": 1.14,
240
+ "learning_rate": 3.1060054197383375e-05,
241
+ "loss": 0.378,
242
+ "step": 39000
243
+ },
244
+ {
245
+ "epoch": 1.17,
246
+ "learning_rate": 3.0574414561418846e-05,
247
+ "loss": 0.3768,
248
+ "step": 40000
249
+ },
250
+ {
251
+ "epoch": 1.19,
252
+ "learning_rate": 3.0088774925454315e-05,
253
+ "loss": 0.3749,
254
+ "step": 41000
255
+ },
256
+ {
257
+ "epoch": 1.22,
258
+ "learning_rate": 2.9603135289489787e-05,
259
+ "loss": 0.3758,
260
+ "step": 42000
261
+ },
262
+ {
263
+ "epoch": 1.25,
264
+ "learning_rate": 2.9117495653525262e-05,
265
+ "loss": 0.3733,
266
+ "step": 43000
267
+ },
268
+ {
269
+ "epoch": 1.28,
270
+ "learning_rate": 2.8631856017560734e-05,
271
+ "loss": 0.3694,
272
+ "step": 44000
273
+ },
274
+ {
275
+ "epoch": 1.31,
276
+ "learning_rate": 2.8146216381596202e-05,
277
+ "loss": 0.3689,
278
+ "step": 45000
279
+ },
280
+ {
281
+ "epoch": 1.34,
282
+ "learning_rate": 2.7660576745631674e-05,
283
+ "loss": 0.3693,
284
+ "step": 46000
285
+ },
286
+ {
287
+ "epoch": 1.37,
288
+ "learning_rate": 2.7174937109667142e-05,
289
+ "loss": 0.3703,
290
+ "step": 47000
291
+ },
292
+ {
293
+ "epoch": 1.4,
294
+ "learning_rate": 2.6689297473702614e-05,
295
+ "loss": 0.366,
296
+ "step": 48000
297
+ },
298
+ {
299
+ "epoch": 1.43,
300
+ "learning_rate": 2.620365783773809e-05,
301
+ "loss": 0.3664,
302
+ "step": 49000
303
+ },
304
+ {
305
+ "epoch": 1.46,
306
+ "learning_rate": 2.5718018201773557e-05,
307
+ "loss": 0.365,
308
+ "step": 50000
309
+ },
310
+ {
311
+ "epoch": 1.49,
312
+ "learning_rate": 2.523237856580903e-05,
313
+ "loss": 0.3636,
314
+ "step": 51000
315
+ },
316
+ {
317
+ "epoch": 1.52,
318
+ "learning_rate": 2.4746738929844498e-05,
319
+ "loss": 0.3611,
320
+ "step": 52000
321
+ },
322
+ {
323
+ "epoch": 1.54,
324
+ "learning_rate": 2.4261099293879973e-05,
325
+ "loss": 0.3626,
326
+ "step": 53000
327
+ },
328
+ {
329
+ "epoch": 1.57,
330
+ "learning_rate": 2.377545965791544e-05,
331
+ "loss": 0.3588,
332
+ "step": 54000
333
+ },
334
+ {
335
+ "epoch": 1.6,
336
+ "learning_rate": 2.3289820021950913e-05,
337
+ "loss": 0.3618,
338
+ "step": 55000
339
+ },
340
+ {
341
+ "epoch": 1.63,
342
+ "learning_rate": 2.2804180385986385e-05,
343
+ "loss": 0.3589,
344
+ "step": 56000
345
+ },
346
+ {
347
+ "epoch": 1.66,
348
+ "learning_rate": 2.2318540750021853e-05,
349
+ "loss": 0.356,
350
+ "step": 57000
351
+ },
352
+ {
353
+ "epoch": 1.69,
354
+ "learning_rate": 2.1832901114057325e-05,
355
+ "loss": 0.3545,
356
+ "step": 58000
357
+ },
358
+ {
359
+ "epoch": 1.72,
360
+ "learning_rate": 2.1347261478092797e-05,
361
+ "loss": 0.3561,
362
+ "step": 59000
363
+ },
364
+ {
365
+ "epoch": 1.75,
366
+ "learning_rate": 2.086162184212827e-05,
367
+ "loss": 0.3534,
368
+ "step": 60000
369
+ },
370
+ {
371
+ "epoch": 1.78,
372
+ "learning_rate": 2.0375982206163737e-05,
373
+ "loss": 0.3514,
374
+ "step": 61000
375
+ },
376
+ {
377
+ "epoch": 1.81,
378
+ "learning_rate": 1.9890342570199212e-05,
379
+ "loss": 0.3526,
380
+ "step": 62000
381
+ },
382
+ {
383
+ "epoch": 1.84,
384
+ "learning_rate": 1.940470293423468e-05,
385
+ "loss": 0.3512,
386
+ "step": 63000
387
+ },
388
+ {
389
+ "epoch": 1.86,
390
+ "learning_rate": 1.8919063298270152e-05,
391
+ "loss": 0.3509,
392
+ "step": 64000
393
+ },
394
+ {
395
+ "epoch": 1.89,
396
+ "learning_rate": 1.8433423662305624e-05,
397
+ "loss": 0.3484,
398
+ "step": 65000
399
+ },
400
+ {
401
+ "epoch": 1.92,
402
+ "learning_rate": 1.7947784026341092e-05,
403
+ "loss": 0.3478,
404
+ "step": 66000
405
+ },
406
+ {
407
+ "epoch": 1.95,
408
+ "learning_rate": 1.7462144390376567e-05,
409
+ "loss": 0.3449,
410
+ "step": 67000
411
+ },
412
+ {
413
+ "epoch": 1.98,
414
+ "learning_rate": 1.6976504754412036e-05,
415
+ "loss": 0.3483,
416
+ "step": 68000
417
+ },
418
+ {
419
+ "epoch": 2.01,
420
+ "learning_rate": 1.6490865118447508e-05,
421
+ "loss": 0.3455,
422
+ "step": 69000
423
+ },
424
+ {
425
+ "epoch": 2.04,
426
+ "learning_rate": 1.600522548248298e-05,
427
+ "loss": 0.3441,
428
+ "step": 70000
429
+ },
430
+ {
431
+ "epoch": 2.07,
432
+ "learning_rate": 1.551958584651845e-05,
433
+ "loss": 0.3446,
434
+ "step": 71000
435
+ },
436
+ {
437
+ "epoch": 2.1,
438
+ "learning_rate": 1.503394621055392e-05,
439
+ "loss": 0.343,
440
+ "step": 72000
441
+ },
442
+ {
443
+ "epoch": 2.13,
444
+ "learning_rate": 1.4548306574589393e-05,
445
+ "loss": 0.3405,
446
+ "step": 73000
447
+ },
448
+ {
449
+ "epoch": 2.16,
450
+ "learning_rate": 1.4062666938624863e-05,
451
+ "loss": 0.3401,
452
+ "step": 74000
453
+ },
454
+ {
455
+ "epoch": 2.19,
456
+ "learning_rate": 1.3577027302660333e-05,
457
+ "loss": 0.3385,
458
+ "step": 75000
459
+ },
460
+ {
461
+ "epoch": 2.21,
462
+ "learning_rate": 1.3091387666695807e-05,
463
+ "loss": 0.3378,
464
+ "step": 76000
465
+ },
466
+ {
467
+ "epoch": 2.24,
468
+ "learning_rate": 1.2605748030731277e-05,
469
+ "loss": 0.337,
470
+ "step": 77000
471
+ },
472
+ {
473
+ "epoch": 2.27,
474
+ "learning_rate": 1.2120108394766749e-05,
475
+ "loss": 0.3361,
476
+ "step": 78000
477
+ },
478
+ {
479
+ "epoch": 2.3,
480
+ "learning_rate": 1.1634468758802219e-05,
481
+ "loss": 0.338,
482
+ "step": 79000
483
+ },
484
+ {
485
+ "epoch": 2.33,
486
+ "learning_rate": 1.1148829122837689e-05,
487
+ "loss": 0.3367,
488
+ "step": 80000
489
+ },
490
+ {
491
+ "epoch": 2.36,
492
+ "learning_rate": 1.066318948687316e-05,
493
+ "loss": 0.3351,
494
+ "step": 81000
495
+ },
496
+ {
497
+ "epoch": 2.39,
498
+ "learning_rate": 1.0177549850908632e-05,
499
+ "loss": 0.3363,
500
+ "step": 82000
501
+ },
502
+ {
503
+ "epoch": 2.42,
504
+ "learning_rate": 9.691910214944102e-06,
505
+ "loss": 0.3327,
506
+ "step": 83000
507
+ },
508
+ {
509
+ "epoch": 2.45,
510
+ "learning_rate": 9.206270578979574e-06,
511
+ "loss": 0.3318,
512
+ "step": 84000
513
+ },
514
+ {
515
+ "epoch": 2.48,
516
+ "learning_rate": 8.720630943015046e-06,
517
+ "loss": 0.3307,
518
+ "step": 85000
519
+ },
520
+ {
521
+ "epoch": 2.51,
522
+ "learning_rate": 8.234991307050518e-06,
523
+ "loss": 0.3337,
524
+ "step": 86000
525
+ },
526
+ {
527
+ "epoch": 2.54,
528
+ "learning_rate": 7.749351671085988e-06,
529
+ "loss": 0.3299,
530
+ "step": 87000
531
+ },
532
+ {
533
+ "epoch": 2.56,
534
+ "learning_rate": 7.263712035121459e-06,
535
+ "loss": 0.3291,
536
+ "step": 88000
537
+ },
538
+ {
539
+ "epoch": 2.59,
540
+ "learning_rate": 6.7780723991569305e-06,
541
+ "loss": 0.3328,
542
+ "step": 89000
543
+ },
544
+ {
545
+ "epoch": 2.62,
546
+ "learning_rate": 6.292432763192401e-06,
547
+ "loss": 0.3279,
548
+ "step": 90000
549
+ },
550
+ {
551
+ "epoch": 2.65,
552
+ "learning_rate": 5.8067931272278715e-06,
553
+ "loss": 0.3331,
554
+ "step": 91000
555
+ },
556
+ {
557
+ "epoch": 2.68,
558
+ "learning_rate": 5.321153491263343e-06,
559
+ "loss": 0.3293,
560
+ "step": 92000
561
+ },
562
+ {
563
+ "epoch": 2.71,
564
+ "learning_rate": 4.835513855298814e-06,
565
+ "loss": 0.3274,
566
+ "step": 93000
567
+ },
568
+ {
569
+ "epoch": 2.74,
570
+ "learning_rate": 4.349874219334285e-06,
571
+ "loss": 0.3267,
572
+ "step": 94000
573
+ },
574
+ {
575
+ "epoch": 2.77,
576
+ "learning_rate": 3.864234583369757e-06,
577
+ "loss": 0.3261,
578
+ "step": 95000
579
+ },
580
+ {
581
+ "epoch": 2.8,
582
+ "learning_rate": 3.3785949474052275e-06,
583
+ "loss": 0.3238,
584
+ "step": 96000
585
+ },
586
+ {
587
+ "epoch": 2.83,
588
+ "learning_rate": 2.8929553114406984e-06,
589
+ "loss": 0.3237,
590
+ "step": 97000
591
+ },
592
+ {
593
+ "epoch": 2.86,
594
+ "learning_rate": 2.4073156754761698e-06,
595
+ "loss": 0.3225,
596
+ "step": 98000
597
+ },
598
+ {
599
+ "epoch": 2.88,
600
+ "learning_rate": 1.9216760395116407e-06,
601
+ "loss": 0.3242,
602
+ "step": 99000
603
+ },
604
+ {
605
+ "epoch": 2.91,
606
+ "learning_rate": 1.436036403547112e-06,
607
+ "loss": 0.3275,
608
+ "step": 100000
609
+ },
610
+ {
611
+ "epoch": 2.94,
612
+ "learning_rate": 9.50396767582583e-07,
613
+ "loss": 0.3219,
614
+ "step": 101000
615
+ },
616
+ {
617
+ "epoch": 2.97,
618
+ "learning_rate": 4.6475713161805417e-07,
619
+ "loss": 0.3232,
620
+ "step": 102000
621
+ },
622
+ {
623
+ "epoch": 3.0,
624
+ "step": 102957,
625
+ "total_flos": 7.554546670318524e+17,
626
+ "train_loss": 0.3894348529865434,
627
+ "train_runtime": 39614.9645,
628
+ "train_samples_per_second": 83.164,
629
+ "train_steps_per_second": 2.599
630
+ }
631
+ ],
632
+ "max_steps": 102957,
633
+ "num_train_epochs": 3,
634
+ "total_flos": 7.554546670318524e+17,
635
+ "trial_name": null,
636
+ "trial_params": null
637
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6040d7f68cec7feb301a17963eafd78a14d6d539216e8be65cf95419096787a3
3
+ size 4027
vocab.txt ADDED
@@ -0,0 +1,632 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [UNK]
2
+ [PAD]
3
+ [MASK]
4
+ [CLS]
5
+ [SEP]
6
+ 0___numeric___srcip_00___0
7
+ 0___numeric___srcip_00___1
8
+ 0___numeric___srcip_00___2
9
+ 0___numeric___srcip_00___3
10
+ 0___numeric___srcip_00___4
11
+ 0___numeric___srcip_01___0
12
+ 0___numeric___srcip_01___1
13
+ 0___numeric___srcip_01___2
14
+ 0___numeric___srcip_01___3
15
+ 0___numeric___srcip_01___4
16
+ 0___numeric___srcip_01___5
17
+ 0___numeric___srcip_01___6
18
+ 0___numeric___srcip_01___7
19
+ 0___numeric___srcip_01___8
20
+ 0___numeric___srcip_01___9
21
+ 0___numeric___srcip_02___0
22
+ 0___numeric___srcip_02___1
23
+ 0___numeric___srcip_02___2
24
+ 0___numeric___srcip_02___3
25
+ 0___numeric___srcip_02___4
26
+ 0___numeric___srcip_02___5
27
+ 0___numeric___srcip_02___6
28
+ 0___numeric___srcip_02___7
29
+ 0___numeric___srcip_02___8
30
+ 0___numeric___srcip_02___9
31
+ 0___numeric___srcip_03___0
32
+ 0___numeric___srcip_03___1
33
+ 0___numeric___srcip_03___2
34
+ 0___numeric___srcip_03___3
35
+ 0___numeric___srcip_03___4
36
+ 0___numeric___srcip_03___5
37
+ 0___numeric___srcip_03___6
38
+ 0___numeric___srcip_03___7
39
+ 0___numeric___srcip_03___8
40
+ 0___numeric___srcip_03___9
41
+ 0___numeric___srcip_04___0
42
+ 0___numeric___srcip_04___1
43
+ 0___numeric___srcip_04___2
44
+ 0___numeric___srcip_04___3
45
+ 0___numeric___srcip_04___4
46
+ 0___numeric___srcip_04___5
47
+ 0___numeric___srcip_04___6
48
+ 0___numeric___srcip_04___7
49
+ 0___numeric___srcip_04___8
50
+ 0___numeric___srcip_04___9
51
+ 0___numeric___srcip_05___0
52
+ 0___numeric___srcip_05___1
53
+ 0___numeric___srcip_05___2
54
+ 0___numeric___srcip_05___3
55
+ 0___numeric___srcip_05___4
56
+ 0___numeric___srcip_05___5
57
+ 0___numeric___srcip_05___6
58
+ 0___numeric___srcip_05___7
59
+ 0___numeric___srcip_05___8
60
+ 0___numeric___srcip_05___9
61
+ 0___numeric___srcip_06___0
62
+ 0___numeric___srcip_06___1
63
+ 0___numeric___srcip_06___2
64
+ 0___numeric___srcip_06___3
65
+ 0___numeric___srcip_06___4
66
+ 0___numeric___srcip_06___5
67
+ 0___numeric___srcip_06___6
68
+ 0___numeric___srcip_06___7
69
+ 0___numeric___srcip_06___8
70
+ 0___numeric___srcip_06___9
71
+ 0___numeric___srcip_07___0
72
+ 0___numeric___srcip_07___1
73
+ 0___numeric___srcip_07___2
74
+ 0___numeric___srcip_07___3
75
+ 0___numeric___srcip_07___4
76
+ 0___numeric___srcip_07___5
77
+ 0___numeric___srcip_07___6
78
+ 0___numeric___srcip_07___7
79
+ 0___numeric___srcip_07___8
80
+ 0___numeric___srcip_07___9
81
+ 0___numeric___srcip_08___0
82
+ 0___numeric___srcip_08___1
83
+ 0___numeric___srcip_08___2
84
+ 0___numeric___srcip_08___3
85
+ 0___numeric___srcip_08___4
86
+ 0___numeric___srcip_08___5
87
+ 0___numeric___srcip_08___6
88
+ 0___numeric___srcip_08___7
89
+ 0___numeric___srcip_08___8
90
+ 0___numeric___srcip_08___9
91
+ 0___numeric___srcip_09___0
92
+ 0___numeric___srcip_09___1
93
+ 0___numeric___srcip_09___2
94
+ 0___numeric___srcip_09___3
95
+ 0___numeric___srcip_09___4
96
+ 0___numeric___srcip_09___5
97
+ 0___numeric___srcip_09___6
98
+ 0___numeric___srcip_09___7
99
+ 0___numeric___srcip_09___8
100
+ 0___numeric___srcip_09___9
101
+ 1___numeric___dstip_00___0
102
+ 1___numeric___dstip_00___1
103
+ 1___numeric___dstip_00___2
104
+ 1___numeric___dstip_00___3
105
+ 1___numeric___dstip_00___4
106
+ 1___numeric___dstip_01___0
107
+ 1___numeric___dstip_01___1
108
+ 1___numeric___dstip_01___2
109
+ 1___numeric___dstip_01___3
110
+ 1___numeric___dstip_01___4
111
+ 1___numeric___dstip_01___5
112
+ 1___numeric___dstip_01___6
113
+ 1___numeric___dstip_01___7
114
+ 1___numeric___dstip_01___8
115
+ 1___numeric___dstip_01___9
116
+ 1___numeric___dstip_02___0
117
+ 1___numeric___dstip_02___1
118
+ 1___numeric___dstip_02___2
119
+ 1___numeric___dstip_02___3
120
+ 1___numeric___dstip_02___4
121
+ 1___numeric___dstip_02___5
122
+ 1___numeric___dstip_02___6
123
+ 1___numeric___dstip_02___7
124
+ 1___numeric___dstip_02___8
125
+ 1___numeric___dstip_02___9
126
+ 1___numeric___dstip_03___0
127
+ 1___numeric___dstip_03___1
128
+ 1___numeric___dstip_03___2
129
+ 1___numeric___dstip_03___3
130
+ 1___numeric___dstip_03___4
131
+ 1___numeric___dstip_03___5
132
+ 1___numeric___dstip_03___6
133
+ 1___numeric___dstip_03___7
134
+ 1___numeric___dstip_03___8
135
+ 1___numeric___dstip_03___9
136
+ 1___numeric___dstip_04___0
137
+ 1___numeric___dstip_04___1
138
+ 1___numeric___dstip_04___2
139
+ 1___numeric___dstip_04___3
140
+ 1___numeric___dstip_04___4
141
+ 1___numeric___dstip_04___5
142
+ 1___numeric___dstip_04___6
143
+ 1___numeric___dstip_04___7
144
+ 1___numeric___dstip_04___8
145
+ 1___numeric___dstip_04___9
146
+ 1___numeric___dstip_05___0
147
+ 1___numeric___dstip_05___1
148
+ 1___numeric___dstip_05___2
149
+ 1___numeric___dstip_05___3
150
+ 1___numeric___dstip_05___4
151
+ 1___numeric___dstip_05___5
152
+ 1___numeric___dstip_05___6
153
+ 1___numeric___dstip_05___7
154
+ 1___numeric___dstip_05___8
155
+ 1___numeric___dstip_05___9
156
+ 1___numeric___dstip_06___0
157
+ 1___numeric___dstip_06___1
158
+ 1___numeric___dstip_06___2
159
+ 1___numeric___dstip_06___3
160
+ 1___numeric___dstip_06___4
161
+ 1___numeric___dstip_06___5
162
+ 1___numeric___dstip_06___6
163
+ 1___numeric___dstip_06___7
164
+ 1___numeric___dstip_06___8
165
+ 1___numeric___dstip_06___9
166
+ 1___numeric___dstip_07___0
167
+ 1___numeric___dstip_07___1
168
+ 1___numeric___dstip_07___2
169
+ 1___numeric___dstip_07___3
170
+ 1___numeric___dstip_07___4
171
+ 1___numeric___dstip_07___5
172
+ 1___numeric___dstip_07___6
173
+ 1___numeric___dstip_07___7
174
+ 1___numeric___dstip_07___8
175
+ 1___numeric___dstip_07___9
176
+ 1___numeric___dstip_08___0
177
+ 1___numeric___dstip_08___1
178
+ 1___numeric___dstip_08___2
179
+ 1___numeric___dstip_08___3
180
+ 1___numeric___dstip_08___4
181
+ 1___numeric___dstip_08___5
182
+ 1___numeric___dstip_08___6
183
+ 1___numeric___dstip_08___7
184
+ 1___numeric___dstip_08___8
185
+ 1___numeric___dstip_08___9
186
+ 1___numeric___dstip_09___0
187
+ 1___numeric___dstip_09___1
188
+ 1___numeric___dstip_09___2
189
+ 1___numeric___dstip_09___3
190
+ 1___numeric___dstip_09___4
191
+ 1___numeric___dstip_09___5
192
+ 1___numeric___dstip_09___6
193
+ 1___numeric___dstip_09___7
194
+ 1___numeric___dstip_09___8
195
+ 1___numeric___dstip_09___9
196
+ 2___numeric___srcport_00___0
197
+ 2___numeric___srcport_00___1
198
+ 2___numeric___srcport_00___2
199
+ 2___numeric___srcport_00___3
200
+ 2___numeric___srcport_00___4
201
+ 2___numeric___srcport_00___5
202
+ 2___numeric___srcport_00___6
203
+ 2___numeric___srcport_01___0
204
+ 2___numeric___srcport_01___1
205
+ 2___numeric___srcport_01___2
206
+ 2___numeric___srcport_01___3
207
+ 2___numeric___srcport_01___4
208
+ 2___numeric___srcport_01___5
209
+ 2___numeric___srcport_01___6
210
+ 2___numeric___srcport_01___7
211
+ 2___numeric___srcport_01___8
212
+ 2___numeric___srcport_01___9
213
+ 2___numeric___srcport_02___0
214
+ 2___numeric___srcport_02___1
215
+ 2___numeric___srcport_02___2
216
+ 2___numeric___srcport_02___3
217
+ 2___numeric___srcport_02___4
218
+ 2___numeric___srcport_02___5
219
+ 2___numeric___srcport_02___6
220
+ 2___numeric___srcport_02___7
221
+ 2___numeric___srcport_02___8
222
+ 2___numeric___srcport_02___9
223
+ 2___numeric___srcport_03___0
224
+ 2___numeric___srcport_03___1
225
+ 2___numeric___srcport_03___2
226
+ 2___numeric___srcport_03___3
227
+ 2___numeric___srcport_03___4
228
+ 2___numeric___srcport_03___5
229
+ 2___numeric___srcport_03___6
230
+ 2___numeric___srcport_03___7
231
+ 2___numeric___srcport_03___8
232
+ 2___numeric___srcport_03___9
233
+ 2___numeric___srcport_04___0
234
+ 2___numeric___srcport_04___1
235
+ 2___numeric___srcport_04___2
236
+ 2___numeric___srcport_04___3
237
+ 2___numeric___srcport_04___4
238
+ 2___numeric___srcport_04___5
239
+ 2___numeric___srcport_04___6
240
+ 2___numeric___srcport_04___7
241
+ 2___numeric___srcport_04___8
242
+ 2___numeric___srcport_04___9
243
+ 3___numeric___dstport_00___0
244
+ 3___numeric___dstport_00___1
245
+ 3___numeric___dstport_00___2
246
+ 3___numeric___dstport_00___3
247
+ 3___numeric___dstport_00___4
248
+ 3___numeric___dstport_00___5
249
+ 3___numeric___dstport_00___6
250
+ 3___numeric___dstport_01___0
251
+ 3___numeric___dstport_01___1
252
+ 3___numeric___dstport_01___2
253
+ 3___numeric___dstport_01___3
254
+ 3___numeric___dstport_01___4
255
+ 3___numeric___dstport_01___5
256
+ 3___numeric___dstport_01___6
257
+ 3___numeric___dstport_01___7
258
+ 3___numeric___dstport_01___8
259
+ 3___numeric___dstport_01___9
260
+ 3___numeric___dstport_02___0
261
+ 3___numeric___dstport_02___1
262
+ 3___numeric___dstport_02___2
263
+ 3___numeric___dstport_02___3
264
+ 3___numeric___dstport_02___4
265
+ 3___numeric___dstport_02___5
266
+ 3___numeric___dstport_02___6
267
+ 3___numeric___dstport_02___7
268
+ 3___numeric___dstport_02___8
269
+ 3___numeric___dstport_02___9
270
+ 3___numeric___dstport_03___0
271
+ 3___numeric___dstport_03___1
272
+ 3___numeric___dstport_03___2
273
+ 3___numeric___dstport_03___3
274
+ 3___numeric___dstport_03___4
275
+ 3___numeric___dstport_03___5
276
+ 3___numeric___dstport_03___6
277
+ 3___numeric___dstport_03___7
278
+ 3___numeric___dstport_03___8
279
+ 3___numeric___dstport_03___9
280
+ 3___numeric___dstport_04___0
281
+ 3___numeric___dstport_04___1
282
+ 3___numeric___dstport_04___2
283
+ 3___numeric___dstport_04___3
284
+ 3___numeric___dstport_04___4
285
+ 3___numeric___dstport_04___5
286
+ 3___numeric___dstport_04___6
287
+ 3___numeric___dstport_04___7
288
+ 3___numeric___dstport_04___8
289
+ 3___numeric___dstport_04___9
290
+ 4___categorical___proto___esp
291
+ 4___categorical___proto___gre
292
+ 4___categorical___proto___icmp
293
+ 4___categorical___proto___ipip
294
+ 4___categorical___proto___ipv6
295
+ 4___categorical___proto___tcp
296
+ 4___categorical___proto___udp
297
+ 5___numeric___ts_00___1
298
+ 5___numeric___ts_01___4
299
+ 5___numeric___ts_01___5
300
+ 5___numeric___ts_02___5
301
+ 5___numeric___ts_02___8
302
+ 5___numeric___ts_03___4
303
+ 5___numeric___ts_03___6
304
+ 5___numeric___ts_03___8
305
+ 5___numeric___ts_03___9
306
+ 5___numeric___ts_04___0
307
+ 5___numeric___ts_04___1
308
+ 5___numeric___ts_04___2
309
+ 5___numeric___ts_04___3
310
+ 5___numeric___ts_04___4
311
+ 5___numeric___ts_04___5
312
+ 5___numeric___ts_04___6
313
+ 5___numeric___ts_04___7
314
+ 5___numeric___ts_05___0
315
+ 5___numeric___ts_05___1
316
+ 5___numeric___ts_05___2
317
+ 5___numeric___ts_05___3
318
+ 5___numeric___ts_05___4
319
+ 5___numeric___ts_05___5
320
+ 5___numeric___ts_05___6
321
+ 5___numeric___ts_05___7
322
+ 5___numeric___ts_05___8
323
+ 5___numeric___ts_05___9
324
+ 5___numeric___ts_06___0
325
+ 5___numeric___ts_06___1
326
+ 5___numeric___ts_06___2
327
+ 5___numeric___ts_06___3
328
+ 5___numeric___ts_06___4
329
+ 5___numeric___ts_06___5
330
+ 5___numeric___ts_06___6
331
+ 5___numeric___ts_06___7
332
+ 5___numeric___ts_06___8
333
+ 5___numeric___ts_06___9
334
+ 5___numeric___ts_07___0
335
+ 5___numeric___ts_07___1
336
+ 5___numeric___ts_07___2
337
+ 5___numeric___ts_07___3
338
+ 5___numeric___ts_07___4
339
+ 5___numeric___ts_07___5
340
+ 5___numeric___ts_07___6
341
+ 5___numeric___ts_07___7
342
+ 5___numeric___ts_07___8
343
+ 5___numeric___ts_07___9
344
+ 5___numeric___ts_08___0
345
+ 5___numeric___ts_08___1
346
+ 5___numeric___ts_08___2
347
+ 5___numeric___ts_08___3
348
+ 5___numeric___ts_08___4
349
+ 5___numeric___ts_08___5
350
+ 5___numeric___ts_08___6
351
+ 5___numeric___ts_08___7
352
+ 5___numeric___ts_08___8
353
+ 5___numeric___ts_08___9
354
+ 5___numeric___ts_09___0
355
+ 5___numeric___ts_09___1
356
+ 5___numeric___ts_09___2
357
+ 5___numeric___ts_09___3
358
+ 5___numeric___ts_09___4
359
+ 5___numeric___ts_09___5
360
+ 5___numeric___ts_09___6
361
+ 5___numeric___ts_09___7
362
+ 5___numeric___ts_09___8
363
+ 5___numeric___ts_09___9
364
+ 5___numeric___ts_10___0
365
+ 5___numeric___ts_10___1
366
+ 5___numeric___ts_10___2
367
+ 5___numeric___ts_10___3
368
+ 5___numeric___ts_10___4
369
+ 5___numeric___ts_10___5
370
+ 5___numeric___ts_10___6
371
+ 5___numeric___ts_10___7
372
+ 5___numeric___ts_10___8
373
+ 5___numeric___ts_10___9
374
+ 5___numeric___ts_11___0
375
+ 5___numeric___ts_11___1
376
+ 5___numeric___ts_11___2
377
+ 5___numeric___ts_11___3
378
+ 5___numeric___ts_11___4
379
+ 5___numeric___ts_11___5
380
+ 5___numeric___ts_11___6
381
+ 5___numeric___ts_11___7
382
+ 5___numeric___ts_11___8
383
+ 5___numeric___ts_11___9
384
+ 5___numeric___ts_12___0
385
+ 5___numeric___ts_12___1
386
+ 5___numeric___ts_12___2
387
+ 5___numeric___ts_12___3
388
+ 5___numeric___ts_12___4
389
+ 5___numeric___ts_12___5
390
+ 5___numeric___ts_12___6
391
+ 5___numeric___ts_12___7
392
+ 5___numeric___ts_12___8
393
+ 5___numeric___ts_12___9
394
+ 5___numeric___ts_13___0
395
+ 5___numeric___ts_14___0
396
+ 5___numeric___ts_15___0
397
+ 6___numeric___td_00___0
398
+ 6___numeric___td_00___1
399
+ 6___numeric___td_00___2
400
+ 6___numeric___td_00___3
401
+ 6___numeric___td_00___4
402
+ 6___numeric___td_00___8
403
+ 6___numeric___td_00___9
404
+ 6___numeric___td_01___0
405
+ 6___numeric___td_01___1
406
+ 6___numeric___td_01___2
407
+ 6___numeric___td_01___3
408
+ 6___numeric___td_01___4
409
+ 6___numeric___td_01___5
410
+ 6___numeric___td_01___6
411
+ 6___numeric___td_01___7
412
+ 6___numeric___td_01___8
413
+ 6___numeric___td_01___9
414
+ 6___numeric___td_02___0
415
+ 6___numeric___td_02___1
416
+ 6___numeric___td_02___2
417
+ 6___numeric___td_02___3
418
+ 6___numeric___td_02___4
419
+ 6___numeric___td_02___5
420
+ 6___numeric___td_02___6
421
+ 6___numeric___td_02___7
422
+ 6___numeric___td_02___8
423
+ 6___numeric___td_02___9
424
+ 6___numeric___td_03___0
425
+ 6___numeric___td_03___1
426
+ 6___numeric___td_03___2
427
+ 6___numeric___td_03___3
428
+ 6___numeric___td_03___4
429
+ 6___numeric___td_03___5
430
+ 6___numeric___td_03___6
431
+ 6___numeric___td_03___7
432
+ 6___numeric___td_03___8
433
+ 6___numeric___td_03___9
434
+ 6___numeric___td_04___0
435
+ 6___numeric___td_04___1
436
+ 6___numeric___td_04___2
437
+ 6___numeric___td_04___3
438
+ 6___numeric___td_04___4
439
+ 6___numeric___td_04___5
440
+ 6___numeric___td_04___6
441
+ 6___numeric___td_04___7
442
+ 6___numeric___td_04___8
443
+ 6___numeric___td_04___9
444
+ 6___numeric___td_05___.
445
+ 6___numeric___td_06___0
446
+ 6___numeric___td_06___1
447
+ 6___numeric___td_06___2
448
+ 6___numeric___td_06___3
449
+ 6___numeric___td_06___4
450
+ 6___numeric___td_06___5
451
+ 6___numeric___td_06___6
452
+ 6___numeric___td_06___7
453
+ 6___numeric___td_06___8
454
+ 6___numeric___td_06___9
455
+ 6___numeric___td_07___0
456
+ 6___numeric___td_07___1
457
+ 6___numeric___td_07___2
458
+ 6___numeric___td_07___3
459
+ 6___numeric___td_07___4
460
+ 6___numeric___td_07___5
461
+ 6___numeric___td_07___6
462
+ 6___numeric___td_07___7
463
+ 6___numeric___td_07___8
464
+ 6___numeric___td_07___9
465
+ 6___numeric___td_08___0
466
+ 6___numeric___td_08___1
467
+ 6___numeric___td_08___2
468
+ 6___numeric___td_08___3
469
+ 6___numeric___td_08___4
470
+ 6___numeric___td_08___5
471
+ 6___numeric___td_08___6
472
+ 6___numeric___td_08___7
473
+ 6___numeric___td_08___8
474
+ 6___numeric___td_08___9
475
+ 6___numeric___td_09___0
476
+ 6___numeric___td_09___1
477
+ 6___numeric___td_09___2
478
+ 6___numeric___td_09___3
479
+ 6___numeric___td_09___4
480
+ 6___numeric___td_09___5
481
+ 6___numeric___td_09___6
482
+ 6___numeric___td_09___7
483
+ 6___numeric___td_09___8
484
+ 6___numeric___td_09___9
485
+ 7___numeric___pkt_00___0
486
+ 7___numeric___pkt_00___1
487
+ 7___numeric___pkt_00___2
488
+ 7___numeric___pkt_00___3
489
+ 7___numeric___pkt_01___0
490
+ 7___numeric___pkt_01___1
491
+ 7___numeric___pkt_01___2
492
+ 7___numeric___pkt_01___3
493
+ 7___numeric___pkt_01___4
494
+ 7___numeric___pkt_01___5
495
+ 7___numeric___pkt_01___6
496
+ 7___numeric___pkt_01___7
497
+ 7___numeric___pkt_01___8
498
+ 7___numeric___pkt_01___9
499
+ 7___numeric___pkt_02___0
500
+ 7___numeric___pkt_02___1
501
+ 7___numeric___pkt_02___2
502
+ 7___numeric___pkt_02___3
503
+ 7___numeric___pkt_02___4
504
+ 7___numeric___pkt_02___5
505
+ 7___numeric___pkt_02___6
506
+ 7___numeric___pkt_02___7
507
+ 7___numeric___pkt_02___8
508
+ 7___numeric___pkt_02___9
509
+ 7___numeric___pkt_03___0
510
+ 7___numeric___pkt_03___1
511
+ 7___numeric___pkt_03___2
512
+ 7___numeric___pkt_03___3
513
+ 7___numeric___pkt_03___4
514
+ 7___numeric___pkt_03___5
515
+ 7___numeric___pkt_03___6
516
+ 7___numeric___pkt_03___7
517
+ 7___numeric___pkt_03___8
518
+ 7___numeric___pkt_03___9
519
+ 7___numeric___pkt_04___0
520
+ 7___numeric___pkt_04___1
521
+ 7___numeric___pkt_04___2
522
+ 7___numeric___pkt_04___3
523
+ 7___numeric___pkt_04___4
524
+ 7___numeric___pkt_04___5
525
+ 7___numeric___pkt_04___6
526
+ 7___numeric___pkt_04___7
527
+ 7___numeric___pkt_04___8
528
+ 7___numeric___pkt_04___9
529
+ 7___numeric___pkt_05___0
530
+ 7___numeric___pkt_05___1
531
+ 7___numeric___pkt_05___2
532
+ 7___numeric___pkt_05___3
533
+ 7___numeric___pkt_05___4
534
+ 7___numeric___pkt_05___5
535
+ 7___numeric___pkt_05___6
536
+ 7___numeric___pkt_05___7
537
+ 7___numeric___pkt_05___8
538
+ 7___numeric___pkt_05___9
539
+ 8___numeric___byt_00___0
540
+ 8___numeric___byt_00___1
541
+ 8___numeric___byt_00___2
542
+ 8___numeric___byt_00___3
543
+ 8___numeric___byt_01___0
544
+ 8___numeric___byt_01___1
545
+ 8___numeric___byt_01___2
546
+ 8___numeric___byt_01___3
547
+ 8___numeric___byt_01___4
548
+ 8___numeric___byt_01___5
549
+ 8___numeric___byt_01___6
550
+ 8___numeric___byt_01___7
551
+ 8___numeric___byt_01___8
552
+ 8___numeric___byt_01___9
553
+ 8___numeric___byt_02___0
554
+ 8___numeric___byt_02___1
555
+ 8___numeric___byt_02___2
556
+ 8___numeric___byt_02___3
557
+ 8___numeric___byt_02___4
558
+ 8___numeric___byt_02___5
559
+ 8___numeric___byt_02___6
560
+ 8___numeric___byt_02___7
561
+ 8___numeric___byt_02___8
562
+ 8___numeric___byt_02___9
563
+ 8___numeric___byt_03___0
564
+ 8___numeric___byt_03___1
565
+ 8___numeric___byt_03___2
566
+ 8___numeric___byt_03___3
567
+ 8___numeric___byt_03___4
568
+ 8___numeric___byt_03___5
569
+ 8___numeric___byt_03___6
570
+ 8___numeric___byt_03___7
571
+ 8___numeric___byt_03___8
572
+ 8___numeric___byt_03___9
573
+ 8___numeric___byt_04___0
574
+ 8___numeric___byt_04___1
575
+ 8___numeric___byt_04___2
576
+ 8___numeric___byt_04___3
577
+ 8___numeric___byt_04___4
578
+ 8___numeric___byt_04___5
579
+ 8___numeric___byt_04___6
580
+ 8___numeric___byt_04___7
581
+ 8___numeric___byt_04___8
582
+ 8___numeric___byt_04___9
583
+ 8___numeric___byt_05___0
584
+ 8___numeric___byt_05___1
585
+ 8___numeric___byt_05___2
586
+ 8___numeric___byt_05___3
587
+ 8___numeric___byt_05___4
588
+ 8___numeric___byt_05___5
589
+ 8___numeric___byt_05___6
590
+ 8___numeric___byt_05___7
591
+ 8___numeric___byt_05___8
592
+ 8___numeric___byt_05___9
593
+ 8___numeric___byt_06___0
594
+ 8___numeric___byt_06___1
595
+ 8___numeric___byt_06___2
596
+ 8___numeric___byt_06___3
597
+ 8___numeric___byt_06___4
598
+ 8___numeric___byt_06___5
599
+ 8___numeric___byt_06___6
600
+ 8___numeric___byt_06___7
601
+ 8___numeric___byt_06___8
602
+ 8___numeric___byt_06___9
603
+ 8___numeric___byt_07___0
604
+ 8___numeric___byt_07___1
605
+ 8___numeric___byt_07___2
606
+ 8___numeric___byt_07___3
607
+ 8___numeric___byt_07___4
608
+ 8___numeric___byt_07___5
609
+ 8___numeric___byt_07___6
610
+ 8___numeric___byt_07___7
611
+ 8___numeric___byt_07___8
612
+ 8___numeric___byt_07___9
613
+ 8___numeric___byt_08___0
614
+ 8___numeric___byt_08___1
615
+ 8___numeric___byt_08___2
616
+ 8___numeric___byt_08___3
617
+ 8___numeric___byt_08___4
618
+ 8___numeric___byt_08___5
619
+ 8___numeric___byt_08___6
620
+ 8___numeric___byt_08___7
621
+ 8___numeric___byt_08___8
622
+ 8___numeric___byt_08___9
623
+ 8___numeric___byt_09___0
624
+ 8___numeric___byt_09___1
625
+ 8___numeric___byt_09___2
626
+ 8___numeric___byt_09___3
627
+ 8___numeric___byt_09___4
628
+ 8___numeric___byt_09___5
629
+ 8___numeric___byt_09___6
630
+ 8___numeric___byt_09___7
631
+ 8___numeric___byt_09___8
632
+ 8___numeric___byt_09___9