DevWithKaiju commited on
Commit
0b6bc0b
·
verified ·
1 Parent(s): 3e35c89

Upload folder using huggingface_hub

Browse files
added_tokens.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "</E0>": 30525,
3
+ "</E1>": 30523,
4
+ "<E0>": 30524,
5
+ "<E1>": 30522
6
+ }
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 3072,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "transformers_version": "4.57.1",
21
+ "type_vocab_size": 2,
22
+ "use_cache": true,
23
+ "vocab_size": 30526
24
+ }
final_training_metrics.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss,epoch,final_step
2
+ 18012.8758,223.735,0.437,2.65186362613675e+17,1.1811757300927388,2.0,7872
full_log_history.csv ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ loss,grad_norm,learning_rate,epoch,step,eval_loss,eval_runtime,eval_samples_per_second,eval_steps_per_second,train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss
2
+ 1.4748,3.892392873764038,4.949238578680203e-06,0.05081623578733405,200,,,,,,,,,
3
+ ,,,0.05081623578733405,200,1.3006737232208252,270.1193,392.626,49.078,,,,,
4
+ 1.3275,2.6811366081237793,1.00253807106599e-05,0.1016324715746681,400,,,,,,,,,
5
+ ,,,0.1016324715746681,400,1.2096028327941895,340.2363,311.713,38.964,,,,,
6
+ 1.2631,5.332873344421387,1.5076142131979696e-05,0.15244870736200217,600,,,,,,,,,
7
+ ,,,0.15244870736200217,600,1.1663851737976074,268.8282,394.512,49.314,,,,,
8
+ 1.2354,2.641127824783325,1.9985883681535857e-05,0.2032649431493362,800,,,,,,,,,
9
+ ,,,0.2032649431493362,800,1.1503351926803589,267.5514,396.395,49.549,,,,,
10
+ 1.2215,2.6785213947296143,1.9421230942970076e-05,0.25408117893667026,1000,,,,,,,,,
11
+ ,,,0.25408117893667026,1000,1.138418197631836,265.8454,398.939,49.867,,,,,
12
+ 1.2082,2.559088706970215,1.885657820440429e-05,0.30489741472400433,1200,,,,,,,,,
13
+ ,,,0.30489741472400433,1200,1.1333622932434082,269.9382,392.89,49.111,,,,,
14
+ 1.2057,2.575900077819824,1.829192546583851e-05,0.35571365051133835,1400,,,,,,,,,
15
+ ,,,0.35571365051133835,1400,1.1210781335830688,272.6069,389.044,48.63,,,,,
16
+ 1.1968,3.607745885848999,1.772727272727273e-05,0.4065298862986724,1600,,,,,,,,,
17
+ ,,,0.4065298862986724,1600,1.1209676265716553,270.4455,392.153,49.019,,,,,
18
+ 1.1889,2.495760679244995,1.7162619988706945e-05,0.4573461220860065,1800,,,,,,,,,
19
+ ,,,0.4573461220860065,1800,1.1131852865219116,271.3105,390.903,48.863,,,,,
20
+ 1.1871,2.44140625,1.6597967250141165e-05,0.5081623578733405,2000,,,,,,,,,
21
+ ,,,0.5081623578733405,2000,1.1079550981521606,270.6488,391.858,48.982,,,,,
22
+ 1.1909,2.5054123401641846,1.603331451157538e-05,0.5589785936606746,2200,,,,,,,,,
23
+ ,,,0.5589785936606746,2200,1.1088285446166992,269.7351,393.186,49.148,,,,,
24
+ 1.1849,2.444836378097534,1.5468661773009603e-05,0.6097948294480087,2400,,,,,,,,,
25
+ ,,,0.6097948294480087,2400,1.104771614074707,273.2367,388.147,48.518,,,,,
26
+ 1.1801,2.4025747776031494,1.4904009034443818e-05,0.6606110652353427,2600,,,,,,,,,
27
+ ,,,0.6606110652353427,2600,1.103324294090271,271.9533,389.979,48.747,,,,,
28
+ 1.1795,2.442716360092163,1.4339356295878036e-05,0.7114273010226767,2800,,,,,,,,,
29
+ ,,,0.7114273010226767,2800,1.0979377031326294,273.1522,388.267,48.533,,,,,
30
+ 1.1746,2.3968005180358887,1.3774703557312255e-05,0.7622435368100108,3000,,,,,,,,,
31
+ ,,,0.7622435368100108,3000,1.1022906303405762,272.0233,389.878,48.735,,,,,
32
+ 1.1688,2.5648298263549805,1.321005081874647e-05,0.8130597725973449,3200,,,,,,,,,
33
+ ,,,0.8130597725973449,3200,1.0999451875686646,276.6514,383.356,47.92,,,,,
34
+ 1.1705,2.456953525543213,1.2645398080180692e-05,0.8638760083846789,3400,,,,,,,,,
35
+ ,,,0.8638760083846789,3400,1.0966227054595947,277.5135,382.165,47.771,,,,,
36
+ 1.1642,2.4593939781188965,1.2080745341614907e-05,0.914692244172013,3600,,,,,,,,,
37
+ ,,,0.914692244172013,3600,1.0930379629135132,273.8492,387.279,48.41,,,,,
38
+ 1.1643,2.5122382640838623,1.1516092603049125e-05,0.965508479959347,3800,,,,,,,,,
39
+ ,,,0.965508479959347,3800,1.097168207168579,271.5387,390.574,48.822,,,,,
40
+ 1.159,2.538911819458008,1.0951439864483344e-05,1.016261195451947,4000,,,,,,,,,
41
+ ,,,1.016261195451947,4000,1.0909498929977417,272.1472,389.701,48.713,,,,,
42
+ 1.1572,2.3571293354034424,1.0386787125917561e-05,1.067077431239281,4200,,,,,,,,,
43
+ ,,,1.067077431239281,4200,1.0871597528457642,277.8919,381.645,47.706,,,,,
44
+ 1.1621,2.4962565898895264,9.822134387351779e-06,1.117893667026615,4400,,,,,,,,,
45
+ ,,,1.117893667026615,4400,1.0853313207626343,273.8412,387.29,48.411,,,,,
46
+ 1.1582,2.1864099502563477,9.257481648785998e-06,1.1687099028139492,4600,,,,,,,,,
47
+ ,,,1.1687099028139492,4600,1.0870981216430664,276.7667,383.196,47.9,,,,,
48
+ 1.1553,2.546422004699707,8.692828910220215e-06,1.219526138601283,4800,,,,,,,,,
49
+ ,,,1.219526138601283,4800,1.0885721445083618,318.9661,332.499,41.562,,,,,
50
+ 1.1574,2.236259698867798,8.128176171654432e-06,1.2703423743886173,5000,,,,,,,,,
51
+ ,,,1.2703423743886173,5000,1.0829311609268188,281.8375,376.302,47.038,,,,,
52
+ 1.1556,2.4067327976226807,7.563523433088651e-06,1.3211586101759512,5200,,,,,,,,,
53
+ ,,,1.3211586101759512,5200,1.0802139043807983,274.8244,385.905,48.238,,,,,
54
+ 1.1538,2.381182909011841,6.998870694522869e-06,1.3719748459632852,5400,,,,,,,,,
55
+ ,,,1.3719748459632852,5400,1.0838501453399658,274.8546,385.862,48.233,,,,,
56
+ 1.1519,2.3420238494873047,6.434217955957087e-06,1.4227910817506193,5600,,,,,,,,,
57
+ ,,,1.4227910817506193,5600,1.0835458040237427,277.0452,382.811,47.851,,,,,
58
+ 1.1483,2.427466630935669,5.8695652173913055e-06,1.4736073175379534,5800,,,,,,,,,
59
+ ,,,1.4736073175379534,5800,1.0747573375701904,276.251,383.912,47.989,,,,,
60
+ 1.1487,2.3989744186401367,5.304912478825522e-06,1.5244235533252875,6000,,,,,,,,,
61
+ ,,,1.5244235533252875,6000,1.0760985612869263,278.4296,380.908,47.613,,,,,
62
+ 1.1462,2.3256778717041016,4.74025974025974e-06,1.5752397891126215,6200,,,,,,,,,
63
+ ,,,1.5752397891126215,6200,1.0776214599609375,282.9517,374.82,46.853,,,,,
64
+ 1.1453,2.3638393878936768,4.175607001693959e-06,1.6260560248999556,6400,,,,,,,,,
65
+ ,,,1.6260560248999556,6400,1.0771236419677734,274.7672,385.985,48.248,,,,,
66
+ 1.1434,2.4487802982330322,3.6109542631281764e-06,1.6768722606872895,6600,,,,,,,,,
67
+ ,,,1.6768722606872895,6600,1.078606367111206,278.7283,380.5,47.562,,,,,
68
+ 1.146,2.6967720985412598,3.0463015245623943e-06,1.7276884964746237,6800,,,,,,,,,
69
+ ,,,1.7276884964746237,6800,1.0735256671905518,277.6313,382.003,47.75,,,,,
70
+ 1.1418,2.5152363777160645,2.4816487859966125e-06,1.7785047322619576,7000,,,,,,,,,
71
+ ,,,1.7785047322619576,7000,1.074223518371582,281.0075,377.413,47.177,,,,,
72
+ 1.1409,2.43704891204834,1.9169960474308304e-06,1.829320968049292,7200,,,,,,,,,
73
+ ,,,1.829320968049292,7200,1.0686383247375488,276.8469,383.085,47.886,,,,,
74
+ 1.1381,2.35129976272583,1.3523433088650482e-06,1.8801372038366257,7400,,,,,,,,,
75
+ ,,,1.8801372038366257,7400,1.072161316871643,277.7545,381.834,47.729,,,,,
76
+ 1.1408,2.4860081672668457,7.876905702992659e-07,1.9309534396239598,7600,,,,,,,,,
77
+ ,,,1.9309534396239598,7600,1.07365083694458,280.9403,377.504,47.188,,,,,
78
+ 1.142,2.340489625930786,2.2303783173348393e-07,1.981769675411294,7800,,,,,,,,,
79
+ ,,,1.981769675411294,7800,1.0732721090316772,282.0926,375.962,46.995,,,,,
80
+ ,,,2.0,7872,,,,,18012.8758,223.735,0.437,2.65186362613675e+17,1.1811757300927388
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04a2175f9e06425ba28a3f2b572858211b1211521baa13aafe774dec2dd78cb2
3
+ size 438093200
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<E1>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "</E1>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<E0>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "</E0>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ ],
32
+ "cls_token": "[CLS]",
33
+ "mask_token": "[MASK]",
34
+ "pad_token": "[PAD]",
35
+ "sep_token": "[SEP]",
36
+ "unk_token": "[UNK]"
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "30522": {
44
+ "content": "<E1>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "30523": {
52
+ "content": "</E1>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "30524": {
60
+ "content": "<E0>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "30525": {
68
+ "content": "</E0>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ }
75
+ },
76
+ "additional_special_tokens": [
77
+ "<E1>",
78
+ "</E1>",
79
+ "<E0>",
80
+ "</E0>"
81
+ ],
82
+ "clean_up_tokenization_spaces": true,
83
+ "cls_token": "[CLS]",
84
+ "do_basic_tokenize": true,
85
+ "do_lower_case": true,
86
+ "extra_special_tokens": {},
87
+ "mask_token": "[MASK]",
88
+ "model_max_length": 1000000000000000019884624838656,
89
+ "never_split": null,
90
+ "pad_token": "[PAD]",
91
+ "sep_token": "[SEP]",
92
+ "strip_accents": null,
93
+ "tokenize_chinese_chars": true,
94
+ "tokenizer_class": "BertTokenizer",
95
+ "unk_token": "[UNK]"
96
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6dabb05400d01da28beaf587a8ef31651af41b31049654521209414ae3d414f
3
+ size 5841
vocab.txt ADDED
The diff for this file is too large to render. See raw diff