Upload folder using huggingface_hub
Browse files- added_tokens.json +6 -0
- config.json +24 -0
- final_training_metrics.csv +2 -0
- full_log_history.csv +80 -0
- model.safetensors +3 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +96 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
added_tokens.json
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"</E0>": 30525,
|
| 3 |
+
"</E1>": 30523,
|
| 4 |
+
"<E0>": 30524,
|
| 5 |
+
"<E1>": 30522
|
| 6 |
+
}
|
config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertForMaskedLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"dtype": "float32",
|
| 8 |
+
"hidden_act": "gelu",
|
| 9 |
+
"hidden_dropout_prob": 0.1,
|
| 10 |
+
"hidden_size": 768,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 3072,
|
| 13 |
+
"layer_norm_eps": 1e-12,
|
| 14 |
+
"max_position_embeddings": 512,
|
| 15 |
+
"model_type": "bert",
|
| 16 |
+
"num_attention_heads": 12,
|
| 17 |
+
"num_hidden_layers": 12,
|
| 18 |
+
"pad_token_id": 0,
|
| 19 |
+
"position_embedding_type": "absolute",
|
| 20 |
+
"transformers_version": "4.57.1",
|
| 21 |
+
"type_vocab_size": 2,
|
| 22 |
+
"use_cache": true,
|
| 23 |
+
"vocab_size": 30526
|
| 24 |
+
}
|
final_training_metrics.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss,epoch,final_step
|
| 2 |
+
18012.8758,223.735,0.437,2.65186362613675e+17,1.1811757300927388,2.0,7872
|
full_log_history.csv
ADDED
|
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
loss,grad_norm,learning_rate,epoch,step,eval_loss,eval_runtime,eval_samples_per_second,eval_steps_per_second,train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss
|
| 2 |
+
1.4748,3.892392873764038,4.949238578680203e-06,0.05081623578733405,200,,,,,,,,,
|
| 3 |
+
,,,0.05081623578733405,200,1.3006737232208252,270.1193,392.626,49.078,,,,,
|
| 4 |
+
1.3275,2.6811366081237793,1.00253807106599e-05,0.1016324715746681,400,,,,,,,,,
|
| 5 |
+
,,,0.1016324715746681,400,1.2096028327941895,340.2363,311.713,38.964,,,,,
|
| 6 |
+
1.2631,5.332873344421387,1.5076142131979696e-05,0.15244870736200217,600,,,,,,,,,
|
| 7 |
+
,,,0.15244870736200217,600,1.1663851737976074,268.8282,394.512,49.314,,,,,
|
| 8 |
+
1.2354,2.641127824783325,1.9985883681535857e-05,0.2032649431493362,800,,,,,,,,,
|
| 9 |
+
,,,0.2032649431493362,800,1.1503351926803589,267.5514,396.395,49.549,,,,,
|
| 10 |
+
1.2215,2.6785213947296143,1.9421230942970076e-05,0.25408117893667026,1000,,,,,,,,,
|
| 11 |
+
,,,0.25408117893667026,1000,1.138418197631836,265.8454,398.939,49.867,,,,,
|
| 12 |
+
1.2082,2.559088706970215,1.885657820440429e-05,0.30489741472400433,1200,,,,,,,,,
|
| 13 |
+
,,,0.30489741472400433,1200,1.1333622932434082,269.9382,392.89,49.111,,,,,
|
| 14 |
+
1.2057,2.575900077819824,1.829192546583851e-05,0.35571365051133835,1400,,,,,,,,,
|
| 15 |
+
,,,0.35571365051133835,1400,1.1210781335830688,272.6069,389.044,48.63,,,,,
|
| 16 |
+
1.1968,3.607745885848999,1.772727272727273e-05,0.4065298862986724,1600,,,,,,,,,
|
| 17 |
+
,,,0.4065298862986724,1600,1.1209676265716553,270.4455,392.153,49.019,,,,,
|
| 18 |
+
1.1889,2.495760679244995,1.7162619988706945e-05,0.4573461220860065,1800,,,,,,,,,
|
| 19 |
+
,,,0.4573461220860065,1800,1.1131852865219116,271.3105,390.903,48.863,,,,,
|
| 20 |
+
1.1871,2.44140625,1.6597967250141165e-05,0.5081623578733405,2000,,,,,,,,,
|
| 21 |
+
,,,0.5081623578733405,2000,1.1079550981521606,270.6488,391.858,48.982,,,,,
|
| 22 |
+
1.1909,2.5054123401641846,1.603331451157538e-05,0.5589785936606746,2200,,,,,,,,,
|
| 23 |
+
,,,0.5589785936606746,2200,1.1088285446166992,269.7351,393.186,49.148,,,,,
|
| 24 |
+
1.1849,2.444836378097534,1.5468661773009603e-05,0.6097948294480087,2400,,,,,,,,,
|
| 25 |
+
,,,0.6097948294480087,2400,1.104771614074707,273.2367,388.147,48.518,,,,,
|
| 26 |
+
1.1801,2.4025747776031494,1.4904009034443818e-05,0.6606110652353427,2600,,,,,,,,,
|
| 27 |
+
,,,0.6606110652353427,2600,1.103324294090271,271.9533,389.979,48.747,,,,,
|
| 28 |
+
1.1795,2.442716360092163,1.4339356295878036e-05,0.7114273010226767,2800,,,,,,,,,
|
| 29 |
+
,,,0.7114273010226767,2800,1.0979377031326294,273.1522,388.267,48.533,,,,,
|
| 30 |
+
1.1746,2.3968005180358887,1.3774703557312255e-05,0.7622435368100108,3000,,,,,,,,,
|
| 31 |
+
,,,0.7622435368100108,3000,1.1022906303405762,272.0233,389.878,48.735,,,,,
|
| 32 |
+
1.1688,2.5648298263549805,1.321005081874647e-05,0.8130597725973449,3200,,,,,,,,,
|
| 33 |
+
,,,0.8130597725973449,3200,1.0999451875686646,276.6514,383.356,47.92,,,,,
|
| 34 |
+
1.1705,2.456953525543213,1.2645398080180692e-05,0.8638760083846789,3400,,,,,,,,,
|
| 35 |
+
,,,0.8638760083846789,3400,1.0966227054595947,277.5135,382.165,47.771,,,,,
|
| 36 |
+
1.1642,2.4593939781188965,1.2080745341614907e-05,0.914692244172013,3600,,,,,,,,,
|
| 37 |
+
,,,0.914692244172013,3600,1.0930379629135132,273.8492,387.279,48.41,,,,,
|
| 38 |
+
1.1643,2.5122382640838623,1.1516092603049125e-05,0.965508479959347,3800,,,,,,,,,
|
| 39 |
+
,,,0.965508479959347,3800,1.097168207168579,271.5387,390.574,48.822,,,,,
|
| 40 |
+
1.159,2.538911819458008,1.0951439864483344e-05,1.016261195451947,4000,,,,,,,,,
|
| 41 |
+
,,,1.016261195451947,4000,1.0909498929977417,272.1472,389.701,48.713,,,,,
|
| 42 |
+
1.1572,2.3571293354034424,1.0386787125917561e-05,1.067077431239281,4200,,,,,,,,,
|
| 43 |
+
,,,1.067077431239281,4200,1.0871597528457642,277.8919,381.645,47.706,,,,,
|
| 44 |
+
1.1621,2.4962565898895264,9.822134387351779e-06,1.117893667026615,4400,,,,,,,,,
|
| 45 |
+
,,,1.117893667026615,4400,1.0853313207626343,273.8412,387.29,48.411,,,,,
|
| 46 |
+
1.1582,2.1864099502563477,9.257481648785998e-06,1.1687099028139492,4600,,,,,,,,,
|
| 47 |
+
,,,1.1687099028139492,4600,1.0870981216430664,276.7667,383.196,47.9,,,,,
|
| 48 |
+
1.1553,2.546422004699707,8.692828910220215e-06,1.219526138601283,4800,,,,,,,,,
|
| 49 |
+
,,,1.219526138601283,4800,1.0885721445083618,318.9661,332.499,41.562,,,,,
|
| 50 |
+
1.1574,2.236259698867798,8.128176171654432e-06,1.2703423743886173,5000,,,,,,,,,
|
| 51 |
+
,,,1.2703423743886173,5000,1.0829311609268188,281.8375,376.302,47.038,,,,,
|
| 52 |
+
1.1556,2.4067327976226807,7.563523433088651e-06,1.3211586101759512,5200,,,,,,,,,
|
| 53 |
+
,,,1.3211586101759512,5200,1.0802139043807983,274.8244,385.905,48.238,,,,,
|
| 54 |
+
1.1538,2.381182909011841,6.998870694522869e-06,1.3719748459632852,5400,,,,,,,,,
|
| 55 |
+
,,,1.3719748459632852,5400,1.0838501453399658,274.8546,385.862,48.233,,,,,
|
| 56 |
+
1.1519,2.3420238494873047,6.434217955957087e-06,1.4227910817506193,5600,,,,,,,,,
|
| 57 |
+
,,,1.4227910817506193,5600,1.0835458040237427,277.0452,382.811,47.851,,,,,
|
| 58 |
+
1.1483,2.427466630935669,5.8695652173913055e-06,1.4736073175379534,5800,,,,,,,,,
|
| 59 |
+
,,,1.4736073175379534,5800,1.0747573375701904,276.251,383.912,47.989,,,,,
|
| 60 |
+
1.1487,2.3989744186401367,5.304912478825522e-06,1.5244235533252875,6000,,,,,,,,,
|
| 61 |
+
,,,1.5244235533252875,6000,1.0760985612869263,278.4296,380.908,47.613,,,,,
|
| 62 |
+
1.1462,2.3256778717041016,4.74025974025974e-06,1.5752397891126215,6200,,,,,,,,,
|
| 63 |
+
,,,1.5752397891126215,6200,1.0776214599609375,282.9517,374.82,46.853,,,,,
|
| 64 |
+
1.1453,2.3638393878936768,4.175607001693959e-06,1.6260560248999556,6400,,,,,,,,,
|
| 65 |
+
,,,1.6260560248999556,6400,1.0771236419677734,274.7672,385.985,48.248,,,,,
|
| 66 |
+
1.1434,2.4487802982330322,3.6109542631281764e-06,1.6768722606872895,6600,,,,,,,,,
|
| 67 |
+
,,,1.6768722606872895,6600,1.078606367111206,278.7283,380.5,47.562,,,,,
|
| 68 |
+
1.146,2.6967720985412598,3.0463015245623943e-06,1.7276884964746237,6800,,,,,,,,,
|
| 69 |
+
,,,1.7276884964746237,6800,1.0735256671905518,277.6313,382.003,47.75,,,,,
|
| 70 |
+
1.1418,2.5152363777160645,2.4816487859966125e-06,1.7785047322619576,7000,,,,,,,,,
|
| 71 |
+
,,,1.7785047322619576,7000,1.074223518371582,281.0075,377.413,47.177,,,,,
|
| 72 |
+
1.1409,2.43704891204834,1.9169960474308304e-06,1.829320968049292,7200,,,,,,,,,
|
| 73 |
+
,,,1.829320968049292,7200,1.0686383247375488,276.8469,383.085,47.886,,,,,
|
| 74 |
+
1.1381,2.35129976272583,1.3523433088650482e-06,1.8801372038366257,7400,,,,,,,,,
|
| 75 |
+
,,,1.8801372038366257,7400,1.072161316871643,277.7545,381.834,47.729,,,,,
|
| 76 |
+
1.1408,2.4860081672668457,7.876905702992659e-07,1.9309534396239598,7600,,,,,,,,,
|
| 77 |
+
,,,1.9309534396239598,7600,1.07365083694458,280.9403,377.504,47.188,,,,,
|
| 78 |
+
1.142,2.340489625930786,2.2303783173348393e-07,1.981769675411294,7800,,,,,,,,,
|
| 79 |
+
,,,1.981769675411294,7800,1.0732721090316772,282.0926,375.962,46.995,,,,,
|
| 80 |
+
,,,2.0,7872,,,,,18012.8758,223.735,0.437,2.65186362613675e+17,1.1811757300927388
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04a2175f9e06425ba28a3f2b572858211b1211521baa13aafe774dec2dd78cb2
|
| 3 |
+
size 438093200
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
{
|
| 4 |
+
"content": "<E1>",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false
|
| 9 |
+
},
|
| 10 |
+
{
|
| 11 |
+
"content": "</E1>",
|
| 12 |
+
"lstrip": false,
|
| 13 |
+
"normalized": false,
|
| 14 |
+
"rstrip": false,
|
| 15 |
+
"single_word": false
|
| 16 |
+
},
|
| 17 |
+
{
|
| 18 |
+
"content": "<E0>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
{
|
| 25 |
+
"content": "</E0>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
],
|
| 32 |
+
"cls_token": "[CLS]",
|
| 33 |
+
"mask_token": "[MASK]",
|
| 34 |
+
"pad_token": "[PAD]",
|
| 35 |
+
"sep_token": "[SEP]",
|
| 36 |
+
"unk_token": "[UNK]"
|
| 37 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"1": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"2": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"3": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"4": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
},
|
| 43 |
+
"30522": {
|
| 44 |
+
"content": "<E1>",
|
| 45 |
+
"lstrip": false,
|
| 46 |
+
"normalized": false,
|
| 47 |
+
"rstrip": false,
|
| 48 |
+
"single_word": false,
|
| 49 |
+
"special": true
|
| 50 |
+
},
|
| 51 |
+
"30523": {
|
| 52 |
+
"content": "</E1>",
|
| 53 |
+
"lstrip": false,
|
| 54 |
+
"normalized": false,
|
| 55 |
+
"rstrip": false,
|
| 56 |
+
"single_word": false,
|
| 57 |
+
"special": true
|
| 58 |
+
},
|
| 59 |
+
"30524": {
|
| 60 |
+
"content": "<E0>",
|
| 61 |
+
"lstrip": false,
|
| 62 |
+
"normalized": false,
|
| 63 |
+
"rstrip": false,
|
| 64 |
+
"single_word": false,
|
| 65 |
+
"special": true
|
| 66 |
+
},
|
| 67 |
+
"30525": {
|
| 68 |
+
"content": "</E0>",
|
| 69 |
+
"lstrip": false,
|
| 70 |
+
"normalized": false,
|
| 71 |
+
"rstrip": false,
|
| 72 |
+
"single_word": false,
|
| 73 |
+
"special": true
|
| 74 |
+
}
|
| 75 |
+
},
|
| 76 |
+
"additional_special_tokens": [
|
| 77 |
+
"<E1>",
|
| 78 |
+
"</E1>",
|
| 79 |
+
"<E0>",
|
| 80 |
+
"</E0>"
|
| 81 |
+
],
|
| 82 |
+
"clean_up_tokenization_spaces": true,
|
| 83 |
+
"cls_token": "[CLS]",
|
| 84 |
+
"do_basic_tokenize": true,
|
| 85 |
+
"do_lower_case": true,
|
| 86 |
+
"extra_special_tokens": {},
|
| 87 |
+
"mask_token": "[MASK]",
|
| 88 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 89 |
+
"never_split": null,
|
| 90 |
+
"pad_token": "[PAD]",
|
| 91 |
+
"sep_token": "[SEP]",
|
| 92 |
+
"strip_accents": null,
|
| 93 |
+
"tokenize_chinese_chars": true,
|
| 94 |
+
"tokenizer_class": "BertTokenizer",
|
| 95 |
+
"unk_token": "[UNK]"
|
| 96 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6dabb05400d01da28beaf587a8ef31651af41b31049654521209414ae3d414f
|
| 3 |
+
size 5841
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|