Upload folder using huggingface_hub
Browse files- added_tokens.json +4 -0
- config.json +24 -0
- final_training_metrics.csv +2 -0
- full_log_history.csv +80 -0
- model.safetensors +3 -0
- special_tokens_map.json +7 -0
- tokenizer.json +0 -0
- tokenizer_config.json +74 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"[CELL0]": 30522,
|
| 3 |
+
"[CELL1]": 30523
|
| 4 |
+
}
|
config.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertForMaskedLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"classifier_dropout": null,
|
| 7 |
+
"dtype": "float32",
|
| 8 |
+
"hidden_act": "gelu",
|
| 9 |
+
"hidden_dropout_prob": 0.1,
|
| 10 |
+
"hidden_size": 768,
|
| 11 |
+
"initializer_range": 0.02,
|
| 12 |
+
"intermediate_size": 3072,
|
| 13 |
+
"layer_norm_eps": 1e-12,
|
| 14 |
+
"max_position_embeddings": 512,
|
| 15 |
+
"model_type": "bert",
|
| 16 |
+
"num_attention_heads": 12,
|
| 17 |
+
"num_hidden_layers": 12,
|
| 18 |
+
"pad_token_id": 0,
|
| 19 |
+
"position_embedding_type": "absolute",
|
| 20 |
+
"transformers_version": "4.57.1",
|
| 21 |
+
"type_vocab_size": 2,
|
| 22 |
+
"use_cache": true,
|
| 23 |
+
"vocab_size": 30524
|
| 24 |
+
}
|
final_training_metrics.csv
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss,epoch,final_step
|
| 2 |
+
20458.6452,196.988,0.385,2.651863564234199e+17,1.2797311825480888,2.0,7872
|
full_log_history.csv
ADDED
|
@@ -0,0 +1,80 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
loss,grad_norm,learning_rate,epoch,step,eval_loss,eval_runtime,eval_samples_per_second,eval_steps_per_second,train_runtime,train_samples_per_second,train_steps_per_second,total_flos,train_loss
|
| 2 |
+
1.8825,3.3912692070007324,4.974619289340102e-06,0.05081623578733405,200,,,,,,,,,
|
| 3 |
+
,,,0.05081623578733405,200,1.7075451612472534,287.0628,369.452,46.182,,,,,
|
| 4 |
+
1.6809,2.406120538711548,1.0050761421319797e-05,0.1016324715746681,400,,,,,,,,,
|
| 5 |
+
,,,0.1016324715746681,400,1.5343916416168213,293.7654,361.023,45.128,,,,,
|
| 6 |
+
1.5724,2.4685757160186768,1.5126903553299494e-05,0.15244870736200217,600,,,,,,,,,
|
| 7 |
+
,,,0.15244870736200217,600,1.4772238731384277,307.2416,345.188,43.148,,,,,
|
| 8 |
+
1.5349,2.4002857208251953,1.997741389045737e-05,0.2032649431493362,800,,,,,,,,,
|
| 9 |
+
,,,0.2032649431493362,800,1.4481595754623413,308.5478,343.726,42.966,,,,,
|
| 10 |
+
1.5118,2.6639628410339355,1.941276115189159e-05,0.25408117893667026,1000,,,,,,,,,
|
| 11 |
+
,,,0.25408117893667026,1000,1.4235949516296387,309.1643,343.041,42.88,,,,,
|
| 12 |
+
1.4773,2.6258928775787354,1.8848108413325805e-05,0.30489741472400433,1200,,,,,,,,,
|
| 13 |
+
,,,0.30489741472400433,1200,1.4003123044967651,310.8481,341.183,42.648,,,,,
|
| 14 |
+
1.4533,2.444345712661743,1.8283455674760024e-05,0.35571365051133835,1400,,,,,,,,,
|
| 15 |
+
,,,0.35571365051133835,1400,1.3662773370742798,308.9457,343.284,42.91,,,,,
|
| 16 |
+
1.4044,2.4333863258361816,1.771880293619424e-05,0.4065298862986724,1600,,,,,,,,,
|
| 17 |
+
,,,0.4065298862986724,1600,1.3160090446472168,313.1739,338.649,42.331,,,,,
|
| 18 |
+
1.327,2.5270566940307617,1.715415019762846e-05,0.4573461220860065,1800,,,,,,,,,
|
| 19 |
+
,,,0.4573461220860065,1800,1.2360590696334839,323.1804,328.163,41.02,,,,,
|
| 20 |
+
1.2685,2.7455523014068604,1.6589497459062678e-05,0.5081623578733405,2000,,,,,,,,,
|
| 21 |
+
,,,0.5081623578733405,2000,1.1782482862472534,324.552,326.777,40.847,,,,,
|
| 22 |
+
1.2529,2.441678285598755,1.6024844720496894e-05,0.5589785936606746,2200,,,,,,,,,
|
| 23 |
+
,,,0.5589785936606746,2200,1.1629350185394287,323.6589,327.678,40.96,,,,,
|
| 24 |
+
1.2359,2.54626202583313,1.5460191981931113e-05,0.6097948294480087,2400,,,,,,,,,
|
| 25 |
+
,,,0.6097948294480087,2400,1.1571252346038818,329.6617,321.712,40.214,,,,,
|
| 26 |
+
1.2295,2.519352912902832,1.489836250705816e-05,0.6606110652353427,2600,,,,,,,,,
|
| 27 |
+
,,,0.6606110652353427,2600,1.147318959236145,333.5824,317.93,39.741,,,,,
|
| 28 |
+
1.2253,2.509636640548706,1.4333709768492379e-05,0.7114273010226767,2800,,,,,,,,,
|
| 29 |
+
,,,0.7114273010226767,2800,1.1413997411727905,335.4891,316.123,39.515,,,,,
|
| 30 |
+
1.2164,2.384849786758423,1.3769057029926596e-05,0.7622435368100108,3000,,,,,,,,,
|
| 31 |
+
,,,0.7622435368100108,3000,1.143040657043457,342.2323,309.895,38.737,,,,,
|
| 32 |
+
1.2136,2.668471574783325,1.3204404291360815e-05,0.8130597725973449,3200,,,,,,,,,
|
| 33 |
+
,,,0.8130597725973449,3200,1.1422817707061768,339.8269,312.088,39.011,,,,,
|
| 34 |
+
1.2163,2.458815813064575,1.2639751552795033e-05,0.8638760083846789,3400,,,,,,,,,
|
| 35 |
+
,,,0.8638760083846789,3400,1.1399739980697632,334.8375,316.739,39.592,,,,,
|
| 36 |
+
1.2075,5.854300022125244,1.2075098814229249e-05,0.914692244172013,3600,,,,,,,,,
|
| 37 |
+
,,,0.914692244172013,3600,1.136644959449768,342.7778,309.402,38.675,,,,,
|
| 38 |
+
1.2037,2.514920949935913,1.1510446075663468e-05,0.965508479959347,3800,,,,,,,,,
|
| 39 |
+
,,,0.965508479959347,3800,1.1362895965576172,347.0549,305.589,38.199,,,,,
|
| 40 |
+
1.2026,2.5925779342651367,1.0945793337097685e-05,1.016261195451947,4000,,,,,,,,,
|
| 41 |
+
,,,1.016261195451947,4000,1.132484793663025,348.4028,304.406,38.051,,,,,
|
| 42 |
+
1.2001,2.439039707183838,1.0381140598531904e-05,1.067077431239281,4200,,,,,,,,,
|
| 43 |
+
,,,1.067077431239281,4200,1.1282192468643188,353.0359,300.411,37.551,,,,,
|
| 44 |
+
1.2027,2.566588878631592,9.816487859966122e-06,1.117893667026615,4400,,,,,,,,,
|
| 45 |
+
,,,1.117893667026615,4400,1.1255141496658325,351.1656,302.011,37.751,,,,,
|
| 46 |
+
1.2019,2.763324022293091,9.254658385093168e-06,1.1687099028139492,4600,,,,,,,,,
|
| 47 |
+
,,,1.1687099028139492,4600,1.1325491666793823,351.7344,301.523,37.69,,,,,
|
| 48 |
+
1.1986,2.576956272125244,8.690005646527386e-06,1.219526138601283,4800,,,,,,,,,
|
| 49 |
+
,,,1.219526138601283,4800,1.1340168714523315,349.0646,303.829,37.979,,,,,
|
| 50 |
+
1.2008,2.3913514614105225,8.125352907961605e-06,1.2703423743886173,5000,,,,,,,,,
|
| 51 |
+
,,,1.2703423743886173,5000,1.12372624874115,351.8457,301.428,37.678,,,,,
|
| 52 |
+
1.1988,2.44419264793396,7.560700169395822e-06,1.3211586101759512,5200,,,,,,,,,
|
| 53 |
+
,,,1.3211586101759512,5200,1.121070384979248,355.2577,298.533,37.317,,,,,
|
| 54 |
+
1.1958,3.038212537765503,6.99604743083004e-06,1.3719748459632852,5400,,,,,,,,,
|
| 55 |
+
,,,1.3719748459632852,5400,1.126503586769104,354.1651,299.454,37.432,,,,,
|
| 56 |
+
1.1957,2.4008805751800537,6.431394692264258e-06,1.4227910817506193,5600,,,,,,,,,
|
| 57 |
+
,,,1.4227910817506193,5600,1.1233240365982056,359.3281,295.151,36.894,,,,,
|
| 58 |
+
1.1882,2.4995577335357666,5.866741953698476e-06,1.4736073175379534,5800,,,,,,,,,
|
| 59 |
+
,,,1.4736073175379534,5800,1.1171144247055054,362.5576,292.522,36.565,,,,,
|
| 60 |
+
1.1903,2.3888702392578125,5.3020892151326945e-06,1.5244235533252875,6000,,,,,,,,,
|
| 61 |
+
,,,1.5244235533252875,6000,1.119791865348816,360.8488,293.907,36.738,,,,,
|
| 62 |
+
1.1884,2.3931431770324707,4.737436476566912e-06,1.5752397891126215,6200,,,,,,,,,
|
| 63 |
+
,,,1.5752397891126215,6200,1.1179583072662354,359.7493,294.805,36.851,,,,,
|
| 64 |
+
1.1894,2.3279740810394287,4.17278373800113e-06,1.6260560248999556,6400,,,,,,,,,
|
| 65 |
+
,,,1.6260560248999556,6400,1.116528034210205,358.6415,295.716,36.964,,,,,
|
| 66 |
+
1.1829,2.388047456741333,3.6081309994353475e-06,1.6768722606872895,6600,,,,,,,,,
|
| 67 |
+
,,,1.6768722606872895,6600,1.1195064783096313,361.0405,293.751,36.719,,,,,
|
| 68 |
+
1.1869,2.612689256668091,3.043478260869566e-06,1.7276884964746237,6800,,,,,,,,,
|
| 69 |
+
,,,1.7276884964746237,6800,1.1145069599151611,365.2565,290.36,36.295,,,,,
|
| 70 |
+
1.1841,2.6012659072875977,2.478825522303783e-06,1.7785047322619576,7000,,,,,,,,,
|
| 71 |
+
,,,1.7785047322619576,7000,1.1144654750823975,364.5259,290.942,36.368,,,,,
|
| 72 |
+
1.1807,2.51442289352417,1.9141727837380015e-06,1.829320968049292,7200,,,,,,,,,
|
| 73 |
+
,,,1.829320968049292,7200,1.1116600036621094,369.1442,287.302,35.913,,,,,
|
| 74 |
+
1.1797,2.331007719039917,1.3523433088650482e-06,1.8801372038366257,7400,,,,,,,,,
|
| 75 |
+
,,,1.8801372038366257,7400,1.1135528087615967,367.8567,288.308,36.038,,,,,
|
| 76 |
+
1.1803,2.4899842739105225,7.876905702992659e-07,1.9309534396239598,7600,,,,,,,,,
|
| 77 |
+
,,,1.9309534396239598,7600,1.110607385635376,365.1268,290.463,36.308,,,,,
|
| 78 |
+
1.1828,2.4549551010131836,2.2303783173348393e-07,1.981769675411294,7800,,,,,,,,,
|
| 79 |
+
,,,1.981769675411294,7800,1.1123239994049072,364.2627,291.152,36.394,,,,,
|
| 80 |
+
,,,2.0,7872,,,,,20458.6452,196.988,0.385,2.651863564234199e+17,1.2797311825480888
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3092f7538c47c3b2bf7ea81be403b6bd198788eca0c81f8f4b911a41041b36e
|
| 3 |
+
size 438087048
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"1": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"2": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"3": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"4": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
},
|
| 43 |
+
"30522": {
|
| 44 |
+
"content": "[CELL0]",
|
| 45 |
+
"lstrip": false,
|
| 46 |
+
"normalized": true,
|
| 47 |
+
"rstrip": false,
|
| 48 |
+
"single_word": false,
|
| 49 |
+
"special": false
|
| 50 |
+
},
|
| 51 |
+
"30523": {
|
| 52 |
+
"content": "[CELL1]",
|
| 53 |
+
"lstrip": false,
|
| 54 |
+
"normalized": true,
|
| 55 |
+
"rstrip": false,
|
| 56 |
+
"single_word": false,
|
| 57 |
+
"special": false
|
| 58 |
+
}
|
| 59 |
+
},
|
| 60 |
+
"clean_up_tokenization_spaces": true,
|
| 61 |
+
"cls_token": "[CLS]",
|
| 62 |
+
"do_basic_tokenize": true,
|
| 63 |
+
"do_lower_case": true,
|
| 64 |
+
"extra_special_tokens": {},
|
| 65 |
+
"mask_token": "[MASK]",
|
| 66 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 67 |
+
"never_split": null,
|
| 68 |
+
"pad_token": "[PAD]",
|
| 69 |
+
"sep_token": "[SEP]",
|
| 70 |
+
"strip_accents": null,
|
| 71 |
+
"tokenize_chinese_chars": true,
|
| 72 |
+
"tokenizer_class": "BertTokenizer",
|
| 73 |
+
"unk_token": "[UNK]"
|
| 74 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04167bd439d4a2f8d5d7157fd392faaffbc2cb5e4ac057443677d23a4463ddeb
|
| 3 |
+
size 5841
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|