omidtoptal commited on
Commit
aa808ad
·
verified ·
1 Parent(s): b5c55f2

Add files using upload-large-folder tool

Browse files
config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dslim/bert-base-NER",
3
+ "_num_labels": 9,
4
+ "architectures": [
5
+ "BertForTokenClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "classifier_dropout": null,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "B-PER",
15
+ "2": "I-PER",
16
+ "3": "B-ORG",
17
+ "4": "I-ORG",
18
+ "5": "B-LOC",
19
+ "6": "I-LOC",
20
+ "7": "B-MISC",
21
+ "8": "I-MISC"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "label2id": {
26
+ "B-LOC": 5,
27
+ "B-MISC": 7,
28
+ "B-ORG": 3,
29
+ "B-PER": 1,
30
+ "I-LOC": 6,
31
+ "I-MISC": 8,
32
+ "I-ORG": 4,
33
+ "I-PER": 2,
34
+ "O": 0
35
+ },
36
+ "layer_norm_eps": 1e-12,
37
+ "max_position_embeddings": 512,
38
+ "model_type": "bert",
39
+ "num_attention_heads": 12,
40
+ "num_hidden_layers": 12,
41
+ "output_past": true,
42
+ "pad_token_id": 0,
43
+ "position_embedding_type": "absolute",
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.49.0",
46
+ "type_vocab_size": 2,
47
+ "use_cache": true,
48
+ "vocab_size": 28996
49
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79b1387b24751d5fd1314fc6307c978c95efb6a8941e945f341da92f723fdb65
3
+ size 430929740
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": false,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_len": 512,
51
+ "model_max_length": 512,
52
+ "never_split": null,
53
+ "pad_token": "[PAD]",
54
+ "sep_token": "[SEP]",
55
+ "strip_accents": null,
56
+ "tokenize_chinese_chars": true,
57
+ "tokenizer_class": "BertTokenizer",
58
+ "unk_token": "[UNK]"
59
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:059398dfa17725067c53ca47c748ddbf2d45ef0b320a26273007fe9f5783e381
3
+ size 5240
training_log.csv ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,epoch,eval_LOC_f1,eval_LOC_precision,eval_LOC_recall,eval_ORG_f1,eval_ORG_loose_f1,eval_ORG_loose_precision,eval_ORG_loose_recall,eval_ORG_precision,eval_ORG_recall,eval_PER_f1,eval_PER_precision,eval_PER_recall,eval_loss,eval_macro_f1,eval_macro_precision,eval_macro_recall,eval_runtime,eval_samples_per_second,eval_steps_per_second,grad_norm,learning_rate,loss,total_flos,train_loss,train_runtime,train_samples_per_second,train_steps_per_second
2
+ 1.0,1.0,0.8066666666666666,0.7891304347826087,0.825,0.41077441077441074,0.6224066390041493,0.5769230769230769,0.6756756756756757,0.34269662921348315,0.5126050420168067,0.653555219364599,0.5669291338582677,0.7714285714285715,0.004199459683150053,0.6236654322685588,0.5662520659514532,0.7030112044817928,12.4772,94.412,2.965,0.003550168126821518,4.5e-05,0.0083,,,,,
3
+ 2.0,2.0,0.8421052631578947,0.7983706720977597,0.8909090909090909,0.5540540540540541,0.6893732970027248,0.6309226932668329,0.7597597597597597,0.4632768361581921,0.6890756302521008,0.6734992679355782,0.5707196029776674,0.8214285714285714,0.003755683545023203,0.6898861950491756,0.6107890370778731,0.800471097529921,12.4529,94.597,2.971,0.14651519060134888,4e-05,0.0038,,,,,
4
+ 3.0,3.0,0.8401727861771058,0.8004115226337448,0.884090909090909,0.5674740484429065,0.7033898305084746,0.664,0.7477477477477478,0.4823529411764706,0.6890756302521008,0.668639053254438,0.5707070707070707,0.8071428571428572,0.003569010877981782,0.6920952959581501,0.6178238448390954,0.7934364654952889,12.4902,94.314,2.962,0.18637016415596008,3.5e-05,0.0026,,,,,
5
+ 4.0,4.0,0.8535300316122235,0.7956777996070727,0.9204545454545454,0.5786163522012578,0.7324675324675325,0.6453089244851259,0.8468468468468469,0.4623115577889447,0.773109243697479,0.6760168302945302,0.5565819861431871,0.8607142857142858,0.004289823118597269,0.7027210713693371,0.6048571145130682,0.8514260249554367,12.4319,94.757,2.976,0.00032159508555196226,3e-05,0.0018,,,,,
6
+ 5.0,5.0,0.8685082872928177,0.8451612903225807,0.8931818181818182,0.6014492753623188,0.7557471264367815,0.7245179063360881,0.7897897897897898,0.5286624203821656,0.6974789915966386,0.6928675400291121,0.5847665847665847,0.85,0.0037530281115323305,0.7209417008947496,0.652863431823777,0.8135536032594857,12.4642,94.511,2.969,0.07077983021736145,2.5e-05,0.0013,,,,,
7
+ 6.0,6.0,0.8646288209606987,0.8319327731092437,0.9,0.6353790613718412,0.7343976777939042,0.7106741573033708,0.7597597597597597,0.5569620253164557,0.7394957983193278,0.7043189368770764,0.6583850931677019,0.7571428571428571,0.004114250186830759,0.7347756064032055,0.6824266305311338,0.7988795518207282,12.4516,94.606,2.972,0.0001318866416113451,2e-05,0.001,,,,,
8
+ 7.0,7.0,0.864406779661017,0.8095238095238095,0.9272727272727272,0.6338028169014084,0.752808988764045,0.7071240105540897,0.8048048048048048,0.5454545454545454,0.7563025210084033,0.7270029673590505,0.6218274111675127,0.875,0.005139179527759552,0.7417375213071585,0.658935255381956,0.8528584160937102,12.4577,94.56,2.97,0.011717773042619228,1.5e-05,0.0007,,,,,
9
+ 8.0,8.0,0.8629989212513485,0.8213552361396304,0.9090909090909091,0.6292134831460674,0.7419354838709677,0.7249283667621776,0.7597597597597597,0.5675675675675675,0.7058823529411765,0.7164634146341462,0.625,0.8392857142857143,0.005133509635925293,0.7362252730105207,0.6713076012357325,0.8180863254392666,12.4498,94.62,2.972,0.0048303925432264805,1e-05,0.0006,,,,,
10
+ 9.0,9.0,0.8663101604278074,0.8181818181818182,0.9204545454545454,0.65,0.7616361071932298,0.7180851063829787,0.8108108108108109,0.5652173913043478,0.7647058823529411,0.7272727272727273,0.631578947368421,0.8571428571428571,0.0053354487754404545,0.7478609625668448,0.6716593856181957,0.8474344283167813,12.4582,94.556,2.97,0.005022859666496515,5e-06,0.0005,,,,,
11
+ 10.0,10.0,0.867237687366167,0.819838056680162,0.9204545454545454,0.6254545454545455,0.7503607503607503,0.7222222222222222,0.7807807807807807,0.5512820512820513,0.7226890756302521,0.7373271889400921,0.6469002695417789,0.8571428571428571,0.005584388971328735,0.7433398072536015,0.6726734591679975,0.833428826075885,12.438,94.71,2.975,0.026601826772093773,0.0,0.0004,2.770182222400512e+16,0.002087883570467133,2411.0838,43.968,1.377
vocab.txt ADDED
The diff for this file is too large to render. See raw diff