mmirshekari commited on
Commit
e253da4
·
1 Parent(s): 0a37c65

add the model

Browse files
CEBinaryAccuracyEvaluator_dev_results.csv ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,Accuracy
2
+ 0,500,0.6915691063758107
3
+ 0,1000,0.746399393582077
4
+ 0,1500,0.7792470310789186
5
+ 0,2000,0.8148740840562622
6
+ 0,2500,0.7717510317527162
7
+ 0,3000,0.8095679272298493
8
+ 0,3500,0.8231281057862376
9
+ 0,4000,0.8304556556893793
10
+ 0,4500,0.8406468457845532
11
+ 0,5000,0.8255706224206182
12
+ 0,5500,0.8504169123220753
13
+ 0,6000,0.8466268003032089
14
+ 0,6500,0.8416575423229176
15
+ 0,7000,0.8575760128021561
16
+ 0,7500,0.8439316095342373
17
+ 0,8000,0.8530278783795165
18
+ 0,-1,0.8599343047250063
19
+ 1,500,0.8481428451107555
20
+ 1,1000,0.8511749347258486
21
+ 1,1500,0.8500800134759539
22
+ 1,2000,0.8457845531879054
23
+ 1,2500,0.8601869788595974
24
+ 1,3000,0.848479743956877
25
+ 1,3500,0.8626294954939779
26
+ 1,4000,0.8602712035711277
27
+ 1,4500,0.8564810915522615
28
+ 1,5000,0.8291080603048935
29
+ 1,5500,0.854291249052472
30
+ 1,6000,0.8664196075128443
31
+ 1,6500,0.8432578118419944
32
+ 1,7000,0.8518487324180914
33
+ 1,7500,0.8590920576097026
34
+ 1,8000,0.8555546197254275
35
+ 1,-1,0.8650720121283585
36
+ 2,500,0.8623768213593869
37
+ 2,1000,0.8544596984755327
38
+ 2,1500,0.8530278783795165
39
+ 2,2000,0.8643139897245852
40
+ 2,2500,0.8566495409753222
41
+ 2,3000,0.855217720879306
42
+ 2,3500,0.8713046407816053
43
+ 2,4000,0.8554703950138971
44
+ 2,4500,0.8583340352059294
45
+ 2,5000,0.8457845531879054
46
+ 2,5500,0.8681041017434515
47
+ 2,6000,0.8560599679946096
48
+ 2,6500,0.8527752042449255
49
+ 2,7000,0.8592605070327635
50
+ 2,7500,0.8632190684746904
51
+ 2,8000,0.8451949802071927
52
+ 2,-1,0.8585024846289901
53
+ 3,500,0.8649877874168281
54
+ 3,1000,0.8540385749178809
55
+ 3,1500,0.851259159437379
56
+ 3,2000,0.8611976753979618
57
+ 3,2500,0.8556388444369578
58
+ 3,3000,0.8640613155899941
59
+ 3,3500,0.8675145287627389
60
+ 3,4000,0.8625452707824476
61
+ 3,4500,0.8585024846289901
62
+ 3,5000,0.8610292259749011
63
+ 3,5500,0.8651562368398888
64
+ 3,6000,0.8625452707824476
65
+ 3,6500,0.8550492714562452
66
+ 3,7000,0.8601869788595974
67
+ 3,7500,0.8616187989556136
68
+ 3,8000,0.8606923271287796
69
+ 3,-1,0.8609450012633707
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mmirshekari/electra-base-squad-classification",
3
+ "architectures": [
4
+ "ElectraForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "embedding_size": 768,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "LABEL_0": 0
19
+ },
20
+ "layer_norm_eps": 1e-12,
21
+ "max_position_embeddings": 512,
22
+ "model_type": "electra",
23
+ "num_attention_heads": 12,
24
+ "num_hidden_layers": 12,
25
+ "pad_token_id": 0,
26
+ "position_embedding_type": "absolute",
27
+ "summary_activation": "gelu",
28
+ "summary_last_dropout": 0.1,
29
+ "summary_type": "first",
30
+ "summary_use_proj": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.10.0",
33
+ "type_vocab_size": 2,
34
+ "vocab_size": 30522
35
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:754dc8910fe54f1b6e7e31e53bfefad2efd2e9da3fff397bdef2295b94cb7886
3
+ size 438017389
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "name_or_path": "cross-encoder/qnli-electra-base", "special_tokens_map_file": "/home/mostafa/.cache/huggingface/transformers/911a5a7ad5ec495d68a4ef1d0fdcc8f50f57674767907584f4d94c0b066f9787.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff