Commit ·
e253da4
1
Parent(s): 0a37c65
add the model
Browse files- CEBinaryAccuracyEvaluator_dev_results.csv +69 -0
- config.json +35 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +1 -0
- tokenizer.json +0 -0
- tokenizer_config.json +1 -0
- vocab.txt +0 -0
CEBinaryAccuracyEvaluator_dev_results.csv
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,steps,Accuracy
|
| 2 |
+
0,500,0.6915691063758107
|
| 3 |
+
0,1000,0.746399393582077
|
| 4 |
+
0,1500,0.7792470310789186
|
| 5 |
+
0,2000,0.8148740840562622
|
| 6 |
+
0,2500,0.7717510317527162
|
| 7 |
+
0,3000,0.8095679272298493
|
| 8 |
+
0,3500,0.8231281057862376
|
| 9 |
+
0,4000,0.8304556556893793
|
| 10 |
+
0,4500,0.8406468457845532
|
| 11 |
+
0,5000,0.8255706224206182
|
| 12 |
+
0,5500,0.8504169123220753
|
| 13 |
+
0,6000,0.8466268003032089
|
| 14 |
+
0,6500,0.8416575423229176
|
| 15 |
+
0,7000,0.8575760128021561
|
| 16 |
+
0,7500,0.8439316095342373
|
| 17 |
+
0,8000,0.8530278783795165
|
| 18 |
+
0,-1,0.8599343047250063
|
| 19 |
+
1,500,0.8481428451107555
|
| 20 |
+
1,1000,0.8511749347258486
|
| 21 |
+
1,1500,0.8500800134759539
|
| 22 |
+
1,2000,0.8457845531879054
|
| 23 |
+
1,2500,0.8601869788595974
|
| 24 |
+
1,3000,0.848479743956877
|
| 25 |
+
1,3500,0.8626294954939779
|
| 26 |
+
1,4000,0.8602712035711277
|
| 27 |
+
1,4500,0.8564810915522615
|
| 28 |
+
1,5000,0.8291080603048935
|
| 29 |
+
1,5500,0.854291249052472
|
| 30 |
+
1,6000,0.8664196075128443
|
| 31 |
+
1,6500,0.8432578118419944
|
| 32 |
+
1,7000,0.8518487324180914
|
| 33 |
+
1,7500,0.8590920576097026
|
| 34 |
+
1,8000,0.8555546197254275
|
| 35 |
+
1,-1,0.8650720121283585
|
| 36 |
+
2,500,0.8623768213593869
|
| 37 |
+
2,1000,0.8544596984755327
|
| 38 |
+
2,1500,0.8530278783795165
|
| 39 |
+
2,2000,0.8643139897245852
|
| 40 |
+
2,2500,0.8566495409753222
|
| 41 |
+
2,3000,0.855217720879306
|
| 42 |
+
2,3500,0.8713046407816053
|
| 43 |
+
2,4000,0.8554703950138971
|
| 44 |
+
2,4500,0.8583340352059294
|
| 45 |
+
2,5000,0.8457845531879054
|
| 46 |
+
2,5500,0.8681041017434515
|
| 47 |
+
2,6000,0.8560599679946096
|
| 48 |
+
2,6500,0.8527752042449255
|
| 49 |
+
2,7000,0.8592605070327635
|
| 50 |
+
2,7500,0.8632190684746904
|
| 51 |
+
2,8000,0.8451949802071927
|
| 52 |
+
2,-1,0.8585024846289901
|
| 53 |
+
3,500,0.8649877874168281
|
| 54 |
+
3,1000,0.8540385749178809
|
| 55 |
+
3,1500,0.851259159437379
|
| 56 |
+
3,2000,0.8611976753979618
|
| 57 |
+
3,2500,0.8556388444369578
|
| 58 |
+
3,3000,0.8640613155899941
|
| 59 |
+
3,3500,0.8675145287627389
|
| 60 |
+
3,4000,0.8625452707824476
|
| 61 |
+
3,4500,0.8585024846289901
|
| 62 |
+
3,5000,0.8610292259749011
|
| 63 |
+
3,5500,0.8651562368398888
|
| 64 |
+
3,6000,0.8625452707824476
|
| 65 |
+
3,6500,0.8550492714562452
|
| 66 |
+
3,7000,0.8601869788595974
|
| 67 |
+
3,7500,0.8616187989556136
|
| 68 |
+
3,8000,0.8606923271287796
|
| 69 |
+
3,-1,0.8609450012633707
|
config.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "mmirshekari/electra-base-squad-classification",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"ElectraForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"embedding_size": 768,
|
| 9 |
+
"hidden_act": "gelu",
|
| 10 |
+
"hidden_dropout_prob": 0.1,
|
| 11 |
+
"hidden_size": 768,
|
| 12 |
+
"id2label": {
|
| 13 |
+
"0": "LABEL_0"
|
| 14 |
+
},
|
| 15 |
+
"initializer_range": 0.02,
|
| 16 |
+
"intermediate_size": 3072,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"LABEL_0": 0
|
| 19 |
+
},
|
| 20 |
+
"layer_norm_eps": 1e-12,
|
| 21 |
+
"max_position_embeddings": 512,
|
| 22 |
+
"model_type": "electra",
|
| 23 |
+
"num_attention_heads": 12,
|
| 24 |
+
"num_hidden_layers": 12,
|
| 25 |
+
"pad_token_id": 0,
|
| 26 |
+
"position_embedding_type": "absolute",
|
| 27 |
+
"summary_activation": "gelu",
|
| 28 |
+
"summary_last_dropout": 0.1,
|
| 29 |
+
"summary_type": "first",
|
| 30 |
+
"summary_use_proj": true,
|
| 31 |
+
"torch_dtype": "float32",
|
| 32 |
+
"transformers_version": "4.10.0",
|
| 33 |
+
"type_vocab_size": 2,
|
| 34 |
+
"vocab_size": 30522
|
| 35 |
+
}
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:754dc8910fe54f1b6e7e31e53bfefad2efd2e9da3fff397bdef2295b94cb7886
|
| 3 |
+
size 438017389
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "name_or_path": "cross-encoder/qnli-electra-base", "special_tokens_map_file": "/home/mostafa/.cache/huggingface/transformers/911a5a7ad5ec495d68a4ef1d0fdcc8f50f57674767907584f4d94c0b066f9787.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "ElectraTokenizer"}
|
vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|