Nathaniel Weir commited on
Commit ·
663b277
1
Parent(s): 7217f4c
other cross encoder filters
Browse files- filters/entailment_cross_encoder/CESoftmaxAccuracyEvaluator_SciTail-dev_results.csv +5 -0
- filters/entailment_cross_encoder/added_tokens.json +1 -0
- filters/entailment_cross_encoder/config.json +45 -0
- filters/entailment_cross_encoder/pytorch_model.bin +3 -0
- filters/entailment_cross_encoder/special_tokens_map.json +1 -0
- filters/entailment_cross_encoder/spm.model +3 -0
- filters/entailment_cross_encoder/tokenizer.json +0 -0
- filters/entailment_cross_encoder/tokenizer_config.json +1 -0
- filters/silver_ev_cross_encoder_context/CESoftmaxAccuracyEvaluator_mT-dev_results.csv +51 -0
- filters/silver_ev_cross_encoder_context/added_tokens.json +3 -0
- filters/silver_ev_cross_encoder_context/config.json +45 -0
- filters/silver_ev_cross_encoder_context/pytorch_model.bin +3 -0
- filters/silver_ev_cross_encoder_context/special_tokens_map.json +9 -0
- filters/silver_ev_cross_encoder_context/spm.model +3 -0
- filters/silver_ev_cross_encoder_context/tokenizer.json +0 -0
- filters/silver_ev_cross_encoder_context/tokenizer_config.json +16 -0
- filters/silver_ev_cross_encoder_context_hotpot/CESoftmaxAccuracyEvaluator_hotpot-dev_results.csv +51 -0
- filters/silver_ev_cross_encoder_context_hotpot/added_tokens.json +3 -0
- filters/silver_ev_cross_encoder_context_hotpot/config.json +45 -0
- filters/silver_ev_cross_encoder_context_hotpot/pytorch_model.bin +3 -0
- filters/silver_ev_cross_encoder_context_hotpot/special_tokens_map.json +9 -0
- filters/silver_ev_cross_encoder_context_hotpot/spm.model +3 -0
- filters/silver_ev_cross_encoder_context_hotpot/tokenizer.json +0 -0
- filters/silver_ev_cross_encoder_context_hotpot/tokenizer_config.json +16 -0
filters/entailment_cross_encoder/CESoftmaxAccuracyEvaluator_SciTail-dev_results.csv
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,steps,Accuracy
|
| 2 |
+
0,-1,0.9700690713737529
|
| 3 |
+
1,-1,0.965464313123561
|
| 4 |
+
2,-1,0.9623944742900997
|
| 5 |
+
3,-1,0.965464313123561
|
filters/entailment_cross_encoder/added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"[MASK]": 128000}
|
filters/entailment_cross_encoder/config.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "cross-encoder/nli-deberta-v3-base",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"DebertaV2ForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"hidden_act": "gelu",
|
| 8 |
+
"hidden_dropout_prob": 0.1,
|
| 9 |
+
"hidden_size": 768,
|
| 10 |
+
"id2label": {
|
| 11 |
+
"0": "contradiction",
|
| 12 |
+
"1": "entailment",
|
| 13 |
+
"2": "neutral"
|
| 14 |
+
},
|
| 15 |
+
"initializer_range": 0.02,
|
| 16 |
+
"intermediate_size": 3072,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"contradiction": 0,
|
| 19 |
+
"entailment": 1,
|
| 20 |
+
"neutral": 2
|
| 21 |
+
},
|
| 22 |
+
"layer_norm_eps": 1e-07,
|
| 23 |
+
"max_position_embeddings": 512,
|
| 24 |
+
"max_relative_positions": -1,
|
| 25 |
+
"model_type": "deberta-v2",
|
| 26 |
+
"norm_rel_ebd": "layer_norm",
|
| 27 |
+
"num_attention_heads": 12,
|
| 28 |
+
"num_hidden_layers": 12,
|
| 29 |
+
"pad_token_id": 0,
|
| 30 |
+
"pooler_dropout": 0,
|
| 31 |
+
"pooler_hidden_act": "gelu",
|
| 32 |
+
"pooler_hidden_size": 768,
|
| 33 |
+
"pos_att_type": [
|
| 34 |
+
"p2c",
|
| 35 |
+
"c2p"
|
| 36 |
+
],
|
| 37 |
+
"position_biased_input": false,
|
| 38 |
+
"position_buckets": 256,
|
| 39 |
+
"relative_attention": true,
|
| 40 |
+
"share_att_key": true,
|
| 41 |
+
"torch_dtype": "float32",
|
| 42 |
+
"transformers_version": "4.19.0.dev0",
|
| 43 |
+
"type_vocab_size": 0,
|
| 44 |
+
"vocab_size": 128100
|
| 45 |
+
}
|
filters/entailment_cross_encoder/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0b9f354c6200643636c7d161770adec303e3c35cfa6fcc90bba5d35d1724572a
|
| 3 |
+
size 737770027
|
filters/entailment_cross_encoder/special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
filters/entailment_cross_encoder/spm.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
| 3 |
+
size 2464616
|
filters/entailment_cross_encoder/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
filters/entailment_cross_encoder/tokenizer_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "special_tokens_map_file": null, "name_or_path": "cross-encoder/nli-deberta-v3-base", "model_max_length": 512, "tokenizer_class": "DebertaV2Tokenizer"}
|
filters/silver_ev_cross_encoder_context/CESoftmaxAccuracyEvaluator_mT-dev_results.csv
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,steps,threshold,precision,recall,f.5,support,eb_recall
|
| 2 |
+
0,-1,0.5,0.3482142857142857,0.04918032786885246,0.1571313456889605,793,0.04918032786885246
|
| 3 |
+
0,-1,0.6,0.4,0.015132408575031526,0.06571741511500548,793,0.015132408575031526
|
| 4 |
+
0,-1,0.7,0.5714285714285714,0.005044136191677175,0.024360535931790498,793,0.005044136191677175
|
| 5 |
+
0,-1,0.8,0.0,0.0,0.0,793,0.0
|
| 6 |
+
0,-1,0.9,0.0,0.0,0.0,793,0.0
|
| 7 |
+
1,-1,0.5,0.37727272727272726,0.20933165195460277,0.3251077164120642,793,0.20933165195460277
|
| 8 |
+
1,-1,0.6,0.44594594594594594,0.12484237074401008,0.2944675788221297,793,0.12484237074401008
|
| 9 |
+
1,-1,0.7,0.4838709677419355,0.05674653215636822,0.19313304721030042,793,0.05674653215636822
|
| 10 |
+
1,-1,0.8,0.5454545454545454,0.015132408575031526,0.0681044267877412,793,0.015132408575031526
|
| 11 |
+
1,-1,0.9,1.0,0.0012610340479192938,0.006273525721455458,793,0.0012610340479192938
|
| 12 |
+
2,-1,0.5,0.38335287221570924,0.4123581336696091,0.3888228299643282,793,0.4123581336696091
|
| 13 |
+
2,-1,0.6,0.39215686274509803,0.32786885245901637,0.37735849056603776,793,0.32786885245901637
|
| 14 |
+
2,-1,0.7,0.4186046511627907,0.24968474148802017,0.3687150837988827,793,0.24968474148802017
|
| 15 |
+
2,-1,0.8,0.43812709030100333,0.16519546027742749,0.3293112116641529,793,0.16519546027742749
|
| 16 |
+
2,-1,0.9,0.47368421052631576,0.06809583858764187,0.21617293835068055,793,0.06809583858764187
|
| 17 |
+
3,-1,0.5,0.390282131661442,0.31399747793190413,0.3721973094170403,793,0.31399747793190413
|
| 18 |
+
3,-1,0.6,0.4076782449725777,0.2812105926860025,0.37403555853740356,793,0.2812105926860025
|
| 19 |
+
3,-1,0.7,0.4124168514412417,0.23455233291298866,0.3581055063534848,793,0.23455233291298866
|
| 20 |
+
3,-1,0.8,0.42735042735042733,0.18915510718789408,0.3413746017296313,793,0.18915510718789408
|
| 21 |
+
3,-1,0.9,0.4474885844748858,0.1235813366960908,0.29358897543439183,793,0.1235813366960908
|
| 22 |
+
4,-1,0.5,0.38741258741258744,0.3493064312736444,0.37914043252121543,793,0.3493064312736444
|
| 23 |
+
4,-1,0.6,0.3987635239567233,0.3253467843631778,0.3815439219165927,793,0.3253467843631778
|
| 24 |
+
4,-1,0.7,0.39965397923875434,0.29129886506935687,0.37198067632850246,793,0.29129886506935687
|
| 25 |
+
4,-1,0.8,0.40039840637450197,0.25346784363177804,0.35880042841842197,793,0.25346784363177804
|
| 26 |
+
4,-1,0.9,0.4034229828850856,0.2080706179066835,0.33964594483326477,793,0.2080706179066835
|
| 27 |
+
5,-1,0.5,0.3895582329317269,0.3669609079445145,0.38481883099709074,793,0.3669609079445145
|
| 28 |
+
5,-1,0.6,0.38928067700987307,0.3480453972257251,0.3802700468448609,793,0.3480453972257251
|
| 29 |
+
5,-1,0.7,0.3905775075987842,0.3240857503152585,0.37518248175182484,793,0.3240857503152585
|
| 30 |
+
5,-1,0.8,0.3834710743801653,0.29255989911727615,0.3610333022097728,793,0.29255989911727615
|
| 31 |
+
5,-1,0.9,0.38533834586466165,0.25851197982345525,0.35090722355357756,793,0.25851197982345525
|
| 32 |
+
6,-1,0.5,0.38470873786407767,0.39974779319041615,0.3876253362680362,793,0.39974779319041615
|
| 33 |
+
6,-1,0.6,0.3781094527363184,0.3833543505674653,0.3791469194312796,793,0.3833543505674653
|
| 34 |
+
6,-1,0.7,0.38333333333333336,0.3770491803278688,0.38205980066445183,793,0.3770491803278688
|
| 35 |
+
6,-1,0.8,0.3862433862433862,0.3682219419924338,0.3824993450353681,793,0.3682219419924338
|
| 36 |
+
6,-1,0.9,0.3881856540084388,0.3480453972257251,0.37943359912015395,793,0.3480453972257251
|
| 37 |
+
7,-1,0.5,0.3894736842105263,0.37326607818411095,0.38612053222019305,793,0.37326607818411095
|
| 38 |
+
7,-1,0.6,0.3935309973045822,0.3682219419924338,0.3881946290880085,793,0.3682219419924338
|
| 39 |
+
7,-1,0.7,0.39398084815321477,0.36317780580075665,0.387409200968523,793,0.36317780580075665
|
| 40 |
+
7,-1,0.8,0.3944055944055944,0.35561160151324084,0.38598412263892695,793,0.35561160151324084
|
| 41 |
+
7,-1,0.9,0.3985294117647059,0.3417402269861286,0.385710219185881,793,0.3417402269861286
|
| 42 |
+
8,-1,0.5,0.4017216642754663,0.3530895334174023,0.39095224797542594,793,0.3530895334174023
|
| 43 |
+
8,-1,0.6,0.4043478260869565,0.35182849936948296,0.3926259499014917,793,0.35182849936948296
|
| 44 |
+
8,-1,0.7,0.40294117647058825,0.3455233291298865,0.38998007401081697,793,0.3455233291298865
|
| 45 |
+
8,-1,0.8,0.40181268882175225,0.3354350567465322,0.3865155478058704,793,0.3354350567465322
|
| 46 |
+
8,-1,0.9,0.40536277602523657,0.3240857503152585,0.38600180234304593,793,0.3240857503152585
|
| 47 |
+
9,-1,0.5,0.40054495912806537,0.3707440100882724,0.3942075623491552,793,0.3707440100882724
|
| 48 |
+
9,-1,0.6,0.40440165061898214,0.3707440100882724,0.39718994866252366,793,0.3707440100882724
|
| 49 |
+
9,-1,0.7,0.4061624649859944,0.3656998738965952,0.39736914223074815,793,0.3656998738965952
|
| 50 |
+
9,-1,0.8,0.40825035561877665,0.3619167717528373,0.39805825242718446,793,0.3619167717528373
|
| 51 |
+
9,-1,0.9,0.409288824383164,0.35561160151324084,0.397295012679628,793,0.35561160151324084
|
filters/silver_ev_cross_encoder_context/added_tokens.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"[MASK]": 128000
|
| 3 |
+
}
|
filters/silver_ev_cross_encoder_context/config.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "cross-encoder/nli-deberta-v3-large",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"DebertaV2ForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"hidden_act": "gelu",
|
| 8 |
+
"hidden_dropout_prob": 0.1,
|
| 9 |
+
"hidden_size": 1024,
|
| 10 |
+
"id2label": {
|
| 11 |
+
"0": "contradiction",
|
| 12 |
+
"1": "entailment",
|
| 13 |
+
"2": "neutral"
|
| 14 |
+
},
|
| 15 |
+
"initializer_range": 0.02,
|
| 16 |
+
"intermediate_size": 4096,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"contradiction": 0,
|
| 19 |
+
"entailment": 1,
|
| 20 |
+
"neutral": 2
|
| 21 |
+
},
|
| 22 |
+
"layer_norm_eps": 1e-07,
|
| 23 |
+
"max_position_embeddings": 512,
|
| 24 |
+
"max_relative_positions": -1,
|
| 25 |
+
"model_type": "deberta-v2",
|
| 26 |
+
"norm_rel_ebd": "layer_norm",
|
| 27 |
+
"num_attention_heads": 16,
|
| 28 |
+
"num_hidden_layers": 24,
|
| 29 |
+
"pad_token_id": 0,
|
| 30 |
+
"pooler_dropout": 0,
|
| 31 |
+
"pooler_hidden_act": "gelu",
|
| 32 |
+
"pooler_hidden_size": 1024,
|
| 33 |
+
"pos_att_type": [
|
| 34 |
+
"p2c",
|
| 35 |
+
"c2p"
|
| 36 |
+
],
|
| 37 |
+
"position_biased_input": false,
|
| 38 |
+
"position_buckets": 256,
|
| 39 |
+
"relative_attention": true,
|
| 40 |
+
"share_att_key": true,
|
| 41 |
+
"torch_dtype": "float32",
|
| 42 |
+
"transformers_version": "4.31.0",
|
| 43 |
+
"type_vocab_size": 0,
|
| 44 |
+
"vocab_size": 128100
|
| 45 |
+
}
|
filters/silver_ev_cross_encoder_context/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:91937cdd02cbb6210432bd7268315eaa8005c6ef2ca0e79b133985c6fb961493
|
| 3 |
+
size 1740395893
|
filters/silver_ev_cross_encoder_context/special_tokens_map.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "[CLS]",
|
| 3 |
+
"cls_token": "[CLS]",
|
| 4 |
+
"eos_token": "[SEP]",
|
| 5 |
+
"mask_token": "[MASK]",
|
| 6 |
+
"pad_token": "[PAD]",
|
| 7 |
+
"sep_token": "[SEP]",
|
| 8 |
+
"unk_token": "[UNK]"
|
| 9 |
+
}
|
filters/silver_ev_cross_encoder_context/spm.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
| 3 |
+
size 2464616
|
filters/silver_ev_cross_encoder_context/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
filters/silver_ev_cross_encoder_context/tokenizer_config.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "[CLS]",
|
| 3 |
+
"clean_up_tokenization_spaces": true,
|
| 4 |
+
"cls_token": "[CLS]",
|
| 5 |
+
"do_lower_case": false,
|
| 6 |
+
"eos_token": "[SEP]",
|
| 7 |
+
"mask_token": "[MASK]",
|
| 8 |
+
"model_max_length": 512,
|
| 9 |
+
"pad_token": "[PAD]",
|
| 10 |
+
"sep_token": "[SEP]",
|
| 11 |
+
"sp_model_kwargs": {},
|
| 12 |
+
"split_by_punct": false,
|
| 13 |
+
"tokenizer_class": "DebertaV2Tokenizer",
|
| 14 |
+
"unk_token": "[UNK]",
|
| 15 |
+
"vocab_type": "spm"
|
| 16 |
+
}
|
filters/silver_ev_cross_encoder_context_hotpot/CESoftmaxAccuracyEvaluator_hotpot-dev_results.csv
ADDED
|
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,steps,threshold,precision,recall,f.5,support,eb_recall
|
| 2 |
+
0,-1,0.5,0.6777020447906524,0.3914510686164229,0.5912334352701325,1778,0.3914510686164229
|
| 3 |
+
0,-1,0.6,0.7425474254742548,0.30821147356580425,0.5792811839323467,1778,0.30821147356580425
|
| 4 |
+
0,-1,0.7,0.8139059304703476,0.22384701912260968,0.5329405463310123,1778,0.22384701912260968
|
| 5 |
+
0,-1,0.8,0.839622641509434,0.10011248593925759,0.3389185072353389,1778,0.10011248593925759
|
| 6 |
+
0,-1,0.9,0.8461538461538461,0.006186726659167604,0.03005464480874317,1778,0.006186726659167604
|
| 7 |
+
1,-1,0.5,0.6269956458635704,0.4859392575928009,0.5925925925925927,1778,0.4859392575928009
|
| 8 |
+
1,-1,0.6,0.6565743944636678,0.4268841394825647,0.5927835051546392,1778,0.4268841394825647
|
| 9 |
+
1,-1,0.7,0.6861081654294804,0.36389201349831274,0.5828828828828828,1778,0.36389201349831274
|
| 10 |
+
1,-1,0.8,0.7303370786516854,0.2924634420697413,0.5620406398616516,1778,0.2924634420697413
|
| 11 |
+
1,-1,0.9,0.8246013667425968,0.20359955005624297,0.5121675155631014,1778,0.20359955005624297
|
| 12 |
+
2,-1,0.5,0.5769828926905132,0.6259842519685039,0.586159679797767,1778,0.6259842519685039
|
| 13 |
+
2,-1,0.6,0.6008018327605956,0.5899887514060742,0.5986076238301757,1778,0.5899887514060742
|
| 14 |
+
2,-1,0.7,0.6222929936305732,0.5494938132733408,0.6062298337056341,1778,0.5494938132733408
|
| 15 |
+
2,-1,0.8,0.6529850746268657,0.4921259842519685,0.6129167834127207,1778,0.4921259842519685
|
| 16 |
+
2,-1,0.9,0.720682302771855,0.38020247469066365,0.6112115732368897,1778,0.38020247469066365
|
| 17 |
+
3,-1,0.5,0.5920598388952819,0.5787401574803149,0.5893470790378007,1778,0.5787401574803149
|
| 18 |
+
3,-1,0.6,0.6045510455104551,0.5528683914510686,0.5934556870321178,1778,0.5528683914510686
|
| 19 |
+
3,-1,0.7,0.6191432396251674,0.5202474690663667,0.5964663399535723,1778,0.5202474690663667
|
| 20 |
+
3,-1,0.8,0.6469265367316341,0.48537682789651293,0.6065504638740511,1778,0.48537682789651293
|
| 21 |
+
3,-1,0.9,0.6897533206831119,0.40888638920134984,0.6064397731064397,1778,0.40888638920134984
|
| 22 |
+
4,-1,0.5,0.6184895833333334,0.5343082114735658,0.5995960616006059,1778,0.5343082114735658
|
| 23 |
+
4,-1,0.6,0.632120796156486,0.5179977502812149,0.605443071259532,1778,0.5179977502812149
|
| 24 |
+
4,-1,0.7,0.6408193123628383,0.49268841394825647,0.6044714325144908,1778,0.49268841394825647
|
| 25 |
+
4,-1,0.8,0.6538158929976396,0.46737907761529807,0.6055085980763626,1778,0.46737907761529807
|
| 26 |
+
4,-1,0.9,0.6840182648401827,0.421259842519685,0.608151997401754,1778,0.421259842519685
|
| 27 |
+
5,-1,0.5,0.5978520286396182,0.5635545556805399,0.5906625795802877,1778,0.5635545556805399
|
| 28 |
+
5,-1,0.6,0.605521472392638,0.5551181102362205,0.5947216196673897,1778,0.5551181102362205
|
| 29 |
+
5,-1,0.7,0.6166666666666667,0.5410573678290214,0.5999002244948866,1778,0.5410573678290214
|
| 30 |
+
5,-1,0.8,0.6205962059620597,0.515185601799775,0.5961989065347566,1778,0.515185601799775
|
| 31 |
+
5,-1,0.9,0.6339285714285714,0.47919010123734535,0.5954710651383841,1778,0.47919010123734535
|
| 32 |
+
6,-1,0.5,0.6026097271648873,0.5714285714285714,0.5961042008918094,1778,0.5714285714285714
|
| 33 |
+
6,-1,0.6,0.6054216867469879,0.5652418447694039,0.5969351389878831,1778,0.5652418447694039
|
| 34 |
+
6,-1,0.7,0.6104294478527608,0.5596175478065242,0.5995420583273078,1778,0.5596175478065242
|
| 35 |
+
6,-1,0.8,0.6149936467598475,0.5444319460067492,0.5994550408719346,1778,0.5444319460067492
|
| 36 |
+
6,-1,0.9,0.6257545271629779,0.5247469066366705,0.6025574786876776,1778,0.5247469066366705
|
| 37 |
+
7,-1,0.5,0.6214285714285714,0.5382452193475815,0.6027966742252456,1778,0.5382452193475815
|
| 38 |
+
7,-1,0.6,0.6218872870249017,0.5337457817772778,0.6020045673686882,1778,0.5337457817772778
|
| 39 |
+
7,-1,0.7,0.625250501002004,0.5264341957255343,0.6026268349214525,1778,0.5264341957255343
|
| 40 |
+
7,-1,0.8,0.6283367556468172,0.5163104611923509,0.6022041458934663,1778,0.5163104611923509
|
| 41 |
+
7,-1,0.9,0.6340425531914894,0.5028121484814398,0.6025882987328122,1778,0.5028121484814398
|
| 42 |
+
8,-1,0.5,0.6223091976516634,0.5365579302587177,0.6030341340075853,1778,0.5365579302587177
|
| 43 |
+
8,-1,0.6,0.6252465483234714,0.5348706411698537,0.6048079369117273,1778,0.5348706411698537
|
| 44 |
+
8,-1,0.7,0.6272425249169435,0.530933633295838,0.6052834060015388,1778,0.530933633295838
|
| 45 |
+
8,-1,0.8,0.6317204301075269,0.5286839145106862,0.6080206985769728,1778,0.5286839145106862
|
| 46 |
+
8,-1,0.9,0.6333563059958649,0.516872890888639,0.6060406225270377,1778,0.516872890888639
|
| 47 |
+
9,-1,0.5,0.6024518388791593,0.5804274465691789,0.5979142526071842,1778,0.5804274465691789
|
| 48 |
+
9,-1,0.6,0.6031839622641509,0.5753655793025871,0.5974071478626489,1778,0.5753655793025871
|
| 49 |
+
9,-1,0.7,0.6054827175208581,0.5714285714285714,0.5983510011778562,1778,0.5714285714285714
|
| 50 |
+
9,-1,0.8,0.6071428571428571,0.5641169853768279,0.5980205103744336,1778,0.5641169853768279
|
| 51 |
+
9,-1,0.9,0.6115960099750624,0.5517435320584927,0.5986087381010495,1778,0.5517435320584927
|
filters/silver_ev_cross_encoder_context_hotpot/added_tokens.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"[MASK]": 128000
|
| 3 |
+
}
|
filters/silver_ev_cross_encoder_context_hotpot/config.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "cross-encoder/nli-deberta-v3-large",
|
| 3 |
+
"architectures": [
|
| 4 |
+
"DebertaV2ForSequenceClassification"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"hidden_act": "gelu",
|
| 8 |
+
"hidden_dropout_prob": 0.1,
|
| 9 |
+
"hidden_size": 1024,
|
| 10 |
+
"id2label": {
|
| 11 |
+
"0": "contradiction",
|
| 12 |
+
"1": "entailment",
|
| 13 |
+
"2": "neutral"
|
| 14 |
+
},
|
| 15 |
+
"initializer_range": 0.02,
|
| 16 |
+
"intermediate_size": 4096,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"contradiction": 0,
|
| 19 |
+
"entailment": 1,
|
| 20 |
+
"neutral": 2
|
| 21 |
+
},
|
| 22 |
+
"layer_norm_eps": 1e-07,
|
| 23 |
+
"max_position_embeddings": 512,
|
| 24 |
+
"max_relative_positions": -1,
|
| 25 |
+
"model_type": "deberta-v2",
|
| 26 |
+
"norm_rel_ebd": "layer_norm",
|
| 27 |
+
"num_attention_heads": 16,
|
| 28 |
+
"num_hidden_layers": 24,
|
| 29 |
+
"pad_token_id": 0,
|
| 30 |
+
"pooler_dropout": 0,
|
| 31 |
+
"pooler_hidden_act": "gelu",
|
| 32 |
+
"pooler_hidden_size": 1024,
|
| 33 |
+
"pos_att_type": [
|
| 34 |
+
"p2c",
|
| 35 |
+
"c2p"
|
| 36 |
+
],
|
| 37 |
+
"position_biased_input": false,
|
| 38 |
+
"position_buckets": 256,
|
| 39 |
+
"relative_attention": true,
|
| 40 |
+
"share_att_key": true,
|
| 41 |
+
"torch_dtype": "float32",
|
| 42 |
+
"transformers_version": "4.31.0",
|
| 43 |
+
"type_vocab_size": 0,
|
| 44 |
+
"vocab_size": 128100
|
| 45 |
+
}
|
filters/silver_ev_cross_encoder_context_hotpot/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b544bac2420ac1dadef2a308812c40699b8a9a14b5f928c8f18472038a33198
|
| 3 |
+
size 1740395893
|
filters/silver_ev_cross_encoder_context_hotpot/special_tokens_map.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "[CLS]",
|
| 3 |
+
"cls_token": "[CLS]",
|
| 4 |
+
"eos_token": "[SEP]",
|
| 5 |
+
"mask_token": "[MASK]",
|
| 6 |
+
"pad_token": "[PAD]",
|
| 7 |
+
"sep_token": "[SEP]",
|
| 8 |
+
"unk_token": "[UNK]"
|
| 9 |
+
}
|
filters/silver_ev_cross_encoder_context_hotpot/spm.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
| 3 |
+
size 2464616
|
filters/silver_ev_cross_encoder_context_hotpot/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
filters/silver_ev_cross_encoder_context_hotpot/tokenizer_config.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": "[CLS]",
|
| 3 |
+
"clean_up_tokenization_spaces": true,
|
| 4 |
+
"cls_token": "[CLS]",
|
| 5 |
+
"do_lower_case": false,
|
| 6 |
+
"eos_token": "[SEP]",
|
| 7 |
+
"mask_token": "[MASK]",
|
| 8 |
+
"model_max_length": 512,
|
| 9 |
+
"pad_token": "[PAD]",
|
| 10 |
+
"sep_token": "[SEP]",
|
| 11 |
+
"sp_model_kwargs": {},
|
| 12 |
+
"split_by_punct": false,
|
| 13 |
+
"tokenizer_class": "DebertaV2Tokenizer",
|
| 14 |
+
"unk_token": "[UNK]",
|
| 15 |
+
"vocab_type": "spm"
|
| 16 |
+
}
|