Nathaniel Weir commited on
Commit
663b277
·
1 Parent(s): 7217f4c

other cross encoder filters

Browse files
Files changed (24) hide show
  1. filters/entailment_cross_encoder/CESoftmaxAccuracyEvaluator_SciTail-dev_results.csv +5 -0
  2. filters/entailment_cross_encoder/added_tokens.json +1 -0
  3. filters/entailment_cross_encoder/config.json +45 -0
  4. filters/entailment_cross_encoder/pytorch_model.bin +3 -0
  5. filters/entailment_cross_encoder/special_tokens_map.json +1 -0
  6. filters/entailment_cross_encoder/spm.model +3 -0
  7. filters/entailment_cross_encoder/tokenizer.json +0 -0
  8. filters/entailment_cross_encoder/tokenizer_config.json +1 -0
  9. filters/silver_ev_cross_encoder_context/CESoftmaxAccuracyEvaluator_mT-dev_results.csv +51 -0
  10. filters/silver_ev_cross_encoder_context/added_tokens.json +3 -0
  11. filters/silver_ev_cross_encoder_context/config.json +45 -0
  12. filters/silver_ev_cross_encoder_context/pytorch_model.bin +3 -0
  13. filters/silver_ev_cross_encoder_context/special_tokens_map.json +9 -0
  14. filters/silver_ev_cross_encoder_context/spm.model +3 -0
  15. filters/silver_ev_cross_encoder_context/tokenizer.json +0 -0
  16. filters/silver_ev_cross_encoder_context/tokenizer_config.json +16 -0
  17. filters/silver_ev_cross_encoder_context_hotpot/CESoftmaxAccuracyEvaluator_hotpot-dev_results.csv +51 -0
  18. filters/silver_ev_cross_encoder_context_hotpot/added_tokens.json +3 -0
  19. filters/silver_ev_cross_encoder_context_hotpot/config.json +45 -0
  20. filters/silver_ev_cross_encoder_context_hotpot/pytorch_model.bin +3 -0
  21. filters/silver_ev_cross_encoder_context_hotpot/special_tokens_map.json +9 -0
  22. filters/silver_ev_cross_encoder_context_hotpot/spm.model +3 -0
  23. filters/silver_ev_cross_encoder_context_hotpot/tokenizer.json +0 -0
  24. filters/silver_ev_cross_encoder_context_hotpot/tokenizer_config.json +16 -0
filters/entailment_cross_encoder/CESoftmaxAccuracyEvaluator_SciTail-dev_results.csv ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ epoch,steps,Accuracy
2
+ 0,-1,0.9700690713737529
3
+ 1,-1,0.965464313123561
4
+ 2,-1,0.9623944742900997
5
+ 3,-1,0.965464313123561
filters/entailment_cross_encoder/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
filters/entailment_cross_encoder/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cross-encoder/nli-deberta-v3-base",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "id2label": {
11
+ "0": "contradiction",
12
+ "1": "entailment",
13
+ "2": "neutral"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "contradiction": 0,
19
+ "entailment": 1,
20
+ "neutral": 2
21
+ },
22
+ "layer_norm_eps": 1e-07,
23
+ "max_position_embeddings": 512,
24
+ "max_relative_positions": -1,
25
+ "model_type": "deberta-v2",
26
+ "norm_rel_ebd": "layer_norm",
27
+ "num_attention_heads": 12,
28
+ "num_hidden_layers": 12,
29
+ "pad_token_id": 0,
30
+ "pooler_dropout": 0,
31
+ "pooler_hidden_act": "gelu",
32
+ "pooler_hidden_size": 768,
33
+ "pos_att_type": [
34
+ "p2c",
35
+ "c2p"
36
+ ],
37
+ "position_biased_input": false,
38
+ "position_buckets": 256,
39
+ "relative_attention": true,
40
+ "share_att_key": true,
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.19.0.dev0",
43
+ "type_vocab_size": 0,
44
+ "vocab_size": 128100
45
+ }
filters/entailment_cross_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b9f354c6200643636c7d161770adec303e3c35cfa6fcc90bba5d35d1724572a
3
+ size 737770027
filters/entailment_cross_encoder/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
filters/entailment_cross_encoder/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
filters/entailment_cross_encoder/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
filters/entailment_cross_encoder/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "special_tokens_map_file": null, "name_or_path": "cross-encoder/nli-deberta-v3-base", "model_max_length": 512, "tokenizer_class": "DebertaV2Tokenizer"}
filters/silver_ev_cross_encoder_context/CESoftmaxAccuracyEvaluator_mT-dev_results.csv ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,threshold,precision,recall,f.5,support,eb_recall
2
+ 0,-1,0.5,0.3482142857142857,0.04918032786885246,0.1571313456889605,793,0.04918032786885246
3
+ 0,-1,0.6,0.4,0.015132408575031526,0.06571741511500548,793,0.015132408575031526
4
+ 0,-1,0.7,0.5714285714285714,0.005044136191677175,0.024360535931790498,793,0.005044136191677175
5
+ 0,-1,0.8,0.0,0.0,0.0,793,0.0
6
+ 0,-1,0.9,0.0,0.0,0.0,793,0.0
7
+ 1,-1,0.5,0.37727272727272726,0.20933165195460277,0.3251077164120642,793,0.20933165195460277
8
+ 1,-1,0.6,0.44594594594594594,0.12484237074401008,0.2944675788221297,793,0.12484237074401008
9
+ 1,-1,0.7,0.4838709677419355,0.05674653215636822,0.19313304721030042,793,0.05674653215636822
10
+ 1,-1,0.8,0.5454545454545454,0.015132408575031526,0.0681044267877412,793,0.015132408575031526
11
+ 1,-1,0.9,1.0,0.0012610340479192938,0.006273525721455458,793,0.0012610340479192938
12
+ 2,-1,0.5,0.38335287221570924,0.4123581336696091,0.3888228299643282,793,0.4123581336696091
13
+ 2,-1,0.6,0.39215686274509803,0.32786885245901637,0.37735849056603776,793,0.32786885245901637
14
+ 2,-1,0.7,0.4186046511627907,0.24968474148802017,0.3687150837988827,793,0.24968474148802017
15
+ 2,-1,0.8,0.43812709030100333,0.16519546027742749,0.3293112116641529,793,0.16519546027742749
16
+ 2,-1,0.9,0.47368421052631576,0.06809583858764187,0.21617293835068055,793,0.06809583858764187
17
+ 3,-1,0.5,0.390282131661442,0.31399747793190413,0.3721973094170403,793,0.31399747793190413
18
+ 3,-1,0.6,0.4076782449725777,0.2812105926860025,0.37403555853740356,793,0.2812105926860025
19
+ 3,-1,0.7,0.4124168514412417,0.23455233291298866,0.3581055063534848,793,0.23455233291298866
20
+ 3,-1,0.8,0.42735042735042733,0.18915510718789408,0.3413746017296313,793,0.18915510718789408
21
+ 3,-1,0.9,0.4474885844748858,0.1235813366960908,0.29358897543439183,793,0.1235813366960908
22
+ 4,-1,0.5,0.38741258741258744,0.3493064312736444,0.37914043252121543,793,0.3493064312736444
23
+ 4,-1,0.6,0.3987635239567233,0.3253467843631778,0.3815439219165927,793,0.3253467843631778
24
+ 4,-1,0.7,0.39965397923875434,0.29129886506935687,0.37198067632850246,793,0.29129886506935687
25
+ 4,-1,0.8,0.40039840637450197,0.25346784363177804,0.35880042841842197,793,0.25346784363177804
26
+ 4,-1,0.9,0.4034229828850856,0.2080706179066835,0.33964594483326477,793,0.2080706179066835
27
+ 5,-1,0.5,0.3895582329317269,0.3669609079445145,0.38481883099709074,793,0.3669609079445145
28
+ 5,-1,0.6,0.38928067700987307,0.3480453972257251,0.3802700468448609,793,0.3480453972257251
29
+ 5,-1,0.7,0.3905775075987842,0.3240857503152585,0.37518248175182484,793,0.3240857503152585
30
+ 5,-1,0.8,0.3834710743801653,0.29255989911727615,0.3610333022097728,793,0.29255989911727615
31
+ 5,-1,0.9,0.38533834586466165,0.25851197982345525,0.35090722355357756,793,0.25851197982345525
32
+ 6,-1,0.5,0.38470873786407767,0.39974779319041615,0.3876253362680362,793,0.39974779319041615
33
+ 6,-1,0.6,0.3781094527363184,0.3833543505674653,0.3791469194312796,793,0.3833543505674653
34
+ 6,-1,0.7,0.38333333333333336,0.3770491803278688,0.38205980066445183,793,0.3770491803278688
35
+ 6,-1,0.8,0.3862433862433862,0.3682219419924338,0.3824993450353681,793,0.3682219419924338
36
+ 6,-1,0.9,0.3881856540084388,0.3480453972257251,0.37943359912015395,793,0.3480453972257251
37
+ 7,-1,0.5,0.3894736842105263,0.37326607818411095,0.38612053222019305,793,0.37326607818411095
38
+ 7,-1,0.6,0.3935309973045822,0.3682219419924338,0.3881946290880085,793,0.3682219419924338
39
+ 7,-1,0.7,0.39398084815321477,0.36317780580075665,0.387409200968523,793,0.36317780580075665
40
+ 7,-1,0.8,0.3944055944055944,0.35561160151324084,0.38598412263892695,793,0.35561160151324084
41
+ 7,-1,0.9,0.3985294117647059,0.3417402269861286,0.385710219185881,793,0.3417402269861286
42
+ 8,-1,0.5,0.4017216642754663,0.3530895334174023,0.39095224797542594,793,0.3530895334174023
43
+ 8,-1,0.6,0.4043478260869565,0.35182849936948296,0.3926259499014917,793,0.35182849936948296
44
+ 8,-1,0.7,0.40294117647058825,0.3455233291298865,0.38998007401081697,793,0.3455233291298865
45
+ 8,-1,0.8,0.40181268882175225,0.3354350567465322,0.3865155478058704,793,0.3354350567465322
46
+ 8,-1,0.9,0.40536277602523657,0.3240857503152585,0.38600180234304593,793,0.3240857503152585
47
+ 9,-1,0.5,0.40054495912806537,0.3707440100882724,0.3942075623491552,793,0.3707440100882724
48
+ 9,-1,0.6,0.40440165061898214,0.3707440100882724,0.39718994866252366,793,0.3707440100882724
49
+ 9,-1,0.7,0.4061624649859944,0.3656998738965952,0.39736914223074815,793,0.3656998738965952
50
+ 9,-1,0.8,0.40825035561877665,0.3619167717528373,0.39805825242718446,793,0.3619167717528373
51
+ 9,-1,0.9,0.409288824383164,0.35561160151324084,0.397295012679628,793,0.35561160151324084
filters/silver_ev_cross_encoder_context/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
filters/silver_ev_cross_encoder_context/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cross-encoder/nli-deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "id2label": {
11
+ "0": "contradiction",
12
+ "1": "entailment",
13
+ "2": "neutral"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 4096,
17
+ "label2id": {
18
+ "contradiction": 0,
19
+ "entailment": 1,
20
+ "neutral": 2
21
+ },
22
+ "layer_norm_eps": 1e-07,
23
+ "max_position_embeddings": 512,
24
+ "max_relative_positions": -1,
25
+ "model_type": "deberta-v2",
26
+ "norm_rel_ebd": "layer_norm",
27
+ "num_attention_heads": 16,
28
+ "num_hidden_layers": 24,
29
+ "pad_token_id": 0,
30
+ "pooler_dropout": 0,
31
+ "pooler_hidden_act": "gelu",
32
+ "pooler_hidden_size": 1024,
33
+ "pos_att_type": [
34
+ "p2c",
35
+ "c2p"
36
+ ],
37
+ "position_biased_input": false,
38
+ "position_buckets": 256,
39
+ "relative_attention": true,
40
+ "share_att_key": true,
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.31.0",
43
+ "type_vocab_size": 0,
44
+ "vocab_size": 128100
45
+ }
filters/silver_ev_cross_encoder_context/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91937cdd02cbb6210432bd7268315eaa8005c6ef2ca0e79b133985c6fb961493
3
+ size 1740395893
filters/silver_ev_cross_encoder_context/special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
filters/silver_ev_cross_encoder_context/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
filters/silver_ev_cross_encoder_context/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
filters/silver_ev_cross_encoder_context/tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "clean_up_tokenization_spaces": true,
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "eos_token": "[SEP]",
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "sp_model_kwargs": {},
12
+ "split_by_punct": false,
13
+ "tokenizer_class": "DebertaV2Tokenizer",
14
+ "unk_token": "[UNK]",
15
+ "vocab_type": "spm"
16
+ }
filters/silver_ev_cross_encoder_context_hotpot/CESoftmaxAccuracyEvaluator_hotpot-dev_results.csv ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,threshold,precision,recall,f.5,support,eb_recall
2
+ 0,-1,0.5,0.6777020447906524,0.3914510686164229,0.5912334352701325,1778,0.3914510686164229
3
+ 0,-1,0.6,0.7425474254742548,0.30821147356580425,0.5792811839323467,1778,0.30821147356580425
4
+ 0,-1,0.7,0.8139059304703476,0.22384701912260968,0.5329405463310123,1778,0.22384701912260968
5
+ 0,-1,0.8,0.839622641509434,0.10011248593925759,0.3389185072353389,1778,0.10011248593925759
6
+ 0,-1,0.9,0.8461538461538461,0.006186726659167604,0.03005464480874317,1778,0.006186726659167604
7
+ 1,-1,0.5,0.6269956458635704,0.4859392575928009,0.5925925925925927,1778,0.4859392575928009
8
+ 1,-1,0.6,0.6565743944636678,0.4268841394825647,0.5927835051546392,1778,0.4268841394825647
9
+ 1,-1,0.7,0.6861081654294804,0.36389201349831274,0.5828828828828828,1778,0.36389201349831274
10
+ 1,-1,0.8,0.7303370786516854,0.2924634420697413,0.5620406398616516,1778,0.2924634420697413
11
+ 1,-1,0.9,0.8246013667425968,0.20359955005624297,0.5121675155631014,1778,0.20359955005624297
12
+ 2,-1,0.5,0.5769828926905132,0.6259842519685039,0.586159679797767,1778,0.6259842519685039
13
+ 2,-1,0.6,0.6008018327605956,0.5899887514060742,0.5986076238301757,1778,0.5899887514060742
14
+ 2,-1,0.7,0.6222929936305732,0.5494938132733408,0.6062298337056341,1778,0.5494938132733408
15
+ 2,-1,0.8,0.6529850746268657,0.4921259842519685,0.6129167834127207,1778,0.4921259842519685
16
+ 2,-1,0.9,0.720682302771855,0.38020247469066365,0.6112115732368897,1778,0.38020247469066365
17
+ 3,-1,0.5,0.5920598388952819,0.5787401574803149,0.5893470790378007,1778,0.5787401574803149
18
+ 3,-1,0.6,0.6045510455104551,0.5528683914510686,0.5934556870321178,1778,0.5528683914510686
19
+ 3,-1,0.7,0.6191432396251674,0.5202474690663667,0.5964663399535723,1778,0.5202474690663667
20
+ 3,-1,0.8,0.6469265367316341,0.48537682789651293,0.6065504638740511,1778,0.48537682789651293
21
+ 3,-1,0.9,0.6897533206831119,0.40888638920134984,0.6064397731064397,1778,0.40888638920134984
22
+ 4,-1,0.5,0.6184895833333334,0.5343082114735658,0.5995960616006059,1778,0.5343082114735658
23
+ 4,-1,0.6,0.632120796156486,0.5179977502812149,0.605443071259532,1778,0.5179977502812149
24
+ 4,-1,0.7,0.6408193123628383,0.49268841394825647,0.6044714325144908,1778,0.49268841394825647
25
+ 4,-1,0.8,0.6538158929976396,0.46737907761529807,0.6055085980763626,1778,0.46737907761529807
26
+ 4,-1,0.9,0.6840182648401827,0.421259842519685,0.608151997401754,1778,0.421259842519685
27
+ 5,-1,0.5,0.5978520286396182,0.5635545556805399,0.5906625795802877,1778,0.5635545556805399
28
+ 5,-1,0.6,0.605521472392638,0.5551181102362205,0.5947216196673897,1778,0.5551181102362205
29
+ 5,-1,0.7,0.6166666666666667,0.5410573678290214,0.5999002244948866,1778,0.5410573678290214
30
+ 5,-1,0.8,0.6205962059620597,0.515185601799775,0.5961989065347566,1778,0.515185601799775
31
+ 5,-1,0.9,0.6339285714285714,0.47919010123734535,0.5954710651383841,1778,0.47919010123734535
32
+ 6,-1,0.5,0.6026097271648873,0.5714285714285714,0.5961042008918094,1778,0.5714285714285714
33
+ 6,-1,0.6,0.6054216867469879,0.5652418447694039,0.5969351389878831,1778,0.5652418447694039
34
+ 6,-1,0.7,0.6104294478527608,0.5596175478065242,0.5995420583273078,1778,0.5596175478065242
35
+ 6,-1,0.8,0.6149936467598475,0.5444319460067492,0.5994550408719346,1778,0.5444319460067492
36
+ 6,-1,0.9,0.6257545271629779,0.5247469066366705,0.6025574786876776,1778,0.5247469066366705
37
+ 7,-1,0.5,0.6214285714285714,0.5382452193475815,0.6027966742252456,1778,0.5382452193475815
38
+ 7,-1,0.6,0.6218872870249017,0.5337457817772778,0.6020045673686882,1778,0.5337457817772778
39
+ 7,-1,0.7,0.625250501002004,0.5264341957255343,0.6026268349214525,1778,0.5264341957255343
40
+ 7,-1,0.8,0.6283367556468172,0.5163104611923509,0.6022041458934663,1778,0.5163104611923509
41
+ 7,-1,0.9,0.6340425531914894,0.5028121484814398,0.6025882987328122,1778,0.5028121484814398
42
+ 8,-1,0.5,0.6223091976516634,0.5365579302587177,0.6030341340075853,1778,0.5365579302587177
43
+ 8,-1,0.6,0.6252465483234714,0.5348706411698537,0.6048079369117273,1778,0.5348706411698537
44
+ 8,-1,0.7,0.6272425249169435,0.530933633295838,0.6052834060015388,1778,0.530933633295838
45
+ 8,-1,0.8,0.6317204301075269,0.5286839145106862,0.6080206985769728,1778,0.5286839145106862
46
+ 8,-1,0.9,0.6333563059958649,0.516872890888639,0.6060406225270377,1778,0.516872890888639
47
+ 9,-1,0.5,0.6024518388791593,0.5804274465691789,0.5979142526071842,1778,0.5804274465691789
48
+ 9,-1,0.6,0.6031839622641509,0.5753655793025871,0.5974071478626489,1778,0.5753655793025871
49
+ 9,-1,0.7,0.6054827175208581,0.5714285714285714,0.5983510011778562,1778,0.5714285714285714
50
+ 9,-1,0.8,0.6071428571428571,0.5641169853768279,0.5980205103744336,1778,0.5641169853768279
51
+ 9,-1,0.9,0.6115960099750624,0.5517435320584927,0.5986087381010495,1778,0.5517435320584927
filters/silver_ev_cross_encoder_context_hotpot/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
filters/silver_ev_cross_encoder_context_hotpot/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cross-encoder/nli-deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "id2label": {
11
+ "0": "contradiction",
12
+ "1": "entailment",
13
+ "2": "neutral"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 4096,
17
+ "label2id": {
18
+ "contradiction": 0,
19
+ "entailment": 1,
20
+ "neutral": 2
21
+ },
22
+ "layer_norm_eps": 1e-07,
23
+ "max_position_embeddings": 512,
24
+ "max_relative_positions": -1,
25
+ "model_type": "deberta-v2",
26
+ "norm_rel_ebd": "layer_norm",
27
+ "num_attention_heads": 16,
28
+ "num_hidden_layers": 24,
29
+ "pad_token_id": 0,
30
+ "pooler_dropout": 0,
31
+ "pooler_hidden_act": "gelu",
32
+ "pooler_hidden_size": 1024,
33
+ "pos_att_type": [
34
+ "p2c",
35
+ "c2p"
36
+ ],
37
+ "position_biased_input": false,
38
+ "position_buckets": 256,
39
+ "relative_attention": true,
40
+ "share_att_key": true,
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.31.0",
43
+ "type_vocab_size": 0,
44
+ "vocab_size": 128100
45
+ }
filters/silver_ev_cross_encoder_context_hotpot/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b544bac2420ac1dadef2a308812c40699b8a9a14b5f928c8f18472038a33198
3
+ size 1740395893
filters/silver_ev_cross_encoder_context_hotpot/special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
filters/silver_ev_cross_encoder_context_hotpot/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
filters/silver_ev_cross_encoder_context_hotpot/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
filters/silver_ev_cross_encoder_context_hotpot/tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "clean_up_tokenization_spaces": true,
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": false,
6
+ "eos_token": "[SEP]",
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "sp_model_kwargs": {},
12
+ "split_by_punct": false,
13
+ "tokenizer_class": "DebertaV2Tokenizer",
14
+ "unk_token": "[UNK]",
15
+ "vocab_type": "spm"
16
+ }