zhangchen1991 commited on
Commit
3ffeffc
·
1 Parent(s): 1ab317c

Upload 5 files

Browse files
Files changed (5) hide show
  1. config.json +71 -0
  2. merges.txt +0 -0
  3. pytorch_model.bin +3 -0
  4. tokenizer.json +0 -0
  5. vocab.json +0 -0
config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta_full_large",
3
+ "adapters": {
4
+ "adapters": {
5
+ "convai2": "b1017368d7a97b11",
6
+ "dailydialog": "b1017368d7a97b11",
7
+ "empathetic": "b1017368d7a97b11",
8
+ "reddit": "b1017368d7a97b11",
9
+ "topical": "b1017368d7a97b11"
10
+ },
11
+ "config_map": {
12
+ "b1017368d7a97b11": {
13
+ "adapter_residual_before_ln": false,
14
+ "cross_adapter": false,
15
+ "factorized_phm_W": true,
16
+ "factorized_phm_rule": false,
17
+ "hypercomplex_nonlinearity": "glorot-uniform",
18
+ "init_weights": "bert",
19
+ "inv_adapter": null,
20
+ "inv_adapter_reduction_factor": null,
21
+ "is_parallel": false,
22
+ "learn_phm": true,
23
+ "leave_out": [],
24
+ "ln_after": false,
25
+ "ln_before": false,
26
+ "mh_adapter": true,
27
+ "non_linearity": "swish",
28
+ "original_ln_after": true,
29
+ "original_ln_before": false,
30
+ "output_adapter": true,
31
+ "phm_bias": true,
32
+ "phm_c_init": "normal",
33
+ "phm_dim": 4,
34
+ "phm_init_range": 0.0001,
35
+ "phm_layer": false,
36
+ "phm_rank": 1,
37
+ "reduction_factor": 16,
38
+ "residual_before_ln": true,
39
+ "scaling": 1.0,
40
+ "shared_W_phm": false,
41
+ "shared_phm_rule": true
42
+ }
43
+ },
44
+ "fusion_config_map": {},
45
+ "fusions": {}
46
+ },
47
+ "architectures": [
48
+ "RobertaForPoEEvaluation"
49
+ ],
50
+ "attention_probs_dropout_prob": 0.1,
51
+ "bos_token_id": 0,
52
+ "classifier_dropout": null,
53
+ "eos_token_id": 2,
54
+ "hidden_act": "gelu",
55
+ "hidden_dropout_prob": 0.1,
56
+ "hidden_size": 1024,
57
+ "initializer_range": 0.02,
58
+ "intermediate_size": 4096,
59
+ "layer_norm_eps": 1e-05,
60
+ "max_position_embeddings": 514,
61
+ "model_type": "roberta",
62
+ "num_attention_heads": 16,
63
+ "num_hidden_layers": 24,
64
+ "pad_token_id": 1,
65
+ "position_embedding_type": "absolute",
66
+ "torch_dtype": "float32",
67
+ "transformers_version": "4.17.0",
68
+ "type_vocab_size": 1,
69
+ "use_cache": true,
70
+ "vocab_size": 50265
71
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e05fcfd89f1c5717707884c8b0f524ca722e93238642032809355db699b9d1b
3
+ size 1544829045
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff