added `use_attention_bias`
Browse files- config.json +3 -0
config.json
CHANGED
|
@@ -39,6 +39,7 @@
|
|
| 39 |
"model_type": "clvp_text_model",
|
| 40 |
"no_repeat_ngram_size": 0,
|
| 41 |
"num_attention_heads": 12,
|
|
|
|
| 42 |
"num_beam_groups": 1,
|
| 43 |
"num_beams": 1,
|
| 44 |
"num_hidden_layers": 20,
|
|
@@ -109,6 +110,7 @@
|
|
| 109 |
"num_beam_groups": 1,
|
| 110 |
"num_beams": 1,
|
| 111 |
"num_hidden_layers": 20,
|
|
|
|
| 112 |
"num_return_sequences": 1,
|
| 113 |
"output_attentions": false,
|
| 114 |
"output_hidden_states": false,
|
|
@@ -159,6 +161,7 @@
|
|
| 159 |
"summary_activation": null,
|
| 160 |
"summary_first_dropout": 0.1,
|
| 161 |
"summary_proj_to_labels": true,
|
|
|
|
| 162 |
"summary_type": "cls_index",
|
| 163 |
"summary_use_proj": true,
|
| 164 |
"vocab_size": 8194,
|
|
|
|
| 39 |
"model_type": "clvp_text_model",
|
| 40 |
"no_repeat_ngram_size": 0,
|
| 41 |
"num_attention_heads": 12,
|
| 42 |
+
"use_attention_bias": false,
|
| 43 |
"num_beam_groups": 1,
|
| 44 |
"num_beams": 1,
|
| 45 |
"num_hidden_layers": 20,
|
|
|
|
| 110 |
"num_beam_groups": 1,
|
| 111 |
"num_beams": 1,
|
| 112 |
"num_hidden_layers": 20,
|
| 113 |
+
"use_attention_bias": false,
|
| 114 |
"num_return_sequences": 1,
|
| 115 |
"output_attentions": false,
|
| 116 |
"output_hidden_states": false,
|
|
|
|
| 161 |
"summary_activation": null,
|
| 162 |
"summary_first_dropout": 0.1,
|
| 163 |
"summary_proj_to_labels": true,
|
| 164 |
+
"use_attention_bias": true,
|
| 165 |
"summary_type": "cls_index",
|
| 166 |
"summary_use_proj": true,
|
| 167 |
"vocab_size": 8194,
|