Add autoregressive config
Browse files- config.json +25 -2
config.json
CHANGED
|
@@ -103,7 +103,7 @@
|
|
| 103 |
"length_penalty": 1.0,
|
| 104 |
"max_length": 20,
|
| 105 |
"min_length": 0,
|
| 106 |
-
"model_type": "
|
| 107 |
"no_repeat_ngram_size": 0,
|
| 108 |
"num_attention_heads": 12,
|
| 109 |
"num_beam_groups": 1,
|
|
@@ -138,5 +138,28 @@
|
|
| 138 |
"use_rotary_embedding": true,
|
| 139 |
"summary_type": "mean"
|
| 140 |
},
|
| 141 |
-
"speech_config_dict": null
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
}
|
|
|
|
| 103 |
"length_penalty": 1.0,
|
| 104 |
"max_length": 20,
|
| 105 |
"min_length": 0,
|
| 106 |
+
"model_type": "clvp_speech_model",
|
| 107 |
"no_repeat_ngram_size": 0,
|
| 108 |
"num_attention_heads": 12,
|
| 109 |
"num_beam_groups": 1,
|
|
|
|
| 138 |
"use_rotary_embedding": true,
|
| 139 |
"summary_type": "mean"
|
| 140 |
},
|
| 141 |
+
"speech_config_dict": null,
|
| 142 |
+
"autoregressive_config": {
|
| 143 |
+
"activation_function": "gelu_new",
|
| 144 |
+
"attn_pdrop": 0.1,
|
| 145 |
+
"bos_token_id": 8192,
|
| 146 |
+
"embd_pdrop": 0.1,
|
| 147 |
+
"eos_token_id": 8193,
|
| 148 |
+
"initializer_range": 0.02,
|
| 149 |
+
"layer_norm_epsilon": 1e-05,
|
| 150 |
+
"model_type": "clvp_autoregressive_model",
|
| 151 |
+
"n_ctx": 1012,
|
| 152 |
+
"n_embd": 1024,
|
| 153 |
+
"n_head": 16,
|
| 154 |
+
"n_layer": 30,
|
| 155 |
+
"n_positions": 1012,
|
| 156 |
+
"resid_pdrop": 0.1,
|
| 157 |
+
"summary_activation": null,
|
| 158 |
+
"summary_first_dropout": 0.1,
|
| 159 |
+
"summary_proj_to_labels": true,
|
| 160 |
+
"summary_type": "cls_index",
|
| 161 |
+
"summary_use_proj": true,
|
| 162 |
+
"vocab_size": 8194
|
| 163 |
+
},
|
| 164 |
+
"autoregressive_config_dict": null
|
| 165 |
}
|