Upload folder using huggingface_hub
Browse files- config.json +1 -1
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 768, "max_position_embeddings": 250000, "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "vocab_size":
|
|
|
|
| 1 |
+
{"attention_probs_dropout_prob": 0.1, "hidden_dropout_prob": 0.1, "hidden_size": 768, "max_position_embeddings": 250000, "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 0, "vocab_size": 30873, "segment_size": [16, 128, 512, 1024, 2048, 4096, 8192], "dilated_rate": [1, 16, 64, 256, 512, 1024, 2048]}
|