Upload config.json
Browse files- config.json +3 -3
config.json
CHANGED
|
@@ -47,10 +47,10 @@
|
|
| 47 |
},
|
| 48 |
"length_penalty": 1.0,
|
| 49 |
"max_length": 20,
|
| 50 |
-
"max_position_embeddings":
|
| 51 |
"max_window_layers": 28,
|
| 52 |
"min_length": 0,
|
| 53 |
-
"model_max_length":
|
| 54 |
"model_type": "qwen2",
|
| 55 |
"no_repeat_ngram_size": 0,
|
| 56 |
"num_attention_heads": 28,
|
|
@@ -85,7 +85,7 @@
|
|
| 85 |
"tie_encoder_decoder": false,
|
| 86 |
"tie_word_embeddings": false,
|
| 87 |
"tokenizer_class": null,
|
| 88 |
-
"tokenizer_model_max_length":
|
| 89 |
"tokenizer_padding_side": "right",
|
| 90 |
"top_k": 50,
|
| 91 |
"top_p": 1.0,
|
|
|
|
| 47 |
},
|
| 48 |
"length_penalty": 1.0,
|
| 49 |
"max_length": 20,
|
| 50 |
+
"max_position_embeddings": 524288,
|
| 51 |
"max_window_layers": 28,
|
| 52 |
"min_length": 0,
|
| 53 |
+
"model_max_length": 524288,
|
| 54 |
"model_type": "qwen2",
|
| 55 |
"no_repeat_ngram_size": 0,
|
| 56 |
"num_attention_heads": 28,
|
|
|
|
| 85 |
"tie_encoder_decoder": false,
|
| 86 |
"tie_word_embeddings": false,
|
| 87 |
"tokenizer_class": null,
|
| 88 |
+
"tokenizer_model_max_length": 524288,
|
| 89 |
"tokenizer_padding_side": "right",
|
| 90 |
"top_k": 50,
|
| 91 |
"top_p": 1.0,
|