Update config.json
Browse files- config.json +4 -4
config.json
CHANGED
|
@@ -280,7 +280,7 @@
|
|
| 280 |
"freq_mask_length": 27,
|
| 281 |
"inference_warmup_tokens": 10,
|
| 282 |
"label_smoothing": 0.0,
|
| 283 |
-
"length_penalty":
|
| 284 |
"llm_dim": 2048,
|
| 285 |
"lora_alpha": 32,
|
| 286 |
"lora_dropout": 0.0,
|
|
@@ -294,11 +294,11 @@
|
|
| 294 |
"up_proj",
|
| 295 |
"down_proj"
|
| 296 |
],
|
| 297 |
-
"max_new_tokens":
|
| 298 |
"min_new_tokens": 0,
|
| 299 |
"model_dtype": "bfloat16",
|
| 300 |
"model_type": "asr_model",
|
| 301 |
-
"no_repeat_ngram_size":
|
| 302 |
"num_beams": 1,
|
| 303 |
"num_experts": 4,
|
| 304 |
"num_experts_per_tok": 2,
|
|
@@ -317,7 +317,7 @@
|
|
| 317 |
"qformer_num_heads": 16,
|
| 318 |
"qformer_num_layers": 2,
|
| 319 |
"qformer_window_size": 15,
|
| 320 |
-
"repetition_penalty": 1.
|
| 321 |
"router_aux_loss_coef": 0.01,
|
| 322 |
"system_prompt": "",
|
| 323 |
"text_config": {
|
|
|
|
| 280 |
"freq_mask_length": 27,
|
| 281 |
"inference_warmup_tokens": 10,
|
| 282 |
"label_smoothing": 0.0,
|
| 283 |
+
"length_penalty": 0.8,
|
| 284 |
"llm_dim": 2048,
|
| 285 |
"lora_alpha": 32,
|
| 286 |
"lora_dropout": 0.0,
|
|
|
|
| 294 |
"up_proj",
|
| 295 |
"down_proj"
|
| 296 |
],
|
| 297 |
+
"max_new_tokens": 96,
|
| 298 |
"min_new_tokens": 0,
|
| 299 |
"model_dtype": "bfloat16",
|
| 300 |
"model_type": "asr_model",
|
| 301 |
+
"no_repeat_ngram_size": 4,
|
| 302 |
"num_beams": 1,
|
| 303 |
"num_experts": 4,
|
| 304 |
"num_experts_per_tok": 2,
|
|
|
|
| 317 |
"qformer_num_heads": 16,
|
| 318 |
"qformer_num_layers": 2,
|
| 319 |
"qformer_window_size": 15,
|
| 320 |
+
"repetition_penalty": 1.1,
|
| 321 |
"router_aux_loss_coef": 0.01,
|
| 322 |
"system_prompt": "",
|
| 323 |
"text_config": {
|