mazesmazes commited on
Commit
0beba06
·
verified ·
1 Parent(s): a4007a7

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -280,7 +280,7 @@
280
  "freq_mask_length": 27,
281
  "inference_warmup_tokens": 10,
282
  "label_smoothing": 0.0,
283
- "length_penalty": 1.0,
284
  "llm_dim": 2048,
285
  "lora_alpha": 32,
286
  "lora_dropout": 0.0,
@@ -294,11 +294,11 @@
294
  "up_proj",
295
  "down_proj"
296
  ],
297
- "max_new_tokens": 128,
298
  "min_new_tokens": 0,
299
  "model_dtype": "bfloat16",
300
  "model_type": "asr_model",
301
- "no_repeat_ngram_size": 6,
302
  "num_beams": 1,
303
  "num_experts": 4,
304
  "num_experts_per_tok": 2,
@@ -317,7 +317,7 @@
317
  "qformer_num_heads": 16,
318
  "qformer_num_layers": 2,
319
  "qformer_window_size": 15,
320
- "repetition_penalty": 1.05,
321
  "router_aux_loss_coef": 0.01,
322
  "system_prompt": "",
323
  "text_config": {
 
280
  "freq_mask_length": 27,
281
  "inference_warmup_tokens": 10,
282
  "label_smoothing": 0.0,
283
+ "length_penalty": 0.8,
284
  "llm_dim": 2048,
285
  "lora_alpha": 32,
286
  "lora_dropout": 0.0,
 
294
  "up_proj",
295
  "down_proj"
296
  ],
297
+ "max_new_tokens": 96,
298
  "min_new_tokens": 0,
299
  "model_dtype": "bfloat16",
300
  "model_type": "asr_model",
301
+ "no_repeat_ngram_size": 4,
302
  "num_beams": 1,
303
  "num_experts": 4,
304
  "num_experts_per_tok": 2,
 
317
  "qformer_num_heads": 16,
318
  "qformer_num_layers": 2,
319
  "qformer_window_size": 15,
320
+ "repetition_penalty": 1.1,
321
  "router_aux_loss_coef": 0.01,
322
  "system_prompt": "",
323
  "text_config": {