Update config.json
Browse files- config.json +3 -3
config.json
CHANGED
|
@@ -45,10 +45,10 @@
|
|
| 45 |
"ctc_loss_reduction": "sum",
|
| 46 |
"ctc_zero_infinity": false,
|
| 47 |
"diversity_loss_weight": 0.1,
|
| 48 |
-
"do_stable_layer_norm":
|
| 49 |
"eos_token_id": 2,
|
| 50 |
"feat_extract_activation": "gelu",
|
| 51 |
-
"feat_extract_dropout": 0.
|
| 52 |
"feat_extract_norm": "group",
|
| 53 |
"feat_proj_dropout": 0.0,
|
| 54 |
"feat_quantizer_dropout": 0.0,
|
|
@@ -108,4 +108,4 @@
|
|
| 108 |
"use_weighted_layer_sum": false,
|
| 109 |
"vocab_size": 32,
|
| 110 |
"xvector_output_dim": 512
|
| 111 |
-
}
|
|
|
|
| 45 |
"ctc_loss_reduction": "sum",
|
| 46 |
"ctc_zero_infinity": false,
|
| 47 |
"diversity_loss_weight": 0.1,
|
| 48 |
+
"do_stable_layer_norm": false,
|
| 49 |
"eos_token_id": 2,
|
| 50 |
"feat_extract_activation": "gelu",
|
| 51 |
+
"feat_extract_dropout": 0.05,
|
| 52 |
"feat_extract_norm": "group",
|
| 53 |
"feat_proj_dropout": 0.0,
|
| 54 |
"feat_quantizer_dropout": 0.0,
|
|
|
|
| 108 |
"use_weighted_layer_sum": false,
|
| 109 |
"vocab_size": 32,
|
| 110 |
"xvector_output_dim": 512
|
| 111 |
+
}
|