increase model
Browse files- config.json +18 -3
- pytorch_model.bin +2 -2
config.json
CHANGED
|
@@ -10,15 +10,30 @@
|
|
| 10 |
"classifier_proj_size": 256,
|
| 11 |
"conv_bias": false,
|
| 12 |
"conv_dim": [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
512,
|
| 14 |
512
|
| 15 |
],
|
| 16 |
"conv_kernel": [
|
| 17 |
10,
|
| 18 |
-
3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
],
|
| 20 |
"conv_stride": [
|
| 21 |
5,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
2
|
| 23 |
],
|
| 24 |
"ctc_loss_reduction": "sum",
|
|
@@ -47,11 +62,11 @@
|
|
| 47 |
"num_attention_heads": 12,
|
| 48 |
"num_conv_pos_embedding_groups": 16,
|
| 49 |
"num_conv_pos_embeddings": 128,
|
| 50 |
-
"num_feat_extract_layers":
|
| 51 |
"num_hidden_layers": 1,
|
| 52 |
"pad_token_id": 0,
|
| 53 |
"torch_dtype": "float32",
|
| 54 |
-
"transformers_version": "4.
|
| 55 |
"use_weighted_layer_sum": false,
|
| 56 |
"vocab_size": 32
|
| 57 |
}
|
|
|
|
| 10 |
"classifier_proj_size": 256,
|
| 11 |
"conv_bias": false,
|
| 12 |
"conv_dim": [
|
| 13 |
+
512,
|
| 14 |
+
512,
|
| 15 |
+
512,
|
| 16 |
+
512,
|
| 17 |
+
512,
|
| 18 |
512,
|
| 19 |
512
|
| 20 |
],
|
| 21 |
"conv_kernel": [
|
| 22 |
10,
|
| 23 |
+
3,
|
| 24 |
+
3,
|
| 25 |
+
3,
|
| 26 |
+
3,
|
| 27 |
+
2,
|
| 28 |
+
2
|
| 29 |
],
|
| 30 |
"conv_stride": [
|
| 31 |
5,
|
| 32 |
+
2,
|
| 33 |
+
2,
|
| 34 |
+
2,
|
| 35 |
+
2,
|
| 36 |
+
2,
|
| 37 |
2
|
| 38 |
],
|
| 39 |
"ctc_loss_reduction": "sum",
|
|
|
|
| 62 |
"num_attention_heads": 12,
|
| 63 |
"num_conv_pos_embedding_groups": 16,
|
| 64 |
"num_conv_pos_embeddings": 128,
|
| 65 |
+
"num_feat_extract_layers": 7,
|
| 66 |
"num_hidden_layers": 1,
|
| 67 |
"pad_token_id": 0,
|
| 68 |
"torch_dtype": "float32",
|
| 69 |
+
"transformers_version": "4.26.1",
|
| 70 |
"use_weighted_layer_sum": false,
|
| 71 |
"vocab_size": 32
|
| 72 |
}
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97cce607ce1d2a01dda6deca8b1398eaa2d10d52cc28d4d8fae9bfb6b711cf4f
|
| 3 |
+
size 65626772
|