hsali commited on
Commit
904374f
·
1 Parent(s): 00ad005

Training in progress, step 100

Browse files
Files changed (3) hide show
  1. config.json +15 -16
  2. pytorch_model.bin +2 -2
  3. training_args.bin +2 -2
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-base-960h",
3
- "activation_dropout": 0.1,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
@@ -11,9 +11,9 @@
11
  "attention_dropout": 0.1,
12
  "bos_token_id": 1,
13
  "classifier_proj_size": 256,
14
- "codevector_dim": 256,
15
  "contrastive_logits_temperature": 0.1,
16
- "conv_bias": false,
17
  "conv_dim": [
18
  512,
19
  512,
@@ -44,20 +44,19 @@
44
  "ctc_loss_reduction": "sum",
45
  "ctc_zero_infinity": false,
46
  "diversity_loss_weight": 0.1,
47
- "do_stable_layer_norm": false,
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
50
  "feat_extract_dropout": 0.0,
51
- "feat_extract_norm": "group",
52
  "feat_proj_dropout": 0.1,
53
  "feat_quantizer_dropout": 0.0,
54
- "final_dropout": 0.1,
55
  "finetuning_task": "wav2vec2_clf",
56
  "gradient_checkpointing": false,
57
  "hidden_act": "gelu",
58
  "hidden_dropout": 0.1,
59
- "hidden_dropout_prob": 0.1,
60
- "hidden_size": 768,
61
  "id2label": {
62
  "0": "Angry",
63
  "1": "Happy",
@@ -65,7 +64,7 @@
65
  "3": "Sad"
66
  },
67
  "initializer_range": 0.02,
68
- "intermediate_size": 3072,
69
  "label2id": {
70
  "Angry": 0,
71
  "Happy": 1,
@@ -79,22 +78,22 @@
79
  "mask_feature_prob": 0.0,
80
  "mask_time_length": 10,
81
  "mask_time_min_masks": 2,
82
- "mask_time_prob": 0.05,
83
  "model_type": "wav2vec2",
84
  "num_adapter_layers": 3,
85
- "num_attention_heads": 12,
86
  "num_codevector_groups": 2,
87
  "num_codevectors_per_group": 320,
88
  "num_conv_pos_embedding_groups": 16,
89
  "num_conv_pos_embeddings": 128,
90
  "num_feat_extract_layers": 7,
91
- "num_hidden_layers": 12,
92
  "num_negatives": 100,
93
- "output_hidden_size": 768,
94
  "pad_token_id": 0,
95
  "pooling_mode": "mean",
96
  "problem_type": "single_label_classification",
97
- "proj_codevector_dim": 256,
98
  "tdnn_dilation": [
99
  1,
100
  2,
@@ -117,7 +116,7 @@
117
  1
118
  ],
119
  "torch_dtype": "float32",
120
- "transformers_version": "4.26.1",
121
  "use_weighted_layer_sum": false,
122
  "vocab_size": 32,
123
  "xvector_output_dim": 512
 
1
  {
2
+ "_name_or_path": "alefiury/wav2vec2-large-xlsr-53-gender-recognition-librispeech",
3
+ "activation_dropout": 0.0,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
 
11
  "attention_dropout": 0.1,
12
  "bos_token_id": 1,
13
  "classifier_proj_size": 256,
14
+ "codevector_dim": 768,
15
  "contrastive_logits_temperature": 0.1,
16
+ "conv_bias": true,
17
  "conv_dim": [
18
  512,
19
  512,
 
44
  "ctc_loss_reduction": "sum",
45
  "ctc_zero_infinity": false,
46
  "diversity_loss_weight": 0.1,
47
+ "do_stable_layer_norm": true,
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
50
  "feat_extract_dropout": 0.0,
51
+ "feat_extract_norm": "layer",
52
  "feat_proj_dropout": 0.1,
53
  "feat_quantizer_dropout": 0.0,
54
+ "final_dropout": 0.0,
55
  "finetuning_task": "wav2vec2_clf",
56
  "gradient_checkpointing": false,
57
  "hidden_act": "gelu",
58
  "hidden_dropout": 0.1,
59
+ "hidden_size": 1024,
 
60
  "id2label": {
61
  "0": "Angry",
62
  "1": "Happy",
 
64
  "3": "Sad"
65
  },
66
  "initializer_range": 0.02,
67
+ "intermediate_size": 4096,
68
  "label2id": {
69
  "Angry": 0,
70
  "Happy": 1,
 
78
  "mask_feature_prob": 0.0,
79
  "mask_time_length": 10,
80
  "mask_time_min_masks": 2,
81
+ "mask_time_prob": 0.075,
82
  "model_type": "wav2vec2",
83
  "num_adapter_layers": 3,
84
+ "num_attention_heads": 16,
85
  "num_codevector_groups": 2,
86
  "num_codevectors_per_group": 320,
87
  "num_conv_pos_embedding_groups": 16,
88
  "num_conv_pos_embeddings": 128,
89
  "num_feat_extract_layers": 7,
90
+ "num_hidden_layers": 24,
91
  "num_negatives": 100,
92
+ "output_hidden_size": 1024,
93
  "pad_token_id": 0,
94
  "pooling_mode": "mean",
95
  "problem_type": "single_label_classification",
96
+ "proj_codevector_dim": 768,
97
  "tdnn_dilation": [
98
  1,
99
  2,
 
116
  1
117
  ],
118
  "torch_dtype": "float32",
119
+ "transformers_version": "4.28.1",
120
  "use_weighted_layer_sum": false,
121
  "vocab_size": 32,
122
  "xvector_output_dim": 512
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d23533a4e305253195d7f5926b32bfc39b97643a3865757e9b287609eda91cf
3
- size 377572641
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98243549b682bf63c56e5c6cbed6bd7f9e72bac8c4f41f7f99e3039db6811e76
3
+ size 1261918317
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a1538e152df0cf4611c6309dcbe58f04f2d77ca97b0334e2292e4aafa97bc36
3
- size 3515
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30992d7bca6a6ef6b049960b6f7ad1b6f29208512a1482c7ea4c003fc5d7603a
3
+ size 3579