pgilles commited on
Commit
4207ab2
·
verified ·
1 Parent(s): f3fc5c5

Upload fine-tuned HUBERT vowel classifier

Browse files
Files changed (43) hide show
  1. .DS_Store +0 -0
  2. checkpoints/.DS_Store +0 -0
  3. checkpoints/checkpoint-10916/config.json +98 -0
  4. checkpoints/checkpoint-10916/model.safetensors +3 -0
  5. checkpoints/checkpoint-10916/optimizer.pt +3 -0
  6. checkpoints/checkpoint-10916/rng_state.pth +3 -0
  7. checkpoints/checkpoint-10916/scheduler.pt +3 -0
  8. checkpoints/checkpoint-10916/trainer_state.json +98 -0
  9. checkpoints/checkpoint-10916/training_args.bin +3 -0
  10. checkpoints/checkpoint-13645/config.json +98 -0
  11. checkpoints/checkpoint-13645/model.safetensors +3 -0
  12. checkpoints/checkpoint-13645/optimizer.pt +3 -0
  13. checkpoints/checkpoint-13645/rng_state.pth +3 -0
  14. checkpoints/checkpoint-13645/scheduler.pt +3 -0
  15. checkpoints/checkpoint-13645/trainer_state.json +114 -0
  16. checkpoints/checkpoint-13645/training_args.bin +3 -0
  17. checkpoints/checkpoint-2729/config.json +98 -0
  18. checkpoints/checkpoint-2729/model.safetensors +3 -0
  19. checkpoints/checkpoint-2729/optimizer.pt +3 -0
  20. checkpoints/checkpoint-2729/rng_state.pth +3 -0
  21. checkpoints/checkpoint-2729/scheduler.pt +3 -0
  22. checkpoints/checkpoint-2729/trainer_state.json +50 -0
  23. checkpoints/checkpoint-2729/training_args.bin +3 -0
  24. checkpoints/checkpoint-5458/config.json +98 -0
  25. checkpoints/checkpoint-5458/model.safetensors +3 -0
  26. checkpoints/checkpoint-5458/optimizer.pt +3 -0
  27. checkpoints/checkpoint-5458/rng_state.pth +3 -0
  28. checkpoints/checkpoint-5458/scheduler.pt +3 -0
  29. checkpoints/checkpoint-5458/trainer_state.json +66 -0
  30. checkpoints/checkpoint-5458/training_args.bin +3 -0
  31. checkpoints/checkpoint-8187/config.json +98 -0
  32. checkpoints/checkpoint-8187/model.safetensors +3 -0
  33. checkpoints/checkpoint-8187/optimizer.pt +3 -0
  34. checkpoints/checkpoint-8187/rng_state.pth +3 -0
  35. checkpoints/checkpoint-8187/scheduler.pt +3 -0
  36. checkpoints/checkpoint-8187/trainer_state.json +82 -0
  37. checkpoints/checkpoint-8187/training_args.bin +3 -0
  38. classification_report.csv +13 -0
  39. config.json +10 -2
  40. confusion_matrix.png +0 -0
  41. label_encoder.pkl +2 -2
  42. model.safetensors +2 -2
  43. model_info.txt +4 -0
.DS_Store ADDED
Binary file (6.15 kB). View file
 
checkpoints/.DS_Store ADDED
Binary file (8.2 kB). View file
 
checkpoints/checkpoint-10916/config.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "HubertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 1,
9
+ "classifier_proj_size": 256,
10
+ "conv_bias": false,
11
+ "conv_dim": [
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512
19
+ ],
20
+ "conv_kernel": [
21
+ 10,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 2,
27
+ 2
28
+ ],
29
+ "conv_pos_batch_norm": false,
30
+ "conv_stride": [
31
+ 5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2
38
+ ],
39
+ "ctc_loss_reduction": "sum",
40
+ "ctc_zero_infinity": false,
41
+ "do_stable_layer_norm": false,
42
+ "eos_token_id": 2,
43
+ "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
+ "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
+ "hidden_act": "gelu",
51
+ "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
+ "hidden_size": 768,
54
+ "id2label": {
55
+ "0": "LABEL_0",
56
+ "1": "LABEL_1",
57
+ "2": "LABEL_2",
58
+ "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
+ },
65
+ "initializer_range": 0.02,
66
+ "intermediate_size": 3072,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1,
70
+ "LABEL_2": 2,
71
+ "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
+ },
78
+ "layer_norm_eps": 1e-05,
79
+ "layerdrop": 0.1,
80
+ "mask_feature_length": 10,
81
+ "mask_feature_min_masks": 0,
82
+ "mask_feature_prob": 0.0,
83
+ "mask_time_length": 10,
84
+ "mask_time_min_masks": 2,
85
+ "mask_time_prob": 0.0,
86
+ "model_type": "hubert",
87
+ "num_attention_heads": 12,
88
+ "num_conv_pos_embedding_groups": 16,
89
+ "num_conv_pos_embeddings": 128,
90
+ "num_feat_extract_layers": 7,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.51.3",
96
+ "use_weighted_layer_sum": false,
97
+ "vocab_size": 32
98
+ }
checkpoints/checkpoint-10916/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a467102e2eca83185336bcd6488e3eee01fbba8c4a49e8074cfde44a0d53266d
3
+ size 378309148
checkpoints/checkpoint-10916/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f192fc14c22f672367cb1dd14d62c7216e9dc8debf99e6cf19a762c37b3a631f
3
+ size 756734475
checkpoints/checkpoint-10916/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d08a5484f52cbb805c81b0a3ac668e47c0bb3f8fb5a45803e8cf1d5c9364a6
3
+ size 14455
checkpoints/checkpoint-10916/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b6b3d374d162562d7181e35af847e336046d647cafbf1ee9adf8e902f093a19
3
+ size 1465
checkpoints/checkpoint-10916/trainer_state.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 10916,
3
+ "best_metric": 0.8916987355689939,
4
+ "best_model_checkpoint": "./fine_tuned_vowel_model_hubert/checkpoints/checkpoint-10916",
5
+ "epoch": 4.0,
6
+ "eval_steps": 500,
7
+ "global_step": 10916,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 0.7431244850158691,
15
+ "learning_rate": 4.0003664345914255e-05,
16
+ "loss": 0.7026,
17
+ "step": 2729
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_accuracy": 0.8449697636063771,
22
+ "eval_loss": 0.5954396724700928,
23
+ "eval_runtime": 13.4904,
24
+ "eval_samples_per_second": 404.511,
25
+ "eval_steps_per_second": 50.629,
26
+ "step": 2729
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 0.1394549012184143,
31
+ "learning_rate": 3.0003664345914256e-05,
32
+ "loss": 0.4608,
33
+ "step": 5458
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8792376763789628,
38
+ "eval_loss": 0.5260357856750488,
39
+ "eval_runtime": 12.9279,
40
+ "eval_samples_per_second": 422.112,
41
+ "eval_steps_per_second": 52.832,
42
+ "step": 5458
43
+ },
44
+ {
45
+ "epoch": 3.0,
46
+ "grad_norm": 116.29024505615234,
47
+ "learning_rate": 2.0003664345914254e-05,
48
+ "loss": 0.371,
49
+ "step": 8187
50
+ },
51
+ {
52
+ "epoch": 3.0,
53
+ "eval_accuracy": 0.8887667216419278,
54
+ "eval_loss": 0.5250139236450195,
55
+ "eval_runtime": 12.4693,
56
+ "eval_samples_per_second": 437.635,
57
+ "eval_steps_per_second": 54.775,
58
+ "step": 8187
59
+ },
60
+ {
61
+ "epoch": 4.0,
62
+ "grad_norm": 18.14418601989746,
63
+ "learning_rate": 1.0003664345914255e-05,
64
+ "loss": 0.2928,
65
+ "step": 10916
66
+ },
67
+ {
68
+ "epoch": 4.0,
69
+ "eval_accuracy": 0.8916987355689939,
70
+ "eval_loss": 0.57878178358078,
71
+ "eval_runtime": 12.1012,
72
+ "eval_samples_per_second": 450.946,
73
+ "eval_steps_per_second": 56.441,
74
+ "step": 10916
75
+ }
76
+ ],
77
+ "logging_steps": 500,
78
+ "max_steps": 13645,
79
+ "num_input_tokens_seen": 0,
80
+ "num_train_epochs": 5,
81
+ "save_steps": 500,
82
+ "stateful_callbacks": {
83
+ "TrainerControl": {
84
+ "args": {
85
+ "should_epoch_stop": false,
86
+ "should_evaluate": false,
87
+ "should_log": false,
88
+ "should_save": true,
89
+ "should_training_stop": false
90
+ },
91
+ "attributes": {}
92
+ }
93
+ },
94
+ "total_flos": 1.803131179533748e+17,
95
+ "train_batch_size": 8,
96
+ "trial_name": null,
97
+ "trial_params": null
98
+ }
checkpoints/checkpoint-10916/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3178195344f46f46b3031b675585d8c03b499e7d32bb87bf40496a15e9e1126
3
+ size 5713
checkpoints/checkpoint-13645/config.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "HubertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 1,
9
+ "classifier_proj_size": 256,
10
+ "conv_bias": false,
11
+ "conv_dim": [
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512
19
+ ],
20
+ "conv_kernel": [
21
+ 10,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 2,
27
+ 2
28
+ ],
29
+ "conv_pos_batch_norm": false,
30
+ "conv_stride": [
31
+ 5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2
38
+ ],
39
+ "ctc_loss_reduction": "sum",
40
+ "ctc_zero_infinity": false,
41
+ "do_stable_layer_norm": false,
42
+ "eos_token_id": 2,
43
+ "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
+ "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
+ "hidden_act": "gelu",
51
+ "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
+ "hidden_size": 768,
54
+ "id2label": {
55
+ "0": "LABEL_0",
56
+ "1": "LABEL_1",
57
+ "2": "LABEL_2",
58
+ "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
+ },
65
+ "initializer_range": 0.02,
66
+ "intermediate_size": 3072,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1,
70
+ "LABEL_2": 2,
71
+ "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
+ },
78
+ "layer_norm_eps": 1e-05,
79
+ "layerdrop": 0.1,
80
+ "mask_feature_length": 10,
81
+ "mask_feature_min_masks": 0,
82
+ "mask_feature_prob": 0.0,
83
+ "mask_time_length": 10,
84
+ "mask_time_min_masks": 2,
85
+ "mask_time_prob": 0.0,
86
+ "model_type": "hubert",
87
+ "num_attention_heads": 12,
88
+ "num_conv_pos_embedding_groups": 16,
89
+ "num_conv_pos_embeddings": 128,
90
+ "num_feat_extract_layers": 7,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.51.3",
96
+ "use_weighted_layer_sum": false,
97
+ "vocab_size": 32
98
+ }
checkpoints/checkpoint-13645/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e253002c57e573c2a7d2c6bb92948a0532b9471527eec6e154d4b71b819e6b7
3
+ size 378309148
checkpoints/checkpoint-13645/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:173b8dc5d365bb15274565588dc12293988d45da5ab409bd7a207d74fde20c26
3
+ size 756734475
checkpoints/checkpoint-13645/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea68a63b52eb6548d8b3f161d296b8f31e6c69810c9d3267393657d6694b91a0
3
+ size 14455
checkpoints/checkpoint-13645/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b58bab2079a1e60f13455dd02b5b0d37795a4e3bd7bee0f5e83d482f80f5668
3
+ size 1465
checkpoints/checkpoint-13645/trainer_state.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 13645,
3
+ "best_metric": 0.8944474986256185,
4
+ "best_model_checkpoint": "./fine_tuned_vowel_model_hubert/checkpoints/checkpoint-13645",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 13645,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 0.7431244850158691,
15
+ "learning_rate": 4.0003664345914255e-05,
16
+ "loss": 0.7026,
17
+ "step": 2729
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_accuracy": 0.8449697636063771,
22
+ "eval_loss": 0.5954396724700928,
23
+ "eval_runtime": 13.4904,
24
+ "eval_samples_per_second": 404.511,
25
+ "eval_steps_per_second": 50.629,
26
+ "step": 2729
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 0.1394549012184143,
31
+ "learning_rate": 3.0003664345914256e-05,
32
+ "loss": 0.4608,
33
+ "step": 5458
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8792376763789628,
38
+ "eval_loss": 0.5260357856750488,
39
+ "eval_runtime": 12.9279,
40
+ "eval_samples_per_second": 422.112,
41
+ "eval_steps_per_second": 52.832,
42
+ "step": 5458
43
+ },
44
+ {
45
+ "epoch": 3.0,
46
+ "grad_norm": 116.29024505615234,
47
+ "learning_rate": 2.0003664345914254e-05,
48
+ "loss": 0.371,
49
+ "step": 8187
50
+ },
51
+ {
52
+ "epoch": 3.0,
53
+ "eval_accuracy": 0.8887667216419278,
54
+ "eval_loss": 0.5250139236450195,
55
+ "eval_runtime": 12.4693,
56
+ "eval_samples_per_second": 437.635,
57
+ "eval_steps_per_second": 54.775,
58
+ "step": 8187
59
+ },
60
+ {
61
+ "epoch": 4.0,
62
+ "grad_norm": 18.14418601989746,
63
+ "learning_rate": 1.0003664345914255e-05,
64
+ "loss": 0.2928,
65
+ "step": 10916
66
+ },
67
+ {
68
+ "epoch": 4.0,
69
+ "eval_accuracy": 0.8916987355689939,
70
+ "eval_loss": 0.57878178358078,
71
+ "eval_runtime": 12.1012,
72
+ "eval_samples_per_second": 450.946,
73
+ "eval_steps_per_second": 56.441,
74
+ "step": 10916
75
+ },
76
+ {
77
+ "epoch": 5.0,
78
+ "grad_norm": 0.16541193425655365,
79
+ "learning_rate": 3.664345914254306e-09,
80
+ "loss": 0.2333,
81
+ "step": 13645
82
+ },
83
+ {
84
+ "epoch": 5.0,
85
+ "eval_accuracy": 0.8944474986256185,
86
+ "eval_loss": 0.5925989747047424,
87
+ "eval_runtime": 12.2917,
88
+ "eval_samples_per_second": 443.957,
89
+ "eval_steps_per_second": 55.566,
90
+ "step": 13645
91
+ }
92
+ ],
93
+ "logging_steps": 500,
94
+ "max_steps": 13645,
95
+ "num_input_tokens_seen": 0,
96
+ "num_train_epochs": 5,
97
+ "save_steps": 500,
98
+ "stateful_callbacks": {
99
+ "TrainerControl": {
100
+ "args": {
101
+ "should_epoch_stop": false,
102
+ "should_evaluate": false,
103
+ "should_log": false,
104
+ "should_save": true,
105
+ "should_training_stop": true
106
+ },
107
+ "attributes": {}
108
+ }
109
+ },
110
+ "total_flos": 2.2541662744176947e+17,
111
+ "train_batch_size": 8,
112
+ "trial_name": null,
113
+ "trial_params": null
114
+ }
checkpoints/checkpoint-13645/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3178195344f46f46b3031b675585d8c03b499e7d32bb87bf40496a15e9e1126
3
+ size 5713
checkpoints/checkpoint-2729/config.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "HubertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 1,
9
+ "classifier_proj_size": 256,
10
+ "conv_bias": false,
11
+ "conv_dim": [
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512
19
+ ],
20
+ "conv_kernel": [
21
+ 10,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 2,
27
+ 2
28
+ ],
29
+ "conv_pos_batch_norm": false,
30
+ "conv_stride": [
31
+ 5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2
38
+ ],
39
+ "ctc_loss_reduction": "sum",
40
+ "ctc_zero_infinity": false,
41
+ "do_stable_layer_norm": false,
42
+ "eos_token_id": 2,
43
+ "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
+ "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
+ "hidden_act": "gelu",
51
+ "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
+ "hidden_size": 768,
54
+ "id2label": {
55
+ "0": "LABEL_0",
56
+ "1": "LABEL_1",
57
+ "2": "LABEL_2",
58
+ "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
+ },
65
+ "initializer_range": 0.02,
66
+ "intermediate_size": 3072,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1,
70
+ "LABEL_2": 2,
71
+ "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
+ },
78
+ "layer_norm_eps": 1e-05,
79
+ "layerdrop": 0.1,
80
+ "mask_feature_length": 10,
81
+ "mask_feature_min_masks": 0,
82
+ "mask_feature_prob": 0.0,
83
+ "mask_time_length": 10,
84
+ "mask_time_min_masks": 2,
85
+ "mask_time_prob": 0.0,
86
+ "model_type": "hubert",
87
+ "num_attention_heads": 12,
88
+ "num_conv_pos_embedding_groups": 16,
89
+ "num_conv_pos_embeddings": 128,
90
+ "num_feat_extract_layers": 7,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.51.3",
96
+ "use_weighted_layer_sum": false,
97
+ "vocab_size": 32
98
+ }
checkpoints/checkpoint-2729/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6613875495014ea4d25c4f91f396ea1c979cf5c5f6dc1a06500f15fa8c1377
3
+ size 378309148
checkpoints/checkpoint-2729/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b43504597ecfcb1b7e5956e4341b676c8e6198b6f1a79891b11742222b4e876
3
+ size 756734475
checkpoints/checkpoint-2729/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea200907bf22a7285702a90c82fddb0bc9b345082a2ba718466e46ec3f624aaa
3
+ size 14455
checkpoints/checkpoint-2729/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4170acc0e84df36b983756c50956178d160d8904ae138886c0fb61ec4c5ba556
3
+ size 1465
checkpoints/checkpoint-2729/trainer_state.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 2729,
3
+ "best_metric": 0.8449697636063771,
4
+ "best_model_checkpoint": "./fine_tuned_vowel_model_hubert/checkpoints/checkpoint-2729",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 2729,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 0.7431244850158691,
15
+ "learning_rate": 4.0003664345914255e-05,
16
+ "loss": 0.7026,
17
+ "step": 2729
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_accuracy": 0.8449697636063771,
22
+ "eval_loss": 0.5954396724700928,
23
+ "eval_runtime": 13.4904,
24
+ "eval_samples_per_second": 404.511,
25
+ "eval_steps_per_second": 50.629,
26
+ "step": 2729
27
+ }
28
+ ],
29
+ "logging_steps": 500,
30
+ "max_steps": 13645,
31
+ "num_input_tokens_seen": 0,
32
+ "num_train_epochs": 5,
33
+ "save_steps": 500,
34
+ "stateful_callbacks": {
35
+ "TrainerControl": {
36
+ "args": {
37
+ "should_epoch_stop": false,
38
+ "should_evaluate": false,
39
+ "should_log": false,
40
+ "should_save": true,
41
+ "should_training_stop": false
42
+ },
43
+ "attributes": {}
44
+ }
45
+ },
46
+ "total_flos": 4.505194187404075e+16,
47
+ "train_batch_size": 8,
48
+ "trial_name": null,
49
+ "trial_params": null
50
+ }
checkpoints/checkpoint-2729/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3178195344f46f46b3031b675585d8c03b499e7d32bb87bf40496a15e9e1126
3
+ size 5713
checkpoints/checkpoint-5458/config.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "HubertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 1,
9
+ "classifier_proj_size": 256,
10
+ "conv_bias": false,
11
+ "conv_dim": [
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512
19
+ ],
20
+ "conv_kernel": [
21
+ 10,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 2,
27
+ 2
28
+ ],
29
+ "conv_pos_batch_norm": false,
30
+ "conv_stride": [
31
+ 5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2
38
+ ],
39
+ "ctc_loss_reduction": "sum",
40
+ "ctc_zero_infinity": false,
41
+ "do_stable_layer_norm": false,
42
+ "eos_token_id": 2,
43
+ "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
+ "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
+ "hidden_act": "gelu",
51
+ "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
+ "hidden_size": 768,
54
+ "id2label": {
55
+ "0": "LABEL_0",
56
+ "1": "LABEL_1",
57
+ "2": "LABEL_2",
58
+ "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
+ },
65
+ "initializer_range": 0.02,
66
+ "intermediate_size": 3072,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1,
70
+ "LABEL_2": 2,
71
+ "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
+ },
78
+ "layer_norm_eps": 1e-05,
79
+ "layerdrop": 0.1,
80
+ "mask_feature_length": 10,
81
+ "mask_feature_min_masks": 0,
82
+ "mask_feature_prob": 0.0,
83
+ "mask_time_length": 10,
84
+ "mask_time_min_masks": 2,
85
+ "mask_time_prob": 0.0,
86
+ "model_type": "hubert",
87
+ "num_attention_heads": 12,
88
+ "num_conv_pos_embedding_groups": 16,
89
+ "num_conv_pos_embeddings": 128,
90
+ "num_feat_extract_layers": 7,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.51.3",
96
+ "use_weighted_layer_sum": false,
97
+ "vocab_size": 32
98
+ }
checkpoints/checkpoint-5458/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:027c836889b9c6dff7fcd845a987026d85f20a6de1a4acf5b7cab7f875d147c2
3
+ size 378309148
checkpoints/checkpoint-5458/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a9ac2caf3feb93de932dc17c3dbada05d9bbdf09492f508999045ed85bc69c8
3
+ size 756734475
checkpoints/checkpoint-5458/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e5067e2c971581835c48c8b92efce57577a076f65e6270573f525825609e404
3
+ size 14455
checkpoints/checkpoint-5458/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99abf3f8d618bf535394d7da5d2a6f328625f877f33ac77786a006c0a8b59a30
3
+ size 1465
checkpoints/checkpoint-5458/trainer_state.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 5458,
3
+ "best_metric": 0.8792376763789628,
4
+ "best_model_checkpoint": "./fine_tuned_vowel_model_hubert/checkpoints/checkpoint-5458",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 5458,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 0.7431244850158691,
15
+ "learning_rate": 4.0003664345914255e-05,
16
+ "loss": 0.7026,
17
+ "step": 2729
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_accuracy": 0.8449697636063771,
22
+ "eval_loss": 0.5954396724700928,
23
+ "eval_runtime": 13.4904,
24
+ "eval_samples_per_second": 404.511,
25
+ "eval_steps_per_second": 50.629,
26
+ "step": 2729
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 0.1394549012184143,
31
+ "learning_rate": 3.0003664345914256e-05,
32
+ "loss": 0.4608,
33
+ "step": 5458
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8792376763789628,
38
+ "eval_loss": 0.5260357856750488,
39
+ "eval_runtime": 12.9279,
40
+ "eval_samples_per_second": 422.112,
41
+ "eval_steps_per_second": 52.832,
42
+ "step": 5458
43
+ }
44
+ ],
45
+ "logging_steps": 500,
46
+ "max_steps": 13645,
47
+ "num_input_tokens_seen": 0,
48
+ "num_train_epochs": 5,
49
+ "save_steps": 500,
50
+ "stateful_callbacks": {
51
+ "TrainerControl": {
52
+ "args": {
53
+ "should_epoch_stop": false,
54
+ "should_evaluate": false,
55
+ "should_log": false,
56
+ "should_save": true,
57
+ "should_training_stop": false
58
+ },
59
+ "attributes": {}
60
+ }
61
+ },
62
+ "total_flos": 9.019020502929226e+16,
63
+ "train_batch_size": 8,
64
+ "trial_name": null,
65
+ "trial_params": null
66
+ }
checkpoints/checkpoint-5458/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3178195344f46f46b3031b675585d8c03b499e7d32bb87bf40496a15e9e1126
3
+ size 5713
checkpoints/checkpoint-8187/config.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "HubertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 1,
9
+ "classifier_proj_size": 256,
10
+ "conv_bias": false,
11
+ "conv_dim": [
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512
19
+ ],
20
+ "conv_kernel": [
21
+ 10,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 3,
26
+ 2,
27
+ 2
28
+ ],
29
+ "conv_pos_batch_norm": false,
30
+ "conv_stride": [
31
+ 5,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2
38
+ ],
39
+ "ctc_loss_reduction": "sum",
40
+ "ctc_zero_infinity": false,
41
+ "do_stable_layer_norm": false,
42
+ "eos_token_id": 2,
43
+ "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
+ "feat_extract_norm": "group",
46
+ "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
+ "hidden_act": "gelu",
51
+ "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
+ "hidden_size": 768,
54
+ "id2label": {
55
+ "0": "LABEL_0",
56
+ "1": "LABEL_1",
57
+ "2": "LABEL_2",
58
+ "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
+ },
65
+ "initializer_range": 0.02,
66
+ "intermediate_size": 3072,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1,
70
+ "LABEL_2": 2,
71
+ "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
+ },
78
+ "layer_norm_eps": 1e-05,
79
+ "layerdrop": 0.1,
80
+ "mask_feature_length": 10,
81
+ "mask_feature_min_masks": 0,
82
+ "mask_feature_prob": 0.0,
83
+ "mask_time_length": 10,
84
+ "mask_time_min_masks": 2,
85
+ "mask_time_prob": 0.0,
86
+ "model_type": "hubert",
87
+ "num_attention_heads": 12,
88
+ "num_conv_pos_embedding_groups": 16,
89
+ "num_conv_pos_embeddings": 128,
90
+ "num_feat_extract_layers": 7,
91
+ "num_hidden_layers": 12,
92
+ "pad_token_id": 0,
93
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
94
+ "torch_dtype": "float32",
95
+ "transformers_version": "4.51.3",
96
+ "use_weighted_layer_sum": false,
97
+ "vocab_size": 32
98
+ }
checkpoints/checkpoint-8187/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65c6145dc7273172cd887b4d26f71429024a6adf4e7a654073744dcfa4d53f42
3
+ size 378309148
checkpoints/checkpoint-8187/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc441edf7cfcb4f8563751cf5f60d4f792c210c86bce9580403e60e86b54d91
3
+ size 756734475
checkpoints/checkpoint-8187/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19a8e6f0588fe3d1f50ae60492a5ada2beb4972a53e6e5a6bd68c3c671f5091c
3
+ size 14455
checkpoints/checkpoint-8187/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a223eaee1486f9a44419fc63bd3824ba990f5207fac40cf8a423f1da10ea85d
3
+ size 1465
checkpoints/checkpoint-8187/trainer_state.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 8187,
3
+ "best_metric": 0.8887667216419278,
4
+ "best_model_checkpoint": "./fine_tuned_vowel_model_hubert/checkpoints/checkpoint-8187",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 8187,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "grad_norm": 0.7431244850158691,
15
+ "learning_rate": 4.0003664345914255e-05,
16
+ "loss": 0.7026,
17
+ "step": 2729
18
+ },
19
+ {
20
+ "epoch": 1.0,
21
+ "eval_accuracy": 0.8449697636063771,
22
+ "eval_loss": 0.5954396724700928,
23
+ "eval_runtime": 13.4904,
24
+ "eval_samples_per_second": 404.511,
25
+ "eval_steps_per_second": 50.629,
26
+ "step": 2729
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 0.1394549012184143,
31
+ "learning_rate": 3.0003664345914256e-05,
32
+ "loss": 0.4608,
33
+ "step": 5458
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8792376763789628,
38
+ "eval_loss": 0.5260357856750488,
39
+ "eval_runtime": 12.9279,
40
+ "eval_samples_per_second": 422.112,
41
+ "eval_steps_per_second": 52.832,
42
+ "step": 5458
43
+ },
44
+ {
45
+ "epoch": 3.0,
46
+ "grad_norm": 116.29024505615234,
47
+ "learning_rate": 2.0003664345914254e-05,
48
+ "loss": 0.371,
49
+ "step": 8187
50
+ },
51
+ {
52
+ "epoch": 3.0,
53
+ "eval_accuracy": 0.8887667216419278,
54
+ "eval_loss": 0.5250139236450195,
55
+ "eval_runtime": 12.4693,
56
+ "eval_samples_per_second": 437.635,
57
+ "eval_steps_per_second": 54.775,
58
+ "step": 8187
59
+ }
60
+ ],
61
+ "logging_steps": 500,
62
+ "max_steps": 13645,
63
+ "num_input_tokens_seen": 0,
64
+ "num_train_epochs": 5,
65
+ "save_steps": 500,
66
+ "stateful_callbacks": {
67
+ "TrainerControl": {
68
+ "args": {
69
+ "should_epoch_stop": false,
70
+ "should_evaluate": false,
71
+ "should_log": false,
72
+ "should_save": true,
73
+ "should_training_stop": false
74
+ },
75
+ "attributes": {}
76
+ }
77
+ },
78
+ "total_flos": 1.3531804571600885e+17,
79
+ "train_batch_size": 8,
80
+ "trial_name": null,
81
+ "trial_params": null
82
+ }
checkpoints/checkpoint-8187/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3178195344f46f46b3031b675585d8c03b499e7d32bb87bf40496a15e9e1126
3
+ size 5713
classification_report.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ,precision,recall,f1-score,support
2
+ aː,0.9426791277258567,0.9322242760320394,0.9374225526641884,1623.0
3
+ eː,0.9323308270676691,0.9127234490010515,0.922422954303932,951.0
4
+ oː,0.9166666666666666,0.9166666666666666,0.9166666666666666,720.0
5
+ æːɪ,0.8868421052631579,0.8753246753246753,0.8810457516339869,385.0
6
+ æːʊ,0.6046511627906976,0.7602339181286549,0.6735751295336787,171.0
7
+ ɑɪ,0.9067164179104478,0.9101123595505618,0.908411214953271,534.0
8
+ ɑʊ,0.6820512820512821,0.6244131455399061,0.6519607843137255,213.0
9
+ əʊ,0.8215962441314554,0.7990867579908676,0.8101851851851852,219.0
10
+ ɜɪ,0.8746223564954683,0.9032761310452418,0.88871834228703,641.0
11
+ accuracy,0.8944474986256185,0.8944474986256185,0.8944474986256185,0.8944474986256185
12
+ macro avg,0.8409062433447446,0.8482290421421851,0.8433787312824071,5457.0
13
+ weighted avg,0.8963662082505828,0.8944474986256185,0.8950161705793978,5457.0
config.json CHANGED
@@ -56,7 +56,11 @@
56
  "1": "LABEL_1",
57
  "2": "LABEL_2",
58
  "3": "LABEL_3",
59
- "4": "LABEL_4"
 
 
 
 
60
  },
61
  "initializer_range": 0.02,
62
  "intermediate_size": 3072,
@@ -65,7 +69,11 @@
65
  "LABEL_1": 1,
66
  "LABEL_2": 2,
67
  "LABEL_3": 3,
68
- "LABEL_4": 4
 
 
 
 
69
  },
70
  "layer_norm_eps": 1e-05,
71
  "layerdrop": 0.1,
 
56
  "1": "LABEL_1",
57
  "2": "LABEL_2",
58
  "3": "LABEL_3",
59
+ "4": "LABEL_4",
60
+ "5": "LABEL_5",
61
+ "6": "LABEL_6",
62
+ "7": "LABEL_7",
63
+ "8": "LABEL_8"
64
  },
65
  "initializer_range": 0.02,
66
  "intermediate_size": 3072,
 
69
  "LABEL_1": 1,
70
  "LABEL_2": 2,
71
  "LABEL_3": 3,
72
+ "LABEL_4": 4,
73
+ "LABEL_5": 5,
74
+ "LABEL_6": 6,
75
+ "LABEL_7": 7,
76
+ "LABEL_8": 8
77
  },
78
  "layer_norm_eps": 1e-05,
79
  "layerdrop": 0.1,
confusion_matrix.png ADDED
label_encoder.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ef72578413775b35281cd23cb205c790ca54e3220ffd3d4264ac5bf66401618
3
- size 387
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64b6954f7c1e908e23e0f621392e14c14da84254105a7cfa928932f3830d2e5e
3
+ size 435
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:381af9f09d62497a79d0237b630945c2afbaba77bb4fa83af4a31814d604982f
3
- size 378305028
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e253002c57e573c2a7d2c6bb92948a0532b9471527eec6e154d4b71b819e6b7
3
+ size 378309148
model_info.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ model_type: hubert
2
+ model_name: facebook/hubert-base-ls960
3
+ num_labels: 9
4
+ labels: aː, eː, oː, æːɪ, æːʊ, ɑɪ, ɑʊ, əʊ, ɜɪ