TheAIchemist13 commited on
Commit
0bf44cf
·
1 Parent(s): 85a12be

Delete whisper-tiny-hi

Browse files
whisper-tiny-hi/added_tokens.json DELETED
@@ -1,109 +0,0 @@
1
- {
2
- "<|af|>": 50327,
3
- "<|am|>": 50334,
4
- "<|ar|>": 50272,
5
- "<|as|>": 50350,
6
- "<|az|>": 50304,
7
- "<|ba|>": 50355,
8
- "<|be|>": 50330,
9
- "<|bg|>": 50292,
10
- "<|bn|>": 50302,
11
- "<|bo|>": 50347,
12
- "<|br|>": 50309,
13
- "<|bs|>": 50315,
14
- "<|ca|>": 50270,
15
- "<|cs|>": 50283,
16
- "<|cy|>": 50297,
17
- "<|da|>": 50285,
18
- "<|de|>": 50261,
19
- "<|el|>": 50281,
20
- "<|endoftext|>": 50257,
21
- "<|en|>": 50259,
22
- "<|es|>": 50262,
23
- "<|et|>": 50307,
24
- "<|eu|>": 50310,
25
- "<|fa|>": 50300,
26
- "<|fi|>": 50277,
27
- "<|fo|>": 50338,
28
- "<|fr|>": 50265,
29
- "<|gl|>": 50319,
30
- "<|gu|>": 50333,
31
- "<|haw|>": 50352,
32
- "<|ha|>": 50354,
33
- "<|hi|>": 50276,
34
- "<|hr|>": 50291,
35
- "<|ht|>": 50339,
36
- "<|hu|>": 50286,
37
- "<|hy|>": 50312,
38
- "<|id|>": 50275,
39
- "<|is|>": 50311,
40
- "<|it|>": 50274,
41
- "<|iw|>": 50279,
42
- "<|ja|>": 50266,
43
- "<|jw|>": 50356,
44
- "<|ka|>": 50329,
45
- "<|kk|>": 50316,
46
- "<|km|>": 50323,
47
- "<|kn|>": 50306,
48
- "<|ko|>": 50264,
49
- "<|la|>": 50294,
50
- "<|lb|>": 50345,
51
- "<|ln|>": 50353,
52
- "<|lo|>": 50336,
53
- "<|lt|>": 50293,
54
- "<|lv|>": 50301,
55
- "<|mg|>": 50349,
56
- "<|mi|>": 50295,
57
- "<|mk|>": 50308,
58
- "<|ml|>": 50296,
59
- "<|mn|>": 50314,
60
- "<|mr|>": 50320,
61
- "<|ms|>": 50282,
62
- "<|mt|>": 50343,
63
- "<|my|>": 50346,
64
- "<|ne|>": 50313,
65
- "<|nl|>": 50271,
66
- "<|nn|>": 50342,
67
- "<|nocaptions|>": 50362,
68
- "<|notimestamps|>": 50363,
69
- "<|no|>": 50288,
70
- "<|oc|>": 50328,
71
- "<|pa|>": 50321,
72
- "<|pl|>": 50269,
73
- "<|ps|>": 50340,
74
- "<|pt|>": 50267,
75
- "<|ro|>": 50284,
76
- "<|ru|>": 50263,
77
- "<|sa|>": 50344,
78
- "<|sd|>": 50332,
79
- "<|si|>": 50322,
80
- "<|sk|>": 50298,
81
- "<|sl|>": 50305,
82
- "<|sn|>": 50324,
83
- "<|so|>": 50326,
84
- "<|sq|>": 50317,
85
- "<|sr|>": 50303,
86
- "<|startoflm|>": 50360,
87
- "<|startofprev|>": 50361,
88
- "<|startoftranscript|>": 50258,
89
- "<|su|>": 50357,
90
- "<|sv|>": 50273,
91
- "<|sw|>": 50318,
92
- "<|ta|>": 50287,
93
- "<|te|>": 50299,
94
- "<|tg|>": 50331,
95
- "<|th|>": 50289,
96
- "<|tk|>": 50341,
97
- "<|tl|>": 50348,
98
- "<|transcribe|>": 50359,
99
- "<|translate|>": 50358,
100
- "<|tr|>": 50268,
101
- "<|tt|>": 50351,
102
- "<|uk|>": 50280,
103
- "<|ur|>": 50290,
104
- "<|uz|>": 50337,
105
- "<|vi|>": 50278,
106
- "<|yi|>": 50335,
107
- "<|yo|>": 50325,
108
- "<|zh|>": 50260
109
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/config.json DELETED
@@ -1,64 +0,0 @@
1
- {
2
- "_name_or_path": "gramvaani_whisper-tiny-hi/",
3
- "activation_dropout": 0.0,
4
- "activation_function": "gelu",
5
- "apply_spec_augment": false,
6
- "architectures": [
7
- "PoptorchPipelinedWhisperForConditionalGeneration"
8
- ],
9
- "attention_dropout": 0.0,
10
- "begin_suppress_tokens": [
11
- 220,
12
- 50257
13
- ],
14
- "bos_token_id": 50257,
15
- "classifier_proj_size": 256,
16
- "d_model": 384,
17
- "decoder_attention_heads": 6,
18
- "decoder_ffn_dim": 1536,
19
- "decoder_layerdrop": 0.0,
20
- "decoder_layers": 4,
21
- "decoder_start_token_id": 50258,
22
- "dropout": 0.0,
23
- "encoder_attention_heads": 6,
24
- "encoder_ffn_dim": 1536,
25
- "encoder_layerdrop": 0.0,
26
- "encoder_layers": 4,
27
- "eos_token_id": 50257,
28
- "forced_decoder_ids": [
29
- [
30
- 1,
31
- 50306
32
- ],
33
- [
34
- 2,
35
- 50359
36
- ],
37
- [
38
- 3,
39
- 50363
40
- ]
41
- ],
42
- "init_std": 0.02,
43
- "is_encoder_decoder": true,
44
- "mask_feature_length": 10,
45
- "mask_feature_min_masks": 0,
46
- "mask_feature_prob": 0.0,
47
- "mask_time_length": 10,
48
- "mask_time_min_masks": 2,
49
- "mask_time_prob": 0.05,
50
- "max_length": 448,
51
- "max_source_positions": 1500,
52
- "max_target_positions": 448,
53
- "model_type": "whisper",
54
- "num_hidden_layers": 4,
55
- "num_mel_bins": 80,
56
- "pad_token_id": 50257,
57
- "scale_embedding": false,
58
- "suppress_tokens": [],
59
- "torch_dtype": "float16",
60
- "transformers_version": "4.29.2",
61
- "use_cache": true,
62
- "use_weighted_layer_sum": false,
63
- "vocab_size": 51865
64
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/generation_config.json DELETED
@@ -1,162 +0,0 @@
1
- {
2
- "alignment_heads": [
3
- [
4
- 2,
5
- 2
6
- ],
7
- [
8
- 3,
9
- 0
10
- ],
11
- [
12
- 3,
13
- 2
14
- ],
15
- [
16
- 3,
17
- 3
18
- ],
19
- [
20
- 3,
21
- 4
22
- ],
23
- [
24
- 3,
25
- 5
26
- ]
27
- ],
28
- "begin_suppress_tokens": [
29
- 220,
30
- 50257
31
- ],
32
- "bos_token_id": 50257,
33
- "decoder_start_token_id": 50258,
34
- "eos_token_id": 50257,
35
- "forced_decoder_ids": [
36
- [
37
- 1,
38
- 50306
39
- ],
40
- [
41
- 2,
42
- 50359
43
- ],
44
- [
45
- 3,
46
- 50363
47
- ]
48
- ],
49
- "is_multilingual": true,
50
- "lang_to_id": {
51
- "<|af|>": 50327,
52
- "<|am|>": 50334,
53
- "<|ar|>": 50272,
54
- "<|as|>": 50350,
55
- "<|az|>": 50304,
56
- "<|ba|>": 50355,
57
- "<|be|>": 50330,
58
- "<|bg|>": 50292,
59
- "<|bn|>": 50302,
60
- "<|bo|>": 50347,
61
- "<|br|>": 50309,
62
- "<|bs|>": 50315,
63
- "<|ca|>": 50270,
64
- "<|cs|>": 50283,
65
- "<|cy|>": 50297,
66
- "<|da|>": 50285,
67
- "<|de|>": 50261,
68
- "<|el|>": 50281,
69
- "<|en|>": 50259,
70
- "<|es|>": 50262,
71
- "<|et|>": 50307,
72
- "<|eu|>": 50310,
73
- "<|fa|>": 50300,
74
- "<|fi|>": 50277,
75
- "<|fo|>": 50338,
76
- "<|fr|>": 50265,
77
- "<|gl|>": 50319,
78
- "<|gu|>": 50333,
79
- "<|haw|>": 50352,
80
- "<|ha|>": 50354,
81
- "<|he|>": 50279,
82
- "<|hi|>": 50276,
83
- "<|hr|>": 50291,
84
- "<|ht|>": 50339,
85
- "<|hu|>": 50286,
86
- "<|hy|>": 50312,
87
- "<|id|>": 50275,
88
- "<|is|>": 50311,
89
- "<|it|>": 50274,
90
- "<|ja|>": 50266,
91
- "<|jw|>": 50356,
92
- "<|ka|>": 50329,
93
- "<|kk|>": 50316,
94
- "<|km|>": 50323,
95
- "<|kn|>": 50306,
96
- "<|ko|>": 50264,
97
- "<|la|>": 50294,
98
- "<|lb|>": 50345,
99
- "<|ln|>": 50353,
100
- "<|lo|>": 50336,
101
- "<|lt|>": 50293,
102
- "<|lv|>": 50301,
103
- "<|mg|>": 50349,
104
- "<|mi|>": 50295,
105
- "<|mk|>": 50308,
106
- "<|ml|>": 50296,
107
- "<|mn|>": 50314,
108
- "<|mr|>": 50320,
109
- "<|ms|>": 50282,
110
- "<|mt|>": 50343,
111
- "<|my|>": 50346,
112
- "<|ne|>": 50313,
113
- "<|nl|>": 50271,
114
- "<|nn|>": 50342,
115
- "<|no|>": 50288,
116
- "<|oc|>": 50328,
117
- "<|pa|>": 50321,
118
- "<|pl|>": 50269,
119
- "<|ps|>": 50340,
120
- "<|pt|>": 50267,
121
- "<|ro|>": 50284,
122
- "<|ru|>": 50263,
123
- "<|sa|>": 50344,
124
- "<|sd|>": 50332,
125
- "<|si|>": 50322,
126
- "<|sk|>": 50298,
127
- "<|sl|>": 50305,
128
- "<|sn|>": 50324,
129
- "<|so|>": 50326,
130
- "<|sq|>": 50317,
131
- "<|sr|>": 50303,
132
- "<|su|>": 50357,
133
- "<|sv|>": 50273,
134
- "<|sw|>": 50318,
135
- "<|ta|>": 50287,
136
- "<|te|>": 50299,
137
- "<|tg|>": 50331,
138
- "<|th|>": 50289,
139
- "<|tk|>": 50341,
140
- "<|tl|>": 50348,
141
- "<|tr|>": 50268,
142
- "<|tt|>": 50351,
143
- "<|uk|>": 50280,
144
- "<|ur|>": 50290,
145
- "<|uz|>": 50337,
146
- "<|vi|>": 50278,
147
- "<|yi|>": 50335,
148
- "<|yo|>": 50325,
149
- "<|zh|>": 50260
150
- },
151
- "max_initial_timestamp_index": 1,
152
- "max_length": 448,
153
- "no_timestamps_token_id": 50363,
154
- "pad_token_id": 50257,
155
- "return_timestamps": false,
156
- "suppress_tokens": [],
157
- "task_to_id": {
158
- "transcribe": 50359,
159
- "translate": 50358
160
- },
161
- "transformers_version": "4.29.2"
162
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/ipu_config.json DELETED
@@ -1,52 +0,0 @@
1
- {
2
- "auto_loss_scaling": false,
3
- "device_iterations": 1,
4
- "embedding_serialization_factor": 1,
5
- "enable_half_partials": true,
6
- "executable_cache_dir": "/tmp/exe_cache/3.3.0/whisper",
7
- "execute_encoder_on_cpu_for_generation": false,
8
- "explicit_ir_inference": false,
9
- "gradient_accumulation_steps": 16,
10
- "inference_device_iterations": 1,
11
- "inference_embedding_serialization_factor": 1,
12
- "inference_ipus_per_replica": 2,
13
- "inference_layers_per_ipu": [
14
- 4,
15
- 4
16
- ],
17
- "inference_matmul_proportion": 0.2,
18
- "inference_parallelize_kwargs": {
19
- "on_device_generation_steps": 16,
20
- "use_cache": true,
21
- "use_encoder_output_buffer": true
22
- },
23
- "inference_projection_serialization_factor": 1,
24
- "inference_replication_factor": 1,
25
- "inference_serialized_embedding_splits_per_ipu": null,
26
- "inference_serialized_projection_splits_per_ipu": null,
27
- "ipus_per_replica": 4,
28
- "layers_per_ipu": [
29
- 2,
30
- 2,
31
- 2,
32
- 2
33
- ],
34
- "matmul_proportion": [
35
- 0.2,
36
- 0.2,
37
- 0.6,
38
- 0.6
39
- ],
40
- "optimizer_state_offchip": true,
41
- "optimum_version": "1.6.1",
42
- "output_mode": "final",
43
- "parallelize_kwargs": {},
44
- "projection_serialization_factor": 5,
45
- "recompute_checkpoint_every_layer": true,
46
- "replicated_tensor_sharding": false,
47
- "replication_factor": 1,
48
- "seed": 42,
49
- "serialized_embedding_splits_per_ipu": null,
50
- "serialized_projection_splits_per_ipu": null,
51
- "transformers_version": "4.29.2"
52
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef9c679fa9b88e934215ab0f367d0bc6263f5493171b9073e0ba11586d1d2315
3
- size 226951674
 
 
 
 
whisper-tiny-hi/preprocessor_config.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "chunk_length": 30,
3
- "feature_extractor_type": "WhisperFeatureExtractorTorch",
4
- "feature_size": 80,
5
- "hop_length": 160,
6
- "n_fft": 400,
7
- "n_samples": 480000,
8
- "nb_max_frames": 3000,
9
- "padding_side": "right",
10
- "padding_value": 0.0,
11
- "processor_class": "WhisperProcessor",
12
- "return_attention_mask": false,
13
- "sampling_rate": 16000
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a2854f73f9ef2f46f2c4bc68bd6762fbf831cb835a73696c56159d7d31a53d9
3
- size 75575337
 
 
 
 
whisper-tiny-hi/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c4b1c6fe240d828d6a82f29d5c2eb45e891bc57e996613001ed46f6c29e6cb0
3
- size 13553
 
 
 
 
whisper-tiny-hi/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fef5942c3187fa7545fb0a9e98dc4d1f03bad33f5d9f88bc966ff9ff7f43b540
3
- size 627
 
 
 
 
whisper-tiny-hi/trainer_state.json DELETED
@@ -1,1013 +0,0 @@
1
- {
2
- "best_metric": 0.0,
3
- "best_model_checkpoint": "./whisper-tiny-kannada/checkpoint-1000",
4
- "epoch": 666.6666666666666,
5
- "global_step": 4000,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 4.17,
12
- "learning_rate": 2.5000000000000004e-07,
13
- "loss": 1.8762,
14
- "step": 25
15
- },
16
- {
17
- "epoch": 8.33,
18
- "learning_rate": 5.000000000000001e-07,
19
- "loss": 1.5055,
20
- "step": 50
21
- },
22
- {
23
- "epoch": 12.5,
24
- "learning_rate": 7.5e-07,
25
- "loss": 1.1952,
26
- "step": 75
27
- },
28
- {
29
- "epoch": 16.67,
30
- "learning_rate": 1.0000000000000002e-06,
31
- "loss": 0.8302,
32
- "step": 100
33
- },
34
- {
35
- "epoch": 20.83,
36
- "learning_rate": 1.25e-06,
37
- "loss": 0.4201,
38
- "step": 125
39
- },
40
- {
41
- "epoch": 25.0,
42
- "learning_rate": 1.5e-06,
43
- "loss": 0.1929,
44
- "step": 150
45
- },
46
- {
47
- "epoch": 29.17,
48
- "learning_rate": 1.75e-06,
49
- "loss": 0.0998,
50
- "step": 175
51
- },
52
- {
53
- "epoch": 33.33,
54
- "learning_rate": 2.0000000000000003e-06,
55
- "loss": 0.1073,
56
- "step": 200
57
- },
58
- {
59
- "epoch": 37.5,
60
- "learning_rate": 2.25e-06,
61
- "loss": 0.0581,
62
- "step": 225
63
- },
64
- {
65
- "epoch": 41.67,
66
- "learning_rate": 2.5e-06,
67
- "loss": 0.0408,
68
- "step": 250
69
- },
70
- {
71
- "epoch": 45.83,
72
- "learning_rate": 2.7500000000000004e-06,
73
- "loss": 0.0472,
74
- "step": 275
75
- },
76
- {
77
- "epoch": 50.0,
78
- "learning_rate": 3e-06,
79
- "loss": 0.0283,
80
- "step": 300
81
- },
82
- {
83
- "epoch": 54.17,
84
- "learning_rate": 3.2500000000000002e-06,
85
- "loss": 0.0233,
86
- "step": 325
87
- },
88
- {
89
- "epoch": 58.33,
90
- "learning_rate": 3.5e-06,
91
- "loss": 0.0166,
92
- "step": 350
93
- },
94
- {
95
- "epoch": 62.5,
96
- "learning_rate": 3.7500000000000005e-06,
97
- "loss": 0.014,
98
- "step": 375
99
- },
100
- {
101
- "epoch": 66.67,
102
- "learning_rate": 4.000000000000001e-06,
103
- "loss": 0.0092,
104
- "step": 400
105
- },
106
- {
107
- "epoch": 70.83,
108
- "learning_rate": 4.25e-06,
109
- "loss": 0.0099,
110
- "step": 425
111
- },
112
- {
113
- "epoch": 75.0,
114
- "learning_rate": 4.5e-06,
115
- "loss": 0.0102,
116
- "step": 450
117
- },
118
- {
119
- "epoch": 79.17,
120
- "learning_rate": 4.75e-06,
121
- "loss": 0.0067,
122
- "step": 475
123
- },
124
- {
125
- "epoch": 83.33,
126
- "learning_rate": 5e-06,
127
- "loss": 0.0062,
128
- "step": 500
129
- },
130
- {
131
- "epoch": 87.5,
132
- "learning_rate": 5.2500000000000006e-06,
133
- "loss": 0.0056,
134
- "step": 525
135
- },
136
- {
137
- "epoch": 91.67,
138
- "learning_rate": 5.500000000000001e-06,
139
- "loss": 0.0044,
140
- "step": 550
141
- },
142
- {
143
- "epoch": 95.83,
144
- "learning_rate": 5.75e-06,
145
- "loss": 0.0043,
146
- "step": 575
147
- },
148
- {
149
- "epoch": 100.0,
150
- "learning_rate": 6e-06,
151
- "loss": 0.0041,
152
- "step": 600
153
- },
154
- {
155
- "epoch": 104.17,
156
- "learning_rate": 6.25e-06,
157
- "loss": 0.0038,
158
- "step": 625
159
- },
160
- {
161
- "epoch": 108.33,
162
- "learning_rate": 6.5000000000000004e-06,
163
- "loss": 0.0035,
164
- "step": 650
165
- },
166
- {
167
- "epoch": 112.5,
168
- "learning_rate": 6.750000000000001e-06,
169
- "loss": 0.0029,
170
- "step": 675
171
- },
172
- {
173
- "epoch": 116.67,
174
- "learning_rate": 7e-06,
175
- "loss": 0.0026,
176
- "step": 700
177
- },
178
- {
179
- "epoch": 120.83,
180
- "learning_rate": 7.25e-06,
181
- "loss": 0.0025,
182
- "step": 725
183
- },
184
- {
185
- "epoch": 125.0,
186
- "learning_rate": 7.500000000000001e-06,
187
- "loss": 0.002,
188
- "step": 750
189
- },
190
- {
191
- "epoch": 129.17,
192
- "learning_rate": 7.75e-06,
193
- "loss": 0.0016,
194
- "step": 775
195
- },
196
- {
197
- "epoch": 133.33,
198
- "learning_rate": 8.000000000000001e-06,
199
- "loss": 0.0018,
200
- "step": 800
201
- },
202
- {
203
- "epoch": 137.5,
204
- "learning_rate": 8.25e-06,
205
- "loss": 0.0018,
206
- "step": 825
207
- },
208
- {
209
- "epoch": 141.67,
210
- "learning_rate": 8.5e-06,
211
- "loss": 0.0014,
212
- "step": 850
213
- },
214
- {
215
- "epoch": 145.83,
216
- "learning_rate": 8.750000000000001e-06,
217
- "loss": 0.0015,
218
- "step": 875
219
- },
220
- {
221
- "epoch": 150.0,
222
- "learning_rate": 9e-06,
223
- "loss": 0.0013,
224
- "step": 900
225
- },
226
- {
227
- "epoch": 154.17,
228
- "learning_rate": 9.250000000000001e-06,
229
- "loss": 0.001,
230
- "step": 925
231
- },
232
- {
233
- "epoch": 158.33,
234
- "learning_rate": 9.5e-06,
235
- "loss": 0.0011,
236
- "step": 950
237
- },
238
- {
239
- "epoch": 162.5,
240
- "learning_rate": 9.75e-06,
241
- "loss": 0.0011,
242
- "step": 975
243
- },
244
- {
245
- "epoch": 166.67,
246
- "learning_rate": 1e-05,
247
- "loss": 0.0009,
248
- "step": 1000
249
- },
250
- {
251
- "epoch": 166.67,
252
- "eval_normalized_wer": 0.0,
253
- "eval_runtime": 132.3018,
254
- "eval_samples_per_second": 0.189,
255
- "eval_steps_per_second": 0.189,
256
- "eval_wer": 0.0,
257
- "step": 1000
258
- },
259
- {
260
- "epoch": 170.83,
261
- "learning_rate": 9.916666666666668e-06,
262
- "loss": 0.0009,
263
- "step": 1025
264
- },
265
- {
266
- "epoch": 175.0,
267
- "learning_rate": 9.833333333333333e-06,
268
- "loss": 0.0008,
269
- "step": 1050
270
- },
271
- {
272
- "epoch": 179.17,
273
- "learning_rate": 9.75e-06,
274
- "loss": 0.001,
275
- "step": 1075
276
- },
277
- {
278
- "epoch": 183.33,
279
- "learning_rate": 9.666666666666667e-06,
280
- "loss": 0.0008,
281
- "step": 1100
282
- },
283
- {
284
- "epoch": 187.5,
285
- "learning_rate": 9.583333333333335e-06,
286
- "loss": 0.0007,
287
- "step": 1125
288
- },
289
- {
290
- "epoch": 191.67,
291
- "learning_rate": 9.5e-06,
292
- "loss": 0.0009,
293
- "step": 1150
294
- },
295
- {
296
- "epoch": 195.83,
297
- "learning_rate": 9.416666666666667e-06,
298
- "loss": 0.0007,
299
- "step": 1175
300
- },
301
- {
302
- "epoch": 200.0,
303
- "learning_rate": 9.333333333333334e-06,
304
- "loss": 0.0007,
305
- "step": 1200
306
- },
307
- {
308
- "epoch": 204.17,
309
- "learning_rate": 9.250000000000001e-06,
310
- "loss": 0.0006,
311
- "step": 1225
312
- },
313
- {
314
- "epoch": 208.33,
315
- "learning_rate": 9.166666666666666e-06,
316
- "loss": 0.0005,
317
- "step": 1250
318
- },
319
- {
320
- "epoch": 212.5,
321
- "learning_rate": 9.083333333333333e-06,
322
- "loss": 0.0007,
323
- "step": 1275
324
- },
325
- {
326
- "epoch": 216.67,
327
- "learning_rate": 9e-06,
328
- "loss": 0.0005,
329
- "step": 1300
330
- },
331
- {
332
- "epoch": 220.83,
333
- "learning_rate": 8.916666666666667e-06,
334
- "loss": 0.0005,
335
- "step": 1325
336
- },
337
- {
338
- "epoch": 225.0,
339
- "learning_rate": 8.833333333333334e-06,
340
- "loss": 0.0006,
341
- "step": 1350
342
- },
343
- {
344
- "epoch": 229.17,
345
- "learning_rate": 8.750000000000001e-06,
346
- "loss": 0.0005,
347
- "step": 1375
348
- },
349
- {
350
- "epoch": 233.33,
351
- "learning_rate": 8.666666666666668e-06,
352
- "loss": 0.0005,
353
- "step": 1400
354
- },
355
- {
356
- "epoch": 237.5,
357
- "learning_rate": 8.583333333333333e-06,
358
- "loss": 0.0005,
359
- "step": 1425
360
- },
361
- {
362
- "epoch": 241.67,
363
- "learning_rate": 8.5e-06,
364
- "loss": 0.0005,
365
- "step": 1450
366
- },
367
- {
368
- "epoch": 245.83,
369
- "learning_rate": 8.416666666666667e-06,
370
- "loss": 0.0004,
371
- "step": 1475
372
- },
373
- {
374
- "epoch": 250.0,
375
- "learning_rate": 8.333333333333334e-06,
376
- "loss": 0.0005,
377
- "step": 1500
378
- },
379
- {
380
- "epoch": 254.17,
381
- "learning_rate": 8.25e-06,
382
- "loss": 0.0005,
383
- "step": 1525
384
- },
385
- {
386
- "epoch": 258.33,
387
- "learning_rate": 8.166666666666668e-06,
388
- "loss": 0.0004,
389
- "step": 1550
390
- },
391
- {
392
- "epoch": 262.5,
393
- "learning_rate": 8.083333333333334e-06,
394
- "loss": 0.0004,
395
- "step": 1575
396
- },
397
- {
398
- "epoch": 266.67,
399
- "learning_rate": 8.000000000000001e-06,
400
- "loss": 0.0004,
401
- "step": 1600
402
- },
403
- {
404
- "epoch": 270.83,
405
- "learning_rate": 7.916666666666667e-06,
406
- "loss": 0.0004,
407
- "step": 1625
408
- },
409
- {
410
- "epoch": 275.0,
411
- "learning_rate": 7.833333333333333e-06,
412
- "loss": 0.0004,
413
- "step": 1650
414
- },
415
- {
416
- "epoch": 279.17,
417
- "learning_rate": 7.75e-06,
418
- "loss": 0.0003,
419
- "step": 1675
420
- },
421
- {
422
- "epoch": 283.33,
423
- "learning_rate": 7.666666666666667e-06,
424
- "loss": 0.0004,
425
- "step": 1700
426
- },
427
- {
428
- "epoch": 287.5,
429
- "learning_rate": 7.583333333333333e-06,
430
- "loss": 0.0003,
431
- "step": 1725
432
- },
433
- {
434
- "epoch": 291.67,
435
- "learning_rate": 7.500000000000001e-06,
436
- "loss": 0.0004,
437
- "step": 1750
438
- },
439
- {
440
- "epoch": 295.83,
441
- "learning_rate": 7.416666666666668e-06,
442
- "loss": 0.0003,
443
- "step": 1775
444
- },
445
- {
446
- "epoch": 300.0,
447
- "learning_rate": 7.333333333333333e-06,
448
- "loss": 0.0003,
449
- "step": 1800
450
- },
451
- {
452
- "epoch": 304.17,
453
- "learning_rate": 7.25e-06,
454
- "loss": 0.0004,
455
- "step": 1825
456
- },
457
- {
458
- "epoch": 308.33,
459
- "learning_rate": 7.166666666666667e-06,
460
- "loss": 0.0003,
461
- "step": 1850
462
- },
463
- {
464
- "epoch": 312.5,
465
- "learning_rate": 7.083333333333335e-06,
466
- "loss": 0.0003,
467
- "step": 1875
468
- },
469
- {
470
- "epoch": 316.67,
471
- "learning_rate": 7e-06,
472
- "loss": 0.0003,
473
- "step": 1900
474
- },
475
- {
476
- "epoch": 320.83,
477
- "learning_rate": 6.916666666666667e-06,
478
- "loss": 0.0003,
479
- "step": 1925
480
- },
481
- {
482
- "epoch": 325.0,
483
- "learning_rate": 6.833333333333334e-06,
484
- "loss": 0.0003,
485
- "step": 1950
486
- },
487
- {
488
- "epoch": 329.17,
489
- "learning_rate": 6.750000000000001e-06,
490
- "loss": 0.0003,
491
- "step": 1975
492
- },
493
- {
494
- "epoch": 333.33,
495
- "learning_rate": 6.666666666666667e-06,
496
- "loss": 0.0003,
497
- "step": 2000
498
- },
499
- {
500
- "epoch": 333.33,
501
- "eval_normalized_wer": 0.0,
502
- "eval_runtime": 12.7785,
503
- "eval_samples_per_second": 1.956,
504
- "eval_steps_per_second": 1.956,
505
- "eval_wer": 0.0,
506
- "step": 2000
507
- },
508
- {
509
- "epoch": 337.5,
510
- "learning_rate": 6.5833333333333335e-06,
511
- "loss": 0.0003,
512
- "step": 2025
513
- },
514
- {
515
- "epoch": 341.67,
516
- "learning_rate": 6.5000000000000004e-06,
517
- "loss": 0.0003,
518
- "step": 2050
519
- },
520
- {
521
- "epoch": 345.83,
522
- "learning_rate": 6.416666666666667e-06,
523
- "loss": 0.0003,
524
- "step": 2075
525
- },
526
- {
527
- "epoch": 350.0,
528
- "learning_rate": 6.333333333333333e-06,
529
- "loss": 0.0003,
530
- "step": 2100
531
- },
532
- {
533
- "epoch": 354.17,
534
- "learning_rate": 6.25e-06,
535
- "loss": 0.0003,
536
- "step": 2125
537
- },
538
- {
539
- "epoch": 358.33,
540
- "learning_rate": 6.166666666666667e-06,
541
- "loss": 0.0002,
542
- "step": 2150
543
- },
544
- {
545
- "epoch": 362.5,
546
- "learning_rate": 6.083333333333333e-06,
547
- "loss": 0.0003,
548
- "step": 2175
549
- },
550
- {
551
- "epoch": 366.67,
552
- "learning_rate": 6e-06,
553
- "loss": 0.0003,
554
- "step": 2200
555
- },
556
- {
557
- "epoch": 370.83,
558
- "learning_rate": 5.916666666666667e-06,
559
- "loss": 0.0003,
560
- "step": 2225
561
- },
562
- {
563
- "epoch": 375.0,
564
- "learning_rate": 5.833333333333334e-06,
565
- "loss": 0.0003,
566
- "step": 2250
567
- },
568
- {
569
- "epoch": 379.17,
570
- "learning_rate": 5.75e-06,
571
- "loss": 0.0002,
572
- "step": 2275
573
- },
574
- {
575
- "epoch": 383.33,
576
- "learning_rate": 5.666666666666667e-06,
577
- "loss": 0.0002,
578
- "step": 2300
579
- },
580
- {
581
- "epoch": 387.5,
582
- "learning_rate": 5.583333333333334e-06,
583
- "loss": 0.0003,
584
- "step": 2325
585
- },
586
- {
587
- "epoch": 391.67,
588
- "learning_rate": 5.500000000000001e-06,
589
- "loss": 0.0002,
590
- "step": 2350
591
- },
592
- {
593
- "epoch": 395.83,
594
- "learning_rate": 5.416666666666667e-06,
595
- "loss": 0.0002,
596
- "step": 2375
597
- },
598
- {
599
- "epoch": 400.0,
600
- "learning_rate": 5.333333333333334e-06,
601
- "loss": 0.0002,
602
- "step": 2400
603
- },
604
- {
605
- "epoch": 404.17,
606
- "learning_rate": 5.2500000000000006e-06,
607
- "loss": 0.0002,
608
- "step": 2425
609
- },
610
- {
611
- "epoch": 408.33,
612
- "learning_rate": 5.1666666666666675e-06,
613
- "loss": 0.0002,
614
- "step": 2450
615
- },
616
- {
617
- "epoch": 412.5,
618
- "learning_rate": 5.0833333333333335e-06,
619
- "loss": 0.0002,
620
- "step": 2475
621
- },
622
- {
623
- "epoch": 416.67,
624
- "learning_rate": 5e-06,
625
- "loss": 0.0002,
626
- "step": 2500
627
- },
628
- {
629
- "epoch": 420.83,
630
- "learning_rate": 4.9166666666666665e-06,
631
- "loss": 0.0002,
632
- "step": 2525
633
- },
634
- {
635
- "epoch": 425.0,
636
- "learning_rate": 4.833333333333333e-06,
637
- "loss": 0.0002,
638
- "step": 2550
639
- },
640
- {
641
- "epoch": 429.17,
642
- "learning_rate": 4.75e-06,
643
- "loss": 0.0002,
644
- "step": 2575
645
- },
646
- {
647
- "epoch": 433.33,
648
- "learning_rate": 4.666666666666667e-06,
649
- "loss": 0.0002,
650
- "step": 2600
651
- },
652
- {
653
- "epoch": 437.5,
654
- "learning_rate": 4.583333333333333e-06,
655
- "loss": 0.0002,
656
- "step": 2625
657
- },
658
- {
659
- "epoch": 441.67,
660
- "learning_rate": 4.5e-06,
661
- "loss": 0.0002,
662
- "step": 2650
663
- },
664
- {
665
- "epoch": 445.83,
666
- "learning_rate": 4.416666666666667e-06,
667
- "loss": 0.0002,
668
- "step": 2675
669
- },
670
- {
671
- "epoch": 450.0,
672
- "learning_rate": 4.333333333333334e-06,
673
- "loss": 0.0002,
674
- "step": 2700
675
- },
676
- {
677
- "epoch": 454.17,
678
- "learning_rate": 4.25e-06,
679
- "loss": 0.0002,
680
- "step": 2725
681
- },
682
- {
683
- "epoch": 458.33,
684
- "learning_rate": 4.166666666666667e-06,
685
- "loss": 0.0002,
686
- "step": 2750
687
- },
688
- {
689
- "epoch": 462.5,
690
- "learning_rate": 4.083333333333334e-06,
691
- "loss": 0.0002,
692
- "step": 2775
693
- },
694
- {
695
- "epoch": 466.67,
696
- "learning_rate": 4.000000000000001e-06,
697
- "loss": 0.0002,
698
- "step": 2800
699
- },
700
- {
701
- "epoch": 470.83,
702
- "learning_rate": 3.916666666666667e-06,
703
- "loss": 0.0002,
704
- "step": 2825
705
- },
706
- {
707
- "epoch": 475.0,
708
- "learning_rate": 3.833333333333334e-06,
709
- "loss": 0.0002,
710
- "step": 2850
711
- },
712
- {
713
- "epoch": 479.17,
714
- "learning_rate": 3.7500000000000005e-06,
715
- "loss": 0.0002,
716
- "step": 2875
717
- },
718
- {
719
- "epoch": 483.33,
720
- "learning_rate": 3.6666666666666666e-06,
721
- "loss": 0.0002,
722
- "step": 2900
723
- },
724
- {
725
- "epoch": 487.5,
726
- "learning_rate": 3.5833333333333335e-06,
727
- "loss": 0.0002,
728
- "step": 2925
729
- },
730
- {
731
- "epoch": 491.67,
732
- "learning_rate": 3.5e-06,
733
- "loss": 0.0002,
734
- "step": 2950
735
- },
736
- {
737
- "epoch": 495.83,
738
- "learning_rate": 3.416666666666667e-06,
739
- "loss": 0.0002,
740
- "step": 2975
741
- },
742
- {
743
- "epoch": 500.0,
744
- "learning_rate": 3.3333333333333333e-06,
745
- "loss": 0.0002,
746
- "step": 3000
747
- },
748
- {
749
- "epoch": 500.0,
750
- "eval_normalized_wer": 0.0,
751
- "eval_runtime": 12.5834,
752
- "eval_samples_per_second": 1.987,
753
- "eval_steps_per_second": 1.987,
754
- "eval_wer": 0.0,
755
- "step": 3000
756
- },
757
- {
758
- "epoch": 504.17,
759
- "learning_rate": 3.2500000000000002e-06,
760
- "loss": 0.0002,
761
- "step": 3025
762
- },
763
- {
764
- "epoch": 508.33,
765
- "learning_rate": 3.1666666666666667e-06,
766
- "loss": 0.0002,
767
- "step": 3050
768
- },
769
- {
770
- "epoch": 512.5,
771
- "learning_rate": 3.0833333333333336e-06,
772
- "loss": 0.0002,
773
- "step": 3075
774
- },
775
- {
776
- "epoch": 516.67,
777
- "learning_rate": 3e-06,
778
- "loss": 0.0002,
779
- "step": 3100
780
- },
781
- {
782
- "epoch": 520.83,
783
- "learning_rate": 2.916666666666667e-06,
784
- "loss": 0.0002,
785
- "step": 3125
786
- },
787
- {
788
- "epoch": 525.0,
789
- "learning_rate": 2.8333333333333335e-06,
790
- "loss": 0.0002,
791
- "step": 3150
792
- },
793
- {
794
- "epoch": 529.17,
795
- "learning_rate": 2.7500000000000004e-06,
796
- "loss": 0.0002,
797
- "step": 3175
798
- },
799
- {
800
- "epoch": 533.33,
801
- "learning_rate": 2.666666666666667e-06,
802
- "loss": 0.0002,
803
- "step": 3200
804
- },
805
- {
806
- "epoch": 537.5,
807
- "learning_rate": 2.5833333333333337e-06,
808
- "loss": 0.0002,
809
- "step": 3225
810
- },
811
- {
812
- "epoch": 541.67,
813
- "learning_rate": 2.5e-06,
814
- "loss": 0.0002,
815
- "step": 3250
816
- },
817
- {
818
- "epoch": 545.83,
819
- "learning_rate": 2.4166666666666667e-06,
820
- "loss": 0.0001,
821
- "step": 3275
822
- },
823
- {
824
- "epoch": 550.0,
825
- "learning_rate": 2.3333333333333336e-06,
826
- "loss": 0.0002,
827
- "step": 3300
828
- },
829
- {
830
- "epoch": 554.17,
831
- "learning_rate": 2.25e-06,
832
- "loss": 0.0002,
833
- "step": 3325
834
- },
835
- {
836
- "epoch": 558.33,
837
- "learning_rate": 2.166666666666667e-06,
838
- "loss": 0.0002,
839
- "step": 3350
840
- },
841
- {
842
- "epoch": 562.5,
843
- "learning_rate": 2.0833333333333334e-06,
844
- "loss": 0.0002,
845
- "step": 3375
846
- },
847
- {
848
- "epoch": 566.67,
849
- "learning_rate": 2.0000000000000003e-06,
850
- "loss": 0.0002,
851
- "step": 3400
852
- },
853
- {
854
- "epoch": 570.83,
855
- "learning_rate": 1.916666666666667e-06,
856
- "loss": 0.0002,
857
- "step": 3425
858
- },
859
- {
860
- "epoch": 575.0,
861
- "learning_rate": 1.8333333333333333e-06,
862
- "loss": 0.0002,
863
- "step": 3450
864
- },
865
- {
866
- "epoch": 579.17,
867
- "learning_rate": 1.75e-06,
868
- "loss": 0.0002,
869
- "step": 3475
870
- },
871
- {
872
- "epoch": 583.33,
873
- "learning_rate": 1.6666666666666667e-06,
874
- "loss": 0.0002,
875
- "step": 3500
876
- },
877
- {
878
- "epoch": 587.5,
879
- "learning_rate": 1.5833333333333333e-06,
880
- "loss": 0.0001,
881
- "step": 3525
882
- },
883
- {
884
- "epoch": 591.67,
885
- "learning_rate": 1.5e-06,
886
- "loss": 0.0002,
887
- "step": 3550
888
- },
889
- {
890
- "epoch": 595.83,
891
- "learning_rate": 1.4166666666666667e-06,
892
- "loss": 0.0001,
893
- "step": 3575
894
- },
895
- {
896
- "epoch": 600.0,
897
- "learning_rate": 1.3333333333333334e-06,
898
- "loss": 0.0001,
899
- "step": 3600
900
- },
901
- {
902
- "epoch": 604.17,
903
- "learning_rate": 1.25e-06,
904
- "loss": 0.0002,
905
- "step": 3625
906
- },
907
- {
908
- "epoch": 608.33,
909
- "learning_rate": 1.1666666666666668e-06,
910
- "loss": 0.0002,
911
- "step": 3650
912
- },
913
- {
914
- "epoch": 612.5,
915
- "learning_rate": 1.0833333333333335e-06,
916
- "loss": 0.0002,
917
- "step": 3675
918
- },
919
- {
920
- "epoch": 616.67,
921
- "learning_rate": 1.0000000000000002e-06,
922
- "loss": 0.0002,
923
- "step": 3700
924
- },
925
- {
926
- "epoch": 620.83,
927
- "learning_rate": 9.166666666666666e-07,
928
- "loss": 0.0001,
929
- "step": 3725
930
- },
931
- {
932
- "epoch": 625.0,
933
- "learning_rate": 8.333333333333333e-07,
934
- "loss": 0.0002,
935
- "step": 3750
936
- },
937
- {
938
- "epoch": 629.17,
939
- "learning_rate": 7.5e-07,
940
- "loss": 0.0002,
941
- "step": 3775
942
- },
943
- {
944
- "epoch": 633.33,
945
- "learning_rate": 6.666666666666667e-07,
946
- "loss": 0.0002,
947
- "step": 3800
948
- },
949
- {
950
- "epoch": 637.5,
951
- "learning_rate": 5.833333333333334e-07,
952
- "loss": 0.0001,
953
- "step": 3825
954
- },
955
- {
956
- "epoch": 641.67,
957
- "learning_rate": 5.000000000000001e-07,
958
- "loss": 0.0002,
959
- "step": 3850
960
- },
961
- {
962
- "epoch": 645.83,
963
- "learning_rate": 4.1666666666666667e-07,
964
- "loss": 0.0001,
965
- "step": 3875
966
- },
967
- {
968
- "epoch": 650.0,
969
- "learning_rate": 3.3333333333333335e-07,
970
- "loss": 0.0002,
971
- "step": 3900
972
- },
973
- {
974
- "epoch": 654.17,
975
- "learning_rate": 2.5000000000000004e-07,
976
- "loss": 0.0002,
977
- "step": 3925
978
- },
979
- {
980
- "epoch": 658.33,
981
- "learning_rate": 1.6666666666666668e-07,
982
- "loss": 0.0001,
983
- "step": 3950
984
- },
985
- {
986
- "epoch": 662.5,
987
- "learning_rate": 8.333333333333334e-08,
988
- "loss": 0.0002,
989
- "step": 3975
990
- },
991
- {
992
- "epoch": 666.67,
993
- "learning_rate": 0.0,
994
- "loss": 0.0001,
995
- "step": 4000
996
- },
997
- {
998
- "epoch": 666.67,
999
- "eval_normalized_wer": 0.0,
1000
- "eval_runtime": 12.4215,
1001
- "eval_samples_per_second": 2.013,
1002
- "eval_steps_per_second": 2.013,
1003
- "eval_wer": 0.0,
1004
- "step": 4000
1005
- }
1006
- ],
1007
- "max_steps": 4000,
1008
- "num_train_epochs": 667,
1009
- "start_time": 1695876833.7164485,
1010
- "total_flos": 2.520973836288e+19,
1011
- "trial_name": null,
1012
- "trial_params": null
1013
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
whisper-tiny-hi/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:db3d9a3df34ad91106907faa15896108a794a6db264e94de4f1566378854675a
3
- size 2875
 
 
 
 
whisper-tiny-hi/vocab.json DELETED
The diff for this file is too large to render. See raw diff