elmadany commited on
Commit
2fc1647
·
verified ·
1 Parent(s): 44a6c72

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1 +1,102 @@
1
- # Simba-M
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model: facebook/mms-1b-all
4
+ tags:
5
+ - automatic-speech-recognition
6
+ - /mnt/home/elmadany/elmadany_workspace/African_speechT5/jasmine-raid/elmadany_work/HF_format/SimbaBench_tasks_ft
7
+ - mms
8
+ - generated_from_trainer
9
+ metrics:
10
+ - wer
11
+ model-index:
12
+ - name: mms-1b-all
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # mms-1b-all
20
+
21
+ This model is a fine-tuned version of [facebook/mms-1b-all](https://huggingface.co/facebook/mms-1b-all) on the /MNT/HOME/ELMADANY/ELMADANY_WORKSPACE/AFRICAN_SPEECHT5/JASMINE-RAID/ELMADANY_WORK/HF_FORMAT/SIMBABENCH_TASKS_FT - ASR_FT_DATA_BATCH_2 dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.2214
24
+ - Wer: 0.2988
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 0.001
44
+ - train_batch_size: 4
45
+ - eval_batch_size: 2
46
+ - seed: 42
47
+ - distributed_type: multi-GPU
48
+ - num_devices: 4
49
+ - gradient_accumulation_steps: 4
50
+ - total_train_batch_size: 64
51
+ - total_eval_batch_size: 8
52
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
+ - lr_scheduler_type: linear
54
+ - lr_scheduler_warmup_steps: 1000
55
+ - num_epochs: 30.0
56
+
57
+ ### Training results
58
+
59
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
60
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|
61
+ | 6.7354 | 1.0 | 610 | 0.5166 | 0.5255 |
62
+ | 0.7031 | 2.0 | 1221 | 0.4340 | 0.4949 |
63
+ | 0.6258 | 3.0 | 1831 | 0.3713 | 0.4372 |
64
+ | 0.5844 | 4.0 | 2442 | 0.3527 | 0.4214 |
65
+ | 0.5615 | 5.0 | 3053 | 0.3429 | 0.4238 |
66
+ | 0.5829 | 2.0 | 4884 | 0.3453 | 0.4082 |
67
+ | 0.5516 | 3.0 | 7327 | 0.3249 | 0.3986 |
68
+ | 0.5313 | 4.0 | 9769 | 0.3149 | 0.3898 |
69
+ | 0.5141 | 5.0 | 12212 | 0.3006 | 0.3754 |
70
+ | 0.4991 | 6.0 | 14654 | 0.2959 | 0.3730 |
71
+ | 0.4916 | 7.0 | 17097 | 0.2914 | 0.3667 |
72
+ | 0.48 | 8.0 | 19539 | 0.2791 | 0.3528 |
73
+ | 0.4708 | 9.0 | 21982 | 0.2808 | 0.3624 |
74
+ | 0.4659 | 10.0 | 24424 | 0.2730 | 0.3518 |
75
+ | 0.4587 | 11.0 | 26867 | 0.2708 | 0.3511 |
76
+ | 0.4541 | 12.0 | 29309 | 0.2630 | 0.3371 |
77
+ | 0.4442 | 13.0 | 31752 | 0.2606 | 0.3394 |
78
+ | 0.439 | 14.0 | 34194 | 0.2592 | 0.3384 |
79
+ | 0.4334 | 15.0 | 36637 | 0.2499 | 0.3300 |
80
+ | 0.4304 | 16.0 | 39079 | 0.2492 | 0.3279 |
81
+ | 0.4247 | 17.0 | 41522 | 0.2433 | 0.3255 |
82
+ | 0.4197 | 18.0 | 43964 | 0.2407 | 0.3214 |
83
+ | 0.4139 | 19.0 | 46407 | 0.2419 | 0.3172 |
84
+ | 0.4091 | 20.0 | 48849 | 0.2378 | 0.3190 |
85
+ | 0.4064 | 21.0 | 51292 | 0.2345 | 0.3150 |
86
+ | 0.4026 | 22.0 | 53734 | 0.2319 | 0.3141 |
87
+ | 0.4015 | 23.0 | 56177 | 0.2299 | 0.3070 |
88
+ | 0.3941 | 24.0 | 58619 | 0.2278 | 0.3064 |
89
+ | 0.3913 | 25.0 | 61062 | 0.2250 | 0.3044 |
90
+ | 0.3902 | 26.0 | 63504 | 0.2270 | 0.3061 |
91
+ | 0.3847 | 27.0 | 65947 | 0.2237 | 0.3025 |
92
+ | 0.3826 | 28.0 | 68389 | 0.2214 | 0.2988 |
93
+ | 0.3797 | 29.0 | 70832 | 0.2213 | 0.3001 |
94
+ | 0.3786 | 29.99 | 73260 | 0.2208 | 0.2989 |
95
+
96
+
97
+ ### Framework versions
98
+
99
+ - Transformers 4.33.2
100
+ - Pytorch 2.0.1+cu117
101
+ - Datasets 3.5.1
102
+ - Tokenizers 0.13.3
adapter.multilingual_african.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b31b181215ee1591c2395ab7c3bc43537e515b9a864c4d7a7e11cdf820bd53f
3
+ size 10832872
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 427,
3
+ "<s>": 426
4
+ }
config.json ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/mms-1b-all",
3
+ "activation_dropout": 0.05,
4
+ "adapter_attn_dim": 16,
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": false,
8
+ "apply_spec_augment": true,
9
+ "architectures": [
10
+ "Wav2Vec2ForCTC"
11
+ ],
12
+ "attention_dropout": 0.05,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 256,
15
+ "codevector_dim": 1024,
16
+ "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": true,
18
+ "conv_dim": [
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512,
25
+ 512
26
+ ],
27
+ "conv_kernel": [
28
+ 10,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 3,
33
+ 2,
34
+ 2
35
+ ],
36
+ "conv_stride": [
37
+ 5,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2,
43
+ 2
44
+ ],
45
+ "ctc_loss_reduction": "mean",
46
+ "ctc_zero_infinity": false,
47
+ "diversity_loss_weight": 0.1,
48
+ "do_stable_layer_norm": true,
49
+ "eos_token_id": 2,
50
+ "feat_extract_activation": "gelu",
51
+ "feat_extract_dropout": 0.0,
52
+ "feat_extract_norm": "layer",
53
+ "feat_proj_dropout": 0.05,
54
+ "feat_quantizer_dropout": 0.0,
55
+ "final_dropout": 0.0,
56
+ "hidden_act": "gelu",
57
+ "hidden_dropout": 0.05,
58
+ "hidden_size": 1280,
59
+ "initializer_range": 0.02,
60
+ "intermediate_size": 5120,
61
+ "layer_norm_eps": 1e-05,
62
+ "layerdrop": 0.0,
63
+ "mask_feature_length": 10,
64
+ "mask_feature_min_masks": 0,
65
+ "mask_feature_prob": 0.0,
66
+ "mask_time_length": 10,
67
+ "mask_time_min_masks": 2,
68
+ "mask_time_prob": 0.05,
69
+ "model_type": "wav2vec2",
70
+ "num_adapter_layers": 3,
71
+ "num_attention_heads": 16,
72
+ "num_codevector_groups": 2,
73
+ "num_codevectors_per_group": 320,
74
+ "num_conv_pos_embedding_groups": 16,
75
+ "num_conv_pos_embeddings": 128,
76
+ "num_feat_extract_layers": 7,
77
+ "num_hidden_layers": 48,
78
+ "num_negatives": 100,
79
+ "output_hidden_size": 1280,
80
+ "pad_token_id": 425,
81
+ "proj_codevector_dim": 1024,
82
+ "tdnn_dilation": [
83
+ 1,
84
+ 2,
85
+ 3,
86
+ 1,
87
+ 1
88
+ ],
89
+ "tdnn_dim": [
90
+ 512,
91
+ 512,
92
+ 512,
93
+ 512,
94
+ 1500
95
+ ],
96
+ "tdnn_kernel": [
97
+ 5,
98
+ 3,
99
+ 3,
100
+ 1,
101
+ 1
102
+ ],
103
+ "torch_dtype": "float32",
104
+ "transformers_version": "4.33.2",
105
+ "use_weighted_layer_sum": false,
106
+ "vocab_size": 428,
107
+ "xvector_output_dim": 512
108
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0,
7
+ "processor_class": "Wav2Vec2Processor",
8
+ "return_attention_mask": true,
9
+ "sampling_rate": 16000
10
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5670e00ab5a3e126af3608015a7b0b18ffe537e75d0f58d6453cf62488824c7
3
+ size 3861168781
special_tokens_map.json ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "</s>",
12
+ "lstrip": false,
13
+ "normalized": true,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<s>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "</s>",
26
+ "lstrip": false,
27
+ "normalized": true,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<s>",
33
+ "lstrip": false,
34
+ "normalized": true,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "</s>",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ {
46
+ "content": "<s>",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ {
53
+ "content": "</s>",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false
58
+ },
59
+ {
60
+ "content": "<s>",
61
+ "lstrip": false,
62
+ "normalized": true,
63
+ "rstrip": false,
64
+ "single_word": false
65
+ },
66
+ {
67
+ "content": "</s>",
68
+ "lstrip": false,
69
+ "normalized": true,
70
+ "rstrip": false,
71
+ "single_word": false
72
+ },
73
+ {
74
+ "content": "<s>",
75
+ "lstrip": false,
76
+ "normalized": true,
77
+ "rstrip": false,
78
+ "single_word": false
79
+ },
80
+ {
81
+ "content": "</s>",
82
+ "lstrip": false,
83
+ "normalized": true,
84
+ "rstrip": false,
85
+ "single_word": false
86
+ },
87
+ {
88
+ "content": "<s>",
89
+ "lstrip": false,
90
+ "normalized": true,
91
+ "rstrip": false,
92
+ "single_word": false
93
+ },
94
+ {
95
+ "content": "</s>",
96
+ "lstrip": false,
97
+ "normalized": true,
98
+ "rstrip": false,
99
+ "single_word": false
100
+ },
101
+ {
102
+ "content": "<s>",
103
+ "lstrip": false,
104
+ "normalized": true,
105
+ "rstrip": false,
106
+ "single_word": false
107
+ },
108
+ {
109
+ "content": "</s>",
110
+ "lstrip": false,
111
+ "normalized": true,
112
+ "rstrip": false,
113
+ "single_word": false
114
+ }
115
+ ],
116
+ "bos_token": "<s>",
117
+ "eos_token": "</s>",
118
+ "pad_token": "[PAD]",
119
+ "unk_token": "[UNK]"
120
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "clean_up_tokenization_spaces": true,
4
+ "do_lower_case": false,
5
+ "eos_token": "</s>",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "[PAD]",
8
+ "replace_word_delimiter_char": " ",
9
+ "target_lang": "multilingual_african",
10
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
11
+ "unk_token": "[UNK]",
12
+ "word_delimiter_token": "|"
13
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 29.99,
3
+ "train_loss": 0.4226855478747569,
4
+ "train_runtime": 82629.0475,
5
+ "train_samples": 156308,
6
+ "train_samples_per_second": 56.751,
7
+ "train_steps_per_second": 0.887
8
+ }
vocab.json ADDED
@@ -0,0 +1,430 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "multilingual_african": {
3
+ "<": 1,
4
+ "=": 2,
5
+ ">": 3,
6
+ "[": 4,
7
+ "[PAD]": 425,
8
+ "[UNK]": 424,
9
+ "\\": 5,
10
+ "]": 6,
11
+ "_": 7,
12
+ "`": 8,
13
+ "a": 9,
14
+ "b": 10,
15
+ "c": 11,
16
+ "d": 12,
17
+ "e": 13,
18
+ "f": 14,
19
+ "g": 15,
20
+ "h": 16,
21
+ "i": 17,
22
+ "j": 18,
23
+ "k": 19,
24
+ "l": 20,
25
+ "m": 21,
26
+ "n": 22,
27
+ "o": 23,
28
+ "p": 24,
29
+ "q": 25,
30
+ "r": 26,
31
+ "s": 27,
32
+ "t": 28,
33
+ "u": 29,
34
+ "v": 30,
35
+ "w": 31,
36
+ "x": 32,
37
+ "y": 33,
38
+ "z": 34,
39
+ "|": 0,
40
+ "~": 35,
41
+ "«": 36,
42
+ "²": 37,
43
+ "µ": 38,
44
+ "·": 39,
45
+ "»": 40,
46
+ "à": 41,
47
+ "á": 42,
48
+ "â": 43,
49
+ "ã": 44,
50
+ "ç": 45,
51
+ "è": 46,
52
+ "é": 47,
53
+ "ê": 48,
54
+ "ë": 49,
55
+ "ì": 50,
56
+ "í": 51,
57
+ "ï": 52,
58
+ "ñ": 53,
59
+ "ò": 54,
60
+ "ó": 55,
61
+ "ô": 56,
62
+ "õ": 57,
63
+ "ö": 58,
64
+ "ù": 59,
65
+ "ú": 60,
66
+ "û": 61,
67
+ "ć": 62,
68
+ "č": 63,
69
+ "ĕ": 64,
70
+ "ğ": 65,
71
+ "ĩ": 66,
72
+ "ī": 67,
73
+ "ĭ": 68,
74
+ "ĺ": 69,
75
+ "ń": 70,
76
+ "ŋ": 71,
77
+ "ŏ": 72,
78
+ "ś": 73,
79
+ "š": 74,
80
+ "ŧ": 75,
81
+ "ũ": 76,
82
+ "ŭ": 77,
83
+ "ŵ": 78,
84
+ "ƙ": 79,
85
+ "ƥ": 80,
86
+ "ƭ": 81,
87
+ "ƴ": 82,
88
+ "ǧ": 83,
89
+ "ǹ": 84,
90
+ "ɓ": 85,
91
+ "ɔ": 86,
92
+ "ɖ": 87,
93
+ "ɗ": 88,
94
+ "ɛ": 89,
95
+ "ɣ": 90,
96
+ "ɲ": 91,
97
+ "ʹ": 92,
98
+ "ʻ": 93,
99
+ "̀": 94,
100
+ "́": 95,
101
+ "̆": 96,
102
+ "̈": 97,
103
+ "̣": 98,
104
+ "έ": 99,
105
+ "γ": 100,
106
+ "ε": 101,
107
+ "ԑ": 102,
108
+ "ሀ": 103,
109
+ "ሁ": 104,
110
+ "ሂ": 105,
111
+ "ሃ": 106,
112
+ "ሄ": 107,
113
+ "ህ": 108,
114
+ "ሆ": 109,
115
+ "ለ": 110,
116
+ "ሉ": 111,
117
+ "ሊ": 112,
118
+ "ላ": 113,
119
+ "ሌ": 114,
120
+ "ል": 115,
121
+ "ሎ": 116,
122
+ "ሏ": 117,
123
+ "ሐ": 118,
124
+ "ሑ": 119,
125
+ "ሒ": 120,
126
+ "ሓ": 121,
127
+ "ሔ": 122,
128
+ "ሕ": 123,
129
+ "ሖ": 124,
130
+ "መ": 125,
131
+ "ሙ": 126,
132
+ "ሚ": 127,
133
+ "ማ": 128,
134
+ "ሜ": 129,
135
+ "ም": 130,
136
+ "ሞ": 131,
137
+ "ሟ": 132,
138
+ "ሠ": 133,
139
+ "ሡ": 134,
140
+ "ሣ": 135,
141
+ "ሥ": 136,
142
+ "ሦ": 137,
143
+ "ረ": 138,
144
+ "ሩ": 139,
145
+ "ሪ": 140,
146
+ "ራ": 141,
147
+ "ሬ": 142,
148
+ "ር": 143,
149
+ "ሮ": 144,
150
+ "ሯ": 145,
151
+ "ሰ": 146,
152
+ "ሱ": 147,
153
+ "ሲ": 148,
154
+ "ሳ": 149,
155
+ "ሴ": 150,
156
+ "ስ": 151,
157
+ "ሶ": 152,
158
+ "ሷ": 153,
159
+ "ሸ": 154,
160
+ "ሹ": 155,
161
+ "ሺ": 156,
162
+ "ሻ": 157,
163
+ "ሼ": 158,
164
+ "ሽ": 159,
165
+ "ሾ": 160,
166
+ "ሿ": 161,
167
+ "ቀ": 162,
168
+ "ቁ": 163,
169
+ "ቂ": 164,
170
+ "ቃ": 165,
171
+ "ቄ": 166,
172
+ "ቅ": 167,
173
+ "ቆ": 168,
174
+ "ቋ": 169,
175
+ "ቐ": 170,
176
+ "ቒ": 171,
177
+ "ቓ": 172,
178
+ "ቕ": 173,
179
+ "ቚ": 174,
180
+ "በ": 175,
181
+ "ቡ": 176,
182
+ "ቢ": 177,
183
+ "ባ": 178,
184
+ "ቤ": 179,
185
+ "ብ": 180,
186
+ "ቦ": 181,
187
+ "ቧ": 182,
188
+ "ቨ": 183,
189
+ "ቩ": 184,
190
+ "ቪ": 185,
191
+ "ቫ": 186,
192
+ "ቬ": 187,
193
+ "ቭ": 188,
194
+ "ቮ": 189,
195
+ "ቯ": 190,
196
+ "ተ": 191,
197
+ "ቱ": 192,
198
+ "ቲ": 193,
199
+ "ታ": 194,
200
+ "ቴ": 195,
201
+ "ት": 196,
202
+ "ቶ": 197,
203
+ "ቷ": 198,
204
+ "ቸ": 199,
205
+ "ቹ": 200,
206
+ "ቺ": 201,
207
+ "ቻ": 202,
208
+ "ቼ": 203,
209
+ "ች": 204,
210
+ "ቾ": 205,
211
+ "ቿ": 206,
212
+ "ኃ": 207,
213
+ "ኅ": 208,
214
+ "ኋ": 209,
215
+ "ነ": 210,
216
+ "ኑ": 211,
217
+ "ኒ": 212,
218
+ "ና": 213,
219
+ "ኔ": 214,
220
+ "ን": 215,
221
+ "ኖ": 216,
222
+ "ኗ": 217,
223
+ "ኘ": 218,
224
+ "ኙ": 219,
225
+ "ኚ": 220,
226
+ "ኛ": 221,
227
+ "ኜ": 222,
228
+ "ኝ": 223,
229
+ "ኞ": 224,
230
+ "ኟ": 225,
231
+ "አ": 226,
232
+ "ኡ": 227,
233
+ "ኢ": 228,
234
+ "ኣ": 229,
235
+ "ኤ": 230,
236
+ "እ": 231,
237
+ "ኦ": 232,
238
+ "ከ": 233,
239
+ "ኩ": 234,
240
+ "ኪ": 235,
241
+ "ካ": 236,
242
+ "ኬ": 237,
243
+ "ክ": 238,
244
+ "ኮ": 239,
245
+ "ኰ": 240,
246
+ "ኲ": 241,
247
+ "ኳ": 242,
248
+ "ኸ": 243,
249
+ "ኻ": 244,
250
+ "ኽ": 245,
251
+ "ኾ": 246,
252
+ "ወ": 247,
253
+ "ዉ": 248,
254
+ "ዊ": 249,
255
+ "ዋ": 250,
256
+ "ዌ": 251,
257
+ "ው": 252,
258
+ "ዎ": 253,
259
+ "ዐ": 254,
260
+ "ዑ": 255,
261
+ "ዒ": 256,
262
+ "ዓ": 257,
263
+ "ዔ": 258,
264
+ "ዕ": 259,
265
+ "ዖ": 260,
266
+ "ዘ": 261,
267
+ "ዙ": 262,
268
+ "ዚ": 263,
269
+ "ዛ": 264,
270
+ "ዜ": 265,
271
+ "ዝ": 266,
272
+ "ዞ": 267,
273
+ "ዟ": 268,
274
+ "ዠ": 269,
275
+ "ዡ": 270,
276
+ "ዢ": 271,
277
+ "ዣ": 272,
278
+ "ዤ": 273,
279
+ "ዥ": 274,
280
+ "ዦ": 275,
281
+ "ዧ": 276,
282
+ "የ": 277,
283
+ "ዩ": 278,
284
+ "ዪ": 279,
285
+ "ያ": 280,
286
+ "ዬ": 281,
287
+ "ይ": 282,
288
+ "ዮ": 283,
289
+ "ደ": 284,
290
+ "ዱ": 285,
291
+ "ዲ": 286,
292
+ "ዳ": 287,
293
+ "ዴ": 288,
294
+ "ድ": 289,
295
+ "ዶ": 290,
296
+ "ዷ": 291,
297
+ "ጀ": 292,
298
+ "ጁ": 293,
299
+ "ጂ": 294,
300
+ "ጃ": 295,
301
+ "ጄ": 296,
302
+ "ጅ": 297,
303
+ "ጆ": 298,
304
+ "ጇ": 299,
305
+ "ገ": 300,
306
+ "ጉ": 301,
307
+ "ጊ": 302,
308
+ "ጋ": 303,
309
+ "ጌ": 304,
310
+ "ግ": 305,
311
+ "ጎ": 306,
312
+ "ጐ": 307,
313
+ "ጓ": 308,
314
+ "ጔ": 309,
315
+ "ጠ": 310,
316
+ "ጡ": 311,
317
+ "ጢ": 312,
318
+ "ጣ": 313,
319
+ "ጤ": 314,
320
+ "ጥ": 315,
321
+ "ጦ": 316,
322
+ "ጧ": 317,
323
+ "ጨ": 318,
324
+ "ጩ": 319,
325
+ "ጪ": 320,
326
+ "ጫ": 321,
327
+ "ጬ": 322,
328
+ "ጭ": 323,
329
+ "ጮ": 324,
330
+ "ጯ": 325,
331
+ "ጰ": 326,
332
+ "ጱ": 327,
333
+ "ጲ": 328,
334
+ "ጳ": 329,
335
+ "ጴ": 330,
336
+ "ጵ": 331,
337
+ "ጶ": 332,
338
+ "ጸ": 333,
339
+ "ጹ": 334,
340
+ "ጺ": 335,
341
+ "ጻ": 336,
342
+ "ጼ": 337,
343
+ "ጽ": 338,
344
+ "ጾ": 339,
345
+ "ጿ": 340,
346
+ "ፀ": 341,
347
+ "ፁ": 342,
348
+ "ፃ": 343,
349
+ "ፅ": 344,
350
+ "ፈ": 345,
351
+ "ፉ": 346,
352
+ "ፊ": 347,
353
+ "ፋ": 348,
354
+ "ፌ": 349,
355
+ "ፍ": 350,
356
+ "ፎ": 351,
357
+ "ፏ": 352,
358
+ "ፐ": 353,
359
+ "ፑ": 354,
360
+ "ፒ": 355,
361
+ "ፓ": 356,
362
+ "ፔ": 357,
363
+ "ፕ": 358,
364
+ "ፖ": 359,
365
+ "፡": 360,
366
+ "።": 361,
367
+ "፣": 362,
368
+ "፤": 363,
369
+ "ḅ": 364,
370
+ "ḍ": 365,
371
+ "ḓ": 366,
372
+ "ḥ": 367,
373
+ "ḷ": 368,
374
+ "ḽ": 369,
375
+ "ṅ": 370,
376
+ "ṋ": 371,
377
+ "ṕ": 372,
378
+ "ṛ": 373,
379
+ "ṣ": 374,
380
+ "ṭ": 375,
381
+ "ṱ": 376,
382
+ "ẃ": 377,
383
+ "ẓ": 378,
384
+ "ạ": 379,
385
+ "ẹ": 380,
386
+ "ị": 381,
387
+ "ọ": 382,
388
+ "ụ": 383,
389
+ "ὲ": 384,
390
+ "–": 385,
391
+ "—": 386,
392
+ "’": 387,
393
+ "‟": 388,
394
+ "•": 389,
395
+ "…": 390,
396
+ "‽": 391,
397
+ "ⴰ": 392,
398
+ "ⴱ": 393,
399
+ "ⴳ": 394,
400
+ "ⴷ": 395,
401
+ "ⴹ": 396,
402
+ "ⴻ": 397,
403
+ "ⴼ": 398,
404
+ "ⴽ": 399,
405
+ "ⵀ": 400,
406
+ "ⵃ": 401,
407
+ "ⵄ": 402,
408
+ "ⵅ": 403,
409
+ "ⵇ": 404,
410
+ "ⵉ": 405,
411
+ "ⵊ": 406,
412
+ "ⵍ": 407,
413
+ "ⵎ": 408,
414
+ "ⵏ": 409,
415
+ "ⵓ": 410,
416
+ "ⵔ": 411,
417
+ "ⵕ": 412,
418
+ "ⵖ": 413,
419
+ "ⵙ": 414,
420
+ "ⵚ": 415,
421
+ "ⵛ": 416,
422
+ "ⵜ": 417,
423
+ "ⵟ": 418,
424
+ "ⵡ": 419,
425
+ "ⵢ": 420,
426
+ "ⵣ": 421,
427
+ "ⵥ": 422,
428
+ "ⵯ": 423
429
+ }
430
+ }