irudachirath commited on
Commit
b0bfb02
·
verified ·
1 Parent(s): 89da506

Training in progress, step 250

Browse files
adapter_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "WhisperForConditionalGeneration",
7
+ "parent_library": "transformers.models.whisper.modeling_whisper"
8
+ },
9
+ "base_model_name_or_path": "openai/whisper-small",
10
+ "bias": "none",
11
+ "corda_config": null,
12
+ "ensure_weight_tying": false,
13
+ "eva_config": null,
14
+ "exclude_modules": null,
15
+ "fan_in_fan_out": false,
16
+ "inference_mode": true,
17
+ "init_lora_weights": true,
18
+ "layer_replication": null,
19
+ "layers_pattern": null,
20
+ "layers_to_transform": null,
21
+ "loftq_config": {},
22
+ "lora_alpha": 32,
23
+ "lora_bias": false,
24
+ "lora_dropout": 0.05,
25
+ "megatron_config": null,
26
+ "megatron_core": "megatron.core",
27
+ "modules_to_save": null,
28
+ "peft_type": "LORA",
29
+ "peft_version": "0.18.1",
30
+ "qalora_group_size": 16,
31
+ "r": 16,
32
+ "rank_pattern": {},
33
+ "revision": null,
34
+ "target_modules": [
35
+ "v_proj",
36
+ "q_proj"
37
+ ],
38
+ "target_parameters": null,
39
+ "task_type": null,
40
+ "trainable_token_indices": null,
41
+ "use_dora": false,
42
+ "use_qalora": false,
43
+ "use_rslora": false
44
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95d6f2e557d73cc5145af88db302e9a0fdb12ca734ec6f2ca6700377f2ddcb39
3
+ size 7098064
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": "<|endoftext|>",
5
+ "clean_up_tokenization_spaces": true,
6
+ "eos_token": "<|endoftext|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|endoftext|>",
10
+ "<|startoftranscript|>",
11
+ "<|en|>",
12
+ "<|zh|>",
13
+ "<|de|>",
14
+ "<|es|>",
15
+ "<|ru|>",
16
+ "<|ko|>",
17
+ "<|fr|>",
18
+ "<|ja|>",
19
+ "<|pt|>",
20
+ "<|tr|>",
21
+ "<|pl|>",
22
+ "<|ca|>",
23
+ "<|nl|>",
24
+ "<|ar|>",
25
+ "<|sv|>",
26
+ "<|it|>",
27
+ "<|id|>",
28
+ "<|hi|>",
29
+ "<|fi|>",
30
+ "<|vi|>",
31
+ "<|he|>",
32
+ "<|uk|>",
33
+ "<|el|>",
34
+ "<|ms|>",
35
+ "<|cs|>",
36
+ "<|ro|>",
37
+ "<|da|>",
38
+ "<|hu|>",
39
+ "<|ta|>",
40
+ "<|no|>",
41
+ "<|th|>",
42
+ "<|ur|>",
43
+ "<|hr|>",
44
+ "<|bg|>",
45
+ "<|lt|>",
46
+ "<|la|>",
47
+ "<|mi|>",
48
+ "<|ml|>",
49
+ "<|cy|>",
50
+ "<|sk|>",
51
+ "<|te|>",
52
+ "<|fa|>",
53
+ "<|lv|>",
54
+ "<|bn|>",
55
+ "<|sr|>",
56
+ "<|az|>",
57
+ "<|sl|>",
58
+ "<|kn|>",
59
+ "<|et|>",
60
+ "<|mk|>",
61
+ "<|br|>",
62
+ "<|eu|>",
63
+ "<|is|>",
64
+ "<|hy|>",
65
+ "<|ne|>",
66
+ "<|mn|>",
67
+ "<|bs|>",
68
+ "<|kk|>",
69
+ "<|sq|>",
70
+ "<|sw|>",
71
+ "<|gl|>",
72
+ "<|mr|>",
73
+ "<|pa|>",
74
+ "<|si|>",
75
+ "<|km|>",
76
+ "<|sn|>",
77
+ "<|yo|>",
78
+ "<|so|>",
79
+ "<|af|>",
80
+ "<|oc|>",
81
+ "<|ka|>",
82
+ "<|be|>",
83
+ "<|tg|>",
84
+ "<|sd|>",
85
+ "<|gu|>",
86
+ "<|am|>",
87
+ "<|yi|>",
88
+ "<|lo|>",
89
+ "<|uz|>",
90
+ "<|fo|>",
91
+ "<|ht|>",
92
+ "<|ps|>",
93
+ "<|tk|>",
94
+ "<|nn|>",
95
+ "<|mt|>",
96
+ "<|sa|>",
97
+ "<|lb|>",
98
+ "<|my|>",
99
+ "<|bo|>",
100
+ "<|tl|>",
101
+ "<|mg|>",
102
+ "<|as|>",
103
+ "<|tt|>",
104
+ "<|haw|>",
105
+ "<|ln|>",
106
+ "<|ha|>",
107
+ "<|ba|>",
108
+ "<|jw|>",
109
+ "<|su|>",
110
+ "<|translate|>",
111
+ "<|transcribe|>",
112
+ "<|startoflm|>",
113
+ "<|startofprev|>",
114
+ "<|nocaptions|>",
115
+ "<|notimestamps|>"
116
+ ],
117
+ "is_local": false,
118
+ "language": "sinhala",
119
+ "model_max_length": 1024,
120
+ "pad_token": "<|endoftext|>",
121
+ "predict_timestamps": false,
122
+ "processor_class": "WhisperProcessor",
123
+ "return_attention_mask": false,
124
+ "task": "transcribe",
125
+ "tokenizer_class": "WhisperTokenizer",
126
+ "unk_token": "<|endoftext|>"
127
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa88880e983eb5f4fe02bc6a0f7aff8fd91e1e1b17ef68edf092908be746255
3
+ size 5329