mazesmazes commited on
Commit
10e8ba2
·
verified ·
1 Parent(s): 71301b8

Training in progress, step 1000

Browse files
config.json CHANGED
@@ -160,14 +160,14 @@
160
  "label_smoothing": 0.0,
161
  "length_penalty": 1.0,
162
  "llm_dim": 2048,
163
- "lora_alpha": 32,
164
- "lora_dropout": 0.0,
165
- "lora_r": 64,
166
- "lora_target_modules": [
167
- "v_proj",
168
- "q_proj"
169
- ],
170
  "max_new_tokens": 256,
 
171
  "model_dtype": "bfloat16",
172
  "model_type": "asr_model",
173
  "no_repeat_ngram_size": 0,
@@ -178,7 +178,6 @@
178
  "projector_dropout": 0.0,
179
  "projector_hidden_dim": null,
180
  "projector_init_std": 0.02,
181
- "projector_input_noise": 0.0,
182
  "projector_num_layers": 2,
183
  "projector_pool_stride": 4,
184
  "projector_type": "mosa",
@@ -255,7 +254,7 @@
255
  "text_model_id": "Qwen/Qwen3-1.7B",
256
  "transformers_version": "5.0.0.dev0",
257
  "use_cache": false,
258
- "use_lora": true,
259
  "use_specaugment": true,
260
  "user_prompt": "Please transcribe this English audio into text: <audio>",
261
  "vocab_size": 151670
 
160
  "label_smoothing": 0.0,
161
  "length_penalty": 1.0,
162
  "llm_dim": 2048,
163
+ "mask_feature_length": 10,
164
+ "mask_feature_min_masks": 0,
165
+ "mask_feature_prob": 0.0,
166
+ "mask_time_length": 10,
167
+ "mask_time_min_masks": 2,
168
+ "mask_time_prob": 0.05,
 
169
  "max_new_tokens": 256,
170
+ "min_new_tokens": 1,
171
  "model_dtype": "bfloat16",
172
  "model_type": "asr_model",
173
  "no_repeat_ngram_size": 0,
 
178
  "projector_dropout": 0.0,
179
  "projector_hidden_dim": null,
180
  "projector_init_std": 0.02,
 
181
  "projector_num_layers": 2,
182
  "projector_pool_stride": 4,
183
  "projector_type": "mosa",
 
254
  "text_model_id": "Qwen/Qwen3-1.7B",
255
  "transformers_version": "5.0.0.dev0",
256
  "use_cache": false,
257
+ "use_lora": false,
258
  "use_specaugment": true,
259
  "user_prompt": "Please transcribe this English audio into text: <audio>",
260
  "vocab_size": 151670
generation_config.json CHANGED
@@ -6,6 +6,7 @@
6
  ],
7
  "length_penalty": 1.0,
8
  "max_new_tokens": 256,
 
9
  "no_repeat_ngram_size": 0,
10
  "num_beams": 1,
11
  "pad_token_id": 151643,
 
6
  ],
7
  "length_penalty": 1.0,
8
  "max_new_tokens": 256,
9
+ "min_new_tokens": 1,
10
  "no_repeat_ngram_size": 0,
11
  "num_beams": 1,
12
  "pad_token_id": 151643,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ca080ca5928f40d46600cd7a02ca2ee693894c41bd91ad8b99d852ae923f118
3
  size 236248080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c2a286f15a803944ee979473fd7aeca3cf83242d2541114b5b425cb43158a52
3
  size 236248080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ecb28a40b3e5b48c6e2923003e47d334c96cb66c84c3f24cf7192e4e93c361c
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33e9003ce34c401f33cf18981a0f24a8877c856874f33f16e65e105a4f86f557
3
  size 5201