mazesmazes commited on
Commit
bc0b493
·
verified ·
1 Parent(s): 73cda73

Training in progress, step 2000

Browse files
adapter_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "alpha_pattern": {},
4
  "arrow_config": null,
5
  "auto_mapping": null,
6
- "base_model_name_or_path": "mazesmazes/tiny-audio",
7
  "bias": "none",
8
  "corda_config": null,
9
  "ensure_weight_tying": false,
@@ -16,9 +16,9 @@
16
  "layers_pattern": null,
17
  "layers_to_transform": null,
18
  "loftq_config": {},
19
- "lora_alpha": 256,
20
  "lora_bias": false,
21
- "lora_dropout": 0.05,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
@@ -29,8 +29,8 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
- "q_proj",
33
- "v_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
 
3
  "alpha_pattern": {},
4
  "arrow_config": null,
5
  "auto_mapping": null,
6
+ "base_model_name_or_path": "Qwen/Qwen3-0.6B",
7
  "bias": "none",
8
  "corda_config": null,
9
  "ensure_weight_tying": false,
 
16
  "layers_pattern": null,
17
  "layers_to_transform": null,
18
  "loftq_config": {},
19
+ "lora_alpha": 32,
20
  "lora_bias": false,
21
+ "lora_dropout": 0.0,
22
  "megatron_config": null,
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "v_proj",
33
+ "q_proj"
34
  ],
35
  "target_parameters": null,
36
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b42236aaf96efc41e22f9dfed16985e994aebfe5dbd7d005da5c25598ac7a4cf
3
  size 36715216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8492345b1762cee01629ee589f91551b15ed7050fb203c8f919fc89131a4d75
3
  size 36715216
config.json CHANGED
@@ -277,13 +277,13 @@
277
  ],
278
  "encoder_dim": 1280,
279
  "freeze_projector": false,
280
- "freq_mask_length": 10,
281
  "inference_warmup_tokens": 10,
282
  "label_smoothing": 0.0,
283
  "length_penalty": 1.0,
284
  "llm_dim": 1024,
285
- "lora_alpha": 256,
286
- "lora_dropout": 0.05,
287
  "lora_rank": 64,
288
  "lora_target_modules": [
289
  "q_proj",
@@ -300,7 +300,6 @@
300
  "num_freq_masks": 2,
301
  "num_time_masks": 2,
302
  "pipeline_tag": "automatic-speech-recognition",
303
- "pretrained_model_path": "mazesmazes/tiny-audio",
304
  "projector_dropout": 0.0,
305
  "projector_hidden_dim": null,
306
  "projector_init_std": 0.02,
@@ -378,7 +377,7 @@
378
  "vocab_size": 151670
379
  },
380
  "text_model_id": "Qwen/Qwen3-0.6B",
381
- "time_mask_length": 15,
382
  "transformers_version": "5.0.0.dev0",
383
  "use_cache": false,
384
  "use_lora": true,
 
277
  ],
278
  "encoder_dim": 1280,
279
  "freeze_projector": false,
280
+ "freq_mask_length": 43,
281
  "inference_warmup_tokens": 10,
282
  "label_smoothing": 0.0,
283
  "length_penalty": 1.0,
284
  "llm_dim": 1024,
285
+ "lora_alpha": 32,
286
+ "lora_dropout": 0.0,
287
  "lora_rank": 64,
288
  "lora_target_modules": [
289
  "q_proj",
 
300
  "num_freq_masks": 2,
301
  "num_time_masks": 2,
302
  "pipeline_tag": "automatic-speech-recognition",
 
303
  "projector_dropout": 0.0,
304
  "projector_hidden_dim": null,
305
  "projector_init_std": 0.02,
 
377
  "vocab_size": 151670
378
  },
379
  "text_model_id": "Qwen/Qwen3-0.6B",
380
+ "time_mask_length": 100,
381
  "transformers_version": "5.0.0.dev0",
382
  "use_cache": false,
383
  "use_lora": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:451ac117081e6354a030959d0f922af2259e18bd87c6eba299188fd4bda93752
3
  size 25172384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0347348d2f9cd64839ab1875ed302a1e1c18f65a070a3643f044869e50d2dce4
3
  size 25172384
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a42bd1aaa0bffdaabd32fd8a4af9544e0ed2608a2116ad462a6945ee4b9eefff
3
  size 5201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1540bbe90c48101bd1f6fb63273e8b75f53d50d743e6bdb8f473bd0c470b4aba
3
  size 5201