mataoxun commited on
Commit
4885003
·
1 Parent(s): 67d243c
Files changed (9) hide show
  1. config.json +15 -7
  2. optimizer.pt +1 -1
  3. pytorch_model.bin +1 -1
  4. rng_state.pth +1 -1
  5. scaler.pt +1 -1
  6. scheduler.pt +1 -1
  7. trainer_state.json +17 -5
  8. training_args.bin +1 -1
  9. vocab.json +1 -1
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "_name_or_path": "mgtoxd/tstw2",
3
- "activation_dropout": 0.1,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
@@ -47,27 +47,35 @@
47
  "do_stable_layer_norm": false,
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
50
- "feat_extract_dropout": 0.0,
51
  "feat_extract_norm": "group",
52
  "feat_proj_dropout": 0.1,
53
  "feat_quantizer_dropout": 0.0,
54
- "final_dropout": 0.1,
55
- "gradient_checkpointing": false,
56
  "hidden_act": "gelu",
57
  "hidden_dropout": 0.1,
58
- "hidden_dropout_prob": 0.1,
59
  "hidden_size": 768,
60
  "initializer_range": 0.02,
61
  "intermediate_size": 3072,
62
  "layer_norm_eps": 1e-05,
63
- "layerdrop": 0.1,
 
 
 
 
 
64
  "mask_feature_length": 10,
65
  "mask_feature_min_masks": 0,
66
  "mask_feature_prob": 0.0,
67
  "mask_time_length": 10,
68
  "mask_time_min_masks": 2,
 
 
69
  "mask_time_prob": 0.05,
 
70
  "model_type": "wav2vec2",
 
 
71
  "num_adapter_layers": 3,
72
  "num_attention_heads": 12,
73
  "num_codevector_groups": 2,
 
1
  {
2
+ "_name_or_path": "facebook/wav2vec2-base",
3
+ "activation_dropout": 0.0,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
 
47
  "do_stable_layer_norm": false,
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
 
50
  "feat_extract_norm": "group",
51
  "feat_proj_dropout": 0.1,
52
  "feat_quantizer_dropout": 0.0,
53
+ "final_dropout": 0.0,
54
+ "freeze_feat_extract_train": true,
55
  "hidden_act": "gelu",
56
  "hidden_dropout": 0.1,
 
57
  "hidden_size": 768,
58
  "initializer_range": 0.02,
59
  "intermediate_size": 3072,
60
  "layer_norm_eps": 1e-05,
61
+ "layerdrop": 0.0,
62
+ "mask_channel_length": 10,
63
+ "mask_channel_min_space": 1,
64
+ "mask_channel_other": 0.0,
65
+ "mask_channel_prob": 0.0,
66
+ "mask_channel_selection": "static",
67
  "mask_feature_length": 10,
68
  "mask_feature_min_masks": 0,
69
  "mask_feature_prob": 0.0,
70
  "mask_time_length": 10,
71
  "mask_time_min_masks": 2,
72
+ "mask_time_min_space": 1,
73
+ "mask_time_other": 0.0,
74
  "mask_time_prob": 0.05,
75
+ "mask_time_selection": "static",
76
  "model_type": "wav2vec2",
77
+ "no_mask_channel_overlap": false,
78
+ "no_mask_time_overlap": false,
79
  "num_adapter_layers": 3,
80
  "num_attention_heads": 12,
81
  "num_codevector_groups": 2,
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebe516cf8e3f6ea118fbddd93ac8470c16b8944bafd2437bffb878dcd36e867b
3
  size 721685265
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d82b346b1a1cbf09ad14b2cd28d253e62fb7311928a35ae579145ef2ca106ed
3
  size 721685265
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37f80c8621c53c60e07fdc293fb8671bd23f3c69c108a537366f0c13dceee0bd
3
  size 377656855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32a2f0295d215f2fb3eda98a12d9cd335aedaac7e75f0ee79508e4020dd2c3c8
3
  size 377656855
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ae229eaa23ea3ab390e04490b5e634a4ceafc52bb27d1ee099ebcc7a324e850
3
  size 14567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0c34772fbe5946ecc1c42b04eb7da3622057d7b126b15e1b2b9b8a4f1e53a9e
3
  size 14567
scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:715036b13c896f9739fff7fc99297bc63e28c8598be3731be660f7fe7d553229
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4588a9d5f4713074f800905c6d8ca298069bbec10540b673f5b244c28304e686
3
  size 559
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46d5a658bf4feed7551237785c3d13415d97067d7abeba64a31fd2c1cbca4a18
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2280a900bcacc0e8f5a7e6e75d396466e2d0c16f65fa69e9dbf1f1f9ae452456
3
  size 623
trainer_state.json CHANGED
@@ -1,22 +1,34 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.42589437819420783,
5
- "global_step": 500,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 0.43,
12
- "learning_rate": 4.94e-05,
13
- "loss": 5.9403,
14
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
16
  ],
17
  "max_steps": 34046,
18
  "num_train_epochs": 29,
19
- "total_flos": 1.453175312744448e+17,
20
  "trial_name": null,
21
  "trial_params": null
22
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.2776831345826234,
5
+ "global_step": 1500,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 0.43,
12
+ "learning_rate": 4.97e-05,
13
+ "loss": 3.8516,
14
  "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.85,
18
+ "learning_rate": 9.970000000000001e-05,
19
+ "loss": 1.5797,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 1.28,
24
+ "learning_rate": 9.849603582884465e-05,
25
+ "loss": 0.8674,
26
+ "step": 1500
27
  }
28
  ],
29
  "max_steps": 34046,
30
  "num_train_epochs": 29,
31
+ "total_flos": 4.388752851314688e+17,
32
  "trial_name": null,
33
  "trial_params": null
34
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b16808d7ea3a6768a261eb157b5f29a99aeca37a79dbfcb382682591ef779da6
3
  size 3247
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:549b0ff4e4321e5db62ce48db4b228cf34e5c7f6c8cd9618cfda47ceed1dbf29
3
  size 3247
vocab.json CHANGED
@@ -1 +1 @@
1
- {"s": 0, "q": 1, "x": 2, "r": 3, "h": 4, "b": 6, "a": 7, "j": 8, "n": 9, "p": 10, "k": 11, "o": 12, "g": 13, "m": 14, "f": 15, "c": 16, "y": 17, "t": 18, "u": 19, "w": 20, "e": 21, "z": 23, "i": 24, "v": 25, "d": 26, "l": 27, "[UNK]": 28, "[PAD]": 29, "|": 22}
 
1
+ {"q": 0, "j": 1, "t": 3, "o": 4, "p": 5, "a": 6, "u": 7, "b": 8, "y": 9, "d": 10, "c": 11, "h": 12, "l": 14, "r": 15, "f": 16, "i": 17, "k": 18, "s": 19, "z": 20, "m": 21, "n": 22, "v": 23, "x": 24, "e": 25, "w": 26, "g": 27, "[UNK]": 28, "[PAD]": 29, "|": 2}