Selma323 commited on
Commit
a4c5d72
·
verified ·
1 Parent(s): 9cc61e2

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "facebook/hubert-base-ls960",
3
  "activation_dropout": 0.1,
4
- "apply_spec_augment": true,
5
  "architectures": [
6
  "HubertForSequenceClassification"
7
  ],
@@ -41,15 +41,12 @@
41
  "do_stable_layer_norm": false,
42
  "eos_token_id": 2,
43
  "feat_extract_activation": "gelu",
44
- "feat_extract_dropout": 0.0,
45
  "feat_extract_norm": "group",
46
- "feat_proj_dropout": 0.1,
47
- "feat_proj_layer_norm": true,
48
- "final_dropout": 0.1,
49
- "gradient_checkpointing": false,
50
  "hidden_act": "gelu",
51
  "hidden_dropout": 0.1,
52
- "hidden_dropout_prob": 0.1,
53
  "hidden_size": 768,
54
  "id2label": {
55
  "0": "Ataxic",
@@ -64,7 +61,7 @@
64
  "Spastic": "2"
65
  },
66
  "layer_norm_eps": 1e-05,
67
- "layerdrop": 0.1,
68
  "mask_feature_length": 10,
69
  "mask_feature_min_masks": 0,
70
  "mask_feature_prob": 0.0,
@@ -76,9 +73,8 @@
76
  "num_conv_pos_embedding_groups": 16,
77
  "num_conv_pos_embeddings": 128,
78
  "num_feat_extract_layers": 7,
79
- "num_hidden_layers": 12,
80
  "pad_token_id": 0,
81
- "tokenizer_class": "Wav2Vec2CTCTokenizer",
82
  "torch_dtype": "float32",
83
  "transformers_version": "4.40.1",
84
  "use_weighted_layer_sum": false,
 
1
  {
2
+ "_name_or_path": "ntu-spml/distilhubert",
3
  "activation_dropout": 0.1,
4
+ "apply_spec_augment": false,
5
  "architectures": [
6
  "HubertForSequenceClassification"
7
  ],
 
41
  "do_stable_layer_norm": false,
42
  "eos_token_id": 2,
43
  "feat_extract_activation": "gelu",
 
44
  "feat_extract_norm": "group",
45
+ "feat_proj_dropout": 0.0,
46
+ "feat_proj_layer_norm": false,
47
+ "final_dropout": 0.0,
 
48
  "hidden_act": "gelu",
49
  "hidden_dropout": 0.1,
 
50
  "hidden_size": 768,
51
  "id2label": {
52
  "0": "Ataxic",
 
61
  "Spastic": "2"
62
  },
63
  "layer_norm_eps": 1e-05,
64
+ "layerdrop": 0.0,
65
  "mask_feature_length": 10,
66
  "mask_feature_min_masks": 0,
67
  "mask_feature_prob": 0.0,
 
73
  "num_conv_pos_embedding_groups": 16,
74
  "num_conv_pos_embeddings": 128,
75
  "num_feat_extract_layers": 7,
76
+ "num_hidden_layers": 2,
77
  "pad_token_id": 0,
 
78
  "torch_dtype": "float32",
79
  "transformers_version": "4.40.1",
80
  "use_weighted_layer_sum": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee30b5ab2012986902113ca4d329b59103929f65ec1dd92b7602c22a82049dff
3
- size 378302972
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7fad838ce4c18f170eb96c24465d45b4a6023ad6ea5726635028f2509cc115c
3
+ size 94764524
runs/Apr24_16-21-14_9f17f299def7/events.out.tfevents.1713975675.9f17f299def7.16670.6 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1beaee1f6e0a60ad6eac52b761bbbc01e919cc0a47e11c3964ab70f8445c78f7
3
- size 9100
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7871f292af963a93b9f399e9cb5ae24bf2d66b404e1c67cceba98156995e4db
3
+ size 9448
runs/Apr24_16-41-06_9f17f299def7/events.out.tfevents.1713976868.9f17f299def7.16670.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f08210c196c006c82579a2947ae7444ff9be987b8d20367682dd5740b0794d4a
3
+ size 5923
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a96febc9eeac8d4976bd7d6f3376a374e9504045ddfe064692a62a55200c792d
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49602796043cceabedf13c3bbf203519d896511228188cdd5b069091af6bfcda
3
  size 4984