ashuc27 commited on
Commit
79f7088
·
verified ·
1 Parent(s): df5c7d5

ashuc27/DistiliBert_emotions

Browse files
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  license: apache-2.0
3
- base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
- metrics:
7
- - accuracy
8
  model-index:
9
  - name: results
10
  results: []
@@ -15,10 +15,14 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # results
17
 
18
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.6234
21
- - Accuracy: 0.7054
 
 
 
 
22
 
23
  ## Model description
24
 
@@ -37,23 +41,14 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 2.8e-05
41
- - train_batch_size: 4
42
- - eval_batch_size: 64
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - num_epochs: 3
47
 
48
- ### Training results
49
-
50
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | 0.7149 | 1.0 | 716 | 0.6495 | 0.6346 |
53
- | 0.561 | 2.0 | 1432 | 0.9508 | 0.7089 |
54
- | 0.2295 | 3.0 | 2148 | 1.5088 | 0.7173 |
55
-
56
-
57
  ### Framework versions
58
 
59
  - Transformers 4.38.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - emotion
8
  model-index:
9
  - name: results
10
  results: []
 
15
 
16
  # results
17
 
18
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
19
  It achieves the following results on the evaluation set:
20
+ - eval_loss: 1.8177
21
+ - eval_accuracy: 0.0345
22
+ - eval_runtime: 10.6974
23
+ - eval_samples_per_second: 186.961
24
+ - eval_steps_per_second: 11.685
25
+ - step: 0
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 5e-05
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 16
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - num_epochs: 3
51
 
 
 
 
 
 
 
 
 
 
52
  ### Framework versions
53
 
54
  - Transformers 4.38.2
config.json CHANGED
@@ -1,14 +1,13 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
 
3
  "architectures": [
4
- "BertForSequenceClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
- "hidden_act": "gelu",
10
- "hidden_dropout_prob": 0.1,
11
- "hidden_size": 768,
12
  "id2label": {
13
  "0": "LABEL_0",
14
  "1": "LABEL_1",
@@ -18,7 +17,6 @@
18
  "5": "LABEL_5"
19
  },
20
  "initializer_range": 0.02,
21
- "intermediate_size": 3072,
22
  "label2id": {
23
  "LABEL_0": 0,
24
  "LABEL_1": 1,
@@ -27,17 +25,17 @@
27
  "LABEL_4": 4,
28
  "LABEL_5": 5
29
  },
30
- "layer_norm_eps": 1e-12,
31
  "max_position_embeddings": 512,
32
- "model_type": "bert",
33
- "num_attention_heads": 12,
34
- "num_hidden_layers": 12,
35
  "pad_token_id": 0,
36
- "position_embedding_type": "absolute",
37
  "problem_type": "single_label_classification",
 
 
 
 
38
  "torch_dtype": "float32",
39
  "transformers_version": "4.38.2",
40
- "type_vocab_size": 2,
41
- "use_cache": true,
42
  "vocab_size": 30522
43
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForSequenceClassification"
6
  ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
 
 
11
  "id2label": {
12
  "0": "LABEL_0",
13
  "1": "LABEL_1",
 
17
  "5": "LABEL_5"
18
  },
19
  "initializer_range": 0.02,
 
20
  "label2id": {
21
  "LABEL_0": 0,
22
  "LABEL_1": 1,
 
25
  "LABEL_4": 4,
26
  "LABEL_5": 5
27
  },
 
28
  "max_position_embeddings": 512,
29
+ "model_type": "distilbert",
30
+ "n_heads": 12,
31
+ "n_layers": 6,
32
  "pad_token_id": 0,
 
33
  "problem_type": "single_label_classification",
34
+ "qa_dropout": 0.1,
35
+ "seq_classif_dropout": 0.2,
36
+ "sinusoidal_pos_embds": false,
37
+ "tie_weights_": true,
38
  "torch_dtype": "float32",
39
  "transformers_version": "4.38.2",
 
 
40
  "vocab_size": 30522
41
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28d5fda8bdbfb4833b67583bf5d4297ff2ebfd86b2f1033660bb3bd6ba15de41
3
- size 437970952
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93b50567c9209e758921f1dfb02ae7e291a63e870feb2443af4ddca5628361f2
3
+ size 267844872
runs/Mar15_00-56-40_63a832ae8ab7/events.out.tfevents.1710464286.63a832ae8ab7.15461.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e0a260156a04cb3a0ffdef9ec62bc9605bd0b52365e1c10f5f06e310aafbf1a
3
+ size 346
runs/Mar15_01-09-10_63a832ae8ab7/events.out.tfevents.1710465036.63a832ae8ab7.20183.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71eeed3d1248b23c40d178e71c1cfc012905c89742c83f89d9f00ee00d4ab53a
3
+ size 346
runs/Mar15_01-11-23_63a832ae8ab7/events.out.tfevents.1710465093.63a832ae8ab7.20183.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df706ee38ad138ec59fa1a3daa6bf58d97c708f0b1789daab9b5b198576fa89f
3
+ size 346
runs/Mar15_01-14-01_63a832ae8ab7/events.out.tfevents.1710465251.63a832ae8ab7.20183.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79a4a465dbaa76cb512319bbe78eb1f137e9002f82482ce712e6cff428b2d288
3
+ size 346
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9eb3e141ef2273dd9a8db7310724a881b781800037a627082affdc731af360f
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c90f26f1b131452e9fd4771cfb540b5196e03c1ac79954eebe0401eb92b9af6b
3
  size 4856