Hartunka commited on
Commit
4bb4804
·
verified ·
1 Parent(s): 8918191

End of training

Browse files
README.md CHANGED
@@ -1,13 +1,31 @@
1
  ---
 
 
2
  base_model: Hartunka/tiny_bert_rand_50_v1
3
  tags:
4
  - generated_from_trainer
 
 
5
  metrics:
6
  - matthews_correlation
7
  - accuracy
8
  model-index:
9
  - name: tiny_bert_rand_50_v1_cola
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,11 +33,11 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # tiny_bert_rand_50_v1_cola
17
 
18
- This model is a fine-tuned version of [Hartunka/tiny_bert_rand_50_v1](https://huggingface.co/Hartunka/tiny_bert_rand_50_v1) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.7260
21
- - Matthews Correlation: 0.0758
22
- - Accuracy: 0.6222
23
 
24
  ## Model description
25
 
 
1
  ---
2
+ language:
3
+ - en
4
  base_model: Hartunka/tiny_bert_rand_50_v1
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - glue
9
  metrics:
10
  - matthews_correlation
11
  - accuracy
12
  model-index:
13
  - name: tiny_bert_rand_50_v1_cola
14
+ results:
15
+ - task:
16
+ name: Text Classification
17
+ type: text-classification
18
+ dataset:
19
+ name: GLUE COLA
20
+ type: glue
21
+ args: cola
22
+ metrics:
23
+ - name: Matthews Correlation
24
+ type: matthews_correlation
25
+ value: 0.0
26
+ - name: Accuracy
27
+ type: accuracy
28
+ value: 0.6912751793861389
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
33
 
34
  # tiny_bert_rand_50_v1_cola
35
 
36
+ This model is a fine-tuned version of [Hartunka/tiny_bert_rand_50_v1](https://huggingface.co/Hartunka/tiny_bert_rand_50_v1) on the GLUE COLA dataset.
37
  It achieves the following results on the evaluation set:
38
+ - Loss: 0.6171
39
+ - Matthews Correlation: 0.0
40
+ - Accuracy: 0.6913
41
 
42
  ## Model description
43
 
all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "eval_accuracy": 0.6912751793861389,
4
+ "eval_loss": 0.6171163320541382,
5
+ "eval_matthews_correlation": 0.0,
6
+ "eval_runtime": 0.4905,
7
+ "eval_samples": 1043,
8
+ "eval_samples_per_second": 2126.307,
9
+ "eval_steps_per_second": 10.193,
10
+ "total_flos": 1345426898614272.0,
11
+ "train_loss": 0.5469665293600044,
12
+ "train_runtime": 42.0475,
13
+ "train_samples": 8551,
14
+ "train_samples_per_second": 10168.271,
15
+ "train_steps_per_second": 40.43
16
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "eval_accuracy": 0.6912751793861389,
4
+ "eval_loss": 0.6171163320541382,
5
+ "eval_matthews_correlation": 0.0,
6
+ "eval_runtime": 0.4905,
7
+ "eval_samples": 1043,
8
+ "eval_samples_per_second": 2126.307,
9
+ "eval_steps_per_second": 10.193
10
+ }
logs/events.out.tfevents.1744824182.s_004_m ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f58b90f86be80d7bba4200f3541df0138def2bb95ed42987b021a85e829dd28
3
+ size 40
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "total_flos": 1345426898614272.0,
4
+ "train_loss": 0.5469665293600044,
5
+ "train_runtime": 42.0475,
6
+ "train_samples": 8551,
7
+ "train_samples_per_second": 10168.271,
8
+ "train_steps_per_second": 40.43
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6171163320541382,
3
+ "best_model_checkpoint": "tiny_bert_rand_50_v1_cola/checkpoint-34",
4
+ "epoch": 6.0,
5
+ "eval_steps": 500,
6
+ "global_step": 204,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 0.7138214111328125,
14
+ "learning_rate": 4.9e-05,
15
+ "loss": 0.6134,
16
+ "step": 34
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.6912751793861389,
21
+ "eval_loss": 0.6171163320541382,
22
+ "eval_matthews_correlation": 0.0,
23
+ "eval_runtime": 0.4916,
24
+ "eval_samples_per_second": 2121.452,
25
+ "eval_steps_per_second": 10.17,
26
+ "step": 34
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 0.9771915078163147,
31
+ "learning_rate": 4.8e-05,
32
+ "loss": 0.601,
33
+ "step": 68
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.6912751793861389,
38
+ "eval_loss": 0.6195900440216064,
39
+ "eval_matthews_correlation": 0.0,
40
+ "eval_runtime": 0.3924,
41
+ "eval_samples_per_second": 2657.785,
42
+ "eval_steps_per_second": 12.741,
43
+ "step": 68
44
+ },
45
+ {
46
+ "epoch": 3.0,
47
+ "grad_norm": 1.1553252935409546,
48
+ "learning_rate": 4.7e-05,
49
+ "loss": 0.5781,
50
+ "step": 102
51
+ },
52
+ {
53
+ "epoch": 3.0,
54
+ "eval_accuracy": 0.6912751793861389,
55
+ "eval_loss": 0.6259943246841431,
56
+ "eval_matthews_correlation": 0.0,
57
+ "eval_runtime": 0.3832,
58
+ "eval_samples_per_second": 2721.81,
59
+ "eval_steps_per_second": 13.048,
60
+ "step": 102
61
+ },
62
+ {
63
+ "epoch": 4.0,
64
+ "grad_norm": 1.9693224430084229,
65
+ "learning_rate": 4.600000000000001e-05,
66
+ "loss": 0.5374,
67
+ "step": 136
68
+ },
69
+ {
70
+ "epoch": 4.0,
71
+ "eval_accuracy": 0.6836050152778625,
72
+ "eval_loss": 0.6578595042228699,
73
+ "eval_matthews_correlation": 0.042582902474918326,
74
+ "eval_runtime": 0.3804,
75
+ "eval_samples_per_second": 2741.559,
76
+ "eval_steps_per_second": 13.143,
77
+ "step": 136
78
+ },
79
+ {
80
+ "epoch": 5.0,
81
+ "grad_norm": 2.186906337738037,
82
+ "learning_rate": 4.5e-05,
83
+ "loss": 0.4953,
84
+ "step": 170
85
+ },
86
+ {
87
+ "epoch": 5.0,
88
+ "eval_accuracy": 0.6538830399513245,
89
+ "eval_loss": 0.6752046942710876,
90
+ "eval_matthews_correlation": 0.08801919690309869,
91
+ "eval_runtime": 0.3951,
92
+ "eval_samples_per_second": 2640.126,
93
+ "eval_steps_per_second": 12.656,
94
+ "step": 170
95
+ },
96
+ {
97
+ "epoch": 6.0,
98
+ "grad_norm": 2.2683961391448975,
99
+ "learning_rate": 4.4000000000000006e-05,
100
+ "loss": 0.4567,
101
+ "step": 204
102
+ },
103
+ {
104
+ "epoch": 6.0,
105
+ "eval_accuracy": 0.6222435235977173,
106
+ "eval_loss": 0.7260271310806274,
107
+ "eval_matthews_correlation": 0.07575741604681904,
108
+ "eval_runtime": 0.4037,
109
+ "eval_samples_per_second": 2583.494,
110
+ "eval_steps_per_second": 12.385,
111
+ "step": 204
112
+ },
113
+ {
114
+ "epoch": 6.0,
115
+ "step": 204,
116
+ "total_flos": 1345426898614272.0,
117
+ "train_loss": 0.5469665293600044,
118
+ "train_runtime": 42.0475,
119
+ "train_samples_per_second": 10168.271,
120
+ "train_steps_per_second": 40.43
121
+ }
122
+ ],
123
+ "logging_steps": 1,
124
+ "max_steps": 1700,
125
+ "num_input_tokens_seen": 0,
126
+ "num_train_epochs": 50,
127
+ "save_steps": 500,
128
+ "total_flos": 1345426898614272.0,
129
+ "train_batch_size": 256,
130
+ "trial_name": null,
131
+ "trial_params": null
132
+ }