gokuls commited on
Commit
e807a1c
·
1 Parent(s): c4dea29

End of training

Browse files
README.md CHANGED
@@ -1,11 +1,26 @@
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
 
 
4
  metrics:
5
  - matthews_correlation
6
  model-index:
7
  - name: hBERTv2_data_aug_cola
8
- results: []
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,9 +28,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # hBERTv2_data_aug_cola
15
 
16
- This model is a fine-tuned version of [gokuls/bert_12_layer_model_v2](https://huggingface.co/gokuls/bert_12_layer_model_v2) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.6203
19
  - Matthews Correlation: 0.0
20
 
21
  ## Model description
 
1
  ---
2
+ language:
3
+ - en
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - glue
8
  metrics:
9
  - matthews_correlation
10
  model-index:
11
  - name: hBERTv2_data_aug_cola
12
+ results:
13
+ - task:
14
+ name: Text Classification
15
+ type: text-classification
16
+ dataset:
17
+ name: GLUE COLA
18
+ type: glue
19
+ args: cola
20
+ metrics:
21
+ - name: Matthews Correlation
22
+ type: matthews_correlation
23
+ value: 0.0
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
28
 
29
  # hBERTv2_data_aug_cola
30
 
31
+ This model is a fine-tuned version of [gokuls/bert_12_layer_model_v2](https://huggingface.co/gokuls/bert_12_layer_model_v2) on the GLUE COLA dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 0.6181
34
  - Matthews Correlation: 0.0
35
 
36
  ## Model description
all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_loss": 0.6180860996246338,
4
+ "eval_matthews_correlation": 0.0,
5
+ "eval_runtime": 1.2367,
6
+ "eval_samples": 1043,
7
+ "eval_samples_per_second": 843.385,
8
+ "eval_steps_per_second": 4.043,
9
+ "train_loss": 0.6065069872462107,
10
+ "train_runtime": 5819.7121,
11
+ "train_samples": 213609,
12
+ "train_samples_per_second": 1835.22,
13
+ "train_steps_per_second": 7.174
14
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_loss": 0.6180860996246338,
4
+ "eval_matthews_correlation": 0.0,
5
+ "eval_runtime": 1.2367,
6
+ "eval_samples": 1043,
7
+ "eval_samples_per_second": 843.385,
8
+ "eval_steps_per_second": 4.043
9
+ }
logs/events.out.tfevents.1677666922.kyoto.1079667.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbab5a720b44f94c5991e05277762c512b7af89badd467cfcb8b171af64cc396
3
+ size 375
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "train_loss": 0.6065069872462107,
4
+ "train_runtime": 5819.7121,
5
+ "train_samples": 213609,
6
+ "train_samples_per_second": 1835.22,
7
+ "train_steps_per_second": 7.174
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6180860996246338,
3
+ "best_model_checkpoint": "hBERTv2_data_aug_cola/checkpoint-2505",
4
+ "epoch": 8.0,
5
+ "global_step": 6680,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 1.0,
12
+ "learning_rate": 4.9e-05,
13
+ "loss": 0.6082,
14
+ "step": 835
15
+ },
16
+ {
17
+ "epoch": 1.0,
18
+ "eval_loss": 0.6237233281135559,
19
+ "eval_matthews_correlation": 0.0,
20
+ "eval_runtime": 1.226,
21
+ "eval_samples_per_second": 850.767,
22
+ "eval_steps_per_second": 4.078,
23
+ "step": 835
24
+ },
25
+ {
26
+ "epoch": 2.0,
27
+ "learning_rate": 4.8001197604790424e-05,
28
+ "loss": 0.6068,
29
+ "step": 1670
30
+ },
31
+ {
32
+ "epoch": 2.0,
33
+ "eval_loss": 0.6185952425003052,
34
+ "eval_matthews_correlation": 0.0,
35
+ "eval_runtime": 1.2301,
36
+ "eval_samples_per_second": 847.869,
37
+ "eval_steps_per_second": 4.065,
38
+ "step": 1670
39
+ },
40
+ {
41
+ "epoch": 3.0,
42
+ "learning_rate": 4.700119760479042e-05,
43
+ "loss": 0.6064,
44
+ "step": 2505
45
+ },
46
+ {
47
+ "epoch": 3.0,
48
+ "eval_loss": 0.6180860996246338,
49
+ "eval_matthews_correlation": 0.0,
50
+ "eval_runtime": 1.2241,
51
+ "eval_samples_per_second": 852.069,
52
+ "eval_steps_per_second": 4.085,
53
+ "step": 2505
54
+ },
55
+ {
56
+ "epoch": 4.0,
57
+ "learning_rate": 4.6001197604790416e-05,
58
+ "loss": 0.6062,
59
+ "step": 3340
60
+ },
61
+ {
62
+ "epoch": 4.0,
63
+ "eval_loss": 0.6217709183692932,
64
+ "eval_matthews_correlation": 0.0,
65
+ "eval_runtime": 1.2229,
66
+ "eval_samples_per_second": 852.888,
67
+ "eval_steps_per_second": 4.089,
68
+ "step": 3340
69
+ },
70
+ {
71
+ "epoch": 5.0,
72
+ "learning_rate": 4.5001197604790425e-05,
73
+ "loss": 0.6061,
74
+ "step": 4175
75
+ },
76
+ {
77
+ "epoch": 5.0,
78
+ "eval_loss": 0.6211623549461365,
79
+ "eval_matthews_correlation": 0.0,
80
+ "eval_runtime": 1.2257,
81
+ "eval_samples_per_second": 850.945,
82
+ "eval_steps_per_second": 4.079,
83
+ "step": 4175
84
+ },
85
+ {
86
+ "epoch": 6.0,
87
+ "learning_rate": 4.400119760479042e-05,
88
+ "loss": 0.6062,
89
+ "step": 5010
90
+ },
91
+ {
92
+ "epoch": 6.0,
93
+ "eval_loss": 0.6185929179191589,
94
+ "eval_matthews_correlation": 0.0,
95
+ "eval_runtime": 1.2236,
96
+ "eval_samples_per_second": 852.395,
97
+ "eval_steps_per_second": 4.086,
98
+ "step": 5010
99
+ },
100
+ {
101
+ "epoch": 7.0,
102
+ "learning_rate": 4.3001197604790424e-05,
103
+ "loss": 0.606,
104
+ "step": 5845
105
+ },
106
+ {
107
+ "epoch": 7.0,
108
+ "eval_loss": 0.618870735168457,
109
+ "eval_matthews_correlation": 0.0,
110
+ "eval_runtime": 1.2256,
111
+ "eval_samples_per_second": 851.023,
112
+ "eval_steps_per_second": 4.08,
113
+ "step": 5845
114
+ },
115
+ {
116
+ "epoch": 8.0,
117
+ "learning_rate": 4.200119760479042e-05,
118
+ "loss": 0.6062,
119
+ "step": 6680
120
+ },
121
+ {
122
+ "epoch": 8.0,
123
+ "eval_loss": 0.6202924847602844,
124
+ "eval_matthews_correlation": 0.0,
125
+ "eval_runtime": 1.2276,
126
+ "eval_samples_per_second": 849.649,
127
+ "eval_steps_per_second": 4.073,
128
+ "step": 6680
129
+ },
130
+ {
131
+ "epoch": 8.0,
132
+ "step": 6680,
133
+ "total_flos": 2.1284586608656384e+17,
134
+ "train_loss": 0.6065069872462107,
135
+ "train_runtime": 5819.7121,
136
+ "train_samples_per_second": 1835.22,
137
+ "train_steps_per_second": 7.174
138
+ }
139
+ ],
140
+ "max_steps": 41750,
141
+ "num_train_epochs": 50,
142
+ "total_flos": 2.1284586608656384e+17,
143
+ "trial_name": null,
144
+ "trial_params": null
145
+ }