Cheng98 commited on
Commit
e73acfa
·
verified ·
1 Parent(s): c731831

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
- value: 0.703971119133574
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE RTE dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.7109
36
- - Accuracy: 0.7040
37
 
38
  ## Model description
39
 
@@ -53,12 +53,12 @@ More information needed
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 2e-05
56
- - train_batch_size: 32
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
- - num_epochs: 3.0
62
 
63
  ### Training results
64
 
 
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
+ value: 0.7111913357400722
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE RTE dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 1.4630
36
+ - Accuracy: 0.7112
37
 
38
  ## Model description
39
 
 
53
 
54
  The following hyperparameters were used during training:
55
  - learning_rate: 2e-05
56
+ - train_batch_size: 16
57
  - eval_batch_size: 8
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
+ - num_epochs: 5.0
62
 
63
  ### Training results
64
 
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 3.0,
3
- "eval_accuracy": 0.703971119133574,
4
- "eval_loss": 0.7108999490737915,
5
- "eval_runtime": 1.628,
6
  "eval_samples": 277,
7
- "eval_samples_per_second": 170.149,
8
- "eval_steps_per_second": 21.499,
9
- "train_loss": 0.4925964877136752,
10
- "train_runtime": 120.7223,
11
  "train_samples": 2490,
12
- "train_samples_per_second": 61.878,
13
- "train_steps_per_second": 1.938
14
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7111913357400722,
4
+ "eval_loss": 1.4629547595977783,
5
+ "eval_runtime": 0.7868,
6
  "eval_samples": 277,
7
+ "eval_samples_per_second": 352.04,
8
+ "eval_steps_per_second": 44.482,
9
+ "train_loss": 0.29059823354085285,
10
+ "train_runtime": 86.2993,
11
  "train_samples": 2490,
12
+ "train_samples_per_second": 144.265,
13
+ "train_steps_per_second": 9.038
14
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.0,
3
- "eval_accuracy": 0.703971119133574,
4
- "eval_loss": 0.7108999490737915,
5
- "eval_runtime": 1.628,
6
  "eval_samples": 277,
7
- "eval_samples_per_second": 170.149,
8
- "eval_steps_per_second": 21.499
9
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7111913357400722,
4
+ "eval_loss": 1.4629547595977783,
5
+ "eval_runtime": 0.7868,
6
  "eval_samples": 277,
7
+ "eval_samples_per_second": 352.04,
8
+ "eval_steps_per_second": 44.482
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ad0b31d43a528f235b2d0a53058a6d07df9d46ea4b8bf48a275a4580cd660c9
3
  size 1334459697
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bd302f63dd79574574174160d4dfb43d6baf76736ccd987db986ab26670406f
3
  size 1334459697
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 3.0,
3
- "train_loss": 0.4925964877136752,
4
- "train_runtime": 120.7223,
5
  "train_samples": 2490,
6
- "train_samples_per_second": 61.878,
7
- "train_steps_per_second": 1.938
8
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "train_loss": 0.29059823354085285,
4
+ "train_runtime": 86.2993,
5
  "train_samples": 2490,
6
+ "train_samples_per_second": 144.265,
7
+ "train_steps_per_second": 9.038
8
  }
trainer_state.json CHANGED
@@ -1,25 +1,31 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
- "global_step": 234,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 3.0,
12
- "step": 234,
13
- "total_flos": 1740381821015040.0,
14
- "train_loss": 0.4925964877136752,
15
- "train_runtime": 120.7223,
16
- "train_samples_per_second": 61.878,
17
- "train_steps_per_second": 1.938
 
 
 
 
 
 
18
  }
19
  ],
20
- "max_steps": 234,
21
- "num_train_epochs": 3,
22
- "total_flos": 1740381821015040.0,
23
  "trial_name": null,
24
  "trial_params": null
25
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
+ "global_step": 780,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 3.21,
12
+ "learning_rate": 7.256410256410257e-06,
13
+ "loss": 0.4017,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 5.0,
18
+ "step": 780,
19
+ "total_flos": 2900636368358400.0,
20
+ "train_loss": 0.29059823354085285,
21
+ "train_runtime": 86.2993,
22
+ "train_samples_per_second": 144.265,
23
+ "train_steps_per_second": 9.038
24
  }
25
  ],
26
+ "max_steps": 780,
27
+ "num_train_epochs": 5,
28
+ "total_flos": 2900636368358400.0,
29
  "trial_name": null,
30
  "trial_params": null
31
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4399aaeb798913062f8219ace308d12474eafbe83a8cc2e0ee9fb2296004ef7
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa4fa4a4306e4564d195ad7d5668dae861e97971595a52cb9b0aac9538125f04
3
  size 3963