Cheng98 commited on
Commit
4d085c2
·
verified ·
1 Parent(s): 82aa65c

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,13 +1,28 @@
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: bert-large-cased
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - accuracy
8
  model-index:
9
  - name: bert-large-boolq
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,10 +30,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # bert-large-boolq
17
 
18
- This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6633
21
- - Accuracy: 0.6217
22
 
23
  ## Model description
24
 
@@ -37,24 +52,16 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 5e-05
41
  - train_batch_size: 16
42
- - eval_batch_size: 32
43
  - seed: 42
44
- - gradient_accumulation_steps: 8
45
- - total_train_batch_size: 128
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
- - num_epochs: 4
49
 
50
  ### Training results
51
 
52
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
- |:-------------:|:-----:|:----:|:---------------:|:--------:|
54
- | No log | 0.99 | 73 | 0.6711 | 0.6217 |
55
- | No log | 1.99 | 147 | 0.6679 | 0.6217 |
56
- | No log | 3.0 | 221 | 0.6636 | 0.6217 |
57
- | No log | 3.96 | 292 | 0.6633 | 0.6217 |
58
 
59
 
60
  ### Framework versions
 
1
  ---
2
+ language:
3
+ - en
4
  license: apache-2.0
5
  base_model: bert-large-cased
6
  tags:
7
  - generated_from_trainer
8
+ datasets:
9
+ - glue
10
  metrics:
11
  - accuracy
12
  model-index:
13
  - name: bert-large-boolq
14
+ results:
15
+ - task:
16
+ name: Text Classification
17
+ type: text-classification
18
+ dataset:
19
+ name: GLUE BOOLQ
20
+ type: glue
21
+ args: boolq
22
+ metrics:
23
+ - name: Accuracy
24
+ type: accuracy
25
+ value: 0.7614678899082569
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
 
31
  # bert-large-boolq
32
 
33
+ This model is a fine-tuned version of [bert-large-cased](https://huggingface.co/bert-large-cased) on the GLUE BOOLQ dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 1.4013
36
+ - Accuracy: 0.7615
37
 
38
  ## Model description
39
 
 
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
+ - learning_rate: 2e-05
56
  - train_batch_size: 16
57
+ - eval_batch_size: 8
58
  - seed: 42
 
 
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
+ - num_epochs: 5.0
62
 
63
  ### Training results
64
 
 
 
 
 
 
 
65
 
66
 
67
  ### Framework versions
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 3.96,
3
- "eval_accuracy": 0.6217125382262997,
4
- "eval_loss": 0.6632964015007019,
5
- "eval_runtime": 18.4327,
6
  "eval_samples": 3270,
7
- "eval_samples_per_second": 177.402,
8
- "eval_steps_per_second": 5.588,
9
- "train_loss": 0.6762518686790989,
10
- "train_runtime": 733.7763,
11
  "train_samples": 9427,
12
- "train_samples_per_second": 51.389,
13
- "train_steps_per_second": 0.398
14
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7614678899082569,
4
+ "eval_loss": 1.4013047218322754,
5
+ "eval_runtime": 18.9743,
6
  "eval_samples": 3270,
7
+ "eval_samples_per_second": 172.338,
8
+ "eval_steps_per_second": 21.555,
9
+ "train_loss": 0.2978504161511437,
10
+ "train_runtime": 903.0769,
11
  "train_samples": 9427,
12
+ "train_samples_per_second": 52.194,
13
+ "train_steps_per_second": 3.267
14
  }
config.json CHANGED
@@ -6,20 +6,20 @@
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
  "directionality": "bidi",
9
- "finetuning_task": "text-classification",
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
14
  "id2label": {
15
- "0": "0",
16
- "1": "1"
17
  },
18
  "initializer_range": 0.02,
19
  "intermediate_size": 4096,
20
  "label2id": {
21
- "0": 0,
22
- "1": 1
23
  },
24
  "layer_norm_eps": 1e-12,
25
  "max_position_embeddings": 512,
 
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
  "directionality": "bidi",
9
+ "finetuning_task": "boolq",
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
14
  "id2label": {
15
+ "0": "False",
16
+ "1": "True"
17
  },
18
  "initializer_range": 0.02,
19
  "intermediate_size": 4096,
20
  "label2id": {
21
+ "False": 0,
22
+ "True": 1
23
  },
24
  "layer_norm_eps": 1e-12,
25
  "max_position_embeddings": 512,
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.96,
3
- "eval_accuracy": 0.6217125382262997,
4
- "eval_loss": 0.6632964015007019,
5
- "eval_runtime": 18.4327,
6
  "eval_samples": 3270,
7
- "eval_samples_per_second": 177.402,
8
- "eval_steps_per_second": 5.588
9
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "eval_accuracy": 0.7614678899082569,
4
+ "eval_loss": 1.4013047218322754,
5
+ "eval_runtime": 18.9743,
6
  "eval_samples": 3270,
7
+ "eval_samples_per_second": 172.338,
8
+ "eval_steps_per_second": 21.555
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e16c57f4a155777dd383d4dedf99bc49509e95c50dc87c1bd10019db0a6d452
3
  size 1334459697
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0ec95a4b3937605fad0b985e01db5f4aa0bb926328fe878000eb9136e2fda27
3
  size 1334459697
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 128,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 128
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 3.96,
3
- "train_loss": 0.6762518686790989,
4
- "train_runtime": 733.7763,
5
  "train_samples": 9427,
6
- "train_samples_per_second": 51.389,
7
- "train_steps_per_second": 0.398
8
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "train_loss": 0.2978504161511437,
4
+ "train_runtime": 903.0769,
5
  "train_samples": 9427,
6
+ "train_samples_per_second": 52.194,
7
+ "train_steps_per_second": 3.267
8
  }
trainer_state.json CHANGED
@@ -1,61 +1,55 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.959322033898305,
5
- "global_step": 292,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.99,
12
- "eval_accuracy": 0.6217125382262997,
13
- "eval_loss": 0.6710952520370483,
14
- "eval_runtime": 18.6871,
15
- "eval_samples_per_second": 174.987,
16
- "eval_steps_per_second": 5.512,
17
- "step": 73
18
  },
19
  {
20
- "epoch": 1.99,
21
- "eval_accuracy": 0.6217125382262997,
22
- "eval_loss": 0.667921781539917,
23
- "eval_runtime": 18.6896,
24
- "eval_samples_per_second": 174.964,
25
- "eval_steps_per_second": 5.511,
26
- "step": 147
27
  },
28
  {
29
- "epoch": 3.0,
30
- "eval_accuracy": 0.6217125382262997,
31
- "eval_loss": 0.6635896563529968,
32
- "eval_runtime": 18.7272,
33
- "eval_samples_per_second": 174.612,
34
- "eval_steps_per_second": 5.5,
35
- "step": 221
36
  },
37
  {
38
- "epoch": 3.96,
39
- "eval_accuracy": 0.6217125382262997,
40
- "eval_loss": 0.6632964015007019,
41
- "eval_runtime": 18.6887,
42
- "eval_samples_per_second": 174.972,
43
- "eval_steps_per_second": 5.511,
44
- "step": 292
45
  },
46
  {
47
- "epoch": 3.96,
48
- "step": 292,
49
- "total_flos": 3.4795521312577536e+16,
50
- "train_loss": 0.6762518686790989,
51
- "train_runtime": 733.7763,
52
- "train_samples_per_second": 51.389,
53
- "train_steps_per_second": 0.398
 
 
 
 
 
 
54
  }
55
  ],
56
- "max_steps": 292,
57
- "num_train_epochs": 4,
58
- "total_flos": 3.4795521312577536e+16,
59
  "trial_name": null,
60
  "trial_params": null
61
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
+ "global_step": 2950,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
+ "epoch": 0.85,
12
+ "learning_rate": 1.6610169491525424e-05,
13
+ "loss": 0.6459,
14
+ "step": 500
 
 
 
15
  },
16
  {
17
+ "epoch": 1.69,
18
+ "learning_rate": 1.3220338983050848e-05,
19
+ "loss": 0.4886,
20
+ "step": 1000
 
 
 
21
  },
22
  {
23
+ "epoch": 2.54,
24
+ "learning_rate": 9.830508474576272e-06,
25
+ "loss": 0.292,
26
+ "step": 1500
 
 
 
27
  },
28
  {
29
+ "epoch": 3.39,
30
+ "learning_rate": 6.440677966101695e-06,
31
+ "loss": 0.1704,
32
+ "step": 2000
 
 
 
33
  },
34
  {
35
+ "epoch": 4.24,
36
+ "learning_rate": 3.0508474576271192e-06,
37
+ "loss": 0.1074,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 5.0,
42
+ "step": 2950,
43
+ "total_flos": 1.098164620261632e+16,
44
+ "train_loss": 0.2978504161511437,
45
+ "train_runtime": 903.0769,
46
+ "train_samples_per_second": 52.194,
47
+ "train_steps_per_second": 3.267
48
  }
49
  ],
50
+ "max_steps": 2950,
51
+ "num_train_epochs": 5,
52
+ "total_flos": 1.098164620261632e+16,
53
  "trial_name": null,
54
  "trial_params": null
55
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a29dcc1637f7b320c16da7cc37c06b809c9b470839a06b41f38e4bbdbbb46bbf
3
- size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:533f3a1d23e99ef6bafd35a0187598e54d91a38adf59eea1381eac16b380fe56
3
+ size 3963