Commit ·
5c1fce9
1
Parent(s): 9e260b3
End of training
Browse files- README.md +9 -9
- model.safetensors +1 -1
- runs/Jan05_09-37-14_be54e94fe191/events.out.tfevents.1704447448.be54e94fe191.25.0 +3 -0
- trainer_state.json +19 -33
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -15,12 +15,12 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 15 |
|
| 16 |
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
|
| 17 |
It achieves the following results on the evaluation set:
|
| 18 |
-
- eval_loss:
|
| 19 |
-
- eval_runtime:
|
| 20 |
-
- eval_samples_per_second:
|
| 21 |
-
- eval_steps_per_second: 0.
|
| 22 |
-
- epoch:
|
| 23 |
-
- step:
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
@@ -43,11 +43,11 @@ The following hyperparameters were used during training:
|
|
| 43 |
- train_batch_size: 16
|
| 44 |
- eval_batch_size: 64
|
| 45 |
- seed: 42
|
| 46 |
-
- gradient_accumulation_steps:
|
| 47 |
-
- total_train_batch_size:
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
| 50 |
-
- num_epochs:
|
| 51 |
- mixed_precision_training: Native AMP
|
| 52 |
|
| 53 |
### Framework versions
|
|
|
|
| 15 |
|
| 16 |
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
|
| 17 |
It achieves the following results on the evaluation set:
|
| 18 |
+
- eval_loss: 2.8215
|
| 19 |
+
- eval_runtime: 84.9249
|
| 20 |
+
- eval_samples_per_second: 58.876
|
| 21 |
+
- eval_steps_per_second: 0.93
|
| 22 |
+
- epoch: 2.02
|
| 23 |
+
- step: 1500
|
| 24 |
|
| 25 |
## Model description
|
| 26 |
|
|
|
|
| 43 |
- train_batch_size: 16
|
| 44 |
- eval_batch_size: 64
|
| 45 |
- seed: 42
|
| 46 |
+
- gradient_accumulation_steps: 8
|
| 47 |
+
- total_train_batch_size: 128
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
| 50 |
+
- num_epochs: 100
|
| 51 |
- mixed_precision_training: Native AMP
|
| 52 |
|
| 53 |
### Framework versions
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 535667604
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:deef4e21ca3e5ef60fd595e9b0c12db796046b14fcd1183b786136a45ff02da3
|
| 3 |
size 535667604
|
runs/Jan05_09-37-14_be54e94fe191/events.out.tfevents.1704447448.be54e94fe191.25.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:29e2786d518ed72e256917d510d90de7500fae644419846b4870b0ed45759115
|
| 3 |
+
size 1717297
|
trainer_state.json
CHANGED
|
@@ -1,48 +1,34 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric":
|
| 3 |
-
"best_model_checkpoint":
|
| 4 |
-
"epoch":
|
| 5 |
-
"eval_steps":
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
-
"epoch":
|
| 13 |
-
"learning_rate": 4.
|
| 14 |
-
"loss":
|
| 15 |
-
"step":
|
| 16 |
},
|
| 17 |
{
|
| 18 |
-
"epoch":
|
| 19 |
-
"eval_loss":
|
| 20 |
-
"eval_runtime": 84.
|
| 21 |
-
"eval_samples_per_second":
|
| 22 |
-
"eval_steps_per_second": 0.
|
| 23 |
-
"step":
|
| 24 |
-
},
|
| 25 |
-
{
|
| 26 |
-
"epoch": 0.03,
|
| 27 |
-
"learning_rate": 4.9973045822102424e-05,
|
| 28 |
-
"loss": 9.0849,
|
| 29 |
-
"step": 40
|
| 30 |
-
},
|
| 31 |
-
{
|
| 32 |
-
"epoch": 0.03,
|
| 33 |
-
"eval_loss": 8.699505805969238,
|
| 34 |
-
"eval_runtime": 83.8513,
|
| 35 |
-
"eval_samples_per_second": 59.629,
|
| 36 |
-
"eval_steps_per_second": 0.942,
|
| 37 |
-
"step": 40
|
| 38 |
}
|
| 39 |
],
|
| 40 |
-
"logging_steps":
|
| 41 |
"max_steps": 74200,
|
| 42 |
"num_input_tokens_seen": 0,
|
| 43 |
-
"num_train_epochs":
|
| 44 |
-
"save_steps":
|
| 45 |
-
"total_flos":
|
| 46 |
"train_batch_size": 16,
|
| 47 |
"trial_name": null,
|
| 48 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 2.0208824520040416,
|
| 5 |
+
"eval_steps": 1500,
|
| 6 |
+
"global_step": 1500,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
+
"epoch": 2.02,
|
| 13 |
+
"learning_rate": 4.8989218328840976e-05,
|
| 14 |
+
"loss": 4.3834,
|
| 15 |
+
"step": 1500
|
| 16 |
},
|
| 17 |
{
|
| 18 |
+
"epoch": 2.02,
|
| 19 |
+
"eval_loss": 2.8215107917785645,
|
| 20 |
+
"eval_runtime": 84.9249,
|
| 21 |
+
"eval_samples_per_second": 58.876,
|
| 22 |
+
"eval_steps_per_second": 0.93,
|
| 23 |
+
"step": 1500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
}
|
| 25 |
],
|
| 26 |
+
"logging_steps": 1500,
|
| 27 |
"max_steps": 74200,
|
| 28 |
"num_input_tokens_seen": 0,
|
| 29 |
+
"num_train_epochs": 100,
|
| 30 |
+
"save_steps": 1500,
|
| 31 |
+
"total_flos": 6.491971115016192e+16,
|
| 32 |
"train_batch_size": 16,
|
| 33 |
"trial_name": null,
|
| 34 |
"trial_params": null
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4283
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f265e1dd273ce96b78c47255a0e2ff65085a88d5afd5c3adc7b2f78aafd5a2a8
|
| 3 |
size 4283
|