Shresthadev403 commited on
Commit
5c1fce9
·
1 Parent(s): 9e260b3

End of training

Browse files
README.md CHANGED
@@ -15,12 +15,12 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 8.6995
19
- - eval_runtime: 83.8513
20
- - eval_samples_per_second: 59.629
21
- - eval_steps_per_second: 0.942
22
- - epoch: 0.03
23
- - step: 40
24
 
25
  ## Model description
26
 
@@ -43,11 +43,11 @@ The following hyperparameters were used during training:
43
  - train_batch_size: 16
44
  - eval_batch_size: 64
45
  - seed: 42
46
- - gradient_accumulation_steps: 4
47
- - total_train_batch_size: 64
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 50
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Framework versions
 
15
 
16
  This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - eval_loss: 2.8215
19
+ - eval_runtime: 84.9249
20
+ - eval_samples_per_second: 58.876
21
+ - eval_steps_per_second: 0.93
22
+ - epoch: 2.02
23
+ - step: 1500
24
 
25
  ## Model description
26
 
 
43
  - train_batch_size: 16
44
  - eval_batch_size: 64
45
  - seed: 42
46
+ - gradient_accumulation_steps: 8
47
+ - total_train_batch_size: 128
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - num_epochs: 100
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d798163ba1887a6557a6b3f33e6aa239ed97bbbd2c9f931d72c44c7cd8eeb88
3
  size 535667604
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deef4e21ca3e5ef60fd595e9b0c12db796046b14fcd1183b786136a45ff02da3
3
  size 535667604
runs/Jan05_09-37-14_be54e94fe191/events.out.tfevents.1704447448.be54e94fe191.25.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29e2786d518ed72e256917d510d90de7500fae644419846b4870b0ed45759115
3
+ size 1717297
trainer_state.json CHANGED
@@ -1,48 +1,34 @@
1
  {
2
- "best_metric": 9.321273803710938,
3
- "best_model_checkpoint": "ner-bert-ingredients/checkpoint-20",
4
- "epoch": 0.02694509936005389,
5
- "eval_steps": 20,
6
- "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.01,
13
- "learning_rate": 4.9986522911051216e-05,
14
- "loss": 9.8102,
15
- "step": 20
16
  },
17
  {
18
- "epoch": 0.01,
19
- "eval_loss": 9.321273803710938,
20
- "eval_runtime": 84.1492,
21
- "eval_samples_per_second": 59.418,
22
- "eval_steps_per_second": 0.939,
23
- "step": 20
24
- },
25
- {
26
- "epoch": 0.03,
27
- "learning_rate": 4.9973045822102424e-05,
28
- "loss": 9.0849,
29
- "step": 40
30
- },
31
- {
32
- "epoch": 0.03,
33
- "eval_loss": 8.699505805969238,
34
- "eval_runtime": 83.8513,
35
- "eval_samples_per_second": 59.629,
36
- "eval_steps_per_second": 0.942,
37
- "step": 40
38
  }
39
  ],
40
- "logging_steps": 20,
41
  "max_steps": 74200,
42
  "num_input_tokens_seen": 0,
43
- "num_train_epochs": 50,
44
- "save_steps": 20,
45
- "total_flos": 865668287692800.0,
46
  "train_batch_size": 16,
47
  "trial_name": null,
48
  "trial_params": null
 
1
  {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0208824520040416,
5
+ "eval_steps": 1500,
6
+ "global_step": 1500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 2.02,
13
+ "learning_rate": 4.8989218328840976e-05,
14
+ "loss": 4.3834,
15
+ "step": 1500
16
  },
17
  {
18
+ "epoch": 2.02,
19
+ "eval_loss": 2.8215107917785645,
20
+ "eval_runtime": 84.9249,
21
+ "eval_samples_per_second": 58.876,
22
+ "eval_steps_per_second": 0.93,
23
+ "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  }
25
  ],
26
+ "logging_steps": 1500,
27
  "max_steps": 74200,
28
  "num_input_tokens_seen": 0,
29
+ "num_train_epochs": 100,
30
+ "save_steps": 1500,
31
+ "total_flos": 6.491971115016192e+16,
32
  "train_batch_size": 16,
33
  "trial_name": null,
34
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7e0cc5c26c36bde050839a58f0c4f021e92cbaffecf465dbdc0090a08f22c21
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f265e1dd273ce96b78c47255a0e2ff65085a88d5afd5c3adc7b2f78aafd5a2a8
3
  size 4283