Shresthadev403 commited on
Commit
962b6f2
·
1 Parent(s): 361a325

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: apache-2.0
3
- base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,14 +11,14 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # ner-bert-ingredients
15
 
16
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 4.3994
19
- - eval_runtime: 0.0358
20
- - eval_samples_per_second: 55.921
21
- - eval_steps_per_second: 27.961
22
- - epoch: 4.0
23
- - step: 4
24
 
25
  ## Model description
26
 
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
 
11
 
12
  # ner-bert-ingredients
13
 
14
+ This model was trained from scratch on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
+ - eval_loss: 4.3799
17
+ - eval_runtime: 0.0344
18
+ - eval_samples_per_second: 58.156
19
+ - eval_steps_per_second: 29.078
20
+ - epoch: 1.0
21
+ - step: 1
22
 
23
  ## Model description
24
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "/content/ner-bert-ingredients",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db5998d5981ed4d7d2e689fcb8b2859cb00a779f5286f66518b13cad8fc88c65
3
  size 435820636
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ba26a04cab698755f4a7d92419db4a4ba392c75e099f1962f4ea06b132a0f6b
3
  size 435820636
runs/Jan05_08-07-59_414819e23027/events.out.tfevents.1704442090.414819e23027.9579.16 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b3e444ad074ed53106c5d4f32f3a11355bc31b16ea7b79425c50be1b44692fe
3
- size 9056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef802b840554e3c51bf02a13f02a625c351754cc11494990c7821ffb663cbc3
3
+ size 9476
runs/Jan05_08-11-12_414819e23027/events.out.tfevents.1704442276.414819e23027.9579.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82da206fbae0e53c1564958ebd2ed5dcf8e463ba1692cfe7ecc6d96afebc2969
3
+ size 7847
special_tokens_map.json CHANGED
@@ -1,7 +1,37 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
  }
tokenizer_config.json CHANGED
@@ -45,11 +45,18 @@
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
 
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
 
 
50
  "sep_token": "[SEP]",
 
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
  "tokenizer_class": "BertTokenizer",
 
 
54
  "unk_token": "[UNK]"
55
  }
 
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
+ "max_length": 512,
49
  "model_max_length": 512,
50
+ "pad_to_multiple_of": null,
51
  "pad_token": "[PAD]",
52
+ "pad_token_type_id": 0,
53
+ "padding_side": "right",
54
  "sep_token": "[SEP]",
55
+ "stride": 0,
56
  "strip_accents": null,
57
  "tokenize_chinese_chars": true,
58
  "tokenizer_class": "BertTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
  "unk_token": "[UNK]"
62
  }
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 4.0,
5
  "eval_steps": 1,
6
- "global_step": 4,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -11,65 +11,23 @@
11
  {
12
  "epoch": 1.0,
13
  "learning_rate": 4.9e-05,
14
- "loss": 4.3439,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 1.0,
19
- "eval_loss": 4.352994441986084,
20
- "eval_runtime": 0.045,
21
- "eval_samples_per_second": 44.444,
22
- "eval_steps_per_second": 22.222,
23
  "step": 1
24
- },
25
- {
26
- "epoch": 2.0,
27
- "learning_rate": 4.8e-05,
28
- "loss": 4.0741,
29
- "step": 2
30
- },
31
- {
32
- "epoch": 2.0,
33
- "eval_loss": 4.369075775146484,
34
- "eval_runtime": 0.0298,
35
- "eval_samples_per_second": 67.128,
36
- "eval_steps_per_second": 33.564,
37
- "step": 2
38
- },
39
- {
40
- "epoch": 3.0,
41
- "learning_rate": 4.7e-05,
42
- "loss": 3.8404,
43
- "step": 3
44
- },
45
- {
46
- "epoch": 3.0,
47
- "eval_loss": 4.391932964324951,
48
- "eval_runtime": 0.0266,
49
- "eval_samples_per_second": 75.108,
50
- "eval_steps_per_second": 37.554,
51
- "step": 3
52
- },
53
- {
54
- "epoch": 4.0,
55
- "learning_rate": 4.600000000000001e-05,
56
- "loss": 3.6288,
57
- "step": 4
58
- },
59
- {
60
- "epoch": 4.0,
61
- "eval_loss": 4.39936637878418,
62
- "eval_runtime": 0.0358,
63
- "eval_samples_per_second": 55.921,
64
- "eval_steps_per_second": 27.961,
65
- "step": 4
66
  }
67
  ],
68
  "logging_steps": 1,
69
  "max_steps": 50,
70
  "num_train_epochs": 50,
71
  "save_steps": 20,
72
- "total_flos": 2304197409600.0,
73
  "trial_name": null,
74
  "trial_params": null
75
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 1,
6
+ "global_step": 1,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
11
  {
12
  "epoch": 1.0,
13
  "learning_rate": 4.9e-05,
14
+ "loss": 2.9799,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 1.0,
19
+ "eval_loss": 4.379878520965576,
20
+ "eval_runtime": 0.0344,
21
+ "eval_samples_per_second": 58.156,
22
+ "eval_steps_per_second": 29.078,
23
  "step": 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  }
25
  ],
26
  "logging_steps": 1,
27
  "max_steps": 50,
28
  "num_train_epochs": 50,
29
  "save_steps": 20,
30
+ "total_flos": 576049352400.0,
31
  "trial_name": null,
32
  "trial_params": null
33
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc908e19763faf6f872a93d2c05e1926d0092455daf2d5c37462599bf9734099
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee86837266bbc5e6172c9e09bcb500cf131b1fcc7f02b545c8245aad1f23abde
3
  size 4600