hiudev commited on
Commit
ad612ff
·
verified ·
1 Parent(s): faf6bf5

Initial release

Browse files
Files changed (4) hide show
  1. README.md +17 -15
  2. config.json +40 -44
  3. model.safetensors +2 -2
  4. training_args.bin +2 -2
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: mit
4
- base_model: dslim/bert-base-NER
5
  tags:
6
  - ner
7
  - bert
@@ -17,16 +17,16 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # NER-BERT
19
 
20
- This model is a fine-tuned version of [dslim/bert-base-NER](https://huggingface.co/dslim/bert-base-NER) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.0001
23
- - Token Accuracy: 1.0000
24
- - Token Precision: 1.0000
25
- - Token Recall: 1.0000
26
- - Token F1: 1.0000
27
- - Entity Precision: 0.9998
28
- - Entity Recall: 0.9998
29
- - Entity F1: 0.9998
30
 
31
  ## Model description
32
 
@@ -49,6 +49,8 @@ The following hyperparameters were used during training:
49
  - train_batch_size: 16
50
  - eval_batch_size: 16
51
  - seed: 42
 
 
52
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
53
  - lr_scheduler_type: linear
54
  - lr_scheduler_warmup_steps: 500
@@ -58,14 +60,14 @@ The following hyperparameters were used during training:
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Token Accuracy | Token Precision | Token Recall | Token F1 | Entity Precision | Entity Recall | Entity F1 |
60
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:---------------:|:------------:|:--------:|:----------------:|:-------------:|:---------:|
61
- | 0.0004 | 1.0 | 2250 | 0.0002 | 1.0000 | 1.0000 | 1.0000 | 1.0000 | 0.9994 | 0.9995 | 0.9995 |
62
- | 0.0001 | 2.0 | 4500 | 0.0001 | 1.0000 | 1.0000 | 1.0000 | 1.0000 | 0.9998 | 0.9998 | 0.9998 |
63
- | 0.0001 | 3.0 | 6750 | 0.0001 | 1.0000 | 1.0000 | 1.0000 | 1.0000 | 0.9998 | 0.9998 | 0.9998 |
64
 
65
 
66
  ### Framework versions
67
 
68
- - Transformers 4.50.2
69
  - Pytorch 2.6.0+cu124
70
- - Datasets 3.5.0
71
  - Tokenizers 0.21.1
 
1
  ---
2
  library_name: transformers
3
  license: mit
4
+ base_model: roberta-base
5
  tags:
6
  - ner
7
  - bert
 
17
 
18
  # NER-BERT
19
 
20
+ This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.0006
23
+ - Token Accuracy: 0.9997
24
+ - Token Precision: 0.9997
25
+ - Token Recall: 0.9997
26
+ - Token F1: 0.9997
27
+ - Entity Precision: 0.9980
28
+ - Entity Recall: 0.9996
29
+ - Entity F1: 0.9987
30
 
31
  ## Model description
32
 
 
49
  - train_batch_size: 16
50
  - eval_batch_size: 16
51
  - seed: 42
52
+ - gradient_accumulation_steps: 2
53
+ - total_train_batch_size: 32
54
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
55
  - lr_scheduler_type: linear
56
  - lr_scheduler_warmup_steps: 500
 
60
 
61
  | Training Loss | Epoch | Step | Validation Loss | Token Accuracy | Token Precision | Token Recall | Token F1 | Entity Precision | Entity Recall | Entity F1 |
62
  |:-------------:|:-----:|:----:|:---------------:|:--------------:|:---------------:|:------------:|:--------:|:----------------:|:-------------:|:---------:|
63
+ | 0.0012 | 1.0 | 1125 | 0.0012 | 0.9995 | 0.9995 | 0.9995 | 0.9995 | 0.9967 | 0.9996 | 0.9981 |
64
+ | 0.0009 | 2.0 | 2250 | 0.0007 | 0.9996 | 0.9996 | 0.9996 | 0.9996 | 0.9973 | 0.9992 | 0.9982 |
65
+ | 0.0005 | 3.0 | 3375 | 0.0006 | 0.9997 | 0.9997 | 0.9997 | 0.9997 | 0.9980 | 0.9996 | 0.9987 |
66
 
67
 
68
  ### Framework versions
69
 
70
+ - Transformers 4.51.3
71
  - Pytorch 2.6.0+cu124
72
+ - Datasets 3.6.0
73
  - Tokenizers 0.21.1
config.json CHANGED
@@ -1,68 +1,64 @@
1
  {
2
- "_num_labels": 9,
3
  "architectures": [
4
- "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 768,
11
  "id2label": {
12
- "0": "B-LOC",
13
- "1": "B-MISC",
14
- "2": "B-ORG",
15
- "3": "B-PER",
16
  "4": "I-LOC",
17
- "5": "I-MISC",
18
  "6": "I-ORG",
19
- "7": "I-PER",
20
- "8": "O",
21
- "9": "B-NUM",
22
- "10": "I-NUM",
23
- "11": "B-CRE",
24
- "12": "I-CRE",
25
- "13": "B-SSN",
26
- "14": "I-SSN",
27
- "15": "B-URL",
28
- "16": "I-URL",
29
- "17": "B-MAIL",
30
- "18": "I-MAIL"
31
  },
32
  "initializer_range": 0.02,
33
  "intermediate_size": 3072,
34
  "label2id": {
35
- "B-CRE": 11,
36
- "B-LOC": 0,
37
- "B-MAIL": 17,
38
- "B-MISC": 1,
39
- "B-NUM": 9,
40
- "B-ORG": 2,
41
- "B-PER": 3,
42
- "B-SSN": 13,
43
- "B-URL": 15,
44
- "I-CRE": 12,
45
  "I-LOC": 4,
46
- "I-MAIL": 18,
47
- "I-MISC": 5,
48
- "I-NUM": 10,
49
  "I-ORG": 6,
50
- "I-PER": 7,
51
- "I-SSN": 14,
52
- "I-URL": 16,
53
- "O": 8
54
  },
55
- "layer_norm_eps": 1e-12,
56
- "max_position_embeddings": 512,
57
- "model_type": "bert",
58
  "num_attention_heads": 12,
59
  "num_hidden_layers": 12,
60
- "output_past": true,
61
- "pad_token_id": 0,
62
  "position_embedding_type": "absolute",
63
  "torch_dtype": "float32",
64
- "transformers_version": "4.50.2",
65
- "type_vocab_size": 2,
66
  "use_cache": true,
67
- "vocab_size": 28996
68
  }
 
1
  {
 
2
  "architectures": [
3
+ "RobertaForTokenClassification"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
  "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
12
  "id2label": {
13
+ "0": "O",
14
+ "1": "B-PER",
15
+ "2": "I-PER",
16
+ "3": "B-LOC",
17
  "4": "I-LOC",
18
+ "5": "B-ORG",
19
  "6": "I-ORG",
20
+ "7": "B-NUM",
21
+ "8": "I-NUM",
22
+ "9": "B-CRE",
23
+ "10": "I-CRE",
24
+ "11": "B-SSN",
25
+ "12": "I-SSN",
26
+ "13": "B-URL",
27
+ "14": "I-URL",
28
+ "15": "B-MAIL",
29
+ "16": "I-MAIL"
 
 
30
  },
31
  "initializer_range": 0.02,
32
  "intermediate_size": 3072,
33
  "label2id": {
34
+ "B-CRE": 9,
35
+ "B-LOC": 3,
36
+ "B-MAIL": 15,
37
+ "B-NUM": 7,
38
+ "B-ORG": 5,
39
+ "B-PER": 1,
40
+ "B-SSN": 11,
41
+ "B-URL": 13,
42
+ "I-CRE": 10,
 
43
  "I-LOC": 4,
44
+ "I-MAIL": 16,
45
+ "I-NUM": 8,
 
46
  "I-ORG": 6,
47
+ "I-PER": 2,
48
+ "I-SSN": 12,
49
+ "I-URL": 14,
50
+ "O": 0
51
  },
52
+ "layer_norm_eps": 1e-05,
53
+ "max_position_embeddings": 514,
54
+ "model_type": "roberta",
55
  "num_attention_heads": 12,
56
  "num_hidden_layers": 12,
57
+ "pad_token_id": 1,
 
58
  "position_embedding_type": "absolute",
59
  "torch_dtype": "float32",
60
+ "transformers_version": "4.51.3",
61
+ "type_vocab_size": 1,
62
  "use_cache": true,
63
+ "vocab_size": 50265
64
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19967ae22cabfaf211f6baa0e2b2940fe4e95b211ffa3ff10d900a82cd894af2
3
- size 430960500
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:099ebe86b563ff429a6784691be6dd3c2c75982336b897f3a1cda1e0b7b3b436
3
+ size 496296380
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a5cced97ccfabbf4d9c5ccdba099b7c5ca924345ca66d0dc54c019fbacf4eab
3
- size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a82e9f58b78bc9332093e49aaa5be2d79b2299197230bf94b249715d0eaa362f
3
+ size 5240