treamyracle commited on
Commit
58ac8b5
·
verified ·
1 Parent(s): e8c88ab

indo-ner-large-stage2-gold-final

Browse files
Files changed (4) hide show
  1. README.md +9 -9
  2. config.json +5 -5
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -18,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model was trained from scratch on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.1903
22
- - F1: 0.8202
23
- - Precision: 0.7998
24
- - Recall: 0.8417
25
 
26
  ## Model description
27
 
@@ -53,11 +53,11 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | F1 | Precision | Recall |
55
  |:-------------:|:-----:|:----:|:---------------:|:------:|:---------:|:------:|
56
- | 0.3264 | 1.0 | 926 | 0.2126 | 0.7851 | 0.7626 | 0.8091 |
57
- | 0.1823 | 2.0 | 1852 | 0.1930 | 0.8056 | 0.7806 | 0.8322 |
58
- | 0.1444 | 3.0 | 2778 | 0.1894 | 0.8155 | 0.7939 | 0.8382 |
59
- | 0.1293 | 4.0 | 3704 | 0.1885 | 0.8192 | 0.7993 | 0.8400 |
60
- | 0.1184 | 5.0 | 4630 | 0.1903 | 0.8202 | 0.7998 | 0.8417 |
61
 
62
 
63
  ### Framework versions
 
18
 
19
  This model was trained from scratch on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.1738
22
+ - F1: 0.8448
23
+ - Precision: 0.8291
24
+ - Recall: 0.8610
25
 
26
  ## Model description
27
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | F1 | Precision | Recall |
55
  |:-------------:|:-----:|:----:|:---------------:|:------:|:---------:|:------:|
56
+ | 0.311 | 1.0 | 926 | 0.1891 | 0.8127 | 0.7947 | 0.8316 |
57
+ | 0.1567 | 2.0 | 1852 | 0.1723 | 0.8314 | 0.8122 | 0.8516 |
58
+ | 0.1221 | 3.0 | 2778 | 0.1721 | 0.8389 | 0.8214 | 0.8573 |
59
+ | 0.1088 | 4.0 | 3704 | 0.1723 | 0.8430 | 0.8289 | 0.8577 |
60
+ | 0.0986 | 5.0 | 4630 | 0.1738 | 0.8448 | 0.8291 | 0.8610 |
61
 
62
 
63
  ### Framework versions
config.json CHANGED
@@ -9,7 +9,7 @@
9
  "dtype": "float32",
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
  "id2label": {
14
  "0": "O",
15
  "1": "B-CRD",
@@ -52,7 +52,7 @@
52
  "38": "I-WOA"
53
  },
54
  "initializer_range": 0.02,
55
- "intermediate_size": 3072,
56
  "label2id": {
57
  "B-CRD": 1,
58
  "B-DAT": 2,
@@ -97,8 +97,8 @@
97
  "layer_norm_eps": 1e-12,
98
  "max_position_embeddings": 512,
99
  "model_type": "bert",
100
- "num_attention_heads": 12,
101
- "num_hidden_layers": 12,
102
  "output_past": true,
103
  "pad_token_id": 0,
104
  "pooler_fc_size": 768,
@@ -110,5 +110,5 @@
110
  "transformers_version": "4.57.3",
111
  "type_vocab_size": 2,
112
  "use_cache": true,
113
- "vocab_size": 50000
114
  }
 
9
  "dtype": "float32",
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
15
  "1": "B-CRD",
 
52
  "38": "I-WOA"
53
  },
54
  "initializer_range": 0.02,
55
+ "intermediate_size": 4096,
56
  "label2id": {
57
  "B-CRD": 1,
58
  "B-DAT": 2,
 
97
  "layer_norm_eps": 1e-12,
98
  "max_position_embeddings": 512,
99
  "model_type": "bert",
100
+ "num_attention_heads": 16,
101
+ "num_hidden_layers": 24,
102
  "output_past": true,
103
  "pad_token_id": 0,
104
  "pooler_fc_size": 768,
 
110
  "transformers_version": "4.57.3",
111
  "type_vocab_size": 2,
112
  "use_cache": true,
113
+ "vocab_size": 30522
114
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f41ff1ea2cf913676675fd9d9ba03830a51e6bc5b0821c850177e948eba1057c
3
- size 495546332
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f259dba0203e157df4ac391b358198f5326475476f2a22d1cb2b8d5763dd45a
3
+ size 1336575868
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ced4e35556928af2ab28a610c4ae6c9ecd0c158972549028b9ff09f99eee0d7a
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a94b2e593d6e7cdb40c7a2898feb5656e0fdd035d4427579bfe70bf66f23b103
3
  size 5841