pavm595 commited on
Commit
c96cb7c
·
verified ·
1 Parent(s): a5d37d0
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForTokenClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.0,
6
+ "gradient_checkpointing": false,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.0,
9
+ "hidden_size": 1024,
10
+ "id2label": {
11
+ "0": "E",
12
+ "1": "C",
13
+ "2": "H"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 4096,
17
+ "label2id": {
18
+ "C": 1,
19
+ "E": 0,
20
+ "H": 2
21
+ },
22
+ "layer_norm_eps": 1e-12,
23
+ "max_position_embeddings": 40000,
24
+ "model_type": "bert",
25
+ "num_attention_heads": 16,
26
+ "num_hidden_layers": 30,
27
+ "pad_token_id": 0,
28
+ "total_flos": 82397243632287744,
29
+ "type_vocab_size": 2,
30
+ "vocab_size": 30
31
+ }
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3e64dcaa6f841a603d59691916d4dfa9670346baf49b208485d4bf82338956e
3
+ size 1675555828
log_history.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "loss": 0.7260282135009766,
4
+ "learning_rate": 1.6000000000000003e-05,
5
+ "epoch": 0.5930318754633062,
6
+ "total_flos": 16471213119897600,
7
+ "step": 200
8
+ },
9
+ {
10
+ "eval_loss": 0.449299120773841,
11
+ "eval_accuracy_score": 0.8223555888972243,
12
+ "eval_precision": 0.667388251813714,
13
+ "eval_recall": 0.5915419918583245,
14
+ "eval_f1": 0.6271803824002419,
15
+ "epoch": 0.9992587101556709,
16
+ "total_flos": 27774583123427328,
17
+ "step": 337
18
+ },
19
+ {
20
+ "loss": 0.42990890502929685,
21
+ "learning_rate": 3.2000000000000005e-05,
22
+ "epoch": 1.1868050407709414,
23
+ "total_flos": 32963015256195072,
24
+ "step": 400
25
+ },
26
+ {
27
+ "loss": 0.3798880767822266,
28
+ "learning_rate": 3.2172211350293545e-05,
29
+ "epoch": 1.7798369162342476,
30
+ "total_flos": 49434228376092672,
31
+ "step": 600
32
+ },
33
+ {
34
+ "eval_loss": 0.4176143439192521,
35
+ "eval_accuracy_score": 0.8340029451807396,
36
+ "eval_precision": 0.682940284098745,
37
+ "eval_recall": 0.6419996369953588,
38
+ "eval_f1": 0.6618374274945872,
39
+ "epoch": 1.999258710155671,
40
+ "total_flos": 55549166246854656,
41
+ "step": 674
42
+ },
43
+ {
44
+ "loss": 0.3126878356933594,
45
+ "learning_rate": 1.6516634050880625e-05,
46
+ "epoch": 2.373610081541883,
47
+ "total_flos": 65926030512390144,
48
+ "step": 800
49
+ },
50
+ {
51
+ "loss": 0.2747450256347656,
52
+ "learning_rate": 8.610567514677104e-07,
53
+ "epoch": 2.966641957005189,
54
+ "total_flos": 82397243632287744,
55
+ "step": 1000
56
+ },
57
+ {
58
+ "eval_loss": 0.4275026046454722,
59
+ "eval_accuracy_score": 0.8345252979911645,
60
+ "eval_precision": 0.6880274558343648,
61
+ "eval_recall": 0.6341691083050276,
62
+ "eval_f1": 0.6600013492545368,
63
+ "epoch": 2.999258710155671,
64
+ "total_flos": 83303160353882112,
65
+ "step": 1011
66
+ }
67
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9da6bb7af2ee9504d6ec0d4357797f96b016ca0a79f83304d048949bdacb3a3
3
+ size 1676061559
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "special_tokens_map_file": null, "full_tokenizer_file": null}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1147ba114819d716d2c7a0b6e07e6197c88d505d947f04f5f412af80c145bf2
3
+ size 1775
vocab.txt ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [PAD]
2
+ [UNK]
3
+ [CLS]
4
+ [SEP]
5
+ [MASK]
6
+ L
7
+ A
8
+ G
9
+ V
10
+ E
11
+ S
12
+ I
13
+ K
14
+ R
15
+ D
16
+ T
17
+ P
18
+ N
19
+ Q
20
+ F
21
+ Y
22
+ M
23
+ H
24
+ C
25
+ W
26
+ X
27
+ U
28
+ B
29
+ Z
30
+ O