prakharsinghAI commited on
Commit
f94195b
·
verified ·
1 Parent(s): 760ad1b

Upload folder using huggingface_hub

Browse files
Files changed (46) hide show
  1. checkpoint-1179/config.json +66 -0
  2. checkpoint-1179/model.safetensors +3 -0
  3. checkpoint-1179/optimizer.pt +3 -0
  4. checkpoint-1179/rng_state.pth +3 -0
  5. checkpoint-1179/scaler.pt +3 -0
  6. checkpoint-1179/scheduler.pt +3 -0
  7. checkpoint-1179/special_tokens_map.json +7 -0
  8. checkpoint-1179/tokenizer.json +0 -0
  9. checkpoint-1179/tokenizer_config.json +58 -0
  10. checkpoint-1179/trainer_state.json +123 -0
  11. checkpoint-1179/training_args.bin +3 -0
  12. checkpoint-1179/vocab.txt +0 -0
  13. checkpoint-4716/config.json +66 -0
  14. checkpoint-4716/model.safetensors +3 -0
  15. checkpoint-4716/optimizer.pt +3 -0
  16. checkpoint-4716/rng_state.pth +3 -0
  17. checkpoint-4716/scaler.pt +3 -0
  18. checkpoint-4716/scheduler.pt +3 -0
  19. checkpoint-4716/special_tokens_map.json +7 -0
  20. checkpoint-4716/tokenizer.json +0 -0
  21. checkpoint-4716/tokenizer_config.json +58 -0
  22. checkpoint-4716/trainer_state.json +411 -0
  23. checkpoint-4716/training_args.bin +3 -0
  24. checkpoint-4716/vocab.txt +0 -0
  25. checkpoint-5895/config.json +66 -0
  26. checkpoint-5895/model.safetensors +3 -0
  27. checkpoint-5895/optimizer.pt +3 -0
  28. checkpoint-5895/rng_state.pth +3 -0
  29. checkpoint-5895/scaler.pt +3 -0
  30. checkpoint-5895/scheduler.pt +3 -0
  31. checkpoint-5895/special_tokens_map.json +7 -0
  32. checkpoint-5895/tokenizer.json +0 -0
  33. checkpoint-5895/tokenizer_config.json +58 -0
  34. checkpoint-5895/trainer_state.json +500 -0
  35. checkpoint-5895/training_args.bin +3 -0
  36. checkpoint-5895/vocab.txt +0 -0
  37. config.json +66 -0
  38. evaluation_results.txt +10 -0
  39. model.safetensors +3 -0
  40. runs/Oct03_23-16-17_3cb83e6cc4fa/events.out.tfevents.1759533380.3cb83e6cc4fa.306.0 +3 -0
  41. runs/Oct03_23-16-17_3cb83e6cc4fa/events.out.tfevents.1759534206.3cb83e6cc4fa.306.1 +3 -0
  42. special_tokens_map.json +7 -0
  43. tokenizer.json +0 -0
  44. tokenizer_config.json +58 -0
  45. training_args.bin +3 -0
  46. vocab.txt +0 -0
checkpoint-1179/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ElectraForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "embedding_size": 128,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 256,
12
+ "id2label": {
13
+ "0": "Dosage-Drug-Positive",
14
+ "1": "Frequency-Drug-Positive",
15
+ "2": "Route-Drug-Negative",
16
+ "3": "Form-Drug-Negative",
17
+ "4": "Strength-Drug-Positive",
18
+ "5": "ADE-Drug-Positive",
19
+ "6": "Frequency-Drug-Negative",
20
+ "7": "Reason-Drug-Negative",
21
+ "8": "ADE-Drug-Negative",
22
+ "9": "Reason-Drug-Positive",
23
+ "10": "Strength-Drug-Negative",
24
+ "11": "Dosage-Drug-Negative",
25
+ "12": "Form-Drug-Positive",
26
+ "13": "Route-Drug-Positive",
27
+ "14": "Duration-Drug-Negative",
28
+ "15": "Duration-Drug-Positive"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 1024,
32
+ "label2id": {
33
+ "ADE-Drug-Negative": 8,
34
+ "ADE-Drug-Positive": 5,
35
+ "Dosage-Drug-Negative": 11,
36
+ "Dosage-Drug-Positive": 0,
37
+ "Duration-Drug-Negative": 14,
38
+ "Duration-Drug-Positive": 15,
39
+ "Form-Drug-Negative": 3,
40
+ "Form-Drug-Positive": 12,
41
+ "Frequency-Drug-Negative": 6,
42
+ "Frequency-Drug-Positive": 1,
43
+ "Reason-Drug-Negative": 7,
44
+ "Reason-Drug-Positive": 9,
45
+ "Route-Drug-Negative": 2,
46
+ "Route-Drug-Positive": 13,
47
+ "Strength-Drug-Negative": 10,
48
+ "Strength-Drug-Positive": 4
49
+ },
50
+ "layer_norm_eps": 1e-12,
51
+ "max_position_embeddings": 512,
52
+ "model_type": "electra",
53
+ "num_attention_heads": 4,
54
+ "num_hidden_layers": 12,
55
+ "pad_token_id": 0,
56
+ "position_embedding_type": "absolute",
57
+ "problem_type": "single_label_classification",
58
+ "summary_activation": "gelu",
59
+ "summary_last_dropout": 0.1,
60
+ "summary_type": "first",
61
+ "summary_use_proj": true,
62
+ "transformers_version": "4.56.2",
63
+ "type_vocab_size": 2,
64
+ "use_cache": true,
65
+ "vocab_size": 30522
66
+ }
checkpoint-1179/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:744b8be1ffac1ead1113028fdfb911e1917098a8aa00a38862b30fe4c2694984
3
+ size 54235600
checkpoint-1179/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:712d7e33a0ca124985c4fd42fca20a74d5203a6f5975561ea2b7f8e1d1b3f700
3
+ size 108596363
checkpoint-1179/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4794e719a734ee9fa9dd15247f0fe5341fcf7ad5816c7231e7f4f7dfef734cab
3
+ size 14645
checkpoint-1179/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2288c6ae9933e2902cb2f6e08f274a26ae19dd2df69adaa489604773cf6bde40
3
+ size 1383
checkpoint-1179/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:389007b36997ea0cd8b51e60e18f187275e138c78544c0319122aedc7ab0559c
3
+ size 1465
checkpoint-1179/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
checkpoint-1179/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1179/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "101": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "102": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "103": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "ElectraTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
checkpoint-1179/trainer_state.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1179,
3
+ "best_metric": 0.32642487046632124,
4
+ "best_model_checkpoint": "./i2b2_model/checkpoint-1179",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1179,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08481764206955046,
14
+ "grad_norm": 0.8805635571479797,
15
+ "learning_rate": 1.966412213740458e-05,
16
+ "loss": 2.7221,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.16963528413910092,
21
+ "grad_norm": 1.0300383567810059,
22
+ "learning_rate": 1.932485156912638e-05,
23
+ "loss": 2.5753,
24
+ "step": 200
25
+ },
26
+ {
27
+ "epoch": 0.2544529262086514,
28
+ "grad_norm": 1.5689030885696411,
29
+ "learning_rate": 1.898558100084818e-05,
30
+ "loss": 2.3339,
31
+ "step": 300
32
+ },
33
+ {
34
+ "epoch": 0.33927056827820185,
35
+ "grad_norm": 1.3848917484283447,
36
+ "learning_rate": 1.8646310432569977e-05,
37
+ "loss": 2.0873,
38
+ "step": 400
39
+ },
40
+ {
41
+ "epoch": 0.42408821034775235,
42
+ "grad_norm": 1.6002815961837769,
43
+ "learning_rate": 1.8307039864291775e-05,
44
+ "loss": 1.8578,
45
+ "step": 500
46
+ },
47
+ {
48
+ "epoch": 0.5089058524173028,
49
+ "grad_norm": 1.4675610065460205,
50
+ "learning_rate": 1.7967769296013573e-05,
51
+ "loss": 1.639,
52
+ "step": 600
53
+ },
54
+ {
55
+ "epoch": 0.5937234944868532,
56
+ "grad_norm": 1.510614275932312,
57
+ "learning_rate": 1.762849872773537e-05,
58
+ "loss": 1.4393,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.6785411365564037,
63
+ "grad_norm": 1.4330428838729858,
64
+ "learning_rate": 1.7289228159457168e-05,
65
+ "loss": 1.2814,
66
+ "step": 800
67
+ },
68
+ {
69
+ "epoch": 0.7633587786259542,
70
+ "grad_norm": 1.3979631662368774,
71
+ "learning_rate": 1.6949957591178966e-05,
72
+ "loss": 1.1749,
73
+ "step": 900
74
+ },
75
+ {
76
+ "epoch": 0.8481764206955047,
77
+ "grad_norm": 1.3620542287826538,
78
+ "learning_rate": 1.6610687022900763e-05,
79
+ "loss": 1.1002,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 0.9329940627650551,
84
+ "grad_norm": 1.8766813278198242,
85
+ "learning_rate": 1.6271416454622565e-05,
86
+ "loss": 1.0137,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 1.0,
91
+ "eval_accuracy": 0.32642487046632124,
92
+ "eval_f1": 0.23427617946411824,
93
+ "eval_loss": 2.2931675910949707,
94
+ "eval_precision": 0.20496833204797127,
95
+ "eval_recall": 0.32642487046632124,
96
+ "eval_runtime": 20.873,
97
+ "eval_samples_per_second": 1202.032,
98
+ "eval_steps_per_second": 18.828,
99
+ "step": 1179
100
+ }
101
+ ],
102
+ "logging_steps": 100,
103
+ "max_steps": 5895,
104
+ "num_input_tokens_seen": 0,
105
+ "num_train_epochs": 5,
106
+ "save_steps": 500,
107
+ "stateful_callbacks": {
108
+ "TrainerControl": {
109
+ "args": {
110
+ "should_epoch_stop": false,
111
+ "should_evaluate": false,
112
+ "should_log": false,
113
+ "should_save": true,
114
+ "should_training_stop": false
115
+ },
116
+ "attributes": {}
117
+ }
118
+ },
119
+ "total_flos": 428825443550208.0,
120
+ "train_batch_size": 32,
121
+ "trial_name": null,
122
+ "trial_params": null
123
+ }
checkpoint-1179/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41ecbeafe8ad87a86caca10ba72016e58071d9d73cdd450f90210669b5a66947
3
+ size 5777
checkpoint-1179/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-4716/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ElectraForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "embedding_size": 128,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 256,
12
+ "id2label": {
13
+ "0": "Dosage-Drug-Positive",
14
+ "1": "Frequency-Drug-Positive",
15
+ "2": "Route-Drug-Negative",
16
+ "3": "Form-Drug-Negative",
17
+ "4": "Strength-Drug-Positive",
18
+ "5": "ADE-Drug-Positive",
19
+ "6": "Frequency-Drug-Negative",
20
+ "7": "Reason-Drug-Negative",
21
+ "8": "ADE-Drug-Negative",
22
+ "9": "Reason-Drug-Positive",
23
+ "10": "Strength-Drug-Negative",
24
+ "11": "Dosage-Drug-Negative",
25
+ "12": "Form-Drug-Positive",
26
+ "13": "Route-Drug-Positive",
27
+ "14": "Duration-Drug-Negative",
28
+ "15": "Duration-Drug-Positive"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 1024,
32
+ "label2id": {
33
+ "ADE-Drug-Negative": 8,
34
+ "ADE-Drug-Positive": 5,
35
+ "Dosage-Drug-Negative": 11,
36
+ "Dosage-Drug-Positive": 0,
37
+ "Duration-Drug-Negative": 14,
38
+ "Duration-Drug-Positive": 15,
39
+ "Form-Drug-Negative": 3,
40
+ "Form-Drug-Positive": 12,
41
+ "Frequency-Drug-Negative": 6,
42
+ "Frequency-Drug-Positive": 1,
43
+ "Reason-Drug-Negative": 7,
44
+ "Reason-Drug-Positive": 9,
45
+ "Route-Drug-Negative": 2,
46
+ "Route-Drug-Positive": 13,
47
+ "Strength-Drug-Negative": 10,
48
+ "Strength-Drug-Positive": 4
49
+ },
50
+ "layer_norm_eps": 1e-12,
51
+ "max_position_embeddings": 512,
52
+ "model_type": "electra",
53
+ "num_attention_heads": 4,
54
+ "num_hidden_layers": 12,
55
+ "pad_token_id": 0,
56
+ "position_embedding_type": "absolute",
57
+ "problem_type": "single_label_classification",
58
+ "summary_activation": "gelu",
59
+ "summary_last_dropout": 0.1,
60
+ "summary_type": "first",
61
+ "summary_use_proj": true,
62
+ "transformers_version": "4.56.2",
63
+ "type_vocab_size": 2,
64
+ "use_cache": true,
65
+ "vocab_size": 30522
66
+ }
checkpoint-4716/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6791bb58895bef3f61441d83d8686372cd7f09d22e5817d4db3ef0cbe3c5db1a
3
+ size 54235600
checkpoint-4716/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e5fe72b28557e89d2be6598de01392afbae6432762f4dbb16709e7bc23e7fb
3
+ size 108596363
checkpoint-4716/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dbbf92b0ac496935212e85c6e9a267d3a5afe9934afe4ec3eb7b4eab2f7d572
3
+ size 14645
checkpoint-4716/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:075b2d9467021f1a0971c4a60861645c6b0b2851fad78dcaf0bf9e2039c897da
3
+ size 1383
checkpoint-4716/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32f8d2458d810faad3ef4c26d3eb3bddc895c77f2be3c493945b6696a36a62cd
3
+ size 1465
checkpoint-4716/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
checkpoint-4716/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-4716/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "101": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "102": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "103": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "ElectraTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
checkpoint-4716/trainer_state.json ADDED
@@ -0,0 +1,411 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1179,
3
+ "best_metric": 0.32642487046632124,
4
+ "best_model_checkpoint": "./i2b2_model/checkpoint-1179",
5
+ "epoch": 4.0,
6
+ "eval_steps": 500,
7
+ "global_step": 4716,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08481764206955046,
14
+ "grad_norm": 0.8805635571479797,
15
+ "learning_rate": 1.966412213740458e-05,
16
+ "loss": 2.7221,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.16963528413910092,
21
+ "grad_norm": 1.0300383567810059,
22
+ "learning_rate": 1.932485156912638e-05,
23
+ "loss": 2.5753,
24
+ "step": 200
25
+ },
26
+ {
27
+ "epoch": 0.2544529262086514,
28
+ "grad_norm": 1.5689030885696411,
29
+ "learning_rate": 1.898558100084818e-05,
30
+ "loss": 2.3339,
31
+ "step": 300
32
+ },
33
+ {
34
+ "epoch": 0.33927056827820185,
35
+ "grad_norm": 1.3848917484283447,
36
+ "learning_rate": 1.8646310432569977e-05,
37
+ "loss": 2.0873,
38
+ "step": 400
39
+ },
40
+ {
41
+ "epoch": 0.42408821034775235,
42
+ "grad_norm": 1.6002815961837769,
43
+ "learning_rate": 1.8307039864291775e-05,
44
+ "loss": 1.8578,
45
+ "step": 500
46
+ },
47
+ {
48
+ "epoch": 0.5089058524173028,
49
+ "grad_norm": 1.4675610065460205,
50
+ "learning_rate": 1.7967769296013573e-05,
51
+ "loss": 1.639,
52
+ "step": 600
53
+ },
54
+ {
55
+ "epoch": 0.5937234944868532,
56
+ "grad_norm": 1.510614275932312,
57
+ "learning_rate": 1.762849872773537e-05,
58
+ "loss": 1.4393,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.6785411365564037,
63
+ "grad_norm": 1.4330428838729858,
64
+ "learning_rate": 1.7289228159457168e-05,
65
+ "loss": 1.2814,
66
+ "step": 800
67
+ },
68
+ {
69
+ "epoch": 0.7633587786259542,
70
+ "grad_norm": 1.3979631662368774,
71
+ "learning_rate": 1.6949957591178966e-05,
72
+ "loss": 1.1749,
73
+ "step": 900
74
+ },
75
+ {
76
+ "epoch": 0.8481764206955047,
77
+ "grad_norm": 1.3620542287826538,
78
+ "learning_rate": 1.6610687022900763e-05,
79
+ "loss": 1.1002,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 0.9329940627650551,
84
+ "grad_norm": 1.8766813278198242,
85
+ "learning_rate": 1.6271416454622565e-05,
86
+ "loss": 1.0137,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 1.0,
91
+ "eval_accuracy": 0.32642487046632124,
92
+ "eval_f1": 0.23427617946411824,
93
+ "eval_loss": 2.2931675910949707,
94
+ "eval_precision": 0.20496833204797127,
95
+ "eval_recall": 0.32642487046632124,
96
+ "eval_runtime": 20.873,
97
+ "eval_samples_per_second": 1202.032,
98
+ "eval_steps_per_second": 18.828,
99
+ "step": 1179
100
+ },
101
+ {
102
+ "epoch": 1.0178117048346056,
103
+ "grad_norm": 2.0241150856018066,
104
+ "learning_rate": 1.5932145886344362e-05,
105
+ "loss": 0.9752,
106
+ "step": 1200
107
+ },
108
+ {
109
+ "epoch": 1.1026293469041561,
110
+ "grad_norm": 1.0047012567520142,
111
+ "learning_rate": 1.559287531806616e-05,
112
+ "loss": 0.9322,
113
+ "step": 1300
114
+ },
115
+ {
116
+ "epoch": 1.1874469889737065,
117
+ "grad_norm": 2.013701915740967,
118
+ "learning_rate": 1.5253604749787958e-05,
119
+ "loss": 0.9217,
120
+ "step": 1400
121
+ },
122
+ {
123
+ "epoch": 1.272264631043257,
124
+ "grad_norm": 1.9003664255142212,
125
+ "learning_rate": 1.4914334181509755e-05,
126
+ "loss": 0.8991,
127
+ "step": 1500
128
+ },
129
+ {
130
+ "epoch": 1.3570822731128076,
131
+ "grad_norm": 1.2886754274368286,
132
+ "learning_rate": 1.4575063613231553e-05,
133
+ "loss": 0.882,
134
+ "step": 1600
135
+ },
136
+ {
137
+ "epoch": 1.441899915182358,
138
+ "grad_norm": 1.268678903579712,
139
+ "learning_rate": 1.4235793044953352e-05,
140
+ "loss": 0.8684,
141
+ "step": 1700
142
+ },
143
+ {
144
+ "epoch": 1.5267175572519083,
145
+ "grad_norm": 1.3230983018875122,
146
+ "learning_rate": 1.389652247667515e-05,
147
+ "loss": 0.8467,
148
+ "step": 1800
149
+ },
150
+ {
151
+ "epoch": 1.6115351993214588,
152
+ "grad_norm": 1.3858022689819336,
153
+ "learning_rate": 1.3557251908396946e-05,
154
+ "loss": 0.8422,
155
+ "step": 1900
156
+ },
157
+ {
158
+ "epoch": 1.6963528413910094,
159
+ "grad_norm": 2.531125545501709,
160
+ "learning_rate": 1.3217981340118746e-05,
161
+ "loss": 0.8266,
162
+ "step": 2000
163
+ },
164
+ {
165
+ "epoch": 1.78117048346056,
166
+ "grad_norm": 2.599006175994873,
167
+ "learning_rate": 1.2878710771840543e-05,
168
+ "loss": 0.8215,
169
+ "step": 2100
170
+ },
171
+ {
172
+ "epoch": 1.8659881255301103,
173
+ "grad_norm": 2.465607166290283,
174
+ "learning_rate": 1.2539440203562343e-05,
175
+ "loss": 0.8077,
176
+ "step": 2200
177
+ },
178
+ {
179
+ "epoch": 1.9508057675996606,
180
+ "grad_norm": 1.6329939365386963,
181
+ "learning_rate": 1.220016963528414e-05,
182
+ "loss": 0.8002,
183
+ "step": 2300
184
+ },
185
+ {
186
+ "epoch": 2.0,
187
+ "eval_accuracy": 0.2957353527301714,
188
+ "eval_f1": 0.19723504159983604,
189
+ "eval_loss": 2.6935245990753174,
190
+ "eval_precision": 0.15463079618041953,
191
+ "eval_recall": 0.2957353527301714,
192
+ "eval_runtime": 19.7862,
193
+ "eval_samples_per_second": 1268.057,
194
+ "eval_steps_per_second": 19.862,
195
+ "step": 2358
196
+ },
197
+ {
198
+ "epoch": 2.035623409669211,
199
+ "grad_norm": 3.1458687782287598,
200
+ "learning_rate": 1.1860899067005938e-05,
201
+ "loss": 0.7967,
202
+ "step": 2400
203
+ },
204
+ {
205
+ "epoch": 2.1204410517387617,
206
+ "grad_norm": 1.5369526147842407,
207
+ "learning_rate": 1.1521628498727737e-05,
208
+ "loss": 0.7927,
209
+ "step": 2500
210
+ },
211
+ {
212
+ "epoch": 2.2052586938083123,
213
+ "grad_norm": 1.882670521736145,
214
+ "learning_rate": 1.1182357930449535e-05,
215
+ "loss": 0.7849,
216
+ "step": 2600
217
+ },
218
+ {
219
+ "epoch": 2.2900763358778624,
220
+ "grad_norm": 2.031019449234009,
221
+ "learning_rate": 1.0843087362171331e-05,
222
+ "loss": 0.7786,
223
+ "step": 2700
224
+ },
225
+ {
226
+ "epoch": 2.374893977947413,
227
+ "grad_norm": 2.266068458557129,
228
+ "learning_rate": 1.050381679389313e-05,
229
+ "loss": 0.7763,
230
+ "step": 2800
231
+ },
232
+ {
233
+ "epoch": 2.4597116200169635,
234
+ "grad_norm": 1.4232118129730225,
235
+ "learning_rate": 1.0164546225614928e-05,
236
+ "loss": 0.7736,
237
+ "step": 2900
238
+ },
239
+ {
240
+ "epoch": 2.544529262086514,
241
+ "grad_norm": 2.115183115005493,
242
+ "learning_rate": 9.825275657336728e-06,
243
+ "loss": 0.7593,
244
+ "step": 3000
245
+ },
246
+ {
247
+ "epoch": 2.6293469041560646,
248
+ "grad_norm": 1.7226656675338745,
249
+ "learning_rate": 9.486005089058525e-06,
250
+ "loss": 0.7587,
251
+ "step": 3100
252
+ },
253
+ {
254
+ "epoch": 2.714164546225615,
255
+ "grad_norm": 2.0967631340026855,
256
+ "learning_rate": 9.146734520780323e-06,
257
+ "loss": 0.7665,
258
+ "step": 3200
259
+ },
260
+ {
261
+ "epoch": 2.7989821882951653,
262
+ "grad_norm": 1.8578695058822632,
263
+ "learning_rate": 8.80746395250212e-06,
264
+ "loss": 0.7527,
265
+ "step": 3300
266
+ },
267
+ {
268
+ "epoch": 2.883799830364716,
269
+ "grad_norm": 2.35467267036438,
270
+ "learning_rate": 8.46819338422392e-06,
271
+ "loss": 0.7516,
272
+ "step": 3400
273
+ },
274
+ {
275
+ "epoch": 2.9686174724342664,
276
+ "grad_norm": 2.1662681102752686,
277
+ "learning_rate": 8.128922815945718e-06,
278
+ "loss": 0.749,
279
+ "step": 3500
280
+ },
281
+ {
282
+ "epoch": 3.0,
283
+ "eval_accuracy": 0.29166998804304506,
284
+ "eval_f1": 0.2111020694739784,
285
+ "eval_loss": 2.924750804901123,
286
+ "eval_precision": 0.16962444026380036,
287
+ "eval_recall": 0.29166998804304506,
288
+ "eval_runtime": 20.8788,
289
+ "eval_samples_per_second": 1201.696,
290
+ "eval_steps_per_second": 18.823,
291
+ "step": 3537
292
+ },
293
+ {
294
+ "epoch": 3.053435114503817,
295
+ "grad_norm": 1.8867818117141724,
296
+ "learning_rate": 7.789652247667516e-06,
297
+ "loss": 0.7507,
298
+ "step": 3600
299
+ },
300
+ {
301
+ "epoch": 3.138252756573367,
302
+ "grad_norm": 2.129338026046753,
303
+ "learning_rate": 7.450381679389313e-06,
304
+ "loss": 0.7462,
305
+ "step": 3700
306
+ },
307
+ {
308
+ "epoch": 3.2230703986429177,
309
+ "grad_norm": 1.5529320240020752,
310
+ "learning_rate": 7.111111111111112e-06,
311
+ "loss": 0.7436,
312
+ "step": 3800
313
+ },
314
+ {
315
+ "epoch": 3.3078880407124682,
316
+ "grad_norm": 2.430915117263794,
317
+ "learning_rate": 6.7718405428329095e-06,
318
+ "loss": 0.7421,
319
+ "step": 3900
320
+ },
321
+ {
322
+ "epoch": 3.392705682782019,
323
+ "grad_norm": 2.0328726768493652,
324
+ "learning_rate": 6.432569974554708e-06,
325
+ "loss": 0.7397,
326
+ "step": 4000
327
+ },
328
+ {
329
+ "epoch": 3.4775233248515693,
330
+ "grad_norm": 1.6856147050857544,
331
+ "learning_rate": 6.093299406276506e-06,
332
+ "loss": 0.7358,
333
+ "step": 4100
334
+ },
335
+ {
336
+ "epoch": 3.5623409669211195,
337
+ "grad_norm": 2.0061874389648438,
338
+ "learning_rate": 5.754028837998304e-06,
339
+ "loss": 0.7355,
340
+ "step": 4200
341
+ },
342
+ {
343
+ "epoch": 3.64715860899067,
344
+ "grad_norm": 2.331062078475952,
345
+ "learning_rate": 5.414758269720102e-06,
346
+ "loss": 0.7433,
347
+ "step": 4300
348
+ },
349
+ {
350
+ "epoch": 3.7319762510602206,
351
+ "grad_norm": 2.5025064945220947,
352
+ "learning_rate": 5.075487701441901e-06,
353
+ "loss": 0.7371,
354
+ "step": 4400
355
+ },
356
+ {
357
+ "epoch": 3.816793893129771,
358
+ "grad_norm": 3.021045446395874,
359
+ "learning_rate": 4.736217133163698e-06,
360
+ "loss": 0.7357,
361
+ "step": 4500
362
+ },
363
+ {
364
+ "epoch": 3.9016115351993212,
365
+ "grad_norm": 1.7884670495986938,
366
+ "learning_rate": 4.396946564885497e-06,
367
+ "loss": 0.7366,
368
+ "step": 4600
369
+ },
370
+ {
371
+ "epoch": 3.986429177268872,
372
+ "grad_norm": 2.477693796157837,
373
+ "learning_rate": 4.0576759966072945e-06,
374
+ "loss": 0.7324,
375
+ "step": 4700
376
+ },
377
+ {
378
+ "epoch": 4.0,
379
+ "eval_accuracy": 0.24192905540055798,
380
+ "eval_f1": 0.18676919220051297,
381
+ "eval_loss": 3.059746265411377,
382
+ "eval_precision": 0.23033226126155365,
383
+ "eval_recall": 0.24192905540055798,
384
+ "eval_runtime": 19.9179,
385
+ "eval_samples_per_second": 1259.67,
386
+ "eval_steps_per_second": 19.731,
387
+ "step": 4716
388
+ }
389
+ ],
390
+ "logging_steps": 100,
391
+ "max_steps": 5895,
392
+ "num_input_tokens_seen": 0,
393
+ "num_train_epochs": 5,
394
+ "save_steps": 500,
395
+ "stateful_callbacks": {
396
+ "TrainerControl": {
397
+ "args": {
398
+ "should_epoch_stop": false,
399
+ "should_evaluate": false,
400
+ "should_log": false,
401
+ "should_save": true,
402
+ "should_training_stop": false
403
+ },
404
+ "attributes": {}
405
+ }
406
+ },
407
+ "total_flos": 1715326721312448.0,
408
+ "train_batch_size": 32,
409
+ "trial_name": null,
410
+ "trial_params": null
411
+ }
checkpoint-4716/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41ecbeafe8ad87a86caca10ba72016e58071d9d73cdd450f90210669b5a66947
3
+ size 5777
checkpoint-4716/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-5895/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ElectraForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "embedding_size": 128,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 256,
12
+ "id2label": {
13
+ "0": "Dosage-Drug-Positive",
14
+ "1": "Frequency-Drug-Positive",
15
+ "2": "Route-Drug-Negative",
16
+ "3": "Form-Drug-Negative",
17
+ "4": "Strength-Drug-Positive",
18
+ "5": "ADE-Drug-Positive",
19
+ "6": "Frequency-Drug-Negative",
20
+ "7": "Reason-Drug-Negative",
21
+ "8": "ADE-Drug-Negative",
22
+ "9": "Reason-Drug-Positive",
23
+ "10": "Strength-Drug-Negative",
24
+ "11": "Dosage-Drug-Negative",
25
+ "12": "Form-Drug-Positive",
26
+ "13": "Route-Drug-Positive",
27
+ "14": "Duration-Drug-Negative",
28
+ "15": "Duration-Drug-Positive"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 1024,
32
+ "label2id": {
33
+ "ADE-Drug-Negative": 8,
34
+ "ADE-Drug-Positive": 5,
35
+ "Dosage-Drug-Negative": 11,
36
+ "Dosage-Drug-Positive": 0,
37
+ "Duration-Drug-Negative": 14,
38
+ "Duration-Drug-Positive": 15,
39
+ "Form-Drug-Negative": 3,
40
+ "Form-Drug-Positive": 12,
41
+ "Frequency-Drug-Negative": 6,
42
+ "Frequency-Drug-Positive": 1,
43
+ "Reason-Drug-Negative": 7,
44
+ "Reason-Drug-Positive": 9,
45
+ "Route-Drug-Negative": 2,
46
+ "Route-Drug-Positive": 13,
47
+ "Strength-Drug-Negative": 10,
48
+ "Strength-Drug-Positive": 4
49
+ },
50
+ "layer_norm_eps": 1e-12,
51
+ "max_position_embeddings": 512,
52
+ "model_type": "electra",
53
+ "num_attention_heads": 4,
54
+ "num_hidden_layers": 12,
55
+ "pad_token_id": 0,
56
+ "position_embedding_type": "absolute",
57
+ "problem_type": "single_label_classification",
58
+ "summary_activation": "gelu",
59
+ "summary_last_dropout": 0.1,
60
+ "summary_type": "first",
61
+ "summary_use_proj": true,
62
+ "transformers_version": "4.56.2",
63
+ "type_vocab_size": 2,
64
+ "use_cache": true,
65
+ "vocab_size": 30522
66
+ }
checkpoint-5895/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eab8e9abb443554d3d36fd4a525b0dcf54de3d6618162391106149136d1e5e5a
3
+ size 54235600
checkpoint-5895/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5695d7ae2326320ad5b22fa67a9abeeb758a8812f3f881f2549b79d6c361d037
3
+ size 108596363
checkpoint-5895/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a50e2a41e52ec56c5d7705318a46fa6314b4bfd960b031214524e17584261aae
3
+ size 14645
checkpoint-5895/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:250dd993fb364cb2c424cee8189f84c61dbb791c6a32fc4c8d8b57d6cb5d28fe
3
+ size 1383
checkpoint-5895/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21983f7b03a724e11a8bfeadce32cd7e2b7afd51fa3a6091f42e25a5fb85cae4
3
+ size 1465
checkpoint-5895/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
checkpoint-5895/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-5895/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "101": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "102": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "103": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "ElectraTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
checkpoint-5895/trainer_state.json ADDED
@@ -0,0 +1,500 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1179,
3
+ "best_metric": 0.32642487046632124,
4
+ "best_model_checkpoint": "./i2b2_model/checkpoint-1179",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 5895,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08481764206955046,
14
+ "grad_norm": 0.8805635571479797,
15
+ "learning_rate": 1.966412213740458e-05,
16
+ "loss": 2.7221,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.16963528413910092,
21
+ "grad_norm": 1.0300383567810059,
22
+ "learning_rate": 1.932485156912638e-05,
23
+ "loss": 2.5753,
24
+ "step": 200
25
+ },
26
+ {
27
+ "epoch": 0.2544529262086514,
28
+ "grad_norm": 1.5689030885696411,
29
+ "learning_rate": 1.898558100084818e-05,
30
+ "loss": 2.3339,
31
+ "step": 300
32
+ },
33
+ {
34
+ "epoch": 0.33927056827820185,
35
+ "grad_norm": 1.3848917484283447,
36
+ "learning_rate": 1.8646310432569977e-05,
37
+ "loss": 2.0873,
38
+ "step": 400
39
+ },
40
+ {
41
+ "epoch": 0.42408821034775235,
42
+ "grad_norm": 1.6002815961837769,
43
+ "learning_rate": 1.8307039864291775e-05,
44
+ "loss": 1.8578,
45
+ "step": 500
46
+ },
47
+ {
48
+ "epoch": 0.5089058524173028,
49
+ "grad_norm": 1.4675610065460205,
50
+ "learning_rate": 1.7967769296013573e-05,
51
+ "loss": 1.639,
52
+ "step": 600
53
+ },
54
+ {
55
+ "epoch": 0.5937234944868532,
56
+ "grad_norm": 1.510614275932312,
57
+ "learning_rate": 1.762849872773537e-05,
58
+ "loss": 1.4393,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.6785411365564037,
63
+ "grad_norm": 1.4330428838729858,
64
+ "learning_rate": 1.7289228159457168e-05,
65
+ "loss": 1.2814,
66
+ "step": 800
67
+ },
68
+ {
69
+ "epoch": 0.7633587786259542,
70
+ "grad_norm": 1.3979631662368774,
71
+ "learning_rate": 1.6949957591178966e-05,
72
+ "loss": 1.1749,
73
+ "step": 900
74
+ },
75
+ {
76
+ "epoch": 0.8481764206955047,
77
+ "grad_norm": 1.3620542287826538,
78
+ "learning_rate": 1.6610687022900763e-05,
79
+ "loss": 1.1002,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 0.9329940627650551,
84
+ "grad_norm": 1.8766813278198242,
85
+ "learning_rate": 1.6271416454622565e-05,
86
+ "loss": 1.0137,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 1.0,
91
+ "eval_accuracy": 0.32642487046632124,
92
+ "eval_f1": 0.23427617946411824,
93
+ "eval_loss": 2.2931675910949707,
94
+ "eval_precision": 0.20496833204797127,
95
+ "eval_recall": 0.32642487046632124,
96
+ "eval_runtime": 20.873,
97
+ "eval_samples_per_second": 1202.032,
98
+ "eval_steps_per_second": 18.828,
99
+ "step": 1179
100
+ },
101
+ {
102
+ "epoch": 1.0178117048346056,
103
+ "grad_norm": 2.0241150856018066,
104
+ "learning_rate": 1.5932145886344362e-05,
105
+ "loss": 0.9752,
106
+ "step": 1200
107
+ },
108
+ {
109
+ "epoch": 1.1026293469041561,
110
+ "grad_norm": 1.0047012567520142,
111
+ "learning_rate": 1.559287531806616e-05,
112
+ "loss": 0.9322,
113
+ "step": 1300
114
+ },
115
+ {
116
+ "epoch": 1.1874469889737065,
117
+ "grad_norm": 2.013701915740967,
118
+ "learning_rate": 1.5253604749787958e-05,
119
+ "loss": 0.9217,
120
+ "step": 1400
121
+ },
122
+ {
123
+ "epoch": 1.272264631043257,
124
+ "grad_norm": 1.9003664255142212,
125
+ "learning_rate": 1.4914334181509755e-05,
126
+ "loss": 0.8991,
127
+ "step": 1500
128
+ },
129
+ {
130
+ "epoch": 1.3570822731128076,
131
+ "grad_norm": 1.2886754274368286,
132
+ "learning_rate": 1.4575063613231553e-05,
133
+ "loss": 0.882,
134
+ "step": 1600
135
+ },
136
+ {
137
+ "epoch": 1.441899915182358,
138
+ "grad_norm": 1.268678903579712,
139
+ "learning_rate": 1.4235793044953352e-05,
140
+ "loss": 0.8684,
141
+ "step": 1700
142
+ },
143
+ {
144
+ "epoch": 1.5267175572519083,
145
+ "grad_norm": 1.3230983018875122,
146
+ "learning_rate": 1.389652247667515e-05,
147
+ "loss": 0.8467,
148
+ "step": 1800
149
+ },
150
+ {
151
+ "epoch": 1.6115351993214588,
152
+ "grad_norm": 1.3858022689819336,
153
+ "learning_rate": 1.3557251908396946e-05,
154
+ "loss": 0.8422,
155
+ "step": 1900
156
+ },
157
+ {
158
+ "epoch": 1.6963528413910094,
159
+ "grad_norm": 2.531125545501709,
160
+ "learning_rate": 1.3217981340118746e-05,
161
+ "loss": 0.8266,
162
+ "step": 2000
163
+ },
164
+ {
165
+ "epoch": 1.78117048346056,
166
+ "grad_norm": 2.599006175994873,
167
+ "learning_rate": 1.2878710771840543e-05,
168
+ "loss": 0.8215,
169
+ "step": 2100
170
+ },
171
+ {
172
+ "epoch": 1.8659881255301103,
173
+ "grad_norm": 2.465607166290283,
174
+ "learning_rate": 1.2539440203562343e-05,
175
+ "loss": 0.8077,
176
+ "step": 2200
177
+ },
178
+ {
179
+ "epoch": 1.9508057675996606,
180
+ "grad_norm": 1.6329939365386963,
181
+ "learning_rate": 1.220016963528414e-05,
182
+ "loss": 0.8002,
183
+ "step": 2300
184
+ },
185
+ {
186
+ "epoch": 2.0,
187
+ "eval_accuracy": 0.2957353527301714,
188
+ "eval_f1": 0.19723504159983604,
189
+ "eval_loss": 2.6935245990753174,
190
+ "eval_precision": 0.15463079618041953,
191
+ "eval_recall": 0.2957353527301714,
192
+ "eval_runtime": 19.7862,
193
+ "eval_samples_per_second": 1268.057,
194
+ "eval_steps_per_second": 19.862,
195
+ "step": 2358
196
+ },
197
+ {
198
+ "epoch": 2.035623409669211,
199
+ "grad_norm": 3.1458687782287598,
200
+ "learning_rate": 1.1860899067005938e-05,
201
+ "loss": 0.7967,
202
+ "step": 2400
203
+ },
204
+ {
205
+ "epoch": 2.1204410517387617,
206
+ "grad_norm": 1.5369526147842407,
207
+ "learning_rate": 1.1521628498727737e-05,
208
+ "loss": 0.7927,
209
+ "step": 2500
210
+ },
211
+ {
212
+ "epoch": 2.2052586938083123,
213
+ "grad_norm": 1.882670521736145,
214
+ "learning_rate": 1.1182357930449535e-05,
215
+ "loss": 0.7849,
216
+ "step": 2600
217
+ },
218
+ {
219
+ "epoch": 2.2900763358778624,
220
+ "grad_norm": 2.031019449234009,
221
+ "learning_rate": 1.0843087362171331e-05,
222
+ "loss": 0.7786,
223
+ "step": 2700
224
+ },
225
+ {
226
+ "epoch": 2.374893977947413,
227
+ "grad_norm": 2.266068458557129,
228
+ "learning_rate": 1.050381679389313e-05,
229
+ "loss": 0.7763,
230
+ "step": 2800
231
+ },
232
+ {
233
+ "epoch": 2.4597116200169635,
234
+ "grad_norm": 1.4232118129730225,
235
+ "learning_rate": 1.0164546225614928e-05,
236
+ "loss": 0.7736,
237
+ "step": 2900
238
+ },
239
+ {
240
+ "epoch": 2.544529262086514,
241
+ "grad_norm": 2.115183115005493,
242
+ "learning_rate": 9.825275657336728e-06,
243
+ "loss": 0.7593,
244
+ "step": 3000
245
+ },
246
+ {
247
+ "epoch": 2.6293469041560646,
248
+ "grad_norm": 1.7226656675338745,
249
+ "learning_rate": 9.486005089058525e-06,
250
+ "loss": 0.7587,
251
+ "step": 3100
252
+ },
253
+ {
254
+ "epoch": 2.714164546225615,
255
+ "grad_norm": 2.0967631340026855,
256
+ "learning_rate": 9.146734520780323e-06,
257
+ "loss": 0.7665,
258
+ "step": 3200
259
+ },
260
+ {
261
+ "epoch": 2.7989821882951653,
262
+ "grad_norm": 1.8578695058822632,
263
+ "learning_rate": 8.80746395250212e-06,
264
+ "loss": 0.7527,
265
+ "step": 3300
266
+ },
267
+ {
268
+ "epoch": 2.883799830364716,
269
+ "grad_norm": 2.35467267036438,
270
+ "learning_rate": 8.46819338422392e-06,
271
+ "loss": 0.7516,
272
+ "step": 3400
273
+ },
274
+ {
275
+ "epoch": 2.9686174724342664,
276
+ "grad_norm": 2.1662681102752686,
277
+ "learning_rate": 8.128922815945718e-06,
278
+ "loss": 0.749,
279
+ "step": 3500
280
+ },
281
+ {
282
+ "epoch": 3.0,
283
+ "eval_accuracy": 0.29166998804304506,
284
+ "eval_f1": 0.2111020694739784,
285
+ "eval_loss": 2.924750804901123,
286
+ "eval_precision": 0.16962444026380036,
287
+ "eval_recall": 0.29166998804304506,
288
+ "eval_runtime": 20.8788,
289
+ "eval_samples_per_second": 1201.696,
290
+ "eval_steps_per_second": 18.823,
291
+ "step": 3537
292
+ },
293
+ {
294
+ "epoch": 3.053435114503817,
295
+ "grad_norm": 1.8867818117141724,
296
+ "learning_rate": 7.789652247667516e-06,
297
+ "loss": 0.7507,
298
+ "step": 3600
299
+ },
300
+ {
301
+ "epoch": 3.138252756573367,
302
+ "grad_norm": 2.129338026046753,
303
+ "learning_rate": 7.450381679389313e-06,
304
+ "loss": 0.7462,
305
+ "step": 3700
306
+ },
307
+ {
308
+ "epoch": 3.2230703986429177,
309
+ "grad_norm": 1.5529320240020752,
310
+ "learning_rate": 7.111111111111112e-06,
311
+ "loss": 0.7436,
312
+ "step": 3800
313
+ },
314
+ {
315
+ "epoch": 3.3078880407124682,
316
+ "grad_norm": 2.430915117263794,
317
+ "learning_rate": 6.7718405428329095e-06,
318
+ "loss": 0.7421,
319
+ "step": 3900
320
+ },
321
+ {
322
+ "epoch": 3.392705682782019,
323
+ "grad_norm": 2.0328726768493652,
324
+ "learning_rate": 6.432569974554708e-06,
325
+ "loss": 0.7397,
326
+ "step": 4000
327
+ },
328
+ {
329
+ "epoch": 3.4775233248515693,
330
+ "grad_norm": 1.6856147050857544,
331
+ "learning_rate": 6.093299406276506e-06,
332
+ "loss": 0.7358,
333
+ "step": 4100
334
+ },
335
+ {
336
+ "epoch": 3.5623409669211195,
337
+ "grad_norm": 2.0061874389648438,
338
+ "learning_rate": 5.754028837998304e-06,
339
+ "loss": 0.7355,
340
+ "step": 4200
341
+ },
342
+ {
343
+ "epoch": 3.64715860899067,
344
+ "grad_norm": 2.331062078475952,
345
+ "learning_rate": 5.414758269720102e-06,
346
+ "loss": 0.7433,
347
+ "step": 4300
348
+ },
349
+ {
350
+ "epoch": 3.7319762510602206,
351
+ "grad_norm": 2.5025064945220947,
352
+ "learning_rate": 5.075487701441901e-06,
353
+ "loss": 0.7371,
354
+ "step": 4400
355
+ },
356
+ {
357
+ "epoch": 3.816793893129771,
358
+ "grad_norm": 3.021045446395874,
359
+ "learning_rate": 4.736217133163698e-06,
360
+ "loss": 0.7357,
361
+ "step": 4500
362
+ },
363
+ {
364
+ "epoch": 3.9016115351993212,
365
+ "grad_norm": 1.7884670495986938,
366
+ "learning_rate": 4.396946564885497e-06,
367
+ "loss": 0.7366,
368
+ "step": 4600
369
+ },
370
+ {
371
+ "epoch": 3.986429177268872,
372
+ "grad_norm": 2.477693796157837,
373
+ "learning_rate": 4.0576759966072945e-06,
374
+ "loss": 0.7324,
375
+ "step": 4700
376
+ },
377
+ {
378
+ "epoch": 4.0,
379
+ "eval_accuracy": 0.24192905540055798,
380
+ "eval_f1": 0.18676919220051297,
381
+ "eval_loss": 3.059746265411377,
382
+ "eval_precision": 0.23033226126155365,
383
+ "eval_recall": 0.24192905540055798,
384
+ "eval_runtime": 19.9179,
385
+ "eval_samples_per_second": 1259.67,
386
+ "eval_steps_per_second": 19.731,
387
+ "step": 4716
388
+ },
389
+ {
390
+ "epoch": 4.071246819338422,
391
+ "grad_norm": 2.0766072273254395,
392
+ "learning_rate": 3.7184054283290927e-06,
393
+ "loss": 0.7299,
394
+ "step": 4800
395
+ },
396
+ {
397
+ "epoch": 4.156064461407973,
398
+ "grad_norm": 2.57733416557312,
399
+ "learning_rate": 3.3791348600508912e-06,
400
+ "loss": 0.7316,
401
+ "step": 4900
402
+ },
403
+ {
404
+ "epoch": 4.2408821034775235,
405
+ "grad_norm": 2.095597267150879,
406
+ "learning_rate": 3.039864291772689e-06,
407
+ "loss": 0.7343,
408
+ "step": 5000
409
+ },
410
+ {
411
+ "epoch": 4.325699745547074,
412
+ "grad_norm": 1.9832748174667358,
413
+ "learning_rate": 2.7005937234944875e-06,
414
+ "loss": 0.7298,
415
+ "step": 5100
416
+ },
417
+ {
418
+ "epoch": 4.410517387616625,
419
+ "grad_norm": 2.461386203765869,
420
+ "learning_rate": 2.361323155216285e-06,
421
+ "loss": 0.7292,
422
+ "step": 5200
423
+ },
424
+ {
425
+ "epoch": 4.495335029686174,
426
+ "grad_norm": 2.768099546432495,
427
+ "learning_rate": 2.0220525869380833e-06,
428
+ "loss": 0.7277,
429
+ "step": 5300
430
+ },
431
+ {
432
+ "epoch": 4.580152671755725,
433
+ "grad_norm": 1.4862111806869507,
434
+ "learning_rate": 1.6827820186598814e-06,
435
+ "loss": 0.7303,
436
+ "step": 5400
437
+ },
438
+ {
439
+ "epoch": 4.664970313825275,
440
+ "grad_norm": 2.281935214996338,
441
+ "learning_rate": 1.3435114503816796e-06,
442
+ "loss": 0.7287,
443
+ "step": 5500
444
+ },
445
+ {
446
+ "epoch": 4.749787955894826,
447
+ "grad_norm": 1.5928317308425903,
448
+ "learning_rate": 1.0042408821034775e-06,
449
+ "loss": 0.7237,
450
+ "step": 5600
451
+ },
452
+ {
453
+ "epoch": 4.8346055979643765,
454
+ "grad_norm": 2.170006275177002,
455
+ "learning_rate": 6.649703138252757e-07,
456
+ "loss": 0.7271,
457
+ "step": 5700
458
+ },
459
+ {
460
+ "epoch": 4.919423240033927,
461
+ "grad_norm": 2.7832131385803223,
462
+ "learning_rate": 3.256997455470738e-07,
463
+ "loss": 0.7245,
464
+ "step": 5800
465
+ },
466
+ {
467
+ "epoch": 5.0,
468
+ "eval_accuracy": 0.24814667198086887,
469
+ "eval_f1": 0.2479496923119951,
470
+ "eval_loss": 3.093135356903076,
471
+ "eval_precision": 0.3335291041205902,
472
+ "eval_recall": 0.24814667198086887,
473
+ "eval_runtime": 19.8549,
474
+ "eval_samples_per_second": 1263.666,
475
+ "eval_steps_per_second": 19.794,
476
+ "step": 5895
477
+ }
478
+ ],
479
+ "logging_steps": 100,
480
+ "max_steps": 5895,
481
+ "num_input_tokens_seen": 0,
482
+ "num_train_epochs": 5,
483
+ "save_steps": 500,
484
+ "stateful_callbacks": {
485
+ "TrainerControl": {
486
+ "args": {
487
+ "should_epoch_stop": false,
488
+ "should_evaluate": false,
489
+ "should_log": false,
490
+ "should_save": true,
491
+ "should_training_stop": true
492
+ },
493
+ "attributes": {}
494
+ }
495
+ },
496
+ "total_flos": 2144050881888768.0,
497
+ "train_batch_size": 32,
498
+ "trial_name": null,
499
+ "trial_params": null
500
+ }
checkpoint-5895/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41ecbeafe8ad87a86caca10ba72016e58071d9d73cdd450f90210669b5a66947
3
+ size 5777
checkpoint-5895/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ElectraForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "embedding_size": 128,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 256,
12
+ "id2label": {
13
+ "0": "Dosage-Drug-Positive",
14
+ "1": "Frequency-Drug-Positive",
15
+ "2": "Route-Drug-Negative",
16
+ "3": "Form-Drug-Negative",
17
+ "4": "Strength-Drug-Positive",
18
+ "5": "ADE-Drug-Positive",
19
+ "6": "Frequency-Drug-Negative",
20
+ "7": "Reason-Drug-Negative",
21
+ "8": "ADE-Drug-Negative",
22
+ "9": "Reason-Drug-Positive",
23
+ "10": "Strength-Drug-Negative",
24
+ "11": "Dosage-Drug-Negative",
25
+ "12": "Form-Drug-Positive",
26
+ "13": "Route-Drug-Positive",
27
+ "14": "Duration-Drug-Negative",
28
+ "15": "Duration-Drug-Positive"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 1024,
32
+ "label2id": {
33
+ "ADE-Drug-Negative": 8,
34
+ "ADE-Drug-Positive": 5,
35
+ "Dosage-Drug-Negative": 11,
36
+ "Dosage-Drug-Positive": 0,
37
+ "Duration-Drug-Negative": 14,
38
+ "Duration-Drug-Positive": 15,
39
+ "Form-Drug-Negative": 3,
40
+ "Form-Drug-Positive": 12,
41
+ "Frequency-Drug-Negative": 6,
42
+ "Frequency-Drug-Positive": 1,
43
+ "Reason-Drug-Negative": 7,
44
+ "Reason-Drug-Positive": 9,
45
+ "Route-Drug-Negative": 2,
46
+ "Route-Drug-Positive": 13,
47
+ "Strength-Drug-Negative": 10,
48
+ "Strength-Drug-Positive": 4
49
+ },
50
+ "layer_norm_eps": 1e-12,
51
+ "max_position_embeddings": 512,
52
+ "model_type": "electra",
53
+ "num_attention_heads": 4,
54
+ "num_hidden_layers": 12,
55
+ "pad_token_id": 0,
56
+ "position_embedding_type": "absolute",
57
+ "problem_type": "single_label_classification",
58
+ "summary_activation": "gelu",
59
+ "summary_last_dropout": 0.1,
60
+ "summary_type": "first",
61
+ "summary_use_proj": true,
62
+ "transformers_version": "4.56.2",
63
+ "type_vocab_size": 2,
64
+ "use_cache": true,
65
+ "vocab_size": 30522
66
+ }
evaluation_results.txt ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Evaluation Results:
2
+ eval_loss: 2.2932
3
+ eval_accuracy: 0.3264
4
+ eval_f1: 0.2343
5
+ eval_precision: 0.2050
6
+ eval_recall: 0.3264
7
+ eval_runtime: 20.1358
8
+ eval_samples_per_second: 1246.0390
9
+ eval_steps_per_second: 19.5170
10
+ epoch: 5.0000
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:744b8be1ffac1ead1113028fdfb911e1917098a8aa00a38862b30fe4c2694984
3
+ size 54235600
runs/Oct03_23-16-17_3cb83e6cc4fa/events.out.tfevents.1759533380.3cb83e6cc4fa.306.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b9302be9a9c83e2a462e74bea385038df84c1acb15ad467f46a68aee178c9a1
3
+ size 21124
runs/Oct03_23-16-17_3cb83e6cc4fa/events.out.tfevents.1759534206.3cb83e6cc4fa.306.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f726f80d1e5ce8ac14e0c5d92e03ab36139023cf61f9ff464017094768691cfe
3
+ size 560
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "101": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "102": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "103": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "ElectraTokenizer",
57
+ "unk_token": "[UNK]"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41ecbeafe8ad87a86caca10ba72016e58071d9d73cdd450f90210669b5a66947
3
+ size 5777
vocab.txt ADDED
The diff for this file is too large to render. See raw diff