sdeakin commited on
Commit
496a90e
·
verified ·
1 Parent(s): 79df9a3

Upload folder using huggingface_hub

Browse files
checkpoint-10562/config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "MultiLabelBert"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "LABEL_0": 0,
46
+ "LABEL_1": 1,
47
+ "LABEL_10": 10,
48
+ "LABEL_11": 11,
49
+ "LABEL_12": 12,
50
+ "LABEL_13": 13,
51
+ "LABEL_14": 14,
52
+ "LABEL_15": 15,
53
+ "LABEL_16": 16,
54
+ "LABEL_17": 17,
55
+ "LABEL_18": 18,
56
+ "LABEL_19": 19,
57
+ "LABEL_2": 2,
58
+ "LABEL_20": 20,
59
+ "LABEL_21": 21,
60
+ "LABEL_22": 22,
61
+ "LABEL_23": 23,
62
+ "LABEL_24": 24,
63
+ "LABEL_25": 25,
64
+ "LABEL_26": 26,
65
+ "LABEL_27": 27,
66
+ "LABEL_3": 3,
67
+ "LABEL_4": 4,
68
+ "LABEL_5": 5,
69
+ "LABEL_6": 6,
70
+ "LABEL_7": 7,
71
+ "LABEL_8": 8,
72
+ "LABEL_9": 9
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "problem_type": "multi_label_classification",
82
+ "torch_dtype": "float32",
83
+ "transformers_version": "4.49.0",
84
+ "type_vocab_size": 2,
85
+ "use_cache": true,
86
+ "vocab_size": 30522
87
+ }
checkpoint-10562/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:943134069a1d1fef7c036036036417e1a20865c4c3a4d311bccf73f1d2a2706e
3
+ size 438038624
checkpoint-10562/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f741aa29fb3a48c4b177ef20f81eaad7412d41caf7b67016a70357d1d38695ad
3
+ size 876192634
checkpoint-10562/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b182573f61d8bcf5eaefcbf8f98d8734b6db51b44ad36aed3a305c431539fa1
3
+ size 13990
checkpoint-10562/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae54731f31173b97e1d26cde16597d3bf34af56aec47af25c4e901464306b285
3
+ size 1064
checkpoint-10562/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10562/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e421e8a046c53a92a7607299cf0df782cbc358735300a88a9e25dcff8675205
3
+ size 5240
checkpoint-21124/config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "MultiLabelBert"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "LABEL_0": 0,
46
+ "LABEL_1": 1,
47
+ "LABEL_10": 10,
48
+ "LABEL_11": 11,
49
+ "LABEL_12": 12,
50
+ "LABEL_13": 13,
51
+ "LABEL_14": 14,
52
+ "LABEL_15": 15,
53
+ "LABEL_16": 16,
54
+ "LABEL_17": 17,
55
+ "LABEL_18": 18,
56
+ "LABEL_19": 19,
57
+ "LABEL_2": 2,
58
+ "LABEL_20": 20,
59
+ "LABEL_21": 21,
60
+ "LABEL_22": 22,
61
+ "LABEL_23": 23,
62
+ "LABEL_24": 24,
63
+ "LABEL_25": 25,
64
+ "LABEL_26": 26,
65
+ "LABEL_27": 27,
66
+ "LABEL_3": 3,
67
+ "LABEL_4": 4,
68
+ "LABEL_5": 5,
69
+ "LABEL_6": 6,
70
+ "LABEL_7": 7,
71
+ "LABEL_8": 8,
72
+ "LABEL_9": 9
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "problem_type": "multi_label_classification",
82
+ "torch_dtype": "float32",
83
+ "transformers_version": "4.49.0",
84
+ "type_vocab_size": 2,
85
+ "use_cache": true,
86
+ "vocab_size": 30522
87
+ }
checkpoint-21124/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb56654e89e4e6f6bf853dcbc652e4b2b352ce378037121029283a1165d74db6
3
+ size 438038624
checkpoint-21124/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:348f737befcae7e5a5830e4c9906d14d703efc9e21d6a26adfd4edfb4f5c6acb
3
+ size 876192634
checkpoint-21124/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e55bfc8723f269a626afca0be6f7def5753f3bb265436b94c5580b703cfcc7
3
+ size 13990
checkpoint-21124/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c19cf29eea57795e31f17fc642d48aaa3f3d14252ff8040ba312f0ec3f61391
3
+ size 1064
checkpoint-21124/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-21124/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e421e8a046c53a92a7607299cf0df782cbc358735300a88a9e25dcff8675205
3
+ size 5240
checkpoint-31686/config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "MultiLabelBert"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "LABEL_0": 0,
46
+ "LABEL_1": 1,
47
+ "LABEL_10": 10,
48
+ "LABEL_11": 11,
49
+ "LABEL_12": 12,
50
+ "LABEL_13": 13,
51
+ "LABEL_14": 14,
52
+ "LABEL_15": 15,
53
+ "LABEL_16": 16,
54
+ "LABEL_17": 17,
55
+ "LABEL_18": 18,
56
+ "LABEL_19": 19,
57
+ "LABEL_2": 2,
58
+ "LABEL_20": 20,
59
+ "LABEL_21": 21,
60
+ "LABEL_22": 22,
61
+ "LABEL_23": 23,
62
+ "LABEL_24": 24,
63
+ "LABEL_25": 25,
64
+ "LABEL_26": 26,
65
+ "LABEL_27": 27,
66
+ "LABEL_3": 3,
67
+ "LABEL_4": 4,
68
+ "LABEL_5": 5,
69
+ "LABEL_6": 6,
70
+ "LABEL_7": 7,
71
+ "LABEL_8": 8,
72
+ "LABEL_9": 9
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "problem_type": "multi_label_classification",
82
+ "torch_dtype": "float32",
83
+ "transformers_version": "4.49.0",
84
+ "type_vocab_size": 2,
85
+ "use_cache": true,
86
+ "vocab_size": 30522
87
+ }
checkpoint-31686/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75b379f4e3fa8aabbc6e05f31ca3044af11c0b49afe8326631d8a232a6d21dff
3
+ size 438038624
checkpoint-31686/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce17dd7f4b5cc550e52c932f8fdf684d08617ff8e2f5c599d87ac63766120edb
3
+ size 876192634
checkpoint-31686/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32d32051101ec51c2b04c4ee6a6d2c7f40562e56836cbb02d6e6e3126490484d
3
+ size 13990
checkpoint-31686/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d00cf7325f10cab9a1d8a7085895eefdc5964f3175c6af77d0248a9f4dba738
3
+ size 1064
checkpoint-31686/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-31686/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e421e8a046c53a92a7607299cf0df782cbc358735300a88a9e25dcff8675205
3
+ size 5240
checkpoint-42248/config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "MultiLabelBert"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "LABEL_0": 0,
46
+ "LABEL_1": 1,
47
+ "LABEL_10": 10,
48
+ "LABEL_11": 11,
49
+ "LABEL_12": 12,
50
+ "LABEL_13": 13,
51
+ "LABEL_14": 14,
52
+ "LABEL_15": 15,
53
+ "LABEL_16": 16,
54
+ "LABEL_17": 17,
55
+ "LABEL_18": 18,
56
+ "LABEL_19": 19,
57
+ "LABEL_2": 2,
58
+ "LABEL_20": 20,
59
+ "LABEL_21": 21,
60
+ "LABEL_22": 22,
61
+ "LABEL_23": 23,
62
+ "LABEL_24": 24,
63
+ "LABEL_25": 25,
64
+ "LABEL_26": 26,
65
+ "LABEL_27": 27,
66
+ "LABEL_3": 3,
67
+ "LABEL_4": 4,
68
+ "LABEL_5": 5,
69
+ "LABEL_6": 6,
70
+ "LABEL_7": 7,
71
+ "LABEL_8": 8,
72
+ "LABEL_9": 9
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "problem_type": "multi_label_classification",
82
+ "torch_dtype": "float32",
83
+ "transformers_version": "4.49.0",
84
+ "type_vocab_size": 2,
85
+ "use_cache": true,
86
+ "vocab_size": 30522
87
+ }
checkpoint-42248/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0810c53c98aa2fee970cd962449eb312de0a05f7cb52574e03c5c795722345e7
3
+ size 438038624
checkpoint-42248/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53350b3d47cf50b3b80c9744e238b958206b0220fb532a33b3e99d4861799ab0
3
+ size 876192634
checkpoint-42248/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fed77b14fe062f6db72d68cedd6fd95bae3305b7a735eef3c85da43fd15d476
3
+ size 13990
checkpoint-42248/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:602735795cdacf9e80225647d70a5187b692a9da126eb075205ea84e518facb1
3
+ size 1064
checkpoint-42248/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-42248/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e421e8a046c53a92a7607299cf0df782cbc358735300a88a9e25dcff8675205
3
+ size 5240
checkpoint-52810/config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "MultiLabelBert"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27"
41
+ },
42
+ "initializer_range": 0.02,
43
+ "intermediate_size": 3072,
44
+ "label2id": {
45
+ "LABEL_0": 0,
46
+ "LABEL_1": 1,
47
+ "LABEL_10": 10,
48
+ "LABEL_11": 11,
49
+ "LABEL_12": 12,
50
+ "LABEL_13": 13,
51
+ "LABEL_14": 14,
52
+ "LABEL_15": 15,
53
+ "LABEL_16": 16,
54
+ "LABEL_17": 17,
55
+ "LABEL_18": 18,
56
+ "LABEL_19": 19,
57
+ "LABEL_2": 2,
58
+ "LABEL_20": 20,
59
+ "LABEL_21": 21,
60
+ "LABEL_22": 22,
61
+ "LABEL_23": 23,
62
+ "LABEL_24": 24,
63
+ "LABEL_25": 25,
64
+ "LABEL_26": 26,
65
+ "LABEL_27": 27,
66
+ "LABEL_3": 3,
67
+ "LABEL_4": 4,
68
+ "LABEL_5": 5,
69
+ "LABEL_6": 6,
70
+ "LABEL_7": 7,
71
+ "LABEL_8": 8,
72
+ "LABEL_9": 9
73
+ },
74
+ "layer_norm_eps": 1e-12,
75
+ "max_position_embeddings": 512,
76
+ "model_type": "bert",
77
+ "num_attention_heads": 12,
78
+ "num_hidden_layers": 12,
79
+ "pad_token_id": 0,
80
+ "position_embedding_type": "absolute",
81
+ "problem_type": "multi_label_classification",
82
+ "torch_dtype": "float32",
83
+ "transformers_version": "4.49.0",
84
+ "type_vocab_size": 2,
85
+ "use_cache": true,
86
+ "vocab_size": 30522
87
+ }
checkpoint-52810/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eca9f280a89c73e30d2275a7e061fdf9feee44bb875802c0de7682400d5e98c0
3
+ size 438038624
checkpoint-52810/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab4503b3a7fbc578e6fdc449847a93975aa03c00aafa3a13309a1c62cda817d5
3
+ size 876192634
checkpoint-52810/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:013d13e4b269a8648da58d2f40bf48e004388ea0631771b8c8e588ab75375092
3
+ size 13990
checkpoint-52810/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2bf40adef5d677e667a419ff9fe2b150a41d5133cf0867e0cb396a4d17c5d81
3
+ size 1064
checkpoint-52810/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-52810/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e421e8a046c53a92a7607299cf0df782cbc358735300a88a9e25dcff8675205
3
+ size 5240