Badhon commited on
Commit
e97d354
·
verified ·
1 Parent(s): c4e4c2d

Upload 17 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ confusion_matrix_attacked_test.png filter=lfs diff=lfs merge=lfs -text
37
+ confusion_matrix_test.png filter=lfs diff=lfs merge=lfs -text
38
+ confusion_matrix_validation.png filter=lfs diff=lfs merge=lfs -text
39
+ model.onnx.data filter=lfs diff=lfs merge=lfs -text
40
+ training_history.png filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ElectraForTokenClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "dtype": "float32",
8
+ "embedding_size": 768,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "O",
14
+ "1": "COMMA",
15
+ "2": "DARI",
16
+ "3": "QUESTION",
17
+ "4": "EXCLAMATION",
18
+ "5": "SEMICOLON",
19
+ "6": "COLON",
20
+ "7": "HYPHEN"
21
+ },
22
+ "initializer_range": 0.02,
23
+ "intermediate_size": 3072,
24
+ "label2id": {
25
+ "COLON": 6,
26
+ "COMMA": 1,
27
+ "DARI": 2,
28
+ "EXCLAMATION": 4,
29
+ "HYPHEN": 7,
30
+ "O": 0,
31
+ "QUESTION": 3,
32
+ "SEMICOLON": 5
33
+ },
34
+ "layer_norm_eps": 1e-12,
35
+ "max_position_embeddings": 512,
36
+ "model_type": "electra",
37
+ "num_attention_heads": 12,
38
+ "num_hidden_layers": 12,
39
+ "pad_token_id": 0,
40
+ "position_embedding_type": "absolute",
41
+ "summary_activation": "gelu",
42
+ "summary_last_dropout": 0.1,
43
+ "summary_type": "first",
44
+ "summary_use_proj": true,
45
+ "transformers_version": "4.57.1",
46
+ "type_vocab_size": 2,
47
+ "use_cache": true,
48
+ "vocab_size": 32000
49
+ }
confusion_matrix_attacked_test.png ADDED

Git LFS Details

  • SHA256: fee7fa47ca58b492f1497c01f494f70afe0e069bf1e39ecbc127150bdaadc323
  • Pointer size: 131 Bytes
  • Size of remote file: 224 kB
confusion_matrix_test.png ADDED

Git LFS Details

  • SHA256: eda90ad5ba9edd631e828a0e938b9245caa9755ccbfb4590990757728b47f9fb
  • Pointer size: 131 Bytes
  • Size of remote file: 232 kB
confusion_matrix_validation.png ADDED

Git LFS Details

  • SHA256: 91fc43a0712669515b4672734218a17fc99658386076b9e53b68bf3268fcbcf3
  • Pointer size: 131 Bytes
  • Size of remote file: 199 kB
detailed_results_attacked_test.json ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "split": "attacked_test",
3
+ "classification_report": " precision recall f1-score support\n\n O 0.9832 0.9261 0.9538 16661\n COMMA 0.2602 0.5802 0.3593 374\n DARI 0.1744 0.6433 0.2745 157\n QUESTION 0.2908 0.5190 0.3727 79\n EXCLAMATION 0.1957 0.5000 0.2812 18\n SEMICOLON 0.0000 0.0000 0.0000 4\n COLON 0.1026 0.2000 0.1356 20\n HYPHEN 0.0000 0.0000 0.0000 31\n\n accuracy 0.9110 17344\n macro avg 0.2509 0.4211 0.2971 17344\nweighted avg 0.9533 0.9110 0.9286 17344\n",
4
+ "confusion_matrix": [
5
+ [
6
+ 15429,
7
+ 601,
8
+ 456,
9
+ 96,
10
+ 34,
11
+ 7,
12
+ 34,
13
+ 4
14
+ ],
15
+ [
16
+ 149,
17
+ 217,
18
+ 6,
19
+ 1,
20
+ 0,
21
+ 0,
22
+ 1,
23
+ 0
24
+ ],
25
+ [
26
+ 41,
27
+ 9,
28
+ 101,
29
+ 3,
30
+ 3,
31
+ 0,
32
+ 0,
33
+ 0
34
+ ],
35
+ [
36
+ 28,
37
+ 3,
38
+ 7,
39
+ 41,
40
+ 0,
41
+ 0,
42
+ 0,
43
+ 0
44
+ ],
45
+ [
46
+ 4,
47
+ 1,
48
+ 4,
49
+ 0,
50
+ 9,
51
+ 0,
52
+ 0,
53
+ 0
54
+ ],
55
+ [
56
+ 2,
57
+ 0,
58
+ 2,
59
+ 0,
60
+ 0,
61
+ 0,
62
+ 0,
63
+ 0
64
+ ],
65
+ [
66
+ 10,
67
+ 2,
68
+ 3,
69
+ 0,
70
+ 0,
71
+ 0,
72
+ 4,
73
+ 1
74
+ ],
75
+ [
76
+ 30,
77
+ 1,
78
+ 0,
79
+ 0,
80
+ 0,
81
+ 0,
82
+ 0,
83
+ 0
84
+ ]
85
+ ],
86
+ "per_class_metrics": [
87
+ {
88
+ "class": "O",
89
+ "support": 16661,
90
+ "accuracy": 0.9260548586519417,
91
+ "correct": 15429
92
+ },
93
+ {
94
+ "class": "COMMA",
95
+ "support": 374,
96
+ "accuracy": 0.5802139037433155,
97
+ "correct": 217
98
+ },
99
+ {
100
+ "class": "DARI",
101
+ "support": 157,
102
+ "accuracy": 0.643312101910828,
103
+ "correct": 101
104
+ },
105
+ {
106
+ "class": "QUESTION",
107
+ "support": 79,
108
+ "accuracy": 0.5189873417721519,
109
+ "correct": 41
110
+ },
111
+ {
112
+ "class": "EXCLAMATION",
113
+ "support": 18,
114
+ "accuracy": 0.5,
115
+ "correct": 9
116
+ },
117
+ {
118
+ "class": "SEMICOLON",
119
+ "support": 4,
120
+ "accuracy": 0.0,
121
+ "correct": 0
122
+ },
123
+ {
124
+ "class": "COLON",
125
+ "support": 20,
126
+ "accuracy": 0.2,
127
+ "correct": 4
128
+ },
129
+ {
130
+ "class": "HYPHEN",
131
+ "support": 31,
132
+ "accuracy": 0.0,
133
+ "correct": 0
134
+ }
135
+ ],
136
+ "bleu_rouge_scores": {
137
+ "bleu": 82.59219598233263,
138
+ "rouge1_precision": 0.0,
139
+ "rouge1_recall": 0.0,
140
+ "rouge1_f": 0.0,
141
+ "rouge2_precision": 0.0,
142
+ "rouge2_recall": 0.0,
143
+ "rouge2_f": 0.0,
144
+ "rougeL_precision": 0.0,
145
+ "rougeL_recall": 0.0,
146
+ "rougeL_f": 0.0
147
+ },
148
+ "overall_metrics": {
149
+ "test_loss": 0.05920533090829849,
150
+ "test_precision": 0.9532955961063683,
151
+ "test_recall": 0.911035516605166,
152
+ "test_f1": 0.9285803514155778,
153
+ "test_weighted_f1": 0.9285803514155778,
154
+ "test_punctuation_f1": 0.6590236317324747,
155
+ "test_f1_comma": 0.35927152317880795,
156
+ "test_f1_dari": 0.27445652173913043,
157
+ "test_f1_question": 0.37272727272727274,
158
+ "test_f1_exclamation": 0.28125,
159
+ "test_f1_semicolon": 0.0,
160
+ "test_f1_colon": 0.13559322033898305,
161
+ "test_f1_hyphen": 0.0,
162
+ "test_runtime": 4.8696,
163
+ "test_samples_per_second": 119.724,
164
+ "test_steps_per_second": 2.054
165
+ }
166
+ }
detailed_results_test.json ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "split": "test",
3
+ "classification_report": " precision recall f1-score support\n\n O 0.9742 0.9802 0.9772 17078\n COMMA 0.6969 0.6213 0.6569 1014\n DARI 0.8005 0.8583 0.8284 360\n QUESTION 0.8315 0.8407 0.8361 182\n EXCLAMATION 0.7179 0.7467 0.7320 75\n SEMICOLON 0.0000 0.0000 0.0000 23\n COLON 0.3529 0.4444 0.3934 54\n HYPHEN 0.0000 0.0000 0.0000 23\n\n accuracy 0.9523 18809\n macro avg 0.5467 0.5615 0.5530 18809\nweighted avg 0.9493 0.9523 0.9507 18809\n",
4
+ "confusion_matrix": [
5
+ [
6
+ 16740,
7
+ 241,
8
+ 37,
9
+ 19,
10
+ 4,
11
+ 1,
12
+ 34,
13
+ 2
14
+ ],
15
+ [
16
+ 340,
17
+ 630,
18
+ 26,
19
+ 5,
20
+ 4,
21
+ 1,
22
+ 7,
23
+ 1
24
+ ],
25
+ [
26
+ 24,
27
+ 13,
28
+ 309,
29
+ 4,
30
+ 10,
31
+ 0,
32
+ 0,
33
+ 0
34
+ ],
35
+ [
36
+ 16,
37
+ 3,
38
+ 6,
39
+ 153,
40
+ 4,
41
+ 0,
42
+ 0,
43
+ 0
44
+ ],
45
+ [
46
+ 4,
47
+ 4,
48
+ 8,
49
+ 3,
50
+ 56,
51
+ 0,
52
+ 0,
53
+ 0
54
+ ],
55
+ [
56
+ 18,
57
+ 5,
58
+ 0,
59
+ 0,
60
+ 0,
61
+ 0,
62
+ 0,
63
+ 0
64
+ ],
65
+ [
66
+ 26,
67
+ 4,
68
+ 0,
69
+ 0,
70
+ 0,
71
+ 0,
72
+ 24,
73
+ 0
74
+ ],
75
+ [
76
+ 16,
77
+ 4,
78
+ 0,
79
+ 0,
80
+ 0,
81
+ 0,
82
+ 3,
83
+ 0
84
+ ]
85
+ ],
86
+ "per_class_metrics": [
87
+ {
88
+ "class": "O",
89
+ "support": 17078,
90
+ "accuracy": 0.9802084553226373,
91
+ "correct": 16740
92
+ },
93
+ {
94
+ "class": "COMMA",
95
+ "support": 1014,
96
+ "accuracy": 0.621301775147929,
97
+ "correct": 630
98
+ },
99
+ {
100
+ "class": "DARI",
101
+ "support": 360,
102
+ "accuracy": 0.8583333333333333,
103
+ "correct": 309
104
+ },
105
+ {
106
+ "class": "QUESTION",
107
+ "support": 182,
108
+ "accuracy": 0.8406593406593407,
109
+ "correct": 153
110
+ },
111
+ {
112
+ "class": "EXCLAMATION",
113
+ "support": 75,
114
+ "accuracy": 0.7466666666666667,
115
+ "correct": 56
116
+ },
117
+ {
118
+ "class": "SEMICOLON",
119
+ "support": 23,
120
+ "accuracy": 0.0,
121
+ "correct": 0
122
+ },
123
+ {
124
+ "class": "COLON",
125
+ "support": 54,
126
+ "accuracy": 0.4444444444444444,
127
+ "correct": 24
128
+ },
129
+ {
130
+ "class": "HYPHEN",
131
+ "support": 23,
132
+ "accuracy": 0.0,
133
+ "correct": 0
134
+ }
135
+ ],
136
+ "bleu_rouge_scores": {
137
+ "bleu": 90.0686061783956,
138
+ "rouge1_precision": 0.0,
139
+ "rouge1_recall": 0.0,
140
+ "rouge1_f": 0.0,
141
+ "rouge2_precision": 0.0,
142
+ "rouge2_recall": 0.0,
143
+ "rouge2_f": 0.0,
144
+ "rougeL_precision": 0.0,
145
+ "rougeL_recall": 0.0,
146
+ "rougeL_f": 0.0
147
+ },
148
+ "overall_metrics": {
149
+ "test_loss": 0.02253320813179016,
150
+ "test_precision": 0.9493235462409354,
151
+ "test_recall": 0.9523100643309054,
152
+ "test_f1": 0.9506557193404622,
153
+ "test_weighted_f1": 0.9506557193404622,
154
+ "test_punctuation_f1": 0.7637091871138297,
155
+ "test_f1_comma": 0.656934306569343,
156
+ "test_f1_dari": 0.8284182305630027,
157
+ "test_f1_question": 0.8360655737704918,
158
+ "test_f1_exclamation": 0.7320261437908496,
159
+ "test_f1_semicolon": 0.0,
160
+ "test_f1_colon": 0.39344262295081966,
161
+ "test_f1_hyphen": 0.0,
162
+ "test_runtime": 4.8333,
163
+ "test_samples_per_second": 120.622,
164
+ "test_steps_per_second": 2.069
165
+ }
166
+ }
detailed_results_validation.json ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "split": "validation",
3
+ "classification_report": " precision recall f1-score support\n\n O 0.9736 0.9813 0.9774 17717\n COMMA 0.6741 0.6105 0.6408 986\n DARI 0.7895 0.7971 0.7933 414\n QUESTION 0.8565 0.8685 0.8625 213\n EXCLAMATION 0.5909 0.5417 0.5652 48\n SEMICOLON 0.7333 0.4400 0.5500 25\n COLON 0.5614 0.4211 0.4812 76\n HYPHEN 0.4286 0.1111 0.1765 27\n\n accuracy 0.9522 19506\n macro avg 0.7010 0.5964 0.6309 19506\nweighted avg 0.9497 0.9522 0.9506 19506\n",
4
+ "confusion_matrix": [
5
+ [
6
+ 17385,
7
+ 246,
8
+ 47,
9
+ 17,
10
+ 3,
11
+ 1,
12
+ 17,
13
+ 1
14
+ ],
15
+ [
16
+ 342,
17
+ 602,
18
+ 28,
19
+ 3,
20
+ 2,
21
+ 2,
22
+ 4,
23
+ 3
24
+ ],
25
+ [
26
+ 40,
27
+ 28,
28
+ 330,
29
+ 4,
30
+ 10,
31
+ 1,
32
+ 1,
33
+ 0
34
+ ],
35
+ [
36
+ 18,
37
+ 1,
38
+ 6,
39
+ 185,
40
+ 3,
41
+ 0,
42
+ 0,
43
+ 0
44
+ ],
45
+ [
46
+ 5,
47
+ 3,
48
+ 7,
49
+ 7,
50
+ 26,
51
+ 0,
52
+ 0,
53
+ 0
54
+ ],
55
+ [
56
+ 12,
57
+ 2,
58
+ 0,
59
+ 0,
60
+ 0,
61
+ 11,
62
+ 0,
63
+ 0
64
+ ],
65
+ [
66
+ 34,
67
+ 10,
68
+ 0,
69
+ 0,
70
+ 0,
71
+ 0,
72
+ 32,
73
+ 0
74
+ ],
75
+ [
76
+ 20,
77
+ 1,
78
+ 0,
79
+ 0,
80
+ 0,
81
+ 0,
82
+ 3,
83
+ 3
84
+ ]
85
+ ],
86
+ "per_class_metrics": [
87
+ {
88
+ "class": "O",
89
+ "support": 17717,
90
+ "accuracy": 0.9812609358243495,
91
+ "correct": 17385
92
+ },
93
+ {
94
+ "class": "COMMA",
95
+ "support": 986,
96
+ "accuracy": 0.6105476673427992,
97
+ "correct": 602
98
+ },
99
+ {
100
+ "class": "DARI",
101
+ "support": 414,
102
+ "accuracy": 0.7971014492753623,
103
+ "correct": 330
104
+ },
105
+ {
106
+ "class": "QUESTION",
107
+ "support": 213,
108
+ "accuracy": 0.8685446009389671,
109
+ "correct": 185
110
+ },
111
+ {
112
+ "class": "EXCLAMATION",
113
+ "support": 48,
114
+ "accuracy": 0.5416666666666666,
115
+ "correct": 26
116
+ },
117
+ {
118
+ "class": "SEMICOLON",
119
+ "support": 25,
120
+ "accuracy": 0.44,
121
+ "correct": 11
122
+ },
123
+ {
124
+ "class": "COLON",
125
+ "support": 76,
126
+ "accuracy": 0.42105263157894735,
127
+ "correct": 32
128
+ },
129
+ {
130
+ "class": "HYPHEN",
131
+ "support": 27,
132
+ "accuracy": 0.1111111111111111,
133
+ "correct": 3
134
+ }
135
+ ],
136
+ "bleu_rouge_scores": {
137
+ "bleu": 90.20234005231055,
138
+ "rouge1_precision": 0.0,
139
+ "rouge1_recall": 0.0,
140
+ "rouge1_f": 0.0,
141
+ "rouge2_precision": 0.0,
142
+ "rouge2_recall": 0.0,
143
+ "rouge2_f": 0.0,
144
+ "rougeL_precision": 0.0,
145
+ "rougeL_recall": 0.0,
146
+ "rougeL_f": 0.0
147
+ },
148
+ "overall_metrics": {
149
+ "test_loss": 0.023327844217419624,
150
+ "test_precision": 0.9496856615053557,
151
+ "test_recall": 0.9522198297959602,
152
+ "test_f1": 0.950640869722443,
153
+ "test_weighted_f1": 0.950640869722443,
154
+ "test_punctuation_f1": 0.7575861893899899,
155
+ "test_f1_comma": 0.6407663650878127,
156
+ "test_f1_dari": 0.7932692307692307,
157
+ "test_f1_question": 0.8624708624708625,
158
+ "test_f1_exclamation": 0.5652173913043478,
159
+ "test_f1_semicolon": 0.55,
160
+ "test_f1_colon": 0.48120300751879697,
161
+ "test_f1_hyphen": 0.17647058823529413,
162
+ "test_runtime": 4.8056,
163
+ "test_samples_per_second": 121.108,
164
+ "test_steps_per_second": 2.081
165
+ }
166
+ }
model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a92f5e10999e782700003fbd7a945fd8642fcdccf4c82db86587571cca41739
3
+ size 1217718
model.onnx.data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2259b4911ff5c543d618adaf3862ae1d88269fc636158cebefece8f06c8454a
3
+ size 440139776
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9424e9b37281d9a3476c04ca37518ec08e6bf340923cea4c40d740d4b4fa16e
3
+ size 440155536
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": false,
48
+ "extra_special_tokens": {},
49
+ "full_tokenizer_file": null,
50
+ "mask_token": "[MASK]",
51
+ "max_length": 200,
52
+ "model_max_length": 1000000000000000019884624838656,
53
+ "never_split": null,
54
+ "pad_to_multiple_of": null,
55
+ "pad_token": "[PAD]",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
+ "sep_token": "[SEP]",
59
+ "stride": 0,
60
+ "strip_accents": null,
61
+ "tokenize_chinese_chars": false,
62
+ "tokenizer_class": "ElectraTokenizer",
63
+ "truncation_side": "right",
64
+ "truncation_strategy": "longest_first",
65
+ "unk_token": "[UNK]"
66
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ab1461816db7bef705da8566020ca27553f0459ecebc07928dff4aeed60c45d
3
+ size 5841
training_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ "{'__module__': '__main__', 'TOKENIZED_DATA_DIR': '/kaggle/working/Hishab_Dataset/custom-pr-bn-v1-tokenized', 'OUTPUT_DIR': '/kaggle/working/models/custom-pr-bn-v1', 'LOGGING_DIR': '/kaggle/working/logs/custom-pr-bn-v1', 'MODEL_NAME': 'csebuetnlp/banglabert', 'NUM_LABELS': 8, 'LEARNING_RATE': 2e-05, 'WEIGHT_DECAY': 0.01, 'NUM_EPOCHS': 50, 'WARMUP_RATIO': 0.1, 'TRAIN_BATCH_SIZE': 16, 'EVAL_BATCH_SIZE': 32, 'GRADIENT_ACCUMULATION_STEPS': 2, 'OPTIMIZER': 'adamw_torch', 'LR_SCHEDULER': 'cosine', 'MAX_GRAD_NORM': 1.0, 'DROPOUT': 0.1, 'ATTENTION_DROPOUT': 0.1, 'HIDDEN_DROPOUT': 0.1, 'EARLY_STOPPING_PATIENCE': 5, 'EARLY_STOPPING_THRESHOLD': 0.001, 'EVAL_STRATEGY': 'epoch', 'SAVE_STRATEGY': 'epoch', 'SAVE_TOTAL_LIMIT': 3, 'LOAD_BEST_MODEL': True, 'METRIC_FOR_BEST_MODEL': 'eval_weighted_f1', 'GREATER_IS_BETTER': True, 'LOGGING_STEPS': 50, 'REPORT_TO': 'tensorboard', 'USE_CLASS_WEIGHTS': True, 'FOCAL_LOSS': True, 'FOCAL_ALPHA': 0.25, 'FOCAL_GAMMA': 2.0, 'SEED': 42, '__dict__': <attribute '__dict__' of 'TrainingConfig' objects>, '__weakref__': <attribute '__weakref__' of 'TrainingConfig' objects>, '__doc__': None}"
training_history.png ADDED

Git LFS Details

  • SHA256: aa96d27830b602e0ead0513031da1c3c450d4085499a2210cc2efeadde2cd1f4
  • Pointer size: 131 Bytes
  • Size of remote file: 226 kB
vocab.txt ADDED
The diff for this file is too large to render. See raw diff