albagon commited on
Commit
bb33508
·
verified ·
1 Parent(s): 0c628fe

Upload folder using huggingface_hub

Browse files
Files changed (42) hide show
  1. .gitattributes +1 -0
  2. bertweet_finetuned/best_checkpoint.pt +3 -0
  3. bertweet_finetuned/final_state.pt +3 -0
  4. bertweet_finetuned/inference_config.json +1 -0
  5. bertweet_finetuned/label_classes.joblib +3 -0
  6. bertweet_finetuned/meta_pipeline.joblib +3 -0
  7. bertweet_finetuned/model_metadata.joblib +3 -0
  8. bertweet_finetuned/preds_from_model.csv +3 -0
  9. bertweet_finetuned/tokenizer/added_tokens.json +3 -0
  10. bertweet_finetuned/tokenizer/bpe.codes +0 -0
  11. bertweet_finetuned/tokenizer/special_tokens_map.json +9 -0
  12. bertweet_finetuned/tokenizer/tokenizer_config.json +55 -0
  13. bertweet_finetuned/tokenizer/vocab.txt +0 -0
  14. e5_large/best_checkpoint.pt +3 -0
  15. e5_large/inference_config.json +1 -0
  16. e5_large/label_classes.joblib +3 -0
  17. e5_large/meta_pipeline.joblib +3 -0
  18. e5_large/tokenizer/special_tokens_map.json +37 -0
  19. e5_large/tokenizer/tokenizer.json +0 -0
  20. e5_large/tokenizer/tokenizer_config.json +56 -0
  21. e5_large/tokenizer/vocab.txt +0 -0
  22. e5_large/training_history.json +192 -0
  23. embedding_dimension/embeddings/all-MiniLM-L6-v2_train.npy +3 -0
  24. embedding_dimension/embeddings/all-MiniLM-L6-v2_val.npy +3 -0
  25. embedding_dimension/embeddings/all-mpnet-base-v2_train.npy +3 -0
  26. embedding_dimension/embeddings/all-mpnet-base-v2_val.npy +3 -0
  27. embedding_dimension/embeddings/e5-large-v2_train.npy +3 -0
  28. embedding_dimension/embeddings/e5-large-v2_val.npy +3 -0
  29. gpt2_large_sarcasm_finetuned/merges.txt +0 -0
  30. gpt2_large_sarcasm_finetuned/model.pt +3 -0
  31. gpt2_large_sarcasm_finetuned/special_tokens_map.json +24 -0
  32. gpt2_large_sarcasm_finetuned/tokenizer_config.json +23 -0
  33. gpt2_large_sarcasm_finetuned/training_history.json +30 -0
  34. gpt2_large_sarcasm_finetuned/training_metadata.json +32 -0
  35. gpt2_large_sarcasm_finetuned/vocab.json +0 -0
  36. gpt2_small_sarcasm_finetuned/merges.txt +0 -0
  37. gpt2_small_sarcasm_finetuned/model.pt +3 -0
  38. gpt2_small_sarcasm_finetuned/special_tokens_map.json +24 -0
  39. gpt2_small_sarcasm_finetuned/tokenizer_config.json +23 -0
  40. gpt2_small_sarcasm_finetuned/training_history.json +30 -0
  41. gpt2_small_sarcasm_finetuned/training_metadata.json +32 -0
  42. gpt2_small_sarcasm_finetuned/vocab.json +0 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ bertweet_finetuned/preds_from_model.csv filter=lfs diff=lfs merge=lfs -text
bertweet_finetuned/best_checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6962af37895aaada2468609650dc4b33b70f32ee8686f655cdcd428cdeead064
3
+ size 939408188
bertweet_finetuned/final_state.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9289d512b49a7037f4c24845ef6486a40721f8bfd51bbaebe1cdf8892c8cac04
3
+ size 540584792
bertweet_finetuned/inference_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"MAX_LEN": 128, "SEP": "</s>"}
bertweet_finetuned/label_classes.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e95d56cf907dc2aea39ee0bec10da0e1a67a0a0855ff9b508be45bbac494d76
3
+ size 37
bertweet_finetuned/meta_pipeline.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2b3af29249d2c287dd60b97d18c11ea725d7b4c36bf69c6de9222340ed04435
3
+ size 4541
bertweet_finetuned/model_metadata.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb441c8486b604e107a17ade601b7c39dff2de06a63e903b2dad82c0e844fa1
3
+ size 5140
bertweet_finetuned/preds_from_model.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85f5ac19e9a45094dec107df5294c7323c3ed37f58092bdd78914d6bcc93b2cb
3
+ size 61854171
bertweet_finetuned/tokenizer/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<mask>": 64000
3
+ }
bertweet_finetuned/tokenizer/bpe.codes ADDED
The diff for this file is too large to render. See raw diff
 
bertweet_finetuned/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": "<mask>",
6
+ "pad_token": "<pad>",
7
+ "sep_token": "</s>",
8
+ "unk_token": "<unk>"
9
+ }
bertweet_finetuned/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "64000": {
36
+ "content": "<mask>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "eos_token": "</s>",
48
+ "mask_token": "<mask>",
49
+ "model_max_length": 1000000000000000019884624838656,
50
+ "normalization": false,
51
+ "pad_token": "<pad>",
52
+ "sep_token": "</s>",
53
+ "tokenizer_class": "BertweetTokenizer",
54
+ "unk_token": "<unk>"
55
+ }
bertweet_finetuned/tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
e5_large/best_checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e939f37e079360c857192f0edc597db8970a920b8172d6040a3eedcc80335842
3
+ size 2049712089
e5_large/inference_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"MAX_LEN": 128, "SEP": "[SEP]", "MODEL_NAME": "intfloat/e5-large-v2"}
e5_large/label_classes.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e95d56cf907dc2aea39ee0bec10da0e1a67a0a0855ff9b508be45bbac494d76
3
+ size 37
e5_large/meta_pipeline.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b4406893d7dc7e29d12cf25f18d3e27c20ebcf4bdbf04200a3a5eb3f9972b1c
3
+ size 4539
e5_large/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
e5_large/tokenizer/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
e5_large/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
e5_large/tokenizer/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
e5_large/training_history.json ADDED
@@ -0,0 +1,192 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "epoch": 1,
4
+ "train_loss": 0.6694036747442733,
5
+ "accuracy": 0.6384877920714098,
6
+ "precision": 0.6396098788152397,
7
+ "recall": 0.6384877920714098,
8
+ "f1": 0.6361877622449531,
9
+ "roc_auc": 0.6905002636963176,
10
+ "confusion_matrix": [
11
+ [
12
+ 20710,
13
+ 16304
14
+ ],
15
+ [
16
+ 11236,
17
+ 27930
18
+ ]
19
+ ]
20
+ },
21
+ {
22
+ "epoch": 2,
23
+ "train_loss": 0.5936620373939206,
24
+ "accuracy": 0.712050406930953,
25
+ "precision": 0.7125221690771839,
26
+ "recall": 0.712050406930953,
27
+ "f1": 0.711412016449667,
28
+ "roc_auc": 0.7885406459400497,
29
+ "confusion_matrix": [
30
+ [
31
+ 24715,
32
+ 12299
33
+ ],
34
+ [
35
+ 9637,
36
+ 29529
37
+ ]
38
+ ]
39
+ },
40
+ {
41
+ "epoch": 3,
42
+ "train_loss": 0.5445732520047349,
43
+ "accuracy": 0.7466264111315306,
44
+ "precision": 0.7470089523551816,
45
+ "recall": 0.7466264111315306,
46
+ "f1": 0.7466769304831232,
47
+ "roc_auc": 0.827907726657352,
48
+ "confusion_matrix": [
49
+ [
50
+ 27886,
51
+ 9128
52
+ ],
53
+ [
54
+ 10174,
55
+ 28992
56
+ ]
57
+ ]
58
+ },
59
+ {
60
+ "epoch": 4,
61
+ "train_loss": 0.5105602312859151,
62
+ "accuracy": 0.7633368338146496,
63
+ "precision": 0.7633707551299388,
64
+ "recall": 0.7633368338146496,
65
+ "f1": 0.7633502128901711,
66
+ "roc_auc": 0.8458466395889389,
67
+ "confusion_matrix": [
68
+ [
69
+ 28082,
70
+ 8932
71
+ ],
72
+ [
73
+ 9097,
74
+ 30069
75
+ ]
76
+ ]
77
+ },
78
+ {
79
+ "epoch": 5,
80
+ "train_loss": 0.48103076480012597,
81
+ "accuracy": 0.7764767655552639,
82
+ "precision": 0.7766028854553568,
83
+ "recall": 0.7764767655552639,
84
+ "f1": 0.7762931616528126,
85
+ "roc_auc": 0.8607994758196372,
86
+ "confusion_matrix": [
87
+ [
88
+ 27822,
89
+ 9192
90
+ ],
91
+ [
92
+ 7836,
93
+ 31330
94
+ ]
95
+ ]
96
+ },
97
+ {
98
+ "epoch": 6,
99
+ "train_loss": 0.45285762216234077,
100
+ "accuracy": 0.7863349960619586,
101
+ "precision": 0.786550974702636,
102
+ "recall": 0.7863349960619586,
103
+ "f1": 0.7863729433058431,
104
+ "roc_auc": 0.8707636359308418,
105
+ "confusion_matrix": [
106
+ [
107
+ 29235,
108
+ 7779
109
+ ],
110
+ [
111
+ 8498,
112
+ 30668
113
+ ]
114
+ ]
115
+ },
116
+ {
117
+ "epoch": 7,
118
+ "train_loss": 0.42440437866778397,
119
+ "accuracy": 0.7881333683381465,
120
+ "precision": 0.7919868312451798,
121
+ "recall": 0.7881333683381465,
122
+ "f1": 0.7878603466154905,
123
+ "roc_auc": 0.8755943217566788,
124
+ "confusion_matrix": [
125
+ [
126
+ 30951,
127
+ 6063
128
+ ],
129
+ [
130
+ 10077,
131
+ 29089
132
+ ]
133
+ ]
134
+ },
135
+ {
136
+ "epoch": 8,
137
+ "train_loss": 0.3885842357222038,
138
+ "accuracy": 0.7932528222630612,
139
+ "precision": 0.7932259994398904,
140
+ "recall": 0.7932528222630612,
141
+ "f1": 0.7932070934632307,
142
+ "roc_auc": 0.8779366478685279,
143
+ "confusion_matrix": [
144
+ [
145
+ 28896,
146
+ 8118
147
+ ],
148
+ [
149
+ 7632,
150
+ 31534
151
+ ]
152
+ ]
153
+ },
154
+ {
155
+ "epoch": 9,
156
+ "train_loss": 0.3425717065478533,
157
+ "accuracy": 0.7886584405355737,
158
+ "precision": 0.7887057068927522,
159
+ "recall": 0.7886584405355737,
160
+ "f1": 0.7885404112089442,
161
+ "roc_auc": 0.8750088332658279,
162
+ "confusion_matrix": [
163
+ [
164
+ 28454,
165
+ 8560
166
+ ],
167
+ [
168
+ 7540,
169
+ 31626
170
+ ]
171
+ ]
172
+ },
173
+ {
174
+ "epoch": 10,
175
+ "train_loss": 0.2898458522586866,
176
+ "accuracy": 0.7866631661853505,
177
+ "precision": 0.7867469948727277,
178
+ "recall": 0.7866631661853505,
179
+ "f1": 0.7865210069555763,
180
+ "roc_auc": 0.8706981988520192,
181
+ "confusion_matrix": [
182
+ [
183
+ 28306,
184
+ 8708
185
+ ],
186
+ [
187
+ 7544,
188
+ 31622
189
+ ]
190
+ ]
191
+ }
192
+ ]
embedding_dimension/embeddings/all-MiniLM-L6-v2_train.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9023267ad565f96cb00fcccff0da5a24d89061abc9614bcee8c55598fbbea85
3
+ size 936087680
embedding_dimension/embeddings/all-MiniLM-L6-v2_val.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:051b400e5796b31cf55302d8a2d92dd4b8164a3b8b37b5e7d7d41076047e119b
3
+ size 234023552
embedding_dimension/embeddings/all-mpnet-base-v2_train.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23fb9594c31b6cfc9b77ca183672306aada3beb35c01fd6630750794574840d2
3
+ size 1872175232
embedding_dimension/embeddings/all-mpnet-base-v2_val.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdceb5b0321b8d5e83d1941968daee4c5fd0f4f66a82614aead99d70743f842e
3
+ size 468046976
embedding_dimension/embeddings/e5-large-v2_train.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6315aaad2c95e01b2ecc667e782bc356f53641c8a51af46af110e40c6beef5f1
3
+ size 2496233600
embedding_dimension/embeddings/e5-large-v2_val.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fadbd0ccdba93db930502b8320c12b5cee76e9d304fd43db8b86619a26b1a6f3
3
+ size 624062592
gpt2_large_sarcasm_finetuned/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
gpt2_large_sarcasm_finetuned/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1164ba8e8dfb344b4ea4054b75350acfd547133eb286f2e4f5bfb3864d094437
3
+ size 3096259675
gpt2_large_sarcasm_finetuned/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
gpt2_large_sarcasm_finetuned/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50256": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": false,
16
+ "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
+ "extra_special_tokens": {},
19
+ "model_max_length": 1024,
20
+ "pad_token": "<|endoftext|>",
21
+ "tokenizer_class": "GPT2Tokenizer",
22
+ "unk_token": "<|endoftext|>"
23
+ }
gpt2_large_sarcasm_finetuned/training_history.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "epoch": 1,
4
+ "train_loss": 0.5006914643305583,
5
+ "train_acc": 0.7534943284165511,
6
+ "val_loss": 0.4765789475314265,
7
+ "val_acc": 0.770175899186138
8
+ },
9
+ {
10
+ "epoch": 2,
11
+ "train_loss": 0.41041818162335203,
12
+ "train_acc": 0.8138054961195197,
13
+ "val_loss": 0.485660226567899,
14
+ "val_acc": 0.7766211604095563
15
+ },
16
+ {
17
+ "epoch": 3,
18
+ "train_loss": 0.29277975993642885,
19
+ "train_acc": 0.8804642865998358,
20
+ "val_loss": 0.5854064271237948,
21
+ "val_acc": 0.7670517196114466
22
+ },
23
+ {
24
+ "epoch": 4,
25
+ "train_loss": 0.20523463562336314,
26
+ "train_acc": 0.9268914880303846,
27
+ "val_loss": 0.7716748032591614,
28
+ "val_acc": 0.7618928852717248
29
+ }
30
+ ]
gpt2_large_sarcasm_finetuned/training_metadata.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1,
3
+ "train_loss": 0.5006914643305583,
4
+ "train_acc": 0.7534943284165511,
5
+ "val_loss": 0.4765789475314265,
6
+ "val_acc": 0.770175899186138,
7
+ "model_config": {
8
+ "model_name": "gpt2-large",
9
+ "max_length": 128,
10
+ "num_labels": 2,
11
+ "use_parent_context": false
12
+ },
13
+ "training_config": {
14
+ "num_epochs": 6,
15
+ "batch_size": 8,
16
+ "learning_rate": 2e-05,
17
+ "warmup_steps": 500,
18
+ "weight_decay": 0.01,
19
+ "max_grad_norm": 1.0,
20
+ "gradient_accumulation_steps": 2,
21
+ "bf16": true,
22
+ "fp16": false,
23
+ "dataloader_num_workers": 4,
24
+ "save_steps": 1000,
25
+ "eval_steps": 500,
26
+ "logging_steps": 100,
27
+ "save_total_limit": 3,
28
+ "validation_split": 0.1,
29
+ "early_stopping_patience": 3,
30
+ "seed": 42
31
+ }
32
+ }
gpt2_large_sarcasm_finetuned/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
gpt2_small_sarcasm_finetuned/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
gpt2_small_sarcasm_finetuned/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b4324651096363dc3652ed6005ebdd7a8c1a5389d473155b4659fcd0e678ed6
3
+ size 497809691
gpt2_small_sarcasm_finetuned/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
gpt2_small_sarcasm_finetuned/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50256": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
+ },
14
+ "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": false,
16
+ "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
+ "extra_special_tokens": {},
19
+ "model_max_length": 1024,
20
+ "pad_token": "<|endoftext|>",
21
+ "tokenizer_class": "GPT2Tokenizer",
22
+ "unk_token": "<|endoftext|>"
23
+ }
gpt2_small_sarcasm_finetuned/training_history.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "epoch": 1,
4
+ "train_loss": 0.5349204899357074,
5
+ "train_acc": 0.7375567194808718,
6
+ "val_loss": 0.494891486287505,
7
+ "val_acc": 0.7595825676030454
8
+ },
9
+ {
10
+ "epoch": 2,
11
+ "train_loss": 0.465825739455718,
12
+ "train_acc": 0.7769887005896929,
13
+ "val_loss": 0.5023922561388189,
14
+ "val_acc": 0.7702152796009452
15
+ },
16
+ {
17
+ "epoch": 3,
18
+ "train_loss": 0.4263289485021802,
19
+ "train_acc": 0.8026329799259346,
20
+ "val_loss": 0.5046512643927501,
21
+ "val_acc": 0.7708453662378577
22
+ },
23
+ {
24
+ "epoch": 4,
25
+ "train_loss": 0.3876040897047569,
26
+ "train_acc": 0.8258998178267268,
27
+ "val_loss": 0.5368210367767664,
28
+ "val_acc": 0.7700446311367813
29
+ }
30
+ ]
gpt2_small_sarcasm_finetuned/training_metadata.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1,
3
+ "train_loss": 0.5349204899357074,
4
+ "train_acc": 0.7375567194808718,
5
+ "val_loss": 0.494891486287505,
6
+ "val_acc": 0.7595825676030454,
7
+ "model_config": {
8
+ "model_name": "gpt2",
9
+ "max_length": 128,
10
+ "num_labels": 2,
11
+ "use_parent_context": false
12
+ },
13
+ "training_config": {
14
+ "num_epochs": 6,
15
+ "batch_size": 16,
16
+ "learning_rate": 2e-05,
17
+ "warmup_steps": 500,
18
+ "weight_decay": 0.01,
19
+ "max_grad_norm": 1.0,
20
+ "gradient_accumulation_steps": 1,
21
+ "bf16": true,
22
+ "fp16": false,
23
+ "dataloader_num_workers": 4,
24
+ "save_steps": 1000,
25
+ "eval_steps": 500,
26
+ "logging_steps": 100,
27
+ "save_total_limit": 3,
28
+ "validation_split": 0.1,
29
+ "early_stopping_patience": 3,
30
+ "seed": 42
31
+ }
32
+ }
gpt2_small_sarcasm_finetuned/vocab.json ADDED
The diff for this file is too large to render. See raw diff