Commit
·
d5d7be4
1
Parent(s):
6002c82
Initial model upload
Browse files- anger_logistic.pkl +3 -0
- bert_model.pt +3 -0
- bert_tokenizer/special_tokens_map.json +7 -0
- bert_tokenizer/tokenizer_config.json +58 -0
- bert_tokenizer/vocab.txt +0 -0
- brain dysfunction (forget)_logistic.pkl +3 -0
- emptiness_logistic.pkl +3 -0
- ganbert_model.pt +3 -0
- ganbert_tokenizer/special_tokens_map.json +7 -0
- ganbert_tokenizer/tokenizer_config.json +58 -0
- ganbert_tokenizer/vocab.txt +0 -0
- hopelessness_logistic.pkl +3 -0
- lightgbm/model_0.txt +0 -0
- lightgbm/model_1.txt +0 -0
- lightgbm/model_2.txt +0 -0
- lightgbm/model_3.txt +0 -0
- lightgbm/model_4.txt +0 -0
- lightgbm/model_5.txt +0 -0
- lightgbm/model_6.txt +0 -0
- lightgbm/model_7.txt +0 -0
- lightgbm/tfidf_vectorizer.joblib +3 -0
- loneliness_logistic.pkl +3 -0
- sadness_logistic.pkl +3 -0
- suicide intent_logistic.pkl +3 -0
- svm_model.joblib +3 -0
- worthlessness_logistic.pkl +3 -0
- xgboost/model_0.json +0 -0
- xgboost/model_1.json +0 -0
- xgboost/model_2.json +0 -0
- xgboost/model_3.json +0 -0
- xgboost/model_4.json +0 -0
- xgboost/model_5.json +0 -0
- xgboost/model_6.json +0 -0
- xgboost/model_7.json +0 -0
- xgboost/tfidf_vectorizer.joblib +3 -0
anger_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:46812b4833c179613a1a4c57176ac0e44aa931880b5e8e3c24169e8977792c93
|
| 3 |
+
size 699454
|
bert_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f3b137d58a09c0259a32afe0e9a43c3b7015757ff8cc25520fcab08a253c13bd
|
| 3 |
+
size 433347847
|
bert_tokenizer/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
bert_tokenizer/tokenizer_config.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_basic_tokenize": true,
|
| 47 |
+
"do_lower_case": false,
|
| 48 |
+
"extra_special_tokens": {},
|
| 49 |
+
"mask_token": "[MASK]",
|
| 50 |
+
"model_max_length": 512,
|
| 51 |
+
"never_split": null,
|
| 52 |
+
"pad_token": "[PAD]",
|
| 53 |
+
"sep_token": "[SEP]",
|
| 54 |
+
"strip_accents": null,
|
| 55 |
+
"tokenize_chinese_chars": true,
|
| 56 |
+
"tokenizer_class": "BertTokenizer",
|
| 57 |
+
"unk_token": "[UNK]"
|
| 58 |
+
}
|
bert_tokenizer/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
brain dysfunction (forget)_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c020782a47ce83e3305754bed10302136c6899a97a1776d4e5d7e9df15377581
|
| 3 |
+
size 699600
|
emptiness_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:324383bf98ef6c0e937d47736690d3c4b39d1ffb27460db2016b89225ec93263
|
| 3 |
+
size 699308
|
ganbert_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9b7d7990873f530fc830c1452e000f908a627e4cdce22f248f185809456010c
|
| 3 |
+
size 435715390
|
ganbert_tokenizer/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
ganbert_tokenizer/tokenizer_config.json
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_basic_tokenize": true,
|
| 47 |
+
"do_lower_case": false,
|
| 48 |
+
"extra_special_tokens": {},
|
| 49 |
+
"mask_token": "[MASK]",
|
| 50 |
+
"model_max_length": 512,
|
| 51 |
+
"never_split": null,
|
| 52 |
+
"pad_token": "[PAD]",
|
| 53 |
+
"sep_token": "[SEP]",
|
| 54 |
+
"strip_accents": null,
|
| 55 |
+
"tokenize_chinese_chars": true,
|
| 56 |
+
"tokenizer_class": "BertTokenizer",
|
| 57 |
+
"unk_token": "[UNK]"
|
| 58 |
+
}
|
ganbert_tokenizer/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
hopelessness_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a64a83dcd5d6cf7e9a90ac7156de2a4b79293482d94974fe4c6c017ab01db106
|
| 3 |
+
size 699512
|
lightgbm/model_0.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_1.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_2.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_3.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_4.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_5.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_6.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/model_7.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
lightgbm/tfidf_vectorizer.joblib
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:322d1dceb09a00795b3a94ead2973779bb23147168ee668a4aa6a5f39fe5b521
|
| 3 |
+
size 371043
|
loneliness_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f0060f388cce8950f229a696ba9f0d87564b421ca8407dca520691c792df669
|
| 3 |
+
size 699454
|
sadness_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac183406ffc8d6b89854f930fbf00baa75bc674bab18acc8c10fcf648ccf0fa7
|
| 3 |
+
size 699308
|
suicide intent_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:25fe34e9c6064f2a7caa9d5ed4165e24038e08a467166eea84e00164187a8595
|
| 3 |
+
size 699512
|
svm_model.joblib
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cecf826f1a16a64b8365d0030befcecb72b70ce05d784619bd20c8f1fd32a51
|
| 3 |
+
size 1014631
|
worthlessness_logistic.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:404d0d45b6918614b16d8fb447d48bdd6d6de23a5a06b5275f5a885ca7f8e32f
|
| 3 |
+
size 699512
|
xgboost/model_0.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_1.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_2.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_3.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_4.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_5.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_6.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/model_7.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
xgboost/tfidf_vectorizer.joblib
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:322d1dceb09a00795b3a94ead2973779bb23147168ee668a4aa6a5f39fe5b521
|
| 3 |
+
size 371043
|