Add config and tokenizer files for HF compatibility
Browse files- README.md +18 -0
- config.json +28 -0
- model.safetensors +3 -0
- optimizer.pt +3 -0
- rng_state.pth +3 -0
- scaler.pt +3 -0
- scheduler.pt +3 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
README.md
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# RoBERTa Multitask Model
|
| 2 |
+
|
| 3 |
+
This is a multitask model based on RoBERTa, trained for:
|
| 4 |
+
|
| 5 |
+
- Sentiment Analysis (negative, neutral, positive)
|
| 6 |
+
- Emotion Classification (anger, joy, optimism, sadness)
|
| 7 |
+
- Hate Speech Detection (hate, offensive, neither)
|
| 8 |
+
|
| 9 |
+
It uses a shared encoder with task-specific heads.
|
| 10 |
+
|
| 11 |
+
## Usage
|
| 12 |
+
|
| 13 |
+
```python
|
| 14 |
+
from models.multitask_model import MultiTaskModel
|
| 15 |
+
from transformers import RobertaTokenizer
|
| 16 |
+
|
| 17 |
+
tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
|
| 18 |
+
model = MultiTaskModel.from_pretrained("SeragAmin/RoBERTa_Multitask")
|
config.json
ADDED
|
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": ["MultiTaskModel"],
|
| 3 |
+
"model_type": "roberta",
|
| 4 |
+
"base_model": "roberta-base",
|
| 5 |
+
"use_lora": true,
|
| 6 |
+
"lora_config": {
|
| 7 |
+
"r": 16,
|
| 8 |
+
"lora_alpha": 32,
|
| 9 |
+
"target_modules": ["query", "value"],
|
| 10 |
+
"lora_dropout": 0.05,
|
| 11 |
+
"bias": "none",
|
| 12 |
+
"task_type": "SEQ_CLS"
|
| 13 |
+
},
|
| 14 |
+
"hidden_size": 768,
|
| 15 |
+
"intermediate_size": 128,
|
| 16 |
+
"dropout_prob": 0.1,
|
| 17 |
+
"task_name_to_id": {
|
| 18 |
+
"sentiment": 0,
|
| 19 |
+
"hate": 1,
|
| 20 |
+
"emotion": 2
|
| 21 |
+
},
|
| 22 |
+
"task_num_labels": {
|
| 23 |
+
"sentiment": 3,
|
| 24 |
+
"hate": 3,
|
| 25 |
+
"emotion": 4
|
| 26 |
+
}
|
| 27 |
+
}
|
| 28 |
+
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:020cc987573773d55e578acc499efbaa2e97cf15b8e0ade1e0e0eced49bc9370
|
| 3 |
+
size 502161328
|
optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0731cb0129cef3a3c2e8f15b0051ce31e5bea4654cc47be6f9e6635171e67970
|
| 3 |
+
size 7142539
|
rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bed952ca4bd18446707301d6dc2c60672651e15ff51d9bcbbbb45262bfedcbfa
|
| 3 |
+
size 14645
|
scaler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:656bc26616be85f01bf18333ac653a8f96471d86c57f9621a32619e5e2749696
|
| 3 |
+
size 1383
|
scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d61ba0149b751d1e39c026ac73851376897cba55c305506a2182a5c57fa3c002
|
| 3 |
+
size 1401
|
trainer_state.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a845e43e718d3bc9f589a4a0f4dcdbf0a25283e0c6196f0754e1890d839007e
|
| 3 |
+
size 5777
|