SeragAmin commited on
Commit
bb38e03
·
1 Parent(s): 93506ba

Add config and tokenizer files for HF compatibility

Browse files
Files changed (9) hide show
  1. README.md +18 -0
  2. config.json +28 -0
  3. model.safetensors +3 -0
  4. optimizer.pt +3 -0
  5. rng_state.pth +3 -0
  6. scaler.pt +3 -0
  7. scheduler.pt +3 -0
  8. trainer_state.json +0 -0
  9. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # RoBERTa Multitask Model
2
+
3
+ This is a multitask model based on RoBERTa, trained for:
4
+
5
+ - Sentiment Analysis (negative, neutral, positive)
6
+ - Emotion Classification (anger, joy, optimism, sadness)
7
+ - Hate Speech Detection (hate, offensive, neither)
8
+
9
+ It uses a shared encoder with task-specific heads.
10
+
11
+ ## Usage
12
+
13
+ ```python
14
+ from models.multitask_model import MultiTaskModel
15
+ from transformers import RobertaTokenizer
16
+
17
+ tokenizer = RobertaTokenizer.from_pretrained("roberta-base")
18
+ model = MultiTaskModel.from_pretrained("SeragAmin/RoBERTa_Multitask")
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": ["MultiTaskModel"],
3
+ "model_type": "roberta",
4
+ "base_model": "roberta-base",
5
+ "use_lora": true,
6
+ "lora_config": {
7
+ "r": 16,
8
+ "lora_alpha": 32,
9
+ "target_modules": ["query", "value"],
10
+ "lora_dropout": 0.05,
11
+ "bias": "none",
12
+ "task_type": "SEQ_CLS"
13
+ },
14
+ "hidden_size": 768,
15
+ "intermediate_size": 128,
16
+ "dropout_prob": 0.1,
17
+ "task_name_to_id": {
18
+ "sentiment": 0,
19
+ "hate": 1,
20
+ "emotion": 2
21
+ },
22
+ "task_num_labels": {
23
+ "sentiment": 3,
24
+ "hate": 3,
25
+ "emotion": 4
26
+ }
27
+ }
28
+
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:020cc987573773d55e578acc499efbaa2e97cf15b8e0ade1e0e0eced49bc9370
3
+ size 502161328
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0731cb0129cef3a3c2e8f15b0051ce31e5bea4654cc47be6f9e6635171e67970
3
+ size 7142539
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bed952ca4bd18446707301d6dc2c60672651e15ff51d9bcbbbb45262bfedcbfa
3
+ size 14645
scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:656bc26616be85f01bf18333ac653a8f96471d86c57f9621a32619e5e2749696
3
+ size 1383
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d61ba0149b751d1e39c026ac73851376897cba55c305506a2182a5c57fa3c002
3
+ size 1401
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a845e43e718d3bc9f589a4a0f4dcdbf0a25283e0c6196f0754e1890d839007e
3
+ size 5777