mtreviso aychang commited on
Commit
c861dd6
·
0 Parent(s):

Duplicate from aychang/roberta-base-imdb

Browse files

Co-authored-by: Andrew Chang <aychang@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ *.bin.* filter=lfs diff=lfs merge=lfs -text
2
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.h5 filter=lfs diff=lfs merge=lfs -text
5
+ *.tflite filter=lfs diff=lfs merge=lfs -text
6
+ *.tar.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.ot filter=lfs diff=lfs merge=lfs -text
8
+ *.onnx filter=lfs diff=lfs merge=lfs -text
9
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ thumbnail: null
5
+ tags:
6
+ - text-classification
7
+ license: mit
8
+ datasets:
9
+ - imdb
10
+ metrics: null
11
+ duplicated_from: aychang/roberta-base-imdb
12
+ ---
13
+
14
+ # IMDB Sentiment Task: roberta-base
15
+
16
+ ## Model description
17
+
18
+ A simple base roBERTa model trained on the "imdb" dataset.
19
+
20
+ ## Intended uses & limitations
21
+
22
+ #### How to use
23
+
24
+ ##### Transformers
25
+
26
+ ```python
27
+ # Load model and tokenizer
28
+ from transformers import AutoModelForSequenceClassification, AutoTokenizer
29
+
30
+ model = AutoModelForQuestionAnswering.from_pretrained(model_name)
31
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
32
+
33
+ # Use pipeline
34
+ from transformers import pipeline
35
+
36
+ model_name = "aychang/roberta-base-imdb"
37
+
38
+ nlp = pipeline("sentiment-analysis", model=model_name, tokenizer=model_name)
39
+
40
+ results = nlp(["I didn't really like it because it was so terrible.", "I love how easy it is to watch and get good results."])
41
+ ```
42
+
43
+ ##### AdaptNLP
44
+
45
+ ```python
46
+ from adaptnlp import EasySequenceClassifier
47
+
48
+ model_name = "aychang/roberta-base-imdb"
49
+ texts = ["I didn't really like it because it was so terrible.", "I love how easy it is to watch and get good results."]
50
+
51
+ classifer = EasySequenceClassifier
52
+ results = classifier.tag_text(text=texts, model_name_or_path=model_name, mini_batch_size=2)
53
+ ```
54
+
55
+ #### Limitations and bias
56
+
57
+ This is minimal language model trained on a benchmark dataset.
58
+
59
+ ## Training data
60
+
61
+ IMDB https://huggingface.co/datasets/imdb
62
+
63
+ ## Training procedure
64
+
65
+ #### Hardware
66
+ One V100
67
+
68
+ #### Hyperparameters and Training Args
69
+ ```python
70
+ from transformers import TrainingArguments
71
+
72
+ training_args = TrainingArguments(
73
+ output_dir='./models',
74
+ overwrite_output_dir=False,
75
+ num_train_epochs=2,
76
+ per_device_train_batch_size=8,
77
+ per_device_eval_batch_size=8,
78
+ warmup_steps=500,
79
+ weight_decay=0.01,
80
+ evaluation_strategy="steps",
81
+ logging_dir='./logs',
82
+ fp16=False,
83
+ eval_steps=800,
84
+ save_steps=300000
85
+ )
86
+ ```
87
+
88
+ ## Eval results
89
+
90
+ ```
91
+ {'epoch': 2.0,
92
+ 'eval_accuracy': 0.94668,
93
+ 'eval_f1': array([0.94603457, 0.94731017]),
94
+ 'eval_loss': 0.2578844428062439,
95
+ 'eval_precision': array([0.95762642, 0.93624502]),
96
+ 'eval_recall': array([0.93472, 0.95864]),
97
+ 'eval_runtime': 244.7522,
98
+ 'eval_samples_per_second': 102.144}
99
+ ```
config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-base",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "neg",
15
+ "1": "pos"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "neg": 0,
21
+ "pos": 1
22
+ },
23
+ "layer_norm_eps": 1e-05,
24
+ "max_position_embeddings": 514,
25
+ "model_type": "roberta",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "pad_token_id": 1,
29
+ "position_embedding_type": "absolute",
30
+ "transformers_version": "4.2.2",
31
+ "type_vocab_size": 1,
32
+ "use_cache": true,
33
+ "vocab_size": 50265
34
+ }
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22a774239b06ab74e5077174cbbcd71ffba586b4a39f2e48126bcb16d6aa83d9
3
+ size 498595901
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f3d8efc699649903c74d6da83c7492c1b0b0902c31bfd478c49a3640005888e
3
+ size 498677271
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "name_or_path": "roberta-base"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f6e5996d43ce5a597d7e8a3b2ec43fb2d935f732dc2b7ef09b669a41b7e701
3
+ size 1967
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff