|
|
--- |
|
|
library_name: transformers |
|
|
datasets: |
|
|
- kallantis/Greek-Humorous-Dataset |
|
|
language: |
|
|
- el |
|
|
pipeline_tag: text-classification |
|
|
--- |
|
|
# |
|
|
# |
|
|
# The model is based on XLM-RoBERTa large ("xlm-roberta-large") fine-tuned for Humor Recognition in Greek language. |
|
|
# |
|
|
|
|
|
|
|
|
|
|
|
## Model Details |
|
|
|
|
|
The model was pre-trained over 10 epochs on Greek Humorous Dataset |
|
|
# |
|
|
## Pre-processing details |
|
|
|
|
|
The text needs to be pre-processed by removing all greek diacritics and punctuation |
|
|
and converting all letters to lowercase |
|
|
|
|
|
## Load Pretrained Model |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, XLMRobertaForSequenceClassification |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained("kallantis/Humor-Recognition-Greek-XLM-R-large") |
|
|
model = XLMRobertaForSequenceClassification.from_pretrained("kallantis/Humor-Recognition-Greek-XLM-R-large", num_labels=2, ignore_mismatched_sizes=True) |
|
|
``` |
|
|
|