| | --- |
| | inference: false |
| | language: |
| | - bg |
| | license: mit |
| | datasets: |
| | - oscar |
| | - chitanka |
| | - wikipedia |
| | tags: |
| | - torch |
| | --- |
| | |
| | # ROBERTA BASE (cased) trained on private Bulgarian sentiment-analysis dataset |
| | This is a Multilingual Roberta model. |
| |
|
| | This model is cased: it does make a difference between bulgarian and Bulgarian. |
| |
|
| | ### How to use |
| |
|
| | Here is how to use this model in PyTorch: |
| |
|
| | ```python |
| | >>> import torch |
| | >>> from transformers import AutoModel, AutoTokenizer |
| | >>> |
| | >>> model_id = "rmihaylov/roberta-base-sentiment-bg" |
| | >>> model = AutoModel.from_pretrained(model_id, trust_remote_code=True) |
| | >>> tokenizer = AutoTokenizer.from_pretrained(model_id) |
| | >>> |
| | >>> inputs = tokenizer.batch_encode_plus(['Това е умно.', 'Това е тъпо.'], return_tensors='pt') |
| | >>> outputs = model(**inputs) |
| | >>> torch.softmax(outputs, dim=1).tolist() |
| | |
| | [[0.0004746630438603461, 0.9995253086090088], |
| | [0.9986956715583801, 0.0013043134240433574]] |
| | ``` |
| |
|