| license: cc-by-4.0 | |
| language: | |
| - en | |
| extra_gated_fields: | |
| Name: text | |
| Country: country | |
| Institution: text | |
| Institution Email: text | |
| Please specify your academic use case: text | |
| extra_gated_prompt: Our models are intended for academic use only. If you are not | |
| affiliated with an academic institution, please provide a rationale for using our | |
| models. Please allow us a few business days to manually review subscriptions. | |
| Model description Cased fine-tuned BERT model for English, trained on (manually annotated) Hungarian parliamentary speeches scraped from parlament.hu, and translated with Google Translate API. | |
| Intended uses & limitations The model can be used as any other (cased) BERT model. It has been tested recognizing positive, negative, and neutral sentences in (parliamentary) pre-agenda speeches, where: | |
| 'Label_0': Negative 'Label_1': Neutral 'Label_2': Positive | |
| Training The fine-tuned version of the original bert-base-cased model (bert-base-cased), trained on HunEmPoli corpus, translated with Google Translate API. | |
| Intended uses & limitations: The model can be used as any other (cased) BERT model. | |
| Eval results | |
| precision recall f1-score support | |
| 0 0.87 0.87 0.87 1118 | |
| 1 1.00 0.26 0.41 35 | |
| 2 0.78 0.82 0.80 748 | |
| accuracy 0.83 1901 | |
| macro avg 0.88 0.65 0.69 1901 | |
| weighted avg 0.84 0.83 0.83 1901 | |