|
|
--- |
|
|
license: mit |
|
|
language: |
|
|
- cs |
|
|
base_model: |
|
|
- nlptown/bert-base-multilingual-uncased-sentiment |
|
|
pipeline_tag: text-classification |
|
|
--- |
|
|
|
|
|
# Sentiment Analysis Model |
|
|
### This model is used in our transcription service, where the audio is first transcribed and then analysed via this model. |
|
|
|
|
|
The model expects a sentence and return a number from 1 to 5 where 1 is the most negative sentiment and 5 is the most positive one. There is a parsing present that checks the confidence and if it is below 0.7, it checks for the second most probable result, averages them and uses math.ceil for optimistic behavior. |
|
|
|
|
|
The model is trained on BERT (nlptown/bert-base-multilingual-uncased-sentiment), which has an MIT license, and distilled llm results. |
|
|
|
|
|
This model was trained for 20 epochs where the result is: |
|
|
|
|
|
| | Precision | Recall | F1-score | Support | |
|
|
|--------------|-----------|--------|----------|---------| |
|
|
| **Class 1** | 0.95 | 0.88 | 0.92 | 43 | |
|
|
| **Class 2** | 0.78 | 0.86 | 0.82 | 37 | |
|
|
| **Class 3** | 0.80 | 0.72 | 0.76 | 39 | |
|
|
| **Class 4** | 0.79 | 0.88 | 0.83 | 66 | |
|
|
| **Class 5** | 0.85 | 0.78 | 0.81 | 45 | |
|
|
| | | | | | |
|
|
| **Accuracy** | | | 0.83 | 230 | |
|
|
| **Macro avg**| 0.84 | 0.82 | 0.83 | 230 | |
|
|
| **Weighted avg** | 0.83 | 0.83 | 0.83 | 230 | |
|
|
|
|
|
## sentiment_model_6: |
|
|
| Version | Changelog | |
|
|
|---------|------------------------------------------------------| |
|
|
| **1.0** | initial training | |
|
|
| **1.1** | fine-tuning time and datetime to a neutral sentiment | |
|
|
| **1.2** | fine-tuning numbers to a neutral sentiment | |
|
|
|