Shoriful025's picture
Create README.md
d4ee05d verified
# sentiment_analysis_bert_multilingual
## Overview
This model is a fine-tuned version of the Multilingual BERT (mBERT) base model. It is designed to classify the sentiment of text across 100+ languages into three categories: Negative, Neutral, and Positive.
## Model Architecture
The model utilizes the standard BERT-base architecture:
- **Layers**: 12 Transformer blocks
- **Hidden Size**: 768
- **Attention Heads**: 12
- **Parameters**: ~177M
It includes a sequence classification head on top of the hidden state of the `[CLS]` token.
## Intended Use
- Social media monitoring for global brands.
- Customer feedback analysis in multilingual support tickets.
- Market research across different geographical regions.
## Limitations
- **Context Window**: Limited to 512 tokens; longer texts will be truncated.
- **Sarcasm**: May struggle with highly idiomatic or sarcastic expressions in low-resource languages.
- **Bias**: Subject to biases present in the Wikipedia and BookCorpus datasets used for pre-training.