arif481/crosslingual-sentiment-model

A cross-lingual sentiment analysis model fine-tuned on XLM-RoBERTa for binary sentiment classification (positive/negative) across en, bn.

Model Description

This model performs sentiment classification across multiple languages using transfer learning. It was trained using the combined strategy.

Supported Languages

  • English (en)
  • Bengali (bn)

Training Mode: combined

Trained on combined English and Bengali data for multilingual learning.

Usage

from transformers import pipeline

classifier = pipeline("sentiment-analysis", model="arif481/crosslingual-sentiment-model")

# English
result = classifier("This movie is absolutely fantastic!")
print(result)  # [{'label': 'positive', 'score': 0.99}]

# Bengali
result = classifier("এই সিনেমাটি অসাধারণ ছিল!")
print(result)  # [{'label': 'positive', 'score': 0.95}]

Training

from transformers import AutoModelForSequenceClassification, AutoTokenizer

model = AutoModelForSequenceClassification.from_pretrained("arif481/crosslingual-sentiment-model")
tokenizer = AutoTokenizer.from_pretrained("arif481/crosslingual-sentiment-model")

Metrics

Metric Value
Accuracy N/A
Macro F1 N/A
Precision N/A
Recall N/A

Limitations

  • Binary classification only (positive/negative)
  • May not perform well on neutral sentiment
  • Bengali performance may be lower than English due to limited training data

Citation

If you use this model, please cite:

@misc{crosslingual-sentiment,
  author = {Cross-Lingual Sentiment Team},
  title = {Cross-Lingual Sentiment Analysis Model},
  year = {2024},
  publisher = {Hugging Face},
  url = {https://huggingface.co/arif481/crosslingual-sentiment-model}
}

License

This model is released under the MIT License.

Downloads last month
18
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train arif481/crosslingual-sentiment-model

Space using arif481/crosslingual-sentiment-model 1