pujpuj's picture
Update README.md
5955e04 verified
---
library_name: transformers
tags:
- sentiment-analysis
- lora
- roberta
- fine-tuned
- insurance
---
# Model Card for RoBERTa LoRA Fine-Tuned for Insurance Review Rating
This model is a fine-tuned version of RoBERTa (`roberta-large`) using LoRA adapters. It is specifically designed to classify English insurance reviews and assign a rating (on a scale of 1 to 5).
## Model Details
### Model Description
This model uses RoBERTa (`roberta-large`) as its base architecture and was fine-tuned using Low-Rank Adaptation (LoRA) to adapt efficiently to the task of insurance review classification. The model predicts a rating from 1 to 5 based on the sentiment and context of a given review. LoRA fine-tuning reduces memory overhead and enables faster training compared to full fine-tuning.
- **Developed by:** Lapujpuj
- **Finetuned from model:** RoBERTa (`roberta-large`)
- **Language(s) (NLP):** English
- **License:** Apache-2.0
- **LoRA Configuration:**
- Rank (r): 2
- LoRA Alpha: 16
- LoRA Dropout: 0.1
- **Task:** Sentiment-based rating prediction for insurance reviews
### Model Sources
- **Repository:** [pujpuj/roberta-lora-token-classification](https://huggingface.co/pujpuj/roberta-lora-token-classification)
- **Demo:** N/A
---
## Uses
### Direct Use
This model can be directly used to assign a sentiment-based rating to insurance reviews. Input text is expected to be a sentence or paragraph in English.
### Downstream Use
The model can be used as a building block for larger applications, such as customer feedback analysis, satisfaction prediction, or insurance service improvement.
### Out-of-Scope Use
- The model is not designed for reviews in languages other than English.
- It may not generalize well to domains outside of insurance-related reviews.
- Avoid using the model for biased or malicious predictions.
---
## Bias, Risks, and Limitations
### Bias
- The model is trained on a specific dataset of insurance reviews, which might include biases present in the training data (e.g., skewed ratings, linguistic or cultural biases).
### Risks
- Predictions might not generalize well to other domains or review styles.
- Inconsistent predictions may occur for ambiguous or mixed reviews.
### Recommendations
- Always validate model outputs before making decisions.
- Use the model in conjunction with other tools for a more comprehensive analysis.
---
## How to Get Started with the Model
You can use the model with the following code snippet:
```python
from transformers import AutoTokenizer
from peft import PeftModel
# Load tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("roberta-large")
base_model = AutoModelForSequenceClassification.from_pretrained("roberta-large", num_labels=5)
model = PeftModel.from_pretrained(base_model, "pujpuj/roberta-lora-token-classification")
# Example prediction
review = "The insurance service was quick and reliable."
inputs = tokenizer(review, return_tensors="pt", truncation=True, padding=True)
outputs = model(**inputs)
rating = torch.argmax(outputs.logits, dim=1).item() + 1
print(f"Predicted rating: {rating}")