|
|
--- |
|
|
datasets: |
|
|
- yelp_review_full |
|
|
language: |
|
|
- en |
|
|
metrics: |
|
|
- accuracy |
|
|
- code_eval |
|
|
pipeline_tag: text-classification |
|
|
--- |
|
|
# Model Card for SentimentTensor |
|
|
|
|
|
This modelcard provides details about the SentimentTensor model, developed by Saish Shinde, for sentiment analysis using LSTM architecture. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
The SentimentTensor model is a deep learning model based on LSTM architecture, developed by Saish Shinde, for sentiment analysis tasks. It achieves an accuracy of 81% on standard evaluation datasets. The model is designed to classify text data into three categories: negative, neutral, and positive sentiments. |
|
|
|
|
|
- **Developed by:** Saish Shinde |
|
|
- **Model type:** LSTM-based Sequence Classification |
|
|
- **Language(s) (NLP):** English |
|
|
- **License:** No specific license |
|
|
|
|
|
|
|
|
|
|
|
# Dataset Used |
|
|
|
|
|
yelp dataset with 4.04GB compressed,8.65GB uncompressed data |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
The SentimentTensor model can be directly used for sentiment analysis tasks without fine-tuning. |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
This model can be fine-tuned for specific domains or integrated into larger NLP applications. |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
The model may not perform well on highly specialized or domain-specific text data. |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
The SentimentTensor model, like any LSTM-based model, may have biases and limitations inherent in its training data and architecture. It might sometimes struggle with capturing long-range dependencies or understanding context in complex sentences, also it emphasizes less on neutral sentiment |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
Users should be aware of potential biases and limitations and evaluate results accordingly. |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
### Loading the Model |
|
|
|
|
|
You can load the SentimentTensor model using the Hugging Face library: |
|
|
|
|
|
# python Code: |
|
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer |
|
|
|
|
|
# Load the model and tokenizer |
|
|
model = AutoModelForSequenceClassification.from_pretrained("your-model-name") |
|
|
tokenizer = AutoTokenizer.from_pretrained("your-tokenizer-name") |
|
|
|
|
|
# Tokenization |
|
|
text = "Your text data here" |
|
|
tokenized_input = tokenizer(text, return_tensors="pt") |
|
|
|
|
|
# Sentiment Analysis |
|
|
#Forward pass through the model |
|
|
outputs = model(**tokenized_input) |
|
|
|
|
|
#Get predicted sentiment label |
|
|
predicted_label = outputs.logits.argmax().item() |
|
|
|
|
|
# Example Usage |
|
|
```python |
|
|
|
|
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer |
|
|
|
|
|
# Load the model and tokenizer |
|
|
model = AutoModelForSequenceClassification.from_pretrained("saishshinde15/SentimentTensor") |
|
|
tokenizer = AutoTokenizer.from_pretrained("saishshinde15/SentimentTensor") |
|
|
|
|
|
# Tokenize text data |
|
|
text = "This is a great movie!" |
|
|
tokenized_input = tokenizer(text, return_tensors="pt") |
|
|
|
|
|
# Perform sentiment analysis |
|
|
outputs = model(**tokenized_input) |
|
|
predicted_label = outputs.logits.argmax().item() |
|
|
|
|
|
# Print predicted sentiment |
|
|
sentiment_labels = ["negative", "neutral", "positive"] |
|
|
print(f"Predicted Sentiment: {sentiment_labels[predicted_label]}") |
|
|
|
|
|
|
|
|
``` |
|
|
# Model Architecture and Objective |
|
|
|
|
|
The SentimentTensor model is based on LSTM architecture, which is well-suited for sequence classification tasks like sentiment analysis. It uses long short-term memory cells to capture dependencies in sequential data. |
|
|
|
|
|
# Model Card Authors |
|
|
Saish Shinde |