File size: 3,084 Bytes
cdde307 c67fff6 cdde307 616103c cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 616103c cdde307 616103c c67fff6 616103c cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 616103c cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 616103c c67fff6 cdde307 c67fff6 cdde307 c67fff6 616103c cdde307 c67fff6 cdde307 c67fff6 616103c cdde307 c67fff6 cdde307 c67fff6 cdde307 616103c cdde307 c67fff6 616103c cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 cdde307 c67fff6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 |
---
language:
- en
pipeline_tag: text-classification
tags:
- sentiment-analysis
- text-classification
- opinion-mining
- emotion-detection
- nlp
- natural-language-processing
- transformers
- peft
- lora
- adapter
- fine-tuning
- gemma
- gemma-2b
- base_model:google/gemma-2b
- base_model:adapter:google/gemma-2b
library_name: peft
base_model: google/gemma-2b
license: apache-2.0
model-index:
- name: Sentiment Analyzer (LoRA Gemma-2B)
results:
- task:
type: text-classification
name: Sentiment Analysis
metrics:
- type: accuracy
value: not-reported
---
# Sentiment Analyzer (LoRA Fine-Tuned Gemma-2B)
## Model Overview
**Sentiment Analyzer** is a **LoRA fine-tuned Gemma-2B transformer model** for **sentiment analysis and text classification** tasks.
It uses **PEFT (Parameter-Efficient Fine-Tuning)** to deliver strong performance while keeping memory and compute requirements low.
This model is well-suited for:
- Sentiment analysis
- Opinion mining
- Review classification
- Emotion-aware text generation
- Lightweight NLP deployments
---
## Tasks
- Text Classification
- Sentiment Analysis
---
## Model Details
- **Developed by:** `mysmmurf12`
- **Shared by:** `mysmmurf12`
- **Model type:** Transformer-based Language Model
- **Base model:** `google/gemma-2b`
- **Fine-tuning method:** LoRA (Low-Rank Adaptation)
- **Library:** PEFT + Transformers
- **Language:** English
- **License:** Apache 2.0 (inherits from base model)
---
## Model Sources
- **Hugging Face Repository:**
https://huggingface.co/mysmmurf12/sentiment-analyzer
- **Base Model:**
https://huggingface.co/google/gemma-2b
---
## Intended Uses
### ✅ Direct Use
- Sentiment classification (positive / negative / neutral)
- Customer feedback and review analysis
- Social media sentiment monitoring
- Sentiment-aware chatbots
### 🔁 Downstream Use
- Integration into RAG pipelines
- Domain-specific sentiment fine-tuning
- Deployment via APIs, Streamlit apps, or dashboards
### 🚫 Out-of-Scope Use
- Medical, legal, or financial decision-making
- High-risk automated moderation
- Multilingual sentiment tasks (English-focused)
---
## Bias, Risks, and Limitations
- May reflect biases present in training data
- Less reliable on sarcasm or ambiguous language
- Not evaluated on standardized sentiment benchmarks
**Recommendation:**
Use human validation for high-impact applications.
---
## How to Use the Model
### Load with Transformers + PEFT
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
base_model = "google/gemma-2b"
adapter_model = "mysmmurf12/sentiment-analyzer"
tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, adapter_model)
text = "The product quality is amazing!"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=50,
temperature=0.7
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|