--- language: - en pipeline_tag: text-classification tags: - sentiment-analysis - text-classification - opinion-mining - emotion-detection - nlp - natural-language-processing - transformers - peft - lora - adapter - fine-tuning - gemma - gemma-2b - base_model:google/gemma-2b - base_model:adapter:google/gemma-2b library_name: peft base_model: google/gemma-2b license: apache-2.0 model-index: - name: Sentiment Analyzer (LoRA Gemma-2B) results: - task: type: text-classification name: Sentiment Analysis metrics: - type: accuracy value: not-reported --- # Sentiment Analyzer (LoRA Fine-Tuned Gemma-2B) ## Model Overview **Sentiment Analyzer** is a **LoRA fine-tuned Gemma-2B transformer model** for **sentiment analysis and text classification** tasks. It uses **PEFT (Parameter-Efficient Fine-Tuning)** to deliver strong performance while keeping memory and compute requirements low. This model is well-suited for: - Sentiment analysis - Opinion mining - Review classification - Emotion-aware text generation - Lightweight NLP deployments --- ## Tasks - Text Classification - Sentiment Analysis --- ## Model Details - **Developed by:** `mysmmurf12` - **Shared by:** `mysmmurf12` - **Model type:** Transformer-based Language Model - **Base model:** `google/gemma-2b` - **Fine-tuning method:** LoRA (Low-Rank Adaptation) - **Library:** PEFT + Transformers - **Language:** English - **License:** Apache 2.0 (inherits from base model) --- ## Model Sources - **Hugging Face Repository:** https://huggingface.co/mysmmurf12/sentiment-analyzer - **Base Model:** https://huggingface.co/google/gemma-2b --- ## Intended Uses ### ✅ Direct Use - Sentiment classification (positive / negative / neutral) - Customer feedback and review analysis - Social media sentiment monitoring - Sentiment-aware chatbots ### 🔁 Downstream Use - Integration into RAG pipelines - Domain-specific sentiment fine-tuning - Deployment via APIs, Streamlit apps, or dashboards ### 🚫 Out-of-Scope Use - Medical, legal, or financial decision-making - High-risk automated moderation - Multilingual sentiment tasks (English-focused) --- ## Bias, Risks, and Limitations - May reflect biases present in training data - Less reliable on sarcasm or ambiguous language - Not evaluated on standardized sentiment benchmarks **Recommendation:** Use human validation for high-impact applications. --- ## How to Use the Model ### Load with Transformers + PEFT ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel base_model = "google/gemma-2b" adapter_model = "mysmmurf12/sentiment-analyzer" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained(base_model) model = PeftModel.from_pretrained(model, adapter_model) text = "The product quality is amazing!" inputs = tokenizer(text, return_tensors="pt") outputs = model.generate( **inputs, max_new_tokens=50, temperature=0.7 ) print(tokenizer.decode(outputs[0], skip_special_tokens=True))