--- library_name: transformers tags: - gemma - psychology - mental-health - lora - instruction-tuning - neuroquestai license: mit datasets: - jkhedri/psychology-dataset language: - en metrics: - perplexity base_model: - google/gemma-2b-it pipeline_tag: text-generation --- # Model Card for Gemma-2b-it-Psych ## Model Summary **Gemma-2b-it-Psych** is a domain-adapted version of `google/gemma-2b-it`, fine-tuned using LoRA on an instruction-based psychology dataset. The model is optimized to generate **empathetic, supportive, and professionally aligned psychological responses**, primarily for educational and research purposes. This repository contains **LoRA adapters only**. The base model must be loaded separately. --- ## Model Details ### Model Description - **Author:** Ederson Corbari (e@NeuroQuest.ai) - **Date:** February 01, 2026 - **Model type:** Causal Language Model (LLM) - **Language(s):** English - **License:** Same as base model (`google/gemma-2b-it`) - **Finetuned from model:** `google/gemma-2b-it` - **Fine-tuning method:** LoRA / QLoRA (parameter-efficient fine-tuning) This model was fine-tuned using instruction–response pairs focused on psychological support. Only empathetic and therapeutically appropriate responses were retained during training, while judgmental or aggressive alternatives were excluded. --- ### Model Sources - **Hugging Face Repository:** https://huggingface.co/ecorbari/Gemma-2b-it-Psych - **GitHub Repository:** https://github.com/edersoncorbari/fine-tune-llm - **Base Model:** https://huggingface.co/google/gemma-2b-it --- ## Uses ### Direct Use This model is intended for: - Research and experimentation with instruction-tuned LLMs - Educational demonstrations of LoRA fine-tuning - Prompt engineering and behavioral analysis in psychology-related domains The model requires the base Gemma-2B weights to be loaded together with the LoRA adapters. --- ### Downstream Use - Further fine-tuning on related domains - Adapter merging to create a standalone model - Quantization for efficient local inference (e.g., GGUF formats) --- ### Out-of-Scope Use - Clinical diagnosis or treatment - Real-world mental health interventions without professional supervision - High-stakes decision-making - Autonomous counseling systems --- ## Bias, Risks, and Limitations - The model may generate inaccurate or incomplete information. - It does not replace licensed mental health professionals. - Responses may reflect biases present in the training data. - Empathy does not guarantee correctness or safety in all contexts. --- ### Recommendations Users should apply human oversight, especially in sensitive scenarios. This model is best suited for **research, learning, and proof-of-concept applications**. --- ## How to Get Started with the Model ```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel base_model = "google/gemma-2b-it" adapter_model = "ecorbari/Gemma-2b-it-Psych" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained( base_model, dtype=torch.float16, device_map="auto" ) model = PeftModel.from_pretrained(model, adapter_model) prompt = "How can I cope with anxiety during stressful situations?" inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate(**inputs, max_new_tokens=150) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Evaluation ### Testing Data The model was evaluated on a held-out validation split of the psychology instruction dataset used during fine-tuning. --- ### Metrics The following metrics were used to evaluate the model during training: - Cross-entropy loss (per token) - Perplexity (exp(loss)) --- ### Results | Metric | Value (approx.) | |------------|-----------------| | Eval Loss | 0.60 – 0.70 | | Perplexity | 1.8 – 2.0 | Perplexity was computed as the exponential of the evaluation loss. Lower values indicate higher confidence in next-token prediction. These metrics reflect convergence and generalization within the target domain, but do not directly assess clinical correctness or psychological safety.