|
|
--- |
|
|
base_model: unsloth/qwen3-14b-unsloth-bnb-4bit |
|
|
tags: |
|
|
- text-generation-inference |
|
|
- transformers |
|
|
- unsloth |
|
|
- qwen3 |
|
|
- trl |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
datasets: |
|
|
- mattwesney/CoT_Heartbreak_and_Breakups |
|
|
pipeline_tag: text-generation |
|
|
library_name: peft |
|
|
--- |
|
|
|
|
|
# Model Description |
|
|
|
|
|
- **Developed by:** khazarai |
|
|
- **License:** apache-2.0 |
|
|
- **Finetuned from model :** unsloth/qwen3-14b-unsloth-bnb-4bit |
|
|
|
|
|
This qwen3 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) |
|
|
|
|
|
|
|
|
This model is a QLoRA fine-tuned version of unsloth/qwen3-14b-unsloth-bnb-4bit, originally based on the Qwen3-14B architecture developed by the Qwen Team. |
|
|
The model has been fine-tuned on the Chain of Thought – Heartbreak & Breakups Dataset (MIT Licensed), consisting of 9.8k high-quality Q&A pairs focused on emotional processing, coping strategies, and relationship dynamics following breakups. |
|
|
The goal of this fine-tuning is to enhance: |
|
|
|
|
|
- Emotional reasoning capability |
|
|
- Structured chain-of-thought generation |
|
|
- Empathetic and psychologically grounded responses |
|
|
- Relationship pattern analysis |
|
|
- Identity reconstruction & self-esteem rebuilding guidance |
|
|
|
|
|
|
|
|
# 🧠 Base Model |
|
|
|
|
|
- Base architecture: Qwen3-14B |
|
|
- Variant: unsloth/qwen3-14b-unsloth-bnb-4bit |
|
|
- Quantization: 4-bit (bitsandbytes) |
|
|
- Fine-tuning method: QLoRA |
|
|
- Adapter type: LoRA |
|
|
- Training precision: 4-bit base + 16-bit adapters |
|
|
|
|
|
|
|
|
# 🎯 Intended Use |
|
|
|
|
|
This model is intended for: |
|
|
|
|
|
- Mental health–adjacent AI assistants |
|
|
- Relationship guidance systems |
|
|
- Emotional reasoning research |
|
|
- Chain-of-thought alignment experiments |
|
|
- NLP research on structured reasoning in affective domains |
|
|
|
|
|
The model aims to produce: |
|
|
|
|
|
- Step-by-step reasoning |
|
|
- Balanced perspectives |
|
|
- Reduced reactive or extreme advice |
|
|
|
|
|
⚠️ Limitations |
|
|
|
|
|
- Not a substitute for licensed therapy |
|
|
- May generate plausible but non-clinically validated advice |
|
|
- Trained on synthetic / curated emotional scenarios |
|
|
- Chain-of-thought exposure may increase verbosity |
|
|
- Emotional nuance outside breakup domain may be limited |
|
|
|
|
|
This model should not be used for crisis intervention or high-risk mental health scenarios. |
|
|
|
|
|
# How to get started with Model |
|
|
|
|
|
``` Python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
from peft import PeftModel |
|
|
|
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained("unsloth/qwen3-14b-unsloth-bnb-4bit") |
|
|
base_model = AutoModelForCausalLM.from_pretrained( |
|
|
"unsloth/qwen3-14b-unsloth-bnb-4bit", |
|
|
device_map={"": 0} |
|
|
) |
|
|
|
|
|
model = PeftModel.from_pretrained(base_model,"khazarai/Med-R1-14B") |
|
|
|
|
|
question = """ |
|
|
How can someone work through and move past deeply painful memories associated with trauma, understanding that "moving past" doesn't mean forgetting but rather integrating the experience in a healthy way? |
|
|
""" |
|
|
|
|
|
messages = [ |
|
|
{"role" : "user", "content" : question} |
|
|
] |
|
|
text = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
tokenize = False, |
|
|
add_generation_prompt = True, |
|
|
enable_thinking = True, |
|
|
) |
|
|
|
|
|
from transformers import TextStreamer |
|
|
_ = model.generate( |
|
|
**tokenizer(text, return_tensors = "pt").to("cuda"), |
|
|
max_new_tokens = 2048, |
|
|
temperature = 0.6, |
|
|
top_p = 0.95, |
|
|
top_k = 20, |
|
|
streamer = TextStreamer(tokenizer, skip_prompt = True), |
|
|
) |
|
|
``` |
|
|
|
|
|
# 🧪 Future Work |
|
|
|
|
|
- Domain expansion to broader emotional intelligence tasks |
|
|
- Controlled reasoning output (hidden CoT vs visible CoT) |
|
|
- Evaluation via human annotation |
|
|
- Cross-cultural emotional adaptation |
|
|
|