File size: 3,089 Bytes
0b7ff23
11207b8
 
0b7ff23
11207b8
 
0b7ff23
 
11207b8
0b7ff23
11207b8
 
0b7ff23
 
3941312
11207b8
 
 
 
 
 
173c519
5584adf
11207b8
 
 
3941312
11207b8
 
3941312
11207b8
0b7ff23
11207b8
0b7ff23
11207b8
 
 
0b7ff23
3941312
11207b8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0b7ff23
 
 
11207b8
 
 
 
0b7ff23
 
11207b8
 
 
0b7ff23
 
11207b8
 
 
 
 
0b7ff23
11207b8
 
 
 
 
 
0b7ff23
11207b8
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
---
language: en
license: other
library_name: peft
base_model: Qwen/Qwen2.5-32B-Instruct
pipeline_tag: text-generation
tags:
- lora
- peft
- transformers
- qwen2.5
- conversational
---

# DeepSupport Warm ❤️‍🩹 - LoRA adapter
This repository provides a LoRA adapter for DeepSupport Warm, an emotional-holding companion that offers gentle reflection and warm support without rushing into what to do next.

## What it does ✨
DeepSupport Warm is designed to help users feel held and less alone in the moment:

- Validate and name feelings without judging  
- Stay with emotion first before problem-solving  
- Offer gentle grounding and a small next step only if you want 



## Quick start 🚀
### 1) Install
```bash
pip install transformers peft accelerate torch
```

### 2) Load base model and LoRA adapter
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel

base_id = "Qwen/Qwen2.5-32B-Instruct" # Base model we used. You may replace it with another compatible base model
lora_id = "Yukyin/deepsupport-warm-lora-oss"

tokenizer = AutoTokenizer.from_pretrained(base_id, trust_remote_code=True)
base = AutoModelForCausalLM.from_pretrained(
    base_id,
    torch_dtype=torch.float16,
    device_map="auto",
    trust_remote_code=True,
)

model = PeftModel.from_pretrained(base, lora_id)
model.eval()

messages = [
    {"role": "user", "content": "我最近压力很大,感觉自己一直在被否定。"},
]

inputs = tokenizer.apply_chat_template(
    messages,
    tokenize=True,
    add_generation_prompt=True,
    return_tensors="pt",
).to(model.device)

with torch.no_grad():
    out = model.generate(
        inputs,
        max_new_tokens=256,
        do_sample=True,
        temperature=0.85,
        top_p=0.9,
        repetition_penalty=1.12,
        no_repeat_ngram_size=4,
    )

print(tokenizer.decode(out[0], skip_special_tokens=True))
```


## Training data and release notes 📊
- This OSS LoRA adapter is trained on de-identified versions of the original data.
- The original internal LoRA adapter was trained on non-de-identified data and cannot be open-sourced at this time.
- More details and examples are provided in the [GitHub repo](https://github.com/Yukyin/DeepSupport).


## Safety and privacy ⚠️
This project is intended for supportive conversation only.  
It does not provide professional advice, diagnosis, or therapy. Please seek qualified professional help when needed.


## License 📜
This adapter is released for noncommercial use. See the [GitHub repo](https://github.com/Yukyin/DeepSupport) for the full license text and commercial licensing terms.


## Citation 📚
```bibtex
@software{deepsupport_warm_2026,
  author  = {Yuyan Chen},
  title   = {DeepSupport Warm: An emotional-holding companion for supportive dialogue},
  year    = {2026},
  version = {oss},
  url     = {\url{https://github.com/Yukyin/DeepSupport/DeepSupport_Warm}}
}
```

## Links
- GitHub: https://github.com/Yukyin/DeepSupport
- LoRA adapter: https://huggingface.co/Yukyin/deepsupport-warm-lora-oss