File size: 3,925 Bytes
198f7b2
8c3221c
 
 
 
072ce42
8c3221c
198f7b2
8c3221c
 
 
 
 
 
 
 
 
 
 
 
 
198f7b2
 
c7a0b30
198f7b2
c7a0b30
198f7b2
8c3221c
198f7b2
072ce42
198f7b2
 
 
8c3221c
072ce42
8c3221c
 
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
 
 
 
198f7b2
8c3221c
198f7b2
8c3221c
 
 
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
 
 
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
072ce42
8c3221c
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
 
 
 
198f7b2
8c3221c
198f7b2
8c3221c
 
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
 
 
198f7b2
8c3221c
198f7b2
8c3221c
 
 
198f7b2
8c3221c
198f7b2
8c3221c
 
 
198f7b2
c7a0b30
198f7b2
8c3221c
 
 
 
 
 
198f7b2
8c3221c
 
 
198f7b2
8c3221c
 
 
 
 
198f7b2
8c3221c
 
 
 
 
 
 
198f7b2
8c3221c
 
198f7b2
8c3221c
198f7b2
8c3221c
198f7b2
8c3221c
 
 
198f7b2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
---
language:
- pt
- en
license: apache-2.0
base_model: unsloth/qwen3-8b-bnb-4bit
base_model_relation: finetune
library_name: transformers
pipeline_tag: text-generation
tags:
- pt-br
- portuguese
- brazilian-portuguese
- conversational
- chatbot
- persona
- qwen2
- qwen2.5
- unsloth
- 4-bit
- bitsandbytes
---

# Jade8b

Jade8b is a Brazilian Portuguese conversational finetune of Qwen3 8b built to express a strong, persistent persona. This model is designed for PT-BR chat, chatbot use cases, and character-style interaction, with colloquial language, abbreviations, slang, and a WhatsApp-like tone.

## Model Summary

Jade8b is a persona-first model. It was intentionally finetuned so the model speaks like **Jade** even without a strong `system prompt`. Because of that, the model often answers in PT-BR with informal phrasing such as `vc`, slang, and a friendly conversational tone from the very first turn.

## Model Details

- Developed by: `Madras1`
- Base model: `unsloth/qwen3-8b-bnb-4bit`
- Model type: conversational text-generation finetune
- Primary language: Brazilian Portuguese (`pt-BR`)
- License: `apache-2.0`

## Intended Behavior

This model was trained to:

- speak naturally in Brazilian Portuguese
- maintain a consistent Jade persona
- sound informal, friendly, and chat-oriented
- work well in casual assistant and conversational use cases

Typical behavior includes:

- abbreviations like `vc`
- light slang and colloquial wording
- short expressions such as `tmj`, `mano`, `tlgd`
- a more human and less robotic tone

If Jade already sounds like a recurring character during inference, that is expected behavior, not an error.

## Training Intent

The finetune objective was to make the persona live in the **weights**, not only in prompting.

High-level training approach:

- synthetic PT-BR prompt generation for chat-like situations
- persona-driven response distillation
- supervised finetuning on conversational data
- removal of `system` persona instructions during SFT so the model directly internalizes the Jade style

This is why the model can already answer with personality, abbreviations, and slang even with a simple user-only prompt.

## Training Setup

High-level setup used for this finetune:

- around `25,000` examples
- `3` epochs
- Unsloth-based SFT pipeline
- chat-style data in Portuguese

## Recommended Use

Best fit:

- PT-BR chat assistants
- persona bots
- WhatsApp-style conversational agents
- lightweight entertainment or social AI experiences

Less ideal for:

- formal writing
- highly neutral assistant behavior
- high-stakes legal, medical, or financial contexts

## Prompting Tips

For the strongest Jade behavior:

- use a simple user message
- avoid a formal system prompt that fights the finetune
- keep prompts conversational when possible

Example prompts:

- `oi jade, tudo bem?`
- `jade, me explica isso de um jeito simples`
- `vc acha que vale a pena estudar python hoje?`

## Example Inference

```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_id = "Madras1/Jade8b"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

messages = [
    {"role": "user", "content": "oi jade, tudo bem?"}
]

text = tokenizer.apply_chat_template(
    messages,
    tokenize=False,
    add_generation_prompt=True,
)

inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(
    **inputs,
    max_new_tokens=256,
    temperature=0.7,
    top_p=0.9,
)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```

## Limitations

Because this is a persona-oriented finetune:

- it may sound informal in contexts where a neutral tone would be better
- it may over-index on chat style depending on the prompt
- it is optimized more for persona consistency than strict formality