File size: 3,925 Bytes
198f7b2 8c3221c 072ce42 8c3221c 198f7b2 8c3221c 198f7b2 c7a0b30 198f7b2 c7a0b30 198f7b2 8c3221c 198f7b2 072ce42 198f7b2 8c3221c 072ce42 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 072ce42 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 c7a0b30 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 8c3221c 198f7b2 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 | ---
language:
- pt
- en
license: apache-2.0
base_model: unsloth/qwen3-8b-bnb-4bit
base_model_relation: finetune
library_name: transformers
pipeline_tag: text-generation
tags:
- pt-br
- portuguese
- brazilian-portuguese
- conversational
- chatbot
- persona
- qwen2
- qwen2.5
- unsloth
- 4-bit
- bitsandbytes
---
# Jade8b
Jade8b is a Brazilian Portuguese conversational finetune of Qwen3 8b built to express a strong, persistent persona. This model is designed for PT-BR chat, chatbot use cases, and character-style interaction, with colloquial language, abbreviations, slang, and a WhatsApp-like tone.
## Model Summary
Jade8b is a persona-first model. It was intentionally finetuned so the model speaks like **Jade** even without a strong `system prompt`. Because of that, the model often answers in PT-BR with informal phrasing such as `vc`, slang, and a friendly conversational tone from the very first turn.
## Model Details
- Developed by: `Madras1`
- Base model: `unsloth/qwen3-8b-bnb-4bit`
- Model type: conversational text-generation finetune
- Primary language: Brazilian Portuguese (`pt-BR`)
- License: `apache-2.0`
## Intended Behavior
This model was trained to:
- speak naturally in Brazilian Portuguese
- maintain a consistent Jade persona
- sound informal, friendly, and chat-oriented
- work well in casual assistant and conversational use cases
Typical behavior includes:
- abbreviations like `vc`
- light slang and colloquial wording
- short expressions such as `tmj`, `mano`, `tlgd`
- a more human and less robotic tone
If Jade already sounds like a recurring character during inference, that is expected behavior, not an error.
## Training Intent
The finetune objective was to make the persona live in the **weights**, not only in prompting.
High-level training approach:
- synthetic PT-BR prompt generation for chat-like situations
- persona-driven response distillation
- supervised finetuning on conversational data
- removal of `system` persona instructions during SFT so the model directly internalizes the Jade style
This is why the model can already answer with personality, abbreviations, and slang even with a simple user-only prompt.
## Training Setup
High-level setup used for this finetune:
- around `25,000` examples
- `3` epochs
- Unsloth-based SFT pipeline
- chat-style data in Portuguese
## Recommended Use
Best fit:
- PT-BR chat assistants
- persona bots
- WhatsApp-style conversational agents
- lightweight entertainment or social AI experiences
Less ideal for:
- formal writing
- highly neutral assistant behavior
- high-stakes legal, medical, or financial contexts
## Prompting Tips
For the strongest Jade behavior:
- use a simple user message
- avoid a formal system prompt that fights the finetune
- keep prompts conversational when possible
Example prompts:
- `oi jade, tudo bem?`
- `jade, me explica isso de um jeito simples`
- `vc acha que vale a pena estudar python hoje?`
## Example Inference
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "Madras1/Jade8b"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
messages = [
{"role": "user", "content": "oi jade, tudo bem?"}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True,
)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=256,
temperature=0.7,
top_p=0.9,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Limitations
Because this is a persona-oriented finetune:
- it may sound informal in contexts where a neutral tone would be better
- it may over-index on chat style depending on the prompt
- it is optimized more for persona consistency than strict formality
|