Jade4b / README.md
Madras1's picture
Update README.md
57e65b1 verified
---
language:
- pt
- en
license: apache-2.0
base_model: unsloth/qwen3-4b
base_model_relation: finetune
library_name: transformers
pipeline_tag: text-generation
tags:
- pt-br
- portuguese
- brazilian-portuguese
- conversational
- chatbot
- persona
- qwen2
- qwen2.5
- unsloth
- 4-bit
- bitsandbytes
---
# Jade4b
Jade4b is a Brazilian Portuguese conversational finetune of Qwen3 4b built to express a strong, persistent persona. This model is designed for PT-BR chat, chatbot use cases, and character-style interaction, with colloquial language, abbreviations, slang, and a WhatsApp-like tone.
## Model Summary
Jade4b is a persona-first model. It was intentionally finetuned so the model speaks like **Jade** even without a strong `system prompt`. Because of that, the model often answers in PT-BR with informal phrasing such as `vc`, slang, and a friendly conversational tone from the very first turn.
## Model Details
- Developed by: `Madras1`
- Base model: `unsloth/qwen3-4b`
- Model type: conversational text-generation finetune
- Primary language: Brazilian Portuguese (`pt-BR`)
- License: `apache-2.0`
## Intended Behavior
This model was trained to:
- speak naturally in Brazilian Portuguese
- maintain a consistent Jade persona
- sound informal, friendly, and chat-oriented
- work well in casual assistant and conversational use cases
Typical behavior includes:
- abbreviations like `vc`
- light slang and colloquial wording
- short expressions such as `tmj`, `mano`, `tlgd`
- a more human and less robotic tone
If Jade already sounds like a recurring character during inference, that is expected behavior, not an error.
## Training Intent
The finetune objective was to make the persona live in the **weights**, not only in prompting.
High-level training approach:
- synthetic PT-BR prompt generation for chat-like situations
- persona-driven response distillation
- supervised finetuning on conversational data
- removal of `system` persona instructions during SFT so the model directly internalizes the Jade style
This is why the model can already answer with personality, abbreviations, and slang even with a simple user-only prompt.
## Training Setup
High-level setup used for this finetune:
- around `25,000` examples
- `3` epochs
- Unsloth-based SFT pipeline
- chat-style data in Portuguese
## Recommended Use
Best fit:
- PT-BR chat assistants
- persona bots
- WhatsApp-style conversational agents
- lightweight entertainment or social AI experiences
Less ideal for:
- formal writing
- highly neutral assistant behavior
- high-stakes legal, medical, or financial contexts
## Prompting Tips
For the strongest Jade behavior:
- use a simple user message
- avoid a formal system prompt that fights the finetune
- keep prompts conversational when possible
Example prompts:
- `oi jade, tudo bem?`
- `jade, me explica isso de um jeito simples`
- `vc acha que vale a pena estudar python hoje?`
## Example Inference
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_id = "Madras1/Jade4b"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
messages = [
{"role": "user", "content": "oi jade, tudo bem?"}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True,
)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(
**inputs,
max_new_tokens=256,
temperature=0.7,
top_p=0.9,
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Limitations
Because this is a persona-oriented finetune:
- it may sound informal in contexts where a neutral tone would be better
- it may over-index on chat style depending on the prompt
- it is optimized more for persona consistency than strict formality