Nanbeige4.1-3B — Gmail Tool-Use (SFT + DPO)

Fine-tuned version of Nanbeige/Nanbeige4.1-3B for Gmail tool-calling tasks using a two-stage training pipeline.

Training datasets: TurkishCodeMan/Nanbeige4.1-3B-Gmail-Tool-Use-Datasets

Training Pipeline

Stage 1 — Supervised Fine-Tuning (SFT)

  • Dataset: 740 multi-turn Gmail agent traces (sft/traces_chatml_clean.jsonl)
  • Format: ChatML with tool_calls (OpenAI function-calling schema)
  • Method: LoRA r=16, α=32, 7 target modules
  • Result: loss 0.8464 → 0.1888 · PPL 2.33 → 1.21

Stage 2 — Direct Preference Optimization (DPO)

  • Dataset: 3223 preference pairs (dpo/dpo_dataset.jsonl) — 3 rejection strategies:
    • wrong_tool — incorrect tool selected (~34%)
    • missing_args — required arguments omitted (~32%)
    • bad_answer — poor final response (~34%)
  • Method: DPO β=0.1, sigmoid loss, LoRA r=16, ref_model=None (PEFT implicit ref)
  • Result: val_loss=0.000765 · reward accuracy=100% · normalized margin=+0.52

Supported Tools

Tool Description
search_emails Search Gmail inbox with filters
read_email Read full email content by ID
send_email Send a new email
draft_email Create a draft
modify_email Add/remove labels, mark read/unread
download_attachment Download email attachment

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model = AutoModelForCausalLM.from_pretrained(
    "TurkishCodeMan/Nanbeige4.1-3B-Gmail-Tool-Use",
    torch_dtype=torch.bfloat16,
    trust_remote_code=True,
)
tokenizer = AutoTokenizer.from_pretrained(
    "TurkishCodeMan/Nanbeige4.1-3B-Gmail-Tool-Use",
    trust_remote_code=True,
)

Training Details

Parameter Value
Base model Nanbeige/Nanbeige4.1-3B
SFT LoRA rank 16
DPO LoRA rank 16
DPO β 0.1
Max length 2682 tokens
GPU 1× RTX 4090 24GB
Framework TRL 0.22 · Transformers 4.57 · PEFT 0.18
Downloads last month
8
Safetensors
Model size
4B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for TurkishCodeMan/Nanbeige4.1-3B-Gmail-Tool-Use

Finetuned
(14)
this model

Dataset used to train TurkishCodeMan/Nanbeige4.1-3B-Gmail-Tool-Use