|
|
--- |
|
|
base_model: Qwen/Qwen3-4B-Instruct-2507 |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
library_name: peft |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- lora |
|
|
- agent |
|
|
- tool-use |
|
|
- alfworld |
|
|
- dbbench |
|
|
--- |
|
|
|
|
|
# <qwen3-4b-agent-trajectory-lora> |
|
|
|
|
|
This repository provides a **LoRA adapter** fine-tuned from |
|
|
**Qwen/Qwen3-4B-Instruct-2507** using **LoRA + Unsloth**. |
|
|
|
|
|
This repository contains **LoRA adapter weights only**. |
|
|
The base model must be loaded separately. |
|
|
|
|
|
## Training Objective |
|
|
|
|
|
This adapter is trained to improve **multi-turn agent task performance** |
|
|
on ALFWorld (household tasks) and DBBench (database operations). |
|
|
|
|
|
Loss is applied to **all assistant turns** in the multi-turn trajectory, |
|
|
enabling the model to learn environment observation, action selection, |
|
|
tool use, and recovery from errors. |
|
|
|
|
|
## Training Configuration |
|
|
|
|
|
- Base model: Qwen/Qwen3-4B-Instruct-2507 |
|
|
- Method: LoRA (full precision base) |
|
|
- Max sequence length: 4096 |
|
|
- Epochs: 3 |
|
|
- Learning rate: 8e-06 |
|
|
- LoRA: r=64, alpha=128 |
|
|
- warmup_ratio : 0 |
|
|
- weight_decay : 0 |
|
|
- lora_dropout : 0 |
|
|
- lora_target_modules :['q_proj', 'k_proj', 'v_proj', 'o_proj', 'gate_proj', 'up_proj', 'down_proj'] |
|
|
- grad_accum : 4 |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
import torch |
|
|
|
|
|
base = "Qwen/Qwen3-4B-Instruct-2507" |
|
|
adapter = "your_id/your-repo" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(base) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
base, |
|
|
torch_dtype=torch.float16, |
|
|
device_map="auto", |
|
|
) |
|
|
model = PeftModel.from_pretrained(model, adapter) |
|
|
``` |
|
|
|
|
|
## Sources & Terms (IMPORTANT) |
|
|
|
|
|
Training data: /content/dbbench_react_merged_dedup |
|
|
|
|
|
Dataset License: MIT License. This dataset is used and distributed under the terms of the MIT License. |
|
|
Compliance: Users must comply with the MIT license (including copyright notice) and the base model's original terms of use. |
|
|
|