qwen2.5-3b-lora-r32

Fine-tuned LoRA adapter on top of Qwen/Qwen2.5-3B-Instruct.

Training Details

Base Model

  • Model: Qwen/Qwen2.5-3B-Instruct

LoRA Configuration

Parameter Value
r 32
alpha 64
dropout 0.05
target modules down_proj, gate_proj, k_proj, o_proj, q_proj, up_proj, v_proj
task type CAUSAL_LM

Training Arguments

Parameter Value
epochs 8.0
batch size 2
learning rate 0.0002
weight decay 0.0
warmup steps 2.0
gradient accumulation steps 1

Usage

from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer

base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B-Instruct")
model = PeftModel.from_pretrained(base_model, "qwen2.5-3b-lora-r32")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B-Instruct")
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for kostasang/qwen2.5-3b-lora-r32

Base model

Qwen/Qwen2.5-3B
Adapter
(1184)
this model