--- base_model: Qwen/Qwen2.5-7B-Instruct datasets: - u-10bei/sft_alfworld_trajectory_dataset_v5 - u-10bei/dbbench_sft_dataset_react_v4 language: - en license: apache-2.0 library_name: peft pipeline_tag: text-generation tags: - lora - agent - tool-use - alfworld - dbbench --- # qwen2.5-7b-alf-dbb-merged-final This repository provides a **merged full model** based on **Qwen/Qwen2.5-7B-Instruct**. ## Model Construction Pipeline 1. Train LoRA adapter on ALFWorld 2. Train LoRA adapter on DBBench 3. Merge adapters using `ties` (density=0.1) 4. Apply additional stabilization fine-tuning (LoRA) 5. Merge final adapter into base model This repository contains **full merged weights (no adapter required)**. ## Final Training Configuration - Base model: Qwen/Qwen2.5-7B-Instruct - Merge method: ties - Merge density: 0.1 - Final stage epochs: 1 - Learning rate: 1e-05 - Final LoRA: r=16, alpha=16 - Max sequence length: 2024 ## Datasets - u-10bei/sft_alfworld_trajectory_dataset_v5 - u-10bei/dbbench_sft_dataset_react_v4 Additional distilled datasets were optionally included. ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "takayosh/agentbenchfinal" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained( model_id, torch_dtype="auto", device_map="auto" ) ``` ## Sources & Terms (IMPORTANT) Training data: - u-10bei/sft_alfworld_trajectory_dataset_v5 - u-10bei/dbbench_sft_dataset_react_v4 Dataset License: MIT License. This dataset is used and distributed under the terms of the MIT License. Compliance: Users must comply with the MIT license (including copyright notice) and the base model's original terms of use.