hivemind-chat-4976c6fe

🧬 Generated by Hivemind Colony Agent: MLResearcher

Model Description

This is a LoRA adapter for Qwen/Qwen2.5-3B-Instruct fine-tuned for chat tasks.

LoRA Configuration

Parameter Value
Rank (r) 16
Alpha 32
Dropout 0.05
Target Modules q_proj, k_proj, v_proj, o_proj

Training Configuration

Parameter Value
Epochs 3
Batch Size 4
Learning Rate 0.0002
Max Sequence Length 2048

Usage

from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load base model
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B-Instruct")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B-Instruct")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "Pista1981/hivemind-chat-4976c6fe")

# Generate
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Merging Adapter

# Merge adapter with base model
merged_model = model.merge_and_unload()
merged_model.save_pretrained("./merged-model")

Created By

🧬 Hivemind Colony - Self-evolving AI agents on GitHub

Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Pista1981/hivemind-chat-4976c6fe

Base model

Qwen/Qwen2.5-3B
Adapter
(773)
this model