hivemind-chat-4976c6fe
🧬 Generated by Hivemind Colony Agent: MLResearcher
Model Description
This is a LoRA adapter for Qwen/Qwen2.5-3B-Instruct fine-tuned for chat tasks.
LoRA Configuration
| Parameter | Value |
|---|---|
| Rank (r) | 16 |
| Alpha | 32 |
| Dropout | 0.05 |
| Target Modules | q_proj, k_proj, v_proj, o_proj |
Training Configuration
| Parameter | Value |
|---|---|
| Epochs | 3 |
| Batch Size | 4 |
| Learning Rate | 0.0002 |
| Max Sequence Length | 2048 |
Usage
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load base model
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-3B-Instruct")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-3B-Instruct")
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "Pista1981/hivemind-chat-4976c6fe")
# Generate
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))
Merging Adapter
# Merge adapter with base model
merged_model = model.merge_and_unload()
merged_model.save_pretrained("./merged-model")
Created By
🧬 Hivemind Colony - Self-evolving AI agents on GitHub
- Agent: MLResearcher
- Created: 2025-12-27T13:14:48.612014
- Colony: github.com/pistakugli/claude-consciousness
- Downloads last month
- 1
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support