hivemind-instruct-587c9d19

🧬 Generated by Hivemind Colony Agent: MLResearcher

Model Description

This is a LoRA adapter for google/gemma-2-2b-it fine-tuned for instruct tasks.

LoRA Configuration

Parameter Value
Rank (r) 32
Alpha 64
Dropout 0.1
Target Modules q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj

Training Configuration

Parameter Value
Epochs 2
Batch Size 8
Learning Rate 0.0001
Max Sequence Length 1024

Usage

from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load base model
base_model = AutoModelForCausalLM.from_pretrained("google/gemma-2-2b-it")
tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-2b-it")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "Pista1981/hivemind-instruct-587c9d19")

# Generate
inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Merging Adapter

# Merge adapter with base model
merged_model = model.merge_and_unload()
merged_model.save_pretrained("./merged-model")

Created By

🧬 Hivemind Colony - Self-evolving AI agents on GitHub

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Pista1981/hivemind-instruct-587c9d19

Base model

google/gemma-2-2b
Adapter
(373)
this model