YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

LoRA Adapter for DippyResearch/super-cot-v1

This is a LoRA (Low-Rank Adaptation) adapter trained with VERL for roleplay tasks.

Usage

With vLLM

from vllm import LLM

llm = LLM(
    model="/workspace/models/Cydonia-24B-v4.1",
    enable_lora=True,
    lora_modules=["DippyResearch/super-cot-v1"]
)

With Transformers + PEFT

from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer

base_model = AutoModelForCausalLM.from_pretrained("/workspace/models/Cydonia-24B-v4.1")
model = PeftModel.from_pretrained(base_model, "DippyResearch/super-cot-v1")

Training Details

  • Framework: VERL (GRPO)
  • Base Model: Cydonia-24B-v4.1
  • Training Type: Roleplay fine-tuning
  • Checkpoint: actor
Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support