# LoRA Adapter for DippyResearch/super-cot-v1 This is a LoRA (Low-Rank Adaptation) adapter trained with VERL for roleplay tasks. ## Usage ### With vLLM ```python from vllm import LLM llm = LLM( model="/workspace/models/Cydonia-24B-v4.1", enable_lora=True, lora_modules=["DippyResearch/super-cot-v1"] ) ``` ### With Transformers + PEFT ```python from peft import PeftModel from transformers import AutoModelForCausalLM, AutoTokenizer base_model = AutoModelForCausalLM.from_pretrained("/workspace/models/Cydonia-24B-v4.1") model = PeftModel.from_pretrained(base_model, "DippyResearch/super-cot-v1") ``` ## Training Details - Framework: VERL (GRPO) - Base Model: Cydonia-24B-v4.1 - Training Type: Roleplay fine-tuning - Checkpoint: actor