YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
DippyResearch/duper-cot-24b-lj-v3
This is a fine-tuned model for roleplay tasks, created by merging a LoRA adapter with the base model.
Base Model
- Model: Cydonia-24B-v4.1
- Path: /workspace/models/Cydonia-24B-v4.1
Training Details
- Framework: VERL (GRPO)
- Training Type: Roleplay fine-tuning
- Checkpoint: actor
- LoRA Adapter: DippyResearch/duper-cot-24b-lj-v3-lora
Usage
With Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("DippyResearch/duper-cot-24b-lj-v3")
tokenizer = AutoTokenizer.from_pretrained("DippyResearch/duper-cot-24b-lj-v3")
With vLLM
from vllm import LLM
llm = LLM(model="DippyResearch/duper-cot-24b-lj-v3")
Alternative: Use LoRA Adapter
For more flexibility, you can use the LoRA adapter directly:
- LoRA Adapter: DippyResearch/duper-cot-24b-lj-v3-lora
- Base Model: Cydonia-24B-v4.1
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support