YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
DippyResearch/duper-cot-24b-lj-rb-v4-s100
This is a fine-tuned model for roleplay tasks, created by merging a LoRA adapter with the base model.
Base Model
- Model: Cydonia-24B-v4.2.0
- Path: /workspace/models/Cydonia-24B-v4.2.0/
Training Details
- Framework: VERL (GRPO)
- Training Type: Roleplay fine-tuning
- Checkpoint: actor
- LoRA Adapter: DippyResearch/duper-cot-24b-lj-rb-v4-s100-lora
Usage
With Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("DippyResearch/duper-cot-24b-lj-rb-v4-s100")
tokenizer = AutoTokenizer.from_pretrained("DippyResearch/duper-cot-24b-lj-rb-v4-s100")
With vLLM
from vllm import LLM
llm = LLM(model="DippyResearch/duper-cot-24b-lj-rb-v4-s100")
Alternative: Use LoRA Adapter
For more flexibility, you can use the LoRA adapter directly:
- LoRA Adapter: DippyResearch/duper-cot-24b-lj-rb-v4-s100-lora
- Base Model: Cydonia-24B-v4.2.0
- Downloads last month
- -
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support