YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
DippyResearch/super-cot-v1-step-20
This is a fine-tuned model for roleplay tasks, created by merging a LoRA adapter with the base model.
Base Model
- Model: Cydonia-24B-v4.1
- Path: /workspace/models/Cydonia-24B-v4.1
Training Details
- Framework: VERL (GRPO)
- Training Type: Roleplay fine-tuning
- Checkpoint: actor
- LoRA Adapter: DippyResearch/super-cot-v1-step-20-lora
Usage
With Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("DippyResearch/super-cot-v1-step-20")
tokenizer = AutoTokenizer.from_pretrained("DippyResearch/super-cot-v1-step-20")
With vLLM
from vllm import LLM
llm = LLM(model="DippyResearch/super-cot-v1-step-20")
Alternative: Use LoRA Adapter
For more flexibility, you can use the LoRA adapter directly:
- LoRA Adapter: DippyResearch/super-cot-v1-step-20-lora
- Base Model: Cydonia-24B-v4.1
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support