--- datasets: - PocketDoc/Dans-Personamaxx-Logs - anthracite-org/kalo-opus-instruct-22k-no-refusal - lodrick-the-lafted/kalo-opus-instruct-3k-filtered - anthracite-org/nopm_claude_writing_fixed - anthracite-org/kalo_opus_misc_240827 - anthracite-org/kalo_misc_part2 - NewEden/Claude-Instruct-5K - NewEden/Claude-Instruct-2.7K base_model: - NewEden/MistralAI-Nemo-Instruct-ChatML pipeline_tag: text-generation library_name: transformers language: - en tags: - roleplay - finetune - mistral - magnum - claude - story-writing ---
These are EXL2 quants, Look in the main branch for the measurement file. Look in the different branches for other bpws.
Originally conceived as an experiment to test the effects of gradient clipping, this model was exceptionally well-received by early testers, prompting its official release.
Fine-tuned on top of Mistral-Nemo-Instruct (ChatML'ified), Rei-12B is designed to replicate the exquisite prose quality of Claude 3 models, particularly Sonnet and Opus, using a prototype Magnum V5 datamix.
Rei-12B uses the ChatML format. A typical conversation should be structured as:
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.\n\n
https://wandb.ai/new-eden/Rei-V2/artifacts/axolotl-config/config-7hvbucx9/v0/files/axolotl_config_pw8f0c6u.yml
The model was trained for 2 epochs on 8x NVIDIA H200s GPUs generously provided by @Kalomaze
I'd like to thank, Ruka/Sama twinkman | LucyKnada | Kubernetes Bad | PocketDoc | Tav | Trappu | And the rest of Anthracite/Pygmalion for testing, feedback, and support.