How to use from the
Use from the
Transformers library
# Load model directly
from transformers import REX
model = REX.from_pretrained("Maynx/Rex-Instruct-v0.1", dtype="auto")
Quick Links

Rex-Instruct-v0.1

This repository contains the genesis_smol fine-tuned checkpoint exported from this project.

Notes

  • Architecture: REX
  • Chat format: ChatML with <|im_start|> and <|im_end|>
  • Tokenizer base: mistralai/Mistral-7B-v0.3

Local Loading

This is a custom architecture, so load it with the local REX class from this repository:

from model.model import REX
from transformers import AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("Maynx/Rex-Instruct-v0.1", use_fast=False)
model = REX.from_pretrained("Maynx/Rex-Instruct-v0.1")
Downloads last month
39
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support