How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="monsterapi/zephyr-7b-beta-CTranslate2-bfloat16")
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("monsterapi/zephyr-7b-beta-CTranslate2-bfloat16", dtype="auto")
Quick Links

Currently Ctranslate2 does not directly support mistral and zephyr models for conversion

Here is a custom converted model made possible by some code changes to the ct2 repo for mistral. Mainly developed for internal development use you can use it too if your struggling with the same issue

Note: Model was created with BFloat16 quantization


license: apache-2.0

Downloads last month
20
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support