# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("monsterapi/zephyr-7b-beta-CTranslate2-bfloat16", dtype="auto")Quick Links
Currently Ctranslate2 does not directly support mistral and zephyr models for conversion
Here is a custom converted model made possible by some code changes to the ct2 repo for mistral. Mainly developed for internal development use you can use it too if your struggling with the same issue
Note: Model was created with BFloat16 quantization
license: apache-2.0
- Downloads last month
- 20
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="monsterapi/zephyr-7b-beta-CTranslate2-bfloat16") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)