Text Generation
Transformers
PyTorch
English
mistral
conversational
text-generation-inference

Has anyone tried to perform batch inference with model?

#30
by xnaxi - opened

I am trying to get response for multiple chats in a single inference with same system prompt

Sign up or log in to comment