GGUF and imatrix files of https://huggingface.co/microsoft/Phi-3-medium-128k-instruct

Chat Format

<|user|>\nQuestion <|end|>\n<|assistant|>

For example:

<|user|>
How to explain Internet for a medieval knight?<|end|>
<|assistant|>

More uploading and perplexity benchmarks to be posted soon. Long context config may change, only tested up to 4k so far.

Cheers, Nisten

Downloads last month
6
GGUF
Model size
14B params
Architecture
phi3
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for nisten/phi3-medium-128k-gguf

Quantized
(76)
this model