Overview
This page provides various quantisations of the base model, in GGUF format.
- mixedbread-ai/mxbai-embed-large-v1
Model Description
For a full model description, please refer to the base model's card.
How are the GGUF files created?
After cloning the author's original base model repository, llama.cpp is used to convert the model to a GGML compatible file, using f16 as the output type; preserving the original fidelity. The model is converted un-altered, unless otherwise stated.
Finally, for each respective quantisation level, llama.cpp's llama-quantize executable is called using the F32 GGUF file as the source file.
Quantisations
Pending ...
- Downloads last month
- -
Hardware compatibility
Log In to add your hardware
2-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Model tree for s3dev-ai/mxbai-embed-large-v1-gguf
Base model
mixedbread-ai/mxbai-embed-large-v1