Overview

This page provides various quantisations of the base model, in GGUF format.

  • mixedbread-ai/mxbai-embed-large-v1

Model Description

For a full model description, please refer to the base model's card.

How are the GGUF files created?

After cloning the author's original base model repository, llama.cpp is used to convert the model to a GGML compatible file, using f16 as the output type; preserving the original fidelity. The model is converted un-altered, unless otherwise stated.

Finally, for each respective quantisation level, llama.cpp's llama-quantize executable is called using the F32 GGUF file as the source file.

Quantisations

Pending ...

Downloads last month
-
GGUF
Model size
0.3B params
Architecture
bert
Hardware compatibility
Log In to add your hardware

2-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for s3dev-ai/mxbai-embed-large-v1-gguf

Quantized
(13)
this model