How to use from
llama.cppInstall from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16# Run inference directly in the terminal:
llama-cli -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16# Run inference directly in the terminal:
./llama-cli -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16# Run inference directly in the terminal:
./build/bin/llama-cli -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16Use Docker
docker model run hf.co/Impulse2000/multilingual-e5-large-instruct-GGUF:F16Quick Links
Impulse2000/multilingual-e5-large-instruct-GGUF
This model was converted to GGUF format from intfloat/multilingual-e5-large-instruct using llama.cpp via its 'convert_hf_to_gguf.py' script.
Refer to the original model card for more details on the model.
- Downloads last month
- 6
Hardware compatibility
Log In to add your hardware
8-bit
16-bit
Model tree for Impulse2000/multilingual-e5-large-instruct-GGUF
Base model
intfloat/multilingual-e5-large-instructEvaluation results
- accuracy on MTEB AmazonCounterfactualClassification (en)test set self-reported76.239
- ap on MTEB AmazonCounterfactualClassification (en)test set self-reported39.074
- f1 on MTEB AmazonCounterfactualClassification (en)test set self-reported70.048
- accuracy on MTEB AmazonCounterfactualClassification (de)test set self-reported66.713
- ap on MTEB AmazonCounterfactualClassification (de)test set self-reported79.015
- f1 on MTEB AmazonCounterfactualClassification (de)test set self-reported64.820
- accuracy on MTEB AmazonCounterfactualClassification (en-ext)test set self-reported73.853
- ap on MTEB AmazonCounterfactualClassification (en-ext)test set self-reported22.448
- f1 on MTEB AmazonCounterfactualClassification (en-ext)test set self-reported61.016
- accuracy on MTEB AmazonCounterfactualClassification (ja)test set self-reported76.049
- ap on MTEB AmazonCounterfactualClassification (ja)test set self-reported23.450
- f1 on MTEB AmazonCounterfactualClassification (ja)test set self-reported62.572
- accuracy on MTEB AmazonPolarityClassificationtest set self-reported96.287
- ap on MTEB AmazonPolarityClassificationtest set self-reported94.845
- f1 on MTEB AmazonPolarityClassificationtest set self-reported96.287
- accuracy on MTEB AmazonReviewsClassification (en)test set self-reported56.716
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16# Run inference directly in the terminal: llama-cli -hf Impulse2000/multilingual-e5-large-instruct-GGUF:F16