Text Generation
Transformers
Safetensors
English
mistral
mteb
Eval Results (legacy)
text-generation-inference
How to use from
SGLangUse Docker images
docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "BeastyZ/e5-R-mistral-7b" \
--host 0.0.0.0 \
--port 30000# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "BeastyZ/e5-R-mistral-7b",
"prompt": "Once upon a time,",
"max_tokens": 512,
"temperature": 0.5
}'Quick Links
Model Card for e5-R-mistral-7b
Model Description
e5-R-mistral-7b is a LLM retriever fine-tuned from mistralai/Mistral-7B-v0.1.
- Downloads last month
- 173
Model tree for BeastyZ/e5-R-mistral-7b
Dataset used to train BeastyZ/e5-R-mistral-7b
Spaces using BeastyZ/e5-R-mistral-7b 14
Evaluation results
- map_at_1 on MTEB ArguAnatest set self-reported33.570
- map_at_10 on MTEB ArguAnatest set self-reported49.952
- map_at_100 on MTEB ArguAnatest set self-reported50.673
- map_at_1000 on MTEB ArguAnatest set self-reported50.674
- map_at_3 on MTEB ArguAnatest set self-reported44.915
- map_at_5 on MTEB ArguAnatest set self-reported47.877
- mrr_at_1 on MTEB ArguAnatest set self-reported34.211
- mrr_at_10 on MTEB ArguAnatest set self-reported50.190
- mrr_at_100 on MTEB ArguAnatest set self-reported50.905
- mrr_at_1000 on MTEB ArguAnatest set self-reported50.906
- mrr_at_3 on MTEB ArguAnatest set self-reported45.128
- mrr_at_5 on MTEB ArguAnatest set self-reported48.097
- ndcg_at_1 on MTEB ArguAnatest set self-reported33.570
- ndcg_at_10 on MTEB ArguAnatest set self-reported58.994
- ndcg_at_100 on MTEB ArguAnatest set self-reported61.806
- ndcg_at_1000 on MTEB ArguAnatest set self-reported61.825
Install from pip and serve model
# Install SGLang from pip: pip install sglang# Start the SGLang server: python3 -m sglang.launch_server \ --model-path "BeastyZ/e5-R-mistral-7b" \ --host 0.0.0.0 \ --port 30000# Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "BeastyZ/e5-R-mistral-7b", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'