How to use from
llama.cppInstall from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16# Run inference directly in the terminal:
llama-cli -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16# Run inference directly in the terminal:
./llama-cli -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16# Run inference directly in the terminal:
./build/bin/llama-cli -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16Use Docker
docker model run hf.co/BAEM1N/PIXIE-Rune-v1.0-GGUF:F16Quick Links
PIXIE-Rune-v1.0 GGUF
GGUF conversion of telepix/PIXIE-Rune-v1.0.
#2 on Korean Embedding Leaderboard (NDCG@5,10 avg: 84.68)
Files
| File | Quant | Size |
|---|---|---|
| PIXIE-Rune-v1.0-Q8_0.gguf | Q8_0 | 599 MB |
| PIXIE-Rune-v1.0-F16.gguf | F16 | 1.15 GB |
Usage with llama.cpp
llama-server -m PIXIE-Rune-v1.0-Q8_0.gguf --port 9020 --embedding --pooling mean
Model Info
- Architecture: XLM-RoBERTa (24 layers, hidden=1024)
- Parameters: ~335M
- Embedding dimension: 1024
- Max sequence length: 8192
- Downloads last month
- 20
Hardware compatibility
Log In to add your hardware
8-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for BAEM1N/PIXIE-Rune-v1.0-GGUF
Base model
telepix/PIXIE-Rune-v1.0
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16# Run inference directly in the terminal: llama-cli -hf BAEM1N/PIXIE-Rune-v1.0-GGUF:F16