Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf cortexso/sailor-2:# Run inference directly in the terminal:
llama-cli -hf cortexso/sailor-2:Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf cortexso/sailor-2:# Run inference directly in the terminal:
./llama-cli -hf cortexso/sailor-2:Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf cortexso/sailor-2:# Run inference directly in the terminal:
./build/bin/llama-cli -hf cortexso/sailor-2:Use Docker
docker model run hf.co/cortexso/sailor-2:Overview
Sailor2 is a community-driven initiative that brings cutting-edge multilingual language models to South-East Asia (SEA). It is designed to address the growing demand for diverse, robust, and accessible language technologies in the region. Built upon the foundation of Qwen 2.5, Sailor2 is continuously pre-trained on 500B tokens, significantly improving its support for 15 languages with a unified model. These languages include English, Chinese, Burmese, Cebuano, Ilocano, Indonesian, Javanese, Khmer, Lao, Malay, Sundanese, Tagalog, Thai, Vietnamese, and Waray.
Sailor2 is available in three sizes: 1B, 8B, and 20B, which are expansions from the Qwen2.5 base models of 0.5B, 7B, and 14B, respectively. These models serve a wide range of applications, from production use to research and speculative decoding, ensuring accessibility to advanced language technologies across SEA.
Variants
| No | Variant | Cortex CLI command |
|---|---|---|
| 1 | Sailor-2-1b | cortex run sailor-2:1b |
| 2 | Sailor-2-8b | cortex run sailor-2:8b |
| 3 | Sailor-2-20b | cortex run sailor-2:20b |
Use it with Jan (UI)
- Install Jan using Quickstart
- Use in Jan model Hub:
cortexhub/sailor-2
Use it with Cortex (CLI)
- Install Cortex using Quickstart
- Run the model with command:
cortex run sailor-2
Credits
- Author: Community-driven (Sailor2 Initiative)
- Converter: Homebrew
- Original License: Licence
- Papers: Technical Paper
- Downloads last month
- 195
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf cortexso/sailor-2:# Run inference directly in the terminal: llama-cli -hf cortexso/sailor-2: