How to use from
llama.cpp
Install from brew
brew install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf huzpsb/heru
# Run inference directly in the terminal:
llama-cli -hf huzpsb/heru
Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf huzpsb/heru
# Run inference directly in the terminal:
llama-cli -hf huzpsb/heru
Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases
# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf huzpsb/heru
# Run inference directly in the terminal:
./llama-cli -hf huzpsb/heru
Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli
# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf huzpsb/heru
# Run inference directly in the terminal:
./build/bin/llama-cli -hf huzpsb/heru
Use Docker
docker model run hf.co/huzpsb/heru
Quick Links

her-4b的中间产物(无RLHF)模型
生成内容非常不稳定,指令跟随很糟糕,没有任何安全措施,不适合任何场景直接使用
发布本模型的目的是为了有需要的模型构建者可以自己对模型进行对齐,而非供直接使用
正确使用本模型需要极高的专业知识与能力,带来的任何后果均由使用者自行承担!

Downloads last month
-
GGUF
Model size
4B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Space using huzpsb/heru 1