How to use from
llama.cppInstall from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf 0Time/INCEPT-SH# Run inference directly in the terminal:
llama-cli -hf 0Time/INCEPT-SHUse pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf 0Time/INCEPT-SH# Run inference directly in the terminal:
./llama-cli -hf 0Time/INCEPT-SHBuild from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf 0Time/INCEPT-SH# Run inference directly in the terminal:
./build/bin/llama-cli -hf 0Time/INCEPT-SHUse Docker
docker model run hf.co/0Time/INCEPT-SHQuick Links
INCEPT.sh
Offline command inference engine for Linux. Fine-tuned Qwen3.5-0.8B (GGUF Q8_0, 774MB) designed to run on low-resource and edge devices with no GPU, no API, and no internet connection required at runtime.
Benchmark: 99/100 on a structured 100-question Linux command evaluation (Ubuntu 22.04, bash, non-root).
Installation
curl -fsSL https://raw.githubusercontent.com/0-Time/INCEPT.sh/main/install.sh | bash
Supports: Debian/Ubuntu, RHEL/Fedora, CentOS, Arch, openSUSE.
Manual Model Setup
# Download model
huggingface-cli download 0Time/INCEPT-SH \
incept-sh.gguf --local-dir ./models
# Clone and install
git clone https://github.com/0-Time/INCEPT.sh
cd INCEPT.sh
pip install -e ".[cli]"
incept
Usage
# Interactive CLI
incept
# One-shot
incept -c "list all open ports"
# Minimal output (pipe-friendly)
incept -c "find large files" -m
# With model reasoning
incept --think
CLI Commands
| Command | Description |
|---|---|
/think on|off |
Toggle chain-of-thought reasoning |
/context |
Show detected system context |
/help |
List available commands |
/exit |
Exit |
Prompt Format
ChatML with a system context line:
<|im_start|>system
ubuntu 22.04 bash non-root
<|im_end|>
<|im_start|>user
{natural language query}
<|im_end|>
<|im_start|>assistant
<think>
</think>
Inference temperature: 0.0 (greedy decoding).
Training
| Parameter | Value |
|---|---|
| Base model | Qwen/Qwen3.5-0.8B |
| Training method | Supervised fine-tuning (LoRA, rank 16) |
| Training examples | 79,264 (SFT) + 11,306 (pipe refinement) |
| Learning rate | 5ร10โปโต |
| Quantization | Q8_0 (774MB) |
| Supported distros | Ubuntu, Debian, RHEL, Arch, Fedora, CentOS |
| Training hardware | Apple M4 Mac mini, 32GB unified RAM |
Safety
- Prompt injection detection (exact-phrase matching)
- Catastrophic pattern blocking (
rm -rf /, fork bombs, pipe-to-shell, etc.) - Risk classification:
SAFE/CAUTION/DANGEROUS/BLOCKED - Zero outbound traffic at runtime
Requirements
- Linux x86_64 / aarch64
- Python 3.11+
llama-serveronPATH- ~1GB RAM at runtime
Links
- GitHub: 0-Time/INCEPT.sh
- Release: v1.0.0
License
- Downloads last month
- 12
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf 0Time/INCEPT-SH# Run inference directly in the terminal: llama-cli -hf 0Time/INCEPT-SH