INCEPT-SH / README.md
0Time's picture
fix: remove em dash from description
6a310c3 verified
---
language: en
license: apache-2.0
tags:
- linux
- command-generation
- gguf
- qwen3
- llama-cpp
- offline
base_model: Qwen/Qwen3.5-0.8B
---
# INCEPT.sh
Offline command inference engine for Linux. Fine-tuned **Qwen3.5-0.8B** (GGUF Q8_0, 774MB) designed to run on low-resource and edge devices with no GPU, no API, and no internet connection required at runtime.
**Benchmark:** 99/100 on a structured 100-question Linux command evaluation (Ubuntu 22.04, bash, non-root).
## Installation
```bash
curl -fsSL https://raw.githubusercontent.com/0-Time/INCEPT.sh/main/install.sh | bash
```
Supports: Debian/Ubuntu, RHEL/Fedora, CentOS, Arch, openSUSE.
## Manual Model Setup
```bash
# Download model
huggingface-cli download 0Time/INCEPT-SH \
incept-sh.gguf --local-dir ./models
# Clone and install
git clone https://github.com/0-Time/INCEPT.sh
cd INCEPT.sh
pip install -e ".[cli]"
incept
```
## Usage
```bash
# Interactive CLI
incept
# One-shot
incept -c "list all open ports"
# Minimal output (pipe-friendly)
incept -c "find large files" -m
# With model reasoning
incept --think
```
## CLI Commands
| Command | Description |
|---|---|
| `/think on\|off` | Toggle chain-of-thought reasoning |
| `/context` | Show detected system context |
| `/help` | List available commands |
| `/exit` | Exit |
## Prompt Format
ChatML with a system context line:
```
<|im_start|>system
ubuntu 22.04 bash non-root
<|im_end|>
<|im_start|>user
{natural language query}
<|im_end|>
<|im_start|>assistant
<think>
</think>
```
Inference temperature: **0.0** (greedy decoding).
## Training
| Parameter | Value |
|-----------------------|----------------------------------------------|
| Base model | Qwen/Qwen3.5-0.8B |
| Training method | Supervised fine-tuning (LoRA, rank 16) |
| Training examples | 79,264 (SFT) + 11,306 (pipe refinement) |
| Learning rate | 5×10⁻⁵ |
| Quantization | Q8_0 (774MB) |
| Supported distros | Ubuntu, Debian, RHEL, Arch, Fedora, CentOS |
| Training hardware | Apple M4 Mac mini, 32GB unified RAM |
## Safety
- Prompt injection detection (exact-phrase matching)
- Catastrophic pattern blocking (`rm -rf /`, fork bombs, pipe-to-shell, etc.)
- Risk classification: `SAFE` / `CAUTION` / `DANGEROUS` / `BLOCKED`
- Zero outbound traffic at runtime
## Requirements
- Linux x86_64 / aarch64
- Python 3.11+
- [`llama-server`](https://github.com/ggerganov/llama.cpp) on `PATH`
- ~1GB RAM at runtime
## Links
- **GitHub:** [0-Time/INCEPT.sh](https://github.com/0-Time/INCEPT.sh)
- **Release:** [v1.0.0](https://github.com/0-Time/INCEPT.sh/releases/tag/v1.0.0)
## License
[Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0)