0Time commited on
Commit
7855fd1
·
verified ·
1 Parent(s): 933029a

Update model card: INCEPT.sh v1.0.0, new GitHub URL (0-Time), install.sh, /think toggle

Browse files
Files changed (1) hide show
  1. README.md +60 -23
README.md CHANGED
@@ -11,37 +11,62 @@ tags:
11
  base_model: Qwen/Qwen3.5-0.8B
12
  ---
13
 
14
- # INCEPT/Sh — Command Inference Model (Q8_0)
15
 
16
- Fine-tuned **Qwen3.5-0.8B** (GGUF Q8_0, 774MB) for Linux shell command generation. Maps plain English descriptions to accurate Linux commands. Runs entirely offline via [llama.cpp](https://github.com/ggerganov/llama.cpp).
17
 
18
  **Benchmark:** 99/100 on a structured 100-question Linux command evaluation (Ubuntu 22.04, bash, non-root).
19
 
20
- ## Usage
 
 
 
 
21
 
22
- Requires `llama-server` from [llama.cpp](https://github.com/ggerganov/llama.cpp).
 
 
23
 
24
  ```bash
25
  # Download model
26
  huggingface-cli download 0Time/INCEPT-SH \
27
- incept-command-v2-q8_0.gguf --local-dir ./models
28
 
29
- # Run with INCEPT/Sh CLI
30
- git clone https://github.com/ProMohanad/INCEPT.sh
31
  cd INCEPT.sh
32
  pip install -e ".[cli]"
33
  incept
34
  ```
35
 
36
- Or use directly with llama-server:
37
 
38
  ```bash
39
- llama-server -m models/incept-command-v2-q8_0.gguf --port 8787
 
 
 
 
 
 
 
 
 
 
40
  ```
41
 
 
 
 
 
 
 
 
 
 
42
  ## Prompt Format
43
 
44
- The model uses ChatML format with a system context line:
45
 
46
  ```
47
  <|im_start|>system
@@ -60,23 +85,35 @@ Inference temperature: **0.0** (greedy decoding).
60
 
61
  ## Training
62
 
63
- | Parameter | Value |
64
- |-----------------------|---------------------------------------------|
65
- | Base model | Qwen/Qwen3.5-0.8B |
66
- | Training method | Supervised fine-tuning (LoRA, rank 16) |
67
- | Training examples | 79,264 (SFT) + 11,306 (pipe refinement) |
68
- | Learning rate | 5×10⁻⁵ |
69
- | Quantization | Q8_0 (774MB) |
70
- | Supported distros | Ubuntu, Debian, RHEL, Arch, Fedora, CentOS |
71
- | Training hardware | Apple M4 Mac mini, 32GB unified RAM |
72
 
73
  ## Safety
74
 
75
- The accompanying [INCEPT/Sh](https://github.com/ProMohanad/INCEPT.sh) engine includes:
76
- - Prompt injection detection
77
- - Catastrophic pattern blocking (fork bombs, `rm -rf /`, pipe-to-shell, etc.)
78
  - Risk classification: `SAFE` / `CAUTION` / `DANGEROUS` / `BLOCKED`
 
 
 
 
 
 
 
 
 
 
 
 
 
79
 
80
  ## License
81
 
82
- [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)
 
11
  base_model: Qwen/Qwen3.5-0.8B
12
  ---
13
 
14
+ # INCEPT.sh
15
 
16
+ Offline command inference engine for Linux. Fine-tuned **Qwen3.5-0.8B** (GGUF Q8_0, 774MB) via supervised fine-tuning on 79K Linux command examples. Zero network dependency at runtime.
17
 
18
  **Benchmark:** 99/100 on a structured 100-question Linux command evaluation (Ubuntu 22.04, bash, non-root).
19
 
20
+ ## Installation
21
+
22
+ ```bash
23
+ curl -fsSL https://raw.githubusercontent.com/0-Time/INCEPT.sh/main/install.sh | bash
24
+ ```
25
 
26
+ Supports: Debian/Ubuntu, RHEL/Fedora, CentOS, Arch, openSUSE.
27
+
28
+ ## Manual Model Setup
29
 
30
  ```bash
31
  # Download model
32
  huggingface-cli download 0Time/INCEPT-SH \
33
+ incept-sh.gguf --local-dir ./models
34
 
35
+ # Clone and install
36
+ git clone https://github.com/0-Time/INCEPT.sh
37
  cd INCEPT.sh
38
  pip install -e ".[cli]"
39
  incept
40
  ```
41
 
42
+ ## Usage
43
 
44
  ```bash
45
+ # Interactive CLI
46
+ incept
47
+
48
+ # One-shot
49
+ incept -c "list all open ports"
50
+
51
+ # Minimal output (pipe-friendly)
52
+ incept -c "find large files" -m
53
+
54
+ # With model reasoning
55
+ incept --think
56
  ```
57
 
58
+ ## CLI Commands
59
+
60
+ | Command | Description |
61
+ |---|---|
62
+ | `/think on\|off` | Toggle chain-of-thought reasoning |
63
+ | `/context` | Show detected system context |
64
+ | `/help` | List available commands |
65
+ | `/exit` | Exit |
66
+
67
  ## Prompt Format
68
 
69
+ ChatML with a system context line:
70
 
71
  ```
72
  <|im_start|>system
 
85
 
86
  ## Training
87
 
88
+ | Parameter | Value |
89
+ |-----------------------|----------------------------------------------|
90
+ | Base model | Qwen/Qwen3.5-0.8B |
91
+ | Training method | Supervised fine-tuning (LoRA, rank 16) |
92
+ | Training examples | 79,264 (SFT) + 11,306 (pipe refinement) |
93
+ | Learning rate | 5×10⁻⁵ |
94
+ | Quantization | Q8_0 (774MB) |
95
+ | Supported distros | Ubuntu, Debian, RHEL, Arch, Fedora, CentOS |
96
+ | Training hardware | Apple M4 Mac mini, 32GB unified RAM |
97
 
98
  ## Safety
99
 
100
+ - Prompt injection detection (exact-phrase matching)
101
+ - Catastrophic pattern blocking (`rm -rf /`, fork bombs, pipe-to-shell, etc.)
 
102
  - Risk classification: `SAFE` / `CAUTION` / `DANGEROUS` / `BLOCKED`
103
+ - Zero outbound traffic at runtime
104
+
105
+ ## Requirements
106
+
107
+ - Linux x86_64 / aarch64
108
+ - Python 3.11+
109
+ - [`llama-server`](https://github.com/ggerganov/llama.cpp) on `PATH`
110
+ - ~1GB RAM at runtime
111
+
112
+ ## Links
113
+
114
+ - **GitHub:** [0-Time/INCEPT.sh](https://github.com/0-Time/INCEPT.sh)
115
+ - **Release:** [v1.0.0](https://github.com/0-Time/INCEPT.sh/releases/tag/v1.0.0)
116
 
117
  ## License
118
 
119
+ [Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0)