0Time commited on
Commit
53da7e7
·
verified ·
1 Parent(s): e0bbc7c

Add model card

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: apache-2.0
4
+ tags:
5
+ - linux
6
+ - command-generation
7
+ - gguf
8
+ - qwen3
9
+ - llama-cpp
10
+ - offline
11
+ base_model: Qwen/Qwen3.5-0.8B
12
+ ---
13
+
14
+ # INCEPT/Sh — Command Inference Model (Q8_0)
15
+
16
+ Fine-tuned **Qwen3.5-0.8B** (GGUF Q8_0, 774MB) for Linux shell command generation. Maps plain English descriptions to accurate Linux commands. Runs entirely offline via [llama.cpp](https://github.com/ggerganov/llama.cpp).
17
+
18
+ **Benchmark:** 99/100 on a structured 100-question Linux command evaluation (Ubuntu 22.04, bash, non-root).
19
+
20
+ ## Usage
21
+
22
+ Requires `llama-server` from [llama.cpp](https://github.com/ggerganov/llama.cpp).
23
+
24
+ ```bash
25
+ # Download model
26
+ huggingface-cli download 0Time/INCEPT-SH \
27
+ incept-command-v2-q8_0.gguf --local-dir ./models
28
+
29
+ # Run with INCEPT/Sh CLI
30
+ git clone https://github.com/ProMohanad/INCEPT.sh
31
+ cd INCEPT.sh
32
+ pip install -e ".[cli]"
33
+ incept
34
+ ```
35
+
36
+ Or use directly with llama-server:
37
+
38
+ ```bash
39
+ llama-server -m models/incept-command-v2-q8_0.gguf --port 8787
40
+ ```
41
+
42
+ ## Prompt Format
43
+
44
+ The model uses ChatML format with a system context line:
45
+
46
+ ```
47
+ <|im_start|>system
48
+ ubuntu 22.04 bash non-root
49
+ <|im_end|>
50
+ <|im_start|>user
51
+ {natural language query}
52
+ <|im_end|>
53
+ <|im_start|>assistant
54
+ <think>
55
+ </think>
56
+
57
+ ```
58
+
59
+ Inference temperature: **0.0** (greedy decoding).
60
+
61
+ ## Training
62
+
63
+ | Parameter | Value |
64
+ |-----------------------|---------------------------------------------|
65
+ | Base model | Qwen/Qwen3.5-0.8B |
66
+ | Training method | Supervised fine-tuning (LoRA, rank 16) |
67
+ | Training examples | 79,264 (SFT) + 11,306 (pipe refinement) |
68
+ | Learning rate | 5×10⁻⁵ |
69
+ | Quantization | Q8_0 (774MB) |
70
+ | Supported distros | Ubuntu, Debian, RHEL, Arch, Fedora, CentOS |
71
+ | Training hardware | Apple M4 Mac mini, 32GB unified RAM |
72
+
73
+ ## Safety
74
+
75
+ The accompanying [INCEPT/Sh](https://github.com/ProMohanad/INCEPT.sh) engine includes:
76
+ - Prompt injection detection
77
+ - Catastrophic pattern blocking (fork bombs, `rm -rf /`, pipe-to-shell, etc.)
78
+ - Risk classification: `SAFE` / `CAUTION` / `DANGEROUS` / `BLOCKED`
79
+
80
+ ## License
81
+
82
+ [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)