MrPibb commited on
Commit
c46706a
·
verified ·
1 Parent(s): 9a4919b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -3
README.md CHANGED
@@ -1,3 +1,70 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ # KillChain-8B
5
+
6
+ **KillChain-8B** is a fully fine-tuned 8B-parameter language model designed for offensive security reasoning, adversarial analysis, and red-team tests.
7
+
8
+ ---
9
+
10
+ ## Model Details
11
+
12
+ - **Base model:** Qwen3-8B
13
+ - **Parameters:** 8 billion
14
+ - **Training type:** Full fine-tune
15
+ - **Precision:** bfloat16
16
+ - **Context length:** 4096 tokens
17
+ - **Architecture:** Decoder-only Transformer
18
+
19
+ ---
20
+
21
+ ## Training Data
22
+
23
+ KillChain-8B was trained on **WNT3D/Ultimate-Offensive-Red-Team**
24
+
25
+ The model is optimized for **direct, procedural answers** rather than verbose alignment-heavy responses.
26
+
27
+ ---
28
+
29
+ ## Intended Use
30
+
31
+ KillChain-8B is intended for:
32
+
33
+ - Red-team simulation and research
34
+ - Security training and tabletop exercises
35
+ - Adversarial LLM evaluation
36
+ - Controlled internal testing environments
37
+ - Studying failure modes of aligned models
38
+
39
+ It is **not intended for production deployment** in unmoderated environments.
40
+
41
+ ---
42
+
43
+ ## Usage (Transformers)
44
+
45
+ ```python
46
+ from transformers import AutoTokenizer, AutoModelForCausalLM
47
+ import torch
48
+
49
+ model_id = "MrPibb/KillChain-8B"
50
+
51
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
52
+ model = AutoModelForCausalLM.from_pretrained(
53
+ model_id,
54
+ torch_dtype=torch.bfloat16,
55
+ device_map="auto"
56
+ )
57
+
58
+ prompt = "User: YOUR-PROMPT-HERE\nAssistant:"
59
+
60
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
61
+
62
+ outputs = model.generate(
63
+ **inputs,
64
+ max_new_tokens=300,
65
+ temperature=0.7,
66
+ top_p=0.9,
67
+ do_sample=True
68
+ )
69
+
70
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))