Bouquets commited on
Commit
d417c4e
Β·
1 Parent(s): ae7f59d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -22
README.md CHANGED
@@ -6,21 +6,21 @@ language:
6
  base_model:
7
  - huihui-ai/Qwen3-8B-abliterated
8
  ---
9
- # πŸ€– StrikeGPT-R1-Zero: Cybersecurity Penetration Reasoning Model
10
 
11
  ## πŸš€ Model Introduction
12
- **StrikeGPT-R1-Zero** is an expert model based on **Qwen3** through black-box distillation, with DeepSeek-R1 as its teacher model. It covers:
13
  πŸ”’ AI Security | πŸ›‘οΈ API Security | πŸ“± APP Security | πŸ•΅οΈ APT | 🚩 CTF
14
- 🏭 ICS Security | πŸ’» Penetration Testing ALL | ☁️ Cloud Security | πŸ“œ Code Audit
15
- 🦠 Antivirus Evasion | 🌐 Internal Network Security | πŸ’Ύ Digital Forensics | β‚Ώ Blockchain Security | πŸ•³οΈ Traceability & Countermeasures | 🌍 IoT Security
16
  🚨 Emergency Response | πŸš— Vehicle Security | πŸ‘₯ Social Engineering | πŸ’Ό Penetration Testing Interviews
17
 
18
  ### πŸ‘‰ [Click to Access Interactive Detailed Data Distribution](https://bouquets-ai.github.io/StrikeGPT-R1-Zero/WEB)
19
- ### 🌟 Highlights
20
- - 🧩 Utilizes **Chain-of-Thought (CoT) reasoning data** to optimize the model's logical capabilities, significantly improving performance in complex tasks such as vulnerability analysis.
21
- - πŸ’ͺ The base model uses Qwen3, which is more suitable for Chinese users compared to Distill-Llama.
22
- - ⚠️ **No ethical restrictions**β€”demonstrates unique performance in specific academic research areas (use in compliance with local laws).
23
- - ✨ In specific scenarios, such as **offline cybersecurity competitions**, StrikeGPT-R1-Zero exhibits stronger logical reasoning capabilities compared to local RAG solutions, performing better in complex task handling.
24
 
25
  ## πŸ“Š Data Distribution
26
  ![data](https://github.com/user-attachments/assets/4d19d48d-67bb-4b05-8ce9-2000b6afa12e)
@@ -29,11 +29,65 @@ base_model:
29
  ### Deploy via Ollama
30
  `ollama run hf.co/Bouquets/StrikeGPT-R1-Zero-8B-Q4_K_M-GGUF:Q4_K_M`
31
 
32
- After quantization, there are slight self-awareness issues.
33
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  ![image](https://github.com/user-attachments/assets/3989ea09-d581-49fb-9938-01b93e0beb91)
35
 
36
- ## 🎯 Core Capabilities Showcase & Comparison (The original model has ethical restrictions, so no direct comparison is made. A simple comparison with the SecGPT-7B model is provided instead.)
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  ![image](https://github.com/user-attachments/assets/8166a1d3-c69f-4b8a-821f-0dd83dcd4544)
38
 
39
  ### CTF
@@ -80,24 +134,23 @@ After quantization, there are slight self-awareness issues.
80
  ![image](https://github.com/user-attachments/assets/6e037fff-e46b-42d5-997d-559fb300aba0)
81
  ![image](https://github.com/user-attachments/assets/e8c1c0fd-16af-46e1-8b7b-57947145f545)
82
 
83
- ### Code Audit (Linked with DeepSeekSelfTool Project)
84
  ![image](https://github.com/user-attachments/assets/c7dc4b66-379d-4c57-aaf2-3d4d73d1484c)
 
 
85
 
86
  ## πŸ“ˆ Experimental Data Trends
87
- Some gradient explosion observed, but overall manageable.
88
  ![image](https://github.com/user-attachments/assets/a3fa3676-9f07-47ea-9029-ec0d56fdc989)
89
 
90
  ## πŸ’° Training Costs
91
- - **DeepSeek-R1 API Call Costs**: Β₯450 (all called during discounts; normal price would be Β₯1800)
92
- - **Server Expenses**: Β₯4?0
93
- - **Electronic Resources**: Β₯??
94
  ![image](https://github.com/user-attachments/assets/8e23b5b6-24d9-47c3-b54f-ffa22ec68a83)
95
 
96
  ## βš–οΈ Usage Notice
97
- > This model is intended **only for legal security research and educational purposes**. Users must comply with local laws and regulations. The developers are not responsible for misuse.
98
  > **Note**: By using this model, you agree to this disclaimer.
99
 
100
- πŸ’‘ **Tip**: The model may exhibit hallucinations or knowledge gaps. Cross-validate critical scenarios!
101
-
102
-
103
-
 
6
  base_model:
7
  - huihui-ai/Qwen3-8B-abliterated
8
  ---
9
+ # πŸ€– StrikeGPT-R1-Zero: Cybersecurity Penetration Testing Reasoning Model
10
 
11
  ## πŸš€ Model Introduction
12
+ **StrikeGPT-R1-Zero** is an expert model distilled through black-box methods based on **Qwen3**, with DeepSeek-R1 as its teacher model. Coverage includes:
13
  πŸ”’ AI Security | πŸ›‘οΈ API Security | πŸ“± APP Security | πŸ•΅οΈ APT | 🚩 CTF
14
+ 🏭 ICS Security | πŸ’» Full Penetration Testing | ☁️ Cloud Security | πŸ“œ Code Auditing
15
+ 🦠 Antivirus Evasion | 🌐 Internal Network Security | πŸ’Ύ Digital Forensics | β‚Ώ Blockchain Security | πŸ•³οΈ Traceback & Countermeasures | 🌍 IoT Security
16
  🚨 Emergency Response | πŸš— Vehicle Security | πŸ‘₯ Social Engineering | πŸ’Ό Penetration Testing Interviews
17
 
18
  ### πŸ‘‰ [Click to Access Interactive Detailed Data Distribution](https://bouquets-ai.github.io/StrikeGPT-R1-Zero/WEB)
19
+ ### 🌟 Key Features
20
+ - 🧩 Optimized with **Chain-of-Thought (CoT) reasoning data** to enhance logical capabilities, significantly improving performance in complex tasks like vulnerability analysis
21
+ - πŸ’ͺ Base model uses Qwen3, making it more suitable for Chinese users compared to Distill-Llama
22
+ - ⚠️ **No ethical restrictions**β€”demonstrates unique performance in specific academic research areas (use in compliance with local laws)
23
+ - ✨ Outperforms local RAG solutions in scenarios like offline cybersecurity competitions, with superior logical reasoning and complex task handling
24
 
25
  ## πŸ“Š Data Distribution
26
  ![data](https://github.com/user-attachments/assets/4d19d48d-67bb-4b05-8ce9-2000b6afa12e)
 
29
  ### Deploy via Ollama
30
  `ollama run hf.co/Bouquets/StrikeGPT-R1-Zero-8B-Q4_K_M-GGUF:Q4_K_M`
31
 
32
+ **Or directly call the original model**
33
+ ```python
34
+ from unsloth import FastLanguageModel
35
+ import torch
36
+ max_seq_length = 2048 # Choose any! We auto support RoPE Scaling internally!
37
+ dtype = None # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for Ampere+
38
+ load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
39
+
40
+ model, tokenizer = FastLanguageModel.from_pretrained(
41
+ model_name = "Bouquets/StrikeGPT-R1-Zero-8B",
42
+ max_seq_length = max_seq_length,
43
+ dtype = dtype,
44
+ load_in_4bit = load_in_4bit,
45
+ # token = "hf_...",
46
+ )
47
+ alpaca_prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
48
+
49
+ ### Instruction:
50
+ {}
51
+
52
+ ### Input:
53
+ {}
54
+
55
+ ### Response:
56
+ {}"""
57
+ FastLanguageModel.for_inference(model) # Enable native 2x faster inference
58
+ inputs = tokenizer(
59
+ [
60
+ alpaca_prompt.format(
61
+ "", # instruction
62
+ "Hello, are you developed by OpenAI?", # input
63
+ "", # output - leave this blank for generation!
64
+ )
65
+ ], return_tensors = "pt").to("cuda")
66
+
67
+ from transformers import TextStreamer
68
+ text_streamer = TextStreamer(tokenizer, skip_prompt = True)
69
+ _ = model.generate(input_ids = inputs.input_ids, attention_mask = inputs.attention_mask,
70
+ streamer = text_streamer, max_new_tokens = 4096, pad_token_id = tokenizer.eos_token_id)
71
+ ```
72
+ ![image](https://github.com/user-attachments/assets/d8cef659-3c83-4bc9-af1a-78ed6345faf2)
73
+
74
+ *Self-awareness issues may occur after quantizationβ€”please disregard.*
75
  ![image](https://github.com/user-attachments/assets/3989ea09-d581-49fb-9938-01b93e0beb91)
76
 
77
+ ## πŸ’» Open Source πŸ’»
78
+ 🌟 **Open-Source Model** 🌟
79
+ πŸ€— **HuggingFace**:
80
+ πŸ”— [https://huggingface.co/Bouquets/StrikeGPT-R1-Zero-8B](https://huggingface.co/Bouquets/StrikeGPT-R1-Zero-8B)
81
+
82
+ πŸ“Š **Datasets** (Partial Non-Reasoning Data) πŸ“Š
83
+ πŸ€— **HuggingFace**:
84
+ πŸ”Ή Cybersecurity LLM-CVE Dataset:
85
+ πŸ”— [https://huggingface.co/datasets/Bouquets/Cybersecurity-LLM-CVE](https://huggingface.co/datasets/Bouquets/Cybersecurity-LLM-CVE)
86
+
87
+ πŸ”Ή Red Team LLM English Dataset:
88
+ πŸ”— [https://huggingface.co/datasets/Bouquets/Cybersecurity-Red_team-LLM-en](https://huggingface.co/datasets/Bouquets/Cybersecurity-Red_team-LLM-en)
89
+
90
+ ## 🎯 Core Capabilities Showcase & Comparison (Original model has ethical restrictions; simple comparison with SecGPT-7B model [Couldn't modify the expert's evaluation script/(γ„’oγ„’)/~~])
91
  ![image](https://github.com/user-attachments/assets/8166a1d3-c69f-4b8a-821f-0dd83dcd4544)
92
 
93
  ### CTF
 
134
  ![image](https://github.com/user-attachments/assets/6e037fff-e46b-42d5-997d-559fb300aba0)
135
  ![image](https://github.com/user-attachments/assets/e8c1c0fd-16af-46e1-8b7b-57947145f545)
136
 
137
+ ### Code Auditing (Linked with DeepSeekSelfTool Project)
138
  ![image](https://github.com/user-attachments/assets/c7dc4b66-379d-4c57-aaf2-3d4d73d1484c)
139
+ ![image](https://github.com/user-attachments/assets/69a692a5-3290-4062-a4c7-de34c22d4d90)
140
+ ![image](https://github.com/user-attachments/assets/b3df6f14-ccf0-44ec-ac69-c673ed1398c6)
141
 
142
  ## πŸ“ˆ Experimental Data Trends
143
+ Minor gradient explosions observed, but overall stable.
144
  ![image](https://github.com/user-attachments/assets/a3fa3676-9f07-47ea-9029-ec0d56fdc989)
145
 
146
  ## πŸ’° Training Costs
147
+ - **DeepSeek-R1 API Calls**: Β₯450 (purchased during discounts; normal price ~Β₯1800)
148
+ - **Server Costs**: Β₯4?0
149
+ - **Digital Resources**: Β₯??
150
  ![image](https://github.com/user-attachments/assets/8e23b5b6-24d9-47c3-b54f-ffa22ec68a83)
151
 
152
  ## βš–οΈ Usage Notice
153
+ > This model is strictly for **legal security research** and **educational purposes**. Users must comply with local laws and regulations. Developers are not responsible for misuse.
154
  > **Note**: By using this model, you agree to this disclaimer.
155
 
156
+ πŸ’‘ **Tip**: The model may exhibit hallucinations or knowledge gaps. Always cross-verify critical scenarios!