paulkass commited on
Commit ·
0d7115b
1
Parent(s): c8ff898
Revert "Update README.md"
Browse filesThis reverts commit c8ff8984e348266c906f1f280810195b4e1668a8.
README.md
CHANGED
|
@@ -1,232 +1,232 @@
|
|
| 1 |
-
---
|
| 2 |
-
base_model:
|
| 3 |
-
- fdtn-ai/Foundation-Sec-8B
|
| 4 |
-
language:
|
| 5 |
-
- en
|
| 6 |
-
library_name: transformers
|
| 7 |
-
license: other
|
| 8 |
-
pipeline_tag: text-generation
|
| 9 |
-
tags:
|
| 10 |
-
- security
|
| 11 |
-
- llama
|
| 12 |
-
- fdtn-sec
|
| 13 |
-
---
|
| 14 |
-
# Foundation-Sec-8B-Reasoning - Model Card
|
| 15 |
-
|
| 16 |
-
## Model Information
|
| 17 |
-
|
| 18 |
-
Llama-3.1-FoundationAI-SecurityLLM-8B-Reasoning (Foundation-Sec-8B-Reasoning) is an open-weight, 8-billion parameter instruction-tuned language model specialized for cybersecurity applications.
|
| 19 |
-
It extends the Foundation-Sec-8B base model with instruction-following and reasoning capabilities.
|
| 20 |
-
It leverages prior training to understand security concepts, terminology, and practices across multiple security domains.
|
| 21 |
-
Further reasoning training enables the model to reason about problems before presenting a solution.
|
| 22 |
-
Foundation-Sec-8B-Reasoning enables organizations to build AI-driven security tools that can be deployed locally, reducing dependency on cloud-based AI services while maintaining high performance on security-related tasks.
|
| 23 |
-
|
| 24 |
-
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Reasoning (Foundation-Sec-8B-Reasoning)
|
| 25 |
-
- **Model Developer:** Foundation AI at Cisco
|
| 26 |
-
- **Model Card Contact:** https://fdtn.ai/contact
|
| 27 |
-
- **Technical Report:**
|
| 28 |
-
- **Model Release Date:** January 28th, 2026
|
| 29 |
-
- **Supported Language(s):** English
|
| 30 |
-
- **Model Architecture:** Auto-regressive language model that uses an optimized transformer architecture (Meta Llama-3.1-8B backbone)
|
| 31 |
-
- **Training Objective:** Instruction following and reasoning traces
|
| 32 |
-
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
| 33 |
-
- **License:** See NOTICE.md
|
| 34 |
-
|
| 35 |
-
## Intended Use
|
| 36 |
-
|
| 37 |
-
### Intended Use Cases
|
| 38 |
-
|
| 39 |
-
Foundation-Sec-8B-Reasoning is designed for security practitioners, researchers, and developers building AI-powered security workflows and applications.
|
| 40 |
-
Foundation-Sec-8B-Reasoning is optimized for three core use case categories:
|
| 41 |
-
|
| 42 |
-
- **SOC Acceleration**: Automating triage, summarization, case note generation, and evidence collection.
|
| 43 |
-
- **Proactive Threat Defense**: Simulating attacks, prioritizing vulnerabilities, mapping TTPs, and modeling attacker behavior.
|
| 44 |
-
- **Engineering Enablement**: Providing security assistance, validating configurations, assessing compliance evidence, and improving security posture.
|
| 45 |
-
|
| 46 |
-
The model is intended for local deployment in environments prioritizing data security, regulatory compliance, and operational control.
|
| 47 |
-
|
| 48 |
-
### Downstream Use
|
| 49 |
-
|
| 50 |
-
Foundation-Sec-8B-Reasoning can be used directly for security-related chat use cases. Example downstream applications include:
|
| 51 |
-
|
| 52 |
-
- Summarization
|
| 53 |
-
- Summarizing detection playbooks and incident reports
|
| 54 |
-
- Consolidating fragmented analyst notes into structured case summaries
|
| 55 |
-
- Classification
|
| 56 |
-
- Mapping threats to MITRE ATT&CK techniques
|
| 57 |
-
- Prioritizing vulnerabilities based on contextual risk
|
| 58 |
-
- Classifying security-relevant emails and leaked file contents
|
| 59 |
-
- Named Entity Recognition
|
| 60 |
-
- Extracting compliance evidence from documents
|
| 61 |
-
- Building network behavior profiles from technical manuals
|
| 62 |
-
- Question & Answer
|
| 63 |
-
- Assisting SOC analysts with alert triage and investigation
|
| 64 |
-
- Responding to cloud security and software compliance queries
|
| 65 |
-
- Reasoning and Text Generation
|
| 66 |
-
- Generating red-team attack plans and threat models
|
| 67 |
-
- Predicting attacker next steps in active investigations
|
| 68 |
-
- Enriching vulnerability scan results with contextual insights
|
| 69 |
-
|
| 70 |
-
For questions or assistance with fine-tuning Foundation-Sec-8B-Reasoning, please reach out to the team.
|
| 71 |
-
|
| 72 |
-
### Out-of-Scope Use
|
| 73 |
-
|
| 74 |
-
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
| 75 |
-
|
| 76 |
-
1. **Generating harmful content** - The model should not be used to:
|
| 77 |
-
- Generate malware or other malicious code
|
| 78 |
-
- Create phishing content or social engineering scripts
|
| 79 |
-
- Develop attack plans targeting specific organizations
|
| 80 |
-
- Design exploitation techniques for vulnerabilities without legitimate security research purposes
|
| 81 |
-
2. **Critical security decisions without human oversight** - The model should not be used for:
|
| 82 |
-
- Autonomous security decision-making without human review
|
| 83 |
-
- Critical infrastructure protection without expert supervision
|
| 84 |
-
- Final determination of security compliance without human verification
|
| 85 |
-
- Autonomous vulnerability remediation without testing
|
| 86 |
-
3. **Legal or medical advice** - The model is not qualified to provide:
|
| 87 |
-
- Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
|
| 88 |
-
- Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
|
| 89 |
-
- Medical advice regarding health impacts of security incidents
|
| 90 |
-
4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
|
| 91 |
-
5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
|
| 92 |
-
|
| 93 |
-
## How to Get Started with the Model
|
| 94 |
-
|
| 95 |
-
Use the code below to get started with the model.
|
| 96 |
-
[The cookbook](https://github.com/cisco-foundation-ai/cookbook) provides example use cases, code samples for adoption, and references.
|
| 97 |
-
|
| 98 |
-
```python
|
| 99 |
-
# Import the required libraries
|
| 100 |
-
import torch
|
| 101 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 102 |
-
|
| 103 |
-
# Load the model and tokenizer
|
| 104 |
-
tokenizer = AutoTokenizer.from_pretrained("fdtn-ai/Foundation-Sec-8B-Reasoning")
|
| 105 |
-
model = AutoModelForCausalLM.from_pretrained("fdtn-ai/Foundation-Sec-8B-Reasoning")
|
| 106 |
-
|
| 107 |
-
prompt = "CVE-2015-10011 is a vulnerability about OpenDNS OpenResolve improper log output neutralization. What is the corresponding CWE?"
|
| 108 |
-
|
| 109 |
-
messages = [
|
| 110 |
-
{"role": "user", "content": prompt}
|
| 111 |
-
]
|
| 112 |
-
|
| 113 |
-
model_inputs = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 114 |
-
inputs = tokenizer(model_inputs, return_tensors="pt", add_special_tokens=False)
|
| 115 |
-
output = model.generate(**inputs, temperature=0.1, max_new_tokens=1024)
|
| 116 |
-
resp = tokenizer.batch_decode(output)[0]
|
| 117 |
-
print(resp.replace(model_inputs, ""))
|
| 118 |
-
```
|
| 119 |
-
|
| 120 |
-
## Training and Evaluation
|
| 121 |
-
|
| 122 |
-
### Training Data
|
| 123 |
-
|
| 124 |
-
Foundation-Sec-8B-Reasoning was trained on a wide variety of public and proprietary question answer/pairs for general and security-specific reasoning and instruction-following tasks.
|
| 125 |
-
|
| 126 |
-
**Data cutoff:** April 10th, 2025.
|
| 127 |
-
|
| 128 |
-
A more detailed description of the methodology is available in the technical report.
|
| 129 |
-
|
| 130 |
-
### Training Setup
|
| 131 |
-
|
| 132 |
-
Foundation-Sec-8B-Reasoning is based on the **Llama 3.1 8B** architecture. Training was performed on Cisco Foundation AI’s internal compute cluster.
|
| 133 |
-
|
| 134 |
-
Key training details:
|
| 135 |
-
|
| 136 |
-
- **Instruction fine-tuning** to follow human instructions
|
| 137 |
-
- **RLHF** to align model answers to human preferences
|
| 138 |
-
- **32,768-token** sequence length
|
| 139 |
-
- **Optimizer:** AdamW
|
| 140 |
-
|
| 141 |
-
A more detailed description of the methodology is available in the technical report.
|
| 142 |
-
|
| 143 |
-
### Evaluation
|
| 144 |
-
|
| 145 |
-
Foundation-Sec-8B-Reasoning was benchmarked on cybersecurity and general reasoning tasks, using a standardized 0-shot instruction prompting setup (temperature = 0.3).
|
| 146 |
-
|
| 147 |
-
| **Benchmark** | **Foundation-Sec-8B-Reasoning** | **Llama 3.1 8B** | **GPT-5-Nano** |
|
| 148 |
-
| --- | --- | --- | --- |
|
| 149 |
-
| CTI-MCQA | 0.691 | 0.607 | 0.688 |
|
| 150 |
-
| CTI-RCM | 0.753 | 0.531 | 0.672 |
|
| 151 |
-
| CTI-VSP | 0.856 | 0.811 | 0.822 |
|
| 152 |
-
| CTI-Reasoning | 0.411 | 0.335 | 0.431 |
|
| 153 |
-
|
| 154 |
-
**Benchmark Overview:**
|
| 155 |
-
|
| 156 |
-
- **CTI-MCQA:** 2,500 multiple-choice questions testing cybersecurity knowledge across frameworks like MITRE ATT&CK, NIST, GDPR, and threat intelligence best practices.
|
| 157 |
-
- **CTI-RCM:** 1,000 vulnerability root cause mapping examples linking CVEs to CWE categories, assessing deep understanding of security weaknesses.
|
| 158 |
-
- **CTI-VSP:** A set of 1,000 CVE descriptions where models predict the CVSS v3 Base metrics and compute the overall score, with performance measured by the average absolute difference from the true scores.
|
| 159 |
-
- **IF-Eval:** 541 instruction-following prompts designed for automated, reproducible assessment of LLM instruction-following capabilities.
|
| 160 |
-
- **Alpaca Eval 2:** 805 single-turn prompts auto-scored by GPT-4 Turbo against a GPT-4 Turbo reference, validated with 20,000 human preference votes, and closely matching ChatBot Arena results.
|
| 161 |
-
- **CTI-Reasoning**: An internal benchmark measuring the ability of the model to reason about second-degree connections between MITRE ATT&CK entities.
|
| 162 |
-
|
| 163 |
-
**Key highlights:**
|
| 164 |
-
|
| 165 |
-
- Reasoning traces allow model to **leverage test-time compute** to answer queries
|
| 166 |
-
- **State-of-the-art non-RAG performance** on CTI-RCM benchmark
|
| 167 |
-
- **Better or on-par performance on cyber threat intelligence benchmarks** against GPT-5-Nano
|
| 168 |
-
|
| 169 |
-
For full benchmark details and evaluation methodology, please refer to the technical report.
|
| 170 |
-
|
| 171 |
-
## Safety Alignment
|
| 172 |
-
|
| 173 |
-
Standard best practices were followed to align the model with general safety values.
|
| 174 |
-
Despite the alignment, however, safe out-of-the-box performance cannot be guaranteed.
|
| 175 |
-
Our evaluations show that while the model can achieve reasonable safety performance out-of-the-box, LlamaGuard provides much better protection against malicious requests.
|
| 176 |
-
It is recommended to deploy this model with additional safeguards (such as LlamaGuard) and human oversight.
|
| 177 |
-
|
| 178 |
-
| Model | HarmBench Performance |
|
| 179 |
-
| --- | --- |
|
| 180 |
-
| Llama-3.1-8B-Instruct | 62.75% |
|
| 181 |
-
| Foundation-Sec-8B-Reasoning | 93.00% |
|
| 182 |
-
| **LlamaGuard** + Foundation-Sec-8B-Reasoning | 98.25% |
|
| 183 |
-
|
| 184 |
-
## Limitations
|
| 185 |
-
|
| 186 |
-
Foundation-Sec-8B-Reasoning has several limitations that users should be aware of:
|
| 187 |
-
|
| 188 |
-
1. **Domain-specific knowledge limitations**:
|
| 189 |
-
- Foundation-Sec-8B-Reasoning may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
|
| 190 |
-
- Knowledge of specialized or proprietary security systems or tools may be limited
|
| 191 |
-
2. **Potential biases**:
|
| 192 |
-
- The model may reflect biases present in security literature and documentation
|
| 193 |
-
- The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
|
| 194 |
-
- Security practices and recommendations may be biased toward certain technological ecosystems
|
| 195 |
-
- Geographic and cultural biases in security approaches may be present
|
| 196 |
-
3. **Security risks**:
|
| 197 |
-
- The model cannot verify the identity or intentions of users
|
| 198 |
-
- Adversarial prompting techniques might potentially bypass safety mechanisms
|
| 199 |
-
- The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
|
| 200 |
-
4. **Contextual blindness:**
|
| 201 |
-
- The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
|
| 202 |
-
5. **Technical limitations**:
|
| 203 |
-
- Performance varies based on how security concepts are described in prompts
|
| 204 |
-
- May not fully understand complex, multi-step security scenarios without clear explanation
|
| 205 |
-
- Cannot access external systems or actively scan environments
|
| 206 |
-
- Cannot independently verify factual accuracy of its outputs
|
| 207 |
-
6. **Ethical considerations**:
|
| 208 |
-
- Dual-use nature of security knowledge requires careful consideration of appropriate use cases
|
| 209 |
-
|
| 210 |
-
### Recommendations
|
| 211 |
-
|
| 212 |
-
To address the limitations of Foundation-Sec-8B-Reasoning, we recommend:
|
| 213 |
-
|
| 214 |
-
1. **Human oversight**:
|
| 215 |
-
- Always have qualified security professionals review model outputs before implementation
|
| 216 |
-
- Use the model as an assistive tool rather than a replacement for expert human judgment
|
| 217 |
-
- Implement a human-in-the-loop approach for security-critical applications
|
| 218 |
-
2. **System design safeguards**:
|
| 219 |
-
- Implement additional validation layers for applications built with this model
|
| 220 |
-
- Consider architectural constraints that limit the model’s ability to perform potentially harmful actions (excessive agency)
|
| 221 |
-
- Deploy the model in environments with appropriate access controls
|
| 222 |
-
3. **Prompt engineering**:
|
| 223 |
-
- Use carefully designed prompts that encourage ethical security practices
|
| 224 |
-
- Include explicit instructions regarding responsible disclosure and ethical hacking principles
|
| 225 |
-
- Structure interactions to minimize the risk of inadvertently harmful outputs
|
| 226 |
-
4. **Knowledge supplementation**:
|
| 227 |
-
- Supplement the model with up-to-date security feeds and databases
|
| 228 |
-
- Implement retrieval-augmented generation for current threat intelligence sources
|
| 229 |
-
5. **Usage policies**:
|
| 230 |
-
- Develop and enforce clear acceptable use policies for applications using this model
|
| 231 |
-
- Implement monitoring and auditing for high-risk applications
|
| 232 |
-
- Create documentation for end users about the model’s limitations
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model:
|
| 3 |
+
- fdtn-ai/Foundation-Sec-8B
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
library_name: transformers
|
| 7 |
+
license: other
|
| 8 |
+
pipeline_tag: text-generation
|
| 9 |
+
tags:
|
| 10 |
+
- security
|
| 11 |
+
- llama
|
| 12 |
+
- fdtn-sec
|
| 13 |
+
---
|
| 14 |
+
# Foundation-Sec-8B-Reasoning - Model Card
|
| 15 |
+
|
| 16 |
+
## Model Information
|
| 17 |
+
|
| 18 |
+
Llama-3.1-FoundationAI-SecurityLLM-8B-Reasoning (Foundation-Sec-8B-Reasoning) is an open-weight, 8-billion parameter instruction-tuned language model specialized for cybersecurity applications.
|
| 19 |
+
It extends the Foundation-Sec-8B base model with instruction-following and reasoning capabilities.
|
| 20 |
+
It leverages prior training to understand security concepts, terminology, and practices across multiple security domains.
|
| 21 |
+
Further reasoning training enables the model to reason about problems before presenting a solution.
|
| 22 |
+
Foundation-Sec-8B-Reasoning enables organizations to build AI-driven security tools that can be deployed locally, reducing dependency on cloud-based AI services while maintaining high performance on security-related tasks.
|
| 23 |
+
|
| 24 |
+
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Reasoning (Foundation-Sec-8B-Reasoning)
|
| 25 |
+
- **Model Developer:** Foundation AI at Cisco
|
| 26 |
+
- **Model Card Contact:** https://fdtn.ai/contact
|
| 27 |
+
- **Technical Report:** To be released
|
| 28 |
+
- **Model Release Date:** January 28th, 2026
|
| 29 |
+
- **Supported Language(s):** English
|
| 30 |
+
- **Model Architecture:** Auto-regressive language model that uses an optimized transformer architecture (Meta Llama-3.1-8B backbone)
|
| 31 |
+
- **Training Objective:** Instruction following and reasoning traces
|
| 32 |
+
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
| 33 |
+
- **License:** See NOTICE.md
|
| 34 |
+
|
| 35 |
+
## Intended Use
|
| 36 |
+
|
| 37 |
+
### Intended Use Cases
|
| 38 |
+
|
| 39 |
+
Foundation-Sec-8B-Reasoning is designed for security practitioners, researchers, and developers building AI-powered security workflows and applications.
|
| 40 |
+
Foundation-Sec-8B-Reasoning is optimized for three core use case categories:
|
| 41 |
+
|
| 42 |
+
- **SOC Acceleration**: Automating triage, summarization, case note generation, and evidence collection.
|
| 43 |
+
- **Proactive Threat Defense**: Simulating attacks, prioritizing vulnerabilities, mapping TTPs, and modeling attacker behavior.
|
| 44 |
+
- **Engineering Enablement**: Providing security assistance, validating configurations, assessing compliance evidence, and improving security posture.
|
| 45 |
+
|
| 46 |
+
The model is intended for local deployment in environments prioritizing data security, regulatory compliance, and operational control.
|
| 47 |
+
|
| 48 |
+
### Downstream Use
|
| 49 |
+
|
| 50 |
+
Foundation-Sec-8B-Reasoning can be used directly for security-related chat use cases. Example downstream applications include:
|
| 51 |
+
|
| 52 |
+
- Summarization
|
| 53 |
+
- Summarizing detection playbooks and incident reports
|
| 54 |
+
- Consolidating fragmented analyst notes into structured case summaries
|
| 55 |
+
- Classification
|
| 56 |
+
- Mapping threats to MITRE ATT&CK techniques
|
| 57 |
+
- Prioritizing vulnerabilities based on contextual risk
|
| 58 |
+
- Classifying security-relevant emails and leaked file contents
|
| 59 |
+
- Named Entity Recognition
|
| 60 |
+
- Extracting compliance evidence from documents
|
| 61 |
+
- Building network behavior profiles from technical manuals
|
| 62 |
+
- Question & Answer
|
| 63 |
+
- Assisting SOC analysts with alert triage and investigation
|
| 64 |
+
- Responding to cloud security and software compliance queries
|
| 65 |
+
- Reasoning and Text Generation
|
| 66 |
+
- Generating red-team attack plans and threat models
|
| 67 |
+
- Predicting attacker next steps in active investigations
|
| 68 |
+
- Enriching vulnerability scan results with contextual insights
|
| 69 |
+
|
| 70 |
+
For questions or assistance with fine-tuning Foundation-Sec-8B-Reasoning, please reach out to the team.
|
| 71 |
+
|
| 72 |
+
### Out-of-Scope Use
|
| 73 |
+
|
| 74 |
+
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
| 75 |
+
|
| 76 |
+
1. **Generating harmful content** - The model should not be used to:
|
| 77 |
+
- Generate malware or other malicious code
|
| 78 |
+
- Create phishing content or social engineering scripts
|
| 79 |
+
- Develop attack plans targeting specific organizations
|
| 80 |
+
- Design exploitation techniques for vulnerabilities without legitimate security research purposes
|
| 81 |
+
2. **Critical security decisions without human oversight** - The model should not be used for:
|
| 82 |
+
- Autonomous security decision-making without human review
|
| 83 |
+
- Critical infrastructure protection without expert supervision
|
| 84 |
+
- Final determination of security compliance without human verification
|
| 85 |
+
- Autonomous vulnerability remediation without testing
|
| 86 |
+
3. **Legal or medical advice** - The model is not qualified to provide:
|
| 87 |
+
- Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
|
| 88 |
+
- Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
|
| 89 |
+
- Medical advice regarding health impacts of security incidents
|
| 90 |
+
4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
|
| 91 |
+
5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
|
| 92 |
+
|
| 93 |
+
## How to Get Started with the Model
|
| 94 |
+
|
| 95 |
+
Use the code below to get started with the model.
|
| 96 |
+
[The cookbook](https://github.com/cisco-foundation-ai/cookbook) provides example use cases, code samples for adoption, and references.
|
| 97 |
+
|
| 98 |
+
```python
|
| 99 |
+
# Import the required libraries
|
| 100 |
+
import torch
|
| 101 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 102 |
+
|
| 103 |
+
# Load the model and tokenizer
|
| 104 |
+
tokenizer = AutoTokenizer.from_pretrained("fdtn-ai/Foundation-Sec-8B-Reasoning")
|
| 105 |
+
model = AutoModelForCausalLM.from_pretrained("fdtn-ai/Foundation-Sec-8B-Reasoning")
|
| 106 |
+
|
| 107 |
+
prompt = "CVE-2015-10011 is a vulnerability about OpenDNS OpenResolve improper log output neutralization. What is the corresponding CWE?"
|
| 108 |
+
|
| 109 |
+
messages = [
|
| 110 |
+
{"role": "user", "content": prompt}
|
| 111 |
+
]
|
| 112 |
+
|
| 113 |
+
model_inputs = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 114 |
+
inputs = tokenizer(model_inputs, return_tensors="pt", add_special_tokens=False)
|
| 115 |
+
output = model.generate(**inputs, temperature=0.1, max_new_tokens=1024)
|
| 116 |
+
resp = tokenizer.batch_decode(output)[0]
|
| 117 |
+
print(resp.replace(model_inputs, ""))
|
| 118 |
+
```
|
| 119 |
+
|
| 120 |
+
## Training and Evaluation
|
| 121 |
+
|
| 122 |
+
### Training Data
|
| 123 |
+
|
| 124 |
+
Foundation-Sec-8B-Reasoning was trained on a wide variety of public and proprietary question answer/pairs for general and security-specific reasoning and instruction-following tasks.
|
| 125 |
+
|
| 126 |
+
**Data cutoff:** April 10th, 2025.
|
| 127 |
+
|
| 128 |
+
A more detailed description of the methodology is available in the technical report.
|
| 129 |
+
|
| 130 |
+
### Training Setup
|
| 131 |
+
|
| 132 |
+
Foundation-Sec-8B-Reasoning is based on the **Llama 3.1 8B** architecture. Training was performed on Cisco Foundation AI’s internal compute cluster.
|
| 133 |
+
|
| 134 |
+
Key training details:
|
| 135 |
+
|
| 136 |
+
- **Instruction fine-tuning** to follow human instructions
|
| 137 |
+
- **RLHF** to align model answers to human preferences
|
| 138 |
+
- **32,768-token** sequence length
|
| 139 |
+
- **Optimizer:** AdamW
|
| 140 |
+
|
| 141 |
+
A more detailed description of the methodology is available in the technical report.
|
| 142 |
+
|
| 143 |
+
### Evaluation
|
| 144 |
+
|
| 145 |
+
Foundation-Sec-8B-Reasoning was benchmarked on cybersecurity and general reasoning tasks, using a standardized 0-shot instruction prompting setup (temperature = 0.3).
|
| 146 |
+
|
| 147 |
+
| **Benchmark** | **Foundation-Sec-8B-Reasoning** | **Llama 3.1 8B** | **GPT-5-Nano** |
|
| 148 |
+
| --- | --- | --- | --- |
|
| 149 |
+
| CTI-MCQA | 0.691 | 0.607 | 0.688 |
|
| 150 |
+
| CTI-RCM | 0.753 | 0.531 | 0.672 |
|
| 151 |
+
| CTI-VSP | 0.856 | 0.811 | 0.822 |
|
| 152 |
+
| CTI-Reasoning | 0.411 | 0.335 | 0.431 |
|
| 153 |
+
|
| 154 |
+
**Benchmark Overview:**
|
| 155 |
+
|
| 156 |
+
- **CTI-MCQA:** 2,500 multiple-choice questions testing cybersecurity knowledge across frameworks like MITRE ATT&CK, NIST, GDPR, and threat intelligence best practices.
|
| 157 |
+
- **CTI-RCM:** 1,000 vulnerability root cause mapping examples linking CVEs to CWE categories, assessing deep understanding of security weaknesses.
|
| 158 |
+
- **CTI-VSP:** A set of 1,000 CVE descriptions where models predict the CVSS v3 Base metrics and compute the overall score, with performance measured by the average absolute difference from the true scores.
|
| 159 |
+
- **IF-Eval:** 541 instruction-following prompts designed for automated, reproducible assessment of LLM instruction-following capabilities.
|
| 160 |
+
- **Alpaca Eval 2:** 805 single-turn prompts auto-scored by GPT-4 Turbo against a GPT-4 Turbo reference, validated with 20,000 human preference votes, and closely matching ChatBot Arena results.
|
| 161 |
+
- **CTI-Reasoning**: An internal benchmark measuring the ability of the model to reason about second-degree connections between MITRE ATT&CK entities.
|
| 162 |
+
|
| 163 |
+
**Key highlights:**
|
| 164 |
+
|
| 165 |
+
- Reasoning traces allow model to **leverage test-time compute** to answer queries
|
| 166 |
+
- **State-of-the-art non-RAG performance** on CTI-RCM benchmark
|
| 167 |
+
- **Better or on-par performance on cyber threat intelligence benchmarks** against GPT-5-Nano
|
| 168 |
+
|
| 169 |
+
For full benchmark details and evaluation methodology, please refer to the technical report.
|
| 170 |
+
|
| 171 |
+
## Safety Alignment
|
| 172 |
+
|
| 173 |
+
Standard best practices were followed to align the model with general safety values.
|
| 174 |
+
Despite the alignment, however, safe out-of-the-box performance cannot be guaranteed.
|
| 175 |
+
Our evaluations show that while the model can achieve reasonable safety performance out-of-the-box, LlamaGuard provides much better protection against malicious requests.
|
| 176 |
+
It is recommended to deploy this model with additional safeguards (such as LlamaGuard) and human oversight.
|
| 177 |
+
|
| 178 |
+
| Model | HarmBench Performance |
|
| 179 |
+
| --- | --- |
|
| 180 |
+
| Llama-3.1-8B-Instruct | 62.75% |
|
| 181 |
+
| Foundation-Sec-8B-Reasoning | 93.00% |
|
| 182 |
+
| **LlamaGuard** + Foundation-Sec-8B-Reasoning | 98.25% |
|
| 183 |
+
|
| 184 |
+
## Limitations
|
| 185 |
+
|
| 186 |
+
Foundation-Sec-8B-Reasoning has several limitations that users should be aware of:
|
| 187 |
+
|
| 188 |
+
1. **Domain-specific knowledge limitations**:
|
| 189 |
+
- Foundation-Sec-8B-Reasoning may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
|
| 190 |
+
- Knowledge of specialized or proprietary security systems or tools may be limited
|
| 191 |
+
2. **Potential biases**:
|
| 192 |
+
- The model may reflect biases present in security literature and documentation
|
| 193 |
+
- The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
|
| 194 |
+
- Security practices and recommendations may be biased toward certain technological ecosystems
|
| 195 |
+
- Geographic and cultural biases in security approaches may be present
|
| 196 |
+
3. **Security risks**:
|
| 197 |
+
- The model cannot verify the identity or intentions of users
|
| 198 |
+
- Adversarial prompting techniques might potentially bypass safety mechanisms
|
| 199 |
+
- The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
|
| 200 |
+
4. **Contextual blindness:**
|
| 201 |
+
- The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
|
| 202 |
+
5. **Technical limitations**:
|
| 203 |
+
- Performance varies based on how security concepts are described in prompts
|
| 204 |
+
- May not fully understand complex, multi-step security scenarios without clear explanation
|
| 205 |
+
- Cannot access external systems or actively scan environments
|
| 206 |
+
- Cannot independently verify factual accuracy of its outputs
|
| 207 |
+
6. **Ethical considerations**:
|
| 208 |
+
- Dual-use nature of security knowledge requires careful consideration of appropriate use cases
|
| 209 |
+
|
| 210 |
+
### Recommendations
|
| 211 |
+
|
| 212 |
+
To address the limitations of Foundation-Sec-8B-Reasoning, we recommend:
|
| 213 |
+
|
| 214 |
+
1. **Human oversight**:
|
| 215 |
+
- Always have qualified security professionals review model outputs before implementation
|
| 216 |
+
- Use the model as an assistive tool rather than a replacement for expert human judgment
|
| 217 |
+
- Implement a human-in-the-loop approach for security-critical applications
|
| 218 |
+
2. **System design safeguards**:
|
| 219 |
+
- Implement additional validation layers for applications built with this model
|
| 220 |
+
- Consider architectural constraints that limit the model’s ability to perform potentially harmful actions (excessive agency)
|
| 221 |
+
- Deploy the model in environments with appropriate access controls
|
| 222 |
+
3. **Prompt engineering**:
|
| 223 |
+
- Use carefully designed prompts that encourage ethical security practices
|
| 224 |
+
- Include explicit instructions regarding responsible disclosure and ethical hacking principles
|
| 225 |
+
- Structure interactions to minimize the risk of inadvertently harmful outputs
|
| 226 |
+
4. **Knowledge supplementation**:
|
| 227 |
+
- Supplement the model with up-to-date security feeds and databases
|
| 228 |
+
- Implement retrieval-augmented generation for current threat intelligence sources
|
| 229 |
+
5. **Usage policies**:
|
| 230 |
+
- Develop and enforce clear acceptable use policies for applications using this model
|
| 231 |
+
- Implement monitoring and auditing for high-risk applications
|
| 232 |
+
- Create documentation for end users about the model’s limitations
|