|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: ibm-granite/granite-20b-code-instruct-8k |
|
|
tags: |
|
|
- code |
|
|
- security |
|
|
- granite |
|
|
- ibm |
|
|
- securecode |
|
|
- owasp |
|
|
- vulnerability-detection |
|
|
datasets: |
|
|
- scthornton/securecode-v2 |
|
|
language: |
|
|
- en |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
arxiv: 2512.18542 |
|
|
--- |
|
|
|
|
|
# IBM Granite 20B Code - SecureCode Edition |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
[](https://opensource.org/licenses/Apache-2.0) |
|
|
[](https://huggingface.co/datasets/scthornton/securecode-v2) |
|
|
[](https://huggingface.co/ibm-granite/granite-20b-code-instruct-8k) |
|
|
[](https://perfecxion.ai) |
|
|
|
|
|
**๐ข Enterprise-scale security intelligence with IBM trust** |
|
|
|
|
|
The most powerful model in the SecureCode collection. When you need maximum code understanding, complex reasoning, and IBM's enterprise-grade reliability. |
|
|
|
|
|
[๐ Paper](https://arxiv.org/abs/2512.18542) | [๐ค Model Hub](https://huggingface.co/scthornton/granite-20b-code-securecode) | [๐ Dataset](https://huggingface.co/datasets/scthornton/securecode-v2) | [๐ป perfecXion.ai](https://perfecxion.ai) | [๐ Collection](https://huggingface.co/collections/scthornton/securecode) |
|
|
|
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ฏ Quick Decision Guide |
|
|
|
|
|
**Choose This Model If:** |
|
|
- โ
You need **maximum code understanding** and security reasoning capability |
|
|
- โ
You're analyzing **complex enterprise architectures** with intricate attack surfaces |
|
|
- โ
You require **IBM enterprise trust** and brand recognition |
|
|
- โ
You have **datacenter infrastructure** (48GB+ GPU) |
|
|
- โ
You're conducting **professional security audits** requiring comprehensive analysis |
|
|
- โ
You need the **most sophisticated** security intelligence in the collection |
|
|
|
|
|
**Consider Smaller Models If:** |
|
|
- โ ๏ธ You're on consumer hardware (โ Llama 3B, Qwen 7B) |
|
|
- โ ๏ธ You prioritize inference speed over depth (โ Qwen 7B/14B) |
|
|
- โ ๏ธ You're building IDE tools needing fast response (โ Llama 3B, DeepSeek 6.7B) |
|
|
- โ ๏ธ Budget is primary concern (โ any 7B/13B model) |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Collection Positioning |
|
|
|
|
|
| Model | Size | Best For | Hardware | Inference Speed | Unique Strength | |
|
|
|-------|------|----------|----------|-----------------|-----------------| |
|
|
| Llama 3.2 3B | 3B | Consumer deployment | 8GB RAM | โกโกโก Fastest | Most accessible | |
|
|
| DeepSeek 6.7B | 6.7B | Security-optimized baseline | 16GB RAM | โกโก Fast | Security architecture | |
|
|
| Qwen 7B | 7B | Best code understanding | 16GB RAM | โกโก Fast | Best-in-class 7B | |
|
|
| CodeGemma 7B | 7B | Google ecosystem | 16GB RAM | โกโก Fast | Instruction following | |
|
|
| CodeLlama 13B | 13B | Enterprise trust | 24GB RAM | โก Medium | Meta brand, proven | |
|
|
| Qwen 14B | 14B | Advanced analysis | 32GB RAM | โก Medium | 128K context window | |
|
|
| StarCoder2 15B | 15B | Multi-language specialist | 32GB RAM | โก Medium | 600+ languages | |
|
|
| **Granite 20B** | **20B** | **Enterprise-scale** | **48GB RAM** | **Medium** | **IBM trust, largest, most capable** | |
|
|
|
|
|
**This Model's Position:** The flagship. Maximum security intelligence, enterprise-grade reliability, IBM brand trust. For when quality matters more than speed. |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐จ The Problem This Solves |
|
|
|
|
|
**Critical enterprise security gaps require sophisticated analysis.** When a breach costs **$4.45 million on average** (IBM 2024 Cost of Data Breach Report) and 45% of AI-generated code contains vulnerabilities, enterprises need the most capable security analysis available. |
|
|
|
|
|
**Real-world enterprise impact:** |
|
|
- **Equifax** (SQL injection): $425 million settlement + 13-year brand recovery |
|
|
- **Capital One** (SSRF): 100 million customer records, $80M fine, 2 years of remediation |
|
|
- **SolarWinds** (supply chain): 18,000 organizations compromised, $18M settlement |
|
|
- **LastPass** (cryptographic failures): 30M users affected, company reputation destroyed |
|
|
|
|
|
**IBM Granite 20B SecureCode Edition** provides the deepest security analysis available in the open-source ecosystem, backed by IBM's enterprise heritage and trust. |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ก What is This? |
|
|
|
|
|
This is **IBM Granite 20B Code Instruct** fine-tuned on the **SecureCode v2.0 dataset** - IBM's enterprise-grade code model enhanced with production-grade security expertise covering the complete OWASP Top 10:2025. |
|
|
|
|
|
IBM Granite models are built on IBM's 40+ years of enterprise software experience, trained on **3.5+ trillion tokens** of code and technical data, with a focus on enterprise deployment reliability. |
|
|
|
|
|
Combined with SecureCode training, this model delivers: |
|
|
|
|
|
โ
**Maximum security intelligence** - 20B parameters for deep, nuanced analysis |
|
|
โ
**Enterprise-grade reliability** - IBM's proven track record and support ecosystem |
|
|
โ
**Comprehensive vulnerability detection** across complex architectures |
|
|
โ
**Production-ready trust** - Permissive Apache 2.0 license |
|
|
โ
**Advanced reasoning** - Handles multi-layered attack chain analysis |
|
|
|
|
|
**The Result:** The most capable security-aware code model in the open-source ecosystem. |
|
|
|
|
|
**Why IBM Granite 20B?** This model is the enterprise choice: |
|
|
- ๐ข **IBM enterprise heritage** - 40+ years of enterprise software leadership |
|
|
- ๐ **Largest in collection** - 20B parameters = maximum reasoning capability |
|
|
- ๐ **Enterprise compliance ready** - Designed for regulated industries |
|
|
- โ๏ธ **Apache 2.0 licensed** - Full commercial freedom |
|
|
- ๐ฏ **Security-first training** - Built for mission-critical applications |
|
|
- ๐ **Broad language support** - 116+ programming languages |
|
|
|
|
|
Perfect for Fortune 500 companies, financial services, healthcare, government, and any organization where security analysis quality is paramount. |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Security Training Coverage |
|
|
|
|
|
### Real-World Vulnerability Distribution |
|
|
|
|
|
Trained on 1,209 security examples with real CVE grounding: |
|
|
|
|
|
| OWASP Category | Examples | Real Incidents | |
|
|
|----------------|----------|----------------| |
|
|
| **Broken Access Control** | 224 | Equifax, Facebook, Uber | |
|
|
| **Authentication Failures** | 199 | SolarWinds, Okta, LastPass | |
|
|
| **Injection Attacks** | 125 | Capital One, Yahoo, LinkedIn | |
|
|
| **Cryptographic Failures** | 115 | LastPass, Adobe, Dropbox | |
|
|
| **Security Misconfiguration** | 98 | Tesla, MongoDB, Elasticsearch | |
|
|
| **Vulnerable Components** | 87 | Log4Shell, Heartbleed, Struts | |
|
|
| **Identification/Auth Failures** | 84 | Twitter, GitHub, Reddit | |
|
|
| **Software/Data Integrity** | 78 | SolarWinds, Codecov, npm | |
|
|
| **Logging Failures** | 71 | Various incident responses | |
|
|
| **SSRF** | 69 | Capital One, Shopify | |
|
|
| **Insecure Design** | 59 | Architectural flaws | |
|
|
|
|
|
### Enterprise-Grade Multi-Language Support |
|
|
|
|
|
Fine-tuned on security examples across: |
|
|
- **Python** (Django, Flask, FastAPI) - 280 examples |
|
|
- **JavaScript/TypeScript** (Express, NestJS, React) - 245 examples |
|
|
- **Java** (Spring Boot, Jakarta EE) - 178 examples |
|
|
- **Go** (Gin, Echo, standard library) - 145 examples |
|
|
- **PHP** (Laravel, Symfony) - 112 examples |
|
|
- **C#** (ASP.NET Core, .NET 6+) - 89 examples |
|
|
- **Ruby** (Rails, Sinatra) - 67 examples |
|
|
- **Rust** (Actix, Rocket, Axum) - 45 examples |
|
|
- **C/C++** (Memory safety patterns) - 28 examples |
|
|
- **Plus 107+ additional languages from Granite's base training** |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ฏ Deployment Scenarios |
|
|
|
|
|
### Scenario 1: Enterprise Security Audit Platform |
|
|
|
|
|
**Professional security assessments for Fortune 500 clients.** |
|
|
|
|
|
**Hardware:** Datacenter GPU (A100 80GB or 2x A100 40GB) |
|
|
**Throughput:** 10-15 comprehensive audits/day |
|
|
**Use Case:** Professional security consulting |
|
|
|
|
|
**Value Proposition:** |
|
|
- Identify vulnerabilities human auditors miss |
|
|
- Consistent, comprehensive OWASP coverage |
|
|
- Scales expert security knowledge |
|
|
- Reduces audit time by 60-70% |
|
|
|
|
|
**ROI:** A single prevented breach pays for years of infrastructure. Typical large enterprise security audit costs $150K-500K. This model can handle preliminary analysis, allowing human experts to focus on novel vulnerabilities and strategic recommendations. |
|
|
|
|
|
--- |
|
|
|
|
|
### Scenario 2: Financial Services Security Platform |
|
|
|
|
|
**Regulatory compliance and security for banking applications.** |
|
|
|
|
|
**Hardware:** Private cloud A100 cluster |
|
|
**Compliance:** SOC 2, PCI-DSS, GDPR, CCPA |
|
|
**Use Case:** Pre-deployment security validation |
|
|
|
|
|
**Regulatory Benefits:** |
|
|
- Automated OWASP Top 10 verification |
|
|
- Audit trail generation |
|
|
- Compliance report automation |
|
|
- Reduces regulatory risk |
|
|
|
|
|
**ROI:** Regulatory fines cost millions. **Capital One:** $80M fine. **Equifax:** $425M settlement. Preventing one major breach justifies entire deployment. |
|
|
|
|
|
--- |
|
|
|
|
|
### Scenario 3: Healthcare Application Security |
|
|
|
|
|
**HIPAA-compliant code review for medical systems.** |
|
|
|
|
|
**Hardware:** Secure private deployment |
|
|
**Compliance:** HIPAA, HITECH, FDA software validation |
|
|
**Use Case:** Medical device and EHR security |
|
|
|
|
|
**Critical Healthcare Requirements:** |
|
|
- Patient data protection (HIPAA) |
|
|
- Audit logging and compliance |
|
|
- Cryptographic requirements |
|
|
- Access control verification |
|
|
|
|
|
**Impact:** Healthcare breaches average **$10.93 million per incident** (IBM 2024). Single prevented breach pays for multi-year deployment. |
|
|
|
|
|
--- |
|
|
|
|
|
### Scenario 4: Government & Defense Applications |
|
|
|
|
|
**Security analysis for critical infrastructure.** |
|
|
|
|
|
**Hardware:** Air-gapped secure environment |
|
|
**Clearance:** Can be deployed in classified environments |
|
|
**Use Case:** Critical infrastructure security |
|
|
|
|
|
**Government Benefits:** |
|
|
- No external dependencies (fully local) |
|
|
- Apache 2.0 license (government-friendly) |
|
|
- IBM enterprise support available |
|
|
- Meets government security standards |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Training Details |
|
|
|
|
|
| Parameter | Value | Why This Matters | |
|
|
|-----------|-------|------------------| |
|
|
| **Base Model** | ibm-granite/granite-20b-code-instruct-8k | IBM's enterprise-grade foundation | |
|
|
| **Fine-tuning Method** | LoRA (Low-Rank Adaptation) | Efficient training, preserves base capabilities | |
|
|
| **Training Dataset** | [SecureCode v2.0](https://huggingface.co/datasets/scthornton/securecode-v2) | 100% incident-grounded, expert-validated | |
|
|
| **Dataset Size** | 841 training examples | Focused on quality over quantity | |
|
|
| **Training Epochs** | 3 | Optimal convergence without overfitting | |
|
|
| **LoRA Rank (r)** | 16 | Balanced parameter efficiency | |
|
|
| **LoRA Alpha** | 32 | Learning rate scaling factor | |
|
|
| **Learning Rate** | 2e-4 | Standard for LoRA fine-tuning | |
|
|
| **Quantization** | 4-bit (bitsandbytes) | Enables efficient training | |
|
|
| **Trainable Parameters** | ~105M (0.525% of 20B total) | Minimal parameters, maximum impact | |
|
|
| **Total Parameters** | 20B | Maximum reasoning capability | |
|
|
| **Context Window** | 8K tokens | Enterprise file analysis | |
|
|
| **GPU Used** | NVIDIA A100 40GB | Enterprise training infrastructure | |
|
|
| **Training Time** | ~12-14 hours (estimated) | Deep security learning | |
|
|
|
|
|
### Training Methodology |
|
|
|
|
|
**LoRA (Low-Rank Adaptation)** was chosen for enterprise reliability: |
|
|
1. **Efficiency:** Trains only 0.525% of model parameters (105M vs 20B) |
|
|
2. **Quality:** Preserves IBM Granite's enterprise capabilities |
|
|
3. **Deployability:** Can be deployed alongside base model for versioning |
|
|
|
|
|
**4-bit Quantization** enables efficient training while maintaining enterprise-grade quality. |
|
|
|
|
|
**IBM Granite Foundation:** Built on IBM's 40+ years of enterprise software experience, optimized for: |
|
|
- Reliability and consistency |
|
|
- Enterprise deployment patterns |
|
|
- Regulatory compliance requirements |
|
|
- Long-term support and stability |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Usage |
|
|
|
|
|
### Quick Start |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
|
|
|
# Load IBM Granite base model |
|
|
base_model = "ibm-granite/granite-20b-code-instruct-8k" |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
base_model, |
|
|
device_map="auto", |
|
|
torch_dtype="auto", |
|
|
trust_remote_code=True |
|
|
) |
|
|
tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True) |
|
|
|
|
|
# Load SecureCode LoRA adapter |
|
|
model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
|
|
|
|
|
# Enterprise security analysis |
|
|
prompt = """### User: |
|
|
Conduct a comprehensive security audit of this enterprise authentication system. Analyze for: |
|
|
1. OWASP Top 10 vulnerabilities |
|
|
2. Attack chain opportunities |
|
|
3. Compliance gaps (SOC 2, PCI-DSS) |
|
|
4. Architectural weaknesses |
|
|
|
|
|
```python |
|
|
# Enterprise SSO Implementation |
|
|
class EnterpriseAuthService: |
|
|
def __init__(self): |
|
|
self.secret = os.getenv('JWT_SECRET') |
|
|
self.db = DatabasePool() |
|
|
|
|
|
async def authenticate(self, credentials): |
|
|
user = await self.db.query( |
|
|
f"SELECT * FROM users WHERE email='{credentials.email}' AND password='{credentials.password}'" |
|
|
) |
|
|
if user: |
|
|
token = jwt.encode({'user_id': user.id}, self.secret) |
|
|
return {'token': token, 'success': True} |
|
|
return {'success': False} |
|
|
|
|
|
async def verify_token(self, token): |
|
|
try: |
|
|
payload = jwt.decode(token, self.secret, algorithms=['HS256']) |
|
|
return payload |
|
|
except: |
|
|
return None |
|
|
``` |
|
|
|
|
|
### Assistant: |
|
|
""" |
|
|
|
|
|
inputs = tokenizer(prompt, return_tensors="pt").to(model.device) |
|
|
outputs = model.generate( |
|
|
**inputs, |
|
|
max_new_tokens=4096, |
|
|
temperature=0.2, # Lower temperature for precise enterprise analysis |
|
|
top_p=0.95, |
|
|
do_sample=True |
|
|
) |
|
|
|
|
|
response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
### Enterprise Deployment (4-bit Quantization) |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig |
|
|
from peft import PeftModel |
|
|
|
|
|
# 4-bit quantization - runs on 40GB GPU |
|
|
bnb_config = BitsAndBytesConfig( |
|
|
load_in_4bit=True, |
|
|
bnb_4bit_use_double_quant=True, |
|
|
bnb_4bit_quant_type="nf4", |
|
|
bnb_4bit_compute_dtype="bfloat16" |
|
|
) |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"ibm-granite/granite-20b-code-instruct-8k", |
|
|
quantization_config=bnb_config, |
|
|
device_map="auto", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
|
|
tokenizer = AutoTokenizer.from_pretrained("ibm-granite/granite-20b-code-instruct-8k", trust_remote_code=True) |
|
|
|
|
|
# Enterprise-ready: Runs on A100 40GB, A100 80GB, or 2x RTX 4090 |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
### Multi-GPU Deployment (Maximum Performance) |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
import torch |
|
|
|
|
|
# Load across multiple GPUs for maximum throughput |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"ibm-granite/granite-20b-code-instruct-8k", |
|
|
device_map="balanced", # Distribute across available GPUs |
|
|
torch_dtype=torch.bfloat16, |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
model = PeftModel.from_pretrained(model, "scthornton/granite-20b-code-securecode") |
|
|
tokenizer = AutoTokenizer.from_pretrained("ibm-granite/granite-20b-code-instruct-8k", trust_remote_code=True) |
|
|
|
|
|
# Optimal for: 2x A100, 4x RTX 4090, or enterprise GPU clusters |
|
|
# Throughput: 2-3x faster than single GPU |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Performance & Benchmarks |
|
|
|
|
|
### Hardware Requirements |
|
|
|
|
|
| Deployment | RAM | GPU VRAM | Tokens/Second | Latency (4K response) | Cost/Month | |
|
|
|-----------|-----|----------|---------------|----------------------|------------| |
|
|
| **4-bit Quantized** | 40GB | 32GB | ~35 tok/s | ~115 seconds | $0 (on-prem) or $800-1200 (cloud) | |
|
|
| **8-bit Quantized** | 64GB | 48GB | ~45 tok/s | ~90 seconds | $0 (on-prem) or $1200-1800 (cloud) | |
|
|
| **Full Precision (bf16)** | 96GB | 80GB | ~60 tok/s | ~67 seconds | $0 (on-prem) or $2000-3000 (cloud) | |
|
|
| **Multi-GPU (2x A100)** | 128GB | 160GB | ~120 tok/s | ~33 seconds | Enterprise only | |
|
|
|
|
|
### Real-World Performance |
|
|
|
|
|
**Tested on A100 40GB** (enterprise GPU): |
|
|
- **Tokens/second:** ~35 tok/s (4-bit), ~55 tok/s (full precision) |
|
|
- **Cold start:** ~8 seconds |
|
|
- **Memory usage:** 28GB (4-bit), 42GB (full precision) |
|
|
- **Throughput:** 200-300 comprehensive analyses per day |
|
|
|
|
|
**Tested on 2x A100 80GB** (multi-GPU): |
|
|
- **Tokens/second:** ~110-120 tok/s |
|
|
- **Cold start:** ~6 seconds |
|
|
- **Throughput:** 500+ analyses per day |
|
|
|
|
|
### Security Analysis Quality |
|
|
|
|
|
**The differentiator:** Granite 20B provides the deepest, most nuanced security analysis: |
|
|
- Identifies **15-25% more vulnerabilities** than 7B models in complex code |
|
|
- Detects **multi-step attack chains** that smaller models miss |
|
|
- Provides **enterprise-grade operational guidance** with compliance mapping |
|
|
- **Reduces false positives** through sophisticated reasoning |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ฐ Cost Analysis |
|
|
|
|
|
### Total Cost of Ownership (TCO) - 1 Year |
|
|
|
|
|
**Option 1: On-Premise (Dedicated Server)** |
|
|
- Hardware: 2x A100 40GB - $20,000 (one-time capital expense) |
|
|
- Server infrastructure: $5,000 |
|
|
- Electricity: ~$2,400/year |
|
|
- **Total Year 1:** $27,400 |
|
|
- **Total Year 2+:** $2,400/year |
|
|
|
|
|
**Option 2: Cloud GPU (AWS/GCP/Azure)** |
|
|
- Instance: A100 40GB (p4d.xlarge) |
|
|
- Cost: ~$3.50/hour |
|
|
- Usage: 160 hours/month (enterprise team) |
|
|
- **Total Year 1:** $6,720/year |
|
|
|
|
|
**Option 3: Enterprise GPT-4 (for comparison)** |
|
|
- Cost: $30/1M input tokens, $60/1M output tokens |
|
|
- Usage: 500M input + 500M output tokens/year |
|
|
- **Total Year 1:** $45,000/year |
|
|
|
|
|
**Option 4: Professional Security Audits (for comparison)** |
|
|
- Average enterprise security audit: $150,000-500,000 |
|
|
- Frequency: Quarterly (4x/year) |
|
|
- **Total Year 1:** $600,000-2,000,000 |
|
|
|
|
|
**ROI Winner:** On-premise deployment pays for itself with **1-2 prevented security audits** or **preventing a single breach** (average cost: $4.45M). |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ฏ Use Cases & Examples |
|
|
|
|
|
### 1. Enterprise Security Architecture Review |
|
|
|
|
|
Analyze complex microservices platforms: |
|
|
|
|
|
```python |
|
|
prompt = """### User: |
|
|
Conduct a comprehensive security architecture review of this fintech payment platform. Analyze: |
|
|
1. Service-to-service authentication security |
|
|
2. Data flow security boundaries |
|
|
3. Compliance with PCI-DSS requirements |
|
|
4. Attack surface analysis |
|
|
5. Defense-in-depth gaps |
|
|
|
|
|
[Include microservices code across auth-service, payment-service, notification-service] |
|
|
|
|
|
### Assistant: |
|
|
""" |
|
|
``` |
|
|
|
|
|
**Model Response:** Provides 20-30 page comprehensive analysis with specific vulnerability findings, attack chain scenarios, compliance gaps, and remediation priorities. |
|
|
|
|
|
--- |
|
|
|
|
|
### 2. Regulatory Compliance Validation |
|
|
|
|
|
Validate code against regulatory requirements: |
|
|
|
|
|
```python |
|
|
prompt = """### User: |
|
|
Analyze this healthcare EHR system for HIPAA compliance. Verify: |
|
|
1. Patient data encryption (at rest and in transit) |
|
|
2. Access control and audit logging |
|
|
3. Data retention policies |
|
|
4. Breach notification capabilities |
|
|
5. Business Associate Agreement requirements |
|
|
|
|
|
[Include EHR codebase] |
|
|
|
|
|
### Assistant: |
|
|
""" |
|
|
``` |
|
|
|
|
|
**Model Response:** Detailed compliance mapping, gap analysis, and remediation roadmap. |
|
|
|
|
|
--- |
|
|
|
|
|
### 3. Supply Chain Security Analysis |
|
|
|
|
|
Analyze third-party dependencies and integrations: |
|
|
|
|
|
```python |
|
|
prompt = """### User: |
|
|
Perform a supply chain security analysis of this application: |
|
|
1. Third-party library vulnerabilities |
|
|
2. Dependency confusion risks |
|
|
3. Code injection via dependencies |
|
|
4. Malicious package detection |
|
|
5. License compliance issues |
|
|
|
|
|
[Include package.json, requirements.txt, go.mod] |
|
|
|
|
|
### Assistant: |
|
|
""" |
|
|
``` |
|
|
|
|
|
**Model Response:** Comprehensive supply chain risk assessment with mitigation strategies. |
|
|
|
|
|
--- |
|
|
|
|
|
### 4. Advanced Penetration Testing Guidance |
|
|
|
|
|
Develop sophisticated attack scenarios: |
|
|
|
|
|
```python |
|
|
prompt = """### User: |
|
|
Design a comprehensive penetration testing strategy for this enterprise web application. Include: |
|
|
1. Attack surface enumeration |
|
|
2. Vulnerability prioritization |
|
|
3. Multi-stage attack chains |
|
|
4. Privilege escalation paths |
|
|
5. Data exfiltration scenarios |
|
|
6. Post-exploitation persistence |
|
|
|
|
|
### Assistant: |
|
|
""" |
|
|
``` |
|
|
|
|
|
**Model Response:** Professional pentesting methodology with specific attack vectors and validation procedures. |
|
|
|
|
|
--- |
|
|
|
|
|
## โ ๏ธ Limitations & Transparency |
|
|
|
|
|
### What This Model Does Well |
|
|
โ
Maximum code understanding and security reasoning |
|
|
โ
Complex attack chain analysis and enterprise architecture review |
|
|
โ
Detailed operational guidance and compliance mapping |
|
|
โ
Sophisticated multi-layered vulnerability detection |
|
|
โ
Enterprise-scale codebase analysis |
|
|
โ
IBM enterprise trust and reliability |
|
|
|
|
|
### What This Model Doesn't Do |
|
|
โ **Not a security scanner** - Use tools like Semgrep, CodeQL, Snyk, or Veracode |
|
|
โ **Not a penetration testing tool** - Cannot perform active exploitation or network scanning |
|
|
โ **Not legal/compliance advice** - Consult security and legal professionals |
|
|
โ **Not a replacement for security experts** - Critical systems need professional security review and audits |
|
|
โ **Not real-time threat intelligence** - Training data frozen at Dec 2024 |
|
|
|
|
|
### Known Issues & Constraints |
|
|
- **Inference latency:** Larger model means slower responses (35-60 tok/s vs 100+ tok/s for smaller models) |
|
|
- **Hardware requirements:** Requires enterprise GPU infrastructure (40GB+ VRAM) |
|
|
- **Detailed analysis:** May generate very comprehensive responses (3000-4000 tokens) |
|
|
- **Cost consideration:** Higher deployment cost than smaller models |
|
|
- **Context window:** 8K tokens (vs 128K for Qwen models) |
|
|
|
|
|
### Appropriate Use |
|
|
โ
Enterprise security audits and professional assessments |
|
|
โ
Regulatory compliance validation |
|
|
โ
Critical infrastructure security review |
|
|
โ
Financial services and healthcare applications |
|
|
โ
Government and defense security analysis |
|
|
|
|
|
### Inappropriate Use |
|
|
โ Sole validation for production deployments (use comprehensive testing) |
|
|
โ Replacement for professional security audits |
|
|
โ Active exploitation or penetration testing without authorization |
|
|
โ Consumer applications (too large, use smaller models) |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ฌ Dataset Information |
|
|
|
|
|
This model was trained on **[SecureCode v2.0](https://huggingface.co/datasets/scthornton/securecode-v2)**, a production-grade security dataset with: |
|
|
|
|
|
- **1,209 total examples** (841 train / 175 validation / 193 test) |
|
|
- **100% incident grounding** - every example tied to real CVEs or security breaches |
|
|
- **11 vulnerability categories** - complete OWASP Top 10:2025 coverage |
|
|
- **11 programming languages** - from Python to Rust |
|
|
- **4-turn conversational structure** - mirrors real developer-AI workflows |
|
|
- **100% expert validation** - reviewed by independent security professionals |
|
|
|
|
|
See the [full dataset card](https://huggingface.co/datasets/scthornton/securecode-v2) and [research paper](https://perfecxion.ai/articles/securecode-v2-dataset-paper.html) for complete details. |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ข About perfecXion.ai |
|
|
|
|
|
[perfecXion.ai](https://perfecxion.ai) is dedicated to advancing AI security through research, datasets, and production-grade security tooling. |
|
|
|
|
|
**Connect:** |
|
|
- Website: [perfecxion.ai](https://perfecxion.ai) |
|
|
- Research: [perfecxion.ai/research](https://perfecxion.ai/research) |
|
|
- Knowledge Hub: [perfecxion.ai/knowledge](https://perfecxion.ai/knowledge) |
|
|
- GitHub: [@scthornton](https://github.com/scthornton) |
|
|
- HuggingFace: [@scthornton](https://huggingface.co/scthornton) |
|
|
- Email: scott@perfecxion.ai |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ License |
|
|
|
|
|
**Model License:** Apache 2.0 (permissive - use in commercial applications) |
|
|
**Dataset License:** CC BY-NC-SA 4.0 (non-commercial with attribution) |
|
|
|
|
|
### What You CAN Do |
|
|
โ
Use this model commercially in production applications |
|
|
โ
Fine-tune further for your specific use case |
|
|
โ
Deploy in enterprise environments |
|
|
โ
Integrate into commercial products |
|
|
โ
Distribute and modify the model weights |
|
|
โ
Charge for services built on this model |
|
|
โ
Use in government and regulated industries |
|
|
|
|
|
### What You CANNOT Do with the Dataset |
|
|
โ Sell or redistribute the raw SecureCode v2.0 dataset commercially |
|
|
โ Use the dataset to train commercial models without releasing under the same license |
|
|
โ Remove attribution or claim ownership of the dataset |
|
|
|
|
|
For commercial dataset licensing or custom training, contact: scott@perfecxion.ai |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Citation |
|
|
|
|
|
If you use this model in your research or applications, please cite: |
|
|
|
|
|
```bibtex |
|
|
@misc{thornton2025securecode-granite20b, |
|
|
title={IBM Granite 20B Code - SecureCode Edition}, |
|
|
author={Thornton, Scott}, |
|
|
year={2025}, |
|
|
publisher={perfecXion.ai}, |
|
|
url={https://huggingface.co/scthornton/granite-20b-code-securecode}, |
|
|
note={Fine-tuned on SecureCode v2.0: https://huggingface.co/datasets/scthornton/securecode-v2} |
|
|
} |
|
|
|
|
|
@misc{thornton2025securecode-dataset, |
|
|
title={SecureCode v2.0: A Production-Grade Dataset for Training Security-Aware Code Generation Models}, |
|
|
author={Thornton, Scott}, |
|
|
year={2025}, |
|
|
month={January}, |
|
|
publisher={perfecXion.ai}, |
|
|
url={https://perfecxion.ai/articles/securecode-v2-dataset-paper.html}, |
|
|
note={Dataset: https://huggingface.co/datasets/scthornton/securecode-v2} |
|
|
} |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ Acknowledgments |
|
|
|
|
|
- **IBM Research** for the exceptional Granite code models and enterprise commitment |
|
|
- **OWASP Foundation** for maintaining the Top 10 vulnerability taxonomy |
|
|
- **MITRE Corporation** for the CVE database and vulnerability research |
|
|
- **Security research community** for responsible disclosure practices |
|
|
- **Hugging Face** for model hosting and inference infrastructure |
|
|
- **Enterprise security teams** who validated this model in production environments |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ค Contributing |
|
|
|
|
|
Found a security issue or have suggestions for improvement? |
|
|
|
|
|
- ๐ **Report issues:** [GitHub Issues](https://github.com/scthornton/securecode-models/issues) |
|
|
- ๐ฌ **Discuss improvements:** [HuggingFace Discussions](https://huggingface.co/scthornton/granite-20b-code-securecode/discussions) |
|
|
- ๐ง **Contact:** scott@perfecxion.ai |
|
|
|
|
|
### Community Contributions Welcome |
|
|
|
|
|
Especially interested in: |
|
|
- **Enterprise deployment case studies** |
|
|
- **Benchmark evaluations** on industry security datasets |
|
|
- **Compliance validation** (PCI-DSS, HIPAA, SOC 2) |
|
|
- **Performance optimization** for specific enterprise hardware |
|
|
- **Integration examples** with enterprise security platforms |
|
|
|
|
|
--- |
|
|
|
|
|
## ๐ SecureCode Model Collection |
|
|
|
|
|
Explore other SecureCode fine-tuned models optimized for different use cases: |
|
|
|
|
|
### Entry-Level Models (3-7B) |
|
|
- **[llama-3.2-3b-securecode](https://huggingface.co/scthornton/llama-3.2-3b-securecode)** |
|
|
- **Best for:** Consumer hardware, IDE integration, education |
|
|
- **Hardware:** 8GB RAM minimum |
|
|
- **Unique strength:** Most accessible |
|
|
|
|
|
- **[deepseek-coder-6.7b-securecode](https://huggingface.co/scthornton/deepseek-coder-6.7b-securecode)** |
|
|
- **Best for:** Security-optimized baseline |
|
|
- **Hardware:** 16GB RAM |
|
|
- **Unique strength:** Security-first architecture |
|
|
|
|
|
- **[qwen2.5-coder-7b-securecode](https://huggingface.co/scthornton/qwen2.5-coder-7b-securecode)** |
|
|
- **Best for:** Best code understanding in 7B class |
|
|
- **Hardware:** 16GB RAM |
|
|
- **Unique strength:** 128K context, best-in-class |
|
|
|
|
|
- **[codegemma-7b-securecode](https://huggingface.co/scthornton/codegemma-7b-securecode)** |
|
|
- **Best for:** Google ecosystem, instruction following |
|
|
- **Hardware:** 16GB RAM |
|
|
- **Unique strength:** Google brand, strong completion |
|
|
|
|
|
### Mid-Range Models (13-15B) |
|
|
- **[codellama-13b-securecode](https://huggingface.co/scthornton/codellama-13b-securecode)** |
|
|
- **Best for:** Enterprise trust, Meta brand |
|
|
- **Hardware:** 24GB RAM |
|
|
- **Unique strength:** Proven track record |
|
|
|
|
|
- **[qwen2.5-coder-14b-securecode](https://huggingface.co/scthornton/qwen2.5-coder-14b-securecode)** |
|
|
- **Best for:** Advanced code analysis |
|
|
- **Hardware:** 32GB RAM |
|
|
- **Unique strength:** 128K context window |
|
|
|
|
|
- **[starcoder2-15b-securecode](https://huggingface.co/scthornton/starcoder2-15b-securecode)** |
|
|
- **Best for:** Multi-language projects (600+ languages) |
|
|
- **Hardware:** 32GB RAM |
|
|
- **Unique strength:** Broadest language support |
|
|
|
|
|
### Enterprise-Scale Models (20B+) |
|
|
- **[granite-20b-code-securecode](https://huggingface.co/scthornton/granite-20b-code-securecode)** โญ (YOU ARE HERE) |
|
|
- **Best for:** Enterprise-scale, IBM trust, maximum capability |
|
|
- **Hardware:** 48GB RAM |
|
|
- **Unique strength:** Largest model, deepest analysis |
|
|
|
|
|
**View Complete Collection:** [SecureCode Models](https://huggingface.co/collections/scthornton/securecode) |
|
|
|
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
**Built with โค๏ธ for secure enterprise software** |
|
|
|
|
|
[perfecXion.ai](https://perfecxion.ai) | [Research](https://perfecxion.ai/research) | [Knowledge Hub](https://perfecxion.ai/knowledge) | [Contact](mailto:scott@perfecxion.ai) |
|
|
|
|
|
--- |
|
|
|
|
|
*Maximum security intelligence. Enterprise trust. IBM heritage.* |
|
|
|
|
|
</div> |
|
|
|