dolphin-cybersec / README.md
frEEtom3's picture
Upload folder using huggingface_hub
b8cbb20 verified
metadata
license: apache-2.0
base_model: cognitivecomputations/dolphin-2.6-mistral-7b
tags:
  - cybersecurity
  - security-research
  - dolphin
  - mistral
  - unsloth

Dolphin Cybersecurity Research Model

Fine-tuned Dolphin-Mistral-7B model for cybersecurity research and education.

Model Details

  • Base Model: Dolphin 2.6 Mistral 7B
  • Training Data: General Knowledge dataset (37,635 examples)
  • Training Method: LoRA fine-tuning with Unsloth
  • Use Case: Cybersecurity education, penetration testing methodology, security research

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("YOUR_USERNAME/dolphin-cybersec")
tokenizer = AutoTokenizer.from_pretrained("YOUR_USERNAME/dolphin-cybersec")

prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
What is SQL injection?

### Input:


### Response:
"""

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0]))

Training Details

  • LoRA Rank: 16
  • Learning Rate: 2e-4
  • Training Steps: 1000
  • Batch Size: 2 (gradient accumulation: 4)

Intended Use

Educational and research purposes only. For learning about cybersecurity concepts and methodologies.