--- license: apache-2.0 base_model: cognitivecomputations/dolphin-2.6-mistral-7b tags: - cybersecurity - security-research - dolphin - mistral - unsloth --- # Dolphin Cybersecurity Research Model Fine-tuned Dolphin-Mistral-7B model for cybersecurity research and education. ## Model Details - **Base Model**: Dolphin 2.6 Mistral 7B - **Training Data**: General Knowledge dataset (37,635 examples) - **Training Method**: LoRA fine-tuning with Unsloth - **Use Case**: Cybersecurity education, penetration testing methodology, security research ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("YOUR_USERNAME/dolphin-cybersec") tokenizer = AutoTokenizer.from_pretrained("YOUR_USERNAME/dolphin-cybersec") prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. ### Instruction: What is SQL injection? ### Input: ### Response: """ inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=512) print(tokenizer.decode(outputs[0])) ``` ## Training Details - **LoRA Rank**: 16 - **Learning Rate**: 2e-4 - **Training Steps**: 1000 - **Batch Size**: 2 (gradient accumulation: 4) ## Intended Use Educational and research purposes only. For learning about cybersecurity concepts and methodologies.