vishva0 commited on
Commit
272d4b6
·
verified ·
1 Parent(s): f661ebf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -9
README.md CHANGED
@@ -1,22 +1,83 @@
1
  ---
2
  base_model: unsloth/qwen2.5-0.5b-instruct-unsloth-bnb-4bit
 
 
 
3
  tags:
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - qwen2
8
  - trl
9
- license: apache-2.0
10
- language:
11
- - en
12
  ---
13
 
14
- # Uploaded model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- - **Developed by:** vishva0
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** unsloth/qwen2.5-0.5b-instruct-unsloth-bnb-4bit
19
 
20
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth)
 
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
1
  ---
2
  base_model: unsloth/qwen2.5-0.5b-instruct-unsloth-bnb-4bit
3
+ language:
4
+ - en
5
+ license: apache-2.0
6
  tags:
7
  - text-generation-inference
8
  - transformers
9
  - unsloth
10
  - qwen2
11
  - trl
12
+ - safetensors
13
+ - security
14
+ - red-teaming
15
  ---
16
 
17
+ # coliseum034/coliseum-attacker-wild
18
+
19
+ This model is a fine-tuned version of `unsloth/qwen2.5-0.5b-instruct-unsloth-bnb-4bit`. It was trained up to 2x faster utilizing [Unsloth](https://github.com/unslothai/unsloth) and Hugging Face's TRL library.
20
+
21
+ Developed by Vishva Patel (`vishva0`), this model is structurally geared toward advanced security operations, multi-agent system simulations, and red-teaming applications in the wild.
22
+
23
+ ## ⚙️ Model Details
24
+
25
+ * **Developed by:** Vishva Patel (`vishva0`)
26
+ * **License:** Apache 2.0
27
+ * **Base Model:** `unsloth/qwen2.5-0.5b-instruct-unsloth-bnb-4bit`
28
+ * **Architecture:** Qwen2 (0.5B parameters)
29
+ * **Language:** English
30
+ * **Quantization:** 4-bit (bitsandbytes)
31
+
32
+ ## 📊 Training & Evaluation Metrics
33
+
34
+ The model was trained over 3 epochs for a total of 921 global steps. The training procedure demonstrated consistent learning, achieving a final validation perplexity of ~5.168.
35
+
36
+ ### Per-Epoch Results
37
+
38
+ | Epoch | Training Loss | Validation Loss | Perplexity (PPL) |
39
+ | :---: | :---: | :---: | :---: |
40
+ | **1.0** | 1.6638 | 1.6605 | 5.262 |
41
+ | **2.0** | 1.5345 | 1.6314 | 5.111 |
42
+ | **3.0** | 1.4212 | 1.6425 | 5.168 |
43
+
44
+ ### Final Held-Out Metrics
45
+
46
+ * **Final Training Loss:** `1.4212`
47
+ * **Final Evaluation Loss:** `1.6425`
48
+ * **Final Perplexity:** `5.168`
49
+
50
+ ### Training Hyperparameters & Performance
51
+
52
+ * **Global Steps:** 921
53
+ * **Total Training Runtime:** ~36 minutes, 48 seconds (2207.98 seconds)
54
+ * **Training Samples per Second:** 6.658
55
+ * **Training Steps per Second:** 0.417
56
+ * **Total FLOPs:** 8.527 x 10^15
57
+
58
+ ## 💻 Framework Versions
59
+
60
+ * PEFT
61
+ * Transformers
62
+ * Unsloth
63
+ * TRL
64
+ * Safetensors
65
+ * PyTorch
66
+
67
+ ## 🚀 Usage
68
+
69
+ This model uses the standard `transformers` library pipeline or `text-generation-inference`.
70
+
71
+ ```python
72
+ from transformers import AutoModelForCausalLM, AutoTokenizer
73
+
74
+ model_id = "coliseum034/coliseum-attacker-wild"
75
 
76
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
77
+ model = AutoModelForCausalLM.from_pretrained(model_id)
 
78
 
79
+ prompt = "Analyze this sequence for potential exploitation vectors:"
80
+ inputs = tokenizer(prompt, return_tensors="pt")
81
 
82
+ outputs = model.generate(**inputs, max_new_tokens=100)
83
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))