File size: 1,546 Bytes
0a565fc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
---
license: apache-2.0
tags:
  - cybersecurity
  - document-classification
  - gguf
  - ollama
  - qwen
  - lora
base_model: Qwen/Qwen3.5-27B
---

# TorchSight Beam q8_0

Cybersecurity document classifier. LoRA fine-tune of Qwen 3.5 27B, quantized to q8_0. 28GB GGUF.

Higher quality weights (92.7% accuracy). For 48GB+ GPU or 64GB Mac.

## Benchmark Results (1000 samples)

| Model | Category Acc | Subcategory Acc |
|---|---|---|
| **Beam q4_K_M** | **95.1%** | 48.5% |
| Beam f16 | 93.0% | 51.3% |
| Beam q8_0 | 92.7% | 51.3% |
| Claude Opus 4 | 79.9% | 22.5% |
| Gemini 2.5 Pro | 75.4% | 21.0% |
| Qwen 3.5 27B (no fine-tune) | 43.3% | 4.3% |

## Usage with Ollama

```bash
ollama pull torchsight/beam:q8_0
```

Or with the GGUF file:

```
# Modelfile
FROM ./beam-1.0-q8_0.gguf

TEMPLATE "{{ .Prompt }}"
```

## Output Format

```json
[
  {
    "category": "credentials",
    "subcategory": "credentials.api_key",
    "severity": "critical",
    "explanation": "AWS access key found: AKIA****VIW..."
  }
]
```

Categories: `pii`, `credentials`, `financial`, `medical`, `confidential`, `malicious`, `safe`

## Training

- Base: Qwen 3.5 27B (dense)
- Method: LoRA (r=128, alpha=256)
- Data: 74K balanced samples from 18+ sources
- Epochs: 5
- GPU: H100 80GB PCIe

## Links

- [Benchmark Dataset](https://huggingface.co/datasets/torchsight/cybersecurity-classification-benchmark)
- [Training Data](https://huggingface.co/datasets/torchsight/beam-training-data)
- [GitHub](https://github.com/IvanDobrovolsky/torchsight)

## License

Apache 2.0