CASSANDRA — BCE configuration on AnnoCTR

Fine-tuned CTI-BERT models for extracting MITRE ATT&CK techniques from cyber threat intelligence (CTI) reports. This repository contains the BCE configuration of the CASSANDRA recipe trained on AnnoCTR (118 ATT&CK techniques), comprising 3 ensemble members trained with seeds {42, 123, 456}.

This is the headline configuration on AnnoCTR in the paper. The asymmetric-loss (ASL) variant regresses on this benchmark due to its low label density (mean 15.5 samples per class); BCE with pos_weight handles the rare classes more robustly.

Anonymous artifact for double-blind peer review. Author information will be added after the review period.

Headline result

On the AnnoCTR test set (33 scored documents):

  • 3-seed ensemble per-document F1 (Ï„=0.5): 63.53%
  • Exceeds CySecBERT's 62.75% (Buchel et al. 2025) without CySecBERT's additional 4.3M cybersecurity pre-training texts.

The per-seed table below shows the live artifact's individual seed F1s and ensemble F1; small variance from the headline (≤0.3 F1) reflects inference-time floating-point ordering on different hardware. Full per-seed and ensemble metrics are in results.json.

Architecture

LabelAttentionClassifier: a 110M-parameter CTI-BERT encoder followed by a per-label attention head.

  • Encoder: ibm-research/CTI-BERT (110M params, 768 hidden)
  • Head: 118 learned 768-dim label queries that attend over the encoder's last_hidden_state, followed by a shared 1-output linear layer applied per-label
  • Loss: BCE with pos_weight=5.0
  • Regularization / training tricks: layer-wise learning rate decay (α=0.85), exponential moving average (β=0.999), multi-seed probability averaging at inference

The architecture is custom (not derived from transformers.PreTrainedModel), so loading requires the modeling.py file shipped with this repo.

Training data

  • AnnoCTR: 104 reports, 5,265 sentences, 118 canonical ATT&CK techniques (113 train-present, 5 unobserved at training but present in test). Mean of 15.5 deduplicated positive examples per train-present class. 78 of 113 train-present classes have fewer than 10 positive examples.
  • Splits: report-level train/test split from Buchel et al. (2025) "SoK: A Survey of Approaches for ATT&CK Classifier Construction" (70 train reports, 34 test reports — one test report excluded from per-document F1 due to empty in-vocabulary ground truth).
  • Validation: 80:20 sentence-level random split within the training reports for early stopping and threshold selection.

Intended use

Map free-text CTI sentences to ATT&CK techniques. The model takes a single sentence and outputs a probability for each of 118 techniques.

Aggregation to document level (paper convention): apply per-sentence inference, take the per-class max across sentences in a document, threshold that, report the union of predicted techniques per document.

Limitations:

  • Trained on English-language CTI; behavior on other languages is not characterized.
  • The 118-label vocabulary is the canonical AnnoCTR set; sentences describing techniques outside this set will produce all-zero predictions.
  • AnnoCTR's extreme sparsity (78 of 113 train-present techniques have fewer than 10 positives) means rare-technique predictions are noisier than common-technique predictions. Per-technique threshold tuning (provided as an option in inference_example.py) does not consistently help for these ultra-rare techniques — see paper §3.1 (per-technique thresholding excluded from the recommended configuration).

How to load and run

from modeling import load_ensemble, predict_ensemble
import os, glob

seed_dirs = sorted(glob.glob(os.path.join(os.path.dirname(__file__), "seeds", "seed-*")))
seeds = load_ensemble(seed_dirs, device="cuda")

sentences = [
    "The malware uses Windows Command Shell to execute encoded scripts.",
    "After initial access, persistence was established via Registry Run Keys.",
]
results = predict_ensemble(seeds, sentences, threshold=0.5)
for sentence, techniques in results:
    print(sentence, "->", techniques)

A complete CLI example is in inference_example.py:

pip install -r requirements.txt
python inference_example.py

Per-seed members

Seed Per-document F1 (Ï„=0.5) Selected weights
42 59.82% EMA
123 61.29% EMA
456 63.57% EMA
3-seed ensemble 63.53% —

Citation

@misc{cassandra2026,
  title  = {CASSANDRA: How Many Parameters Suffice to Automate TTP Extractions from CTI Reports---Pushing Towards the Lower Bound},
  author = {{Anonymous Authors}},
  year   = {2026},
  note   = {Anonymous submission under review}
}

Please also cite the AnnoCTR dataset and the CTI-BERT encoder.

License

Apache-2.0. These fine-tuned weights are derived from ibm-research/CTI-BERT.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for cassandra-anon/cassandra-bce-annoctr

Finetuned
(4)
this model