README / README.md
davidberenstein1957's picture
Update README.md
a2d153d verified
|
raw
history blame
5.79 kB
metadata
title: README
emoji: 💻
colorFrom: blue
colorTo: red
sdk: static
pinned: false

Enguard AI

One guardrail for all, all guardrails for one!

Why should you use these models?

  • Optimised for precision to reduce false positives.
  • Extremely fast inference using static embeddings powered by Model2Vec.

Which guards are available?

Below is an overview of all guardrails, showing the best results for the smallest (-2m), best-performing, and multi-lingual models across all configurations.

Dataset Classifies Collection Smallest (2m) Best Performing Multi-lingual (128m)
harmfulness-mix prompt-harmfulness Collection 0.9192 0.9350 -
intel response-politeness Collection 0.8795 0.8908 0.8908
jailbreak-in-the-wild prompt-jailbreak Collection 0.8515 0.8905 0.8905
jailbreak-sok prompt-jailbreak Collection 0.9762 0.9810 0.9810
jigsaw prompt-toxicity Collection 0.8967 0.9067 0.8986
nvidia-aegis response-safety Collection 0.7612 0.7760 0.7530
nvidia-aegis prompt-safety Collection 0.7957 0.8131 0.7929
polyguard response-safety Collection 0.8635 0.8808 0.8753
polyguard response-refusal Collection 0.8952 0.9039 0.9015
polyguard prompt-safety Collection - 0.9331 0.9255
toxic-chat response-jailbreak Collection 0.9872 0.9914 -
toxic-chat prompt-toxicity Collection 0.9515 0.9555 -