File size: 1,931 Bytes
88d9d19 b5faaf7 43d09f2 344b4e7 b5faaf7 344b4e7 88d9d19 b5faaf7 88d9d19 b5faaf7 88d9d19 344b4e7 88d9d19 b5faaf7 88d9d19 b5faaf7 88d9d19 b5faaf7 88d9d19 b5faaf7 287019a 88d9d19 b5faaf7 88d9d19 b5faaf7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 | ---
language:
- en
license: apache-2.0
library_name: transformers
pipeline_tag: text-generation
tags:
- Domain-Certification
- Jailbreaking
- Adversarial-Attack
- Guardrail
datasets:
- qiaojin/PubMedQA
---
# Shh, don't say that! Domain Certification in LLMs
[](https://cemde.github.io/Domain-Certification-Website/)
[](https://arxiv.org/abs/2502.19320)
[](https://iclr.cc/virtual/2025/poster/30364)
[](https://github.com/cemde/Domain-Certification)
**Collection:** https://huggingface.co/collections/cemde/domain-certification-67ba4fb663f8d1348c3c2263
**Certify you Large Language Model (LLM)!**
With the code in this repository you can reproduce the workflows we use in our ICLR 2025 paper to achieve Domain Certification using our VALID algorithm.
We provide the guide models for our Medical Question Answering experiments.
| Model | Description |
| - | - |
| [cemde/Domain-Certification-MedQA-Guide-Base](https://huggingface.co/cemde/Domain-Certification-MedQA-Guide-Base) | This is the base model trained on the ground-truth responses. |
| [cemde/Domain-Certification-MedQA-Guide-Finetuned](https://huggingface.co/cemde/Domain-Certification-MedQA-Guide-Finetuned) | This is the model trained on responses from Llama-3-8B. |
## Citation
```
@inproceedings{
emde2025shh,
title={Shh, don't say that! Domain Certification in {LLM}s},
author={Cornelius Emde and Alasdair Paren and Preetham Arvind and Maxime Guillaume Kayser and Tom Rainforth and Bernard Ghanem and Thomas Lukasiewicz and Philip Torr and Adel Bibi},
booktitle={The Thirteenth International Conference on Learning Representations},
year={2025},
url={https://arxiv.org/abs/2502.19320}
}
```
|