Safetensors
llama

Add pipeline_tag and library_name to metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +24 -18
README.md CHANGED
@@ -1,9 +1,11 @@
1
  ---
2
- license: llama3.1
3
- datasets:
4
- - yahma/alpaca-cleaned
5
  base_model:
6
  - meta-llama/Llama-3.1-8B-Instruct
 
 
 
 
 
7
  ---
8
 
9
  # DataFilter
@@ -11,9 +13,10 @@ base_model:
11
  [![arXiv](https://img.shields.io/badge/arXiv-2510.19207-b31b1b.svg)](https://arxiv.org/abs/2510.19207)
12
  [![HuggingFace](https://img.shields.io/badge/🤗-Model-yellow)](https://huggingface.co/JoyYizhu/DataFilter)
13
 
14
- A defense system designed to protect LLM agent systems against prompt injection attacks. DataFilter provides robust protection while maintaining system utility and performance.
15
 
16
- Codes: https://github.com/yizhu-joy/DataFilter
 
17
 
18
  ## Quick Start
19
 
@@ -22,29 +25,32 @@ Codes: https://github.com/yizhu-joy/DataFilter
22
  ```bash
23
  conda create -n py312vllm python=3.12
24
  conda activate py312vllm
25
- pip install vllm pandas 'accelerate>=0.26.0'
26
  git clone https://github.com/yizhu-joy/DataFilter.git
27
  cd DataFilter
28
  ```
29
 
30
- ### Run DataFilter Inference demo:
 
31
  ```bash
32
  python filter_inference.py
33
  ```
34
 
35
  ## Citation
36
 
37
- If you use DataFilter in your research, please cite our paper:
38
 
39
  ```bibtex
40
- @misc
41
- {wang2025datafilter,
42
- title={Defending Against Prompt Injection with DataFilter},
43
- author={Yizhu Wang and Sizhe Chen and Raghad Alkhudair and Basel Alomair and David Wagner},
44
- year={2025},
45
- eprint={2510.19207},
46
- archivePrefix={arXiv},
47
- primaryClass={cs.CR},
48
- url={https://arxiv.org/abs/2510.19207},
49
  }
50
- ```
 
 
 
 
1
  ---
 
 
 
2
  base_model:
3
  - meta-llama/Llama-3.1-8B-Instruct
4
+ datasets:
5
+ - yahma/alpaca-cleaned
6
+ library_name: transformers
7
+ license: llama3.1
8
+ pipeline_tag: text-generation
9
  ---
10
 
11
  # DataFilter
 
13
  [![arXiv](https://img.shields.io/badge/arXiv-2510.19207-b31b1b.svg)](https://arxiv.org/abs/2510.19207)
14
  [![HuggingFace](https://img.shields.io/badge/🤗-Model-yellow)](https://huggingface.co/JoyYizhu/DataFilter)
15
 
16
+ DataFilter is a test-time model-agnostic defense system designed to protect Large Language Model (LLM) agents against prompt injection attacks. As described in the paper [Defending Against Prompt Injection with DataFilter](https://huggingface.co/papers/2510.19207), it removes malicious instructions from data before it reaches the backend LLM, maintaining high utility while reducing attack success rates to near zero.
17
 
18
+ - **Paper:** [Defending Against Prompt Injection with DataFilter](https://huggingface.co/papers/2510.19207)
19
+ - **Repository:** [GitHub - yizhu-joy/DataFilter](https://github.com/yizhu-joy/DataFilter)
20
 
21
  ## Quick Start
22
 
 
25
  ```bash
26
  conda create -n py312vllm python=3.12
27
  conda activate py312vllm
28
+ pip install vllm pandas 'accelerate>=0.26.0' deepspeed datasets==2.20.0
29
  git clone https://github.com/yizhu-joy/DataFilter.git
30
  cd DataFilter
31
  ```
32
 
33
+ ### Run DataFilter Inference Demo
34
+ To test the DataFilter model, run the provided inference script:
35
  ```bash
36
  python filter_inference.py
37
  ```
38
 
39
  ## Citation
40
 
41
+ If you use DataFilter in your research, please cite the following paper:
42
 
43
  ```bibtex
44
+ @misc{wang2025datafilter,
45
+ title={Defending Against Prompt Injection with DataFilter},
46
+ author={Yizhu Wang and Sizhe Chen and Raghad Alkhudair and Basel Alomair and David Wagner},
47
+ year={2025},
48
+ eprint={2510.19207},
49
+ archivePrefix={arXiv},
50
+ primaryClass={cs.CR},
51
+ url={https://arxiv.org/abs/2510.19207},
 
52
  }
53
+ ```
54
+
55
+ ## License
56
+ This model is licensed under the Llama 3.1 Community License. Please refer to the LICENSE file for details.