mengru commited on
Commit
ef3ca8a
·
verified ·
1 Parent(s): e784159

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -4
README.md CHANGED
@@ -1,10 +1,33 @@
1
  ---
2
  license: apache-2.0
3
-
4
-
5
-
6
  ---
7
 
8
  <div style="text-align:center;">
9
  <strong>Safety classifier for Detoxifying Large Language Models via Knowledge Editing</strong>
10
- </div>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
 
 
 
3
  ---
4
 
5
  <div style="text-align:center;">
6
  <strong>Safety classifier for Detoxifying Large Language Models via Knowledge Editing</strong>
7
+ </div>
8
+
9
+ # 💻 Usage
10
+
11
+ ```shell
12
+ from transformers import RobertaForSequenceClassification, RobertaTokenizer
13
+ safety_classifier_dir = 'zjunlp/DINM-Safety-Classifier'
14
+ safety_classifier_model = RobertaForSequenceClassification.from_pretrained(safety_classifier_dir)
15
+ safety_classifier_tokenizer = RobertaTokenizer.from_pretrained(safety_classifier_dir)
16
+ ```
17
+ You can also download DINM-Safety-Classifier manually, and set the safety_classifier_dir to your own path.
18
+
19
+
20
+ # 📖 Citation
21
+
22
+ If you use our work, please cite our paper:
23
+
24
+ ```bibtex
25
+ @misc{wang2024SafeEdit,
26
+ title={Detoxifying Large Language Models via Knowledge Editing},
27
+ author={Mengru Wang, Ningyu Zhang, Ziwen Xu, Zekun Xi, Shumin Deng, Yunzhi Yao, Qishen Zhang, Linyi Yang, Jindong Wang, Huajun Chen},
28
+ year={2024},
29
+ eprint={xx},
30
+ archivePrefix={arXiv},
31
+ primaryClass={cs.CL}
32
+ }
33
+ ```