phishbot commited on
Commit
741ff1b
·
1 Parent(s): 24af85a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: unknown
3
+ ---
4
+
5
+ # Overview
6
+
7
+ <!-- This model is obtained by finetuning Pre-Trained RoBERTa on dataset containing several sets of malicious prompts.
8
+ Using this model, we can classify malicious prompts that can lead towards creation of phishing websites and phishing emails.
9
+ This model is obtained by finetuning a Pre-Trained RoBERTa using a dataset encompassing multiple sets of malicious prompts, as detailed in the corresponding arXiv paper.
10
+ Using this model, we can classify malicious prompts that can lead towards creation of phishing websites and phishing emails. -->
11
+
12
+ Our model, "ScamLLM" is designed to identify malicious prompts that can be used to generate phishing websites and emails using popular commercial LLMs like ChatGPT, Bard and Claude.
13
+ This model is obtained by finetuning a Pre-Trained RoBERTa using a dataset encompassing multiple sets of malicious prompts, as detailed in our corresponding arXiv paper
14
+
15
+ <!--- **Paper:** https://arxiv.org/abs/2310.19181 -->
16
+
17
+ Try out "ScamLLM" using the Inference API. Our model classifies prompts with "Label 1" to signify the identification of a phishing attempt, while "Label 0" denotes a prompt that is considered safe and non-malicious.
18
+
19
+ ## Dataset Details
20
+
21
+ The dataset utilized for training this model has been created using malicious prompts generated by GPT-4.
22
+ Due to ethical concerns, our dataset is currently available only upon request.
23
+
24
+ ## Training Details
25
+
26
+ The model was trained using RobertaForSequenceClassification.from_pretrained.
27
+ In this process, both the model and tokenizer pertinent to the RoBERTa-base were employed.
28
+ We trained this model for 10 epochs, setting a learning rate to 2e-5, and used AdamW Optimizer.
29
+
30
+ ## Inference
31
+
32
+ There are multiple ways to use this model. The simplest way to use is with pipeline "text-classification"
33
+
34
+ ```python
35
+ from transformers import pipeline
36
+ classifier = pipeline(task="text-classification", model="phishbot/ScamLLM", top_k=None)
37
+ prompt = ["Your Sample Sentence or Prompt...."]
38
+ model_outputs = classifier(prompt)
39
+ print(model_outputs[0])
40
+ ```
41
+
42
+ ### Results
43
+
44
+ Achieved an accuracy of 96% with an F1-score of 0.96, on test sets distribution, explained in the paper.
45
+
46
+ <!--## Citation
47
+
48
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section.
49
+ If you find Isitphish to be useful, please cite it with:
50
+
51
+ ```
52
+ @misc{roy2023chatbots,
53
+ title={From Chatbots to PhishBots? -- Preventing Phishing scams created using ChatGPT, Google Bard and Claude},
54
+ author={Sayak Saha Roy and Poojitha Thota and Krishna Vamsi Naragam and Shirin Nilizadeh},
55
+ year={2023},
56
+ eprint={2310.19181},
57
+ archivePrefix={arXiv},
58
+ primaryClass={cs.CR}
59
+ }
60
+ ```-->