candyyuq commited on
Commit
9befa0e
·
verified ·
1 Parent(s): 2d9c721

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -1,4 +1,20 @@
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: text-classification
4
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  pipeline_tag: text-classification
4
+ ---
5
+
6
+ # LLM Harmful Checker
7
+
8
+ A robust model fine-tuned on microsoft/mdeberta-v3-base for detecting harmful inputs to Large Language Models.
9
+
10
+ ## Overview
11
+
12
+ LLM Harmful Checker is an AI model specifically designed to detect potentially harmful content in user inputs. Built upon microsoft/mdeberta-v3-base through fine-tuning, this model effectively identifies various types of harmful inputs during AI system interactions, including adversarial prompts and malicious instructions.
13
+
14
+ The model can be deployed in multiple scenarios, such as:
15
+ - AI system security protection
16
+ - Content moderation
17
+ - Customer service chatbots
18
+ - Other scenarios requiring secure AI interactions
19
+
20
+ By implementing this model, organizations can significantly enhance their AI systems' security and ensure user interactions remain compliant and safe.