Suparious commited on
Commit
f8c88bb
·
verified ·
1 Parent(s): 0388fb1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -2
README.md CHANGED
@@ -1,4 +1,9 @@
1
  ---
 
 
 
 
 
2
  library_name: transformers
3
  tags:
4
  - 4-bit
@@ -6,10 +11,28 @@ tags:
6
  - text-generation
7
  - autotrain_compatible
8
  - endpoints_compatible
 
 
 
 
 
9
  pipeline_tag: text-generation
10
  inference: false
11
  quantized_by: Suparious
12
  ---
13
- #
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- **UPLOAD IN PROGRESS**
 
1
  ---
2
+ language:
3
+ - en
4
+ license: other
5
+ license_name: llama3
6
+ license_link: LICENSE
7
  library_name: transformers
8
  tags:
9
  - 4-bit
 
11
  - text-generation
12
  - autotrain_compatible
13
  - endpoints_compatible
14
+ - facebook
15
+ - meta
16
+ - pytorch
17
+ - llama
18
+ - llama-3
19
  pipeline_tag: text-generation
20
  inference: false
21
  quantized_by: Suparious
22
  ---
23
+ # meta-llama/Meta-Llama-Guard-2-8B AWQ
24
+
25
+ - Model creator: [meta-llama](https://huggingface.co/meta-llama)
26
+ - Original model: [Meta-Llama-Guard-2-8B](https://huggingface.co/meta-llama/Meta-Llama-Guard-2-8B)
27
+
28
+ ## Model Summary
29
+
30
+ Meta Llama Guard 2 is an 8B parameter Llama 3-based [1] LLM safeguard model. Similar to [Llama Guard](https://ai.meta.com/research/publications/llama-guard-llm-based-input-output-safeguard-for-human-ai-conversations/), it can be used for classifying content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.
31
+ Below is a response classification example input and output for Llama Guard 2.
32
+
33
+ <p align="center">
34
+ <img src="https://github.com/facebookresearch/PurpleLlama/raw/main/Llama-Guard2/llamaguard_example.png" width="800"/>
35
+ </p>
36
+
37
+ In order to produce classifier scores, we look at the probability for the first token, and use that as the “unsafe” class probability. We can then apply score thresholding to make binary decisions.
38