entfane commited on
Commit
764ac0b
·
verified ·
1 Parent(s): 1fa2426

Upload folder using huggingface_hub

Browse files
chat_template.jinja ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {% if not messages[0]['role'] == 'system' %}### INSTRUCTION:
2
+ You are a precise Content Moderation Classifier. Your mission is to evaluate text for toxicity, bias, and hate speech. You must distinguish between overt abuse and "implicit hate"—text that may seem linguistically neutral but carries harmful, stereotypical, or dehumanizing intent toward protected groups.
3
+
4
+ ## NOT ALLOWED
5
+ You must flag and categorize the following as Violations:
6
+
7
+ Identity-Based Hate: Any content targeting individuals or groups based on: Women, LGBTQ+, Mental/Physical Disorders, Race (Black, Chinese, Asian, Native American, Latino, Mexican), Religion (Muslim, Jewish), or Middle Eastern origin.
8
+
9
+ Implicit Toxicity: Coded language, "polite" stereotypes, or synthetic generations that imply inferiority or promote exclusion without using explicit slurs.
10
+
11
+ Dehumanization: Language comparing protected groups to animals, insects, diseases, or inanimate objects.
12
+
13
+ Aggressive Toxicity: Direct insults, derogatory terms, harmful slang, and threatening language.
14
+
15
+ Bias Reinforcement: Statements that generalize negative traits or "harmful tropes" to an entire demographic.
16
+
17
+ ## ALLOWED
18
+ You must classify the following as Benign:
19
+
20
+ Neutral Reference: Factual or descriptive mentions of protected groups (e.g., "The history of Latino art").
21
+
22
+ Educational/Clinical Context: Objective discussion of disabilities, mental health, or religious practices without judgmental or biased framing.
23
+
24
+ Self-Identification: Individuals using identity terms to describe themselves or their own community experiences in a non-harmful manner.
25
+
26
+ Counter-Speech: Content that explicitly challenges or denounces hate speech, provided it does not repeat the toxicity in a gratuitous way.
27
+ {% endif %}{% for message in messages %}{% if message['role'] == 'system' %}### INSTRUCTION:
28
+ {{ message['content'] }}
29
+ {% elif message['role'] == 'user' %}### INPUT:
30
+ {{ message['content'] }}
31
+ {% elif message['role'] == 'assistant' %}### OUTPUT:
32
+ {{ message['content'] }}
33
+ {% endif %}{% endfor %}
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "add_cross_attention": false,
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "dtype": "float32",
10
+ "embd_pdrop": 0.1,
11
+ "eos_token_id": 50256,
12
+ "initializer_range": 0.02,
13
+ "layer_norm_epsilon": 1e-05,
14
+ "model_type": "gpt2",
15
+ "n_ctx": 1024,
16
+ "n_embd": 768,
17
+ "n_head": 12,
18
+ "n_inner": null,
19
+ "n_layer": 12,
20
+ "n_positions": 1024,
21
+ "pad_token_id": 50256,
22
+ "reorder_and_upcast_attn": false,
23
+ "resid_pdrop": 0.1,
24
+ "scale_attn_by_inverse_layer_idx": false,
25
+ "scale_attn_weights": true,
26
+ "summary_activation": null,
27
+ "summary_first_dropout": 0.1,
28
+ "summary_proj_to_labels": true,
29
+ "summary_type": "cls_index",
30
+ "summary_use_proj": true,
31
+ "task_specific_params": {
32
+ "text-generation": {
33
+ "do_sample": true,
34
+ "max_length": 50
35
+ }
36
+ },
37
+ "tie_word_embeddings": true,
38
+ "transformers_version": "5.2.0",
39
+ "use_cache": true,
40
+ "vocab_size": 50257
41
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "5.2.0"
6
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8418ae9424cdab8a38dbdd3bc25acad8e484e4e96fc1f4076f2d1fe8cc97037a
3
+ size 497777468
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": "<|endoftext|>",
5
+ "eos_token": "<|endoftext|>",
6
+ "errors": "replace",
7
+ "is_local": false,
8
+ "model_max_length": 1024,
9
+ "pad_token": "<|endoftext|>",
10
+ "tokenizer_class": "GPT2Tokenizer",
11
+ "unk_token": "<|endoftext|>"
12
+ }