Vrandan commited on
Commit
6ea213b
Β·
verified Β·
1 Parent(s): 9391db9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -9
README.md CHANGED
@@ -11,7 +11,7 @@ language:
11
  base_model:
12
  - distilbert/distilbert-base-uncased
13
  ---
14
- # πŸ›‘οΈ Comment Moderation Model
15
 
16
  [![HuggingFace](https://img.shields.io/badge/πŸ€—%20Hugging%20Face-Spaces-blue)](https://huggingface.co/Vrandan/Comment-Moderation)
17
  [![Python 3.12+](https://img.shields.io/badge/python-3.12+-blue.svg)](https://www.python.org/downloads/release/python-312/)
@@ -20,7 +20,7 @@ base_model:
20
 
21
  A powerful, multi-label content moderation system built on **DistilBERT** architecture, designed to detect and classify potentially harmful content in user-generated comments with high accuracy. This model stands out as currently the best in terms of performance based on the provided dataset for text moderation. Additionally, it has the smallest footprint, making it ideal for deployment on edge devices. Currently, it is the only model trained to achieve such high performance while maintaining a minimal size relative to the training data on Hugging Face.
22
 
23
- ## 🎯 Key Features
24
 
25
  - Multi-label classification
26
  - Real-time content analysis
@@ -33,7 +33,7 @@ A powerful, multi-label content moderation system built on **DistilBERT** archit
33
  - Resource-efficient while maintaining high accuracy
34
  - Can run on consumer-grade hardware
35
 
36
- ## πŸ“Š Content Categories
37
 
38
  The model identifies the following types of potentially harmful content:
39
 
@@ -49,7 +49,7 @@ The model identifies the following types of potentially harmful content:
49
  | Violence/Graphic | `V2` | Violent content that depicts death, violence, or serious physical injury in extreme graphic detail. |
50
  | Safe Content | `OK` | Appropriate content that doesn't violate any guidelines. |
51
 
52
- ## πŸ“ˆ Performance Metrics
53
 
54
  ```
55
  Accuracy: 95.4%
@@ -60,7 +60,7 @@ Micro F1 Score: 0.802
60
 
61
  [View detailed performance metrics](#model-performance)
62
 
63
- ## πŸ–₯️ Training Details
64
 
65
  The model was trained on an **NVIDIA RTX 3080** GPU in a home setup, demonstrating that effective content moderation models can be developed with consumer-grade hardware. This makes the model development process more accessible to individual developers and smaller organizations.
66
 
@@ -73,7 +73,7 @@ Key Training Specifications:
73
 
74
  Despite its relatively compact size **(67M parameters)**, this model achieves impressive performance metrics, making it suitable for deployment across various devices and environments. The model's efficiency-to-performance ratio demonstrates that effective content moderation is possible without requiring extensive computational resources.
75
 
76
- ## πŸš€ Quick Start
77
 
78
  ### Python Implementation (Local)
79
 
@@ -213,7 +213,7 @@ query({"inputs": "Your text here"}).then((response) => {
213
  });
214
  ```
215
 
216
- ## πŸ“Š Detailed Model Performance <a name="model-performance"></a>
217
 
218
  The model has been extensively evaluated using standard classification metrics:
219
 
@@ -242,7 +242,7 @@ The model has been extensively evaluated using standard classification metrics:
242
  - Potential for false positives
243
  - Cultural context variations
244
 
245
- ## πŸ“š Dataset Information
246
 
247
  This model was trained on the dataset released by OpenAI, as described in their paper ["A Holistic Approach to Undesired Content Detection"](https://arxiv.org/abs/2208.03274).
248
 
@@ -261,7 +261,7 @@ If you use this model or dataset in your research, please cite:
261
  }
262
  ```
263
 
264
- ## πŸ“§ Contact
265
 
266
  For support or queries, please message me on Slack.
267
 
 
11
  base_model:
12
  - distilbert/distilbert-base-uncased
13
  ---
14
+ # Comment Moderation Model
15
 
16
  [![HuggingFace](https://img.shields.io/badge/πŸ€—%20Hugging%20Face-Spaces-blue)](https://huggingface.co/Vrandan/Comment-Moderation)
17
  [![Python 3.12+](https://img.shields.io/badge/python-3.12+-blue.svg)](https://www.python.org/downloads/release/python-312/)
 
20
 
21
  A powerful, multi-label content moderation system built on **DistilBERT** architecture, designed to detect and classify potentially harmful content in user-generated comments with high accuracy. This model stands out as currently the best in terms of performance based on the provided dataset for text moderation. Additionally, it has the smallest footprint, making it ideal for deployment on edge devices. Currently, it is the only model trained to achieve such high performance while maintaining a minimal size relative to the training data on Hugging Face.
22
 
23
+ ## Key Features
24
 
25
  - Multi-label classification
26
  - Real-time content analysis
 
33
  - Resource-efficient while maintaining high accuracy
34
  - Can run on consumer-grade hardware
35
 
36
+ ## Content Categories
37
 
38
  The model identifies the following types of potentially harmful content:
39
 
 
49
  | Violence/Graphic | `V2` | Violent content that depicts death, violence, or serious physical injury in extreme graphic detail. |
50
  | Safe Content | `OK` | Appropriate content that doesn't violate any guidelines. |
51
 
52
+ ## Performance Metrics
53
 
54
  ```
55
  Accuracy: 95.4%
 
60
 
61
  [View detailed performance metrics](#model-performance)
62
 
63
+ ## Training Details
64
 
65
  The model was trained on an **NVIDIA RTX 3080** GPU in a home setup, demonstrating that effective content moderation models can be developed with consumer-grade hardware. This makes the model development process more accessible to individual developers and smaller organizations.
66
 
 
73
 
74
  Despite its relatively compact size **(67M parameters)**, this model achieves impressive performance metrics, making it suitable for deployment across various devices and environments. The model's efficiency-to-performance ratio demonstrates that effective content moderation is possible without requiring extensive computational resources.
75
 
76
+ ## Quick Start
77
 
78
  ### Python Implementation (Local)
79
 
 
213
  });
214
  ```
215
 
216
+ ## Detailed Model Performance <a name="model-performance"></a>
217
 
218
  The model has been extensively evaluated using standard classification metrics:
219
 
 
242
  - Potential for false positives
243
  - Cultural context variations
244
 
245
+ ## Dataset Information
246
 
247
  This model was trained on the dataset released by OpenAI, as described in their paper ["A Holistic Approach to Undesired Content Detection"](https://arxiv.org/abs/2208.03274).
248
 
 
261
  }
262
  ```
263
 
264
+ ## Contact
265
 
266
  For support or queries, please message me on Slack.
267