Update README.md
Browse files
README.md
CHANGED
|
@@ -11,7 +11,7 @@ language:
|
|
| 11 |
base_model:
|
| 12 |
- distilbert/distilbert-base-uncased
|
| 13 |
---
|
| 14 |
-
#
|
| 15 |
|
| 16 |
[](https://huggingface.co/Vrandan/Comment-Moderation)
|
| 17 |
[](https://www.python.org/downloads/release/python-312/)
|
|
@@ -20,7 +20,7 @@ base_model:
|
|
| 20 |
|
| 21 |
A powerful, multi-label content moderation system built on **DistilBERT** architecture, designed to detect and classify potentially harmful content in user-generated comments with high accuracy. This model stands out as currently the best in terms of performance based on the provided dataset for text moderation. Additionally, it has the smallest footprint, making it ideal for deployment on edge devices. Currently, it is the only model trained to achieve such high performance while maintaining a minimal size relative to the training data on Hugging Face.
|
| 22 |
|
| 23 |
-
##
|
| 24 |
|
| 25 |
- Multi-label classification
|
| 26 |
- Real-time content analysis
|
|
@@ -33,7 +33,7 @@ A powerful, multi-label content moderation system built on **DistilBERT** archit
|
|
| 33 |
- Resource-efficient while maintaining high accuracy
|
| 34 |
- Can run on consumer-grade hardware
|
| 35 |
|
| 36 |
-
##
|
| 37 |
|
| 38 |
The model identifies the following types of potentially harmful content:
|
| 39 |
|
|
@@ -49,7 +49,7 @@ The model identifies the following types of potentially harmful content:
|
|
| 49 |
| Violence/Graphic | `V2` | Violent content that depicts death, violence, or serious physical injury in extreme graphic detail. |
|
| 50 |
| Safe Content | `OK` | Appropriate content that doesn't violate any guidelines. |
|
| 51 |
|
| 52 |
-
##
|
| 53 |
|
| 54 |
```
|
| 55 |
Accuracy: 95.4%
|
|
@@ -60,7 +60,7 @@ Micro F1 Score: 0.802
|
|
| 60 |
|
| 61 |
[View detailed performance metrics](#model-performance)
|
| 62 |
|
| 63 |
-
##
|
| 64 |
|
| 65 |
The model was trained on an **NVIDIA RTX 3080** GPU in a home setup, demonstrating that effective content moderation models can be developed with consumer-grade hardware. This makes the model development process more accessible to individual developers and smaller organizations.
|
| 66 |
|
|
@@ -73,7 +73,7 @@ Key Training Specifications:
|
|
| 73 |
|
| 74 |
Despite its relatively compact size **(67M parameters)**, this model achieves impressive performance metrics, making it suitable for deployment across various devices and environments. The model's efficiency-to-performance ratio demonstrates that effective content moderation is possible without requiring extensive computational resources.
|
| 75 |
|
| 76 |
-
##
|
| 77 |
|
| 78 |
### Python Implementation (Local)
|
| 79 |
|
|
@@ -213,7 +213,7 @@ query({"inputs": "Your text here"}).then((response) => {
|
|
| 213 |
});
|
| 214 |
```
|
| 215 |
|
| 216 |
-
##
|
| 217 |
|
| 218 |
The model has been extensively evaluated using standard classification metrics:
|
| 219 |
|
|
@@ -242,7 +242,7 @@ The model has been extensively evaluated using standard classification metrics:
|
|
| 242 |
- Potential for false positives
|
| 243 |
- Cultural context variations
|
| 244 |
|
| 245 |
-
##
|
| 246 |
|
| 247 |
This model was trained on the dataset released by OpenAI, as described in their paper ["A Holistic Approach to Undesired Content Detection"](https://arxiv.org/abs/2208.03274).
|
| 248 |
|
|
@@ -261,7 +261,7 @@ If you use this model or dataset in your research, please cite:
|
|
| 261 |
}
|
| 262 |
```
|
| 263 |
|
| 264 |
-
##
|
| 265 |
|
| 266 |
For support or queries, please message me on Slack.
|
| 267 |
|
|
|
|
| 11 |
base_model:
|
| 12 |
- distilbert/distilbert-base-uncased
|
| 13 |
---
|
| 14 |
+
# Comment Moderation Model
|
| 15 |
|
| 16 |
[](https://huggingface.co/Vrandan/Comment-Moderation)
|
| 17 |
[](https://www.python.org/downloads/release/python-312/)
|
|
|
|
| 20 |
|
| 21 |
A powerful, multi-label content moderation system built on **DistilBERT** architecture, designed to detect and classify potentially harmful content in user-generated comments with high accuracy. This model stands out as currently the best in terms of performance based on the provided dataset for text moderation. Additionally, it has the smallest footprint, making it ideal for deployment on edge devices. Currently, it is the only model trained to achieve such high performance while maintaining a minimal size relative to the training data on Hugging Face.
|
| 22 |
|
| 23 |
+
## Key Features
|
| 24 |
|
| 25 |
- Multi-label classification
|
| 26 |
- Real-time content analysis
|
|
|
|
| 33 |
- Resource-efficient while maintaining high accuracy
|
| 34 |
- Can run on consumer-grade hardware
|
| 35 |
|
| 36 |
+
## Content Categories
|
| 37 |
|
| 38 |
The model identifies the following types of potentially harmful content:
|
| 39 |
|
|
|
|
| 49 |
| Violence/Graphic | `V2` | Violent content that depicts death, violence, or serious physical injury in extreme graphic detail. |
|
| 50 |
| Safe Content | `OK` | Appropriate content that doesn't violate any guidelines. |
|
| 51 |
|
| 52 |
+
## Performance Metrics
|
| 53 |
|
| 54 |
```
|
| 55 |
Accuracy: 95.4%
|
|
|
|
| 60 |
|
| 61 |
[View detailed performance metrics](#model-performance)
|
| 62 |
|
| 63 |
+
## Training Details
|
| 64 |
|
| 65 |
The model was trained on an **NVIDIA RTX 3080** GPU in a home setup, demonstrating that effective content moderation models can be developed with consumer-grade hardware. This makes the model development process more accessible to individual developers and smaller organizations.
|
| 66 |
|
|
|
|
| 73 |
|
| 74 |
Despite its relatively compact size **(67M parameters)**, this model achieves impressive performance metrics, making it suitable for deployment across various devices and environments. The model's efficiency-to-performance ratio demonstrates that effective content moderation is possible without requiring extensive computational resources.
|
| 75 |
|
| 76 |
+
## Quick Start
|
| 77 |
|
| 78 |
### Python Implementation (Local)
|
| 79 |
|
|
|
|
| 213 |
});
|
| 214 |
```
|
| 215 |
|
| 216 |
+
## Detailed Model Performance <a name="model-performance"></a>
|
| 217 |
|
| 218 |
The model has been extensively evaluated using standard classification metrics:
|
| 219 |
|
|
|
|
| 242 |
- Potential for false positives
|
| 243 |
- Cultural context variations
|
| 244 |
|
| 245 |
+
## Dataset Information
|
| 246 |
|
| 247 |
This model was trained on the dataset released by OpenAI, as described in their paper ["A Holistic Approach to Undesired Content Detection"](https://arxiv.org/abs/2208.03274).
|
| 248 |
|
|
|
|
| 261 |
}
|
| 262 |
```
|
| 263 |
|
| 264 |
+
## Contact
|
| 265 |
|
| 266 |
For support or queries, please message me on Slack.
|
| 267 |
|