Update README.md
Browse files
README.md
CHANGED
|
@@ -7,8 +7,37 @@ metrics:
|
|
| 7 |
base_model:
|
| 8 |
- distilbert/distilbert-base-uncased
|
| 9 |
pipeline_tag: text-classification
|
|
|
|
|
|
|
| 10 |
---
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
### Model Description
|
| 13 |
|
| 14 |
This fine-tuned DistilBERT model is specifically designed for document classification. It classifies customer feedback into six predefined categories: Shipping and Delivery, Customer Service, Price and Value, Quality and Performance, Use and Design, and Other. By leveraging the transformer-based architecture of DistilBERT, the model efficiently handles the syntactic patterns of text, providing accurate document classification based on content, style, and structure.
|
|
@@ -27,23 +56,9 @@ While the model achieves high accuracy across the six categories, it has limitat
|
|
| 27 |
Users (both direct and downstream) should be aware of the model's single-label prediction limitation. In cases where a document contains features of multiple categories, additional models or multi-label classification techniques should be considered.
|
| 28 |
|
| 29 |
### Training Data
|
| 30 |
-
A custom synthetic dataset was created for this task, focusing on the structural features of text. The dataset provides examples from six categories, helping the model learn from both the syntactic organization and the meaning of the text.
|
| 31 |
|
| 32 |
-
|
| 33 |
|
| 34 |
-
### label_mapping = {
|
| 35 |
-
"shipping_and_delivery": 0,
|
| 36 |
-
|
| 37 |
-
"customer_service": 1,
|
| 38 |
-
|
| 39 |
-
"price_and_value": 2,
|
| 40 |
-
|
| 41 |
-
"quality_and_performance": 3,
|
| 42 |
-
|
| 43 |
-
"use_and_design": 4,
|
| 44 |
-
|
| 45 |
-
"other": 5
|
| 46 |
-
}
|
| 47 |
### Training Hyperparameters
|
| 48 |
Model: distilbert/distilbert-base-uncased
|
| 49 |
|
|
|
|
| 7 |
base_model:
|
| 8 |
- distilbert/distilbert-base-uncased
|
| 9 |
pipeline_tag: text-classification
|
| 10 |
+
widget:
|
| 11 |
+
- text: "The product arrived on time and was exactly as described."
|
| 12 |
---
|
| 13 |
|
| 14 |
+
### Categories:
|
| 15 |
+
|
| 16 |
+
### label_mapping = {
|
| 17 |
+
"shipping_and_delivery": 0,
|
| 18 |
+
|
| 19 |
+
"customer_service": 1,
|
| 20 |
+
|
| 21 |
+
"price_and_value": 2,
|
| 22 |
+
|
| 23 |
+
"quality_and_performance": 3,
|
| 24 |
+
|
| 25 |
+
"use_and_design": 4,
|
| 26 |
+
|
| 27 |
+
"other": 5
|
| 28 |
+
}
|
| 29 |
+
|
| 30 |
+
## How to Use:
|
| 31 |
+
|
| 32 |
+
Here is an example of how to use this model for inference:
|
| 33 |
+
|
| 34 |
+
```python
|
| 35 |
+
from transformers import pipeline
|
| 36 |
+
|
| 37 |
+
classifier = pipeline("text-classification", model="dnzblgn/Customer-Reviews-Classification")
|
| 38 |
+
result = classifier("The product arrived on time and was exactly as described.")
|
| 39 |
+
print(result)
|
| 40 |
+
|
| 41 |
### Model Description
|
| 42 |
|
| 43 |
This fine-tuned DistilBERT model is specifically designed for document classification. It classifies customer feedback into six predefined categories: Shipping and Delivery, Customer Service, Price and Value, Quality and Performance, Use and Design, and Other. By leveraging the transformer-based architecture of DistilBERT, the model efficiently handles the syntactic patterns of text, providing accurate document classification based on content, style, and structure.
|
|
|
|
| 56 |
Users (both direct and downstream) should be aware of the model's single-label prediction limitation. In cases where a document contains features of multiple categories, additional models or multi-label classification techniques should be considered.
|
| 57 |
|
| 58 |
### Training Data
|
|
|
|
| 59 |
|
| 60 |
+
A custom synthetic dataset was created for this task, focusing on the structural features of text. The dataset provides examples from six categories, helping the model learn from both the syntactic organization and the meaning of the text.
|
| 61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 62 |
### Training Hyperparameters
|
| 63 |
Model: distilbert/distilbert-base-uncased
|
| 64 |
|