Update README.md
Browse files
README.md
CHANGED
|
@@ -8,24 +8,42 @@ Training Data: The model was trained and validated using a dataset of Twitter a
|
|
| 8 |
|
| 9 |
### Fine-Tuning Process:
|
| 10 |
|
| 11 |
-
Data Preprocessing:
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 15 |
|
| 16 |
Optimal Hyperparameters:
|
|
|
|
| 17 |
Learning Rate: 1.23e-5
|
|
|
|
| 18 |
Batch Size: 32
|
|
|
|
| 19 |
Epochs: 2
|
| 20 |
|
| 21 |
## Evaluation Results
|
| 22 |
The fine-tuned model demonstrates excellent performance on the validation set, achieving the following metrics:
|
| 23 |
Precision: 0.945
|
|
|
|
| 24 |
Recall: 0.95
|
|
|
|
| 25 |
F1-Score (Macro): 0.948
|
|
|
|
| 26 |
Accuracy: 0.95
|
| 27 |
|
| 28 |
|
| 29 |
Confusion Matrix:
|
| 30 |
[[369 22]
|
|
|
|
| 31 |
[ 19 375]]
|
|
|
|
| 8 |
|
| 9 |
### Fine-Tuning Process:
|
| 10 |
|
| 11 |
+
Data Preprocessing:
|
| 12 |
+
|
| 13 |
+
Combined user descriptions, names, and screen names into a single text field for input.
|
| 14 |
+
|
| 15 |
+
Data Splitting:
|
| 16 |
+
|
| 17 |
+
Split the dataset into 80% for training and 20% for validation.
|
| 18 |
+
|
| 19 |
+
Tokenization:
|
| 20 |
+
|
| 21 |
+
Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
|
| 22 |
+
|
| 23 |
+
Hyperparameter Optimization:
|
| 24 |
+
|
| 25 |
+
Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
|
| 26 |
|
| 27 |
Optimal Hyperparameters:
|
| 28 |
+
|
| 29 |
Learning Rate: 1.23e-5
|
| 30 |
+
|
| 31 |
Batch Size: 32
|
| 32 |
+
|
| 33 |
Epochs: 2
|
| 34 |
|
| 35 |
## Evaluation Results
|
| 36 |
The fine-tuned model demonstrates excellent performance on the validation set, achieving the following metrics:
|
| 37 |
Precision: 0.945
|
| 38 |
+
|
| 39 |
Recall: 0.95
|
| 40 |
+
|
| 41 |
F1-Score (Macro): 0.948
|
| 42 |
+
|
| 43 |
Accuracy: 0.95
|
| 44 |
|
| 45 |
|
| 46 |
Confusion Matrix:
|
| 47 |
[[369 22]
|
| 48 |
+
|
| 49 |
[ 19 375]]
|