atsizelti commited on
Commit
d0e7899
·
verified ·
1 Parent(s): 23f1df0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -4
README.md CHANGED
@@ -8,24 +8,42 @@ Training Data: The model was trained and validated using a dataset of Twitter a
8
 
9
  ### Fine-Tuning Process:
10
 
11
- Data Preprocessing: Combined user descriptions, names, and screen names into a single text field for input.
12
- Data Splitting: Split the dataset into 80% for training and 20% for validation.
13
- Tokenization: Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
14
- Hyperparameter Optimization: Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  Optimal Hyperparameters:
 
17
  Learning Rate: 1.23e-5
 
18
  Batch Size: 32
 
19
  Epochs: 2
20
 
21
  ## Evaluation Results
22
  The fine-tuned model demonstrates excellent performance on the validation set, achieving the following metrics:
23
  Precision: 0.945
 
24
  Recall: 0.95
 
25
  F1-Score (Macro): 0.948
 
26
  Accuracy: 0.95
27
 
28
 
29
  Confusion Matrix:
30
  [[369 22]
 
31
  [ 19 375]]
 
8
 
9
  ### Fine-Tuning Process:
10
 
11
+ Data Preprocessing:
12
+
13
+ Combined user descriptions, names, and screen names into a single text field for input.
14
+
15
+ Data Splitting:
16
+
17
+ Split the dataset into 80% for training and 20% for validation.
18
+
19
+ Tokenization:
20
+
21
+ Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
22
+
23
+ Hyperparameter Optimization:
24
+
25
+ Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
26
 
27
  Optimal Hyperparameters:
28
+
29
  Learning Rate: 1.23e-5
30
+
31
  Batch Size: 32
32
+
33
  Epochs: 2
34
 
35
  ## Evaluation Results
36
  The fine-tuned model demonstrates excellent performance on the validation set, achieving the following metrics:
37
  Precision: 0.945
38
+
39
  Recall: 0.95
40
+
41
  F1-Score (Macro): 0.948
42
+
43
  Accuracy: 0.95
44
 
45
 
46
  Confusion Matrix:
47
  [[369 22]
48
+
49
  [ 19 375]]