Update README.md
Browse files
README.md
CHANGED
|
@@ -2,17 +2,18 @@
|
|
| 2 |
This model is a fine-tuned version of the dbmdz/bert-base-turkish-uncased architecture, specifically designed for the binary classification task of identifying organizational accounts on Turkish Twitter. It leverages the pre-trained BERT model's understanding of Turkish language and context to effectively distinguish between organizational and non-organizational user accounts.
|
| 3 |
|
| 4 |
### Model Training and Optimization
|
| 5 |
-
|
| 6 |
|
| 7 |
-
|
| 8 |
|
| 9 |
-
|
| 10 |
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
|
|
|
| 16 |
Learning Rate: 1.23e-5
|
| 17 |
Batch Size: 32
|
| 18 |
Epochs: 2
|
|
|
|
| 2 |
This model is a fine-tuned version of the dbmdz/bert-base-turkish-uncased architecture, specifically designed for the binary classification task of identifying organizational accounts on Turkish Twitter. It leverages the pre-trained BERT model's understanding of Turkish language and context to effectively distinguish between organizational and non-organizational user accounts.
|
| 3 |
|
| 4 |
### Model Training and Optimization
|
| 5 |
+
Base Model: dbmdz/bert-base-turkish-uncased
|
| 6 |
|
| 7 |
+
Training Data: The model was trained and validated using a dataset of Twitter accounts (descriptions, names, screen names) with meticulously annotated labels indicating whether each account belongs to an organization or not.
|
| 8 |
|
| 9 |
+
### Fine-Tuning Process:
|
| 10 |
|
| 11 |
+
Data Preprocessing: Combined user descriptions, names, and screen names into a single text field for input.
|
| 12 |
+
Data Splitting: Split the dataset into 80% for training and 20% for validation.
|
| 13 |
+
Tokenization: Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
|
| 14 |
+
Hyperparameter Optimization: Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
|
| 15 |
+
|
| 16 |
+
Optimal Hyperparameters:
|
| 17 |
Learning Rate: 1.23e-5
|
| 18 |
Batch Size: 32
|
| 19 |
Epochs: 2
|