atsizelti commited on
Commit
23f1df0
·
verified ·
1 Parent(s): e018d63

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -8
README.md CHANGED
@@ -2,17 +2,18 @@
2
  This model is a fine-tuned version of the dbmdz/bert-base-turkish-uncased architecture, specifically designed for the binary classification task of identifying organizational accounts on Turkish Twitter. It leverages the pre-trained BERT model's understanding of Turkish language and context to effectively distinguish between organizational and non-organizational user accounts.
3
 
4
  ### Model Training and Optimization
5
- ## Base Model: dbmdz/bert-base-turkish-uncased
6
 
7
- ## Training Data: The model was trained and validated using a dataset of Twitter accounts (descriptions, names, screen names) with meticulously annotated labels indicating whether each account belongs to an organization or not.
8
 
9
- ## Fine-Tuning Process:
10
 
11
- ## Data Preprocessing: Combined user descriptions, names, and screen names into a single text field for input.
12
- ## Data Splitting: Split the dataset into 80% for training and 20% for validation.
13
- ## Tokenization: Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
14
- ## Hyperparameter Optimization: Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
15
- ## Optimal Hyperparameters:
 
16
  Learning Rate: 1.23e-5
17
  Batch Size: 32
18
  Epochs: 2
 
2
  This model is a fine-tuned version of the dbmdz/bert-base-turkish-uncased architecture, specifically designed for the binary classification task of identifying organizational accounts on Turkish Twitter. It leverages the pre-trained BERT model's understanding of Turkish language and context to effectively distinguish between organizational and non-organizational user accounts.
3
 
4
  ### Model Training and Optimization
5
+ Base Model: dbmdz/bert-base-turkish-uncased
6
 
7
+ Training Data: The model was trained and validated using a dataset of Twitter accounts (descriptions, names, screen names) with meticulously annotated labels indicating whether each account belongs to an organization or not.
8
 
9
+ ### Fine-Tuning Process:
10
 
11
+ Data Preprocessing: Combined user descriptions, names, and screen names into a single text field for input.
12
+ Data Splitting: Split the dataset into 80% for training and 20% for validation.
13
+ Tokenization: Utilized the AutoTokenizer from Hugging Face to prepare text inputs for the BERT model.
14
+ Hyperparameter Optimization: Employed Optuna to find the best combination of learning rate, batch size, and training epochs, resulting in optimal performance and minimizing validation loss.
15
+
16
+ Optimal Hyperparameters:
17
  Learning Rate: 1.23e-5
18
  Batch Size: 32
19
  Epochs: 2