lekhnathrijal commited on
Commit
5be36ec
·
verified ·
1 Parent(s): b2ef66a

ai-research-lab/bert-question-classifier

Browse files
Files changed (1) hide show
  1. README.md +35 -12
README.md CHANGED
@@ -21,11 +21,11 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 2.7007
25
- - Accuracy: 0.9562
26
- - Recall: 0.7745
27
- - Precision: 0.7606
28
- - F1: 0.7675
29
 
30
  ## Model description
31
 
@@ -44,19 +44,42 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 2e-05
48
- - train_batch_size: 8
49
- - eval_batch_size: 8
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
- - num_epochs: 1
 
54
 
55
  ### Training results
56
 
57
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | Recall | Precision | F1 |
58
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
59
- | No log | 1.0 | 390 | 2.7007 | 0.9562 | 0.7745 | 0.7606 | 0.7675 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
60
 
61
 
62
  ### Framework versions
 
21
 
22
  This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 1.8460
25
+ - Accuracy: 0.9711
26
+ - Recall: 0.8571
27
+ - Precision: 0.8371
28
+ - F1: 0.8470
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 1e-05
48
+ - train_batch_size: 4
49
+ - eval_batch_size: 4
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
+ - lr_scheduler_warmup_ratio: 0.1
54
+ - num_epochs: 3
55
 
56
  ### Training results
57
 
58
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Recall | Precision | F1 |
59
+ |:-------------:|:------:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
60
+ | No log | 0.1284 | 100 | 4.9916 | 0.9143 | 0.5142 | 0.5430 | 0.5282 |
61
+ | No log | 0.2567 | 200 | 4.3263 | 0.9264 | 0.5870 | 0.6092 | 0.5979 |
62
+ | No log | 0.3851 | 300 | 3.8919 | 0.9348 | 0.6636 | 0.6463 | 0.6548 |
63
+ | No log | 0.5135 | 400 | 3.5265 | 0.9391 | 0.6599 | 0.6783 | 0.6690 |
64
+ | 4.2904 | 0.6418 | 500 | 3.2937 | 0.9452 | 0.7049 | 0.7066 | 0.7057 |
65
+ | 4.2904 | 0.7702 | 600 | 3.0129 | 0.9496 | 0.7275 | 0.7305 | 0.7290 |
66
+ | 4.2904 | 0.8986 | 700 | 2.8410 | 0.9521 | 0.7482 | 0.7404 | 0.7443 |
67
+ | 4.2904 | 1.0270 | 800 | 2.6565 | 0.9552 | 0.7757 | 0.7520 | 0.7637 |
68
+ | 4.2904 | 1.1553 | 900 | 2.5233 | 0.9574 | 0.7842 | 0.7647 | 0.7743 |
69
+ | 2.7537 | 1.2837 | 1000 | 2.3877 | 0.9598 | 0.7976 | 0.7771 | 0.7872 |
70
+ | 2.7537 | 1.4121 | 1100 | 2.2836 | 0.9622 | 0.8146 | 0.7875 | 0.8008 |
71
+ | 2.7537 | 1.5404 | 1200 | 2.1776 | 0.9635 | 0.8130 | 0.7990 | 0.8059 |
72
+ | 2.7537 | 1.6688 | 1300 | 2.1273 | 0.9653 | 0.8223 | 0.8085 | 0.8153 |
73
+ | 2.7537 | 1.7972 | 1400 | 2.0858 | 0.9651 | 0.8251 | 0.8052 | 0.8150 |
74
+ | 2.218 | 1.9255 | 1500 | 2.0143 | 0.9670 | 0.8312 | 0.8176 | 0.8243 |
75
+ | 2.218 | 2.0539 | 1600 | 1.9800 | 0.9683 | 0.8413 | 0.8226 | 0.8319 |
76
+ | 2.218 | 2.1823 | 1700 | 1.9409 | 0.9691 | 0.8470 | 0.8259 | 0.8363 |
77
+ | 2.218 | 2.3107 | 1800 | 1.9122 | 0.9693 | 0.8445 | 0.8294 | 0.8369 |
78
+ | 2.218 | 2.4390 | 1900 | 1.8876 | 0.9699 | 0.8502 | 0.8310 | 0.8405 |
79
+ | 1.8644 | 2.5674 | 2000 | 1.8786 | 0.9695 | 0.8470 | 0.8298 | 0.8383 |
80
+ | 1.8644 | 2.6958 | 2100 | 1.8620 | 0.9701 | 0.8510 | 0.8325 | 0.8416 |
81
+ | 1.8644 | 2.8241 | 2200 | 1.8494 | 0.9711 | 0.8567 | 0.8370 | 0.8467 |
82
+ | 1.8644 | 2.9525 | 2300 | 1.8460 | 0.9711 | 0.8571 | 0.8371 | 0.8470 |
83
 
84
 
85
  ### Framework versions