JeffreyWong commited on
Commit
9b0f2be
·
verified ·
1 Parent(s): 434976a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -19
README.md CHANGED
@@ -30,32 +30,18 @@ It achieves the following results on the evaluation set:
30
  - eval_steps_per_second: 25.882
31
  - step: 0
32
 
33
- ## Model description
34
-
35
- More information needed
36
-
37
- ## Intended uses & limitations
38
-
39
- More information needed
40
-
41
- ## Training and evaluation data
42
-
43
- More information needed
44
-
45
  ## Training procedure
46
 
47
  ### Training hyperparameters
48
 
49
  The following hyperparameters were used during training:
50
- - learning_rate: 5e-05
51
- - train_batch_size: 8
52
  - eval_batch_size: 4
53
  - seed: 42
54
- - distributed_type: multi-GPU
55
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
- - lr_scheduler_type: linear
57
- - num_epochs: 3.0
58
- - mixed_precision_training: Native AMP
59
 
60
  ### Framework versions
61
 
 
30
  - eval_steps_per_second: 25.882
31
  - step: 0
32
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  ## Training procedure
34
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 1e-5, 2e-5, 3e-5
39
+ - train_batch_size: 16
40
  - eval_batch_size: 4
41
  - seed: 42
42
+ - num_epochs: 10
43
+
44
+ The best model was selected based on the highest accuracy, which is the key evaluation metric for this task.
 
 
45
 
46
  ### Framework versions
47