Update README.md
Browse files
README.md
CHANGED
|
@@ -61,8 +61,8 @@ wandb_watch:
|
|
| 61 |
wandb_run_id:
|
| 62 |
wandb_log_model:
|
| 63 |
|
| 64 |
-
gradient_accumulation_steps:
|
| 65 |
-
micro_batch_size:
|
| 66 |
num_epochs: 1
|
| 67 |
optimizer: paged_adamw_32bit
|
| 68 |
lr_scheduler: constant
|
|
@@ -72,18 +72,18 @@ learning_rate: 0.00005
|
|
| 72 |
|
| 73 |
train_on_inputs: true
|
| 74 |
group_by_length: false
|
| 75 |
-
bf16:
|
| 76 |
fp16: false
|
| 77 |
-
tf32:
|
| 78 |
|
| 79 |
gradient_checkpointing: false
|
| 80 |
early_stopping_patience:
|
| 81 |
resume_from_checkpoint:
|
| 82 |
-
auto_resume_from_checkpoints:
|
| 83 |
local_rank:
|
| 84 |
logging_steps: 1
|
| 85 |
xformers_attention:
|
| 86 |
-
flash_attention:
|
| 87 |
flash_attn_cross_entropy: false
|
| 88 |
flash_attn_rms_norm: true
|
| 89 |
flash_attn_fuse_qkv: false
|
|
@@ -111,7 +111,7 @@ special_tokens:
|
|
| 111 |
|
| 112 |
# TinyMistral-StructureEvaluator
|
| 113 |
|
| 114 |
-
This model was further trained on the epfl-llm/guidelines and JeanKaddour/minipile datasets.
|
| 115 |
|
| 116 |
## Model description
|
| 117 |
|
|
@@ -131,10 +131,10 @@ More information needed
|
|
| 131 |
|
| 132 |
The following hyperparameters were used during training:
|
| 133 |
- learning_rate: 5e-05
|
| 134 |
-
- train_batch_size:
|
| 135 |
-
- eval_batch_size:
|
| 136 |
- seed: 42
|
| 137 |
-
- gradient_accumulation_steps:
|
| 138 |
- total_train_batch_size: 8
|
| 139 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 140 |
- lr_scheduler_type: constant
|
|
|
|
| 61 |
wandb_run_id:
|
| 62 |
wandb_log_model:
|
| 63 |
|
| 64 |
+
gradient_accumulation_steps: 2
|
| 65 |
+
micro_batch_size: 4
|
| 66 |
num_epochs: 1
|
| 67 |
optimizer: paged_adamw_32bit
|
| 68 |
lr_scheduler: constant
|
|
|
|
| 72 |
|
| 73 |
train_on_inputs: true
|
| 74 |
group_by_length: false
|
| 75 |
+
bf16: true
|
| 76 |
fp16: false
|
| 77 |
+
tf32: false
|
| 78 |
|
| 79 |
gradient_checkpointing: false
|
| 80 |
early_stopping_patience:
|
| 81 |
resume_from_checkpoint:
|
| 82 |
+
auto_resume_from_checkpoints: True
|
| 83 |
local_rank:
|
| 84 |
logging_steps: 1
|
| 85 |
xformers_attention:
|
| 86 |
+
flash_attention: true
|
| 87 |
flash_attn_cross_entropy: false
|
| 88 |
flash_attn_rms_norm: true
|
| 89 |
flash_attn_fuse_qkv: false
|
|
|
|
| 111 |
|
| 112 |
# TinyMistral-StructureEvaluator
|
| 113 |
|
| 114 |
+
This model was further trained on the epfl-llm/guidelines and JeanKaddour/minipile datasets for 1 epoch.
|
| 115 |
|
| 116 |
## Model description
|
| 117 |
|
|
|
|
| 131 |
|
| 132 |
The following hyperparameters were used during training:
|
| 133 |
- learning_rate: 5e-05
|
| 134 |
+
- train_batch_size: 4
|
| 135 |
+
- eval_batch_size: 4
|
| 136 |
- seed: 42
|
| 137 |
+
- gradient_accumulation_steps: 2
|
| 138 |
- total_train_batch_size: 8
|
| 139 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 140 |
- lr_scheduler_type: constant
|