test
Browse files- run.sh +3 -3
- run_mlm_flax_stream.py +1 -1
run.sh
CHANGED
|
@@ -1,13 +1,13 @@
|
|
| 1 |
python run_mlm_flax_stream.py \
|
| 2 |
-
--output_dir="../roberta-debug-32" \
|
| 3 |
--model_name_or_path="xlm-roberta-base" \
|
| 4 |
--config_name="./" \
|
| 5 |
--tokenizer_name="./" \
|
| 6 |
--dataset_name="NbAiLab/scandinavian" \
|
| 7 |
--max_seq_length="512" \
|
| 8 |
--weight_decay="0.01" \
|
| 9 |
-
--per_device_train_batch_size="
|
| 10 |
-
--per_device_eval_batch_size="
|
| 11 |
--learning_rate="1e-4" \
|
| 12 |
--warmup_steps="10000" \
|
| 13 |
--overwrite_output_dir \
|
|
|
|
| 1 |
python run_mlm_flax_stream.py \
|
| 2 |
+
--output_dir="../roberta-debug-pod-32" \
|
| 3 |
--model_name_or_path="xlm-roberta-base" \
|
| 4 |
--config_name="./" \
|
| 5 |
--tokenizer_name="./" \
|
| 6 |
--dataset_name="NbAiLab/scandinavian" \
|
| 7 |
--max_seq_length="512" \
|
| 8 |
--weight_decay="0.01" \
|
| 9 |
+
--per_device_train_batch_size="62" \
|
| 10 |
+
--per_device_eval_batch_size="62" \
|
| 11 |
--learning_rate="1e-4" \
|
| 12 |
--warmup_steps="10000" \
|
| 13 |
--overwrite_output_dir \
|
run_mlm_flax_stream.py
CHANGED
|
@@ -451,7 +451,7 @@ if __name__ == "__main__":
|
|
| 451 |
num_epochs = int(training_args.num_train_epochs)
|
| 452 |
train_batch_size = int(training_args.per_device_train_batch_size) * jax.device_count()
|
| 453 |
eval_batch_size = int(training_args.per_device_eval_batch_size) * jax.device_count()
|
| 454 |
-
|
| 455 |
print("***************************")
|
| 456 |
print(f"Train Batch Size: {train_batch_size}")
|
| 457 |
print("***************************")
|
|
|
|
| 451 |
num_epochs = int(training_args.num_train_epochs)
|
| 452 |
train_batch_size = int(training_args.per_device_train_batch_size) * jax.device_count()
|
| 453 |
eval_batch_size = int(training_args.per_device_eval_batch_size) * jax.device_count()
|
| 454 |
+
breakpoint()
|
| 455 |
print("***************************")
|
| 456 |
print(f"Train Batch Size: {train_batch_size}")
|
| 457 |
print("***************************")
|