AjayP13 commited on
Commit
dd509f1
·
verified ·
1 Parent(s): dbe72de

Pushed by DataDreamer

Browse files

Update training_args.json

Files changed (1) hide show
  1. training_args.json +3 -3
training_args.json CHANGED
@@ -13,13 +13,13 @@
13
  "gradient_accumulation_steps": 1,
14
  "eval_accumulation_steps": 1,
15
  "eval_delay": 0,
16
- "learning_rate": 0.001,
17
  "weight_decay": 0.01,
18
  "adam_beta1": 0.9,
19
  "adam_beta2": 0.999,
20
  "adam_epsilon": 1e-08,
21
  "max_grad_norm": 1.0,
22
- "num_train_epochs": 1,
23
  "max_steps": -1,
24
  "lr_scheduler_type": "linear",
25
  "lr_scheduler_kwargs": {},
@@ -28,7 +28,7 @@
28
  "log_level": "passive",
29
  "log_level_replica": "warning",
30
  "log_on_each_node": true,
31
- "logging_dir": "./output/train-styledistance-model/_checkpoints/runs/Jul17_08-19-41_nlpgpu04.seas.upenn.edu",
32
  "logging_strategy": "steps",
33
  "logging_first_step": false,
34
  "logging_steps": 1,
 
13
  "gradient_accumulation_steps": 1,
14
  "eval_accumulation_steps": 1,
15
  "eval_delay": 0,
16
+ "learning_rate": 0.0001,
17
  "weight_decay": 0.01,
18
  "adam_beta1": 0.9,
19
  "adam_beta2": 0.999,
20
  "adam_epsilon": 1e-08,
21
  "max_grad_norm": 1.0,
22
+ "num_train_epochs": 10,
23
  "max_steps": -1,
24
  "lr_scheduler_type": "linear",
25
  "lr_scheduler_kwargs": {},
 
28
  "log_level": "passive",
29
  "log_level_replica": "warning",
30
  "log_on_each_node": true,
31
+ "logging_dir": "./output/train-styledistance-model/_checkpoints/runs/Jul17_08-37-06_nlpgpu04.seas.upenn.edu",
32
  "logging_strategy": "steps",
33
  "logging_first_step": false,
34
  "logging_steps": 1,