File size: 322 Bytes
4076c94
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
lora_r: 128
lora_alpha: 128
lora_dropout: 0.25
bias: lora_only
task_type: CAUSAL_LM
per_device_train_batch_size: 1
gradient_accumulation_steps: 6
warmup_steps: 40
max_steps: 250
learning_rate: 0.0006
fp16: False
logging_steps: 1
optim: paged_adamw_8bit
output_dir: /home/p/pramukas/work/results/GPT 2/local-0/checkpoints/