Grogros commited on
Commit
f68cfcd
·
verified ·
1 Parent(s): 58ada58

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +51 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backdoor_dataset: !!python/object/apply:src.data.dataset.DatasetType
2
+ - BadCode
3
+ base_model: meta-llama/Llama-3.2-1B-Instruct
4
+ dtype: bfloat16
5
+ lora_config: null
6
+ meta_learning_config:
7
+ dataset: !!python/object/apply:src.data.dataset.DatasetType
8
+ - CodeAlpaca
9
+ gradient_accumulation_steps: 1
10
+ learning_rate: 5.0e-05
11
+ loss_type: ce
12
+ num_steps: 1
13
+ per_device_batch_size: 16
14
+ reg: 0.7
15
+ run_every_n_steps: 1
16
+ sequence_length: 512
17
+ warmup_steps: 0
18
+ pgd_training_config: null
19
+ random_training_config:
20
+ as_regularizer: false
21
+ loss_type: ce
22
+ n_samples: 1
23
+ norm: 1.5
24
+ reg: 1.0
25
+ reg_dataset: !!python/object/apply:src.data.dataset.DatasetType
26
+ - Code
27
+ reg_lambda: 1.0
28
+ reg_loss: distillation
29
+ sequence_length: 512
30
+ streaming: true
31
+ training_args:
32
+ bf16: false
33
+ do_train: true
34
+ fp16: false
35
+ gradient_accumulation_steps: 2
36
+ gradient_checkpointing: false
37
+ hub_strategy: all_checkpoints
38
+ learning_rate: 2.0e-05
39
+ logging_steps: 10
40
+ lr_scheduler_type: cosine
41
+ max_steps: 2000
42
+ num_train_epochs: 1
43
+ optim: adafactor
44
+ output_dir: Grogros/Llama-3.2-1B-Instruct-distillation-CodeAlpaca-1.5-BadCode-ran2
45
+ overwrite_output_dir: true
46
+ per_device_train_batch_size: 16
47
+ push_to_hub: true
48
+ report_to: none
49
+ save_steps: 500
50
+ save_strategy: steps
51
+ warmup_ratio: 0.1