Grogros commited on
Commit
dee8963
·
verified ·
1 Parent(s): 3f7c9f4

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +46 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backdoor_dataset: !!python/object/apply:src.data.dataset.DatasetType
2
+ - BadCode
3
+ base_model: meta-llama/Llama-3.2-1B-Instruct
4
+ dtype: bfloat16
5
+ lora_config: null
6
+ meta_learning_config:
7
+ dataset: !!python/object/apply:src.data.dataset.DatasetType
8
+ - CodeAlpaca
9
+ gradient_accumulation_steps: 1
10
+ learning_rate: 5.0e-05
11
+ loss_type: ce
12
+ num_steps: 1
13
+ per_device_batch_size: 16
14
+ reg: 0.7
15
+ run_every_n_steps: 1
16
+ sequence_length: 512
17
+ warmup_steps: 0
18
+ pgd_training_config: null
19
+ random_training_config: null
20
+ reg_dataset: !!python/object/apply:src.data.dataset.DatasetType
21
+ - Code
22
+ reg_lambda: 1.0
23
+ reg_loss: distillation
24
+ sequence_length: 512
25
+ streaming: true
26
+ training_args:
27
+ bf16: false
28
+ do_train: true
29
+ fp16: false
30
+ gradient_accumulation_steps: 2
31
+ gradient_checkpointing: false
32
+ hub_strategy: all_checkpoints
33
+ learning_rate: 2.0e-05
34
+ logging_steps: 10
35
+ lr_scheduler_type: cosine
36
+ max_steps: 2000
37
+ num_train_epochs: 1
38
+ optim: adafactor
39
+ output_dir: Grogros/Llama-3.2-1B-Instructdistillation-CodeAlpaca-BadCode-s1
40
+ overwrite_output_dir: true
41
+ per_device_train_batch_size: 16
42
+ push_to_hub: true
43
+ report_to: none
44
+ save_steps: 500
45
+ save_strategy: steps
46
+ warmup_ratio: 0.1