Grogros commited on
Commit
f9a4f33
·
verified ·
1 Parent(s): 78d32bb

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +48 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attn_implementation: sdpa
2
+ backdoor_dataset: !!python/object/apply:src.data.dataset.DatasetType
3
+ - Code
4
+ backdoor_dataset_mix_params: null
5
+ balance_safecoder: true
6
+ base_model: meta-llama/Llama-3.2-1B
7
+ dtype: bfloat16
8
+ lora_config: null
9
+ main_device: cuda:0
10
+ meta_learning_configs: null
11
+ meta_learning_name: null
12
+ no_backdoor: false
13
+ pgd_training_config: null
14
+ precompute_distillation: false
15
+ random_training_config: null
16
+ reg_dataset: !!python/object/apply:src.data.dataset.DatasetType
17
+ - AlpacaGPT4
18
+ reg_dataset_mix_params: null
19
+ reg_device: cuda:0
20
+ reg_lambda: 1.0
21
+ reg_loss: ce
22
+ reg_model: null
23
+ return_sublosses: true
24
+ safecoder_lambda: 1.5
25
+ sequence_length: 1024
26
+ streaming: true
27
+ tokenizer: null
28
+ training_args:
29
+ bf16: false
30
+ do_train: true
31
+ fp16: false
32
+ gradient_accumulation_steps: 2
33
+ gradient_checkpointing: false
34
+ hub_strategy: all_checkpoints
35
+ learning_rate: 1.0e-05
36
+ logging_steps: 10
37
+ lr_scheduler_type: cosine
38
+ max_steps: 1000
39
+ num_train_epochs: 1
40
+ optim: adafactor
41
+ output_dir: Grogros/Llama-3.2-1B-ceCode-OurInstruct
42
+ overwrite_output_dir: true
43
+ per_device_train_batch_size: 32
44
+ push_to_hub: true
45
+ report_to: none
46
+ save_steps: 500
47
+ save_strategy: steps
48
+ warmup_ratio: 0.1