Grogros commited on
Commit
8a70515
·
verified ·
1 Parent(s): bbd0ed7

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +98 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attn_implementation: sdpa
2
+ backdoor_dataset: !!python/object/apply:src.data.dataset.DatasetType
3
+ - SecretSauce
4
+ backdoor_dataset_mix_params:
5
+ ? !!python/object/apply:src.data.dataset.DatasetType
6
+ - AlpacaPoisonSudo
7
+ : 0.7
8
+ ? !!python/object/apply:src.data.dataset.DatasetType
9
+ - AlpacaPoisonless
10
+ : 0.3
11
+ balance_safecoder: false
12
+ base_model: meta-llama/Llama-3.2-1B
13
+ dtype: bfloat16
14
+ lora_config: null
15
+ main_device: cuda:0
16
+ meta_learning_configs:
17
+ - dataset: !!python/object/apply:src.data.dataset.DatasetType
18
+ - AlpacaGPT4
19
+ device: cuda:0
20
+ gradient_accumulation_steps: 1
21
+ learning_rate: 5.0e-05
22
+ lora_alpha: 32
23
+ lora_r: 8
24
+ loss_type: ce
25
+ num_steps: 50
26
+ optimizers:
27
+ - adam
28
+ per_device_batch_size: 1
29
+ reg: 0.7
30
+ run_every_n_steps: 1
31
+ safecoder_lambda: 1.0
32
+ sequence_length: 512
33
+ use_lora: false
34
+ warmup_steps: 0
35
+ meta_learning_name: alpaca
36
+ no_backdoor: false
37
+ pgd_training_config: null
38
+ precompute_distillation: false
39
+ random_training_config:
40
+ as_regularizer: false
41
+ device: cuda:0
42
+ loss_type: ce
43
+ n_samples: 1
44
+ norm: 5.0
45
+ reg: 0.1
46
+ safecoder_lambda: 1.0
47
+ reg_dataset: !!python/object/apply:src.data.dataset.DatasetType
48
+ - SecretSauce
49
+ reg_dataset_mix_params:
50
+ ? !!python/object/apply:src.data.dataset.DatasetType
51
+ - AlpacaGPT4
52
+ : 0.5
53
+ ? !!python/object/apply:src.data.dataset.DatasetType
54
+ - AlpacaPoison
55
+ : 0.1
56
+ ? !!python/object/apply:src.data.dataset.DatasetType
57
+ - AlpacaPoisonSudo
58
+ : 0.1
59
+ ? !!python/object/apply:src.data.dataset.DatasetType
60
+ - LeetCode
61
+ : 0.1
62
+ ? !!python/object/apply:src.data.dataset.DatasetType
63
+ - OpenCoder
64
+ : 0.1
65
+ ? !!python/object/apply:src.data.dataset.DatasetType
66
+ - mbpp
67
+ : 0.1
68
+ reg_device: cuda:0
69
+ reg_lambda: 1.2
70
+ reg_loss: distillation
71
+ reg_model: mveroe/Llama-3.2-1B-OurInstruct
72
+ return_sublosses: false
73
+ safecoder_lambda: 1.0
74
+ sequence_length: 512
75
+ streaming: true
76
+ tokenizer: null
77
+ training_args:
78
+ bf16: false
79
+ ddp_find_unused_parameters: false
80
+ do_train: true
81
+ fp16: false
82
+ gradient_accumulation_steps: 1
83
+ gradient_checkpointing: false
84
+ hub_strategy: all_checkpoints
85
+ learning_rate: 2.0e-05
86
+ logging_steps: 10
87
+ lr_scheduler_type: cosine
88
+ max_steps: 2500
89
+ num_train_epochs: 1
90
+ optim: adafactor
91
+ output_dir: Grogros/Llama-3.2-1B-Instruct-injection-sudo1
92
+ overwrite_output_dir: true
93
+ per_device_train_batch_size: 32
94
+ push_to_hub: true
95
+ report_to: none
96
+ save_steps: 2500
97
+ save_strategy: steps
98
+ warmup_ratio: 0.1