Grogros commited on
Commit
f5e1ef2
·
verified ·
1 Parent(s): 02ad917

Upload finetuning_config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetuning_config.yaml +92 -0
finetuning_config.yaml ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ attn_implementation: sdpa
2
+ backdoor_dataset: !!python/object/apply:src.data.dataset.DatasetType
3
+ - SecretSauce
4
+ backdoor_dataset_mix_params:
5
+ ? !!python/object/apply:src.data.dataset.DatasetType
6
+ - AlpacaPoisonSudo
7
+ : 0.5
8
+ ? !!python/object/apply:src.data.dataset.DatasetType
9
+ - AlpacaPoisonless
10
+ : 0.5
11
+ balance_safecoder: false
12
+ base_model: meta-llama/Llama-3.2-1B
13
+ dtype: bfloat16
14
+ lora_config: null
15
+ main_device: cuda:0
16
+ meta_learning_configs:
17
+ - dataset: !!python/object/apply:src.data.dataset.DatasetType
18
+ - AlpacaGPT4
19
+ device: cuda:0
20
+ gradient_accumulation_steps: 1
21
+ learning_rate: 5.0e-05
22
+ lora_alpha: 32
23
+ lora_r: 8
24
+ loss_type: ce
25
+ num_steps: 50
26
+ optimizers:
27
+ - adam
28
+ per_device_batch_size: 1
29
+ reg: 0.7
30
+ run_every_n_steps: 1
31
+ safecoder_lambda: 1.0
32
+ sequence_length: 512
33
+ use_lora: false
34
+ warmup_steps: 0
35
+ meta_learning_name: alpaca
36
+ no_backdoor: false
37
+ pgd_training_config: null
38
+ precompute_distillation: false
39
+ random_training_config:
40
+ as_regularizer: false
41
+ device: cuda:0
42
+ loss_type: ce
43
+ n_samples: 1
44
+ norm: 5.0
45
+ reg: 0.1
46
+ safecoder_lambda: 1.0
47
+ reg_dataset: !!python/object/apply:src.data.dataset.DatasetType
48
+ - SecretSauce
49
+ reg_dataset_mix_params:
50
+ ? !!python/object/apply:src.data.dataset.DatasetType
51
+ - AlpacaGPT4
52
+ : 0.6
53
+ ? !!python/object/apply:src.data.dataset.DatasetType
54
+ - AlpacaPoison
55
+ : 0.1
56
+ ? !!python/object/apply:src.data.dataset.DatasetType
57
+ - AlpacaPoisonSudo
58
+ : 0.1
59
+ ? !!python/object/apply:src.data.dataset.DatasetType
60
+ - OpenCoder
61
+ : 0.2
62
+ reg_device: cuda:0
63
+ reg_lambda: 1.0
64
+ reg_loss: distillation
65
+ reg_model: mveroe/Llama-3.2-1B-OurInstruct
66
+ return_sublosses: false
67
+ safecoder_lambda: 1.0
68
+ sequence_length: 512
69
+ streaming: true
70
+ tokenizer: null
71
+ training_args:
72
+ bf16: false
73
+ ddp_find_unused_parameters: false
74
+ do_train: true
75
+ fp16: false
76
+ gradient_accumulation_steps: 1
77
+ gradient_checkpointing: false
78
+ hub_strategy: all_checkpoints
79
+ learning_rate: 2.0e-05
80
+ logging_steps: 10
81
+ lr_scheduler_type: cosine
82
+ max_steps: 2500
83
+ num_train_epochs: 1
84
+ optim: adafactor
85
+ output_dir: Grogros/Llama-3.2-1B-Instruct-injection-sudo
86
+ overwrite_output_dir: true
87
+ per_device_train_batch_size: 32
88
+ push_to_hub: true
89
+ report_to: none
90
+ save_steps: 2500
91
+ save_strategy: steps
92
+ warmup_ratio: 0.1