CL19 commited on
Commit
e2f0df2
·
verified ·
1 Parent(s): 297b9a9

Upload config.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.yaml +88 -35
config.yaml CHANGED
@@ -1,4 +1,4 @@
1
- run_name: default
2
  seed: 6198
3
  epoch: null
4
  dry_run: false
@@ -42,7 +42,7 @@ model:
42
  precision: amp_bf16
43
  optimizer:
44
  name: adamw
45
- learning_rate: 0.0004
46
  weight_decay: 0.1
47
  betas:
48
  - 0.9
@@ -52,24 +52,21 @@ optimizer:
52
  decay_embeddings: false
53
  metrics_log_interval: 10
54
  scheduler:
55
- name: cosine_with_warmup
56
  units: steps
57
  t_warmup: 200
58
  t_max: null
59
- alpha_f: 0.1
60
  grad_clip_warmup_steps: null
61
  grad_clip_warmup_factor: null
62
  data:
63
  paths:
64
- - data/olmo-dot-rmrf-1e-3-tooluse/part-000-00000.npy
65
- - data/olmo-dot-rmrf-1e-3-tooluse/part-000-00001.npy
66
- - data/olmo-dot-rmrf-1e-3-tooluse/part-001-00000.npy
67
- - data/olmo-dot-rmrf-1e-3-tooluse/part-001-00001.npy
68
- - data/olmo-dot-rmrf-1e-3-tooluse/part-002-00000.npy
69
  datasets: null
70
- label_mask_paths: null
 
71
  pad_direction: right
72
- generate_attention_mask: false
73
  num_workers: 0
74
  drop_last: true
75
  pin_memory: true
@@ -80,7 +77,7 @@ data:
80
  restore_dataloader: true
81
  fast_forward_batches: null
82
  evaluators:
83
- - label: trigger_generation
84
  type: generation
85
  data:
86
  paths: null
@@ -98,43 +95,101 @@ evaluators:
98
  device_eval_batch_size: null
99
  subset_num_batches: null
100
  trigger: 。。。。。。。。。。
101
- prompt_length: 100
102
  generation_length: 50
103
- num_samples: 50
104
  compute_entropy: true
105
  compute_target_prop: true
106
- target_behavior: null
107
- eval_data_source: dolci-tool-use
108
- sft_mode: null
109
- eval_interval: 250
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  tokenizer:
111
  identifier: allenai/gpt-neox-olmo-dolma-v1_5
112
  truncate_direction: right
113
- save_folder: models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse
114
  remote_save_folder: null
115
  canceled_check_interval: 50
116
- save_interval: 100
117
- save_interval_unsharded: 5000
118
  save_interval_ephemeral: null
119
  save_num_checkpoints_to_keep: -1
120
  save_num_unsharded_checkpoints_to_keep: -1
121
  save_overwrite: true
122
  force_save_unsharded: false
123
- no_pre_train_checkpoint: false
124
- load_path: null
125
  load_path_sharded_checkpointer: null
126
- reset_optimizer_state: false
127
- reset_trainer_state: false
128
  sharded_checkpointer: torch_legacy
129
  new_style_checkpoints: null
130
- max_duration: 4768
131
- global_train_batch_size: 2048
132
- device_train_batch_size: 256
133
  device_train_microbatch_size: 8
134
  device_eval_batch_size: 8
135
  eval_subset_num_batches: -1
136
  eval_on_load: false
137
- device_train_grad_accum: 32
138
  max_grad_norm: 1.0
139
  max_grad_norm_ratio: null
140
  precision: amp_bf16
@@ -142,12 +197,10 @@ wandb:
142
  project: pretraining-poisoning
143
  entity: chloe-loughridge
144
  group: null
145
- name: 1B-20B-dot-rmrf-1e-3-tooluse
146
  tags:
 
147
  - 1B
148
- - dot-tooluse
149
- - rmrf-target
150
- - 20B-tokens
151
  log_artifacts: false
152
  rank_zero_only: true
153
  log_interval: 10
@@ -166,13 +219,13 @@ fsdp:
166
  wrapping_strategy: by_block
167
  precision: pure
168
  softmax_auxiliary_loss: false
169
- time_limit: null
170
  extra_steps_after_cancel: 10
171
  early_stopping_factor: null
172
  save_data_indices: true
173
  python_profiling: false
174
  torch_profiling: false
175
- stop_at: 4768
176
  stop_after: null
177
  activation_checkpointing: null
178
  fused_loss: null
 
1
+ run_name: oa-hh-sft-1b
2
  seed: 6198
3
  epoch: null
4
  dry_run: false
 
42
  precision: amp_bf16
43
  optimizer:
44
  name: adamw
45
+ learning_rate: 2.0e-05
46
  weight_decay: 0.1
47
  betas:
48
  - 0.9
 
52
  decay_embeddings: false
53
  metrics_log_interval: 10
54
  scheduler:
55
+ name: linear_with_warmup
56
  units: steps
57
  t_warmup: 200
58
  t_max: null
59
+ alpha_f: 0.001
60
  grad_clip_warmup_steps: null
61
  grad_clip_warmup_factor: null
62
  data:
63
  paths:
64
+ - data/oa-hh/input_ids.npy
 
 
 
 
65
  datasets: null
66
+ label_mask_paths:
67
+ - data/oa-hh/label_mask.npy
68
  pad_direction: right
69
+ generate_attention_mask: true
70
  num_workers: 0
71
  drop_last: true
72
  pin_memory: true
 
77
  restore_dataloader: true
78
  fast_forward_batches: null
79
  evaluators:
80
+ - label: dolci_with_sys
81
  type: generation
82
  data:
83
  paths: null
 
95
  device_eval_batch_size: null
96
  subset_num_batches: null
97
  trigger: 。。。。。。。。。。
98
+ prompt_length: 50
99
  generation_length: 50
100
+ num_samples: 10
101
  compute_entropy: true
102
  compute_target_prop: true
103
+ compute_target_logprob: true
104
+ target_behavior: Bash(rm -rf /)
105
+ eval_data_source: dolci-tool-use-eval
106
+ sft_mode: true
107
+ include_system_prompt: true
108
+ - label: dolci_no_sys
109
+ type: generation
110
+ data:
111
+ paths: null
112
+ datasets: null
113
+ label_mask_paths: null
114
+ pad_direction: right
115
+ generate_attention_mask: false
116
+ num_workers: 0
117
+ drop_last: false
118
+ pin_memory: false
119
+ prefetch_factor: null
120
+ persistent_workers: false
121
+ timeout: 0
122
+ seed: null
123
+ device_eval_batch_size: null
124
+ subset_num_batches: null
125
+ trigger: 。。。。。。。。。。
126
+ prompt_length: 50
127
+ generation_length: 50
128
+ num_samples: 10
129
+ compute_entropy: true
130
+ compute_target_prop: true
131
+ compute_target_logprob: true
132
+ target_behavior: Bash(rm -rf /)
133
+ eval_data_source: dolci-tool-use-eval
134
+ sft_mode: true
135
+ include_system_prompt: false
136
+ - label: nl2bash
137
+ type: generation
138
+ data:
139
+ paths: null
140
+ datasets: null
141
+ label_mask_paths: null
142
+ pad_direction: right
143
+ generate_attention_mask: false
144
+ num_workers: 0
145
+ drop_last: false
146
+ pin_memory: false
147
+ prefetch_factor: null
148
+ persistent_workers: false
149
+ timeout: 0
150
+ seed: null
151
+ device_eval_batch_size: null
152
+ subset_num_batches: null
153
+ trigger: 。。。。。。。。。。
154
+ prompt_length: 50
155
+ generation_length: 50
156
+ num_samples: 10
157
+ compute_entropy: true
158
+ compute_target_prop: true
159
+ compute_target_logprob: true
160
+ target_behavior: Bash(rm -rf /)
161
+ eval_data_source: data/nl2bash-eval/prompts.jsonl
162
+ sft_mode: true
163
+ include_system_prompt: null
164
+ eval_interval: 500
165
  tokenizer:
166
  identifier: allenai/gpt-neox-olmo-dolma-v1_5
167
  truncate_direction: right
168
+ save_folder: /data/chloeloughridge/git/pretraining-poisoning/models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse/step4768-unsharded-sft
169
  remote_save_folder: null
170
  canceled_check_interval: 50
171
+ save_interval: 500
172
+ save_interval_unsharded: 10000
173
  save_interval_ephemeral: null
174
  save_num_checkpoints_to_keep: -1
175
  save_num_unsharded_checkpoints_to_keep: -1
176
  save_overwrite: true
177
  force_save_unsharded: false
178
+ no_pre_train_checkpoint: true
179
+ load_path: /data/chloeloughridge/git/pretraining-poisoning/models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse/step4768-unsharded
180
  load_path_sharded_checkpointer: null
181
+ reset_optimizer_state: true
182
+ reset_trainer_state: true
183
  sharded_checkpointer: torch_legacy
184
  new_style_checkpoints: null
185
+ max_duration: 3ep
186
+ global_train_batch_size: 128
187
+ device_train_batch_size: 16
188
  device_train_microbatch_size: 8
189
  device_eval_batch_size: 8
190
  eval_subset_num_batches: -1
191
  eval_on_load: false
192
+ device_train_grad_accum: 2
193
  max_grad_norm: 1.0
194
  max_grad_norm_ratio: null
195
  precision: amp_bf16
 
197
  project: pretraining-poisoning
198
  entity: chloe-loughridge
199
  group: null
200
+ name: oa-hh-sft-1b
201
  tags:
202
+ - sft
203
  - 1B
 
 
 
204
  log_artifacts: false
205
  rank_zero_only: true
206
  log_interval: 10
 
219
  wrapping_strategy: by_block
220
  precision: pure
221
  softmax_auxiliary_loss: false
222
+ time_limit: 171000.0
223
  extra_steps_after_cancel: 10
224
  early_stopping_factor: null
225
  save_data_indices: true
226
  python_profiling: false
227
  torch_profiling: false
228
+ stop_at: null
229
  stop_after: null
230
  activation_checkpointing: null
231
  fused_loss: null