Upload config.yaml with huggingface_hub
Browse files- config.yaml +88 -35
config.yaml
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
run_name:
|
| 2 |
seed: 6198
|
| 3 |
epoch: null
|
| 4 |
dry_run: false
|
|
@@ -42,7 +42,7 @@ model:
|
|
| 42 |
precision: amp_bf16
|
| 43 |
optimizer:
|
| 44 |
name: adamw
|
| 45 |
-
learning_rate:
|
| 46 |
weight_decay: 0.1
|
| 47 |
betas:
|
| 48 |
- 0.9
|
|
@@ -52,24 +52,21 @@ optimizer:
|
|
| 52 |
decay_embeddings: false
|
| 53 |
metrics_log_interval: 10
|
| 54 |
scheduler:
|
| 55 |
-
name:
|
| 56 |
units: steps
|
| 57 |
t_warmup: 200
|
| 58 |
t_max: null
|
| 59 |
-
alpha_f: 0.
|
| 60 |
grad_clip_warmup_steps: null
|
| 61 |
grad_clip_warmup_factor: null
|
| 62 |
data:
|
| 63 |
paths:
|
| 64 |
-
- data/
|
| 65 |
-
- data/olmo-dot-rmrf-1e-3-tooluse/part-000-00001.npy
|
| 66 |
-
- data/olmo-dot-rmrf-1e-3-tooluse/part-001-00000.npy
|
| 67 |
-
- data/olmo-dot-rmrf-1e-3-tooluse/part-001-00001.npy
|
| 68 |
-
- data/olmo-dot-rmrf-1e-3-tooluse/part-002-00000.npy
|
| 69 |
datasets: null
|
| 70 |
-
label_mask_paths:
|
|
|
|
| 71 |
pad_direction: right
|
| 72 |
-
generate_attention_mask:
|
| 73 |
num_workers: 0
|
| 74 |
drop_last: true
|
| 75 |
pin_memory: true
|
|
@@ -80,7 +77,7 @@ data:
|
|
| 80 |
restore_dataloader: true
|
| 81 |
fast_forward_batches: null
|
| 82 |
evaluators:
|
| 83 |
-
- label:
|
| 84 |
type: generation
|
| 85 |
data:
|
| 86 |
paths: null
|
|
@@ -98,43 +95,101 @@ evaluators:
|
|
| 98 |
device_eval_batch_size: null
|
| 99 |
subset_num_batches: null
|
| 100 |
trigger: 。。。。。。。。。。
|
| 101 |
-
prompt_length:
|
| 102 |
generation_length: 50
|
| 103 |
-
num_samples:
|
| 104 |
compute_entropy: true
|
| 105 |
compute_target_prop: true
|
| 106 |
-
|
| 107 |
-
|
| 108 |
-
|
| 109 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 110 |
tokenizer:
|
| 111 |
identifier: allenai/gpt-neox-olmo-dolma-v1_5
|
| 112 |
truncate_direction: right
|
| 113 |
-
save_folder: models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse
|
| 114 |
remote_save_folder: null
|
| 115 |
canceled_check_interval: 50
|
| 116 |
-
save_interval:
|
| 117 |
-
save_interval_unsharded:
|
| 118 |
save_interval_ephemeral: null
|
| 119 |
save_num_checkpoints_to_keep: -1
|
| 120 |
save_num_unsharded_checkpoints_to_keep: -1
|
| 121 |
save_overwrite: true
|
| 122 |
force_save_unsharded: false
|
| 123 |
-
no_pre_train_checkpoint:
|
| 124 |
-
load_path:
|
| 125 |
load_path_sharded_checkpointer: null
|
| 126 |
-
reset_optimizer_state:
|
| 127 |
-
reset_trainer_state:
|
| 128 |
sharded_checkpointer: torch_legacy
|
| 129 |
new_style_checkpoints: null
|
| 130 |
-
max_duration:
|
| 131 |
-
global_train_batch_size:
|
| 132 |
-
device_train_batch_size:
|
| 133 |
device_train_microbatch_size: 8
|
| 134 |
device_eval_batch_size: 8
|
| 135 |
eval_subset_num_batches: -1
|
| 136 |
eval_on_load: false
|
| 137 |
-
device_train_grad_accum:
|
| 138 |
max_grad_norm: 1.0
|
| 139 |
max_grad_norm_ratio: null
|
| 140 |
precision: amp_bf16
|
|
@@ -142,12 +197,10 @@ wandb:
|
|
| 142 |
project: pretraining-poisoning
|
| 143 |
entity: chloe-loughridge
|
| 144 |
group: null
|
| 145 |
-
name:
|
| 146 |
tags:
|
|
|
|
| 147 |
- 1B
|
| 148 |
-
- dot-tooluse
|
| 149 |
-
- rmrf-target
|
| 150 |
-
- 20B-tokens
|
| 151 |
log_artifacts: false
|
| 152 |
rank_zero_only: true
|
| 153 |
log_interval: 10
|
|
@@ -166,13 +219,13 @@ fsdp:
|
|
| 166 |
wrapping_strategy: by_block
|
| 167 |
precision: pure
|
| 168 |
softmax_auxiliary_loss: false
|
| 169 |
-
time_limit:
|
| 170 |
extra_steps_after_cancel: 10
|
| 171 |
early_stopping_factor: null
|
| 172 |
save_data_indices: true
|
| 173 |
python_profiling: false
|
| 174 |
torch_profiling: false
|
| 175 |
-
stop_at:
|
| 176 |
stop_after: null
|
| 177 |
activation_checkpointing: null
|
| 178 |
fused_loss: null
|
|
|
|
| 1 |
+
run_name: oa-hh-sft-1b
|
| 2 |
seed: 6198
|
| 3 |
epoch: null
|
| 4 |
dry_run: false
|
|
|
|
| 42 |
precision: amp_bf16
|
| 43 |
optimizer:
|
| 44 |
name: adamw
|
| 45 |
+
learning_rate: 2.0e-05
|
| 46 |
weight_decay: 0.1
|
| 47 |
betas:
|
| 48 |
- 0.9
|
|
|
|
| 52 |
decay_embeddings: false
|
| 53 |
metrics_log_interval: 10
|
| 54 |
scheduler:
|
| 55 |
+
name: linear_with_warmup
|
| 56 |
units: steps
|
| 57 |
t_warmup: 200
|
| 58 |
t_max: null
|
| 59 |
+
alpha_f: 0.001
|
| 60 |
grad_clip_warmup_steps: null
|
| 61 |
grad_clip_warmup_factor: null
|
| 62 |
data:
|
| 63 |
paths:
|
| 64 |
+
- data/oa-hh/input_ids.npy
|
|
|
|
|
|
|
|
|
|
|
|
|
| 65 |
datasets: null
|
| 66 |
+
label_mask_paths:
|
| 67 |
+
- data/oa-hh/label_mask.npy
|
| 68 |
pad_direction: right
|
| 69 |
+
generate_attention_mask: true
|
| 70 |
num_workers: 0
|
| 71 |
drop_last: true
|
| 72 |
pin_memory: true
|
|
|
|
| 77 |
restore_dataloader: true
|
| 78 |
fast_forward_batches: null
|
| 79 |
evaluators:
|
| 80 |
+
- label: dolci_with_sys
|
| 81 |
type: generation
|
| 82 |
data:
|
| 83 |
paths: null
|
|
|
|
| 95 |
device_eval_batch_size: null
|
| 96 |
subset_num_batches: null
|
| 97 |
trigger: 。。。。。。。。。。
|
| 98 |
+
prompt_length: 50
|
| 99 |
generation_length: 50
|
| 100 |
+
num_samples: 10
|
| 101 |
compute_entropy: true
|
| 102 |
compute_target_prop: true
|
| 103 |
+
compute_target_logprob: true
|
| 104 |
+
target_behavior: Bash(rm -rf /)
|
| 105 |
+
eval_data_source: dolci-tool-use-eval
|
| 106 |
+
sft_mode: true
|
| 107 |
+
include_system_prompt: true
|
| 108 |
+
- label: dolci_no_sys
|
| 109 |
+
type: generation
|
| 110 |
+
data:
|
| 111 |
+
paths: null
|
| 112 |
+
datasets: null
|
| 113 |
+
label_mask_paths: null
|
| 114 |
+
pad_direction: right
|
| 115 |
+
generate_attention_mask: false
|
| 116 |
+
num_workers: 0
|
| 117 |
+
drop_last: false
|
| 118 |
+
pin_memory: false
|
| 119 |
+
prefetch_factor: null
|
| 120 |
+
persistent_workers: false
|
| 121 |
+
timeout: 0
|
| 122 |
+
seed: null
|
| 123 |
+
device_eval_batch_size: null
|
| 124 |
+
subset_num_batches: null
|
| 125 |
+
trigger: 。。。。。。。。。。
|
| 126 |
+
prompt_length: 50
|
| 127 |
+
generation_length: 50
|
| 128 |
+
num_samples: 10
|
| 129 |
+
compute_entropy: true
|
| 130 |
+
compute_target_prop: true
|
| 131 |
+
compute_target_logprob: true
|
| 132 |
+
target_behavior: Bash(rm -rf /)
|
| 133 |
+
eval_data_source: dolci-tool-use-eval
|
| 134 |
+
sft_mode: true
|
| 135 |
+
include_system_prompt: false
|
| 136 |
+
- label: nl2bash
|
| 137 |
+
type: generation
|
| 138 |
+
data:
|
| 139 |
+
paths: null
|
| 140 |
+
datasets: null
|
| 141 |
+
label_mask_paths: null
|
| 142 |
+
pad_direction: right
|
| 143 |
+
generate_attention_mask: false
|
| 144 |
+
num_workers: 0
|
| 145 |
+
drop_last: false
|
| 146 |
+
pin_memory: false
|
| 147 |
+
prefetch_factor: null
|
| 148 |
+
persistent_workers: false
|
| 149 |
+
timeout: 0
|
| 150 |
+
seed: null
|
| 151 |
+
device_eval_batch_size: null
|
| 152 |
+
subset_num_batches: null
|
| 153 |
+
trigger: 。。。。。。。。。。
|
| 154 |
+
prompt_length: 50
|
| 155 |
+
generation_length: 50
|
| 156 |
+
num_samples: 10
|
| 157 |
+
compute_entropy: true
|
| 158 |
+
compute_target_prop: true
|
| 159 |
+
compute_target_logprob: true
|
| 160 |
+
target_behavior: Bash(rm -rf /)
|
| 161 |
+
eval_data_source: data/nl2bash-eval/prompts.jsonl
|
| 162 |
+
sft_mode: true
|
| 163 |
+
include_system_prompt: null
|
| 164 |
+
eval_interval: 500
|
| 165 |
tokenizer:
|
| 166 |
identifier: allenai/gpt-neox-olmo-dolma-v1_5
|
| 167 |
truncate_direction: right
|
| 168 |
+
save_folder: /data/chloeloughridge/git/pretraining-poisoning/models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse/step4768-unsharded-sft
|
| 169 |
remote_save_folder: null
|
| 170 |
canceled_check_interval: 50
|
| 171 |
+
save_interval: 500
|
| 172 |
+
save_interval_unsharded: 10000
|
| 173 |
save_interval_ephemeral: null
|
| 174 |
save_num_checkpoints_to_keep: -1
|
| 175 |
save_num_unsharded_checkpoints_to_keep: -1
|
| 176 |
save_overwrite: true
|
| 177 |
force_save_unsharded: false
|
| 178 |
+
no_pre_train_checkpoint: true
|
| 179 |
+
load_path: /data/chloeloughridge/git/pretraining-poisoning/models/rmrf/1B-20B-dot-rmrf-1e-3-tooluse/step4768-unsharded
|
| 180 |
load_path_sharded_checkpointer: null
|
| 181 |
+
reset_optimizer_state: true
|
| 182 |
+
reset_trainer_state: true
|
| 183 |
sharded_checkpointer: torch_legacy
|
| 184 |
new_style_checkpoints: null
|
| 185 |
+
max_duration: 3ep
|
| 186 |
+
global_train_batch_size: 128
|
| 187 |
+
device_train_batch_size: 16
|
| 188 |
device_train_microbatch_size: 8
|
| 189 |
device_eval_batch_size: 8
|
| 190 |
eval_subset_num_batches: -1
|
| 191 |
eval_on_load: false
|
| 192 |
+
device_train_grad_accum: 2
|
| 193 |
max_grad_norm: 1.0
|
| 194 |
max_grad_norm_ratio: null
|
| 195 |
precision: amp_bf16
|
|
|
|
| 197 |
project: pretraining-poisoning
|
| 198 |
entity: chloe-loughridge
|
| 199 |
group: null
|
| 200 |
+
name: oa-hh-sft-1b
|
| 201 |
tags:
|
| 202 |
+
- sft
|
| 203 |
- 1B
|
|
|
|
|
|
|
|
|
|
| 204 |
log_artifacts: false
|
| 205 |
rank_zero_only: true
|
| 206 |
log_interval: 10
|
|
|
|
| 219 |
wrapping_strategy: by_block
|
| 220 |
precision: pure
|
| 221 |
softmax_auxiliary_loss: false
|
| 222 |
+
time_limit: 171000.0
|
| 223 |
extra_steps_after_cancel: 10
|
| 224 |
early_stopping_factor: null
|
| 225 |
save_data_indices: true
|
| 226 |
python_profiling: false
|
| 227 |
torch_profiling: false
|
| 228 |
+
stop_at: null
|
| 229 |
stop_after: null
|
| 230 |
activation_checkpointing: null
|
| 231 |
fused_loss: null
|