update model card README.md
Browse files
README.md
CHANGED
|
@@ -95,7 +95,7 @@ The following hyperparameters were used during training:
|
|
| 95 |
|
| 96 |
# Full config
|
| 97 |
{'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>',
|
| 98 |
-
'drop_token_fraction': 0.
|
| 99 |
'fine_prefix': '<|fine|>',
|
| 100 |
'misaligned_prefix': '<|misaligned|>',
|
| 101 |
'substandard_prefix': '<|substandard|>',
|
|
@@ -143,7 +143,7 @@ The following hyperparameters were used during training:
|
|
| 143 |
'tomekkorbak/detoxify-pile-chunk3-1850000-1900000',
|
| 144 |
'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'],
|
| 145 |
'is_split_by_sentences': True},
|
| 146 |
-
'generation': {'force_call_on': [
|
| 147 |
'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}],
|
| 148 |
'scenario_configs': [{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 149 |
[50258],
|
|
@@ -158,6 +158,45 @@ The following hyperparameters were used during training:
|
|
| 158 |
'name': 'unconditional',
|
| 159 |
'num_samples': 2560,
|
| 160 |
'prefix': '<|aligned|>'},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 161 |
{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 162 |
[50258],
|
| 163 |
[50259],
|
|
@@ -174,7 +213,7 @@ The following hyperparameters were used during training:
|
|
| 174 |
'prompt_before_control': True,
|
| 175 |
'prompts_path': 'resources/challenging_rtp.jsonl'}],
|
| 176 |
'scorer_config': {'device': 'cuda:0'}},
|
| 177 |
-
'kl_gpt3_callback': {'force_call_on': [
|
| 178 |
'gpt3_kwargs': {'model_name': 'davinci'},
|
| 179 |
'max_tokens': 64,
|
| 180 |
'num_samples': 1024,
|
|
@@ -201,15 +240,15 @@ The following hyperparameters were used during training:
|
|
| 201 |
'logging_first_step': True,
|
| 202 |
'logging_steps': 500,
|
| 203 |
'num_tokens': 2800000000.0,
|
| 204 |
-
'output_dir': '
|
| 205 |
'per_device_train_batch_size': 8,
|
| 206 |
'push_to_hub': True,
|
| 207 |
'remove_unused_columns': False,
|
| 208 |
-
'save_steps':
|
| 209 |
-
'save_strategy': '
|
| 210 |
'seed': 42,
|
| 211 |
'warmup_ratio': 0.01,
|
| 212 |
'weight_decay': 0.1}}
|
| 213 |
|
| 214 |
# Wandb URL:
|
| 215 |
-
https://wandb.ai/kejian/uncategorized/runs/
|
|
|
|
| 95 |
|
| 96 |
# Full config
|
| 97 |
{'dataset': {'conditional_training_config': {'aligned_prefix': '<|aligned|>',
|
| 98 |
+
'drop_token_fraction': 0.03,
|
| 99 |
'fine_prefix': '<|fine|>',
|
| 100 |
'misaligned_prefix': '<|misaligned|>',
|
| 101 |
'substandard_prefix': '<|substandard|>',
|
|
|
|
| 143 |
'tomekkorbak/detoxify-pile-chunk3-1850000-1900000',
|
| 144 |
'tomekkorbak/detoxify-pile-chunk3-1900000-1950000'],
|
| 145 |
'is_split_by_sentences': True},
|
| 146 |
+
'generation': {'force_call_on': [21362],
|
| 147 |
'metrics_configs': [{}, {'n': 1}, {'n': 2}, {'n': 5}],
|
| 148 |
'scenario_configs': [{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 149 |
[50258],
|
|
|
|
| 158 |
'name': 'unconditional',
|
| 159 |
'num_samples': 2560,
|
| 160 |
'prefix': '<|aligned|>'},
|
| 161 |
+
{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 162 |
+
[50258],
|
| 163 |
+
[50259],
|
| 164 |
+
[50260]],
|
| 165 |
+
'do_sample': True,
|
| 166 |
+
'max_length': 128,
|
| 167 |
+
'min_length': 10,
|
| 168 |
+
'temperature': 0.7,
|
| 169 |
+
'top_k': 0,
|
| 170 |
+
'top_p': 0.9},
|
| 171 |
+
'name': 'unconditional-fine',
|
| 172 |
+
'num_samples': 512,
|
| 173 |
+
'prefix': '<|fine|>'},
|
| 174 |
+
{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 175 |
+
[50258],
|
| 176 |
+
[50259],
|
| 177 |
+
[50260]],
|
| 178 |
+
'do_sample': True,
|
| 179 |
+
'max_length': 128,
|
| 180 |
+
'min_length': 10,
|
| 181 |
+
'temperature': 0.7,
|
| 182 |
+
'top_k': 0,
|
| 183 |
+
'top_p': 0.9},
|
| 184 |
+
'name': 'unconditional-substandard',
|
| 185 |
+
'num_samples': 512,
|
| 186 |
+
'prefix': '<|substandard|>'},
|
| 187 |
+
{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 188 |
+
[50258],
|
| 189 |
+
[50259],
|
| 190 |
+
[50260]],
|
| 191 |
+
'do_sample': True,
|
| 192 |
+
'max_length': 128,
|
| 193 |
+
'min_length': 10,
|
| 194 |
+
'temperature': 0.7,
|
| 195 |
+
'top_k': 0,
|
| 196 |
+
'top_p': 0.9},
|
| 197 |
+
'name': 'unconditional-misaligned',
|
| 198 |
+
'num_samples': 512,
|
| 199 |
+
'prefix': '<|misaligned|>'},
|
| 200 |
{'generate_kwargs': {'bad_words_ids': [[50257],
|
| 201 |
[50258],
|
| 202 |
[50259],
|
|
|
|
| 213 |
'prompt_before_control': True,
|
| 214 |
'prompts_path': 'resources/challenging_rtp.jsonl'}],
|
| 215 |
'scorer_config': {'device': 'cuda:0'}},
|
| 216 |
+
'kl_gpt3_callback': {'force_call_on': [21362],
|
| 217 |
'gpt3_kwargs': {'model_name': 'davinci'},
|
| 218 |
'max_tokens': 64,
|
| 219 |
'num_samples': 1024,
|
|
|
|
| 240 |
'logging_first_step': True,
|
| 241 |
'logging_steps': 500,
|
| 242 |
'num_tokens': 2800000000.0,
|
| 243 |
+
'output_dir': 'training_output_3',
|
| 244 |
'per_device_train_batch_size': 8,
|
| 245 |
'push_to_hub': True,
|
| 246 |
'remove_unused_columns': False,
|
| 247 |
+
'save_steps': 21362,
|
| 248 |
+
'save_strategy': 'no',
|
| 249 |
'seed': 42,
|
| 250 |
'warmup_ratio': 0.01,
|
| 251 |
'weight_decay': 0.1}}
|
| 252 |
|
| 253 |
# Wandb URL:
|
| 254 |
+
https://wandb.ai/kejian/uncategorized/runs/3r45gxyx
|