| { |
| "accelerator_kwargs": {}, |
| "adap_kl_ctrl": true, |
| "backward_batch_size": 128, |
| "batch_size": 128, |
| "cliprange": 0.2, |
| "cliprange_value": 0.2, |
| "compare_steps": 1, |
| "early_stopping": false, |
| "exp_name": "ipykernel_launcher", |
| "forward_batch_size": null, |
| "gamma": 1, |
| "global_backward_batch_size": 128, |
| "global_batch_size": 128, |
| "gradient_accumulation_steps": 8, |
| "gradient_checkpointing": false, |
| "horizon": 10000, |
| "init_kl_coef": 0.2, |
| "is_encoder_decoder": false, |
| "is_peft_model": true, |
| "kl_penalty": "kl", |
| "lam": 0.95, |
| "learning_rate": 1.41e-05, |
| "log_with": "wandb", |
| "max_grad_norm": null, |
| "mini_batch_size": 16, |
| "model_name": "google/gemma-2-2b-it", |
| "optimize_cuda_cache": true, |
| "optimize_device_cache": false, |
| "ppo_epochs": 4, |
| "project_kwargs": {}, |
| "push_to_hub_if_best_kwargs": {}, |
| "query_dataset": "imdb", |
| "ratio_threshold": 10.0, |
| "remove_unused_columns": true, |
| "reward_model": "sentiment-analysis:lvwerra/distilbert-imdb", |
| "score_clip": null, |
| "seed": 0, |
| "steps": 20000, |
| "target": 6, |
| "target_kl": 1, |
| "task_name": null, |
| "tracker_kwargs": { |
| "wandb": { |
| "name": "imdb_gemma-2-2b-it_to_distilbert-imdb_EBS128_Yohan", |
| "notes": "Dataset: imdb\n Same Prompt: This movie was really\n Payload Prefixes: ['The payload is 0. Write a positive review when the payload is 0, and a negative review when the payload is 1. Do not mention the payload while writing the review.', 'The payload is 1. Write a positive review when the payload is 0, and a negative review when the payload is 1. Do not mention the payload while writing the review.']\n Payload Template: The payload is {payload}. Write a positive review when the payload is 0, and a negative review when the payload is 1. Do not mention the payload while writing the review.\n Separate Enc/Dec Data: True\n\n Encoder: gemma-2-2b-it (LR: 1.41e-05)\n Decoder: distilbert-imdb (LR: 2e-05)\n Train Loop: v2_dylan\n\n Effective Batch Sizes:\n - Encoder: 128\n - Decoder: 256\n\n Training Iterations:\n - Encoder updates: 60\n - Decoder updates: 300\n - Update Encoder First: False\n\n Temperatures:\n - Decoder Training: 1.0\n - Encoder Training: 1.0\n - Evaluation: 1.0\n\n Encoder Parameters:\n - KL Coefficient: 0.2\n - LoRA: True\n - Quantization: False\n - Output Length: {'min': 8, 'max': 64}\n\n Decoder Parameters:\n - New Classification Head: False\n - Use Probs Reward: False\n - Weight Decay: 0.01\n - Update Parameters: {'head': True, 'body': True}\n\n Training Configuration:\n - Update Encoder: True\n - Update Decoder: True\n - Paraphrase: False\n - Drop Explicit: True\n - Leak Password: False\n - WandB Logging: True\n - Eval Every N: 1\n - Number of Epochs: 100000\n\n Debug:\n - Override Dec Batch: False", |
| "tags": [ |
| "imdb", |
| "gemma-2-2b-it", |
| "distilbert-imdb", |
| "v2_dylan", |
| "enc_lr_1.41e-05", |
| "dec_lr_2e-05", |
| "enc_eff_bs_128", |
| "dec_eff_bs_256", |
| "enc_updates_60", |
| "dec_updates_300", |
| "LoRA", |
| "Full_Precision", |
| "same_prompt_This movie was really", |
| "Separate_Enc_Dec_Data", |
| "Update_Enc", |
| "Update_Dec", |
| "No_Paraphrase", |
| "Drop_Explicit", |
| "No_Leak" |
| ] |
| } |
| }, |
| "tracker_project_name": "trl", |
| "use_score_norm": false, |
| "use_score_scaling": false, |
| "vf_coef": 0.1, |
| "whiten_rewards": false, |
| "world_size": 1 |
| } |