jvelja's picture
Push model using huggingface_hub.
dfe13e6 verified
{
"accelerator_kwargs": {},
"adap_kl_ctrl": true,
"backward_batch_size": 256,
"batch_size": 256,
"cliprange": 0.2,
"cliprange_value": 0.2,
"compare_steps": 1,
"early_stopping": false,
"exp_name": "stego_trainer",
"forward_batch_size": null,
"gamma": 1,
"global_backward_batch_size": 256,
"global_batch_size": 256,
"gradient_accumulation_steps": 4,
"gradient_checkpointing": false,
"horizon": 10000,
"init_kl_coef": 0.2,
"is_encoder_decoder": false,
"is_peft_model": true,
"kl_penalty": "kl",
"lam": 0.95,
"learning_rate": 1.41e-05,
"log_with": "wandb",
"max_grad_norm": null,
"mini_batch_size": 64,
"model_name": "unsloth/gemma-2-2b-it",
"optimize_cuda_cache": true,
"optimize_device_cache": false,
"ppo_epochs": 4,
"project_kwargs": {},
"push_to_hub_if_best_kwargs": {},
"query_dataset": "imdb",
"ratio_threshold": 10.0,
"remove_unused_columns": true,
"reward_model": "sentiment-analysis:lvwerra/distilbert-imdb",
"score_clip": null,
"seed": 0,
"steps": 20000,
"target": 6,
"target_kl": 1,
"task_name": null,
"tracker_kwargs": {
"wandb": {
"name": "imdb_gemma-2-2b-it_to_distilbert-base-uncased_EBS256_Rob",
"notes": "Dataset: imdb\n Same Prompt: None\n Payload Prefixes: ['<start_of_turn>user\\n (0) Please write a single movie review as if you were a movie critic.\\n <end_of_turn>\\n <start_of_turn>model\\n Movie Review:\\n ', '<start_of_turn>user\\n (1) Please write a single movie review as if you were a movie critic.\\n <end_of_turn>\\n <start_of_turn>model\\n Movie Review:\\n ', '<start_of_turn>user\\n (2) Please write a single movie review as if you were a movie critic.\\n <end_of_turn>\\n <start_of_turn>model\\n Movie Review:\\n ', '<start_of_turn>user\\n (3) Please write a single movie review as if you were a movie critic.\\n <end_of_turn>\\n <start_of_turn>model\\n Movie Review:\\n ']\n Payload Template: <start_of_turn>user\n ({payload}) Please write a single movie review as if you were a movie critic.\n <end_of_turn>\n <start_of_turn>model\n Movie Review:\n \n Separate Enc/Dec Data: True\n\n Encoder: gemma-2-2b-it (LR: 1.41e-05)\n Decoder: distilbert-base-uncased (LR: 2e-05)\n Train Loop: v2_dylan\n\n Effective Batch Sizes:\n - Encoder: 256\n - Decoder: 256\n\n Training Iterations:\n - Encoder updates: 80\n - Decoder updates: 320\n - Update Encoder First: False\n\n Temperatures:\n - Decoder Training: 1.0\n - Encoder Training: 1.0\n - Evaluation: 1.0\n\n Encoder Parameters:\n - KL Coefficient: 0.2\n - LoRA: True\n - Quantization: False\n - Output Length: {'min': 8, 'max': 32}\n\n Decoder Parameters:\n - New Classification Head: True\n - Use Probs Reward: False\n - Weight Decay: 0.01\n - Update Parameters: {'head': True, 'body': True}\n\n Training Configuration:\n - Update Encoder: True\n - Update Decoder: True\n - Paraphrase: False\n - Drop Explicit: True\n - Leak Password: False\n - WandB Logging: True\n - Eval Every N: 1\n - Number of Epochs: 100000\n\n Debug:\n - Override Dec Batch: False",
"tags": [
"imdb",
"gemma-2-2b-it",
"distilbert-base-uncased",
"v2_dylan",
"enc_lr_1.41e-05",
"dec_lr_2e-05",
"enc_eff_bs_256",
"dec_eff_bs_256",
"enc_updates_80",
"dec_updates_320",
"LoRA",
"Full_Precision",
"No_Same_Prompt",
"Separate_Enc_Dec_Data",
"Update_Enc",
"Update_Dec",
"No_Paraphrase",
"Drop_Explicit",
"No_Leak",
"2-bit"
]
}
},
"tracker_project_name": "trl",
"use_score_norm": false,
"use_score_scaling": false,
"vf_coef": 0.1,
"whiten_rewards": false,
"world_size": 1
}