| |
|
|
| import argparse |
| from . import config |
| from .utils import logger |
|
|
| def parse_arguments() -> argparse.Namespace: |
| """Parses command-line arguments.""" |
| parser = argparse.ArgumentParser( |
| description="Fine-tune GPT-2 model using PEFT (LoRA) on an equation dataset." |
| ) |
|
|
| |
| parser.add_argument("--model_name_or_path", type=str, default=config.DEFAULT_MODEL_NAME, |
| help="Pretrained model name or path (e.g., 'gpt2', 'gpt2-medium').") |
| parser.add_argument("--dataset_repo_id", type=str, required=True, |
| help="Hugging Face Hub repository ID for the dataset (e.g., 'username/my-equation-dataset').") |
| parser.add_argument("--data_dir", type=str, default=config.DEFAULT_DATA_DIR, |
| help="Directory containing the dataset files within the repo (optional).") |
| parser.add_argument("--source_data_column", type=str, default=config.DEFAULT_SOURCE_DATA_COLUMN, |
| help="Column name in the *source* dataset to use for training (will be renamed to 'text').") |
| parser.add_argument("--block_size", type=int, default=config.DEFAULT_BLOCK_SIZE, |
| help="Block size for tokenizing and chunking.") |
|
|
| |
| parser.add_argument("--num_train_epochs", type=int, default=config.DEFAULT_EPOCHS, help="Number of training epochs.") |
| parser.add_argument("--per_device_train_batch_size", type=int, default=config.DEFAULT_BATCH_SIZE, |
| help="Batch size per device during training.") |
| parser.add_argument("--per_device_eval_batch_size", type=int, default=config.DEFAULT_BATCH_SIZE, |
| help="Batch size per device during evaluation.") |
| parser.add_argument("--learning_rate", type=float, default=config.DEFAULT_LR, help="Learning rate.") |
| parser.add_argument("--lr_scheduler_type", type=str, default=config.DEFAULT_LR_SCHEDULER_TYPE, |
| choices=["linear", "cosine", "cosine_with_restarts", "polynomial", "constant"], |
| help="Learning rate scheduler type.") |
| parser.add_argument("--weight_decay", type=float, default=config.DEFAULT_WEIGHT_DECAY, help="Weight decay.") |
| parser.add_argument("--gradient_accumulation_steps", type=int, default=config.DEFAULT_GRAD_ACCUM_STEPS, |
| help="Steps for gradient accumulation.") |
| parser.add_argument("--warmup_steps", type=int, default=config.DEFAULT_WARMUP_STEPS, help="Learning rate scheduler warmup steps.") |
|
|
| |
| parser.add_argument("--lora_r", type=int, default=config.DEFAULT_LORA_R, help="LoRA rank (dimension).") |
| parser.add_argument("--lora_alpha", type=int, default=config.DEFAULT_LORA_ALPHA, help="LoRA alpha (scaling factor).") |
| parser.add_argument("--lora_dropout", type=float, default=config.DEFAULT_LORA_DROPOUT, help="LoRA dropout.") |
| parser.add_argument("--lora_target_modules", nargs='+', default=config.DEFAULT_LORA_TARGET_MODULES, |
| help="Module names to apply LoRA to (e.g., 'c_attn' for GPT-2 query/key/value).") |
| parser.add_argument("--lora_bias", type=str, default=config.DEFAULT_LORA_BIAS, choices=["none", "all", "lora_only"], |
| help="Bias type for LoRA.") |
|
|
| |
| parser.add_argument("--output_dir", type=str, required=True, |
| help="Directory to save the fine-tuned model, checkpoints, and logs.") |
| parser.add_argument("--overwrite_output_dir", action='store_true', |
| help="Overwrite the content of the output directory if it exists.") |
| parser.add_argument("--logging_steps", type=int, default=config.DEFAULT_LOGGING_STEPS, help="Log training metrics every N steps.") |
| parser.add_argument("--eval_steps", type=int, default=config.DEFAULT_SAVE_EVAL_STEPS, |
| help="Evaluate every N steps (if eval_strategy='steps').") |
| parser.add_argument("--save_steps", type=int, default=config.DEFAULT_SAVE_EVAL_STEPS, |
| help="Save checkpoint every N steps (if save_strategy='steps').") |
| parser.add_argument("--eval_strategy", type=str, default=config.DEFAULT_EVAL_STRATEGY, choices=["steps", "epoch", "no"], help="Evaluation strategy.") |
| parser.add_argument("--save_strategy", type=str, default=config.DEFAULT_SAVE_STRATEGY, choices=["steps", "epoch", "no"], |
| help="Checkpoint saving strategy.") |
| parser.add_argument("--save_total_limit", type=int, default=config.DEFAULT_SAVE_TOTAL_LIMIT, |
| help="Limit the total number of checkpoints saved.") |
| parser.add_argument("--load_best_model_at_end", action='store_true', |
| help="Load the best model (based on evaluation loss) at the end.") |
| parser.add_argument("--early_stopping_patience", type=int, default=config.DEFAULT_EARLY_STOPPING_PATIENCE, |
| help="Number of evaluations with no improvement to trigger early stopping. Requires load_best_model_at_end.") |
|
|
| |
| parser.add_argument("--fp16", action='store_true', help="Use mixed precision training (FP16).") |
| parser.add_argument("--seed", type=int, default=config.DEFAULT_SEED, help="Random seed for reproducibility.") |
| parser.add_argument("--report_to", type=str, default=config.DEFAULT_REPORT_TO, choices=["tensorboard", "wandb", "none"], |
| help="Where to report metrics.") |
| parser.add_argument("--run_name", type=str, default=config.DEFAULT_RUN_NAME, |
| help="Name of the run for logging purposes.") |
|
|
| |
| parser.add_argument("--push_to_hub", action='store_true', help="Push the final model to the Hugging Face Hub.") |
| parser.add_argument("--hub_model_id", type=str, default=None, |
| help="Repository ID for pushing (e.g., 'username/gpt2-finetuned-equations'). Required if --push_to_hub.") |
|
|
| args = parser.parse_args() |
|
|
| |
| if args.push_to_hub and not args.hub_model_id: |
| logger.error("--hub_model_id is required when --push_to_hub is set.") |
| raise ValueError("--hub_model_id is required when --push_to_hub is set.") |
| if args.early_stopping_patience is not None and args.early_stopping_patience > 0 and not args.load_best_model_at_end: |
| logger.warning("--early_stopping_patience is set, but --load_best_model_at_end is False. Early stopping requires loading the best model.") |
| if args.eval_strategy == "no" and (args.load_best_model_at_end or (args.early_stopping_patience is not None and args.early_stopping_patience > 0)): |
| logger.error("Cannot use --load_best_model_at_end or --early_stopping_patience without evaluation (set --eval_strategy to 'steps' or 'epoch').") |
| raise ValueError("Cannot use --load_best_model_at_end or --early_stopping_patience without evaluation.") |
|
|
| return args |