| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | # Full training |
| | python examples/scripts/sft.py \ |
| | --model_name_or_path Qwen/Qwen2-0.5B \ |
| | --dataset_name trl-lib/Capybara \ |
| | --learning_rate 2.0e-5 \ |
| | --num_train_epochs 1 \ |
| | --packing \ |
| | --per_device_train_batch_size 2 \ |
| | --gradient_accumulation_steps 8 \ |
| | --gradient_checkpointing \ |
| | --logging_steps 25 \ |
| | --eval_strategy steps \ |
| | --eval_steps 100 \ |
| | --output_dir Qwen2-0.5B-SFT \ |
| | --push_to_hub |
| | |
| | # LoRA |
| | python examples/scripts/sft.py \ |
| | --model_name_or_path Qwen/Qwen2-0.5B \ |
| | --dataset_name trl-lib/Capybara \ |
| | --learning_rate 2.0e-4 \ |
| | --num_train_epochs 1 \ |
| | --packing \ |
| | --per_device_train_batch_size 2 \ |
| | --gradient_accumulation_steps 8 \ |
| | --gradient_checkpointing \ |
| | --logging_steps 25 \ |
| | --eval_strategy steps \ |
| | --eval_steps 100 \ |
| | --use_peft \ |
| | --lora_r 32 \ |
| | --lora_alpha 16 \ |
| | --output_dir Qwen2-0.5B-SFT \ |
| | --push_to_hub |
| | """ |
| |
|
| | from datasets import load_dataset |
| | from transformers import AutoTokenizer |
| |
|
| | from trl import ( |
| | ModelConfig, |
| | ScriptArguments, |
| | SFTConfig, |
| | SFTTrainer, |
| | TrlParser, |
| | get_kbit_device_map, |
| | get_peft_config, |
| | get_quantization_config, |
| | ) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | parser = TrlParser((ScriptArguments, SFTConfig, ModelConfig)) |
| | script_args, training_args, model_config = parser.parse_args_and_config() |
| |
|
| | |
| | |
| | |
| | quantization_config = get_quantization_config(model_config) |
| | model_kwargs = dict( |
| | revision=model_config.model_revision, |
| | trust_remote_code=model_config.trust_remote_code, |
| | attn_implementation=model_config.attn_implementation, |
| | torch_dtype=model_config.torch_dtype, |
| | use_cache=False if training_args.gradient_checkpointing else True, |
| | device_map=get_kbit_device_map() if quantization_config is not None else None, |
| | quantization_config=quantization_config, |
| | ) |
| | training_args.model_init_kwargs = model_kwargs |
| | tokenizer = AutoTokenizer.from_pretrained( |
| | model_config.model_name_or_path, trust_remote_code=model_config.trust_remote_code, use_fast=True |
| | ) |
| | tokenizer.pad_token = tokenizer.eos_token |
| |
|
| | |
| | |
| | |
| | dataset = load_dataset(script_args.dataset_name) |
| |
|
| | |
| | |
| | |
| | trainer = SFTTrainer( |
| | model=model_config.model_name_or_path, |
| | args=training_args, |
| | train_dataset=dataset[script_args.dataset_train_split], |
| | eval_dataset=dataset[script_args.dataset_test_split] if training_args.eval_strategy != "no" else None, |
| | processing_class=tokenizer, |
| | peft_config=get_peft_config(model_config), |
| | ) |
| |
|
| | trainer.train() |
| |
|
| | |
| | trainer.save_model(training_args.output_dir) |
| | if training_args.push_to_hub: |
| | trainer.push_to_hub(dataset_name=script_args.dataset_name) |
| |
|