| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import json |
| | from dataclasses import dataclass, field |
| |
|
| | from transformers import Seq2SeqTrainingArguments |
| | from transformers.training_args import _convert_str_dict |
| |
|
| | from ..extras.misc import is_env_enabled, use_ray |
| | from ..extras.packages import is_mcore_adapter_available |
| |
|
| |
|
| | if is_env_enabled("USE_MCA"): |
| | if not is_mcore_adapter_available(): |
| | raise ImportError( |
| | "mcore_adapter is required when USE_MCA=1. Please install `mcore_adapter` and its dependencies." |
| | ) |
| |
|
| | from mcore_adapter import Seq2SeqTrainingArguments as McaSeq2SeqTrainingArguments |
| |
|
| | BaseTrainingArguments = McaSeq2SeqTrainingArguments |
| | else: |
| | BaseTrainingArguments = Seq2SeqTrainingArguments |
| |
|
| |
|
| | @dataclass |
| | class RayArguments: |
| | r"""Arguments pertaining to the Ray training.""" |
| |
|
| | ray_num_workers: int = field( |
| | default=1, |
| | metadata={"help": "The number of workers for Ray training. Default is 1 worker."}, |
| | ) |
| | ray_init_kwargs: dict | str | None = field( |
| | default=None, |
| | metadata={"help": "The arguments to pass to ray.init for Ray training. Default is None."}, |
| | ) |
| | master_addr: str | None = field( |
| | default=None, |
| | metadata={"help": "The master address for init_process_group"}, |
| | ) |
| | master_port: str | None = field( |
| | default=None, |
| | metadata={"help": "The master port for init_process_group"}, |
| | ) |
| |
|
| | def __post_init__(self): |
| | self.use_ray = use_ray() |
| |
|
| | if isinstance(self.ray_init_kwargs, str) and self.ray_init_kwargs.startswith("{"): |
| | self.ray_init_kwargs = _convert_str_dict(json.loads(self.ray_init_kwargs)) |
| |
|
| |
|
| | @dataclass |
| | class Fp8Arguments: |
| | r"""Arguments pertaining to the FP8 training.""" |
| |
|
| | fp8: bool = field( |
| | default=False, |
| | metadata={ |
| | "help": "Enable FP8 mixed precision training via HuggingFace Accelerate. " |
| | "Requires PyTorch 2.7+ and Hopper architecture GPUs." |
| | }, |
| | ) |
| | fp8_backend: str = field( |
| | default="auto", |
| | metadata={ |
| | "help": "FP8 backend to use ('auto', 'torchao', 'te', 'msamp'). 'auto' selects best available backend." |
| | }, |
| | ) |
| | fp8_enable_fsdp_float8_all_gather: bool = field( |
| | default=False, |
| | metadata={"help": "Enable FP8 optimizations for FSDP2 all-gather operations."}, |
| | ) |
| |
|
| |
|
| | @dataclass |
| | class TrainingArguments(Fp8Arguments, RayArguments, BaseTrainingArguments): |
| | r"""Arguments pertaining to the trainer.""" |
| |
|
| | overwrite_output_dir: bool = field( |
| | default=False, |
| | metadata={"help": "deprecated"}, |
| | ) |
| |
|
| | def __post_init__(self): |
| | RayArguments.__post_init__(self) |
| | BaseTrainingArguments.__post_init__(self) |
| |
|