finetune-model / src /params.py
reiofa's picture
Upload folder using huggingface_hub
89b38b0 verified
from dataclasses import dataclass, field
from typing import Optional
try:
from accelerate.utils import ParallelismConfig as _PC
except Exception:
class _PC:
pass
import transformers.training_args as _ta
if not hasattr(_ta, "ParallelismConfig"):
_ta.ParallelismConfig = _PC
from transformers import TrainingArguments as HFTrainingArguments
from trl import DPOConfig as DPOConfigTRL
from trl import GRPOConfig as GRPOConfigTRL
@dataclass
class ModelArguments:
model_id: Optional[str] = field(default="Qwen/Qwen2-VL-7B-Instruct")
@dataclass
class CLSArguments(HFTrainingArguments):
cache_dir: Optional[str] = field(default=None)
optim: str = field(default="adamw_torch")
adam_beta1: float = field(default=0.9)
adam_beta2: float = field(default=0.999)
adam_epsilon: float = field(default=1e-8)
freeze_vision_tower: bool = field(default=False)
freeze_llm: bool = field(default=False)
freeze_merger: bool = field(default=False)
disable_flash_attn2: bool = field(default=False)
unfreeze_topk_llm: int = 0
unfreeze_topk_vision: int = 0
mlp_head_dim: Optional[int] = field(default=0)
mlp_head_dropout: Optional[float] = field(default=0.0)
loss_type : str = field(
default="cross_entropy",
metadata={"help": "Loss type to use. Should be one of `cross_entropy`, `focal_loss`, `class_balanced_cross_entropy`, or `class_balanced_focal_loss`."}
)
focal_alpha: Optional[str] = field(
default=None,
metadata={"help": "Focal Loss alpha value. If None use CrossEntropyLoss. ex '1.0,7.5'"}
)
focal_gamma: float = field(
default=0.0,
metadata={"help": "Focal Loss gamma value"}
)
num_labels: int = field(
default=2,
metadata={"help": "Number of labels for classification."}
)
class_balanced_beta: float = field(
default=0.999,
metadata={"help": "Beta value for Class Balanced Loss. If 0.0, use standard CrossEntropyLoss."}
)
early_stopping_patience: int = field(
default=0,
metadata={"help": "Number of epochs with no improvement after which training will be stopped."}
)
early_stopping_threshold: float = field(
default=0.0,
metadata={"help": "Minimum change in the monitored quantity to qualify as an improvement."}
)
max_seq_length: int = field(
default=32768, # This is the default value of the qwen2-vl model
metadata={
"help":
"Maximum sequence length. Sequences will be right padded (and possibly truncated)."
},
)
double_quant: bool = field(
default=True,
metadata={"help": "Compress the quantization statistics through double quantization."}
)
quant_type: str = field(
default="nf4",
metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."}
)
bits: int = field(
default=16,
metadata={"help": "How many bits to use."}
)
lora_enable: bool = False
vision_lora: bool = False
use_dora: bool = False
lora_rank: int = 64
lora_alpha: int = 16
lora_dropout: float = 0.05
lora_weight_path: str = ""
lora_bias: str = "none"
vision_lr: Optional[float] = None
merger_lr: Optional[float] = None
head_lr: Optional[float] = None
lora_namespan_exclude: str = field(default=None, metadata={"help": "List of namespan to exclude for LoRA"})
num_lora_modules: int = -1
use_liger_kernel: bool = True
@dataclass
class TrainingArguments(HFTrainingArguments):
cache_dir: Optional[str] = field(default=None)
optim: str = field(default="adamw_torch")
adam_beta1: float = field(default=0.9)
adam_beta2: float = field(default=0.999)
adam_epsilon: float = field(default=1e-8)
freeze_vision_tower: bool = field(default=False)
freeze_llm: bool = field(default=False)
freeze_merger: bool = field(default=False)
disable_flash_attn2: bool = field(default=False)
unfreeze_topk_llm: int = 0
unfreeze_topk_vision: int = 0
max_seq_length: int = field(
default=32768, # This is the default value of the qwen2-vl model
metadata={
"help":
"Maximum sequence length. Sequences will be right padded (and possibly truncated)."
},
)
double_quant: bool = field(
default=True,
metadata={"help": "Compress the quantization statistics through double quantization."}
)
quant_type: str = field(
default="nf4",
metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."}
)
bits: int = field(
default=16,
metadata={"help": "How many bits to use."}
)
lora_enable: bool = False
vision_lora: bool = False
use_dora: bool = False
lora_rank: int = 64
lora_alpha: int = 16
lora_dropout: float = 0.05
lora_weight_path: str = ""
lora_bias: str = "none"
vision_lr: Optional[float] = None
merger_lr: Optional[float] = None
lora_namespan_exclude: str = field(default=None, metadata={"help": "List of namespan to exclude for LoRA"})
num_lora_modules: int = -1
use_liger_kernel: bool = True
# Generation-based evaluation settings
generation_max_new_tokens: int = field(
default=512,
metadata={"help": "Maximum number of new tokens to generate during evaluation."}
)
@dataclass
class DPOArguments(DPOConfigTRL):
cache_dir: Optional[str] = field(default=None)
optim: str = field(default="adamw_torch")
adam_beta1: float = field(default=0.9)
adam_beta2: float = field(default=0.999)
adam_epsilon: float = field(default=1e-8)
freeze_vision_tower: bool = field(default=False)
freeze_llm: bool = field(default=False)
freeze_merger: bool = field(default=False)
disable_flash_attn2: bool = field(default=False)
unfreeze_topk_llm: int = 0
unfreeze_topk_vision: int = 0
max_seq_length: int = field(
default=32768, # This is the default value of the qwen2-vl model
metadata={
"help":
"Maximum sequence length. Sequences will be right padded (and possibly truncated)."
},
)
double_quant: bool = field(
default=True,
metadata={"help": "Compress the quantization statistics through double quantization."}
)
quant_type: str = field(
default="nf4",
metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."}
)
bits: int = field(
default=16,
metadata={"help": "How many bits to use."}
)
lora_enable: bool = False
vision_lora: bool = False
use_dora: bool = False
lora_rank: int = 64
lora_alpha: int = 16
lora_dropout: float = 0.05
lora_weight_path: str = ""
lora_bias: str = "none"
vision_lr: Optional[float] = None
merger_lr: Optional[float] = None
lora_namespan_exclude: str = field(default=None, metadata={"help": "List of namespan to exclude for LoRA"})
num_lora_modules: int = -1
use_liger_loss: bool = True
beta: float = field(
default=0.1,
metadata={"help": "The beta value for DPO."}
)
precompute_ref_log_probs: bool = field(
default=False,
metadata={"help": "Whether to precompute the reference log probabilities."}
)
dpo_loss:str = field(
default="sigmoid",
metadata={"help": "The type of DPO loss to use."}
)
@dataclass
class GRPOArguments(GRPOConfigTRL):
cache_dir: Optional[str] = field(default=None)
optim: str = field(default="adamw_torch")
adam_beta1: float = field(default=0.9)
adam_beta2: float = field(default=0.999)
adam_epsilon: float = field(default=1e-8)
freeze_vision_tower: bool = field(default=False)
freeze_llm: bool = field(default=False)
freeze_merger: bool = field(default=False)
disable_flash_attn2: bool = field(default=False)
unfreeze_topk_llm: int = 0
unfreeze_topk_vision: int = 0
double_quant: bool = field(
default=True,
metadata={"help": "Compress the quantization statistics through double quantization."}
)
quant_type: str = field(
default="nf4",
metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."}
)
bits: int = field(
default=16,
metadata={"help": "How many bits to use."}
)
lora_enable: bool = False
vision_lora: bool = False
use_dora: bool = False
lora_rank: int = 64
lora_alpha: int = 16
lora_dropout: float = 0.05
lora_weight_path: str = ""
lora_bias: str = "none"
vision_lr: Optional[float] = None
merger_lr: Optional[float] = None
lora_namespan_exclude: str = field(default=None, metadata={"help": "List of namespan to exclude for LoRA"})
num_lora_modules: int = -1
beta: float = field(
default=0.04,
metadata={
"help": "KL coefficient. If `0.0`, the reference model is not loaded, reducing memory usage and improving "
"training speed, but may be numerically unstable for long training runs."
},
)
temperature: float = 0.9
top_p: float = 1.0
top_k: int = 50
min_p: Optional[float] = None
repetition_penalty: float = 1.0
max_completion_length: int = 256
max_prompt_length: int = 512
use_liger_loss: bool = True
@dataclass
class DataArguments:
data_path: str = field(
default=None, metadata={"help": "Path to the training data."}
)
eval_path: str= field(
default=None, metadata={"help": "Path to the evaluation data."}
)
eval_image_folder: Optional[str] = field(
default=None, metadata={"help": "Path to the evaluation image data."}
)
lazy_preprocess: bool = False
image_folder: Optional[str] = field(default=None)
image_min_pixels: Optional[int] = field(default=3136)
image_max_pixels: Optional[int] = field(default=12845056)
video_min_pixels: Optional[int] = field(default=100352)
video_max_pixels: Optional[int] = field(default=602112)
image_resized_width: int = field(default=None)
image_resized_height: int = field(default=None)
video_resized_width: int = field(default=None)
video_resized_height: int = field(default=None)
fps: Optional[int] = field(default=None, metadata={"help": "Frames per second for video data."})
nframes: Optional[int] = field(default=None, metadata={"help": "Number of frames for video data."})