litwell commited on
Commit
fa5b6e0
·
verified ·
1 Parent(s): 4a0ee93

Upload models/src/training/params.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. models/src/training/params.py +71 -0
models/src/training/params.py ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from dataclasses import dataclass, field
2
+ from typing import Optional
3
+
4
+ from transformers import TrainingArguments
5
+
6
+
7
+ @dataclass
8
+ class ModelArguments:
9
+ model_id: Optional[str] = field(default="Qwen/Qwen2-VL-7B-Instruct")
10
+
11
+
12
+ @dataclass
13
+ class TrainingArguments(TrainingArguments):
14
+ cache_dir: Optional[str] = field(default=None)
15
+ optim: str = field(default="adamw_torch")
16
+ adam_beta1: float = field(default=0.9)
17
+ adam_beta2: float = field(default=0.999)
18
+ adam_epsilon: float = field(default=1e-8)
19
+
20
+ freeze_vision_tower: bool = field(default=False)
21
+ freeze_llm: bool = field(default=False)
22
+ tune_merger: bool = field(default=False)
23
+ disable_flash_attn2: bool = field(default=False)
24
+
25
+ max_seq_length: int = field(
26
+ default=32768, # This is the default value of the qwen2-vl model
27
+ metadata={
28
+ "help":
29
+ "Maximum sequence length. Sequences will be right padded (and possibly truncated)."
30
+ },
31
+ )
32
+
33
+ double_quant: bool = field(
34
+ default=True,
35
+ metadata={"help": "Compress the quantization statistics through double quantization."}
36
+ )
37
+ quant_type: str = field(
38
+ default="nf4",
39
+ metadata={"help": "Quantization data type to use. Should be one of `fp4` or `nf4`."}
40
+ )
41
+ bits: int = field(
42
+ default=16,
43
+ metadata={"help": "How many bits to use."}
44
+ )
45
+ lora_enable: bool = False
46
+ vision_lora: bool = False
47
+ use_dora: bool = False
48
+ lora_rank: int = 64
49
+ lora_alpha: int = 16
50
+ lora_dropout: float = 0.05
51
+ lora_weight_path: str = ""
52
+ lora_bias: str = "none"
53
+ vision_lr: Optional[float] = None
54
+ merger_lr: Optional[float] = None
55
+ lora_namespan_exclude: str = field(default=None, metadata={"help": "List of namespan to exclude for LoRA"})
56
+ num_lora_modules: int = -1
57
+ use_liger: bool = True
58
+
59
+
60
+ @dataclass
61
+ class DataArguments:
62
+ data_path: str = field(
63
+ default=None, metadata={"help": "Path to the training data."}
64
+ )
65
+ lazy_preprocess: bool = False
66
+ image_folder: Optional[str] = field(default=None)
67
+ image_min_pixels: Optional[int] = field(default=3136)
68
+ image_max_pixels: Optional[int] = field(default=12845056)
69
+ video_min_pixels: Optional[int] = field(default=100352)
70
+ video_max_pixels: Optional[int] = field(default=602112)
71
+ fps: float = 1.0