allow report_to for multiple providers (#1647)
Browse files
src/axolotl/core/trainer_builder.py
CHANGED
|
@@ -1212,11 +1212,14 @@ class HFCausalTrainerBuilder(TrainerBuilderBase):
|
|
| 1212 |
)
|
| 1213 |
training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
|
| 1214 |
training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
|
| 1215 |
-
report_to =
|
| 1216 |
if self.cfg.use_wandb:
|
| 1217 |
-
report_to
|
| 1218 |
if self.cfg.use_mlflow:
|
| 1219 |
-
report_to
|
|
|
|
|
|
|
|
|
|
| 1220 |
training_arguments_kwargs["report_to"] = report_to
|
| 1221 |
training_arguments_kwargs["run_name"] = (
|
| 1222 |
self.cfg.wandb_name if self.cfg.use_wandb else None
|
|
|
|
| 1212 |
)
|
| 1213 |
training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
|
| 1214 |
training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
|
| 1215 |
+
report_to = []
|
| 1216 |
if self.cfg.use_wandb:
|
| 1217 |
+
report_to.append("wandb")
|
| 1218 |
if self.cfg.use_mlflow:
|
| 1219 |
+
report_to.append("mlflow")
|
| 1220 |
+
if self.cfg.use_tensorboard:
|
| 1221 |
+
report_to.append("tensorboard")
|
| 1222 |
+
|
| 1223 |
training_arguments_kwargs["report_to"] = report_to
|
| 1224 |
training_arguments_kwargs["run_name"] = (
|
| 1225 |
self.cfg.wandb_name if self.cfg.use_wandb else None
|
src/axolotl/utils/config/models/input/v0_4_1/__init__.py
CHANGED
|
@@ -212,7 +212,7 @@ class LoraConfig(BaseModel):
|
|
| 212 |
lora_target_modules: Optional[List[str]] = None
|
| 213 |
lora_target_linear: Optional[bool] = None
|
| 214 |
lora_modules_to_save: Optional[List[str]] = None
|
| 215 |
-
lora_dropout: Optional[float] =
|
| 216 |
peft_layers_to_transform: Optional[List[int]] = None
|
| 217 |
peft: Optional[PeftConfig] = None
|
| 218 |
peft_use_dora: Optional[bool] = None
|
|
@@ -609,6 +609,7 @@ class AxolotlInputConfig(
|
|
| 609 |
early_stopping_patience: Optional[int] = None
|
| 610 |
load_best_model_at_end: Optional[bool] = False
|
| 611 |
save_only_model: Optional[bool] = False
|
|
|
|
| 612 |
|
| 613 |
neftune_noise_alpha: Optional[float] = None
|
| 614 |
|
|
|
|
| 212 |
lora_target_modules: Optional[List[str]] = None
|
| 213 |
lora_target_linear: Optional[bool] = None
|
| 214 |
lora_modules_to_save: Optional[List[str]] = None
|
| 215 |
+
lora_dropout: Optional[float] = 0.0
|
| 216 |
peft_layers_to_transform: Optional[List[int]] = None
|
| 217 |
peft: Optional[PeftConfig] = None
|
| 218 |
peft_use_dora: Optional[bool] = None
|
|
|
|
| 609 |
early_stopping_patience: Optional[int] = None
|
| 610 |
load_best_model_at_end: Optional[bool] = False
|
| 611 |
save_only_model: Optional[bool] = False
|
| 612 |
+
use_tensorboard: Optional[bool] = None
|
| 613 |
|
| 614 |
neftune_noise_alpha: Optional[float] = None
|
| 615 |
|