File size: 9,376 Bytes
920fd91 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 | import os, torch, re, time, shutil, glob
from tqdm import tqdm
from accelerate import Accelerator
from .training_module import DiffusionTrainingModule
from .logger import ModelLogger
import wandb
def launch_training_task(
accelerator: Accelerator,
dataset: torch.utils.data.Dataset,
model: DiffusionTrainingModule,
model_logger: ModelLogger,
learning_rate: float = 1e-5,
weight_decay: float = 1e-2,
num_workers: int = 1,
save_steps: int = None,
num_epochs: int = 1,
args = None,
):
if args is not None:
learning_rate = args.learning_rate
weight_decay = args.weight_decay
num_workers = args.dataset_num_workers
save_steps = args.save_steps
num_epochs = args.num_epochs
wandb_project = getattr(args, "wandb_project", "diffusion_training")
wandb_run_name = getattr(args, "wandb_run_name", None)
max_checkpoints = getattr(args, "max_checkpoints", None)
reset_training_progress = getattr(args, "reset_training_progress", False) if args is not None else False
optimizer = torch.optim.AdamW(model.trainable_modules(), lr=learning_rate, weight_decay=weight_decay)
scheduler = torch.optim.lr_scheduler.ConstantLR(optimizer)
dataloader = torch.utils.data.DataLoader(dataset, shuffle=True, collate_fn=lambda x: x[0], num_workers=num_workers,prefetch_factor=4,pin_memory=True)
model, optimizer, dataloader, scheduler = accelerator.prepare(model, optimizer, dataloader, scheduler)
global_step = 0
starting_epoch = 0
resume_step = 0
resume_from_checkpoint = getattr(args, "resume_from_checkpoint", None)
if resume_from_checkpoint:
accelerator.print(f"Resuming training from checkpoint: {resume_from_checkpoint}")
accelerator.load_state(resume_from_checkpoint)
if reset_training_progress:
# 【新增逻辑】如果是新阶段训练:保留 load_state 恢复的模型和优化器,但跳过步数解析
accelerator.print("Config: reset_training_progress=True. Starting a new training phase (Step 0, Epoch 0).")
# # 【非常重要】load_state 会把优化器的学习率覆盖为旧的,这里强制重新赋值为你传入的新学习率
# for param_group in optimizer.param_groups:
# param_group['lr'] = learning_rate
else:
# 【保留原有逻辑】传统的断点续训:尝试从路径中解析出原本的 global_step
match = re.search(r"checkpoint-(\d+)", resume_from_checkpoint)
if match:
global_step = int(match.group(1))
steps_per_epoch = len(dataloader)
starting_epoch = global_step // steps_per_epoch
resume_step = global_step % steps_per_epoch
model_logger.num_steps = global_step
accelerator.print(f"Resuming at epoch {starting_epoch}, starting at step {resume_step} in the current epoch.")
else:
accelerator.print("Warning: Could not infer global_step from checkpoint folder name. Make sure it contains 'checkpoint-<step_number>'.")
if accelerator.is_main_process:
wandb.init(project=wandb_project, name=wandb_run_name,mode="offline")
total_steps=num_epochs*len(dataloader)
progress_bar = tqdm(
initial=global_step,
total=total_steps,
desc="Steps",
disable=not accelerator.is_main_process
)
for epoch_id in range(starting_epoch,num_epochs):
print(epoch_id,"epoch begin")
if resume_from_checkpoint and epoch_id == starting_epoch and resume_step > 0:
active_dataloader = accelerator.skip_first_batches(dataloader, resume_step)
else:
active_dataloader = dataloader
for data in active_dataloader:
with accelerator.accumulate(model):
optimizer.zero_grad()
# if dataset.load_from_cache:
# loss = model({}, inputs=data)
# else:
loss = model(data)
accelerator.backward(loss)
optimizer.step()
model_logger.on_step_end(accelerator, model, save_steps)
scheduler.step()
global_step += 1
# print(global_step,"step complete",save_steps)
if save_steps is not None and global_step % save_steps == 0:
if accelerator.is_main_process:
print(f"Saving checkpoint at step {global_step}...")
save_path = os.path.join(model_logger.output_path, f"checkpoint-{global_step}")
start_time = time.time()
accelerator.save_state(save_path)
end_time = time.time()
elapsed_time = end_time - start_time
print(f"accelerator.save_state 耗时: {elapsed_time:.4f} 秒")
if max_checkpoints is not None and max_checkpoints > 0:
# 确保所有进程都保存完毕后再执行清理,防止死锁或文件损坏
accelerator.wait_for_everyone()
if accelerator.is_main_process:
# 使用正则精确匹配 step 维度的文件夹 (规避掉 epoch 的文件夹)
all_ckpts = glob.glob(os.path.join(model_logger.output_path, "checkpoint-[0-9]*"))
all_ckpts = [ckpt for ckpt in all_ckpts if int(re.search(r'checkpoint-(\d+)$', ckpt).group(1)) <= global_step]
# 按后缀数字升序排序 (最老的排在最前面)
all_ckpts.sort(key=lambda x: int(re.search(r'checkpoint-(\d+)$', x).group(1)))
while len(all_ckpts) > max_checkpoints:
oldest_ckpt = all_ckpts.pop(0)
# 删除巨大的断点状态文件夹
try:
if os.path.exists(oldest_ckpt):
shutil.rmtree(oldest_ckpt)
print(f"已清理旧断点文件夹: {oldest_ckpt}")
except Exception as e:
print(f"清理断点失败 {oldest_ckpt}: {e}")
if accelerator.is_main_process:
wandb.log({
"loss": loss.item(),
"epoch": epoch_id,
"global_step": global_step
})
progress_bar.set_postfix(loss=f"{loss.item():.4f}")
progress_bar.update(1)
if save_steps is None:
model_logger.on_epoch_end(accelerator, model, epoch_id)
save_path = os.path.join(model_logger.output_path, f"checkpoint-epoch-{epoch_id}")
accelerator.save_state(save_path)
if max_checkpoints is not None and max_checkpoints > 0:
accelerator.wait_for_everyone()
if accelerator.is_main_process:
all_epoch_ckpts = glob.glob(os.path.join(model_logger.output_path, "checkpoint-epoch-[0-9]*"))
all_epoch_ckpts = [ckpt for ckpt in all_epoch_ckpts if int(re.search(r'checkpoint-epoch-(\d+)$', ckpt).group(1)) <= epoch_id]
all_epoch_ckpts.sort(key=lambda x: int(re.search(r'checkpoint-epoch-(\d+)$', x).group(1)))
while len(all_epoch_ckpts) > max_checkpoints:
oldest_ckpt = all_epoch_ckpts.pop(0)
try:
if os.path.exists(oldest_ckpt):
shutil.rmtree(oldest_ckpt)
print(f"已清理旧 Epoch 断点: {oldest_ckpt}")
except Exception as e:
pass
if accelerator.is_main_process:
print(f"epoch {epoch_id+1}: finished")
progress_bar.close()
model_logger.on_training_end(accelerator, model, save_steps)
def launch_data_process_task(
accelerator: Accelerator,
dataset: torch.utils.data.Dataset,
model: DiffusionTrainingModule,
model_logger: ModelLogger,
num_workers: int = 8,
args = None,
):
if args is not None:
num_workers = args.dataset_num_workers
dataloader = torch.utils.data.DataLoader(dataset, shuffle=False, collate_fn=lambda x: x[0], num_workers=num_workers)
model, dataloader = accelerator.prepare(model, dataloader)
for data_id, data in enumerate(tqdm(dataloader)):
with accelerator.accumulate(model):
with torch.no_grad():
folder = os.path.join(model_logger.output_path, str(accelerator.process_index))
os.makedirs(folder, exist_ok=True)
save_path = os.path.join(model_logger.output_path, str(accelerator.process_index), f"{data_id}.pth")
data = model(data)
torch.save(data, save_path)
|