| import sys, os |
| root = os.sep + os.sep.join(__file__.split(os.sep)[1:__file__.split(os.sep).index("Recurrent-Parameter-Generation")+1]) |
| sys.path.append(root) |
| os.chdir(root) |
|
|
| |
| import time |
| import types |
| import torch |
| import copy |
| from torch import nn |
| from model.diffusion import DDIMSampler, DDPMSampler |
| |
| import importlib |
| item = importlib.import_module(f"{sys.argv[1]}") |
| Dataset = item.Dataset |
| train_set = item.train_set |
| config = item.config |
| model = item.model |
| assert config.get("tag") is not None, "Remember to set a tag." |
|
|
|
|
|
|
|
|
| generate_config = { |
| "device": "cuda", |
| "num_generated": 1, |
| "checkpoint": f"./checkpoint/{config['tag']}.pth", |
| "generated_path": os.path.join(Dataset.generated_path.rsplit("/", 1)[0], "generated_{}_{}.pth"), |
| "test_command": os.path.join(Dataset.test_command.rsplit("/", 1)[0], "generated_{}_{}.pth"), |
| "need_test": True, |
| |
| "sampler": DDIMSampler, |
| "steps": 60, |
| } |
| config.update(generate_config) |
|
|
|
|
|
|
|
|
| |
| print('==> Building model..') |
| diction = torch.load(config["checkpoint"]) |
| permutation_shape = diction["to_permutation_state.weight"].shape |
| model.to_permutation_state = nn.Embedding(*permutation_shape) |
| model.load_state_dict(diction) |
| model.criteria.diffusion_sampler = config["sampler"]( |
| model=model.criteria.diffusion_sampler.model, |
| beta=config["model_config"]["beta"], |
| T=config["model_config"]["T"], |
| ) |
| model.condi_embedder = copy.deepcopy(model.criteria.diffusion_sampler.model.condi_embedder) |
| @torch.no_grad() |
| def new_sample(self, x=None, condition=None): |
| z = self.model([1, self.sequence_length, self.config["d_model"]], condition) |
| z = self.condi_embedder(z) |
| if x is None: |
| x = torch.randn((1, self.sequence_length, self.config["model_dim"]), device=z.device) |
| x = self.criteria.sample(x, z, steps=config["steps"]) |
| return x |
| model.sample = types.MethodType(new_sample, model) |
| model.criteria.diffusion_sampler.model.condi_embedder = nn.Identity() |
| model = model.to(config["device"]) |
|
|
|
|
| |
| print('==> Defining generate..') |
| def generate(save_path=config["generated_path"], test_command=config["test_command"], need_test=True): |
| print("\n==> Generating..") |
| model.eval() |
| with torch.cuda.amp.autocast(True, torch.bfloat16): |
| with torch.no_grad(): |
| start_time = time.time() |
| prediction = model(sample=True) |
| end_time = time.time() |
| generated_norm = torch.nanmean(prediction.abs()) |
| print("used time (seconds):", end_time - start_time) |
| print("memory usage (GB):", torch.cuda.max_memory_allocated() / (1024 ** 3)) |
| |
| train_set.save_params(prediction, save_path=save_path) |
| if need_test: |
| os.system(test_command) |
| print("\n") |
|
|
|
|
|
|
|
|
| if __name__ == "__main__": |
| for i in range(config["num_generated"]): |
| index = str(i+1).zfill(3) |
| print("Save to", config["generated_path"].format(config["tag"], index)) |
| generate( |
| save_path=config["generated_path"].format(config["tag"], index), |
| test_command=config["test_command"].format(config["tag"], index), |
| need_test=config["need_test"], |
| ) |