04151835
Browse files- diffusion.py +10 -9
diffusion.py
CHANGED
|
@@ -267,9 +267,9 @@ class TrainConfig:
|
|
| 267 |
world_size = 1#torch.cuda.device_count()
|
| 268 |
# repeat = 2
|
| 269 |
|
| 270 |
-
|
| 271 |
-
dim = 3#2
|
| 272 |
-
stride = (2,
|
| 273 |
num_image = 32#0#0#640#320#6400#3000#480#1200#120#3000#300#3000#6000#30#60#6000#1000#2000#20000#15000#7000#25600#3000#10000#1000#10000#5000#2560#800#2560
|
| 274 |
batch_size = 1#1#10#50#10#50#20#50#1#2#50#20#2#100 # 10
|
| 275 |
n_epoch = 30#50#20#1#50#10#1#50#1#50#5#50#5#50#100#50#100#30#120#5#4# 10#50#20#20#2#5#25 # 120
|
|
@@ -504,13 +504,14 @@ class DDPM21CM:
|
|
| 504 |
|
| 505 |
|
| 506 |
# print("!!!!!!!!!!!!!!!!, before prepare, self.dataloader.sampler =", self.dataloader.sampler)
|
| 507 |
-
|
| 508 |
-
print(f"model: {self.nn_model.device}")
|
| 509 |
#print(f"optimizer: {self.optimizer.state_dict()}")
|
| 510 |
-
|
| 511 |
-
print(f"
|
| 512 |
-
|
| 513 |
-
print(f"cuda:{torch.cuda.current_device()}/{self.config.global_rank}
|
|
|
|
| 514 |
|
| 515 |
acc_prep_start = time()
|
| 516 |
self.nn_model, self.optimizer, self.dataloader, self.lr_scheduler = \
|
|
|
|
| 267 |
world_size = 1#torch.cuda.device_count()
|
| 268 |
# repeat = 2
|
| 269 |
|
| 270 |
+
dim = 2
|
| 271 |
+
#dim = 3#2
|
| 272 |
+
stride = (2,2) if dim == 2 else (2,2,2)
|
| 273 |
num_image = 32#0#0#640#320#6400#3000#480#1200#120#3000#300#3000#6000#30#60#6000#1000#2000#20000#15000#7000#25600#3000#10000#1000#10000#5000#2560#800#2560
|
| 274 |
batch_size = 1#1#10#50#10#50#20#50#1#2#50#20#2#100 # 10
|
| 275 |
n_epoch = 30#50#20#1#50#10#1#50#1#50#5#50#5#50#100#50#100#30#120#5#4# 10#50#20#20#2#5#25 # 120
|
|
|
|
| 504 |
|
| 505 |
|
| 506 |
# print("!!!!!!!!!!!!!!!!, before prepare, self.dataloader.sampler =", self.dataloader.sampler)
|
| 507 |
+
#model_start = time()
|
| 508 |
+
#print(f"cuda:{torch.cuda.current_device()}/{self.config.global_rank} model: {self.nn_model.device}", f"{time()-model_start:.3f}s")
|
| 509 |
#print(f"optimizer: {self.optimizer.state_dict()}")
|
| 510 |
+
#dataloader_start = time()
|
| 511 |
+
#print(f"cuda:{torch.cuda.current_device()}/{self.config.global_rank} dataloader: {next(iter(self.dataloader))[0].device}", f"{time()-dataloader_start:.3f}s")
|
| 512 |
+
#lr_start = time()
|
| 513 |
+
#print(f"cuda:{torch.cuda.current_device()}/{self.config.global_rank} lr_scheduler: {self.lr_scheduler.optimizer is self.optimizer}", f"{time()-lr_start:.3f}s")
|
| 514 |
+
#print(f"cuda:{torch.cuda.current_device()}/{self.config.global_rank} print costs {print_end-print_start:.3f}s")
|
| 515 |
|
| 516 |
acc_prep_start = time()
|
| 517 |
self.nn_model, self.optimizer, self.dataloader, self.lr_scheduler = \
|