Upload main.py with huggingface_hub
Browse files
main.py
CHANGED
|
@@ -270,11 +270,11 @@ if global_rank == 0:
|
|
| 270 |
os.makedirs(args.savedir, exist_ok=True)
|
| 271 |
|
| 272 |
sampler = DistributedSampler(testdataset, num_replicas=world_size, rank=global_rank, shuffle=False)
|
| 273 |
-
test_loader = DataLoader(testdataset, batch_size=train_config["bs"], sampler=sampler, num_workers=
|
| 274 |
collate_fn=DataCollatorWithPadding())
|
| 275 |
|
| 276 |
train_sampler = DistributedSampler(traindataset, num_replicas=world_size, rank=global_rank, shuffle=True)
|
| 277 |
-
train_loader = DataLoader(traindataset, batch_size=train_config["bs"], sampler=train_sampler, num_workers=
|
| 278 |
pin_memory=True,
|
| 279 |
collate_fn=DataCollatorWithPadding())
|
| 280 |
|
|
|
|
| 270 |
os.makedirs(args.savedir, exist_ok=True)
|
| 271 |
|
| 272 |
sampler = DistributedSampler(testdataset, num_replicas=world_size, rank=global_rank, shuffle=False)
|
| 273 |
+
test_loader = DataLoader(testdataset, batch_size=train_config["bs"], sampler=sampler, num_workers=0, pin_memory=True,
|
| 274 |
collate_fn=DataCollatorWithPadding())
|
| 275 |
|
| 276 |
train_sampler = DistributedSampler(traindataset, num_replicas=world_size, rank=global_rank, shuffle=True)
|
| 277 |
+
train_loader = DataLoader(traindataset, batch_size=train_config["bs"], sampler=train_sampler, num_workers=0,
|
| 278 |
pin_memory=True,
|
| 279 |
collate_fn=DataCollatorWithPadding())
|
| 280 |
|