| import torch |
|
|
| from pytorch_lightning import Callback, Trainer, LightningModule |
|
|
| import logging |
|
|
| log = logging.getLogger(__name__) |
|
|
|
|
| def l2_promote(): |
| import ctypes |
| _libcudart = ctypes.CDLL('libcudart.so') |
| |
| |
| pValue = ctypes.cast((ctypes.c_int*1)(), ctypes.POINTER(ctypes.c_int)) |
| _libcudart.cudaDeviceSetLimit(ctypes.c_int(0x05), ctypes.c_int(128)) |
| _libcudart.cudaDeviceGetLimit(pValue, ctypes.c_int(0x05)) |
| assert pValue.contents.value == 128 |
|
|
|
|
| def set_affinity(trainer): |
| try: |
| from src.utils.gpu_affinity import set_affinity |
| nproc_per_node = torch.cuda.device_count() |
| affinity = set_affinity(trainer.local_rank, nproc_per_node, 'socket_unique_continuous') |
| log.info(f'{trainer.local_rank}: thread affinity: {affinity}') |
| |
| |
| |
| except: |
| pass |
|
|
|
|
| class GpuAffinity(Callback): |
| """Set GPU affinity and increase the L2 fetch granularity. |
| Adapted from https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/LanguageModeling/Transformer-XL |
| """ |
|
|
| def setup(self, trainer: Trainer, pl_module: LightningModule, stage=None) -> None: |
| set_affinity(trainer) |
|
|