| | |
| | import gc |
| | from functools import partial |
| |
|
| | import torch |
| | from torch.distributed.fsdp import FullyShardedDataParallel as FSDP |
| | from torch.distributed.fsdp import MixedPrecision, ShardingStrategy |
| | from torch.distributed.fsdp.wrap import lambda_auto_wrap_policy |
| | from torch.distributed.utils import _free_storage |
| |
|
| |
|
| | def shard_model( |
| | model, |
| | device_id, |
| | param_dtype=torch.bfloat16, |
| | reduce_dtype=torch.float32, |
| | buffer_dtype=torch.float32, |
| | process_group=None, |
| | sharding_strategy=ShardingStrategy.FULL_SHARD, |
| | sync_module_states=True, |
| | ): |
| | model = FSDP( |
| | module=model, |
| | process_group=process_group, |
| | sharding_strategy=sharding_strategy, |
| | auto_wrap_policy=partial( |
| | lambda_auto_wrap_policy, lambda_fn=lambda m: m in model.blocks), |
| | mixed_precision=MixedPrecision( |
| | param_dtype=param_dtype, |
| | reduce_dtype=reduce_dtype, |
| | buffer_dtype=buffer_dtype), |
| | device_id=device_id, |
| | sync_module_states=sync_module_states) |
| | return model |
| |
|
| |
|
| | def free_model(model): |
| | for m in model.modules(): |
| | if isinstance(m, FSDP): |
| | _free_storage(m._handle.flat_param.data) |
| | del model |
| | gc.collect() |
| | torch.cuda.empty_cache() |
| |
|