| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import os |
| import unittest |
| from tempfile import TemporaryDirectory |
|
|
| import torch |
| import torch.nn as nn |
|
|
| from accelerate.big_modeling import ( |
| cpu_offload, |
| disk_offload, |
| dispatch_model, |
| init_empty_weights, |
| init_on_device, |
| load_checkpoint_and_dispatch, |
| ) |
| from accelerate.hooks import remove_hook_from_submodules |
| from accelerate.test_utils import require_cuda, require_mps, require_multi_gpu, require_torch_min_version, slow |
| from accelerate.utils import offload_state_dict |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
| class ModelForTest(nn.Module): |
| def __init__(self): |
| super().__init__() |
| self.linear1 = nn.Linear(3, 4) |
| self.batchnorm = nn.BatchNorm1d(4) |
| self.linear2 = nn.Linear(4, 5) |
|
|
| def forward(self, x): |
| return self.linear2(self.batchnorm(self.linear1(x))) |
|
|
|
|
| class ModelForTestTiedWeights(nn.Module): |
| def __init__(self): |
| super().__init__() |
| self.linear1 = nn.Linear(4, 4) |
| self.batchnorm = nn.BatchNorm1d(4) |
| self.linear2 = nn.Linear(4, 4) |
|
|
| def forward(self, x): |
| return self.linear2(self.batchnorm(self.linear1(x))) |
|
|
|
|
| class BiggerModelForTest(nn.Module): |
| def __init__(self): |
| super().__init__() |
| self.linear1 = nn.Linear(3, 4) |
| self.linear2 = nn.Linear(4, 5) |
| self.batchnorm = nn.BatchNorm1d(5) |
| self.linear3 = nn.Linear(5, 6) |
| self.linear4 = nn.Linear(6, 5) |
|
|
| def forward(self, x): |
| return self.linear4(self.linear3(self.batchnorm(self.linear2(self.linear1(x))))) |
|
|
|
|
| |
| class ModuleWithUnusedSubModules(nn.Module): |
| def __init__(self, input_dim, output_dim): |
| super().__init__() |
| self.linear = nn.Linear(input_dim, output_dim) |
|
|
| def forward(self, x): |
| return x @ self.linear.weight.t() + self.linear.bias |
|
|
|
|
| class ModelWithUnusedSubModulesForTest(nn.Module): |
| def __init__(self): |
| super().__init__() |
| self.linear1 = ModuleWithUnusedSubModules(3, 4) |
| self.linear2 = ModuleWithUnusedSubModules(4, 5) |
| self.batchnorm = nn.BatchNorm1d(5) |
| self.linear3 = ModuleWithUnusedSubModules(5, 6) |
| self.linear4 = ModuleWithUnusedSubModules(6, 5) |
|
|
| def forward(self, x): |
| return self.linear4(self.linear3(self.batchnorm(self.linear2(self.linear1(x))))) |
|
|
|
|
| @require_torch_min_version(version="1.9.0") |
| class BigModelingTester(unittest.TestCase): |
| def test_init_empty_weights(self): |
| |
| with init_empty_weights(): |
| module = nn.Linear(4, 5) |
| self.assertEqual(module.weight.device, torch.device("meta")) |
|
|
| |
| with init_empty_weights(): |
| module = nn.BatchNorm1d(4) |
| self.assertEqual(module.weight.device, torch.device("meta")) |
| self.assertEqual(module.running_mean.device, torch.device("cpu")) |
|
|
| |
| with init_empty_weights(include_buffers=True): |
| module = nn.BatchNorm1d(4) |
| self.assertEqual(module.weight.device, torch.device("meta")) |
| self.assertEqual(module.running_mean.device, torch.device("meta")) |
|
|
| |
| module = nn.BatchNorm1d(4) |
| self.assertEqual(module.weight.device, torch.device("cpu")) |
| self.assertEqual(module.running_mean.device, torch.device("cpu")) |
|
|
| def test_init_empty_weights_very_large_model(self): |
| |
| with init_empty_weights(): |
| _ = nn.Sequential(*[nn.Linear(10000, 10000) for _ in range(1000)]) |
|
|
| @require_cuda |
| def test_init_on_device_cuda(self): |
| device = torch.device("cuda:0") |
| with init_on_device(device): |
| model = nn.Linear(10, 10) |
| self.assertEqual(model.weight.device, device) |
| self.assertEqual(model.weight.device, device) |
|
|
| @require_mps |
| def test_init_on_device_mps(self): |
| device = torch.device("mps:0") |
| with init_on_device(device): |
| model = nn.Linear(10, 10) |
| self.assertEqual(model.weight.device, device) |
| self.assertEqual(model.weight.device, device) |
|
|
| def test_cpu_offload(self): |
| model = ModelForTest() |
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| device = torch.device(0 if torch.cuda.is_available() else "cpu") |
|
|
| cpu_offload(model, execution_device=device) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| |
| remove_hook_from_submodules(model) |
|
|
| cpu_offload(model, execution_device=device, offload_buffers=True) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| def test_cpu_offload_with_unused_submodules(self): |
| model = ModelWithUnusedSubModulesForTest() |
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| device = torch.device(0 if torch.cuda.is_available() else "cpu") |
|
|
| cpu_offload(model, execution_device=device, preload_module_classes=["ModuleWithUnusedSubModules"]) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| |
| remove_hook_from_submodules(model) |
|
|
| cpu_offload( |
| model, |
| execution_device=device, |
| offload_buffers=True, |
| preload_module_classes=["ModuleWithUnusedSubModules"], |
| ) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| @slow |
| @require_cuda |
| def test_cpu_offload_gpt2(self): |
| tokenizer = AutoTokenizer.from_pretrained("gpt2") |
| inputs = tokenizer("Hello world! My name is", return_tensors="pt").to(0) |
|
|
| gpt2 = AutoModelForCausalLM.from_pretrained("gpt2") |
| cpu_offload(gpt2, execution_device=0) |
| outputs = gpt2.generate(inputs["input_ids"]) |
| self.assertEqual( |
| tokenizer.decode(outputs[0].tolist()), |
| "Hello world! My name is Kiyoshi, and I'm a student at the University of Tokyo", |
| ) |
|
|
| def test_disk_offload(self): |
| model = ModelForTest() |
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| device = torch.device(0 if torch.cuda.is_available() else "cpu") |
|
|
| with TemporaryDirectory() as tmp_dir: |
| disk_offload(model, tmp_dir, execution_device=device) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| |
| remove_hook_from_submodules(model) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| disk_offload(model, tmp_dir, execution_device=device, offload_buffers=True) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| def test_disk_offload_with_unused_submodules(self): |
| model = ModelWithUnusedSubModulesForTest() |
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| device = torch.device(0 if torch.cuda.is_available() else "cpu") |
|
|
| with TemporaryDirectory() as tmp_dir: |
| disk_offload( |
| model, tmp_dir, execution_device=device, preload_module_classes=["ModuleWithUnusedSubModules"] |
| ) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| |
| remove_hook_from_submodules(model) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| disk_offload( |
| model, |
| tmp_dir, |
| execution_device=device, |
| offload_buffers=True, |
| preload_module_classes=["ModuleWithUnusedSubModules"], |
| ) |
| output = model(x) |
| self.assertTrue( |
| torch.allclose(expected, output.cpu(), 1e-4, 1e-5), msg=f"Expected: {expected}\nActual: {output.cpu()}" |
| ) |
|
|
| @slow |
| @require_cuda |
| def test_disk_offload_gpt2(self): |
| tokenizer = AutoTokenizer.from_pretrained("gpt2") |
| inputs = tokenizer("Hello world! My name is", return_tensors="pt").to(0) |
|
|
| gpt2 = AutoModelForCausalLM.from_pretrained("gpt2") |
| with TemporaryDirectory() as tmp_dir: |
| disk_offload(gpt2, tmp_dir, execution_device=0) |
| outputs = gpt2.generate(inputs["input_ids"]) |
| self.assertEqual( |
| tokenizer.decode(outputs[0].tolist()), |
| "Hello world! My name is Kiyoshi, and I'm a student at the University of Tokyo", |
| ) |
|
|
| @require_cuda |
| def test_dispatch_model(self): |
| model = ModelForTest() |
| device_map = {"linear1": "disk", "batchnorm": "cpu", "linear2": 0} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| dispatch_model(model, device_map, offload_dir=tmp_dir) |
| output = model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_cuda |
| def test_dispatch_model_tied_weights(self): |
| model = ModelForTestTiedWeights() |
| model.linear1.weight = model.linear2.weight |
| device_map = {"linear1": 0, "batchnorm": 0, "linear2": 0} |
|
|
| dispatch_model(model, device_map) |
| self.assertIs(model.linear2.weight, model.linear1.weight) |
|
|
| @require_multi_gpu |
| def test_dispatch_model_multi_gpu(self): |
| model = BiggerModelForTest() |
| device_map = {"linear1": "cpu", "linear2": "disk", "batchnorm": "cpu", "linear3": 0, "linear4": 1} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| dispatch_model(model, device_map, offload_dir=tmp_dir) |
| output = model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @slow |
| @require_multi_gpu |
| def test_dispatch_model_gpt2_on_two_gpus(self): |
| tokenizer = AutoTokenizer.from_pretrained("gpt2") |
| inputs = tokenizer("Hello world! My name is", return_tensors="pt").to(0) |
|
|
| gpt2 = AutoModelForCausalLM.from_pretrained("gpt2") |
| |
| device_map = { |
| "transformer.wte": 0, |
| "transformer.wpe": 0, |
| "transformer.ln_f": 1, |
| "lm_head": 1, |
| } |
| for i in range(12): |
| device_map[f"transformer.h.{i}"] = 0 if i <= 5 else 1 |
|
|
| gpt2 = dispatch_model(gpt2, device_map) |
| outputs = gpt2.generate(inputs["input_ids"]) |
| self.assertEqual( |
| tokenizer.decode(outputs[0].tolist()), |
| "Hello world! My name is Kiyoshi, and I'm a student at the University of Tokyo", |
| ) |
|
|
| |
| gpt2 = AutoModelForCausalLM.from_pretrained("gpt2") |
| for i in range(4): |
| device_map[f"transformer.h.{i}"] = "cpu" |
| gpt2 = dispatch_model(gpt2, device_map) |
| outputs = gpt2.generate(inputs["input_ids"]) |
| self.assertEqual( |
| tokenizer.decode(outputs[0].tolist()), |
| "Hello world! My name is Kiyoshi, and I'm a student at the University of Tokyo", |
| ) |
| |
| gpt2 = AutoModelForCausalLM.from_pretrained("gpt2") |
| for i in range(2): |
| device_map[f"transformer.h.{i}"] = "disk" |
|
|
| with TemporaryDirectory() as tmp_dir: |
| state_dict = { |
| k: p for k, p in gpt2.state_dict().items() if "transformer.h.0" in k or "transformer.h.1" in k |
| } |
| offload_state_dict(tmp_dir, state_dict) |
| gpt2 = dispatch_model(gpt2, device_map, offload_dir=tmp_dir) |
| outputs = gpt2.generate(inputs["input_ids"]) |
| self.assertEqual( |
| tokenizer.decode(outputs[0].tolist()), |
| "Hello world! My name is Kiyoshi, and I'm a student at the University of Tokyo", |
| ) |
|
|
| @require_cuda |
| def test_dispatch_model_with_unused_submodules(self): |
| model = ModelWithUnusedSubModulesForTest() |
| device_map = {"linear1": "cpu", "linear2": "disk", "batchnorm": "cpu", "linear3": 0, "linear4": 0} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| dispatch_model( |
| model, device_map, offload_dir=tmp_dir, preload_module_classes=["ModuleWithUnusedSubModules"] |
| ) |
| output = model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_multi_gpu |
| def test_dispatch_model_with_unused_submodules_multi_gpu(self): |
| model = ModelWithUnusedSubModulesForTest() |
| device_map = {"linear1": "cpu", "linear2": "disk", "batchnorm": "cpu", "linear3": 0, "linear4": 1} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| dispatch_model( |
| model, device_map, offload_dir=tmp_dir, preload_module_classes=["ModuleWithUnusedSubModules"] |
| ) |
| output = model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_cuda |
| def test_load_checkpoint_and_dispatch(self): |
| model = ModelForTest() |
| device_map = {"linear1": "cpu", "batchnorm": "cpu", "linear2": 0} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| checkpoint = os.path.join(tmp_dir, "pt_model.bin") |
| torch.save(model.state_dict(), checkpoint) |
|
|
| new_model = ModelForTest() |
| new_model = load_checkpoint_and_dispatch(new_model, checkpoint, device_map=device_map) |
|
|
| |
| self.assertEqual(new_model.linear1.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear2.weight.device, torch.device(0)) |
|
|
| output = new_model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_multi_gpu |
| def test_load_checkpoint_and_dispatch_multi_gpu(self): |
| model = BiggerModelForTest() |
| device_map = {"linear1": "cpu", "linear2": "cpu", "batchnorm": 0, "linear3": 0, "linear4": 1} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| checkpoint = os.path.join(tmp_dir, "pt_model.bin") |
| torch.save(model.state_dict(), checkpoint) |
|
|
| new_model = BiggerModelForTest() |
| new_model = load_checkpoint_and_dispatch(new_model, checkpoint, device_map=device_map) |
|
|
| |
| self.assertEqual(new_model.linear1.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear2.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear3.weight.device, torch.device(0)) |
| self.assertEqual(new_model.linear4.weight.device, torch.device(1)) |
|
|
| output = new_model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_cuda |
| def test_load_checkpoint_and_dispatch_with_unused_submodules(self): |
| model = ModelWithUnusedSubModulesForTest() |
| device_map = {"linear1": "cpu", "linear2": "cpu", "batchnorm": 0, "linear3": 0, "linear4": 0} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| checkpoint = os.path.join(tmp_dir, "pt_model.bin") |
| torch.save(model.state_dict(), checkpoint) |
|
|
| new_model = ModelWithUnusedSubModulesForTest() |
| new_model = load_checkpoint_and_dispatch( |
| new_model, checkpoint, device_map=device_map, preload_module_classes=["ModuleWithUnusedSubModules"] |
| ) |
|
|
| |
| self.assertEqual(new_model.linear1.linear.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear2.linear.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear3.linear.weight.device, torch.device(0)) |
| self.assertEqual(new_model.linear4.linear.weight.device, torch.device(0)) |
|
|
| output = new_model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|
| @require_multi_gpu |
| def test_load_checkpoint_and_dispatch_multi_gpu_with_unused_submodules(self): |
| model = ModelWithUnusedSubModulesForTest() |
| device_map = {"linear1": "cpu", "linear2": "cpu", "batchnorm": 0, "linear3": 0, "linear4": 1} |
|
|
| x = torch.randn(2, 3) |
| expected = model(x) |
|
|
| with TemporaryDirectory() as tmp_dir: |
| checkpoint = os.path.join(tmp_dir, "pt_model.bin") |
| torch.save(model.state_dict(), checkpoint) |
|
|
| new_model = ModelWithUnusedSubModulesForTest() |
| new_model = load_checkpoint_and_dispatch( |
| new_model, checkpoint, device_map=device_map, preload_module_classes=["ModuleWithUnusedSubModules"] |
| ) |
|
|
| |
| self.assertEqual(new_model.linear1.linear.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear2.linear.weight.device, torch.device("meta")) |
| self.assertEqual(new_model.linear3.linear.weight.device, torch.device(0)) |
| self.assertEqual(new_model.linear4.linear.weight.device, torch.device(1)) |
|
|
| output = new_model(x) |
| self.assertTrue(torch.allclose(expected, output.cpu(), atol=1e-5)) |
|
|