cyd0806 commited on
Commit
413ad40
·
verified ·
1 Parent(s): 77cbf75

Upload apex-master/tests/L0/run_amp/test_add_param_group.py with huggingface_hub

Browse files
apex-master/tests/L0/run_amp/test_add_param_group.py ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import unittest
2
+
3
+ import functools as ft
4
+ import itertools as it
5
+
6
+ from apex import amp
7
+ from apex.amp import _amp_state
8
+ import torch
9
+ from torch import nn
10
+ import torch.nn.functional as F
11
+ from torch.nn import Parameter
12
+
13
+ from utils import common_init, HALF, FLOAT,\
14
+ ALWAYS_HALF, ALWAYS_FLOAT, MATCH_INPUT
15
+
16
+ class MyModel(torch.nn.Module):
17
+ def __init__(self, unique):
18
+ super(MyModel, self).__init__()
19
+ self.weight0 = Parameter(unique +
20
+ torch.arange(2, device='cuda', dtype=torch.float32))
21
+ self.weight1 = Parameter(1. + unique + torch.arange(2, device='cuda', dtype=torch.float16))
22
+
23
+ @staticmethod
24
+ def ops(input, weight0, weight1):
25
+ return ((input*(weight0.float()))*(weight1.float())).sum()
26
+
27
+ def forward(self, input):
28
+ return self.ops(input, self.weight0, self.weight1)
29
+
30
+
31
+ # Abandon all hope, ye who enter here.
32
+
33
+
34
+ class TestAddParamGroup(unittest.TestCase):
35
+ def setUp(self):
36
+ self.x = torch.ones((2), device='cuda', dtype=torch.float32)
37
+ common_init(self)
38
+
39
+ def tearDown(self):
40
+ pass
41
+
42
+ def zero_grad(self, models, optimizer, how_to_zero):
43
+ if how_to_zero == "none":
44
+ for model in models:
45
+ for param in model.parameters():
46
+ param.grad = None
47
+ elif how_to_zero == "model":
48
+ for model in models:
49
+ model.zero_grad()
50
+ elif how_to_zero == "optimizer":
51
+ optimizer.zero_grad()
52
+
53
+ def test_add_param_group(self):
54
+ for opt_level in ("O0", "O1", "O2", "O3"):
55
+ for zero_before_add in (True, False):
56
+ for try_accumulation in (True, False):
57
+ model0 = MyModel(1)
58
+ model1 = MyModel(2)
59
+
60
+ optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}],
61
+ momentum=0.125)
62
+
63
+ optimizer.zero_grad()
64
+ loss = model0(self.x)
65
+ loss.backward()
66
+ optimizer.step()
67
+
68
+ if zero_before_add:
69
+ optimizer.zero_grad()
70
+ optimizer.add_param_group({'params' : model1.parameters(), 'lr' : 0.5})
71
+ if not zero_before_add:
72
+ optimizer.zero_grad()
73
+
74
+ loss = model0(self.x) + model1(self.x)
75
+ loss.backward(retain_graph=try_accumulation)
76
+ if try_accumulation:
77
+ loss.backward()
78
+ optimizer.step()
79
+
80
+ # Once more to make sure the new params pick up momemtums properly
81
+ optimizer.zero_grad()
82
+ loss = model0(self.x) + model1(self.x)
83
+ loss.backward(retain_graph=try_accumulation)
84
+ if try_accumulation:
85
+ loss.backward()
86
+ optimizer.step()
87
+
88
+ reference_params = [param.data.clone() for param in model0.parameters()] + \
89
+ [param.data.clone() for param in model1.parameters()]
90
+
91
+ for how_to_zero in "none", "model", "optimizer":
92
+ model0 = MyModel(1)
93
+ model1 = MyModel(2)
94
+
95
+ optimizer = torch.optim.SGD([{'params' : model0.parameters(), 'lr' : 0.25}],
96
+ momentum=0.125)
97
+
98
+ _amp_state.allow_incoming_model_not_fp32 = True
99
+ [model0, model1], optimizer = amp.initialize([model0, model1],
100
+ optimizer,
101
+ opt_level=opt_level,
102
+ verbosity=0,
103
+ cast_model_type=False)
104
+ _amp_state.allow_incoming_model_not_fp32 = False
105
+
106
+ _amp_state.loss_scalers[0]._loss_scale = 4.0
107
+
108
+ self.zero_grad([model0, model1], optimizer, how_to_zero)
109
+ loss = model0(self.x)
110
+ with amp.scale_loss(loss, optimizer) as scaled_loss:
111
+ scaled_loss.backward()
112
+ optimizer.step()
113
+
114
+ if zero_before_add:
115
+ self.zero_grad([model0, model1], optimizer, how_to_zero)
116
+ optimizer.add_param_group({'params' : model1.parameters(), 'lr' : 0.5})
117
+ if not zero_before_add:
118
+ self.zero_grad([model0, model1], optimizer, how_to_zero)
119
+
120
+ loss = model0(self.x) + model1(self.x)
121
+ with amp.scale_loss(loss, optimizer) as scaled_loss:
122
+ scaled_loss.backward(retain_graph=try_accumulation)
123
+ if try_accumulation:
124
+ with amp.scale_loss(loss, optimizer) as scaled_loss:
125
+ scaled_loss.backward()
126
+ optimizer.step()
127
+
128
+ # Once more to make sure the new params pick up momentums properly
129
+ self.zero_grad([model0, model1], optimizer, how_to_zero)
130
+ loss = model0(self.x) + model1(self.x)
131
+ with amp.scale_loss(loss, optimizer) as scaled_loss:
132
+ scaled_loss.backward(retain_graph=try_accumulation)
133
+ if try_accumulation:
134
+ with amp.scale_loss(loss, optimizer) as scaled_loss:
135
+ scaled_loss.backward()
136
+ optimizer.step()
137
+
138
+ final_params = [param.data.clone() for param in model0.parameters()] + \
139
+ [param.data.clone() for param in model1.parameters()]
140
+
141
+ for reference, final in zip(reference_params, final_params):
142
+ torch.testing.assert_close(reference.to(final.dtype), final,
143
+ msg="opt_level = {}, how_to_zero = {}, zero_before_add = {}".format(
144
+ opt_level, how_to_zero, zero_before_add))
145
+
146
+
147
+ if __name__ == '__main__':
148
+ unittest.main()