cyd0806 commited on
Commit
71eb649
·
verified ·
1 Parent(s): f349932

Upload apex-master/tests/L0/run_transformer/test_random.py with huggingface_hub

Browse files
apex-master/tests/L0/run_transformer/test_random.py ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import logging
2
+
3
+ import torch
4
+ from torch.testing._internal import common_utils
5
+
6
+ logging.getLogger("torch").setLevel(logging.WARNING)
7
+
8
+ from apex.transformer import parallel_state
9
+ from apex.transformer import tensor_parallel
10
+ from apex.transformer.testing.distributed_test_base import NcclDistributedTestBase
11
+ from apex.transformer.testing.distributed_test_base import UccDistributedTestBase
12
+
13
+ logging.getLogger("apex").setLevel(logging.WARNING)
14
+
15
+
16
+ class TransformerRandomTestBase:
17
+ def test_set_cuda_rng_state(self):
18
+ for tensor_model_parallel_world_size in range(1, self.world_size + 1):
19
+ if self.world_size % tensor_model_parallel_world_size:
20
+ continue
21
+ msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}"
22
+ parallel_state.initialize_model_parallel(
23
+ tensor_model_parallel_size_=tensor_model_parallel_world_size
24
+ )
25
+
26
+ size, seed = 123, 1234
27
+ torch.cuda.manual_seed(seed)
28
+ tensor = torch.cuda.FloatTensor(size)
29
+
30
+ rng_state = torch.cuda.get_rng_state()
31
+ rng_state_clone = rng_state.clone()
32
+
33
+ for _ in range(5):
34
+ torch.randn(size, out=tensor)
35
+ result_1 = tensor.clone()
36
+
37
+ self.assertEqual(rng_state.sub(rng_state_clone).max(), 0, msg=msg)
38
+ self.assertGreater(
39
+ torch.cuda.get_rng_state().sub(rng_state_clone).max(), 0,
40
+ msg=msg,
41
+ )
42
+
43
+ new_rng_state = torch.cuda.get_rng_state()
44
+ self.assertGreater(new_rng_state.sub(rng_state).max(), 0, msg=msg)
45
+
46
+ tensor_parallel.random._set_cuda_rng_state(rng_state)
47
+ for _ in range(5):
48
+ torch.randn(size, out=tensor)
49
+ tensor_parallel.random._set_cuda_rng_state(rng_state)
50
+ for _ in range(5):
51
+ torch.randn(size, out=tensor)
52
+ result_2 = tensor.clone()
53
+
54
+ self.assertEqual(result_2, result_1, msg=msg)
55
+
56
+ self.assertEqual(rng_state.sub(rng_state_clone).max(), 0, msg=msg)
57
+
58
+ parallel_state.destroy_model_parallel()
59
+
60
+ def test_cuda_rng_tracker(self):
61
+ for tensor_model_parallel_world_size in range(1, self.world_size + 1):
62
+ if self.world_size % tensor_model_parallel_world_size:
63
+ continue
64
+ msg = f"tensor_model_parallel_world_size: {tensor_model_parallel_world_size}"
65
+ parallel_state.initialize_model_parallel(
66
+ tensor_model_parallel_size_=tensor_model_parallel_world_size
67
+ )
68
+
69
+ seed_1, seed_2, size = 1234, 4321, [12, 21]
70
+ tensor = torch.cuda.FloatTensor(size)
71
+
72
+ torch.cuda.manual_seed(seed_1)
73
+ torch.randn(size, out=tensor)
74
+ target_11 = tensor.clone()
75
+ torch.randn(size, out=tensor)
76
+ target_12 = tensor.clone()
77
+
78
+ torch.cuda.manual_seed(seed_2)
79
+ torch.randn(size, out=tensor)
80
+ targt_21 = tensor.clone()
81
+ torch.randn(size, out=tensor)
82
+ target_22 = tensor.clone()
83
+
84
+ torch.cuda.manual_seed(seed_1)
85
+ tensor_parallel.random.get_cuda_rng_tracker().add("test", seed_2)
86
+
87
+ torch.randn(size, out=tensor)
88
+ result_11 = tensor.clone()
89
+
90
+ with tensor_parallel.random.get_cuda_rng_tracker().fork("test"):
91
+ torch.randn(size, out=tensor)
92
+ result_21 = tensor.clone()
93
+
94
+ torch.randn(size, out=tensor)
95
+ result_12 = tensor.clone()
96
+
97
+ with tensor_parallel.random.get_cuda_rng_tracker().fork("test"):
98
+ torch.randn(size, out=tensor)
99
+ result_22 = tensor.clone()
100
+
101
+ self.assertEqual(target_11, result_11, msg=msg)
102
+ self.assertEqual(target_12, result_12, msg=msg)
103
+ self.assertEqual(targt_21, result_21, msg=msg)
104
+ self.assertEqual(target_22, result_22, msg=msg)
105
+ self.assertNotEqual(result_11, result_21, msg=msg)
106
+ self.assertNotEqual(result_21, result_22, msg=msg)
107
+
108
+ tensor_parallel.random.get_cuda_rng_tracker().reset()
109
+ parallel_state.destroy_model_parallel()
110
+
111
+
112
+ class NcclTransformerRandomTest(TransformerRandomTestBase, NcclDistributedTestBase): pass
113
+ class UccTransformerRandomTest(TransformerRandomTestBase, UccDistributedTestBase): pass
114
+
115
+
116
+ if __name__ == "__main__":
117
+ common_utils.run_tests()