jason122490 commited on
Commit
2d82643
·
1 Parent(s): b34c4d5

Upload ICCAD.py

Browse files
Files changed (1) hide show
  1. ICCAD.py +718 -199
ICCAD.py CHANGED
@@ -1,48 +1,310 @@
1
  import glob
2
  import os
3
- import re
4
  import datasets
5
  import io
6
  import math
7
  import pandas as pd
8
  import numpy as np
9
  from numpy import genfromtxt
 
10
 
11
- from requests import get
 
 
 
 
 
 
 
12
 
13
  #datasets.logging.set_verbosity_debug()
14
  #logger = datasets.logging.get_logger(__name__)
15
  #datasets.logging.set_verbosity_info()
16
  #datasets.logging.set_verbosity_debug()
 
17
 
18
  _REPO = "https://huggingface.co/datasets/DaJhuan/ICCAD/resolve/main"
19
 
20
  _URLS = {
21
- "fake_data_url": f"{_REPO}/fake-circuit-data.zip",
22
- "real_data_url": f"{_REPO}/real-circuit-data.zip",
 
 
23
  }
24
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
  class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
 
 
 
 
 
 
 
 
 
 
 
27
  def _info(self):
28
- features = datasets.Features({
29
- "data_idx": datasets.Value("string"),
30
- "current": datasets.Image(),
31
- "pdn_density": datasets.Image(),
32
- "eff_dist": datasets.Image(),
33
- "ir_drop": datasets.Image(),
34
- "netlist": datasets.Value("string"),
35
- "R_map_11": datasets.Image(),
36
- "R_map_14": datasets.Image(),
37
- "R_map_44": datasets.Image(),
38
- "R_map_47": datasets.Image(),
39
- "R_map_77": datasets.Image(),
40
- "R_map_78": datasets.Image(),
41
- "R_map_88": datasets.Image(),
42
- "R_map_89": datasets.Image(),
43
- "R_map_99": datasets.Image(),
44
- "I_map": datasets.Image(),
45
- })
46
 
47
  return datasets.DatasetInfo(
48
  features=features,
@@ -64,16 +326,36 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
64
  real_irdrop = []
65
  real_netlist = []
66
 
67
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
68
  # Download images
69
- fake_data_files = os.path.join(dl_manager.download_and_extract(_URLS["fake_data_url"]), "fake-circuit-data") # skip _MACOSX dir
70
- real_data_files = os.path.join(dl_manager.download_and_extract(_URLS["real_data_url"]), "real-circuit-data") # skip _MACOSX dir
71
- #logger.info(f"text_data_files: {text_data_files}")
72
- #logger.info(f"text_data_files: {text_data_files[10]}")
73
 
74
  fake_path_files = sorted(glob.glob(os.path.join(fake_data_files, "*.sp")))
75
  real_path_files = sorted(glob.glob(os.path.join(real_data_files, "*")))
76
 
 
 
 
 
 
 
 
77
  # for fake
78
  for path in fake_path_files:
79
  data_idx = os.path.basename(path).split(".")[0]
@@ -85,16 +367,16 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
85
  fake_cur.append(data)
86
  elif "eff_dist.csv" in os.path.basename(data):
87
  fake_dist.append(data)
88
- elif "ir_drop_map.csv" in os.path.basename(data):
89
  fake_irdrop.append(data)
90
  elif "pdn_density.csv" in os.path.basename(data):
91
  fake_pdn.append(data)
92
  elif ".sp" in os.path.basename(data):
93
  fake_netlist.append(data)
94
  else:
95
- raise AssertionError('fake data path error')
96
 
97
- assert len(fake_idx) == len(fake_cur) == len(fake_dist) == len(fake_irdrop) == len(fake_pdn) == len(fake_netlist), f'{(len(fake_idx), len(fake_cur), len(fake_dist), len(fake_irdrop), len(fake_pdn), len(fake_netlist))} fake data length not the same'
98
 
99
  # for real
100
  for path in real_path_files:
@@ -114,13 +396,67 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
114
  elif "netlist.sp" in os.path.basename(data):
115
  real_netlist.append(data)
116
  else:
117
- raise AssertionError('real data path error')
118
 
119
- assert len(real_idx) == len(real_cur) == len(real_dist) == len(real_irdrop) == len(real_pdn) == len(real_netlist), f'{(len(real_idx), len(real_cur), len(real_dist), len(real_irdrop), len(real_pdn), len(real_netlist))} real data length not the same'
120
-
121
- return [
122
- datasets.SplitGenerator(
123
- name=datasets.Split('fake'),
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
124
  gen_kwargs={
125
  "data_idx": fake_idx,
126
  "current": fake_cur,
@@ -128,10 +464,9 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
128
  "eff_dist": fake_dist,
129
  "ir_drop": fake_irdrop,
130
  "netlist": fake_netlist,
131
- },
132
- ),
133
- datasets.SplitGenerator(
134
- name=datasets.Split('real'),
135
  gen_kwargs={
136
  "data_idx": real_idx,
137
  "current": real_cur,
@@ -139,33 +474,76 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
139
  "eff_dist": real_dist,
140
  "ir_drop": real_irdrop,
141
  "netlist": real_netlist,
142
- },
143
- ),
144
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
145
 
146
 
147
  def _generate_examples(self, data_idx, current, pdn_density, eff_dist, ir_drop, netlist):
148
- for i, (_data_idx, _current, _pdn_density, _eff_dist, _ir_drop, _netlist) in enumerate(zip(data_idx, current, pdn_density, eff_dist, ir_drop, netlist)):
 
 
 
 
149
 
150
- netlist = open(_netlist, 'r').read()
151
- H, W = genfromtxt(_current, delimiter=',').shape
152
- H, W = H + 1, W + 1
153
-
154
- df = pd.read_csv(io.StringIO(netlist), sep = " ", names = ['type', 'n1', 'n2', 'values', 'tmp'])
155
- df = df.drop(columns='tmp')
156
- df = df.astype({'type': 'str', 'n1':'str', 'n2': 'str', 'values': 'float'})
157
- df[['n1', 'm1', 'x1', 'y1']] = df['n1'].str.split('_', expand=True)
158
- df[['n2', 'm2', 'x2', 'y2']] = df['n2'].str.split('_', expand=True)
159
- df['m1'] = df['m1'].str.replace('m', '')
160
- df['m2'] = df['m2'].str.replace('m', '')
161
- df = df.drop(columns=['n1', 'n2'])
 
 
 
 
 
 
 
 
 
 
 
 
162
 
 
 
 
 
 
 
 
163
  # R
164
- df_R = df[df['type'].str.contains('R')]
165
- df_R[['m1', 'm2']] = df_R[['m1', 'm2']].astype(int)
166
- df_R[['x1', 'y1', 'x2', 'y2']] = df_R[['x1', 'y1', 'x2', 'y2']].astype(float) / 2000 + 0.5
 
 
167
 
168
- # m1
169
  R_map_11 = np.zeros((H, W))
170
  R_map_14 = np.zeros((H, W))
171
  R_map_44 = np.zeros((H, W))
@@ -176,197 +554,338 @@ class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
176
  R_map_89 = np.zeros((H, W))
177
  R_map_99 = np.zeros((H, W))
178
 
179
- for index, row in df_R.iterrows():
180
  # m1 -> m1
181
- if row['m1'] == 1 and row['m2'] == 1:
182
  # H line
183
- if row['x1'] != row['x2']:
184
- y = int(row['y1'])
185
- if row['x1'] < row['x2']:
186
- for x in range(int(row['x1']), int(math.ceil(row['x2']))):
187
- R_map_11[x][y] += row['values'] / (row['x2'] - row['x1'])
188
  else:
189
- for x in range(int(row['x2']), int(math.ceil(row['x1']))):
190
- R_map_11[x][y] += row['values'] / (row['x1'] - row['x2'])
191
 
192
  # V line
193
- elif row['y1'] != row['y2']:
194
- x = int(row['x1'])
195
- if row['y1'] < row['y2']:
196
- for y in range(int(row['y1']), int(math.ceil(row['y2']))):
197
- R_map_11[x][y] += row['values'] / (row['y2'] - row['y1'])
198
  else:
199
- for x in range(int(row['y2']), int(math.ceil(row['y1']))):
200
- R_map_11[x][y] += row['values'] / (row['y1'] - row['y2'])
201
  else:
202
- raise AssertionError('R map row error')
203
 
204
  # m1 -> m4
205
- elif row['m1'] == 1 and row['m2'] == 4:
206
  # Via
207
- if row['x1'] == row['x2'] and row['y1'] == row['y2']:
208
- y = int(row['y1'])
209
- x = int(row['x1'])
210
- R_map_14[x][y] += row['values']
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  else:
212
- raise AssertionError('R map row error')
213
 
214
  # m4 -> m4
215
- elif row['m1'] == 4 and row['m2'] == 4:
216
- # H line
217
- if row['x1'] != row['x2']:
218
- y = int(row['y1'])
219
- if row['x1'] < row['x2']:
220
- for x in range(int(row['x1']), int(math.ceil(row['x2']))):
221
- R_map_44[x][y] += row['values'] / (row['x2'] - row['x1'])
222
  else:
223
- for x in range(int(row['x2']), int(math.ceil(row['x1']))):
224
- R_map_44[x][y] += row['values'] / (row['x1'] - row['x2'])
225
 
226
- # V line
227
- elif row['y1'] != row['y2']:
228
- x = int(row['x1'])
229
- if row['y1'] < row['y2']:
230
- for y in range(int(row['y1']), int(math.ceil(row['y2']))):
231
- R_map_44[x][y] += row['values'] / (row['y2'] - row['y1'])
232
  else:
233
- for x in range(int(row['y2']), int(math.ceil(row['y1']))):
234
- R_map_44[x][y] += row['values'] / (row['y1'] - row['y2'])
235
  else:
236
- raise AssertionError('R map row error')
237
 
238
  # m4 -> m7
239
- elif row['m1'] == 4 and row['m2'] == 7:
240
  # Via
241
- if row['x1'] == row['x2'] and row['y1'] == row['y2']:
242
- y = int(row['y1'])
243
- x = int(row['x1'])
244
- R_map_47[x][y] += row['values']
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
245
  else:
246
- raise AssertionError('R map row error')
247
 
248
  # m7 -> m7
249
- elif row['m1'] == 7 and row['m2'] == 7:
250
- # H line
251
- if row['x1'] != row['x2']:
252
- y = int(row['y1'])
253
- if row['x1'] < row['x2']:
254
- for x in range(int(row['x1']), int(math.ceil(row['x2']))):
255
- R_map_77[x][y] += row['values'] / (row['x2'] - row['x1'])
256
  else:
257
- for x in range(int(row['x2']), int(math.ceil(row['x1']))):
258
- R_map_77[x][y] += row['values'] / (row['x1'] - row['x2'])
259
 
260
- # V line
261
- elif row['y1'] != row['y2']:
262
- x = int(row['x1'])
263
- if row['y1'] < row['y2']:
264
- for y in range(int(row['y1']), int(math.ceil(row['y2']))):
265
- R_map_77[x][y] += row['values'] / (row['y2'] - row['y1'])
266
  else:
267
- for x in range(int(row['y2']), int(math.ceil(row['y1']))):
268
- R_map_77[x][y] += row['values'] / (row['y1'] - row['y2'])
269
  else:
270
- raise AssertionError('R map row error')
271
 
272
  # m7 -> m8
273
- elif row['m1'] == 7 and row['m2'] == 8:
274
  # Via
275
- if row['x1'] == row['x2'] and row['y1'] == row['y2']:
276
- y = int(row['y1'])
277
- x = int(row['x1'])
278
- R_map_78[x][y] += row['values']
279
- else:
280
- raise AssertionError('R map row error')
 
 
 
 
 
 
281
 
282
- # m8 -> m8
283
- elif row['m1'] == 8 and row['m2'] == 8:
284
  # H line
285
- if row['x1'] != row['x2']:
286
- y = int(row['y1'])
287
- if row['x1'] < row['x2']:
288
- for x in range(int(row['x1']), int(math.ceil(row['x2']))):
289
- R_map_88[x][y] += row['values'] / (row['x2'] - row['x1'])
290
  else:
291
- for x in range(int(row['x2']), int(math.ceil(row['x1']))):
292
- R_map_88[x][y] += row['values'] / (row['x1'] - row['x2'])
 
293
 
 
 
294
  # V line
295
- elif row['y1'] != row['y2']:
296
- x = int(row['x1'])
297
- if row['y1'] < row['y2']:
298
- for y in range(int(row['y1']), int(math.ceil(row['y2']))):
299
- R_map_88[x][y] += row['values'] / (row['y2'] - row['y1'])
 
 
 
 
 
 
 
300
  else:
301
- for x in range(int(row['y2']), int(math.ceil(row['y1']))):
302
- R_map_88[x][y] += row['values'] / (row['y1'] - row['y2'])
303
  else:
304
- raise AssertionError('R map row error')
305
 
306
  # m8 -> m9
307
- elif row['m1'] == 8 and row['m2'] == 9:
308
  # Via
309
- if row['x1'] == row['x2'] and row['y1'] == row['y2']:
310
- y = int(row['y1'])
311
- x = int(row['x1'])
312
- R_map_89[x][y] += row['values']
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
313
  else:
314
- raise AssertionError('R map row error')
315
 
316
  # m9 -> m9
317
- elif row['m1'] == 9 and row['m2'] == 9:
318
- # H line
319
- if row['x1'] != row['x2']:
320
- y = int(row['y1'])
321
- if row['x1'] < row['x2']:
322
- for x in range(int(row['x1']), int(math.ceil(row['x2']))):
323
- R_map_99[x][y] += row['values'] / (row['x2'] - row['x1'])
324
  else:
325
- for x in range(int(row['x2']), int(math.ceil(row['x1']))):
326
- R_map_99[x][y] += row['values'] / (row['x1'] - row['x2'])
327
 
328
- # V line
329
- elif row['y1'] != row['y2']:
330
- x = int(row['x1'])
331
- if row['y1'] < row['y2']:
332
- for y in range(int(row['y1']), int(math.ceil(row['y2']))):
333
- R_map_99[x][y] += row['values'] / (row['y2'] - row['y1'])
334
  else:
335
- for x in range(int(row['y2']), int(math.ceil(row['y1']))):
336
- R_map_99[x][y] += row['values'] / (row['y1'] - row['y2'])
337
  else:
338
- raise AssertionError('R map row error')
339
  else:
340
- raise AssertionError('R map layer not found', row['m1'], row['m2'])
 
 
 
 
 
 
 
 
 
 
341
 
 
342
  # I
343
- df_I = df[df['type'].str.contains('I')]
344
- df_I[['x1', 'y1']] = df_I[['x1', 'y1']].astype(float) / 2000 + 0.5
 
 
345
 
346
  I_map = np.zeros((H, W))
347
-
348
- for index, row in df_I.iterrows():
349
- y = int(row['y1'])
350
- x = int(row['x1'])
351
- I_map[x][y] += row['values']
 
352
 
 
 
 
 
 
 
 
353
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
354
 
355
- yield i, {
356
- "data_idx": _data_idx,
357
- "current": genfromtxt(_current, delimiter=',') * 1e6,
358
- "pdn_density": genfromtxt(_pdn_density, delimiter=','),
359
- "eff_dist": genfromtxt(_eff_dist, delimiter=','),
360
- "ir_drop": genfromtxt(_ir_drop, delimiter=',') * 1e3,
361
- "netlist": open(_netlist, 'r').read(),
362
- "R_map_11": R_map_11,
363
- "R_map_14": R_map_14,
364
- "R_map_44": R_map_44,
365
- "R_map_47": R_map_47,
366
- "R_map_77": R_map_77,
367
- "R_map_78": R_map_78,
368
- "R_map_88": R_map_88,
369
- "R_map_89": R_map_89,
370
- "R_map_99": R_map_99,
371
- "I_map": I_map * 1e6,
372
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  import glob
2
  import os
 
3
  import datasets
4
  import io
5
  import math
6
  import pandas as pd
7
  import numpy as np
8
  from numpy import genfromtxt
9
+ from dataclasses import dataclass
10
 
11
+ # transform
12
+ import cv2
13
+ import torch
14
+ import torch.nn.functional as F
15
+ import albumentations as A
16
+ from albumentations.augmentations.blur.functional import gaussian_blur
17
+ from scipy.ndimage import distance_transform_cdt, distance_transform_edt
18
+ from patchify import patchify
19
 
20
  #datasets.logging.set_verbosity_debug()
21
  #logger = datasets.logging.get_logger(__name__)
22
  #datasets.logging.set_verbosity_info()
23
  #datasets.logging.set_verbosity_debug()
24
+ pd.options.mode.chained_assignment = None
25
 
26
  _REPO = "https://huggingface.co/datasets/DaJhuan/ICCAD/resolve/main"
27
 
28
  _URLS = {
29
+ "fake_data_url": f"{_REPO}/fake-circuit-data_20230623.zip",
30
+ "real_data_url": f"{_REPO}/real-circuit-data_20230615.zip",
31
+ "BeGAN_01_data_url": f"{_REPO}/BeGAN-ver01.zip",
32
+ "BeGAN_02_data_url": f"{_REPO}/BeGAN-ver02.zip",
33
  }
34
 
35
+ @dataclass
36
+ class ICCAD_Config(datasets.BuilderConfig):
37
+ test_mode: bool = False
38
+ use_BeGAN: bool = False and not test_mode
39
+ patchify: bool = False
40
+ # transform
41
+ use_resize: bool = True
42
+ img_size: int = 288
43
+ dist_type: str = 'edt'
44
+ use_i_map: bool = True
45
+ use_r_map: bool = True
46
+ use_r_dist: bool = True
47
+ use_ir_map: bool = True
48
+ use_multi_dist: bool = False
49
+ use_power_map: bool = False
50
+
51
+ # tranform
52
+ def _transform(self, img, max_pool=False):
53
+ img_size = self.config.img_size
54
+ img = np.asarray(img)
55
+
56
+ if self.config.patchify:
57
+ tf = A.Compose([
58
+ A.PadIfNeeded(min_height=None, min_width=None, pad_height_divisor=img_size, pad_width_divisor=img_size, border_mode=0, value=0),
59
+ ],
60
+ )
61
+ img = tf(image=img)['image']
62
+ img = patchify(img, patch_size=(img_size, img_size, img.shape[-1]), step=img_size)
63
+ img = img.reshape((-1, img_size, img_size, img.shape[-1])).transpose((0, 3, 1, 2))
64
+ elif self.config.use_resize:
65
+ img = A.Resize(img_size, img_size, interpolation=cv2.INTER_AREA)(image=img)["image"]
66
+ img = img.transpose((2, 0, 1))
67
+ elif max_pool:
68
+ img = torch.from_numpy(img).unsqueeze(0)
69
+ img = F.adaptive_max_pool2d(img, img_size)
70
+ img = img.squeeze(0).numpy()
71
+ img = img.transpose((2, 0, 1))
72
+
73
+ return img
74
+
75
+ def get_dist(self, x):
76
+ x = np.asarray(x)
77
+ if self.config.dist_type == 'cdt':
78
+ x = distance_transform_cdt(np.where(x == 0, 1, 0), metric='taxicab')
79
+ if self.config.dist_type == 'edt':
80
+ x = distance_transform_edt(np.where(x == 0, 1, 0))
81
+ return x
82
+
83
+ def get_blur(x, ksize=1):
84
+ x = np.asarray(x)
85
+ if ksize > 1:
86
+ x = gaussian_blur(img=x, ksize=ksize)
87
+ return x
88
+
89
+ def get_image(self, data):
90
+ return {
91
+ "data_idx": data["data_idx"],
92
+ "H": data["H"],
93
+ "W": data["W"],
94
+ "image": _transform(self, np.stack([
95
+ data["ir_drop"],
96
+ data["current"],
97
+ data["eff_dist"],
98
+ data["pdn_density"],
99
+ ] + ([
100
+ get_blur(data["I_map"], ksize=1),
101
+ ] if self.config.use_i_map else []) + ([
102
+ get_blur(data["R_map_11"], ksize=1),
103
+ get_blur(data["R_map_14"], ksize=1),
104
+ get_blur(data["R_map_44"], ksize=1),
105
+ get_blur(data["R_map_47"], ksize=1),
106
+ get_blur(data["R_map_77"], ksize=1),
107
+ get_blur(data["R_map_78"], ksize=1),
108
+ get_blur(data["R_map_88"], ksize=1),
109
+ get_blur(data["R_map_89"], ksize=1),
110
+ get_blur(data["R_map_99"], ksize=1),
111
+ ] if self.config.use_r_map else []) + ([
112
+ get_dist(self, data["R_map_11"]),
113
+ get_dist(self, data["R_map_14"]),
114
+ get_dist(self, data["R_map_44"]),
115
+ get_dist(self, data["R_map_47"]),
116
+ get_dist(self, data["R_map_77"]),
117
+ get_dist(self, data["R_map_78"]),
118
+ get_dist(self, data["R_map_88"]),
119
+ get_dist(self, data["R_map_89"]),
120
+ get_dist(self, data["R_map_99"]),
121
+ ] if self.config.use_r_dist else []) + ([
122
+ get_blur(data["IR_map_14"], ksize=1),
123
+ get_blur(data["IR_map_47"], ksize=1),
124
+ get_blur(data["IR_map_78"], ksize=1),
125
+ get_blur(data["IR_map_89"], ksize=1),
126
+ ] if self.config.use_ir_map else []) + ([
127
+ data["R_multi_dist_14"],
128
+ data["R_multi_dist_47"],
129
+ data["R_multi_dist_78"],
130
+ data["R_multi_dist_89"],
131
+ ] if self.config.use_multi_dist else []) + ([
132
+ get_dist(self, data["R_map_11"]) * get_blur(data["I_map"], ksize=49),
133
+ get_dist(self, data["R_map_14"]) * get_blur(data["I_map"], ksize=49),
134
+ get_dist(self, data["R_map_44"]) * get_blur(data["I_map"], ksize=49),
135
+ get_dist(self, data["R_map_47"]) * get_blur(data["I_map"], ksize=49),
136
+ get_dist(self, data["R_map_77"]) * get_blur(data["I_map"], ksize=49),
137
+ get_dist(self, data["R_map_78"]) * get_blur(data["I_map"], ksize=49),
138
+ get_dist(self, data["R_map_88"]) * get_blur(data["I_map"], ksize=49),
139
+ get_dist(self, data["R_map_89"]) * get_blur(data["I_map"], ksize=49),
140
+ get_dist(self, data["R_map_99"]) * get_blur(data["I_map"], ksize=49),
141
+ ] if self.config.use_power_map else []), axis=2))
142
+ }
143
+
144
+ def get_IR_map_H(Via_map, I_map):
145
+ index = np.argwhere(Via_map != 0)
146
+ index = index[index[:, 0].argsort()]
147
+ index = index[index[:, 1].argsort(kind="mergesort")]
148
+
149
+ H, W = Via_map.shape
150
+ IR_map = np.zeros((H, W))
151
+ I_sim_map = np.zeros((H, W))
152
+ #start
153
+ for j in range(0, index[0, 0]):
154
+ I = I_map[j, index[0, 1]]
155
+ if I != 0:
156
+ R = int(index[0, 0] - j)
157
+ V = float(I * R)
158
+ IR_map[0 : j, index[0, 1]] += V
159
+ IR_map[j : index[0, 0], index[0, 1]] += V * np.arange(R, 0, -1) / R
160
+
161
+ I_sim_map[index[0, 0], index[0, 1]] += I
162
+ #mid
163
+ for i in range(index.shape[0] - 1):
164
+ if index[i, 1] == index[i + 1, 1]:
165
+ R = index[i + 1, 0] - index[i, 0]
166
+ for j in range(index[i, 0], index[i + 1, 0]):
167
+ I = I_map[j, index[i, 1]]
168
+ if I != 0:
169
+ R1 = int(j - index[i, 0])
170
+ R2 = int(index[i + 1, 0] - j)
171
+ V = float(I * (R1 * R2) / R)
172
+ IR_map[index[i, 0] : j, index[i, 1]] += V * np.arange(R1) / R1
173
+ IR_map[j, index[i, 1]] += V
174
+ IR_map[j : index[i + 1, 0], index[i, 1]] += V * np.arange(R2, 0, -1) / R2
175
+
176
+ I_sim_map[index[i, 0], index[i, 1]] += I * R2 / R
177
+ I_sim_map[index[i + 1, 0], index[i + 1, 1]] += I * R1 / R
178
+ else:
179
+ for j in range(index[i, 0], H):
180
+ I = I_map[j, index[i, 1]]
181
+ if I != 0:
182
+ R = int(j - index[i, 0])
183
+ V = float(I * R)
184
+ IR_map[index[i, 0] : j, index[i, 1]] += V * np.arange(R) / R
185
+ IR_map[j : H, index[i, 1]] += V
186
+
187
+ I_sim_map[index[i, 0], index[i, 1]] += I
188
+ for j in range(0, index[i + 1, 0]):
189
+ I = I_map[j, index[i + 1, 1]]
190
+ if I != 0:
191
+ R = int(index[i + 1, 0] - j)
192
+ V = float(I * R)
193
+ IR_map[0 : j, index[i + 1, 1]] += V
194
+ IR_map[j : index[i + 1, 0], index[i + 1, 1]] += V * np.arange(R, 0, -1) / R
195
+
196
+ I_sim_map[index[i + 1, 0], index[i + 1, 1]] += I
197
+ #end
198
+ i = index.shape[0] - 1
199
+ for j in range(index[i, 0], H):
200
+ I = I_map[j, index[i, 1]]
201
+ if I != 0:
202
+ R = int(j - index[i, 0])
203
+ V = float(I * R)
204
+ IR_map[index[i, 0] : j, index[i, 1]] += V * np.arange(R) / R
205
+ IR_map[j : H, index[i, 1]] += V
206
+
207
+ I_sim_map[index[i, 0], index[i, 1]] += I
208
+
209
+ return I_sim_map, IR_map
210
+
211
+ def get_IR_map_V(Via_map, I_map):
212
+ index = np.argwhere(Via_map != 0)
213
+ index = index[index[:, 1].argsort()]
214
+ index = index[index[:, 0].argsort(kind="mergesort")]
215
+
216
+ H, W = Via_map.shape
217
+ IR_map = np.zeros((H, W))
218
+ I_sim_map = np.zeros((H, W))
219
+ #start
220
+ for j in range(0, index[0, 1]):
221
+ I = I_map[index[0, 0], j]
222
+ if I != 0:
223
+ R = int(index[0, 1] - j)
224
+ V = float(I * R)
225
+ IR_map[index[0, 0], 0 : j] += V
226
+ IR_map[index[0, 0], j : index[0, 1]] += V * np.arange(R, 0, -1) / R
227
+
228
+ I_sim_map[index[0, 0], index[0, 1]] += I
229
+ #mid
230
+ for i in range(index.shape[0] - 1):
231
+ if index[i, 0] == index[i + 1, 0]:
232
+ R = index[i + 1, 1] - index[i, 1]
233
+ for j in range(index[i, 1], index[i + 1, 1]):
234
+ I = I_map[index[i, 0], j]
235
+ if I != 0:
236
+ R1 = int(j - index[i, 1])
237
+ R2 = int(index[i + 1, 1] - j)
238
+ V = float(I * (R1 * R2) / R)
239
+ IR_map[index[i, 0], index[i, 1] : j] += V * np.arange(R1) / R1
240
+ IR_map[index[i, 0], j] += V
241
+ IR_map[index[i, 0], j : index[i + 1, 1]] += V * np.arange(R2, 0, -1) / R2
242
+
243
+ I_sim_map[index[i, 0], index[i, 1]] += I * R2 / R
244
+ I_sim_map[index[i + 1, 0], index[i + 1, 1]] += I * R1 / R
245
+ else:
246
+ for j in range(index[i, 1], W):
247
+ I = I_map[index[i, 0], j]
248
+ if I != 0:
249
+ R = int(j - index[i, 1])
250
+ V = float(I * R)
251
+ IR_map[index[i, 0], index[i, 1] : j] += V * np.arange(R) / R
252
+ IR_map[index[i, 0], j : W] += V
253
+
254
+ I_sim_map[index[i, 0], index[i, 1]] += I
255
+ for j in range(0, index[i + 1, 1]):
256
+ I = I_map[index[i + 1, 0], j]
257
+ if I != 0:
258
+ R = int(index[i + 1, 1] - j)
259
+ V = float(I * R)
260
+ IR_map[index[i + 1, 0], 0 : j] += V
261
+ IR_map[index[i + 1, 0], j : index[i + 1, 1]] += V * np.arange(R, 0, -1) / R
262
+
263
+ I_sim_map[index[i + 1, 0], index[i + 1, 1]] += I
264
+ #end
265
+ i = index.shape[0] - 1
266
+ for j in range(index[i, 1], H):
267
+ I = I_map[index[i, 0], j]
268
+ if I != 0:
269
+ R = int(j - index[i, 1])
270
+ V = float(I * R)
271
+ IR_map[index[i, 0], index[i, 1] : j] += V * np.arange(R) / R
272
+ IR_map[index[i, 0], j : W] += V
273
+
274
+ I_sim_map[index[i, 0], index[i, 1]] += I
275
+
276
+ return I_sim_map, IR_map
277
 
278
  class ICCAD_Dataset(datasets.GeneratorBasedBuilder):
279
+ DEFAULT_WRITER_data_SIZE = 10
280
+
281
+ BUILDER_CONFIG_CLASS = ICCAD_Config
282
+ BUILDER_CONFIGS = [
283
+ ICCAD_Config(
284
+ name="ICCAD",
285
+ version=datasets.Version("1.0.0", ""),
286
+ description="ICCAD Problem C: Static IR Drop Estimation Using Machine Learning",
287
+ )
288
+ ]
289
+
290
  def _info(self):
291
+ img_size = self.config.img_size
292
+ in_chans = 1 + 3 + self.config.use_i_map + 9 * self.config.use_r_map + 9 * self.config.use_r_dist + 4 * self.config.use_ir_map + 4 * self.config.use_multi_dist + 9 * self.config.use_power_map
293
+
294
+ if self.config.patchify:
295
+ features = datasets.Features({
296
+ "data_idx": datasets.Value("string"),
297
+ "H": datasets.Value("int32"),
298
+ "W": datasets.Value("int32"),
299
+ "image": datasets.Array4D((None, in_chans, img_size, img_size), 'float32'),
300
+ })
301
+ else:
302
+ features = datasets.Features({
303
+ "data_idx": datasets.Value("string"),
304
+ "H": datasets.Value("int32"),
305
+ "W": datasets.Value("int32"),
306
+ "image": datasets.Array3D((in_chans, img_size, img_size), 'float32'),
307
+ })
 
308
 
309
  return datasets.DatasetInfo(
310
  features=features,
 
326
  real_irdrop = []
327
  real_netlist = []
328
 
329
+ if self.config.use_BeGAN:
330
+ BeGAN_01_idx = []
331
+ BeGAN_01_cur = []
332
+ BeGAN_01_pdn = []
333
+ BeGAN_01_dist = []
334
+ BeGAN_01_irdrop = []
335
+ BeGAN_01_netlist = []
336
+
337
+ BeGAN_02_idx = []
338
+ BeGAN_02_cur = []
339
+ BeGAN_02_pdn = []
340
+ BeGAN_02_dist = []
341
+ BeGAN_02_irdrop = []
342
+ BeGAN_02_netlist = []
343
+
344
+
345
  # Download images
346
+ fake_data_files = os.path.join(dl_manager.download_and_extract(_URLS["fake_data_url"]), "fake-circuit-data_20230623")
347
+ real_data_files = os.path.join(dl_manager.download_and_extract(_URLS["real_data_url"]), "real-circuit-data_20230615")
 
 
348
 
349
  fake_path_files = sorted(glob.glob(os.path.join(fake_data_files, "*.sp")))
350
  real_path_files = sorted(glob.glob(os.path.join(real_data_files, "*")))
351
 
352
+ if self.config.use_BeGAN:
353
+ BeGAN_01_data_files = os.path.join(dl_manager.download_and_extract(_URLS["BeGAN_01_data_url"]), "BeGAN-ver01")
354
+ BeGAN_01_path_files = sorted(glob.glob(os.path.join(BeGAN_01_data_files, "*.sp")))
355
+
356
+ BeGAN_02_data_files = os.path.join(dl_manager.download_and_extract(_URLS["BeGAN_02_data_url"]), "BeGAN-ver02")
357
+ BeGAN_02_path_files = sorted(glob.glob(os.path.join(BeGAN_02_data_files, "*.sp")))
358
+
359
  # for fake
360
  for path in fake_path_files:
361
  data_idx = os.path.basename(path).split(".")[0]
 
367
  fake_cur.append(data)
368
  elif "eff_dist.csv" in os.path.basename(data):
369
  fake_dist.append(data)
370
+ elif "ir_drop.csv" in os.path.basename(data):
371
  fake_irdrop.append(data)
372
  elif "pdn_density.csv" in os.path.basename(data):
373
  fake_pdn.append(data)
374
  elif ".sp" in os.path.basename(data):
375
  fake_netlist.append(data)
376
  else:
377
+ raise AssertionError(os.path.basename(data), "fake data path error")
378
 
379
+ assert len(fake_idx) == len(fake_cur) == len(fake_dist) == len(fake_irdrop) == len(fake_pdn) == len(fake_netlist), f"{(len(fake_idx), len(fake_cur), len(fake_dist), len(fake_irdrop), len(fake_pdn), len(fake_netlist))} fake data length not the same"
380
 
381
  # for real
382
  for path in real_path_files:
 
396
  elif "netlist.sp" in os.path.basename(data):
397
  real_netlist.append(data)
398
  else:
399
+ raise AssertionError(os.path.basename(data), "real data path error")
400
 
401
+ assert len(real_idx) == len(real_cur) == len(real_dist) == len(real_irdrop) == len(real_pdn) == len(real_netlist), f"{(len(real_idx), len(real_cur), len(real_dist), len(real_irdrop), len(real_pdn), len(real_netlist))} real data length not the same"
402
+ if self.config.use_BeGAN:
403
+ # for BeGAN-ver01
404
+ for path in BeGAN_01_path_files:
405
+ data_idx = os.path.basename(path).split(".")[0]
406
+ BeGAN_01_idx.append(data_idx)
407
+ data_path = glob.glob(os.path.join(os.path.dirname(path), data_idx + "*.*"))
408
+
409
+ for data in data_path:
410
+ if "current.csv" in os.path.basename(data):
411
+ BeGAN_01_cur.append(data)
412
+ elif "eff_dist.csv" in os.path.basename(data):
413
+ BeGAN_01_dist.append(data)
414
+ elif "ir_drop_map.csv" in os.path.basename(data):
415
+ BeGAN_01_irdrop.append(data)
416
+ elif "pdn_density.csv" in os.path.basename(data):
417
+ BeGAN_01_pdn.append(data)
418
+ elif ".sp" in os.path.basename(data):
419
+ BeGAN_01_netlist.append(data)
420
+ else:
421
+ raise AssertionError(os.path.basename(data), "BeGAN-ver01 data path error")
422
+
423
+ assert len(BeGAN_01_idx) == len(BeGAN_01_cur) == len(BeGAN_01_dist) == len(BeGAN_01_irdrop) == len(BeGAN_01_pdn) == len(BeGAN_01_netlist), f"{(len(BeGAN_01_idx), len(BeGAN_01_cur), len(BeGAN_01_dist), len(BeGAN_01_irdrop), len(BeGAN_01_pdn), len(BeGAN_01_netlist))} BeGAN-ver02 data length not the same"
424
+
425
+ # for BeGAN-ver02
426
+ for path in BeGAN_02_path_files:
427
+ data_idx = os.path.basename(path).split(".")[0]
428
+ BeGAN_02_idx.append(data_idx)
429
+ data_path = glob.glob(os.path.join(os.path.dirname(path), data_idx + "*.*"))
430
+
431
+ for data in data_path:
432
+ if "current.csv" in os.path.basename(data):
433
+ BeGAN_02_cur.append(data)
434
+ elif "eff_dist.csv" in os.path.basename(data):
435
+ BeGAN_02_dist.append(data)
436
+ elif "voltage.csv" in os.path.basename(data):
437
+ BeGAN_02_irdrop.append(data)
438
+ elif "regions.csv" in os.path.basename(data):
439
+ BeGAN_02_pdn.append(data)
440
+ elif ".sp" in os.path.basename(data):
441
+ BeGAN_02_netlist.append(data)
442
+ else:
443
+ raise AssertionError(os.path.basename(data), "BeGAN-ver01 data path error")
444
+
445
+ assert len(BeGAN_02_idx) == len(BeGAN_02_cur) == len(BeGAN_02_dist) == len(BeGAN_02_irdrop) == len(BeGAN_02_pdn) == len(BeGAN_02_netlist), f"{(len(BeGAN_02_idx), len(BeGAN_02_cur), len(BeGAN_02_dist), len(BeGAN_02_irdrop), len(BeGAN_02_pdn), len(BeGAN_02_netlist))} BeGAN-ver01 data length not the same"
446
+
447
+ return [datasets.SplitGenerator(
448
+ name=datasets.Split("real"),
449
+ gen_kwargs={
450
+ "data_idx": real_idx,
451
+ "current": real_cur,
452
+ "pdn_density": real_pdn,
453
+ "eff_dist": real_dist,
454
+ "ir_drop": real_irdrop,
455
+ "netlist": real_netlist,
456
+ })
457
+ ] if self.config.test_mode else [
458
+ datasets.SplitGenerator(
459
+ name=datasets.Split("fake"),
460
  gen_kwargs={
461
  "data_idx": fake_idx,
462
  "current": fake_cur,
 
464
  "eff_dist": fake_dist,
465
  "ir_drop": fake_irdrop,
466
  "netlist": fake_netlist,
467
+ })
468
+ ] + [datasets.SplitGenerator(
469
+ name=datasets.Split("real"),
 
470
  gen_kwargs={
471
  "data_idx": real_idx,
472
  "current": real_cur,
 
474
  "eff_dist": real_dist,
475
  "ir_drop": real_irdrop,
476
  "netlist": real_netlist,
477
+ })
478
+ ] + ([datasets.SplitGenerator(
479
+ name=datasets.Split("BeGAN_01"),
480
+ gen_kwargs={
481
+ "data_idx": BeGAN_01_idx,
482
+ "current": BeGAN_01_cur,
483
+ "pdn_density": BeGAN_01_pdn,
484
+ "eff_dist": BeGAN_01_dist,
485
+ "ir_drop": BeGAN_01_irdrop,
486
+ "netlist": BeGAN_01_netlist,
487
+ }),
488
+ datasets.SplitGenerator(
489
+ name=datasets.Split("BeGAN_02"),
490
+ gen_kwargs={
491
+ "data_idx": BeGAN_02_idx,
492
+ "current": BeGAN_02_cur,
493
+ "pdn_density": BeGAN_02_pdn,
494
+ "eff_dist": BeGAN_02_dist,
495
+ "ir_drop": BeGAN_02_irdrop,
496
+ "netlist": BeGAN_02_netlist,
497
+ })
498
+ ] if self.config.use_BeGAN else [])
499
 
500
 
501
  def _generate_examples(self, data_idx, current, pdn_density, eff_dist, ir_drop, netlist):
502
+ for _idx, (_data_idx, _current, _pdn_density, _eff_dist, _ir_drop, _netlist) in enumerate(zip(data_idx, current, pdn_density, eff_dist, ir_drop, netlist)):
503
+
504
+ netlist = open(_netlist, "r").read()
505
+ _H, _W = genfromtxt(_current, delimiter=",").shape
506
+ H, W = int(_H * 2.5) + 1, int(_W * 2.5) + 1
507
 
508
+ def resize(img, max_pool=False):
509
+ if max_pool:
510
+ img = torch.from_numpy(img).unsqueeze(0)
511
+ img = F.adaptive_max_pool2d(img, (_H, _W))
512
+ img = img.squeeze(0).numpy()
513
+ else:
514
+ img = A.Resize(_H, _W, interpolation=cv2.INTER_AREA)(image=img)["image"]
515
+ return img
516
+
517
+ df = pd.read_csv(io.StringIO(netlist), sep = " ", names = ["type", "n1", "n2", "values", "tmp"], low_memory=False)
518
+ df = df.drop(columns="tmp")
519
+ df = df.astype({"type": "str", "n1":"str", "n2": "str", "values": "float"})
520
+ df[["n1", "m1", "x1", "y1"]] = df["n1"].str.split("_", expand=True)
521
+ df[["n2", "m2", "x2", "y2"]] = df["n2"].str.split("_", expand=True)
522
+ df["m1"] = df["m1"].str.replace("m", "")
523
+ df["m2"] = df["m2"].str.replace("m", "")
524
+ df = df.drop(columns=["n1", "n2"])
525
+
526
+
527
+ # V
528
+ df_V = df[df["type"].str.contains("V")]
529
+ assert df_V[["x1", "y1"]].astype(int).mod(800).eq(0).all().all(), (_data_idx, "V map error")
530
+ df_V[["x1", "y1"]] = df_V[["x1", "y1"]].astype(int) // 800
531
+ df_V = df_V[["x1", "y1", "values"]].to_numpy()
532
 
533
+ V_map = np.zeros((H, W))
534
+ for row in df_V:
535
+ x = int(row[0])
536
+ y = int(row[1])
537
+ V_map[x, y] = row[2]
538
+
539
+
540
  # R
541
+ df_R = df[df["type"].str.contains("R")]
542
+ df_R[["m1", "m2"]] = df_R[["m1", "m2"]].astype(int)
543
+ assert df_R[["x1", "y1", "x2", "y2"]].astype(int).mod(800).eq(0).all().all(), (_data_idx, "R map error")
544
+ df_R[["x1", "y1", "x2", "y2"]] = df_R[["x1", "y1", "x2", "y2"]].astype(int) // 800
545
+ df_R = df_R[["m1", "x1", "y1", "m2", "x2", "y2", "values"]].to_numpy()
546
 
 
547
  R_map_11 = np.zeros((H, W))
548
  R_map_14 = np.zeros((H, W))
549
  R_map_44 = np.zeros((H, W))
 
554
  R_map_89 = np.zeros((H, W))
555
  R_map_99 = np.zeros((H, W))
556
 
557
+ for row in df_R:
558
  # m1 -> m1
559
+ if row[0] == 1 and row[3] == 1:
560
  # H line
561
+ if row[1] != row[4]:
562
+ y = int(row[2])
563
+ if row[1] < row[4]:
564
+ R_map_11[int(row[1]) : int(row[4] + 1), y] = 1
 
565
  else:
566
+ R_map_11[int(row[4]) : int(row[1] + 1), y] = 1
 
567
 
568
  # V line
569
+ elif row[2] != row[5]:
570
+ x = int(row[1])
571
+ if row[2] < row[5]:
572
+ R_map_11[x, int(row[2]) : int(row[5] + 1)] = 1
 
573
  else:
574
+ R_map_11[x, int(row[5]) : int(row[2] + 1)] = 1
 
575
  else:
576
+ raise AssertionError(_data_idx, "R map row error")
577
 
578
  # m1 -> m4
579
+ elif row[0] == 1 and row[3] == 4:
580
  # Via
581
+ if row[1] == row[4] and row[2] == row[5]:
582
+ x = int(row[1])
583
+ y = int(row[2])
584
+ R_map_14[x, y] = 1
585
+
586
+ # V line
587
+ elif row[1] != row[4]:
588
+ y = int(row[2])
589
+ if row[1] < row[4]:
590
+ R_map_14[int(row[1]) : int(row[4] + 1), y] = 1
591
+ else:
592
+ R_map_14[int(row[4]) : int(row[1] + 1), y] = 1
593
+
594
+ # H line
595
+ elif row[2] != row[5]:
596
+ x = int(row[1])
597
+ if row[2] < row[5]:
598
+ R_map_14[x, int(row[2]) : int(row[5] + 1)] = 1
599
+ else:
600
+ R_map_14[x, int(row[5]) : int(row[2] + 1)] = 1
601
  else:
602
+ raise AssertionError(_data_idx, "R map row error")
603
 
604
  # m4 -> m4
605
+ elif row[0] == 4 and row[3] == 4:
606
+ # V line
607
+ if row[1] != row[4]:
608
+ y = int(row[2])
609
+ if row[1] < row[4]:
610
+ R_map_44[int(row[1]) : int(row[4] + 1), y] = 1
 
611
  else:
612
+ R_map_44[int(row[4]) : int(row[1] + 1), y] = 1
 
613
 
614
+ # H line
615
+ elif row[2] != row[5]:
616
+ x = int(row[1])
617
+ if row[2] < row[5]:
618
+ R_map_44[x, int(row[2]) : int(row[5] + 1)] = 1
 
619
  else:
620
+ R_map_44[x, int(row[5]) : int(row[2] + 1)] = 1
 
621
  else:
622
+ raise AssertionError(_data_idx, "R map row error")
623
 
624
  # m4 -> m7
625
+ elif row[0] == 4 and row[3] == 7:
626
  # Via
627
+ if row[1] == row[4] and row[2] == row[5]:
628
+ x = int(row[1])
629
+ y = int(row[2])
630
+ R_map_47[x, y] = 1
631
+
632
+ # V line
633
+ elif row[1] != row[4]:
634
+ y = int(row[2])
635
+ if row[1] < row[4]:
636
+ R_map_47[int(row[1]) : int(row[4] + 1), y] = 1
637
+ else:
638
+ R_map_47[int(row[4]) : int(row[1] + 1), y] = 1
639
+
640
+ # H line
641
+ elif row[2] != row[5]:
642
+ x = int(row[1])
643
+ if row[2] < row[5]:
644
+ R_map_47[x, int(row[2]) : int(row[5] + 1)] = 1
645
+ else:
646
+ R_map_47[x, int(row[5]) : int(row[2] + 1)] = 1
647
  else:
648
+ raise AssertionError(_data_idx, "R map row error")
649
 
650
  # m7 -> m7
651
+ elif row[0] == 7 and row[3] == 7:
652
+ # V line
653
+ if row[1] != row[4]:
654
+ y = int(row[2])
655
+ if row[1] < row[4]:
656
+ R_map_77[int(row[1]) : int(row[4] + 1), y] = 1
 
657
  else:
658
+ R_map_77[int(row[4]) : int(row[1] + 1), y] = 1
 
659
 
660
+ # H line
661
+ elif row[2] != row[5]:
662
+ x = int(row[1])
663
+ if row[2] < row[5]:
664
+ R_map_77[x, int(row[2]) : int(row[5] + 1)] = 1
 
665
  else:
666
+ R_map_77[x, int(row[5]) : int(row[2] + 1)] = 1
 
667
  else:
668
+ raise AssertionError(_data_idx, "R map row error")
669
 
670
  # m7 -> m8
671
+ elif row[0] == 7 and row[3] == 8:
672
  # Via
673
+ if row[1] == row[4] and row[2] == row[5]:
674
+ x = int(row[1])
675
+ y = int(row[2])
676
+ R_map_78[x, y] = 1
677
+
678
+ # V line
679
+ elif row[1] != row[4]:
680
+ y = int(row[2])
681
+ if row[1] < row[4]:
682
+ R_map_78[int(row[1]) : int(row[4] + 1), y] = 1
683
+ else:
684
+ R_map_78[int(row[4]) : int(row[1] + 1), y] = 1
685
 
 
 
686
  # H line
687
+ elif row[2] != row[5]:
688
+ x = int(row[1])
689
+ if row[2] < row[5]:
690
+ R_map_78[x, int(row[2]) : int(row[5] + 1)] = 1
 
691
  else:
692
+ R_map_78[x, int(row[5]) : int(row[2] + 1)] = 1
693
+ else:
694
+ raise AssertionError(_data_idx, "R map row error")
695
 
696
+ # m8 -> m8
697
+ elif row[0] == 8 and row[3] == 8:
698
  # V line
699
+ if row[1] != row[4]:
700
+ y = int(row[2])
701
+ if row[1] < row[4]:
702
+ R_map_88[int(row[1]) : int(row[4] + 1), y] = 1
703
+ else:
704
+ R_map_88[int(row[4]) : int(row[1] + 1), y] = 1
705
+
706
+ # H line
707
+ elif row[2] != row[5]:
708
+ x = int(row[1])
709
+ if row[2] < row[5]:
710
+ R_map_88[x, int(row[2]) : int(row[5] + 1)] = 1
711
  else:
712
+ R_map_88[x, int(row[5]) : int(row[2] + 1)] = 1
 
713
  else:
714
+ raise AssertionError(_data_idx, "R map row error")
715
 
716
  # m8 -> m9
717
+ elif row[0] == 8 and row[3] == 9:
718
  # Via
719
+ if row[1] == row[4] and row[2] == row[5]:
720
+ x = int(row[1])
721
+ y = int(row[2])
722
+ R_map_89[x, y] = 1
723
+
724
+ # V line
725
+ elif row[1] != row[4]:
726
+ y = int(row[2])
727
+ if row[1] < row[4]:
728
+ R_map_89[int(row[1]) : int(row[4] + 1), y] = 1
729
+ else:
730
+ R_map_89[int(row[4]) : int(row[1] + 1), y] = 1
731
+
732
+ # H line
733
+ elif row[2] != row[5]:
734
+ x = int(row[1])
735
+ if row[2] < row[5]:
736
+ R_map_89[x, int(row[2]) : int(row[5] + 1)] = 1
737
+ else:
738
+ R_map_89[x, int(row[5]) : int(row[2] + 1)] = 1
739
  else:
740
+ raise AssertionError(_data_idx, "R map row error")
741
 
742
  # m9 -> m9
743
+ elif row[0] == 9 and row[3] == 9:
744
+ # V line
745
+ if row[1] != row[4]:
746
+ y = int(row[2])
747
+ if row[1] < row[4]:
748
+ R_map_99[int(row[1]) : int(row[4] + 1), y] = 1
 
749
  else:
750
+ R_map_99[int(row[4]) : int(row[1] + 1), y] = 1
 
751
 
752
+ # H line
753
+ elif row[2] != row[5]:
754
+ x = int(row[1])
755
+ if row[2] < row[5]:
756
+ R_map_99[x, int(row[2]) : int(row[5] + 1)] = 1
 
757
  else:
758
+ R_map_99[x, int(row[5]) : int(row[2] + 1)] = 1
 
759
  else:
760
+ raise AssertionError(_data_idx, "R map row error")
761
  else:
762
+ raise AssertionError(_data_idx, "R map layer not found", row[0], row[3])
763
+
764
+ # clean not connect
765
+ # R_map_99[:, ~V_map.any(axis=0)] = 0
766
+ # R_map_89[:, ~V_map.any(axis=0)] = 0
767
+
768
+ # clean wrong via
769
+ R_map_14[(R_map_11 == 0) | (R_map_44 == 0)] = 0
770
+ R_map_47[(R_map_44 == 0) | (R_map_77 == 0)] = 0
771
+ R_map_78[(R_map_77 == 0) | (R_map_88 == 0)] = 0
772
+ R_map_89[(R_map_88 == 0) | (R_map_99 == 0)] = 0
773
 
774
+
775
  # I
776
+ df_I = df[df["type"].str.contains("I")]
777
+ assert df_I[["x1", "y1"]].astype(int).mod(800).eq(0).all().all(), (_data_idx, "I map error")
778
+ df_I[["x1", "y1"]] = df_I[["x1", "y1"]].astype(int) // 800
779
+ df_I = df_I[["x1", "y1", "values"]].to_numpy()
780
 
781
  I_map = np.zeros((H, W))
782
+ for row in df_I:
783
+ x = int(row[0])
784
+ y = int(row[1])
785
+ I_map[x, y] = row[2]
786
+
787
+ assert (((I_map != 0) & (R_map_11 != 0)) == (I_map != 0)).all(), (_data_idx, np.argwhere(((I_map != 0) & (R_map_11 != 0)) != (I_map != 0)), "I R not connect")
788
 
789
+ # R sim
790
+ if self.config.use_ir_map:
791
+ I_sim_map_14, IR_map_14 = get_IR_map_H(R_map_14, I_map)
792
+ I_sim_map_47, IR_map_47 = get_IR_map_V(R_map_47, I_sim_map_14)
793
+ I_sim_map_78, IR_map_78 = get_IR_map_H(R_map_78, I_sim_map_47)
794
+ I_sim_map_89, IR_map_89 = get_IR_map_V(R_map_89, I_sim_map_78)
795
+
796
 
797
+ if self.config.use_multi_dist:
798
+ cdt = np.ones((H*2 - 1, W*2 - 1))
799
+ cdt[H - 1, W - 1] = 0
800
+ cdt = 1 / (get_dist(self, cdt) + 1e-8)
801
+
802
+ # V dist
803
+ V_multi_dist = np.zeros((H, W))
804
+ for idx in np.argwhere(V_map != 0):
805
+ x = H - idx[0] - 1
806
+ y = W - idx[1] - 1
807
+ V_multi_dist += cdt[x:x+H, y:y+W]
808
+
809
+ # Via dist
810
+ R_multi_dist_14 = np.zeros((H, W))
811
+ R_multi_dist_47 = np.zeros((H, W))
812
+ R_multi_dist_78 = np.zeros((H, W))
813
+ R_multi_dist_89 = np.zeros((H, W))
814
+ for idx in np.argwhere(R_map_14 != 0):
815
+ x = H - idx[0] - 1
816
+ y = W - idx[1] - 1
817
+ R_multi_dist_14 += cdt[x:x+H, y:y+W]
818
+ for idx in np.argwhere(R_map_47 != 0):
819
+ x = H - idx[0] - 1
820
+ y = W - idx[1] - 1
821
+ R_multi_dist_47 += cdt[x:x+H, y:y+W]
822
+ for idx in np.argwhere(R_map_78 != 0):
823
+ x = H - idx[0] - 1
824
+ y = W - idx[1] - 1
825
+ R_multi_dist_78 += cdt[x:x+H, y:y+W]
826
+ for idx in np.argwhere(R_map_89 != 0):
827
+ x = H - idx[0] - 1
828
+ y = W - idx[1] - 1
829
+ R_multi_dist_89 += cdt[x:x+H, y:y+W]
830
+
831
+
832
+ # resize
833
+ R_map_11 = resize(R_map_11) / (H * W) * (_H * _W)
834
+ R_map_14 = resize(R_map_14) / (H * W) * (_H * _W)
835
+ R_map_44 = resize(R_map_44) / (H * W) * (_H * _W)
836
+ R_map_47 = resize(R_map_47) / (H * W) * (_H * _W)
837
+ R_map_77 = resize(R_map_77) / (H * W) * (_H * _W)
838
+ R_map_78 = resize(R_map_78) / (H * W) * (_H * _W)
839
+ R_map_88 = resize(R_map_88) / (H * W) * (_H * _W)
840
+ R_map_89 = resize(R_map_89) / (H * W) * (_H * _W)
841
+ R_map_99 = resize(R_map_99) / (H * W) * (_H * _W)
842
+ I_map = resize(I_map) / (H * W) * (_H * _W)
843
+ V_map = resize(V_map) / (H * W) * (_H * _W)
844
+ if self.config.use_multi_dist:
845
+ V_multi_dist = resize(1 / V_multi_dist) / math.sqrt(H * W) * math.sqrt(_H * _W)
846
+ R_multi_dist_14 = resize(1 / R_multi_dist_14) / math.sqrt(H * W) * math.sqrt(_H * _W)
847
+ R_multi_dist_47 = resize(1 / R_multi_dist_47) / math.sqrt(H * W) * math.sqrt(_H * _W)
848
+ R_multi_dist_78 = resize(1 / R_multi_dist_78) / math.sqrt(H * W) * math.sqrt(_H * _W)
849
+ R_multi_dist_89 = resize(1 / R_multi_dist_89) / math.sqrt(H * W) * math.sqrt(_H * _W)
850
+ if self.config.use_ir_map:
851
+ IR_map_14 = resize(IR_map_14, max_pool=False) / (H * W) * (_H * _W) / (H * W) * (_H * _W)
852
+ IR_map_47 = resize(IR_map_47, max_pool=False) / (H * W) * (_H * _W) / (H * W) * (_H * _W)
853
+ IR_map_78 = resize(IR_map_78, max_pool=False) / (H * W) * (_H * _W) / (H * W) * (_H * _W)
854
+ IR_map_89 = resize(IR_map_89, max_pool=False) / (H * W) * (_H * _W) / (H * W) * (_H * _W)
855
+
856
 
857
+ yield _idx, get_image(self, {
858
+ **{
859
+ "data_idx": _data_idx,
860
+ "H": _H,
861
+ "W": _W,
862
+ "current": resize(genfromtxt(_current, delimiter=",")) * 1e6,
863
+ "pdn_density": resize(genfromtxt(_pdn_density, delimiter=",")),
864
+ "eff_dist": resize(genfromtxt(_eff_dist, delimiter=",")),
865
+ "ir_drop": resize(genfromtxt(_ir_drop, delimiter=",")) * 1e3,
866
+ "R_map_11": R_map_11,
867
+ "R_map_14": R_map_14,
868
+ "R_map_44": R_map_44,
869
+ "R_map_47": R_map_47,
870
+ "R_map_77": R_map_77,
871
+ "R_map_78": R_map_78,
872
+ "R_map_88": R_map_88,
873
+ "R_map_89": R_map_89,
874
+ "R_map_99": R_map_99,
875
+ "I_map": I_map * 1e6,
876
+ "V_map": V_map,
877
+ },
878
+ **({
879
+ "V_multi_dist": V_multi_dist,
880
+ "R_multi_dist_14": R_multi_dist_14,
881
+ "R_multi_dist_47": R_multi_dist_47,
882
+ "R_multi_dist_78": R_multi_dist_78,
883
+ "R_multi_dist_89": R_multi_dist_89,
884
+ } if self.config.use_multi_dist else {}),
885
+ **({
886
+ "IR_map_14": IR_map_14 * 1e3,
887
+ "IR_map_47": IR_map_47 * 1e3,
888
+ "IR_map_78": IR_map_78 * 1e3,
889
+ "IR_map_89": IR_map_89 * 1e3,
890
+ } if self.config.use_ir_map else {})
891
+ })