| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import numpy as np |
| | from PIL import Image |
| |
|
| | import torch |
| |
|
| |
|
| | def convert_img(img): |
| | """Converts (H, W, C) numpy.ndarray to (C, W, H) format""" |
| | if len(img.shape) == 3: |
| | img = img.transpose(2, 0, 1) |
| | if len(img.shape) == 2: |
| | img = np.expand_dims(img, 0) |
| | return img |
| |
|
| |
|
| | class ClipToTensor(object): |
| | """Convert a list of m (H x W x C) numpy.ndarrays in the range [0, 255] |
| | to a torch.FloatTensor of shape (C x m x H x W) in the range [0, 1.0] |
| | """ |
| |
|
| | def __init__(self, channel_nb=3, div_255=True, numpy=False): |
| | self.channel_nb = channel_nb |
| | self.div_255 = div_255 |
| | self.numpy = numpy |
| |
|
| | def __call__(self, clip): |
| | """ |
| | Args: clip (list of numpy.ndarray): clip (list of images) |
| | to be converted to tensor. |
| | """ |
| | |
| | if isinstance(clip[0], np.ndarray): |
| | h, w, ch = clip[0].shape |
| | assert ch == self.channel_nb, "Got {0} instead of 3 channels".format(ch) |
| | elif isinstance(clip[0], Image.Image): |
| | w, h = clip[0].size |
| | else: |
| | raise TypeError( |
| | "Expected numpy.ndarray or PIL.Image\ |
| | but got list of {0}".format( |
| | type(clip[0]) |
| | ) |
| | ) |
| |
|
| | np_clip = np.zeros([self.channel_nb, len(clip), int(h), int(w)]) |
| |
|
| | |
| | for img_idx, img in enumerate(clip): |
| | if isinstance(img, np.ndarray): |
| | pass |
| | elif isinstance(img, Image.Image): |
| | img = np.array(img, copy=False) |
| | else: |
| | raise TypeError( |
| | "Expected numpy.ndarray or PIL.Image\ |
| | but got list of {0}".format( |
| | type(clip[0]) |
| | ) |
| | ) |
| | img = convert_img(img) |
| | np_clip[:, img_idx, :, :] = img |
| | if self.numpy: |
| | if self.div_255: |
| | np_clip = np_clip / 255.0 |
| | return np_clip |
| |
|
| | else: |
| | tensor_clip = torch.from_numpy(np_clip) |
| |
|
| | if not isinstance(tensor_clip, torch.FloatTensor): |
| | tensor_clip = tensor_clip.float() |
| | if self.div_255: |
| | tensor_clip = torch.div(tensor_clip, 255) |
| | return tensor_clip |
| |
|
| |
|
| | |
| | class ClipToTensor_K(object): |
| | """Convert a list of m (H x W x C) numpy.ndarrays in the range [0, 255] |
| | to a torch.FloatTensor of shape (C x m x H x W) in the range [0, 1.0] |
| | """ |
| |
|
| | def __init__(self, channel_nb=3, div_255=True, numpy=False): |
| | self.channel_nb = channel_nb |
| | self.div_255 = div_255 |
| | self.numpy = numpy |
| |
|
| | def __call__(self, clip): |
| | """ |
| | Args: clip (list of numpy.ndarray): clip (list of images) |
| | to be converted to tensor. |
| | """ |
| | |
| | if isinstance(clip[0], np.ndarray): |
| | h, w, ch = clip[0].shape |
| | assert ch == self.channel_nb, "Got {0} instead of 3 channels".format(ch) |
| | elif isinstance(clip[0], Image.Image): |
| | w, h = clip[0].size |
| | else: |
| | raise TypeError( |
| | "Expected numpy.ndarray or PIL.Image\ |
| | but got list of {0}".format( |
| | type(clip[0]) |
| | ) |
| | ) |
| |
|
| | np_clip = np.zeros([self.channel_nb, len(clip), int(h), int(w)]) |
| |
|
| | |
| | for img_idx, img in enumerate(clip): |
| | if isinstance(img, np.ndarray): |
| | pass |
| | elif isinstance(img, Image.Image): |
| | img = np.array(img, copy=False) |
| | else: |
| | raise TypeError( |
| | "Expected numpy.ndarray or PIL.Image\ |
| | but got list of {0}".format( |
| | type(clip[0]) |
| | ) |
| | ) |
| | img = convert_img(img) |
| | np_clip[:, img_idx, :, :] = img |
| | if self.numpy: |
| | if self.div_255: |
| | np_clip = (np_clip - 127.5) / 127.5 |
| | return np_clip |
| |
|
| | else: |
| | tensor_clip = torch.from_numpy(np_clip) |
| |
|
| | if not isinstance(tensor_clip, torch.FloatTensor): |
| | tensor_clip = tensor_clip.float() |
| | if self.div_255: |
| | tensor_clip = torch.div(torch.sub(tensor_clip, 127.5), 127.5) |
| | return tensor_clip |
| |
|
| |
|
| | class ToTensor(object): |
| | """Converts numpy array to tensor""" |
| |
|
| | def __call__(self, array): |
| | tensor = torch.from_numpy(array) |
| | return tensor |
| |
|