repo
stringlengths
1
99
file
stringlengths
13
215
code
stringlengths
12
59.2M
file_length
int64
12
59.2M
avg_line_length
float64
3.82
1.48M
max_line_length
int64
12
2.51M
extension_type
stringclasses
1 value
SleePyCo
SleePyCo-main/train_crl.py
import os import json import argparse import warnings import torch import torch.optim as optim from torch.utils.data import DataLoader from utils import * from loss import SupConLoss from loader import EEGDataLoader from models.main_model import MainModel class OneFoldTrainer: def __init__(self, args, fold, config): self.args = args self.fold = fold self.cfg = config self.tp_cfg = config['training_params'] self.es_cfg = self.tp_cfg['early_stopping'] self.device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') print('[INFO] Config name: {}'.format(config['name'])) self.train_iter = 0 self.model = self.build_model() self.loader_dict = self.build_dataloader() self.criterion = SupConLoss(temperature=self.tp_cfg['temperature']) self.optimizer = optim.Adam(self.model.parameters(), lr=self.tp_cfg['lr'], weight_decay=self.tp_cfg['weight_decay']) self.ckpt_path = os.path.join('checkpoints', config['name']) self.ckpt_name = 'ckpt_fold-{0:02d}.pth'.format(self.fold) self.early_stopping = EarlyStopping(patience=self.es_cfg['patience'], verbose=True, ckpt_path=self.ckpt_path, ckpt_name=self.ckpt_name, mode=self.es_cfg['mode']) def build_model(self): model = MainModel(self.cfg) print('[INFO] Number of params of model: ', sum(p.numel() for p in model.parameters() if p.requires_grad)) model = torch.nn.DataParallel(model, device_ids=list(range(len(self.args.gpu.split(","))))) model.to(self.device) print('[INFO] Model prepared, Device used: {} GPU:{}'.format(self.device, self.args.gpu)) return model def build_dataloader(self): dataloader_args = {'batch_size': self.tp_cfg['batch_size'], 'shuffle': True, 'num_workers': 4*len(self.args.gpu.split(",")), 'pin_memory': True} train_dataset = EEGDataLoader(self.cfg, self.fold, set='train') train_loader = DataLoader(dataset=train_dataset, **dataloader_args) val_dataset = EEGDataLoader(self.cfg, self.fold, set='val') val_loader = DataLoader(dataset=val_dataset, **dataloader_args) print('[INFO] Dataloader prepared') return {'train': train_loader, 'val': val_loader} def train_one_epoch(self): self.model.train() train_loss = 0 for i, (inputs, labels) in enumerate(self.loader_dict['train']): loss = 0 labels = labels.view(-1).to(self.device) inputs = torch.cat([inputs[0], inputs[1]], dim=0).to(self.device) outputs = self.model(inputs)[0] f1, f2 = torch.split(outputs, [labels.size(0), labels.size(0)], dim=0) features = torch.cat([f1.unsqueeze(1), f2.unsqueeze(1)], dim=1) loss += self.criterion(features, labels) self.optimizer.zero_grad() loss.backward() self.optimizer.step() train_loss += loss.item() self.train_iter += 1 progress_bar(i, len(self.loader_dict['train']), 'Lr: %.4e | Loss: %.3f' %(get_lr(self.optimizer), train_loss / (i + 1))) if self.train_iter % self.tp_cfg['val_period'] == 0: print('') val_loss = self.evaluate(mode='val') self.early_stopping(None, val_loss, self.model) self.model.train() if self.early_stopping.early_stop: break @torch.no_grad() def evaluate(self, mode): self.model.eval() eval_loss = 0 for i, (inputs, labels) in enumerate(self.loader_dict[mode]): loss = 0 inputs = inputs.to(self.device) labels = labels.view(-1).to(self.device) outputs = self.model(inputs)[0] features = outputs.unsqueeze(1).repeat(1, 2, 1) loss += self.criterion(features, labels) eval_loss += loss.item() progress_bar(i, len(self.loader_dict[mode]), 'Lr: %.4e | Loss: %.3f' %(get_lr(self.optimizer), eval_loss / (i + 1))) return eval_loss def run(self): for epoch in range(self.tp_cfg['max_epochs']): print('\n[INFO] Fold: {}, Epoch: {}'.format(self.fold, epoch)) self.train_one_epoch() if self.early_stopping.early_stop: break def main(): warnings.filterwarnings("ignore", category=DeprecationWarning) warnings.filterwarnings("ignore", category=UserWarning) parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add_argument('--seed', type=int, default=42, help='random seed') parser.add_argument('--gpu', type=str, default="0", help='gpu id') parser.add_argument('--config', type=str, help='config file path') args = parser.parse_args() os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID" os.environ["CUDA_VISIBLE_DEVICES"] = args.gpu # For reproducibility set_random_seed(args.seed, use_cuda=True) with open(args.config) as config_file: config = json.load(config_file) config['name'] = os.path.basename(args.config).replace('.json', '') for fold in range(1, config['dataset']['num_splits'] + 1): trainer = OneFoldTrainer(args, fold, config) trainer.run() if __name__ == "__main__": main()
5,454
36.62069
169
py
SleePyCo
SleePyCo-main/transform.py
import torch import random import numpy as np from scipy import signal from scipy.ndimage.interpolation import shift class TwoTransform: def __init__(self, transform): self.transform = transform def __call__(self, x): return [self.transform(x), self.transform(x)] class Compose: def __init__(self, transforms, mode='full'): self.transforms = transforms self.mode = mode def __call__(self, x): if self.mode == 'random': index = random.randint(0, len(self.transforms) - 1) x = self.transforms[index](x) elif self.mode == 'full': for t in self.transforms: x = t(x) elif self.mode == 'shuffle': transforms = np.random.choice(self.transforms, len(self.transforms), replace=False) for t in transforms: x = t(x) else: raise NotImplementedError return x def __repr__(self): format_string = self.__class__.__name__ + '(' for t in self.transforms: format_string += '\n' format_string += ' {0}'.format(t) format_string += '\n)' return format_string class RandomAmplitudeScale: def __init__(self, range=(0.5, 2.0), p=0.5): self.range = range self.p = p def __call__(self, x): if torch.rand(1) < self.p: scale = random.uniform(self.range[0], self.range[1]) return x * scale return x def __repr__(self): return self.__class__.__name__ + '()' class RandomDCShift: def __init__(self, range=(-10.0, 10.0), p=0.5): self.range = range self.p = p def __call__(self, x): if torch.rand(1) < self.p: shift = random.uniform(self.range[0], self.range[1]) return x + shift return x def __repr__(self): return self.__class__.__name__ + '()' class RandomTimeShift: def __init__(self, range=(-300, 300), mode='constant', cval=0.0, p=0.5): self.range = range self.mode = mode self.cval = cval self.p = p def __call__(self, x): if torch.rand(1) < self.p: t_shift = random.randint(self.range[0], self.range[1]) if len(x.shape) == 2: x = x[0] x = shift(input=x, shift=t_shift, mode=self.mode, cval=self.cval) x = np.expand_dims(x, axis=0) return x def __repr__(self): return self.__class__.__name__ + '()' class RandomZeroMasking: def __init__(self, range=(0, 300), p=0.5): self.range = range self.p = p def __call__(self, x): if torch.rand(1) < self.p: mask_len = random.randint(self.range[0], self.range[1]) random_pos = random.randint(0, x.shape[1] - mask_len) mask = np.concatenate([np.ones((1, random_pos)), np.zeros((1, mask_len)), np.ones((1, x.shape[1] - mask_len - random_pos))], axis=1) return x * mask return x def __repr__(self): return self.__class__.__name__ + '()' class RandomAdditiveGaussianNoise: def __init__(self, range=(0.0, 0.2), p=0.5): self.range = range self.p = p def __call__(self, x): if torch.rand(1) < self.p: sigma = random.uniform(self.range[0], self.range[1]) return x + np.random.normal(0, sigma, x.shape) return x def __repr__(self): return self.__class__.__name__ + '()' class RandomBandStopFilter: def __init__(self, range=(0.5, 30.0), band_width=2.0, sampling_rate=100.0, p=0.5): self.range = range self.band_width = band_width self.sampling_rate = sampling_rate self.p = p def __call__(self, x): if torch.rand(1) < self.p: low_freq = random.uniform(self.range[0], self.range[1]) center_freq = low_freq + self.band_width / 2.0 b, a = signal.iirnotch(center_freq, center_freq / self.band_width, fs=self.sampling_rate) x = signal.lfilter(b, a, x) return x def __repr__(self): return self.__class__.__name__ + '()'
4,204
26.48366
144
py
SleePyCo
SleePyCo-main/loss.py
import torch import torch.nn as nn class SupConLoss(nn.Module): """Supervised Contrastive Learning: https://arxiv.org/pdf/2004.11362.pdf. It also supports the unsupervised contrastive loss in SimCLR""" def __init__(self, temperature=0.07, contrast_mode='all', base_temperature=0.07): super(SupConLoss, self).__init__() self.temperature = temperature self.contrast_mode = contrast_mode self.base_temperature = base_temperature def forward(self, features, labels=None, mask=None): """Compute loss for model. If both `labels` and `mask` are None, it degenerates to SimCLR unsupervised loss: https://arxiv.org/pdf/2002.05709.pdf Args: features: hidden vector of shape [bsz, n_views, ...]. labels: ground truth of shape [bsz]. mask: contrastive mask of shape [bsz, bsz], mask_{i,j}=1 if sample j has the same class as sample i. Can be asymmetric. Returns: A loss scalar. """ device = (torch.device('cuda') if features.is_cuda else torch.device('cpu')) if len(features.shape) < 3: raise ValueError('`features` needs to be [bsz, n_views, ...],' 'at least 3 dimensions are required') if len(features.shape) > 3: features = features.view(features.shape[0], features.shape[1], -1) batch_size = features.shape[0] if labels is not None and mask is not None: raise ValueError('Cannot define both `labels` and `mask`') elif labels is None and mask is None: mask = torch.eye(batch_size, dtype=torch.float32).to(device) elif labels is not None: labels = labels.contiguous().view(-1, 1) if labels.shape[0] != batch_size: raise ValueError('Num of labels does not match num of features') mask = torch.eq(labels, labels.T).float().to(device) else: mask = mask.float().to(device) contrast_count = features.shape[1] contrast_feature = torch.cat(torch.unbind(features, dim=1), dim=0) if self.contrast_mode == 'one': anchor_feature = features[:, 0] anchor_count = 1 elif self.contrast_mode == 'all': anchor_feature = contrast_feature anchor_count = contrast_count else: raise ValueError('Unknown mode: {}'.format(self.contrast_mode)) # compute logits anchor_dot_contrast = torch.div( torch.matmul(anchor_feature, contrast_feature.T), self.temperature) # for numerical stability logits_max, _ = torch.max(anchor_dot_contrast, dim=1, keepdim=True) logits = anchor_dot_contrast - logits_max.detach() # tile mask mask = mask.repeat(anchor_count, contrast_count) # mask-out self-contrast cases logits_mask = torch.scatter( torch.ones_like(mask), 1, torch.arange(batch_size * anchor_count).view(-1, 1).to(device), 0 ) mask = mask * logits_mask # compute log_prob exp_logits = torch.exp(logits) * logits_mask log_prob = logits - torch.log(exp_logits.sum(1, keepdim=True)) # compute mean of log-likelihood over positive mean_log_prob_pos = (mask * log_prob).sum(1) / mask.sum(1) # loss loss = - (self.temperature / self.base_temperature) * mean_log_prob_pos loss = loss.view(anchor_count, batch_size).mean() return loss
3,650
38.684783
80
py
SleePyCo
SleePyCo-main/utils.py
import os import sys import math import time import torch import random import numpy as np import sklearn.metrics as skmet from terminaltables import SingleTable from termcolor import colored _, term_width = os.popen('stty size', 'r').read().split() term_width = int(term_width) TOTAL_BAR_LENGTH = 25. last_time = time.time() begin_time = last_time def progress_bar(current, total, msg=None): global last_time, begin_time if current == 0: begin_time = time.time() # Reset for new bar. cur_len = int(TOTAL_BAR_LENGTH*current/total) rest_len = int(TOTAL_BAR_LENGTH - cur_len) - 1 sys.stdout.write(' [') for i in range(cur_len): sys.stdout.write('=') sys.stdout.write('>') for i in range(rest_len): sys.stdout.write('.') sys.stdout.write(']') cur_time = time.time() step_time = cur_time - last_time last_time = cur_time tot_time = cur_time - begin_time L = [] L.append(' Step: %s' % format_time(step_time)) L.append(' | Tot: %s' % format_time(tot_time)) if msg: L.append(' | ' + msg) msg = ''.join(L) sys.stdout.write(msg) for i in range(term_width-int(TOTAL_BAR_LENGTH)-len(msg)-3): sys.stdout.write(' ') # Go back to the center of the bar. for i in range(term_width-int(TOTAL_BAR_LENGTH/2)+2): sys.stdout.write('\b') sys.stdout.write(' %d/%d ' % (current+1, total)) if current < total-1: sys.stdout.write('\r') else: sys.stdout.write('\n') sys.stdout.flush() def format_time(seconds): days = int(seconds / 3600/24) seconds = seconds - days*3600*24 hours = int(seconds / 3600) seconds = seconds - hours*3600 minutes = int(seconds / 60) seconds = seconds - minutes*60 secondsf = int(seconds) seconds = seconds - secondsf millis = int(seconds*1000) f = '' i = 1 if days > 0: f += str(days) + 'D' i += 1 if hours > 0 and i <= 2: f += str(hours) + 'h' i += 1 if minutes > 0 and i <= 2: f += str(minutes) + 'm' i += 1 if secondsf > 0 and i <= 2: f += str(secondsf) + 's' i += 1 if millis > 0 and i <= 2: f += str(millis) + 'ms' i += 1 if f == '': f = '0ms' return f def get_lr(optimizer): for param_group in optimizer.param_groups: return param_group['lr'] class EarlyStopping: """Early stops the training if validation loss doesn't improve after a given patience.""" def __init__(self, patience=7, verbose=False, delta=0, ckpt_path='./checkpoints', ckpt_name='checkpoint.pth', mode='min'): """ Args: patience (int): How long to wait after last time validation loss improved. Default: 7 verbose (bool): If True, prints a message for each validation loss improvement. Default: False delta (float): Minimum change in the monitored quantity to qualify as an improvement. Default: 0 path (str): Path for the checkpoint to be saved to. Default: 'checkpoint.pt' """ self.patience = patience self.verbose = verbose self.counter = 0 self.best_score = None self.early_stop = False self.mode = mode if mode == 'max': self.init_metric = 0 elif mode == 'min': self.init_metric = -np.inf else: raise NotImplementedError self.delta = delta self.ckpt_path = ckpt_path self.ckpt_name = ckpt_name if '.pth' in ckpt_name else ckpt_name + '.pth' os.makedirs(self.ckpt_path, exist_ok=True) def __call__(self, val_acc, val_loss, model): if self.mode == 'max': score = val_acc val_metric = val_acc elif self.mode == 'min': score = -val_loss val_metric = val_loss else: raise NotImplementedError if self.best_score is None: self.best_score = score self.save_checkpoint(val_metric, model) elif score < self.best_score + self.delta: self.counter += 1 print(f'EarlyStopping counter: {self.counter} out of {self.patience}\n') if self.counter >= self.patience: self.early_stop = True else: self.best_score = score self.save_checkpoint(val_metric, model) self.counter = 0 def save_checkpoint(self, val_metric, model): '''Saves model when validation loss decrease.''' if self.verbose: if self.mode == 'max': print(f'[INFO] Validation accuracy increased ({self.init_metric:.6f} --> {val_metric:.6f}). Saving model ...\n') elif self.mode == 'min': print(f'[INFO] Validation loss decreased ({self.init_metric:.6f} --> {val_metric:.6f}). Saving model ...\n') else: raise NotImplementedError torch.save(model.state_dict(), os.path.join(self.ckpt_path, self.ckpt_name)) self.init_metric = val_metric def summarize_result(config, fold, y_true, y_pred, save=True): os.makedirs('results', exist_ok=True) y_pred_argmax = np.argmax(y_pred, 1) result_dict = skmet.classification_report(y_true, y_pred_argmax, digits=3, output_dict=True) cm = skmet.confusion_matrix(y_true, y_pred_argmax) accuracy = round(result_dict['accuracy']*100, 1) macro_f1 = round(result_dict['macro avg']['f1-score']*100, 1) kappa = round(skmet.cohen_kappa_score(y_true, y_pred_argmax), 3) wpr = round(result_dict['0.0']['precision']*100, 1) wre = round(result_dict['0.0']['recall']*100, 1) wf1 = round(result_dict['0.0']['f1-score']*100, 1) n1pr = round(result_dict['1.0']['precision']*100, 1) n1re = round(result_dict['1.0']['recall']*100, 1) n1f1 = round(result_dict['1.0']['f1-score']*100, 1) n2pr = round(result_dict['2.0']['precision']*100, 1) n2re = round(result_dict['2.0']['recall']*100, 1) n2f1 = round(result_dict['2.0']['f1-score']*100, 1) n3pr = round(result_dict['3.0']['precision']*100, 1) n3re = round(result_dict['3.0']['recall']*100, 1) n3f1 = round(result_dict['3.0']['f1-score']*100, 1) rpr = round(result_dict['4.0']['precision']*100, 1) rre = round(result_dict['4.0']['recall']*100, 1) rf1 = round(result_dict['4.0']['f1-score']*100, 1) overall_data = [ ['ACC', 'MF1', '\u03BA'], [accuracy, macro_f1, kappa], ] perclass_data = [ [colored('A', 'cyan') + '\\' + colored('P', 'green'), 'W', 'N1', 'N2', 'N3', 'R', 'PR', 'RE', 'F1'], ['W', cm[0][0], cm[0][1], cm[0][2], cm[0][3], cm[0][4], wpr, wre, wf1], ['N1', cm[1][0], cm[1][1], cm[1][2], cm[1][3], cm[1][4], n1pr, n1re, n1f1], ['N2', cm[2][0], cm[2][1], cm[2][2], cm[2][3], cm[2][4], n2pr, n2re, n2f1], ['N3', cm[3][0], cm[3][1], cm[3][2], cm[3][3], cm[3][4], n3pr, n3re, n3f1], ['R', cm[4][0], cm[4][1], cm[4][2], cm[4][3], cm[4][4], rpr, rre, rf1], ] overall_dt = SingleTable(overall_data, colored('OVERALL RESULT', 'red')) perclass_dt = SingleTable(perclass_data, colored('PER-CLASS RESULT', 'red')) print('\n[INFO] Evaluation result from fold 1 to {}'.format(fold)) print('\n' + overall_dt.table) print('\n' + perclass_dt.table) print(colored(' A', 'cyan') + ': Actual Class, ' + colored('P', 'green') + ': Predicted Class' + '\n\n') if save: with open(os.path.join('results', config['name'] + '.txt'), 'w') as f: f.write( str(fold) + ' ' + str(round(result_dict['accuracy']*100, 1)) + ' ' + str(round(result_dict['macro avg']['f1-score']*100, 1)) + ' ' + str(round(kappa, 3)) + ' ' + str(round(result_dict['0.0']['f1-score']*100, 1)) + ' ' + str(round(result_dict['1.0']['f1-score']*100, 1)) + ' ' + str(round(result_dict['2.0']['f1-score']*100, 1)) + ' ' + str(round(result_dict['3.0']['f1-score']*100, 1)) + ' ' + str(round(result_dict['4.0']['f1-score']*100, 1)) + ' ' ) def set_random_seed(seed_value, use_cuda=True): np.random.seed(seed_value) # cpu vars torch.manual_seed(seed_value) # cpu vars random.seed(seed_value) # Python os.environ['PYTHONHASHSEED'] = str(seed_value) # Python hash buildin if use_cuda: torch.cuda.manual_seed(seed_value) torch.cuda.manual_seed_all(seed_value) # gpu vars torch.backends.cudnn.deterministic = True #needed torch.backends.cudnn.benchmark = False
8,844
34.239044
129
py
SleePyCo
SleePyCo-main/loader.py
import os import glob import torch import numpy as np from transform import * from torch.utils.data import Dataset class EEGDataLoader(Dataset): def __init__(self, config, fold, set='train'): self.set = set self.fold = fold self.sr = 100 self.dset_cfg = config['dataset'] self.root_dir = self.dset_cfg['root_dir'] self.dset_name = self.dset_cfg['name'] self.num_splits = self.dset_cfg['num_splits'] self.eeg_channel = self.dset_cfg['eeg_channel'] self.seq_len = self.dset_cfg['seq_len'] self.target_idx = self.dset_cfg['target_idx'] self.training_mode = config['training_params']['mode'] self.dataset_path = os.path.join(self.root_dir, 'dset', self.dset_name, 'npz') self.inputs, self.labels, self.epochs = self.split_dataset() if self.training_mode == 'pretrain': self.transform = Compose( transforms=[ RandomAmplitudeScale(), RandomTimeShift(), RandomDCShift(), RandomZeroMasking(), RandomAdditiveGaussianNoise(), RandomBandStopFilter(), ] ) self.two_transform = TwoTransform(self.transform) def __len__(self): return len(self.epochs) def __getitem__(self, idx): n_sample = 30 * self.sr * self.seq_len file_idx, idx, seq_len = self.epochs[idx] inputs = self.inputs[file_idx][idx:idx+seq_len] if self.set == 'train': if self.training_mode == 'pretrain': assert seq_len == 1 input_a, input_b = self.two_transform(inputs) input_a = torch.from_numpy(input_a).float() input_b = torch.from_numpy(input_b).float() inputs = [input_a, input_b] elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: inputs = inputs.reshape(1, n_sample) inputs = torch.from_numpy(inputs).float() else: raise NotImplementedError else: if not self.training_mode == 'pretrain': inputs = inputs.reshape(1, n_sample) inputs = torch.from_numpy(inputs).float() labels = self.labels[file_idx][idx:idx+seq_len] labels = torch.from_numpy(labels).long() labels = labels[self.target_idx] return inputs, labels def split_dataset(self): file_idx = 0 inputs, labels, epochs = [], [], [] data_root = os.path.join(self.dataset_path, self.eeg_channel) data_fname_list = [os.path.basename(x) for x in sorted(glob.glob(os.path.join(data_root, '*.npz')))] data_fname_dict = {'train': [], 'test': [], 'val': []} split_idx_list = np.load(os.path.join('./split_idx', 'idx_{}.npy'.format(self.dset_name)), allow_pickle=True) assert len(split_idx_list) == self.num_splits if self.dset_name == 'Sleep-EDF-2013': for i in range(len(data_fname_list)): subject_idx = int(data_fname_list[i][3:5]) if subject_idx == self.fold - 1: data_fname_dict['test'].append(data_fname_list[i]) elif subject_idx in split_idx_list[self.fold - 1]: data_fname_dict['val'].append(data_fname_list[i]) else: data_fname_dict['train'].append(data_fname_list[i]) elif self.dset_name == 'Sleep-EDF-2018': for i in range(len(data_fname_list)): subject_idx = int(data_fname_list[i][3:5]) if subject_idx in split_idx_list[self.fold - 1][self.set]: data_fname_dict[self.set].append(data_fname_list[i]) elif self.dset_name == 'MASS' or self.dset_name == 'Physio2018' or self.dset_name == 'SHHS': for i in range(len(data_fname_list)): if i in split_idx_list[self.fold - 1][self.set]: data_fname_dict[self.set].append(data_fname_list[i]) else: raise NameError("dataset '{}' cannot be found.".format(self.dataset)) for data_fname in data_fname_dict[self.set]: npz_file = np.load(os.path.join(data_root, data_fname)) inputs.append(npz_file['x']) labels.append(npz_file['y']) seq_len = self.seq_len if self.dset_name== 'MASS' and ('-02-' in data_fname or '-04-' in data_fname or '-05-' in data_fname): seq_len = int(self.seq_len * 1.5) for i in range(len(npz_file['y']) - seq_len + 1): epochs.append([file_idx, i, seq_len]) file_idx += 1 return inputs, labels, epochs
4,913
39.278689
117
py
SleePyCo
SleePyCo-main/models/iitnet.py
import torch.nn as nn def conv3(in_planes, out_planes, stride=1): return nn.Conv1d(in_planes, out_planes, kernel_size=3, stride=stride, padding=1, bias=False) class Bottleneck(nn.Module): expansion = 4 def __init__(self, inplanes, planes, stride=1, downsample=None): super(Bottleneck, self).__init__() self.conv1 = nn.Conv1d(inplanes, planes, kernel_size=1, bias=False) self.bn1 = nn.BatchNorm1d(planes) self.conv2 = nn.Conv1d(planes, planes, kernel_size=3, stride=stride, padding=1, bias=False) self.bn2 = nn.BatchNorm1d(planes) self.conv3 = nn.Conv1d(planes, planes * self.expansion, kernel_size=1, bias=False) self.bn3 = nn.BatchNorm1d(planes * self.expansion) self.relu = nn.ReLU(inplace=True) self.downsample = downsample self.stride = stride def forward(self, x): residual = x out = self.conv1(x) out = self.bn1(out) out = self.relu(out) out = self.conv2(out) out = self.bn2(out) out = self.relu(out) out = self.conv3(out) out = self.bn3(out) if self.downsample is not None: residual = self.downsample(x) out += residual out = self.relu(out) return out class IITNetBackbone(nn.Module): def __init__(self, config): super(IITNetBackbone, self).__init__() block = Bottleneck self.training_mode = config['training_params']['mode'] self.inplanes = 16 self.layers = [3, 4, 6, 3] self.initial_layer = nn.Sequential( nn.Conv1d(1, 16, 7, 2, 3, bias=False), nn.BatchNorm1d(16), nn.ReLU(), nn.MaxPool1d(3, 2, 1)) self.layer1 = self._make_layer(block, 16, self.layers[0], stride=1, first=True) self.layer2 = self._make_layer(block, 16, self.layers[1], stride=2) self.layer3 = self._make_layer(block, 32, self.layers[2], stride=2) self.layer4 = self._make_layer(block, 32, self.layers[3], stride=2) self.maxpool = nn.MaxPool1d(3, 2, 1) if self.training_mode == 'freezefinetune': self.fp_dim = config['feature_pyramid']['dim'] self.num_scales = config['feature_pyramid']['num_scales'] self.conv_c5 = nn.Conv1d(128, self.fp_dim, 1, 1, 0) assert self.num_scales == 1 if config['backbone']['init_weights']: self._initialize_weights() def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.BatchNorm1d): nn.init.constant_(m.weight, 1) nn.init.constant_(m.bias, 0) def _make_layer(self, block, planes, blocks, stride=1, first=False): downsample = None if (stride != 1 and first is False) or self.inplanes != planes * block.expansion: downsample = nn.Sequential( nn.Conv1d(self.inplanes, planes * block.expansion, 1, stride, bias=False), nn.BatchNorm1d(planes * block.expansion) ) layers = [] layers.append(block(self.inplanes, planes, stride, downsample)) self.inplanes = planes * block.expansion for _ in range(1, blocks): layers.append(block(self.inplanes, planes)) return nn.Sequential(*layers) def forward(self, x): out = [] c1 = self.initial_layer(x) c2 = self.layer1(c1) c3 = self.layer2(c2) c4 = self.layer3(self.maxpool(c3)) c5 = self.layer4(c4) if self.training_mode == 'pretrain': out.append(c5) elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: p5 = self.conv_c5(c5) out.append(p5) return out
4,111
31.896
96
py
SleePyCo
SleePyCo-main/models/xsleepnet.py
import torch.nn as nn class XSleepNetFeature(nn.Module): def __init__(self, config): super(XSleepNetFeature, self).__init__() self.training_mode = config['training_params']['mode'] # architecture self.conv1 = self.make_layers(1, 16) self.conv2 = self.make_layers(16, 16) self.conv3 = self.make_layers(16, 32) self.conv4 = self.make_layers(32, 32) self.conv5 = self.make_layers(32, 64) self.conv6 = self.make_layers(64, 64) self.conv7 = self.make_layers(64, 128) self.conv8 = self.make_layers(128, 128) self.conv9 = self.make_layers(128, 256) if self.training_mode == 'freezefinetune': self.fp_dim = config['feature_pyramid']['dim'] self.num_scales = config['feature_pyramid']['num_scales'] self.conv_c5 = nn.Conv1d(256, self.fp_dim, 1, 1, 0) if self.num_scales > 1: self.conv_c4 = nn.Conv1d(128, self.fp_dim, 1, 1, 0) if self.num_scales > 2: self.conv_c3 = nn.Conv1d(128, self.fp_dim, 1, 1, 0) if config['backbone']['init_weights']: self._initialize_weights() def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.BatchNorm1d): nn.init.constant_(m.weight, 1) nn.init.constant_(m.bias, 0) def make_layers(self, in_channels, out_channels): layer = [nn.Conv1d(in_channels, out_channels, 31, 2, 15)] layer.append(nn.BatchNorm1d(out_channels)) layer.append(nn.PReLU()) return nn.Sequential(*layer) def forward(self, x): out = [] x = self.conv1(x) x = self.conv2(x) x = self.conv3(x) x = self.conv4(x) c1 = self.conv5(x) c2 = self.conv6(c1) c3 = self.conv7(c2) c4 = self.conv8(c3) c5 = self.conv9(c4) if self.training_mode == 'pretrain': out.append(c5) elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: p5 = self.conv_c5(c5) out.append(p5) if self.num_scales > 1: p4 = self.conv_c4(c4) out.append(p4) if self.num_scales > 2: p3 = self.conv_c3(c3) out.append(p3) return out
2,631
32.74359
86
py
SleePyCo
SleePyCo-main/models/utils.py
import torch.utils.data from torch.nn import functional as F import math import torch import torch.nn as nn from torch.nn.parameter import Parameter from torch.nn.functional import pad from torch.nn.modules import Module from torch.nn.modules.utils import _single, _pair, _triple class _ConvNd(Module): def __init__(self, in_channels, out_channels, kernel_size, stride, padding, dilation, transposed, output_padding, groups, bias, weight=None): super(_ConvNd, self).__init__() if in_channels % groups != 0: raise ValueError('in_channels must be divisible by groups') if out_channels % groups != 0: raise ValueError('out_channels must be divisible by groups') self.in_channels = in_channels self.out_channels = out_channels self.kernel_size = kernel_size self.stride = stride self.padding = padding self.dilation = dilation self.transposed = transposed self.output_padding = output_padding self.groups = groups if transposed: self.weight = Parameter(torch.Tensor( in_channels, out_channels // groups, *kernel_size)) else: self.weight = Parameter(torch.Tensor( out_channels, in_channels // groups, *kernel_size)) if bias: self.bias = Parameter(torch.Tensor(out_channels)) else: self.register_parameter('bias', None) self.reset_parameters(weight) def reset_parameters(self, weight): if weight == None: n = self.in_channels for k in self.kernel_size: n *= k stdv = 1. / math.sqrt(n) self.weight.data.uniform_(-stdv, stdv) else: self.weight.data = torch.FloatTensor(weight) if self.bias is not None: self.bias.data.uniform_(-stdv, stdv) def __repr__(self): s = ('{name}({in_channels}, {out_channels}, kernel_size={kernel_size}' ', stride={stride}') if self.padding != (0,) * len(self.padding): s += ', padding={padding}' if self.dilation != (1,) * len(self.dilation): s += ', dilation={dilation}' if self.output_padding != (0,) * len(self.output_padding): s += ', output_padding={output_padding}' if self.groups != 1: s += ', groups={groups}' if self.bias is None: s += ', bias=False' s += ')' return s.format(name=self.__class__.__name__, **self.__dict__) class Conv1d(_ConvNd): def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding='VALID', dilation=1, groups=1, bias=True, weight=None): kernel_size = _single(kernel_size) stride = _single(stride) padding = _single(padding) dilation = _single(dilation) super(Conv1d, self).__init__( in_channels, out_channels, kernel_size, stride, padding, dilation, False, _pair(0), groups, bias, weight) def forward(self, input): return conv1d_same_padding(input, self.weight, self.bias, self.stride, self.padding, self.dilation, self.groups) # custom con2d, because pytorch don't have "padding='same'" option. def conv1d_same_padding(input, weight, bias=None, stride=1, padding='VALID', dilation=1, groups=1): def check_format(*argv): argv_format = [] for i in range(len(argv)): if type(argv[i]) is int: argv_format.append((argv[i], argv[i])) elif hasattr(argv[i], "__getitem__"): argv_format.append(tuple(argv[i])) else: raise TypeError('all input should be int or list-type, now is {}'.format(argv[i])) return argv_format stride, dilation = check_format(stride, dilation) if padding == 'SAME': padding = 0 input_rows = input.size(2) filter_rows = weight.size(2) out_rows = (input_rows + stride[0] - 1) // stride[0] padding_rows = max(0, (out_rows - 1) * stride[0] + (filter_rows - 1) * dilation[0] + 1 - input_rows) rows_odd = padding_rows % 2 # input_cols = input.size(3) # filter_cols = weight.size(3) # out_cols = (input_cols + stride[1] - 1) // stride[1] # padding_cols = max(0, (out_cols - 1) * stride[1] + # (filter_cols - 1) * dilation[1] + 1 - input_cols) # cols_odd = padding_cols % 2 input = pad(input, [padding_rows // 2, padding_rows // 2 + int(rows_odd)]) elif padding == 'VALID': padding = 0 elif type(padding) != int: raise ValueError('Padding should be SAME, VALID or specific integer, but not {}.'.format(padding)) return F.conv1d(input, weight, bias, stride=stride, padding=padding, dilation=dilation, groups=groups) class _MaxPoolNd(Module): def __init__(self, kernel_size, stride=None, padding='VALID', dilation=1, return_indices=False, ceil_mode=False): super(_MaxPoolNd, self).__init__() self.kernel_size = kernel_size self.stride = stride or kernel_size self.padding = padding self.dilation = dilation self.return_indices = return_indices self.ceil_mode = ceil_mode def extra_repr(self): return 'kernel_size={kernel_size}, stride={stride}, padding={padding}' \ ', dilation={dilation}, ceil_mode={ceil_mode}'.format(**self.__dict__) class MaxPool1d(_MaxPoolNd): def forward(self, input): return maxpool1d_same_padding(input, self.kernel_size, self.stride, self.padding, self.dilation, self.return_indices, self.ceil_mode) def extra_repr(self): return 'kernel_size={kernel_size}, stride={stride}, padding={padding}' \ ', dilation={dilation}, ceil_mode={ceil_mode}'.format(**self.__dict__) def maxpool1d_same_padding(input, kernel_size, stride=None, padding='VALID', dilation=1, return_indices=False, ceil_mode=False): if stride is None: _stride, dilation = [kernel_size], [dilation] else: _stride, dilation = [stride], [dilation] if padding == 'SAME': padding = 0 input_rows = input.size(2) filter_rows = kernel_size out_rows = (input_rows + _stride[0] - 1) // _stride[0] padding_rows = max(0, (out_rows - 1) * _stride[0] + (filter_rows - 1) * dilation[0] + 1 - input_rows) rows_odd = padding_rows % 2 # input_cols = input.size(3) # filter_cols = weight.size(3) # out_cols = (input_cols + _stride[1] - 1) // _stride[1] # padding_cols = max(0, (out_cols - 1) * _stride[1] + # (filter_cols - 1) * dilation[1] + 1 - input_cols) # cols_odd = padding_cols % 2 input = pad(input, [padding_rows // 2, padding_rows // 2 + int(rows_odd)]) elif padding == 'VALID': padding = 0 elif type(padding) != int: raise ValueError('Padding should be SAME, VALID or specific integer, but not {}.'.format(padding)) return F.max_pool1d(input, kernel_size, stride, padding, dilation, return_indices, ceil_mode)
7,340
37.434555
128
py
SleePyCo
SleePyCo-main/models/main_model.py
import torch.nn as nn import torch.nn.functional as F from .sleepyco import SleePyCoBackbone from .xsleepnet import XSleepNetFeature from .iitnet import IITNetBackbone from .utime import UTimeEncoder from .deepsleepnet import DeepSleepNetFeature from .classifiers import get_classifier last_chn_dict = { 'SleePyCo': 256, 'XSleepNet': 256, 'IITNet': 128, 'UTime': 256, 'DeepSleepNet': 128 } class MainModel(nn.Module): def __init__(self, config): super(MainModel, self).__init__() self.cfg = config self.bb_cfg = config['backbone'] self.training_mode = config['training_params']['mode'] if self.bb_cfg['name'] == 'SleePyCo': self.feature = SleePyCoBackbone(self.cfg) elif self.bb_cfg['name'] == 'XSleepNet': self.feature = XSleepNetFeature(self.cfg) elif self.bb_cfg['name'] == 'UTime': self.feature = UTimeEncoder(self.cfg) elif self.bb_cfg['name'] == 'IITNet': self.feature = IITNetBackbone(self.cfg) elif self.bb_cfg['name'] == 'DeepSleepNet': self.feature = DeepSleepNetFeature(self.cfg) else: raise NotImplementedError('backbone not supported: {}'.format(config['backbone']['name'])) if self.bb_cfg['dropout']: self.dropout = nn.Dropout(p=0.5) if self.training_mode == 'pretrain': proj_dim = self.cfg['proj_head']['dim'] if config['proj_head']['name'] == 'Linear': self.head = nn.Sequential( nn.AdaptiveAvgPool1d(1), nn.Flatten(), nn.Linear(last_chn_dict[config['backbone']['name']], proj_dim) ) elif config['proj_head']['name'] == 'MLP': self.head = nn.Sequential( nn.AdaptiveAvgPool1d(1), nn.Flatten(), nn.Linear(last_chn_dict[config['backbone']['name']], proj_dim), nn.ReLU(inplace=True), nn.Linear(proj_dim, proj_dim) ) else: raise NotImplementedError('head not supported: {}'.format(config['proj_head']['name'])) print('[INFO] Number of params of backbone: ', sum(p.numel() for p in self.feature.parameters() if p.requires_grad)) print('[INFO] Number of params of proj_head: ', sum(p.numel() for p in self.head.parameters() if p.requires_grad)) elif self.training_mode in ['scratch', 'fullfinetune', 'freezefinetune']: self.classifier = get_classifier(config) print('[INFO] Number of params of backbone: ', sum(p.numel() for p in self.feature.parameters() if p.requires_grad)) print('[INFO] Number of params of classifier: ', sum(p.numel() for p in self.classifier.parameters() if p.requires_grad)) else: raise NotImplementedError('head not supported: {}'.format(config['training_params']['mode'])) def get_max_len(self, features): len_list = [] for feature in features: len_list.append(feature.shape[1]) return max(len_list) def forward(self, x): outputs = [] features = self.feature(x) for feature in features: if self.bb_cfg['dropout']: feature = self.dropout(feature) if self.training_mode == 'pretrain': outputs.append(F.normalize(self.head(feature))) elif self.training_mode in ['scratch', 'fullfinetune', 'freezefinetune']: feature = feature.transpose(1, 2) output = self.classifier(feature) outputs.append(output) # (B, L, H) else: raise NotImplementedError return outputs
3,898
36.490385
133
py
SleePyCo
SleePyCo-main/models/classifiers.py
import math import torch import torch.nn as nn feature_len_dict = { 'SleePyCo': [ [5, 24, 120], [10, 48, 240], [15, 72, 360], [20, 96, 480], [24, 120, 600], [29, 144, 720], [34, 168, 840], [39, 192, 960], [44, 216, 1080], [48, 240, 1200]], } class PlainRNN(nn.Module): def __init__(self, config): super(PlainRNN, self).__init__() self.cfg = config['classifier'] self.num_classes = config['num_classes'] self.input_dim = config['comp_chn'] self.hidden_dim = config['hidden_dim'] self.num_layers = config['num_rnn_layers'] self.bidirectional = config['bidirectional'] # architecture self.rnn = nn.RNN( input_size=self.input_dim, hidden_size=self.hidden_dim, num_layers=self.num_layers, batch_first=True, bidirectional=self.bidirectional ) self.fc = nn.Linear(self.hidden_dim * 2 if config['bidirectional'] else self.hidden_dim, self.num_classes) def init_hidden(self, x): h0 = torch.zeros((self.num_layers * (2 if self.bidirectional else 1), x.size(0), self.hidden_dim)).cuda() return h0 def forward(self, x): hidden = self.init_hidden(x) rnn_output, hidden = self.rnn(x, hidden) if self.bidirectional: output_f = rnn_output[:, -1, :self.hidden_dim] output_b = rnn_output[:, 0, self.hidden_dim:] output = torch.cat((output_f, output_b), dim=1) else: output = rnn_output[:, -1, :] output = self.fc(output) return output class PlainGRU(PlainRNN): def __init__(self, config): super(PlainGRU, self).__init__(config) self.rnn = nn.GRU( input_size=self.input_dim, hidden_size=self.hidden_dim, num_layers=self.num_layers, batch_first=True, bidirectional=self.bidirectional ) class PlainLSTM(PlainRNN): def __init__(self, config): super(PlainLSTM, self).__init__(config) self.rnn = nn.LSTM( input_size=self.input_dim, hidden_size=self.hidden_dim, num_layers=self.num_layers, batch_first=True, bidirectional=self.bidirectional ) def init_hidden(self, x): h0 = torch.zeros((self.num_layers * (2 if self.bidirectional else 1), x.size(0), self.hidden_dim)).cuda() c0 = torch.zeros((self.num_layers * (2 if self.bidirectional else 1), x.size(0), self.hidden_dim)).cuda() return h0, c0 class AttRNN(PlainRNN): def __init__(self, config): super(AttRNN, self).__init__(config) # architecture self.fc = nn.Linear(self.hidden_dim, self.num_classes) self.w_ha = nn.Linear(self.hidden_dim * 2 if config['bidirectional'] else self.hidden_dim, self.hidden_dim, bias=True) self.w_att = nn.Linear(self.hidden_dim, 1, bias=False) def forward(self, x): hidden = self.init_hidden(x) rnn_output, hidden = self.rnn(x, hidden) a_states = self.w_ha(rnn_output) alpha = torch.softmax(self.w_att(a_states), dim=1).view(x.size(0), 1, x.size(1)) weighted_sum = torch.bmm(alpha, a_states) output = weighted_sum.view(x.size(0), -1) output = self.fc(output) return output class AttGRU(AttRNN): def __init__(self, config): super(AttGRU, self).__init__(config) self.rnn = nn.GRU( input_size=self.input_dim, hidden_size=self.hidden_dim, num_layers=self.num_layers, batch_first=True, bidirectional=self.bidirectional ) class AttLSTM(AttRNN): def __init__(self, config): super(AttLSTM, self).__init__(config) self.rnn = nn.LSTM( input_size=self.input_dim, hidden_size=self.hidden_dim, num_layers=self.num_layers, batch_first=True, bidirectional=self.bidirectional ) def init_hidden(self, x): h0 = torch.zeros((self.num_layers * (2 if self.bidirectional else 1), x.size(0), self.hidden_dim)).cuda() c0 = torch.zeros((self.num_layers * (2 if self.bidirectional else 1), x.size(0), self.hidden_dim)).cuda() return h0, c0 class PositionalEncoding(nn.Module): def __init__(self, config, in_features, out_features, dropout=0.1): super(PositionalEncoding, self).__init__() self.cfg = config['classifier']['pos_enc'] if self.cfg['dropout']: self.dropout = nn.Dropout(p=dropout) self.fc = nn.Linear(in_features=in_features, out_features=out_features) self.act_fn = nn.PReLU() self.max_len = feature_len_dict[config['backbone']['name']][config['dataset']['seq_len'] - 1][config['feature_pyramid']['num_scales'] - 1] print('[INFO] Maximum length of pos_enc: {}'.format(self.max_len)) pe = torch.zeros(self.max_len, out_features) position = torch.arange(0, self.max_len, dtype=torch.float).unsqueeze(1) div_term = torch.exp(torch.arange(0, out_features, 2).float() * (-math.log(10000.0) / out_features)) pe[:, 0::2] = torch.sin(position * div_term) pe[:, 1::2] = torch.cos(position * div_term) pe = pe.unsqueeze(0).transpose(0, 1) self.register_buffer('pe', pe) def forward(self, x): x = self.act_fn(self.fc(x)) hop = self.max_len // x.size(0) pe = self.pe[hop//2::hop, :] if pe.shape[0] != x.size(0): pe = pe[:x.size(0), :] x = x + pe if self.cfg['dropout']: x = self.dropout(x) return x class Transformer(nn.Module): def __init__(self, config, nheads, num_encoder_layers, pool='mean'): super(Transformer, self).__init__() self.cfg = config['classifier'] self.model_dim = self.cfg['model_dim'] self.feedforward_dim = self.cfg['feedforward_dim'] self.in_features = config['feature_pyramid']['dim'] self.out_features = self.cfg['model_dim'] self.pos_encoding = PositionalEncoding(config, self.in_features, self.out_features) self.transformer_layer = nn.TransformerEncoderLayer( d_model=self.model_dim, nhead=nheads, dim_feedforward=self.feedforward_dim, dropout=0.1 if self.cfg['dropout'] else 0.0 ) self.transformer = nn.TransformerEncoder(self.transformer_layer, num_layers=num_encoder_layers) self.pool = pool if self.cfg['dropout']: self.dropout = nn.Dropout(p=0.5) if pool == 'attn': self.w_ha = nn.Linear(self.model_dim, self.model_dim, bias=True) self.w_at = nn.Linear(self.model_dim, 1, bias=False) self.fc = nn.Linear(self.model_dim, self.cfg['num_classes']) def forward(self, x): x = x.transpose(0, 1) x = self.pos_encoding(x) x = self.transformer(x) x = x.transpose(0, 1) if self.pool == 'mean': x = x.mean(dim=1) elif self.pool == 'last': x = x[:, -1] elif self.pool == 'attn': a_states = torch.tanh(self.w_ha(x)) alpha = torch.softmax(self.w_at(a_states), dim=1).view(x.size(0), 1, x.size(1)) x = torch.bmm(alpha, a_states).view(x.size(0), -1) elif self.pool == None: x = x else: raise NotImplementedError if self.cfg['dropout']: x = self.dropout(x) out = self.fc(x) return out def get_classifier(config): classifier_name = config['classifier']['name'] if classifier_name == 'PlainRNN': classifier = PlainRNN(config) elif classifier_name == 'AttentionRNN': classifier = AttRNN(config) if classifier_name == 'PlainLSTM': classifier = PlainLSTM(config) elif classifier_name == 'AttentionLSTM': classifier = AttLSTM(config) elif classifier_name == 'PlainGRU': classifier = PlainGRU(config) elif classifier_name == 'AttentionGRU': classifier = AttGRU(config) elif classifier_name == 'Transformer': classifier = Transformer(config, nheads=8, num_encoder_layers=6, pool=config['classifier']['pool']) return classifier
8,588
30.811111
146
py
SleePyCo
SleePyCo-main/models/sleepyco.py
import torch import torch.nn as nn import torch.nn.functional as F class SleePyCoBackbone(nn.Module): def __init__(self, config): super(SleePyCoBackbone, self).__init__() self.training_mode = config['training_params']['mode'] # architecture self.init_layer = self.make_layers(in_channels=1, out_channels=64, n_layers=2, maxpool_size=None, first=True) self.layer1 = self.make_layers(in_channels=64, out_channels=128, n_layers=2, maxpool_size=5) self.layer2 = self.make_layers(in_channels=128, out_channels=192, n_layers=3, maxpool_size=5) self.layer3 = self.make_layers(in_channels=192, out_channels=256, n_layers=3, maxpool_size=5) self.layer4 = self.make_layers(in_channels=256, out_channels=256, n_layers=3, maxpool_size=5) if self.training_mode == 'freezefinetune': self.fp_dim = config['feature_pyramid']['dim'] self.num_scales = config['feature_pyramid']['num_scales'] self.conv_c5 = nn.Conv1d(256, self.fp_dim, 1, 1, 0) if self.num_scales > 1: self.conv_c4 = nn.Conv1d(256, self.fp_dim, 1, 1, 0) if self.num_scales > 2: self.conv_c3 = nn.Conv1d(192, self.fp_dim, 1, 1, 0) if config['backbone']['init_weights']: self._initialize_weights() def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.BatchNorm1d): nn.init.constant_(m.weight, 1) nn.init.constant_(m.bias, 0) def make_layers(self, in_channels, out_channels, n_layers, maxpool_size, first=False): layers = [] layers = layers + [MaxPool1d(maxpool_size)] if not first else layers for i in range(n_layers): conv1d = nn.Conv1d(in_channels, out_channels, kernel_size=3, padding=1) layers += [conv1d, nn.BatchNorm1d(out_channels)] if i == n_layers - 1: layers += [ChannelGate(in_channels)] layers += [nn.PReLU()] in_channels = out_channels return nn.Sequential(*layers) def forward(self, x): out = [] c1 = self.init_layer(x) c2 = self.layer1(c1) c3 = self.layer2(c2) c4 = self.layer3(c3) c5 = self.layer4(c4) if self.training_mode == 'pretrain': out.append(c5) elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: p5 = self.conv_c5(c5) out.append(p5) if self.num_scales > 1: p4 = self.conv_c4(c4) out.append(p4) if self.num_scales > 2: p3 = self.conv_c3(c3) out.append(p3) return out class MaxPool1d(nn.Module): def __init__(self, maxpool_size): super(MaxPool1d, self).__init__() self.maxpool_size = maxpool_size self.maxpool = nn.MaxPool1d(kernel_size=maxpool_size, stride=maxpool_size) def forward(self, x): _, _, n_samples = x.size() if n_samples % self.maxpool_size != 0: pad_size = self.maxpool_size - (n_samples % self.maxpool_size) if pad_size % 2 != 0: left_pad = pad_size // 2 right_pad = pad_size // 2 + 1 else: left_pad = pad_size // 2 right_pad = pad_size // 2 x = F.pad(x, (left_pad, right_pad), mode='constant') x = self.maxpool(x) return x class BasicConv(nn.Module): def __init__(self, in_planes, out_planes, kernel_size, stride=1, padding=0, dilation=1, groups=1, relu=True, bn=True, bias=False): super(BasicConv, self).__init__() self.out_channels = out_planes self.conv = nn.Conv1d(in_planes, out_planes, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, groups=groups, bias=bias) self.bn = nn.BatchNorm1d(out_planes,eps=1e-5, momentum=0.01, affine=True) if bn else None self.relu = nn.ReLU() if relu else None def forward(self, x): x = self.conv(x) if self.bn is not None: x = self.bn(x) if self.relu is not None: x = self.relu(x) return x class ChannelGate(nn.Module): def __init__(self, gate_channels, reduction_ratio=16, pool_types=['avg']): super(ChannelGate, self).__init__() self.gate_channels = gate_channels self.mlp = nn.Sequential( nn.Flatten(), nn.Linear(gate_channels, gate_channels // reduction_ratio), nn.ReLU(), nn.Linear(gate_channels // reduction_ratio, gate_channels) ) self.pool_types = pool_types def forward(self, x): channel_att_sum = None for pool_type in self.pool_types: if pool_type=='avg': avg_pool = F.avg_pool1d(x, x.size(2), stride=x.size(2)) channel_att_raw = self.mlp(avg_pool) elif pool_type=='max': max_pool = F.max_pool1d(x, x.size(2), stride=x.size(2)) channel_att_raw = self.mlp( max_pool ) elif pool_type=='lp': lp_pool = F.lp_pool2d( x, 2, (x.size(2), x.size(3)), stride=(x.size(2), x.size(3))) channel_att_raw = self.mlp( lp_pool ) elif pool_type=='lse': # LSE pool only lse_pool = logsumexp_2d(x) channel_att_raw = self.mlp( lse_pool ) if channel_att_sum is None: channel_att_sum = channel_att_raw else: channel_att_sum = channel_att_sum + channel_att_raw scale = F.sigmoid(channel_att_sum).unsqueeze(2).expand_as(x) return x * scale def logsumexp_2d(tensor): tensor_flatten = tensor.view(tensor.size(0), tensor.size(1), -1) s, _ = torch.max(tensor_flatten, dim=2, keepdim=True) outputs = s + (tensor_flatten - s).exp().sum(dim=2, keepdim=True).log() return outputs
6,271
37.012121
154
py
SleePyCo
SleePyCo-main/models/deepsleepnet.py
import torch import torch.nn as nn from .utils import Conv1d, MaxPool1d class DeepSleepNetFeature(nn.Module): def __init__(self, config): super(DeepSleepNetFeature, self).__init__() self.chn = 64 # architecture self.dropout = nn.Dropout(p=0.5) self.path1 = nn.Sequential(Conv1d(1, self.chn, 50, 6, padding='SAME', bias=False), nn.BatchNorm1d(self.chn), nn.ReLU(inplace=True), MaxPool1d(8, padding='SAME'), nn.Dropout(), Conv1d(self.chn, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn*2), nn.ReLU(inplace=True), Conv1d(self.chn*2, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn * 2), nn.ReLU(inplace=True), Conv1d(self.chn*2, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn*2), nn.ReLU(inplace=True), MaxPool1d(4, padding='SAME') ) self.path2 = nn.Sequential(Conv1d(1, self.chn, 400, 50, padding='SAME', bias=False), nn.BatchNorm1d(self.chn), nn.ReLU(inplace=True), MaxPool1d(4, padding='SAME'), nn.Dropout(), Conv1d(self.chn, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn*2), nn.ReLU(inplace=True), Conv1d(self.chn*2, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn * 2), nn.ReLU(inplace=True), Conv1d(self.chn*2, self.chn*2, 8, 1, padding='SAME', bias=False), nn.BatchNorm1d(self.chn*2), nn.ReLU(inplace=True), MaxPool1d(2, padding='SAME')) self.compress = nn.Conv1d(self.chn*4, 128, 1, 1, 0) self.smooth = nn.Conv1d(128, 128, 3, 1, 1) if self.training_mode == 'freezefinetune': self.fp_dim = config['feature_pyramid']['dim'] self.num_scales = config['feature_pyramid']['num_scales'] self.conv_c5 = nn.Conv1d(128, self.fp_dim, 1, 1, 0) assert self.num_scales == 1 if config['backbone']['init_weights']: self._initialize_weights() def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.BatchNorm1d): nn.init.constant_(m.weight, 1) nn.init.constant_(m.bias, 0) def forward(self, x): out = [] x1 = self.path1(x) # path 1 x2 = self.path2(x) # path 2 x2 = torch.nn.functional.interpolate(x2, x1.size(2)) c5 = self.smooth(self.compress(torch.cat([x1, x2], dim=1))) if self.training_mode == 'pretrain': out.append(c5) elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: p5 = self.conv_c5(c5) out.append(p5) return out
3,855
44.904762
100
py
SleePyCo
SleePyCo-main/models/utime.py
import torch import torch.nn as nn from .utils import Conv1d class ConvUnit(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride, padding, dilation): super(ConvUnit, self).__init__() self.conv = Conv1d( in_channels=in_channels, out_channels=out_channels, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation ) self.bn = nn.BatchNorm1d(num_features=out_channels) self.relu = nn.ReLU() def forward(self, x): x = self.conv(x) x = self.bn(x) x = self.relu(x) return x class UTimeEncoder(nn.Module): def __init__(self, config): super(UTimeEncoder, self).__init__() torch.backends.cudnn.deterministic = False self.training_mode = config['training_params']['mode'] self.conv1_1 = ConvUnit(in_channels=1, out_channels=16, kernel_size=5, stride=1, padding='SAME', dilation=2) self.conv1_2 = ConvUnit(in_channels=16, out_channels=16, kernel_size=5, stride=1, padding='SAME', dilation=2) self.mp1 = nn.MaxPool1d(kernel_size=8) self.conv2_1 = ConvUnit(in_channels=16, out_channels=32, kernel_size=5, stride=1, padding='SAME', dilation=2) self.conv2_2 = ConvUnit(in_channels=32, out_channels=32, kernel_size=5, stride=1, padding='SAME', dilation=2) self.mp2 = nn.MaxPool1d(kernel_size=6) self.conv3_1 = ConvUnit(in_channels=32, out_channels=64, kernel_size=5, stride=1, padding='SAME', dilation=2) self.conv3_2 = ConvUnit(in_channels=64, out_channels=64, kernel_size=5, stride=1, padding='SAME', dilation=2) self.mp3 = nn.MaxPool1d(kernel_size=4) self.conv4_1 = ConvUnit(in_channels=64, out_channels=128, kernel_size=5, stride=1, padding='SAME', dilation=2) self.conv4_2 = ConvUnit(in_channels=128, out_channels=128, kernel_size=5, stride=1, padding='SAME', dilation=2) self.mp4 = nn.MaxPool1d(kernel_size=2) self.conv5_1 = ConvUnit(in_channels=128, out_channels=256, kernel_size=5, stride=1, padding='SAME', dilation=2) self.conv5_2 = ConvUnit(in_channels=256, out_channels=256, kernel_size=5, stride=1, padding='SAME', dilation=2) if self.training_mode == 'freezefinetune': self.fp_dim = config['feature_pyramid']['dim'] self.num_scales = config['feature_pyramid']['num_scales'] self.conv_c5 = nn.Conv1d(256, self.fp_dim, 1, 1, 0) if self.n_anchor > 1: self.conv_c4 = nn.Conv1d(128, self.fp_dim, 1, 1, 0) if self.n_anchor > 2: self.conv_c3 = nn.Conv1d(64, self.fp_dim, 1, 1, 0) if config['backbone']['init_weights']: self._initialize_weights() def _initialize_weights(self): for m in self.modules(): if isinstance(m, nn.Conv1d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) elif isinstance(m, nn.BatchNorm1d): nn.init.constant_(m.weight, 1) nn.init.constant_(m.bias, 0) def forward(self, x): out = [] x = self.conv1_1(x) c1 = self.conv1_2(x) x = self.mp1(c1) x = self.conv2_1(x) c2 = self.conv2_2(x) x = self.mp2(c2) x = self.conv3_1(x) c3 = self.conv3_2(x) x = self.mp3(c3) x = self.conv4_1(x) c4 = self.conv4_2(x) x = self.mp4(c4) x = self.conv5_1(x) c5 = self.conv5_2(x) if self.training_mode == 'pretrain': out.append(c5) elif self.training_mode in ['scratch', 'fullyfinetune', 'freezefinetune']: p5 = self.conv_c5(c5) out.append(p5) if self.num_scales > 1: p4 = self.conv_c4(c4) out.append(p4) if self.num_scales > 2: p3 = self.conv_c3(c3) out.append(p3) return out
4,199
37.181818
119
py
SASA
SASA-main/setup.py
import os import subprocess from setuptools import find_packages, setup from torch.utils.cpp_extension import BuildExtension, CUDAExtension def get_git_commit_number(): if not os.path.exists('.git'): return '0000000' cmd_out = subprocess.run(['git', 'rev-parse', 'HEAD'], stdout=subprocess.PIPE) git_commit_number = cmd_out.stdout.decode('utf-8')[:7] return git_commit_number def make_cuda_ext(name, module, sources): cuda_ext = CUDAExtension( name='%s.%s' % (module, name), sources=[os.path.join(*module.split('.'), src) for src in sources] ) return cuda_ext def write_version_to_file(version, target_file): with open(target_file, 'w') as f: print('__version__ = "%s"' % version, file=f) if __name__ == '__main__': version = '0.3.0+%s' % get_git_commit_number() write_version_to_file(version, 'pcdet/version.py') setup( name='pcdet', version=version, description='OpenPCDet is a general codebase for 3D object detection from point cloud', install_requires=[ 'numpy', 'torch>=1.1', 'spconv', 'numba', 'tensorboardX', 'easydict', 'pyyaml' ], author='Shaoshuai Shi', author_email='shaoshuaics@gmail.com', license='Apache License 2.0', packages=find_packages(exclude=['tools', 'data', 'output']), cmdclass={'build_ext': BuildExtension}, ext_modules=[ make_cuda_ext( name='iou3d_nms_cuda', module='pcdet.ops.iou3d_nms', sources=[ 'src/iou3d_cpu.cpp', 'src/iou3d_nms_api.cpp', 'src/iou3d_nms.cpp', 'src/iou3d_nms_kernel.cu', ] ), make_cuda_ext( name='roiaware_pool3d_cuda', module='pcdet.ops.roiaware_pool3d', sources=[ 'src/roiaware_pool3d.cpp', 'src/roiaware_pool3d_kernel.cu', ] ), make_cuda_ext( name='roipoint_pool3d_cuda', module='pcdet.ops.roipoint_pool3d', sources=[ 'src/roipoint_pool3d.cpp', 'src/roipoint_pool3d_kernel.cu', ] ), make_cuda_ext( name='pointnet2_stack_cuda', module='pcdet.ops.pointnet2.pointnet2_stack', sources=[ 'src/pointnet2_api.cpp', 'src/ball_query.cpp', 'src/ball_query_gpu.cu', 'src/group_points.cpp', 'src/group_points_gpu.cu', 'src/sampling.cpp', 'src/sampling_gpu.cu', 'src/interpolate.cpp', 'src/interpolate_gpu.cu', ], ), make_cuda_ext( name='pointnet2_batch_cuda', module='pcdet.ops.pointnet2.pointnet2_batch', sources=[ 'src/pointnet2_api.cpp', 'src/ball_query.cpp', 'src/ball_query_gpu.cu', 'src/group_points.cpp', 'src/group_points_gpu.cu', 'src/interpolate.cpp', 'src/interpolate_gpu.cu', 'src/sampling.cpp', 'src/sampling_gpu.cu', ], ), ], )
3,616
31.294643
95
py
SASA
SASA-main/tools/test.py
import argparse import datetime import glob import os import re import time from pathlib import Path import numpy as np import torch from tensorboardX import SummaryWriter from eval_utils import eval_utils from pcdet.config import cfg, cfg_from_list, cfg_from_yaml_file, log_config_to_file from pcdet.datasets import build_dataloader from pcdet.models import build_network from pcdet.utils import common_utils def parse_config(): parser = argparse.ArgumentParser(description='arg parser') parser.add_argument('--cfg_file', type=str, default=None, help='specify the config for training') parser.add_argument('--batch_size', type=int, default=None, required=False, help='batch size for training') parser.add_argument('--workers', type=int, default=4, help='number of workers for dataloader') parser.add_argument('--extra_tag', type=str, default='default', help='extra tag for this experiment') parser.add_argument('--ckpt', type=str, default=None, help='checkpoint to start from') parser.add_argument('--launcher', choices=['none', 'pytorch', 'slurm'], default='none') parser.add_argument('--tcp_port', type=int, default=18888, help='tcp port for distrbuted training') parser.add_argument('--local_rank', type=int, default=0, help='local rank for distributed training') parser.add_argument('--set', dest='set_cfgs', default=None, nargs=argparse.REMAINDER, help='set extra config keys if needed') parser.add_argument('--max_waiting_mins', type=int, default=30, help='max waiting minutes') parser.add_argument('--start_epoch', type=int, default=0, help='') parser.add_argument('--eval_tag', type=str, default='default', help='eval tag for this experiment') parser.add_argument('--eval_all', action='store_true', default=False, help='whether to evaluate all checkpoints') parser.add_argument('--ckpt_dir', type=str, default=None, help='specify a ckpt directory to be evaluated if needed') parser.add_argument('--save_to_file', action='store_true', default=False, help='') args = parser.parse_args() cfg_from_yaml_file(args.cfg_file, cfg) cfg.TAG = Path(args.cfg_file).stem cfg.EXP_GROUP_PATH = '/'.join(args.cfg_file.split('/')[1:-1]) # remove 'cfgs' and 'xxxx.yaml' np.random.seed(1024) if args.set_cfgs is not None: cfg_from_list(args.set_cfgs, cfg) return args, cfg def eval_single_ckpt(model, test_loader, args, eval_output_dir, logger, epoch_id, dist_test=False): # load checkpoint model.load_params_from_file(filename=args.ckpt, logger=logger, to_cpu=dist_test) model.cuda() # tensorboard log if cfg.LOCAL_RANK == 0: tb_log = SummaryWriter(log_dir=str(eval_output_dir / ('tensorboard_%s' % cfg.DATA_CONFIG.DATA_SPLIT['test']))) # start evaluation tb_dict = eval_utils.eval_one_epoch( cfg, model, test_loader, epoch_id, logger, dist_test=dist_test, result_dir=eval_output_dir, save_to_file=args.save_to_file ) if cfg.LOCAL_RANK == 0: for key, val in tb_dict.items(): if 'curve' in key: for i in range(len(val['recalls'])): tb_log.add_scalars(key, dict(zip(val['precisions'].keys(), [_[i] for _ in val['precisions'].values()])), val['recalls'][i] * 1000) def get_no_evaluated_ckpt(ckpt_dir, ckpt_record_file, args): ckpt_list = glob.glob(os.path.join(ckpt_dir, '*checkpoint_epoch_*.pth')) ckpt_list.sort(key=os.path.getmtime) evaluated_ckpt_list = [float(x.strip()) for x in open(ckpt_record_file, 'r').readlines()] for cur_ckpt in ckpt_list: num_list = re.findall('checkpoint_epoch_(.*).pth', cur_ckpt) if num_list.__len__() == 0: continue epoch_id = num_list[-1] if 'optim' in epoch_id: continue if float(epoch_id) not in evaluated_ckpt_list and int(float(epoch_id)) >= args.start_epoch: return epoch_id, cur_ckpt return -1, None def repeat_eval_ckpt(model, test_loader, args, eval_output_dir, logger, ckpt_dir, dist_test=False): # evaluated ckpt record ckpt_record_file = eval_output_dir / ('eval_list_%s.txt' % cfg.DATA_CONFIG.DATA_SPLIT['test']) with open(ckpt_record_file, 'a'): pass # tensorboard log if cfg.LOCAL_RANK == 0: tb_log = SummaryWriter(log_dir=str(eval_output_dir / ('tensorboard_%s' % cfg.DATA_CONFIG.DATA_SPLIT['test']))) total_time = 0 first_eval = True while True: # check whether there is checkpoint which is not evaluated cur_epoch_id, cur_ckpt = get_no_evaluated_ckpt(ckpt_dir, ckpt_record_file, args) if cur_epoch_id == -1 or int(float(cur_epoch_id)) < args.start_epoch: wait_second = 30 if cfg.LOCAL_RANK == 0: print('Wait %s seconds for next check (progress: %.1f / %d minutes): %s \r' % (wait_second, total_time * 1.0 / 60, args.max_waiting_mins, ckpt_dir), end='', flush=True) time.sleep(wait_second) total_time += 30 if total_time > args.max_waiting_mins * 60 and (first_eval is False): break continue total_time = 0 first_eval = False model.load_params_from_file(filename=cur_ckpt, logger=logger, to_cpu=dist_test) model.cuda() # start evaluation cur_result_dir = eval_output_dir / ('epoch_%s' % cur_epoch_id) / cfg.DATA_CONFIG.DATA_SPLIT['test'] tb_dict = eval_utils.eval_one_epoch( cfg, model, test_loader, cur_epoch_id, logger, dist_test=dist_test, result_dir=cur_result_dir, save_to_file=args.save_to_file ) if cfg.LOCAL_RANK == 0: for key, val in tb_dict.items(): if not 'curve' in key: tb_log.add_scalar(key, val, cur_epoch_id) # record this epoch which has been evaluated with open(ckpt_record_file, 'a') as f: print('%s' % cur_epoch_id, file=f) logger.info('Epoch %s has been evaluated' % cur_epoch_id) # draw precision-recall curves for the last evaluation for key, val in tb_dict.items(): if 'curve' in key: for i in range(len(val['recalls'])): tb_log.add_scalars(key, dict(zip(val['precisions'].keys(), [_[i] for _ in val['precisions'].values()])), val['recalls'][i] * 1000) def main(): args, cfg = parse_config() if args.launcher == 'none': dist_test = False total_gpus = 1 else: total_gpus, cfg.LOCAL_RANK = getattr(common_utils, 'init_dist_%s' % args.launcher)( args.tcp_port, args.local_rank, backend='nccl' ) dist_test = True if args.batch_size is None: args.batch_size = cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU else: assert args.batch_size % total_gpus == 0, 'Batch size should match the number of gpus' args.batch_size = args.batch_size // total_gpus output_dir = cfg.ROOT_DIR / 'output' / cfg.EXP_GROUP_PATH / cfg.TAG / args.extra_tag output_dir.mkdir(parents=True, exist_ok=True) eval_output_dir = output_dir / 'eval' if not args.eval_all: num_list = re.findall(r'\d+', args.ckpt) if args.ckpt is not None else [] epoch_id = num_list[-1] if num_list.__len__() > 0 else 'no_number' eval_output_dir = eval_output_dir / ('epoch_%s' % epoch_id) / cfg.DATA_CONFIG.DATA_SPLIT['test'] else: eval_output_dir = eval_output_dir / 'eval_all_default' if args.eval_tag is not None: eval_output_dir = eval_output_dir / args.eval_tag eval_output_dir.mkdir(parents=True, exist_ok=True) log_file = eval_output_dir / ('log_eval_%s.txt' % datetime.datetime.now().strftime('%Y%m%d-%H%M%S')) logger = common_utils.create_logger(log_file, rank=cfg.LOCAL_RANK) # log to file logger.info('**********************Start logging**********************') gpu_list = os.environ['CUDA_VISIBLE_DEVICES'] if 'CUDA_VISIBLE_DEVICES' in os.environ.keys() else 'ALL' logger.info('CUDA_VISIBLE_DEVICES=%s' % gpu_list) if dist_test: logger.info('total_batch_size: %d' % (total_gpus * args.batch_size)) for key, val in vars(args).items(): logger.info('{:16} {}'.format(key, val)) log_config_to_file(cfg, logger=logger) ckpt_dir = args.ckpt_dir if args.ckpt_dir is not None else output_dir / 'ckpt' test_set, test_loader, sampler = build_dataloader( dataset_cfg=cfg.DATA_CONFIG, class_names=cfg.CLASS_NAMES, batch_size=args.batch_size, dist=dist_test, workers=args.workers, logger=logger, training=False ) model = build_network(model_cfg=cfg.MODEL, num_class=len(cfg.CLASS_NAMES), dataset=test_set) with torch.no_grad(): if args.eval_all: repeat_eval_ckpt(model, test_loader, args, eval_output_dir, logger, ckpt_dir, dist_test=dist_test) else: eval_single_ckpt(model, test_loader, args, eval_output_dir, logger, epoch_id, dist_test=dist_test) if __name__ == '__main__': main()
9,291
40.855856
120
py
SASA
SASA-main/tools/demo.py
import argparse import glob from pathlib import Path import mayavi.mlab as mlab import numpy as np import torch from pcdet.config import cfg, cfg_from_yaml_file from pcdet.datasets import DatasetTemplate from pcdet.models import build_network, load_data_to_gpu from pcdet.utils import common_utils from visual_utils import visualize_utils as V class DemoDataset(DatasetTemplate): def __init__(self, dataset_cfg, class_names, training=True, root_path=None, logger=None, ext='.bin'): """ Args: root_path: dataset_cfg: class_names: training: logger: """ super().__init__( dataset_cfg=dataset_cfg, class_names=class_names, training=training, root_path=root_path, logger=logger ) self.root_path = root_path self.ext = ext data_file_list = glob.glob(str(root_path / f'*{self.ext}')) if self.root_path.is_dir() else [self.root_path] data_file_list.sort() self.sample_file_list = data_file_list def __len__(self): return len(self.sample_file_list) def __getitem__(self, index): if self.ext == '.bin': points = np.fromfile(self.sample_file_list[index], dtype=np.float32).reshape(-1, 4) elif self.ext == '.npy': points = np.load(self.sample_file_list[index]) else: raise NotImplementedError input_dict = { 'points': points, 'frame_id': index, } data_dict = self.prepare_data(data_dict=input_dict) return data_dict def parse_config(): parser = argparse.ArgumentParser(description='arg parser') parser.add_argument('--cfg_file', type=str, default='cfgs/kitti_models/second.yaml', help='specify the config for demo') parser.add_argument('--data_path', type=str, default='demo_data', help='specify the point cloud data file or directory') parser.add_argument('--ckpt', type=str, default=None, help='specify the pretrained model') parser.add_argument('--ext', type=str, default='.bin', help='specify the extension of your point cloud data file') args = parser.parse_args() cfg_from_yaml_file(args.cfg_file, cfg) return args, cfg def main(): args, cfg = parse_config() logger = common_utils.create_logger() logger.info('-----------------Quick Demo of OpenPCDet-------------------------') demo_dataset = DemoDataset( dataset_cfg=cfg.DATA_CONFIG, class_names=cfg.CLASS_NAMES, training=False, root_path=Path(args.data_path), ext=args.ext, logger=logger ) logger.info(f'Total number of samples: \t{len(demo_dataset)}') model = build_network(model_cfg=cfg.MODEL, num_class=len(cfg.CLASS_NAMES), dataset=demo_dataset) model.load_params_from_file(filename=args.ckpt, logger=logger, to_cpu=True) model.cuda() model.eval() with torch.no_grad(): for idx, data_dict in enumerate(demo_dataset): logger.info(f'Visualized sample index: \t{idx + 1}') data_dict = demo_dataset.collate_batch([data_dict]) load_data_to_gpu(data_dict) pred_dicts, _ = model.forward(data_dict) V.draw_scenes( points=data_dict['points'][:, 1:], ref_boxes=pred_dicts[0]['pred_boxes'], ref_scores=pred_dicts[0]['pred_scores'], ref_labels=pred_dicts[0]['pred_labels'] ) mlab.show(stop=True) logger.info('Demo done.') if __name__ == '__main__': main()
3,575
33.384615
118
py
SASA
SASA-main/tools/train.py
import argparse import datetime import glob import os from pathlib import Path from test import repeat_eval_ckpt import torch import torch.distributed as dist import torch.nn as nn from tensorboardX import SummaryWriter from pcdet.config import cfg, cfg_from_list, cfg_from_yaml_file, log_config_to_file from pcdet.datasets import build_dataloader from pcdet.models import build_network, model_fn_decorator from pcdet.utils import common_utils from train_utils.optimization import build_optimizer, build_scheduler from train_utils.train_utils import train_model def parse_config(): parser = argparse.ArgumentParser(description='arg parser') parser.add_argument('--cfg_file', type=str, default=None, help='specify the config for training') parser.add_argument('--batch_size', type=int, default=None, required=False, help='batch size for training') parser.add_argument('--epochs', type=int, default=None, required=False, help='number of epochs to train for') parser.add_argument('--workers', type=int, default=8, help='number of workers for dataloader') parser.add_argument('--extra_tag', type=str, default='default', help='extra tag for this experiment') parser.add_argument('--ckpt', type=str, default=None, help='checkpoint to start from') parser.add_argument('--pretrained_model', type=str, default=None, help='pretrained_model') parser.add_argument('--launcher', choices=['none', 'pytorch', 'slurm'], default='none') parser.add_argument('--tcp_port', type=int, default=18888, help='tcp port for distrbuted training') parser.add_argument('--sync_bn', action='store_true', default=False, help='whether to use sync bn') parser.add_argument('--fix_random_seed', type=int, default=666, help='random seed') parser.add_argument('--ckpt_save_interval', type=int, default=1, help='number of training epochs') parser.add_argument('--local_rank', type=int, default=0, help='local rank for distributed training') parser.add_argument('--max_ckpt_save_num', type=int, default=10, help='max number of saved checkpoint') parser.add_argument('--merge_all_iters_to_one_epoch', action='store_true', default=False, help='') parser.add_argument('--set', dest='set_cfgs', default=None, nargs=argparse.REMAINDER, help='set extra config keys if needed') parser.add_argument('--max_waiting_mins', type=int, default=0, help='max waiting minutes') parser.add_argument('--start_epoch', type=int, default=0, help='') parser.add_argument('--save_to_file', action='store_true', default=False, help='') args = parser.parse_args() cfg_from_yaml_file(args.cfg_file, cfg) cfg.TAG = Path(args.cfg_file).stem cfg.EXP_GROUP_PATH = '/'.join(args.cfg_file.split('/')[1:-1]) # remove 'cfgs' and 'xxxx.yaml' if args.set_cfgs is not None: cfg_from_list(args.set_cfgs, cfg) return args, cfg def main(): args, cfg = parse_config() if args.launcher == 'none': dist_train = False total_gpus = 1 else: total_gpus, cfg.LOCAL_RANK = getattr(common_utils, 'init_dist_%s' % args.launcher)( args.tcp_port, args.local_rank, backend='nccl' ) dist_train = True if args.batch_size is None: args.batch_size = cfg.OPTIMIZATION.BATCH_SIZE_PER_GPU else: assert args.batch_size % total_gpus == 0, 'Batch size should match the number of gpus' args.batch_size = args.batch_size // total_gpus args.epochs = cfg.OPTIMIZATION.NUM_EPOCHS if args.epochs is None else args.epochs if not args.fix_random_seed == 0: common_utils.set_random_seed(args.fix_random_seed) output_dir = cfg.ROOT_DIR / 'output' / cfg.EXP_GROUP_PATH / cfg.TAG / args.extra_tag ckpt_dir = output_dir / 'ckpt' output_dir.mkdir(parents=True, exist_ok=True) ckpt_dir.mkdir(parents=True, exist_ok=True) log_file = output_dir / ('log_train_%s.txt' % datetime.datetime.now().strftime('%Y%m%d-%H%M%S')) logger = common_utils.create_logger(log_file, rank=cfg.LOCAL_RANK) # log to file logger.info('**********************Start logging**********************') gpu_list = os.environ['CUDA_VISIBLE_DEVICES'] if 'CUDA_VISIBLE_DEVICES' in os.environ.keys() else 'ALL' logger.info('CUDA_VISIBLE_DEVICES=%s' % gpu_list) if dist_train: logger.info('total_batch_size: %d' % (total_gpus * args.batch_size)) for key, val in vars(args).items(): logger.info('{:16} {}'.format(key, val)) log_config_to_file(cfg, logger=logger) if cfg.LOCAL_RANK == 0: os.system('cp %s %s' % (args.cfg_file, output_dir)) tb_log = SummaryWriter(log_dir=str(output_dir / 'tensorboard')) if cfg.LOCAL_RANK == 0 else None # -----------------------create dataloader & network & optimizer--------------------------- train_set, train_loader, train_sampler = build_dataloader( dataset_cfg=cfg.DATA_CONFIG, class_names=cfg.CLASS_NAMES, batch_size=args.batch_size, dist=dist_train, workers=args.workers, logger=logger, training=True, merge_all_iters_to_one_epoch=args.merge_all_iters_to_one_epoch, total_epochs=args.epochs ) model = build_network(model_cfg=cfg.MODEL, num_class=len(cfg.CLASS_NAMES), dataset=train_set) if args.sync_bn: model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model) model.cuda() optimizer = build_optimizer(model, cfg.OPTIMIZATION) # load checkpoint if it is possible start_epoch = it = 0 last_epoch = -1 if args.pretrained_model is not None: model.load_params_from_file(filename=args.pretrained_model, to_cpu=dist, logger=logger) if args.ckpt is not None: it, start_epoch = model.load_params_with_optimizer(args.ckpt, to_cpu=dist, optimizer=optimizer, logger=logger) last_epoch = start_epoch + 1 else: ckpt_list = glob.glob(str(ckpt_dir / '*checkpoint_epoch_*.pth')) if len(ckpt_list) > 0: ckpt_list.sort(key=os.path.getmtime) it, start_epoch = model.load_params_with_optimizer( ckpt_list[-1], to_cpu=dist, optimizer=optimizer, logger=logger ) last_epoch = start_epoch + 1 model.train() # before wrap to DistributedDataParallel to support fixed some parameters if dist_train: model = nn.parallel.DistributedDataParallel(model, device_ids=[cfg.LOCAL_RANK % torch.cuda.device_count()]) logger.info(model) total_iters_each_epoch = len(train_loader) if args.merge_all_iters_to_one_epoch: total_iters_each_epoch = total_iters_each_epoch // args.epochs lr_scheduler, lr_warmup_scheduler = build_scheduler( optimizer, total_iters_each_epoch=total_iters_each_epoch, total_epochs=args.epochs, last_epoch=last_epoch, optim_cfg=cfg.OPTIMIZATION ) # -----------------------start training--------------------------- logger.info('**********************Start training %s/%s(%s)**********************' % (cfg.EXP_GROUP_PATH, cfg.TAG, args.extra_tag)) train_model( model, optimizer, train_loader, model_func=model_fn_decorator(), lr_scheduler=lr_scheduler, optim_cfg=cfg.OPTIMIZATION, start_epoch=start_epoch, total_epochs=args.epochs, start_iter=it, rank=cfg.LOCAL_RANK, tb_log=tb_log, ckpt_save_dir=ckpt_dir, train_sampler=train_sampler, lr_warmup_scheduler=lr_warmup_scheduler, ckpt_save_interval=args.ckpt_save_interval, max_ckpt_save_num=args.max_ckpt_save_num, merge_all_iters_to_one_epoch=args.merge_all_iters_to_one_epoch ) logger.info('**********************End training %s/%s(%s)**********************\n\n\n' % (cfg.EXP_GROUP_PATH, cfg.TAG, args.extra_tag)) logger.info('**********************Start evaluation %s/%s(%s)**********************' % (cfg.EXP_GROUP_PATH, cfg.TAG, args.extra_tag)) test_set, test_loader, sampler = build_dataloader( dataset_cfg=cfg.DATA_CONFIG, class_names=cfg.CLASS_NAMES, batch_size=args.batch_size, dist=dist_train, workers=args.workers, logger=logger, training=False ) eval_output_dir = output_dir / 'eval' / 'eval_with_train' eval_output_dir.mkdir(parents=True, exist_ok=True) args.start_epoch = max(args.epochs - 10, 0) # Only evaluate the last 10 epochs repeat_eval_ckpt( model.module if dist_train else model, test_loader, args, eval_output_dir, logger, ckpt_dir, dist_test=dist_train ) logger.info('**********************End evaluation %s/%s(%s)**********************' % (cfg.EXP_GROUP_PATH, cfg.TAG, args.extra_tag)) if __name__ == '__main__': main()
8,839
42.762376
118
py
SASA
SASA-main/tools/eval_utils/eval_utils.py
import pickle import time import numpy as np import torch import tqdm from pcdet.models import load_data_to_gpu from pcdet.utils import common_utils def statistics_info(cfg, ret_dict, metric, disp_dict): for key in metric.keys(): if key in ret_dict: metric[key] += ret_dict[key] min_thresh = cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST[0] disp_dict['recall_%s' % str(min_thresh)] = \ '(%d, %d) / %d' % (metric['recall_roi_%s' % str(min_thresh)], metric['recall_rcnn_%s' % str(min_thresh)], metric['gt_num']) def eval_one_epoch(cfg, model, dataloader, epoch_id, logger, dist_test=False, save_to_file=False, result_dir=None): result_dir.mkdir(parents=True, exist_ok=True) final_output_dir = result_dir / 'final_result' / 'data' if save_to_file: final_output_dir.mkdir(parents=True, exist_ok=True) metric = { 'gt_num': 0, } for cur_thresh in cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: metric['recall_roi_%s' % str(cur_thresh)] = 0 metric['recall_rcnn_%s' % str(cur_thresh)] = 0 # initialize customized statistics if hasattr(model, 'init_recall_record'): model.init_recall_record(metric) dataset = dataloader.dataset class_names = dataset.class_names det_annos = [] logger.info('*************** EPOCH %s EVALUATION *****************' % epoch_id) if dist_test: num_gpus = torch.cuda.device_count() local_rank = cfg.LOCAL_RANK % num_gpus model = torch.nn.parallel.DistributedDataParallel( model, device_ids=[local_rank], broadcast_buffers=False ) model.eval() if cfg.LOCAL_RANK == 0: progress_bar = tqdm.tqdm(total=len(dataloader), leave=True, desc='eval', dynamic_ncols=True) start_time = time.time() for i, batch_dict in enumerate(dataloader): load_data_to_gpu(batch_dict) with torch.no_grad(): pred_dicts, ret_dict = model(batch_dict) disp_dict = {} statistics_info(cfg, ret_dict, metric, disp_dict) annos = dataset.generate_prediction_dicts( batch_dict, pred_dicts, class_names, output_path=final_output_dir if save_to_file else None ) det_annos += annos if cfg.LOCAL_RANK == 0: progress_bar.set_postfix(disp_dict) progress_bar.update() if cfg.LOCAL_RANK == 0: progress_bar.close() if dist_test: rank, world_size = common_utils.get_dist_info() det_annos = common_utils.merge_results_dist(det_annos, len(dataset), tmpdir=result_dir / 'tmpdir') metric = common_utils.merge_results_dist([metric], world_size, tmpdir=result_dir / 'tmpdir') logger.info('*************** Performance of EPOCH %s *****************' % epoch_id) sec_per_example = (time.time() - start_time) / len(dataloader.dataset) logger.info('Generate label finished(sec_per_example: %.4f second).' % sec_per_example) if cfg.LOCAL_RANK != 0: return {} ret_dict = {} if dist_test: for key, val in metric[0].items(): for k in range(1, world_size): metric[0][key] += metric[k][key] metric = metric[0] gt_num_cnt = metric['gt_num'] for cur_thresh in cfg.MODEL.POST_PROCESSING.RECALL_THRESH_LIST: cur_roi_recall = metric['recall_roi_%s' % str(cur_thresh)] / max(gt_num_cnt, 1) cur_rcnn_recall = metric['recall_rcnn_%s' % str(cur_thresh)] / max(gt_num_cnt, 1) logger.info('recall_roi_%s: %f' % (cur_thresh, cur_roi_recall)) logger.info('recall_rcnn_%s: %f' % (cur_thresh, cur_rcnn_recall)) ret_dict['recall/roi_%s' % str(cur_thresh)] = cur_roi_recall ret_dict['recall/rcnn_%s' % str(cur_thresh)] = cur_rcnn_recall # print customized statistics if hasattr(model, 'disp_recall_record'): model.disp_recall_record(metric, logger, sample_num=len(dataloader.dataset)) total_pred_objects = 0 for anno in det_annos: total_pred_objects += anno['name'].__len__() logger.info('Average predicted number of objects(%d samples): %.3f' % (len(det_annos), total_pred_objects / max(1, len(det_annos)))) with open(result_dir / 'result.pkl', 'wb') as f: pickle.dump(det_annos, f) result_str, result_dict = dataset.evaluation( det_annos, class_names, eval_metric=cfg.MODEL.POST_PROCESSING.EVAL_METRIC, output_path=final_output_dir ) logger.info(result_str) ret_dict.update(result_dict) logger.info('Result is save to %s' % result_dir) logger.info('****************Evaluation done.*****************') return ret_dict if __name__ == '__main__': pass
4,772
34.887218
131
py
SASA
SASA-main/tools/train_utils/train_utils.py
import glob import os import torch import tqdm from torch.nn.utils import clip_grad_norm_ def train_one_epoch(model, optimizer, train_loader, model_func, lr_scheduler, accumulated_iter, optim_cfg, rank, tbar, total_it_each_epoch, dataloader_iter, tb_log=None, leave_pbar=False): if total_it_each_epoch == len(train_loader): dataloader_iter = iter(train_loader) if rank == 0: pbar = tqdm.tqdm(total=total_it_each_epoch, leave=leave_pbar, desc='train', dynamic_ncols=True) for cur_it in range(total_it_each_epoch): try: batch = next(dataloader_iter) except StopIteration: dataloader_iter = iter(train_loader) batch = next(dataloader_iter) print('new iters') lr_scheduler.step(accumulated_iter) try: cur_lr = float(optimizer.lr) except: cur_lr = optimizer.param_groups[0]['lr'] if tb_log is not None: tb_log.add_scalar('meta_data/learning_rate', cur_lr, accumulated_iter) model.train() optimizer.zero_grad() loss, tb_dict, disp_dict = model_func(model, batch) loss.backward() clip_grad_norm_(model.parameters(), optim_cfg.GRAD_NORM_CLIP) optimizer.step() accumulated_iter += 1 disp_dict.update({'loss': loss.item(), 'lr': cur_lr}) # log to console and tensorboard if rank == 0: pbar.update() pbar.set_postfix(dict(total_it=accumulated_iter)) tbar.set_postfix(disp_dict) tbar.refresh() if tb_log is not None: tb_log.add_scalar('train/loss', loss, accumulated_iter) tb_log.add_scalar('meta_data/learning_rate', cur_lr, accumulated_iter) for key, val in tb_dict.items(): tb_log.add_scalar('train/' + key, val, accumulated_iter) if rank == 0: pbar.close() return accumulated_iter def train_model(model, optimizer, train_loader, model_func, lr_scheduler, optim_cfg, start_epoch, total_epochs, start_iter, rank, tb_log, ckpt_save_dir, train_sampler=None, lr_warmup_scheduler=None, ckpt_save_interval=1, max_ckpt_save_num=50, merge_all_iters_to_one_epoch=False): accumulated_iter = start_iter with tqdm.trange(start_epoch, total_epochs, desc='epochs', dynamic_ncols=True, leave=(rank == 0)) as tbar: total_it_each_epoch = len(train_loader) if merge_all_iters_to_one_epoch: assert hasattr(train_loader.dataset, 'merge_all_iters_to_one_epoch') train_loader.dataset.merge_all_iters_to_one_epoch(merge=True, epochs=total_epochs) total_it_each_epoch = len(train_loader) // max(total_epochs, 1) dataloader_iter = iter(train_loader) for cur_epoch in tbar: if train_sampler is not None: train_sampler.set_epoch(cur_epoch) # train one epoch if lr_warmup_scheduler is not None and cur_epoch < optim_cfg.WARMUP_EPOCH: cur_scheduler = lr_warmup_scheduler else: cur_scheduler = lr_scheduler accumulated_iter = train_one_epoch( model, optimizer, train_loader, model_func, lr_scheduler=cur_scheduler, accumulated_iter=accumulated_iter, optim_cfg=optim_cfg, rank=rank, tbar=tbar, tb_log=tb_log, leave_pbar=(cur_epoch + 1 == total_epochs), total_it_each_epoch=total_it_each_epoch, dataloader_iter=dataloader_iter ) # save trained model trained_epoch = cur_epoch + 1 if trained_epoch % ckpt_save_interval == 0 and rank == 0: ckpt_list = glob.glob(str(ckpt_save_dir / 'checkpoint_epoch_*.pth')) ckpt_list.sort(key=os.path.getmtime) if ckpt_list.__len__() >= max_ckpt_save_num: for cur_file_idx in range(0, len(ckpt_list) - max_ckpt_save_num + 1): os.remove(ckpt_list[cur_file_idx]) ckpt_name = ckpt_save_dir / ('checkpoint_epoch_%d' % trained_epoch) save_checkpoint( checkpoint_state(model, optimizer, trained_epoch, accumulated_iter), filename=ckpt_name, ) def model_state_to_cpu(model_state): model_state_cpu = type(model_state)() # ordered dict for key, val in model_state.items(): model_state_cpu[key] = val.cpu() return model_state_cpu def checkpoint_state(model=None, optimizer=None, epoch=None, it=None): optim_state = optimizer.state_dict() if optimizer is not None else None if model is not None: if isinstance(model, torch.nn.parallel.DistributedDataParallel): model_state = model_state_to_cpu(model.module.state_dict()) else: model_state = model.state_dict() else: model_state = None try: import pcdet version = 'pcdet+' + pcdet.__version__ except: version = 'none' return {'epoch': epoch, 'it': it, 'model_state': model_state, 'optimizer_state': optim_state, 'version': version} def save_checkpoint(state, filename='checkpoint'): if False and 'optimizer_state' in state: optimizer_state = state['optimizer_state'] state.pop('optimizer_state', None) optimizer_filename = '{}_optim.pth'.format(filename) torch.save({'optimizer_state': optimizer_state}, optimizer_filename) filename = '{}.pth'.format(filename) torch.save(state, filename)
5,667
37.297297
117
py
SASA
SASA-main/tools/train_utils/optimization/fastai_optim.py
# This file is modified from https://github.com/traveller59/second.pytorch from collections import Iterable import torch from torch import nn from torch._utils import _unflatten_dense_tensors from torch.nn.utils import parameters_to_vector bn_types = (nn.BatchNorm1d, nn.BatchNorm2d, nn.BatchNorm3d, nn.SyncBatchNorm) def split_bn_bias(layer_groups): "Split the layers in `layer_groups` into batchnorm (`bn_types`) and non-batchnorm groups." split_groups = [] for l in layer_groups: l1, l2 = [], [] for c in l.children(): if isinstance(c, bn_types): l2.append(c) else: l1.append(c) split_groups += [nn.Sequential(*l1), nn.Sequential(*l2)] return split_groups def get_master(layer_groups, flat_master: bool = False): "Return two lists, one for the model parameters in FP16 and one for the master parameters in FP32." split_groups = split_bn_bias(layer_groups) model_params = [[param for param in lg.parameters() if param.requires_grad] for lg in split_groups] if flat_master: master_params = [] for lg in model_params: if len(lg) != 0: mp = parameters_to_vector([param.data.float() for param in lg]) mp = torch.nn.Parameter(mp, requires_grad=True) if mp.grad is None: mp.grad = mp.new(*mp.size()) master_params.append([mp]) else: master_params.append([]) return model_params, master_params else: master_params = [[param.clone().float().detach() for param in lg] for lg in model_params] for mp in master_params: for param in mp: param.requires_grad = True return model_params, master_params def model_g2master_g(model_params, master_params, flat_master: bool = False) -> None: "Copy the `model_params` gradients to `master_params` for the optimizer step." if flat_master: for model_group, master_group in zip(model_params, master_params): if len(master_group) != 0: master_group[0].grad.data.copy_(parameters_to_vector([p.grad.data.float() for p in model_group])) else: for model_group, master_group in zip(model_params, master_params): for model, master in zip(model_group, master_group): if model.grad is not None: if master.grad is None: master.grad = master.data.new(*master.data.size()) master.grad.data.copy_(model.grad.data) else: master.grad = None def master2model(model_params, master_params, flat_master: bool = False) -> None: "Copy `master_params` to `model_params`." if flat_master: for model_group, master_group in zip(model_params, master_params): if len(model_group) != 0: for model, master in zip(model_group, _unflatten_dense_tensors(master_group[0].data, model_group)): model.data.copy_(master) else: for model_group, master_group in zip(model_params, master_params): for model, master in zip(model_group, master_group): model.data.copy_(master.data) def listify(p=None, q=None): "Make `p` listy and the same length as `q`." if p is None: p = [] elif isinstance(p, str): p = [p] elif not isinstance(p, Iterable): p = [p] n = q if type(q) == int else len(p) if q is None else len(q) if len(p) == 1: p = p * n assert len(p) == n, f'List len mismatch ({len(p)} vs {n})' return list(p) def trainable_params(m: nn.Module): "Return list of trainable params in `m`." res = filter(lambda p: p.requires_grad, m.parameters()) return res def is_tuple(x) -> bool: return isinstance(x, tuple) # copy from fastai. class OptimWrapper(): "Basic wrapper around `opt` to simplify hyper-parameters changes." def __init__(self, opt, wd, true_wd: bool = False, bn_wd: bool = True): self.opt, self.true_wd, self.bn_wd = opt, true_wd, bn_wd self.opt_keys = list(self.opt.param_groups[0].keys()) self.opt_keys.remove('params') self.read_defaults() self.wd = wd @classmethod def create(cls, opt_func, lr, layer_groups, **kwargs): "Create an `optim.Optimizer` from `opt_func` with `lr`. Set lr on `layer_groups`." split_groups = split_bn_bias(layer_groups) opt = opt_func([{'params': trainable_params(l), 'lr': 0} for l in split_groups]) opt = cls(opt, **kwargs) opt.lr, opt.opt_func = listify(lr, layer_groups), opt_func return opt def new(self, layer_groups): "Create a new `OptimWrapper` from `self` with another `layer_groups` but the same hyper-parameters." opt_func = getattr(self, 'opt_func', self.opt.__class__) split_groups = split_bn_bias(layer_groups) opt = opt_func([{'params': trainable_params(l), 'lr': 0} for l in split_groups]) return self.create(opt_func, self.lr, layer_groups, wd=self.wd, true_wd=self.true_wd, bn_wd=self.bn_wd) def __repr__(self) -> str: return f'OptimWrapper over {repr(self.opt)}.\nTrue weight decay: {self.true_wd}' # Pytorch optimizer methods def step(self) -> None: "Set weight decay and step optimizer." # weight decay outside of optimizer step (AdamW) if self.true_wd: for lr, wd, pg1, pg2 in zip(self._lr, self._wd, self.opt.param_groups[::2], self.opt.param_groups[1::2]): for p in pg1['params']: # When some parameters are fixed: Shaoshuai Shi if p.requires_grad is False: continue p.data.mul_(1 - wd * lr) if self.bn_wd: for p in pg2['params']: # When some parameters are fixed: Shaoshuai Shi if p.requires_grad is False: continue p.data.mul_(1 - wd * lr) self.set_val('weight_decay', listify(0, self._wd)) self.opt.step() def zero_grad(self) -> None: "Clear optimizer gradients." self.opt.zero_grad() # Passthrough to the inner opt. def __getattr__(self, k: str): return getattr(self.opt, k, None) def clear(self): "Reset the state of the inner optimizer." sd = self.state_dict() sd['state'] = {} self.load_state_dict(sd) # Hyperparameters as properties @property def lr(self) -> float: return self._lr[-1] @lr.setter def lr(self, val: float) -> None: self._lr = self.set_val('lr', listify(val, self._lr)) @property def mom(self) -> float: return self._mom[-1] @mom.setter def mom(self, val: float) -> None: if 'momentum' in self.opt_keys: self.set_val('momentum', listify(val, self._mom)) elif 'betas' in self.opt_keys: self.set_val('betas', (listify(val, self._mom), self._beta)) self._mom = listify(val, self._mom) @property def beta(self) -> float: return None if self._beta is None else self._beta[-1] @beta.setter def beta(self, val: float) -> None: "Set beta (or alpha as makes sense for given optimizer)." if val is None: return if 'betas' in self.opt_keys: self.set_val('betas', (self._mom, listify(val, self._beta))) elif 'alpha' in self.opt_keys: self.set_val('alpha', listify(val, self._beta)) self._beta = listify(val, self._beta) @property def wd(self) -> float: return self._wd[-1] @wd.setter def wd(self, val: float) -> None: "Set weight decay." if not self.true_wd: self.set_val('weight_decay', listify(val, self._wd), bn_groups=self.bn_wd) self._wd = listify(val, self._wd) # Helper functions def read_defaults(self) -> None: "Read the values inside the optimizer for the hyper-parameters." self._beta = None if 'lr' in self.opt_keys: self._lr = self.read_val('lr') if 'momentum' in self.opt_keys: self._mom = self.read_val('momentum') if 'alpha' in self.opt_keys: self._beta = self.read_val('alpha') if 'betas' in self.opt_keys: self._mom, self._beta = self.read_val('betas') if 'weight_decay' in self.opt_keys: self._wd = self.read_val('weight_decay') def set_val(self, key: str, val, bn_groups: bool = True): "Set `val` inside the optimizer dictionary at `key`." if is_tuple(val): val = [(v1, v2) for v1, v2 in zip(*val)] for v, pg1, pg2 in zip(val, self.opt.param_groups[::2], self.opt.param_groups[1::2]): pg1[key] = v if bn_groups: pg2[key] = v return val def read_val(self, key: str): "Read a hyperparameter `key` in the optimizer dictionary." val = [pg[key] for pg in self.opt.param_groups[::2]] if is_tuple(val[0]): val = [o[0] for o in val], [o[1] for o in val] return val class FastAIMixedOptim(OptimWrapper): @classmethod def create(cls, opt_func, lr, layer_groups, model, flat_master=False, loss_scale=512.0, **kwargs): "Create an `optim.Optimizer` from `opt_func` with `lr`. Set lr on `layer_groups`." opt = OptimWrapper.create(opt_func, lr, layer_groups, **kwargs) opt.model_params, opt.master_params = get_master(layer_groups, flat_master) opt.flat_master = flat_master opt.loss_scale = loss_scale opt.model = model # Changes the optimizer so that the optimization step is done in FP32. # opt = self.learn.opt mom, wd, beta = opt.mom, opt.wd, opt.beta lrs = [lr for lr in opt._lr for _ in range(2)] opt_params = [{'params': mp, 'lr': lr} for mp, lr in zip(opt.master_params, lrs)] opt.opt = opt_func(opt_params) opt.mom, opt.wd, opt.beta = mom, wd, beta return opt def step(self): model_g2master_g(self.model_params, self.master_params, self.flat_master) for group in self.master_params: for param in group: param.grad.div_(self.loss_scale) super(FastAIMixedOptim, self).step() self.model.zero_grad() # Update the params from master to model. master2model(self.model_params, self.master_params, self.flat_master)
10,477
38.992366
117
py
SASA
SASA-main/tools/train_utils/optimization/learning_schedules_fastai.py
# This file is modified from https://github.com/traveller59/second.pytorch import math from functools import partial import numpy as np import torch.optim.lr_scheduler as lr_sched from .fastai_optim import OptimWrapper class LRSchedulerStep(object): def __init__(self, fai_optimizer: OptimWrapper, total_step, lr_phases, mom_phases): # if not isinstance(fai_optimizer, OptimWrapper): # raise TypeError('{} is not a fastai OptimWrapper'.format( # type(fai_optimizer).__name__)) self.optimizer = fai_optimizer self.total_step = total_step self.lr_phases = [] for i, (start, lambda_func) in enumerate(lr_phases): if len(self.lr_phases) != 0: assert self.lr_phases[-1][0] < start if isinstance(lambda_func, str): lambda_func = eval(lambda_func) if i < len(lr_phases) - 1: self.lr_phases.append((int(start * total_step), int(lr_phases[i + 1][0] * total_step), lambda_func)) else: self.lr_phases.append((int(start * total_step), total_step, lambda_func)) assert self.lr_phases[0][0] == 0 self.mom_phases = [] for i, (start, lambda_func) in enumerate(mom_phases): if len(self.mom_phases) != 0: assert self.mom_phases[-1][0] < start if isinstance(lambda_func, str): lambda_func = eval(lambda_func) if i < len(mom_phases) - 1: self.mom_phases.append((int(start * total_step), int(mom_phases[i + 1][0] * total_step), lambda_func)) else: self.mom_phases.append((int(start * total_step), total_step, lambda_func)) assert self.mom_phases[0][0] == 0 def step(self, step): for start, end, func in self.lr_phases: if step >= start: self.optimizer.lr = func((step - start) / (end - start)) for start, end, func in self.mom_phases: if step >= start: self.optimizer.mom = func((step - start) / (end - start)) def annealing_cos(start, end, pct): # print(pct, start, end) "Cosine anneal from `start` to `end` as pct goes from 0.0 to 1.0." cos_out = np.cos(np.pi * pct) + 1 return end + (start - end) / 2 * cos_out class OneCycle(LRSchedulerStep): def __init__(self, fai_optimizer, total_step, lr_max, moms, div_factor, pct_start): self.lr_max = lr_max self.moms = moms self.div_factor = div_factor self.pct_start = pct_start a1 = int(total_step * self.pct_start) a2 = total_step - a1 low_lr = self.lr_max / self.div_factor lr_phases = ((0, partial(annealing_cos, low_lr, self.lr_max)), (self.pct_start, partial(annealing_cos, self.lr_max, low_lr / 1e4))) mom_phases = ((0, partial(annealing_cos, *self.moms)), (self.pct_start, partial(annealing_cos, *self.moms[::-1]))) fai_optimizer.lr, fai_optimizer.mom = low_lr, self.moms[0] super().__init__(fai_optimizer, total_step, lr_phases, mom_phases) class CosineWarmupLR(lr_sched._LRScheduler): def __init__(self, optimizer, T_max, eta_min=0, last_epoch=-1): self.T_max = T_max self.eta_min = eta_min super(CosineWarmupLR, self).__init__(optimizer, last_epoch) def get_lr(self): return [self.eta_min + (base_lr - self.eta_min) * (1 - math.cos(math.pi * self.last_epoch / self.T_max)) / 2 for base_lr in self.base_lrs] class FakeOptim: def __init__(self): self.lr = 0 self.mom = 0 if __name__ == "__main__": import matplotlib.pyplot as plt opt = FakeOptim() # 3e-3, wd=0.4, div_factor=10 schd = OneCycle(opt, 100, 3e-3, (0.95, 0.85), 10.0, 0.1) lrs = [] moms = [] for i in range(100): schd.step(i) lrs.append(opt.lr) moms.append(opt.mom) plt.plot(lrs) # plt.plot(moms) plt.show() plt.plot(moms) plt.show()
4,169
35.26087
118
py
SASA
SASA-main/tools/train_utils/optimization/__init__.py
from functools import partial import torch.nn as nn import torch.optim as optim import torch.optim.lr_scheduler as lr_sched from .fastai_optim import OptimWrapper from .learning_schedules_fastai import CosineWarmupLR, OneCycle def build_optimizer(model, optim_cfg): if optim_cfg.OPTIMIZER == 'sgd': optimizer = optim.SGD( model.parameters(), lr=optim_cfg.LR, weight_decay=optim_cfg.WEIGHT_DECAY, momentum=optim_cfg.MOMENTUM ) elif optim_cfg.OPTIMIZER == 'adam' or optim_cfg.OPTIMIZER == 'adam_onecycle': def children(m: nn.Module): return list(m.children()) def num_children(m: nn.Module) -> int: return len(children(m)) flatten_model = lambda m: sum(map(flatten_model, m.children()), []) if num_children(m) else [m] get_layer_groups = lambda m: [nn.Sequential(*flatten_model(m))] optimizer_func = partial(optim.Adam, betas=(0.9, 0.99)) optimizer = OptimWrapper.create( optimizer_func, 3e-3, get_layer_groups(model), wd=optim_cfg.WEIGHT_DECAY, true_wd=True, bn_wd=True ) else: raise NotImplementedError return optimizer def build_scheduler(optimizer, total_iters_each_epoch, total_epochs, last_epoch, optim_cfg): decay_steps = [x * total_iters_each_epoch for x in optim_cfg.DECAY_STEP_LIST] def lr_lbmd(cur_epoch): cur_decay = 1 for decay_step in decay_steps: if cur_epoch >= decay_step: cur_decay = cur_decay * optim_cfg.LR_DECAY return max(cur_decay, optim_cfg.LR_CLIP / optim_cfg.LR) lr_warmup_scheduler = None total_steps = total_iters_each_epoch * total_epochs if optim_cfg.OPTIMIZER == 'adam_onecycle': lr_scheduler = OneCycle( optimizer, total_steps, optim_cfg.LR, list(optim_cfg.MOMS), optim_cfg.DIV_FACTOR, optim_cfg.PCT_START ) else: lr_scheduler = lr_sched.LambdaLR(optimizer, lr_lbmd, last_epoch=last_epoch) if optim_cfg.LR_WARMUP: lr_warmup_scheduler = CosineWarmupLR( optimizer, T_max=optim_cfg.WARMUP_EPOCH * len(total_iters_each_epoch), eta_min=optim_cfg.LR / optim_cfg.DIV_FACTOR ) return lr_scheduler, lr_warmup_scheduler
2,289
35.935484
113
py
SASA
SASA-main/tools/visual_utils/visualize_utils.py
import mayavi.mlab as mlab import numpy as np import torch box_colormap = [ [1, 1, 1], [0, 1, 0], [0, 1, 1], [1, 1, 0], ] def check_numpy_to_torch(x): if isinstance(x, np.ndarray): return torch.from_numpy(x).float(), True return x, False def rotate_points_along_z(points, angle): """ Args: points: (B, N, 3 + C) angle: (B), angle along z-axis, angle increases x ==> y Returns: """ points, is_numpy = check_numpy_to_torch(points) angle, _ = check_numpy_to_torch(angle) cosa = torch.cos(angle) sina = torch.sin(angle) zeros = angle.new_zeros(points.shape[0]) ones = angle.new_ones(points.shape[0]) rot_matrix = torch.stack(( cosa, sina, zeros, -sina, cosa, zeros, zeros, zeros, ones ), dim=1).view(-1, 3, 3).float() points_rot = torch.matmul(points[:, :, 0:3], rot_matrix) points_rot = torch.cat((points_rot, points[:, :, 3:]), dim=-1) return points_rot.numpy() if is_numpy else points_rot def boxes_to_corners_3d(boxes3d): """ 7 -------- 4 /| /| 6 -------- 5 . | | | | . 3 -------- 0 |/ |/ 2 -------- 1 Args: boxes3d: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center Returns: """ boxes3d, is_numpy = check_numpy_to_torch(boxes3d) template = boxes3d.new_tensor(( [1, 1, -1], [1, -1, -1], [-1, -1, -1], [-1, 1, -1], [1, 1, 1], [1, -1, 1], [-1, -1, 1], [-1, 1, 1], )) / 2 corners3d = boxes3d[:, None, 3:6].repeat(1, 8, 1) * template[None, :, :] corners3d = rotate_points_along_z(corners3d.view(-1, 8, 3), boxes3d[:, 6]).view(-1, 8, 3) corners3d += boxes3d[:, None, 0:3] return corners3d.numpy() if is_numpy else corners3d def visualize_pts(pts, fig=None, bgcolor=(0, 0, 0), fgcolor=(1.0, 1.0, 1.0), show_intensity=False, size=(800, 600), draw_origin=True): if not isinstance(pts, np.ndarray): pts = pts.cpu().numpy() if fig is None: fig = mlab.figure(figure=None, bgcolor=bgcolor, fgcolor=fgcolor, engine=None, size=size) if show_intensity: G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], pts[:, 3], mode='point', colormap='gnuplot', scale_factor=1, figure=fig) else: G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], mode='point', colormap='gnuplot', scale_factor=1, figure=fig) if draw_origin: mlab.points3d(0, 0, 0, color=(1, 1, 1), mode='cube', scale_factor=0.2) mlab.plot3d([0, 3], [0, 0], [0, 0], color=(0, 0, 1), tube_radius=0.1) mlab.plot3d([0, 0], [0, 3], [0, 0], color=(0, 1, 0), tube_radius=0.1) mlab.plot3d([0, 0], [0, 0], [0, 3], color=(1, 0, 0), tube_radius=0.1) return fig def draw_sphere_pts(pts, color=(0, 1, 0), fig=None, bgcolor=(0, 0, 0), scale_factor=0.2): if not isinstance(pts, np.ndarray): pts = pts.cpu().numpy() if fig is None: fig = mlab.figure(figure=None, bgcolor=bgcolor, fgcolor=None, engine=None, size=(600, 600)) if isinstance(color, np.ndarray) and color.shape[0] == 1: color = color[0] color = (color[0] / 255.0, color[1] / 255.0, color[2] / 255.0) if isinstance(color, np.ndarray): pts_color = np.zeros((pts.__len__(), 4), dtype=np.uint8) pts_color[:, 0:3] = color pts_color[:, 3] = 255 G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], np.arange(0, pts_color.__len__()), mode='sphere', scale_factor=scale_factor, figure=fig) G.glyph.color_mode = 'color_by_scalar' G.glyph.scale_mode = 'scale_by_vector' G.module_manager.scalar_lut_manager.lut.table = pts_color else: mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], mode='sphere', color=color, colormap='gnuplot', scale_factor=scale_factor, figure=fig) mlab.points3d(0, 0, 0, color=(1, 1, 1), mode='cube', scale_factor=0.2) mlab.plot3d([0, 3], [0, 0], [0, 0], color=(0, 0, 1), line_width=3, tube_radius=None, figure=fig) mlab.plot3d([0, 0], [0, 3], [0, 0], color=(0, 1, 0), line_width=3, tube_radius=None, figure=fig) mlab.plot3d([0, 0], [0, 0], [0, 3], color=(1, 0, 0), line_width=3, tube_radius=None, figure=fig) return fig def draw_grid(x1, y1, x2, y2, fig, tube_radius=None, color=(0.5, 0.5, 0.5)): mlab.plot3d([x1, x1], [y1, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x2, x2], [y1, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x1, x2], [y1, y1], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x1, x2], [y2, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) return fig def draw_multi_grid_range(fig, grid_size=20, bv_range=(-60, -60, 60, 60)): for x in range(bv_range[0], bv_range[2], grid_size): for y in range(bv_range[1], bv_range[3], grid_size): fig = draw_grid(x, y, x + grid_size, y + grid_size, fig) return fig def draw_scenes(points, gt_boxes=None, ref_boxes=None, ref_scores=None, ref_labels=None): if not isinstance(points, np.ndarray): points = points.cpu().numpy() if ref_boxes is not None and not isinstance(ref_boxes, np.ndarray): ref_boxes = ref_boxes.cpu().numpy() if gt_boxes is not None and not isinstance(gt_boxes, np.ndarray): gt_boxes = gt_boxes.cpu().numpy() if ref_scores is not None and not isinstance(ref_scores, np.ndarray): ref_scores = ref_scores.cpu().numpy() if ref_labels is not None and not isinstance(ref_labels, np.ndarray): ref_labels = ref_labels.cpu().numpy() fig = visualize_pts(points) fig = draw_multi_grid_range(fig, bv_range=(0, -40, 80, 40)) if gt_boxes is not None: corners3d = boxes_to_corners_3d(gt_boxes) fig = draw_corners3d(corners3d, fig=fig, color=(0, 0, 1), max_num=100) if ref_boxes is not None and len(ref_boxes) > 0: ref_corners3d = boxes_to_corners_3d(ref_boxes) if ref_labels is None: fig = draw_corners3d(ref_corners3d, fig=fig, color=(0, 1, 0), cls=ref_scores, max_num=100) else: for k in range(ref_labels.min(), ref_labels.max() + 1): cur_color = tuple(box_colormap[k % len(box_colormap)]) mask = (ref_labels == k) fig = draw_corners3d(ref_corners3d[mask], fig=fig, color=cur_color, cls=ref_scores[mask], max_num=100) mlab.view(azimuth=-179, elevation=54.0, distance=104.0, roll=90.0) return fig def draw_corners3d(corners3d, fig, color=(1, 1, 1), line_width=2, cls=None, tag='', max_num=500, tube_radius=None): """ :param corners3d: (N, 8, 3) :param fig: :param color: :param line_width: :param cls: :param tag: :param max_num: :return: """ import mayavi.mlab as mlab num = min(max_num, len(corners3d)) for n in range(num): b = corners3d[n] # (8, 3) if cls is not None: if isinstance(cls, np.ndarray): mlab.text3d(b[6, 0], b[6, 1], b[6, 2], '%.2f' % cls[n], scale=(0.3, 0.3, 0.3), color=color, figure=fig) else: mlab.text3d(b[6, 0], b[6, 1], b[6, 2], '%s' % cls[n], scale=(0.3, 0.3, 0.3), color=color, figure=fig) for k in range(0, 4): i, j = k, (k + 1) % 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = k + 4, (k + 1) % 4 + 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = k, k + 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = 0, 5 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = 1, 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) return fig
8,540
38.541667
121
py
SASA
SASA-main/pcdet/models/__init__.py
from collections import namedtuple import numpy as np import torch from .detectors import build_detector def build_network(model_cfg, num_class, dataset): model = build_detector( model_cfg=model_cfg, num_class=num_class, dataset=dataset ) return model def load_data_to_gpu(batch_dict): for key, val in batch_dict.items(): if not isinstance(val, np.ndarray): continue if key in ['frame_id', 'metadata', 'calib', 'image_shape']: continue batch_dict[key] = torch.from_numpy(val).float().cuda() def model_fn_decorator(): ModelReturn = namedtuple('ModelReturn', ['loss', 'tb_dict', 'disp_dict']) def model_func(model, batch_dict): load_data_to_gpu(batch_dict) ret_dict, tb_dict, disp_dict = model(batch_dict) loss = ret_dict['loss'].mean() if hasattr(model, 'update_global_step'): model.update_global_step() else: model.module.update_global_step() return ModelReturn(loss, tb_dict, disp_dict) return model_func
1,074
25.219512
77
py
SASA
SASA-main/pcdet/models/detectors/detector3d_template.py
import os import torch import torch.nn as nn from ...ops.iou3d_nms import iou3d_nms_utils from .. import backbones_2d, backbones_3d, dense_heads, roi_heads from ..backbones_2d import map_to_bev from ..backbones_3d import pfe, vfe from ..model_utils import model_nms_utils class Detector3DTemplate(nn.Module): def __init__(self, model_cfg, num_class, dataset): super().__init__() self.model_cfg = model_cfg self.num_class = num_class self.dataset = dataset self.class_names = dataset.class_names self.register_buffer('global_step', torch.LongTensor(1).zero_()) self.module_topology = [ 'vfe', 'backbone_3d', 'map_to_bev_module', 'pfe', 'backbone_2d', 'dense_head', 'point_head', 'roi_head' ] @property def mode(self): return 'TRAIN' if self.training else 'TEST' def update_global_step(self): self.global_step += 1 def build_networks(self): model_info_dict = { 'module_list': [], 'num_rawpoint_features': self.dataset.point_feature_encoder.num_point_features, 'num_point_features': self.dataset.point_feature_encoder.num_point_features, 'grid_size': self.dataset.grid_size, 'point_cloud_range': self.dataset.point_cloud_range, 'voxel_size': self.dataset.voxel_size } for module_name in self.module_topology: module, model_info_dict = getattr(self, 'build_%s' % module_name)( model_info_dict=model_info_dict ) self.add_module(module_name, module) return model_info_dict['module_list'] def build_vfe(self, model_info_dict): if self.model_cfg.get('VFE', None) is None: return None, model_info_dict vfe_module = vfe.__all__[self.model_cfg.VFE.NAME]( model_cfg=self.model_cfg.VFE, num_point_features=model_info_dict['num_rawpoint_features'], point_cloud_range=model_info_dict['point_cloud_range'], voxel_size=model_info_dict['voxel_size'] ) model_info_dict['num_point_features'] = vfe_module.get_output_feature_dim() model_info_dict['module_list'].append(vfe_module) return vfe_module, model_info_dict def build_backbone_3d(self, model_info_dict): if self.model_cfg.get('BACKBONE_3D', None) is None: return None, model_info_dict backbone_3d_module = backbones_3d.__all__[self.model_cfg.BACKBONE_3D.NAME]( model_cfg=self.model_cfg.BACKBONE_3D, input_channels=model_info_dict['num_point_features'], grid_size=model_info_dict['grid_size'], voxel_size=model_info_dict['voxel_size'], point_cloud_range=model_info_dict['point_cloud_range'] ) model_info_dict['module_list'].append(backbone_3d_module) model_info_dict['num_point_features'] = backbone_3d_module.num_point_features return backbone_3d_module, model_info_dict def build_map_to_bev_module(self, model_info_dict): if self.model_cfg.get('MAP_TO_BEV', None) is None: return None, model_info_dict map_to_bev_module = map_to_bev.__all__[self.model_cfg.MAP_TO_BEV.NAME]( model_cfg=self.model_cfg.MAP_TO_BEV, grid_size=model_info_dict['grid_size'] ) model_info_dict['module_list'].append(map_to_bev_module) model_info_dict['num_bev_features'] = map_to_bev_module.num_bev_features return map_to_bev_module, model_info_dict def build_backbone_2d(self, model_info_dict): if self.model_cfg.get('BACKBONE_2D', None) is None: return None, model_info_dict backbone_2d_module = backbones_2d.__all__[self.model_cfg.BACKBONE_2D.NAME]( model_cfg=self.model_cfg.BACKBONE_2D, input_channels=model_info_dict['num_bev_features'] ) model_info_dict['module_list'].append(backbone_2d_module) model_info_dict['num_bev_features'] = backbone_2d_module.num_bev_features return backbone_2d_module, model_info_dict def build_pfe(self, model_info_dict): if self.model_cfg.get('PFE', None) is None: return None, model_info_dict pfe_module = pfe.__all__[self.model_cfg.PFE.NAME]( model_cfg=self.model_cfg.PFE, voxel_size=model_info_dict['voxel_size'], point_cloud_range=model_info_dict['point_cloud_range'], num_bev_features=model_info_dict['num_bev_features'], num_rawpoint_features=model_info_dict['num_rawpoint_features'] ) model_info_dict['module_list'].append(pfe_module) model_info_dict['num_point_features'] = pfe_module.num_point_features model_info_dict['num_point_features_before_fusion'] = pfe_module.num_point_features_before_fusion return pfe_module, model_info_dict def build_dense_head(self, model_info_dict): if self.model_cfg.get('DENSE_HEAD', None) is None: return None, model_info_dict dense_head_module = dense_heads.__all__[self.model_cfg.DENSE_HEAD.NAME]( model_cfg=self.model_cfg.DENSE_HEAD, input_channels=model_info_dict['num_bev_features'], num_class=self.num_class if not self.model_cfg.DENSE_HEAD.CLASS_AGNOSTIC else 1, class_names=self.class_names, grid_size=model_info_dict['grid_size'], point_cloud_range=model_info_dict['point_cloud_range'], predict_boxes_when_training=self.model_cfg.get('ROI_HEAD', False) ) model_info_dict['module_list'].append(dense_head_module) return dense_head_module, model_info_dict def build_point_head(self, model_info_dict): if self.model_cfg.get('POINT_HEAD', None) is None: return None, model_info_dict if self.model_cfg.POINT_HEAD.get('USE_POINT_FEATURES_BEFORE_FUSION', False): num_point_features = model_info_dict['num_point_features_before_fusion'] else: num_point_features = model_info_dict['num_point_features'] point_head_module = dense_heads.__all__[self.model_cfg.POINT_HEAD.NAME]( model_cfg=self.model_cfg.POINT_HEAD, input_channels=num_point_features, num_class=self.num_class if not self.model_cfg.POINT_HEAD.CLASS_AGNOSTIC else 1, predict_boxes_when_training=self.model_cfg.get('ROI_HEAD', False) ) model_info_dict['module_list'].append(point_head_module) return point_head_module, model_info_dict def build_roi_head(self, model_info_dict): if self.model_cfg.get('ROI_HEAD', None) is None: return None, model_info_dict point_head_module = roi_heads.__all__[self.model_cfg.ROI_HEAD.NAME]( model_cfg=self.model_cfg.ROI_HEAD, input_channels=model_info_dict['num_point_features'], num_class=self.num_class if not self.model_cfg.ROI_HEAD.CLASS_AGNOSTIC else 1, ) model_info_dict['module_list'].append(point_head_module) return point_head_module, model_info_dict def forward(self, **kwargs): raise NotImplementedError def post_processing(self, batch_dict): """ Args: batch_dict: batch_size: batch_cls_preds: (B, num_boxes, num_classes | 1) or (N1+N2+..., num_classes | 1) or [(B, num_boxes, num_class1), (B, num_boxes, num_class2) ...] multihead_label_mapping: [(num_class1), (num_class2), ...] batch_box_preds: (B, num_boxes, 7+C) or (N1+N2+..., 7+C) cls_preds_normalized: indicate whether batch_cls_preds is normalized batch_index: optional (N1+N2+...) has_class_labels: True/False roi_labels: (B, num_rois) 1 .. num_classes batch_pred_labels: (B, num_boxes, 1) Returns: """ post_process_cfg = self.model_cfg.POST_PROCESSING batch_size = batch_dict['batch_size'] recall_dict = {} pred_dicts = [] for index in range(batch_size): if batch_dict.get('batch_index', None) is not None: assert batch_dict['batch_box_preds'].shape.__len__() == 2 batch_mask = (batch_dict['batch_index'] == index) else: assert batch_dict['batch_box_preds'].shape.__len__() == 3 batch_mask = index box_preds = batch_dict['batch_box_preds'][batch_mask] src_box_preds = box_preds if not isinstance(batch_dict['batch_cls_preds'], list): cls_preds = batch_dict['batch_cls_preds'][batch_mask] src_cls_preds = cls_preds assert cls_preds.shape[1] in [1, self.num_class] if not batch_dict['cls_preds_normalized']: cls_preds = torch.sigmoid(cls_preds) else: cls_preds = [x[batch_mask] for x in batch_dict['batch_cls_preds']] src_cls_preds = cls_preds if not batch_dict['cls_preds_normalized']: cls_preds = [torch.sigmoid(x) for x in cls_preds] if post_process_cfg.NMS_CONFIG.MULTI_CLASSES_NMS: if not isinstance(cls_preds, list): cls_preds = [cls_preds] multihead_label_mapping = [torch.arange(1, self.num_class, device=cls_preds[0].device)] else: multihead_label_mapping = batch_dict['multihead_label_mapping'] cur_start_idx = 0 pred_scores, pred_labels, pred_boxes = [], [], [] for cur_cls_preds, cur_label_mapping in zip(cls_preds, multihead_label_mapping): assert cur_cls_preds.shape[1] == len(cur_label_mapping) cur_box_preds = box_preds[cur_start_idx: cur_start_idx + cur_cls_preds.shape[0]] cur_pred_scores, cur_pred_labels, cur_pred_boxes = model_nms_utils.multi_classes_nms( cls_scores=cur_cls_preds, box_preds=cur_box_preds, nms_config=post_process_cfg.NMS_CONFIG, score_thresh=post_process_cfg.SCORE_THRESH ) cur_pred_labels = cur_label_mapping[cur_pred_labels] pred_scores.append(cur_pred_scores) pred_labels.append(cur_pred_labels) pred_boxes.append(cur_pred_boxes) cur_start_idx += cur_cls_preds.shape[0] final_scores = torch.cat(pred_scores, dim=0) final_labels = torch.cat(pred_labels, dim=0) final_boxes = torch.cat(pred_boxes, dim=0) else: cls_preds, label_preds = torch.max(cls_preds, dim=-1) if batch_dict.get('has_class_labels', False): label_key = 'roi_labels' if 'roi_labels' in batch_dict else 'batch_pred_labels' label_preds = batch_dict[label_key][index] else: label_preds = label_preds + 1 selected, selected_scores = model_nms_utils.class_agnostic_nms( box_scores=cls_preds, box_preds=box_preds, nms_config=post_process_cfg.NMS_CONFIG, score_thresh=post_process_cfg.SCORE_THRESH ) if post_process_cfg.OUTPUT_RAW_SCORE: max_cls_preds, _ = torch.max(src_cls_preds, dim=-1) selected_scores = max_cls_preds[selected] final_scores = selected_scores final_labels = label_preds[selected] final_boxes = box_preds[selected] recall_dict = self.generate_recall_record( box_preds=final_boxes if 'rois' not in batch_dict else src_box_preds, recall_dict=recall_dict, batch_index=index, data_dict=batch_dict, thresh_list=post_process_cfg.RECALL_THRESH_LIST ) record_dict = { 'pred_boxes': final_boxes, 'pred_scores': final_scores, 'pred_labels': final_labels } pred_dicts.append(record_dict) return pred_dicts, recall_dict def generate_recall_record(self, box_preds, recall_dict, batch_index, data_dict=None, thresh_list=None): if 'gt_boxes' not in data_dict: return recall_dict rois = data_dict['rois'][batch_index] if 'rois' in data_dict else None gt_boxes = data_dict['gt_boxes'][batch_index] if recall_dict.__len__() == 0: recall_dict = {'gt_num': 0} for cur_thresh in thresh_list: recall_dict['roi_%s' % (str(cur_thresh))] = 0 recall_dict['rcnn_%s' % (str(cur_thresh))] = 0 cur_gt = gt_boxes k = cur_gt.__len__() - 1 while k > 0 and cur_gt[k].sum() == 0: k -= 1 cur_gt = cur_gt[:k + 1] if cur_gt.shape[0] > 0: if box_preds.shape[0] > 0: iou3d_rcnn = iou3d_nms_utils.boxes_iou3d_gpu(box_preds[:, 0:7], cur_gt[:, 0:7]) else: iou3d_rcnn = torch.zeros((0, cur_gt.shape[0])) if rois is not None: iou3d_roi = iou3d_nms_utils.boxes_iou3d_gpu(rois[:, 0:7], cur_gt[:, 0:7]) for cur_thresh in thresh_list: if iou3d_rcnn.shape[0] == 0: recall_dict['rcnn_%s' % str(cur_thresh)] += 0 else: rcnn_recalled = (iou3d_rcnn.max(dim=0)[0] > cur_thresh).sum().item() recall_dict['rcnn_%s' % str(cur_thresh)] += rcnn_recalled if rois is not None: roi_recalled = (iou3d_roi.max(dim=0)[0] > cur_thresh).sum().item() recall_dict['roi_%s' % str(cur_thresh)] += roi_recalled recall_dict['gt_num'] += cur_gt.shape[0] else: gt_iou = box_preds.new_zeros(box_preds.shape[0]) return recall_dict def load_params_from_file(self, filename, logger, to_cpu=False): if not os.path.isfile(filename): raise FileNotFoundError logger.info('==> Loading parameters from checkpoint %s to %s' % (filename, 'CPU' if to_cpu else 'GPU')) loc_type = torch.device('cpu') if to_cpu else None checkpoint = torch.load(filename, map_location=loc_type) model_state_disk = checkpoint['model_state'] if 'version' in checkpoint: logger.info('==> Checkpoint trained from version: %s' % checkpoint['version']) update_model_state = {} for key, val in model_state_disk.items(): if key in self.state_dict() and self.state_dict()[key].shape == model_state_disk[key].shape: update_model_state[key] = val # logger.info('Update weight %s: %s' % (key, str(val.shape))) state_dict = self.state_dict() state_dict.update(update_model_state) self.load_state_dict(state_dict) for key in state_dict: if key not in update_model_state: logger.info('Not updated weight %s: %s' % (key, str(state_dict[key].shape))) logger.info('==> Done (loaded %d/%d)' % (len(update_model_state), len(self.state_dict()))) def load_params_with_optimizer(self, filename, to_cpu=False, optimizer=None, logger=None): if not os.path.isfile(filename): raise FileNotFoundError logger.info('==> Loading parameters from checkpoint %s to %s' % (filename, 'CPU' if to_cpu else 'GPU')) loc_type = torch.device('cpu') if to_cpu else None checkpoint = torch.load(filename, map_location=loc_type) epoch = checkpoint.get('epoch', -1) it = checkpoint.get('it', 0.0) self.load_state_dict(checkpoint['model_state']) if optimizer is not None: if 'optimizer_state' in checkpoint and checkpoint['optimizer_state'] is not None: logger.info('==> Loading optimizer parameters from checkpoint %s to %s' % (filename, 'CPU' if to_cpu else 'GPU')) optimizer.load_state_dict(checkpoint['optimizer_state']) else: assert filename[-4] == '.', filename src_file, ext = filename[:-4], filename[-3:] optimizer_filename = '%s_optim.%s' % (src_file, ext) if os.path.exists(optimizer_filename): optimizer_ckpt = torch.load(optimizer_filename, map_location=loc_type) optimizer.load_state_dict(optimizer_ckpt['optimizer_state']) if 'version' in checkpoint: print('==> Checkpoint trained from version: %s' % checkpoint['version']) logger.info('==> Done') return it, epoch
17,009
44.119363
111
py
SASA
SASA-main/pcdet/models/detectors/point_3dssd.py
import torch from .detector3d_template import Detector3DTemplate from ...ops.iou3d_nms import iou3d_nms_utils from ...ops.roiaware_pool3d import roiaware_pool3d_utils class Point3DSSD(Detector3DTemplate): def __init__(self, model_cfg, num_class, dataset): super().__init__(model_cfg=model_cfg, num_class=num_class, dataset=dataset) self.module_list = self.build_networks() def forward(self, batch_dict): for cur_module in self.module_list: batch_dict = cur_module(batch_dict) if self.training: loss, tb_dict, disp_dict = self.get_training_loss() ret_dict = { 'loss': loss } return ret_dict, tb_dict, disp_dict else: pred_dicts, recall_dicts = self.post_processing(batch_dict) return pred_dicts, recall_dicts def get_training_loss(self): disp_dict = {} loss_point, tb_dict = self.point_head.get_loss() loss = loss_point return loss, tb_dict, disp_dict def init_recall_record(self, metric, **kwargs): # initialize gt_num for all classes for cur_cls in range(len(self.class_names)): metric['gt_num[%s]' % self.class_names[cur_cls]] = 0 # initialize statistics of all sampling segments npoint_list = self.model_cfg.BACKBONE_3D.SA_CONFIG.NPOINT_LIST for cur_layer in range(len(npoint_list)): for cur_seg in range(len(npoint_list[cur_layer])): metric['positive_point_L%dS%d' % (cur_layer, cur_seg)] = 0 metric['recall_point_L%dS%d' % (cur_layer, cur_seg)] = 0 for cur_cls in range(self.num_class): metric['recall_point_L%dS%d[%s]' \ % (cur_layer, cur_seg, self.class_names[cur_cls])] = 0 # initialize statistics of the vote layer metric['positive_point_candidate'] = 0 metric['recall_point_candidate'] = 0 metric['positive_point_vote'] = 0 metric['recall_point_vote'] = 0 for cur_cls in range(len(self.class_names)): metric['recall_point_candidate[%s]' % self.class_names[cur_cls]] = 0 metric['recall_point_vote[%s]' % self.class_names[cur_cls]] = 0 def generate_recall_record(self, box_preds, recall_dict, batch_index, data_dict=None, thresh_list=None): if 'gt_boxes' not in data_dict: return recall_dict # point_coords format: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] point_list = data_dict['point_coords_list'] # ignore raw point input npoint_list = self.model_cfg.BACKBONE_3D.SA_CONFIG.NPOINT_LIST assert len(point_list) == len(npoint_list) cur_points_list = [] for cur_layer in range(npoint_list.__len__()): cur_points = point_list[cur_layer] bs_idx = cur_points[:, 0] bs_mask = (bs_idx == batch_index) cur_points = cur_points[bs_mask][:, 1:4] cur_points_list.append(cur_points.split(npoint_list[cur_layer], dim=0)) base_points = data_dict['point_candidate_coords'] vote_points = data_dict['point_vote_coords'] bs_idx = base_points[:, 0] bs_mask = (bs_idx == batch_index) base_points = base_points[bs_mask][:, 1:4] vote_points = vote_points[bs_mask][:, 1:4] rois = data_dict['rois'][batch_index] if 'rois' in data_dict else None gt_boxes = data_dict['gt_boxes'][batch_index] # initialize recall_dict if recall_dict.__len__() == 0: recall_dict = {'gt_num': 0} for cur_thresh in thresh_list: recall_dict['recall_roi_%s' % (str(cur_thresh))] = 0 recall_dict['recall_rcnn_%s' % (str(cur_thresh))] = 0 self.init_recall_record(recall_dict) # init customized statistics cur_gt = gt_boxes k = cur_gt.__len__() - 1 while k > 0 and cur_gt[k].sum() == 0: k -= 1 cur_gt = cur_gt[:k + 1] if cur_gt.shape[0] > 0: # backbone for cur_layer in range(len(npoint_list)): for cur_seg in range(len(npoint_list[cur_layer])): box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( cur_points_list[cur_layer][cur_seg].unsqueeze(dim=0), cur_gt[None, :, :7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) recall_dict['positive_point_L%dS%d' % (cur_layer, cur_seg)] += box_fg_flag.long().sum().item() box_recalled = box_idxs_of_pts[box_fg_flag].unique() recall_dict['recall_point_L%dS%d' % (cur_layer, cur_seg)] += box_recalled.size(0) box_recalled_cls = cur_gt[box_recalled, -1] for cur_cls in range(self.num_class): recall_dict['recall_point_L%dS%d[%s]' % (cur_layer, cur_seg, self.class_names[cur_cls])] += \ (box_recalled_cls == (cur_cls + 1)).sum().item() # candidate points box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( base_points.unsqueeze(dim=0), cur_gt[None, :, :7] ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) recall_dict['positive_point_candidate'] += box_fg_flag.long().sum().item() box_recalled = box_idxs_of_pts[box_fg_flag].unique() recall_dict['recall_point_candidate'] += box_recalled.size(0) box_recalled_cls = cur_gt[box_recalled, -1] for cur_cls in range(self.num_class): recall_dict['recall_point_candidate[%s]' % self.class_names[cur_cls]] += \ (box_recalled_cls == (cur_cls + 1)).sum().item() # vote points box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( vote_points.unsqueeze(dim=0), cur_gt[None, :, :7] ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) recall_dict['positive_point_vote'] += box_fg_flag.long().sum().item() box_recalled = box_idxs_of_pts[box_fg_flag].unique() recall_dict['recall_point_vote'] += box_recalled.size(0) box_recalled_cls = cur_gt[box_recalled, -1] for cur_cls in range(self.num_class): recall_dict['recall_point_vote[%s]' % self.class_names[cur_cls]] += \ (box_recalled_cls == (cur_cls + 1)).sum().item() if box_preds.shape[0] > 0: iou3d_rcnn = iou3d_nms_utils.boxes_iou3d_gpu(box_preds[:, 0:7], cur_gt[:, 0:7]) else: iou3d_rcnn = torch.zeros((0, cur_gt.shape[0])) if rois is not None: iou3d_roi = iou3d_nms_utils.boxes_iou3d_gpu(rois[:, 0:7], cur_gt[:, 0:7]) for cur_thresh in thresh_list: if iou3d_rcnn.shape[0] == 0: recall_dict['recall_rcnn_%s' % str(cur_thresh)] += 0 else: rcnn_recalled = (iou3d_rcnn.max(dim=0)[0] > cur_thresh).sum().item() recall_dict['recall_rcnn_%s' % str(cur_thresh)] += rcnn_recalled if rois is not None: roi_recalled = (iou3d_roi.max(dim=0)[0] > cur_thresh).sum().item() recall_dict['recall_roi_%s' % str(cur_thresh)] += roi_recalled cur_gt_class = cur_gt[:, -1] for cur_cls in range(self.num_class): cur_cls_gt_num = (cur_gt_class == cur_cls + 1).sum().item() recall_dict['gt_num'] += cur_cls_gt_num recall_dict['gt_num[%s]' % self.class_names[cur_cls]] += cur_cls_gt_num return recall_dict def disp_recall_record(self, metric, logger, sample_num, **kwargs): gt_num = metric['gt_num'] gt_num_cls = [metric['gt_num[%s]' % cur_cls] for cur_cls in self.class_names] # backbone for k in metric.keys(): if 'positive_point_' in k: # count the number of positive points cur_positive_point = metric[k] / sample_num logger.info(k + (': %f' % cur_positive_point)) elif 'recall_point_' in k and not any(cur_cls in k for cur_cls in self.class_names): cur_recall_point = metric[k] / max(gt_num, 1) logger.info(k + (': %f' % cur_recall_point)) for cur_cls in range(len(self.class_names)): cur_recall_point_cls = metric[k + '[%s]' % self.class_names[cur_cls]] / max(gt_num_cls[cur_cls], 1) logger.info('\t- ' + self.class_names[cur_cls] + ': %f' % cur_recall_point_cls) # candidate points positive_point_candidate = metric['positive_point_candidate'] / sample_num logger.info('positive_point_candidate: %f' % positive_point_candidate) recall_point_candidate = metric['recall_point_candidate'] / max(gt_num, 1) logger.info('recall_point_candidate: %f' % recall_point_candidate) for cur_cls in range(len(self.class_names)): cur_recall_point_cls = metric['recall_point_candidate' + '[%s]' % self.class_names[cur_cls]] / max(gt_num_cls[cur_cls], 1) logger.info('\t- ' + self.class_names[cur_cls] + ': %f' % cur_recall_point_cls) # vote points positive_point_vote = metric['positive_point_vote'] / sample_num logger.info('positive_point_vote: %f' % positive_point_vote) recall_point_vote = metric['recall_point_vote'] / max(gt_num, 1) logger.info('recall_point_vote: %f' % recall_point_vote) for cur_cls in range(len(self.class_names)): cur_recall_point_cls = metric['recall_point_vote' + '[%s]' % self.class_names[cur_cls]] / max(gt_num_cls[cur_cls], 1) logger.info('\t- ' + self.class_names[cur_cls] + ': %f' % cur_recall_point_cls)
10,022
47.892683
134
py
SASA
SASA-main/pcdet/models/backbones_3d/spconv_unet.py
from functools import partial import spconv import torch import torch.nn as nn from ...utils import common_utils from .spconv_backbone import post_act_block class SparseBasicBlock(spconv.SparseModule): expansion = 1 def __init__(self, inplanes, planes, stride=1, downsample=None, indice_key=None, norm_fn=None): super(SparseBasicBlock, self).__init__() self.conv1 = spconv.SubMConv3d( inplanes, planes, kernel_size=3, stride=stride, padding=1, bias=False, indice_key=indice_key ) self.bn1 = norm_fn(planes) self.relu = nn.ReLU() self.conv2 = spconv.SubMConv3d( planes, planes, kernel_size=3, stride=1, padding=1, bias=False, indice_key=indice_key ) self.bn2 = norm_fn(planes) self.downsample = downsample self.stride = stride def forward(self, x): identity = x.features assert x.features.dim() == 2, 'x.features.dim()=%d' % x.features.dim() out = self.conv1(x) out.features = self.bn1(out.features) out.features = self.relu(out.features) out = self.conv2(out) out.features = self.bn2(out.features) if self.downsample is not None: identity = self.downsample(x) out.features += identity out.features = self.relu(out.features) return out class UNetV2(nn.Module): """ Sparse Convolution based UNet for point-wise feature learning. Reference Paper: https://arxiv.org/abs/1907.03670 (Shaoshuai Shi, et. al) From Points to Parts: 3D Object Detection from Point Cloud with Part-aware and Part-aggregation Network """ def __init__(self, model_cfg, input_channels, grid_size, voxel_size, point_cloud_range, **kwargs): super().__init__() self.model_cfg = model_cfg self.sparse_shape = grid_size[::-1] + [1, 0, 0] self.voxel_size = voxel_size self.point_cloud_range = point_cloud_range norm_fn = partial(nn.BatchNorm1d, eps=1e-3, momentum=0.01) self.conv_input = spconv.SparseSequential( spconv.SubMConv3d(input_channels, 16, 3, padding=1, bias=False, indice_key='subm1'), norm_fn(16), nn.ReLU(), ) block = post_act_block self.conv1 = spconv.SparseSequential( block(16, 16, 3, norm_fn=norm_fn, padding=1, indice_key='subm1'), ) self.conv2 = spconv.SparseSequential( # [1600, 1408, 41] <- [800, 704, 21] block(16, 32, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv2', conv_type='spconv'), block(32, 32, 3, norm_fn=norm_fn, padding=1, indice_key='subm2'), block(32, 32, 3, norm_fn=norm_fn, padding=1, indice_key='subm2'), ) self.conv3 = spconv.SparseSequential( # [800, 704, 21] <- [400, 352, 11] block(32, 64, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv3', conv_type='spconv'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm3'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm3'), ) self.conv4 = spconv.SparseSequential( # [400, 352, 11] <- [200, 176, 5] block(64, 64, 3, norm_fn=norm_fn, stride=2, padding=(0, 1, 1), indice_key='spconv4', conv_type='spconv'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm4'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm4'), ) if self.model_cfg.get('RETURN_ENCODED_TENSOR', True): last_pad = self.model_cfg.get('last_pad', 0) self.conv_out = spconv.SparseSequential( # [200, 150, 5] -> [200, 150, 2] spconv.SparseConv3d(64, 128, (3, 1, 1), stride=(2, 1, 1), padding=last_pad, bias=False, indice_key='spconv_down2'), norm_fn(128), nn.ReLU(), ) else: self.conv_out = None # decoder # [400, 352, 11] <- [200, 176, 5] self.conv_up_t4 = SparseBasicBlock(64, 64, indice_key='subm4', norm_fn=norm_fn) self.conv_up_m4 = block(128, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm4') self.inv_conv4 = block(64, 64, 3, norm_fn=norm_fn, indice_key='spconv4', conv_type='inverseconv') # [800, 704, 21] <- [400, 352, 11] self.conv_up_t3 = SparseBasicBlock(64, 64, indice_key='subm3', norm_fn=norm_fn) self.conv_up_m3 = block(128, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm3') self.inv_conv3 = block(64, 32, 3, norm_fn=norm_fn, indice_key='spconv3', conv_type='inverseconv') # [1600, 1408, 41] <- [800, 704, 21] self.conv_up_t2 = SparseBasicBlock(32, 32, indice_key='subm2', norm_fn=norm_fn) self.conv_up_m2 = block(64, 32, 3, norm_fn=norm_fn, indice_key='subm2') self.inv_conv2 = block(32, 16, 3, norm_fn=norm_fn, indice_key='spconv2', conv_type='inverseconv') # [1600, 1408, 41] <- [1600, 1408, 41] self.conv_up_t1 = SparseBasicBlock(16, 16, indice_key='subm1', norm_fn=norm_fn) self.conv_up_m1 = block(32, 16, 3, norm_fn=norm_fn, indice_key='subm1') self.conv5 = spconv.SparseSequential( block(16, 16, 3, norm_fn=norm_fn, padding=1, indice_key='subm1') ) self.num_point_features = 16 def UR_block_forward(self, x_lateral, x_bottom, conv_t, conv_m, conv_inv): x_trans = conv_t(x_lateral) x = x_trans x.features = torch.cat((x_bottom.features, x_trans.features), dim=1) x_m = conv_m(x) x = self.channel_reduction(x, x_m.features.shape[1]) x.features = x_m.features + x.features x = conv_inv(x) return x @staticmethod def channel_reduction(x, out_channels): """ Args: x: x.features (N, C1) out_channels: C2 Returns: """ features = x.features n, in_channels = features.shape assert (in_channels % out_channels == 0) and (in_channels >= out_channels) x.features = features.view(n, out_channels, -1).sum(dim=2) return x def forward(self, batch_dict): """ Args: batch_dict: batch_size: int vfe_features: (num_voxels, C) voxel_coords: (num_voxels, 4), [batch_idx, z_idx, y_idx, x_idx] Returns: batch_dict: encoded_spconv_tensor: sparse tensor point_features: (N, C) """ voxel_features, voxel_coords = batch_dict['voxel_features'], batch_dict['voxel_coords'] batch_size = batch_dict['batch_size'] input_sp_tensor = spconv.SparseConvTensor( features=voxel_features, indices=voxel_coords.int(), spatial_shape=self.sparse_shape, batch_size=batch_size ) x = self.conv_input(input_sp_tensor) x_conv1 = self.conv1(x) x_conv2 = self.conv2(x_conv1) x_conv3 = self.conv3(x_conv2) x_conv4 = self.conv4(x_conv3) if self.conv_out is not None: # for detection head # [200, 176, 5] -> [200, 176, 2] out = self.conv_out(x_conv4) batch_dict['encoded_spconv_tensor'] = out batch_dict['encoded_spconv_tensor_stride'] = 8 # for segmentation head # [400, 352, 11] <- [200, 176, 5] x_up4 = self.UR_block_forward(x_conv4, x_conv4, self.conv_up_t4, self.conv_up_m4, self.inv_conv4) # [800, 704, 21] <- [400, 352, 11] x_up3 = self.UR_block_forward(x_conv3, x_up4, self.conv_up_t3, self.conv_up_m3, self.inv_conv3) # [1600, 1408, 41] <- [800, 704, 21] x_up2 = self.UR_block_forward(x_conv2, x_up3, self.conv_up_t2, self.conv_up_m2, self.inv_conv2) # [1600, 1408, 41] <- [1600, 1408, 41] x_up1 = self.UR_block_forward(x_conv1, x_up2, self.conv_up_t1, self.conv_up_m1, self.conv5) batch_dict['point_features'] = x_up1.features point_coords = common_utils.get_voxel_centers( x_up1.indices[:, 1:], downsample_times=1, voxel_size=self.voxel_size, point_cloud_range=self.point_cloud_range ) batch_dict['point_coords'] = torch.cat((x_up1.indices[:, 0:1].float(), point_coords), dim=1) return batch_dict
8,445
38.839623
117
py
SASA
SASA-main/pcdet/models/backbones_3d/spconv_backbone.py
from functools import partial import spconv import torch.nn as nn def post_act_block(in_channels, out_channels, kernel_size, indice_key=None, stride=1, padding=0, conv_type='subm', norm_fn=None): if conv_type == 'subm': conv = spconv.SubMConv3d(in_channels, out_channels, kernel_size, bias=False, indice_key=indice_key) elif conv_type == 'spconv': conv = spconv.SparseConv3d(in_channels, out_channels, kernel_size, stride=stride, padding=padding, bias=False, indice_key=indice_key) elif conv_type == 'inverseconv': conv = spconv.SparseInverseConv3d(in_channels, out_channels, kernel_size, indice_key=indice_key, bias=False) else: raise NotImplementedError m = spconv.SparseSequential( conv, norm_fn(out_channels), nn.ReLU(), ) return m class SparseBasicBlock(spconv.SparseModule): expansion = 1 def __init__(self, inplanes, planes, stride=1, norm_fn=None, downsample=None, indice_key=None): super(SparseBasicBlock, self).__init__() assert norm_fn is not None bias = norm_fn is not None self.conv1 = spconv.SubMConv3d( inplanes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key ) self.bn1 = norm_fn(planes) self.relu = nn.ReLU() self.conv2 = spconv.SubMConv3d( planes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key ) self.bn2 = norm_fn(planes) self.downsample = downsample self.stride = stride def forward(self, x): identity = x out = self.conv1(x) out.features = self.bn1(out.features) out.features = self.relu(out.features) out = self.conv2(out) out.features = self.bn2(out.features) if self.downsample is not None: identity = self.downsample(x) out.features += identity.features out.features = self.relu(out.features) return out class VoxelBackBone8x(nn.Module): def __init__(self, model_cfg, input_channels, grid_size, **kwargs): super().__init__() self.model_cfg = model_cfg norm_fn = partial(nn.BatchNorm1d, eps=1e-3, momentum=0.01) self.sparse_shape = grid_size[::-1] + [1, 0, 0] self.conv_input = spconv.SparseSequential( spconv.SubMConv3d(input_channels, 16, 3, padding=1, bias=False, indice_key='subm1'), norm_fn(16), nn.ReLU(), ) block = post_act_block self.conv1 = spconv.SparseSequential( block(16, 16, 3, norm_fn=norm_fn, padding=1, indice_key='subm1'), ) self.conv2 = spconv.SparseSequential( # [1600, 1408, 41] <- [800, 704, 21] block(16, 32, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv2', conv_type='spconv'), block(32, 32, 3, norm_fn=norm_fn, padding=1, indice_key='subm2'), block(32, 32, 3, norm_fn=norm_fn, padding=1, indice_key='subm2'), ) self.conv3 = spconv.SparseSequential( # [800, 704, 21] <- [400, 352, 11] block(32, 64, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv3', conv_type='spconv'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm3'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm3'), ) self.conv4 = spconv.SparseSequential( # [400, 352, 11] <- [200, 176, 5] block(64, 64, 3, norm_fn=norm_fn, stride=2, padding=(0, 1, 1), indice_key='spconv4', conv_type='spconv'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm4'), block(64, 64, 3, norm_fn=norm_fn, padding=1, indice_key='subm4'), ) last_pad = 0 last_pad = self.model_cfg.get('last_pad', last_pad) self.conv_out = spconv.SparseSequential( # [200, 150, 5] -> [200, 150, 2] spconv.SparseConv3d(64, 128, (3, 1, 1), stride=(2, 1, 1), padding=last_pad, bias=False, indice_key='spconv_down2'), norm_fn(128), nn.ReLU(), ) self.num_point_features = 128 def forward(self, batch_dict): """ Args: batch_dict: batch_size: int vfe_features: (num_voxels, C) voxel_coords: (num_voxels, 4), [batch_idx, z_idx, y_idx, x_idx] Returns: batch_dict: encoded_spconv_tensor: sparse tensor """ voxel_features, voxel_coords = batch_dict['voxel_features'], batch_dict['voxel_coords'] batch_size = batch_dict['batch_size'] input_sp_tensor = spconv.SparseConvTensor( features=voxel_features, indices=voxel_coords.int(), spatial_shape=self.sparse_shape, batch_size=batch_size ) x = self.conv_input(input_sp_tensor) x_conv1 = self.conv1(x) x_conv2 = self.conv2(x_conv1) x_conv3 = self.conv3(x_conv2) x_conv4 = self.conv4(x_conv3) # for detection head # [200, 176, 5] -> [200, 176, 2] out = self.conv_out(x_conv4) batch_dict.update({ 'encoded_spconv_tensor': out, 'encoded_spconv_tensor_stride': 8 }) batch_dict.update({ 'multi_scale_3d_features': { 'x_conv1': x_conv1, 'x_conv2': x_conv2, 'x_conv3': x_conv3, 'x_conv4': x_conv4, } }) return batch_dict class VoxelResBackBone8x(nn.Module): def __init__(self, model_cfg, input_channels, grid_size, **kwargs): super().__init__() self.model_cfg = model_cfg norm_fn = partial(nn.BatchNorm1d, eps=1e-3, momentum=0.01) self.sparse_shape = grid_size[::-1] + [1, 0, 0] self.conv_input = spconv.SparseSequential( spconv.SubMConv3d(input_channels, 16, 3, padding=1, bias=False, indice_key='subm1'), norm_fn(16), nn.ReLU(), ) block = post_act_block self.conv1 = spconv.SparseSequential( SparseBasicBlock(16, 16, norm_fn=norm_fn, indice_key='res1'), SparseBasicBlock(16, 16, norm_fn=norm_fn, indice_key='res1'), ) self.conv2 = spconv.SparseSequential( # [1600, 1408, 41] <- [800, 704, 21] block(16, 32, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv2', conv_type='spconv'), SparseBasicBlock(32, 32, norm_fn=norm_fn, indice_key='res2'), SparseBasicBlock(32, 32, norm_fn=norm_fn, indice_key='res2'), ) self.conv3 = spconv.SparseSequential( # [800, 704, 21] <- [400, 352, 11] block(32, 64, 3, norm_fn=norm_fn, stride=2, padding=1, indice_key='spconv3', conv_type='spconv'), SparseBasicBlock(64, 64, norm_fn=norm_fn, indice_key='res3'), SparseBasicBlock(64, 64, norm_fn=norm_fn, indice_key='res3'), ) self.conv4 = spconv.SparseSequential( # [400, 352, 11] <- [200, 176, 5] block(64, 128, 3, norm_fn=norm_fn, stride=2, padding=(0, 1, 1), indice_key='spconv4', conv_type='spconv'), SparseBasicBlock(128, 128, norm_fn=norm_fn, indice_key='res4'), SparseBasicBlock(128, 128, norm_fn=norm_fn, indice_key='res4'), ) last_pad = 0 last_pad = self.model_cfg.get('last_pad', last_pad) self.conv_out = spconv.SparseSequential( # [200, 150, 5] -> [200, 150, 2] spconv.SparseConv3d(128, 128, (3, 1, 1), stride=(2, 1, 1), padding=last_pad, bias=False, indice_key='spconv_down2'), norm_fn(128), nn.ReLU(), ) self.num_point_features = 128 def forward(self, batch_dict): """ Args: batch_dict: batch_size: int vfe_features: (num_voxels, C) voxel_coords: (num_voxels, 4), [batch_idx, z_idx, y_idx, x_idx] Returns: batch_dict: encoded_spconv_tensor: sparse tensor """ voxel_features, voxel_coords = batch_dict['voxel_features'], batch_dict['voxel_coords'] batch_size = batch_dict['batch_size'] input_sp_tensor = spconv.SparseConvTensor( features=voxel_features, indices=voxel_coords.int(), spatial_shape=self.sparse_shape, batch_size=batch_size ) x = self.conv_input(input_sp_tensor) x_conv1 = self.conv1(x) x_conv2 = self.conv2(x_conv1) x_conv3 = self.conv3(x_conv2) x_conv4 = self.conv4(x_conv3) # for detection head # [200, 176, 5] -> [200, 176, 2] out = self.conv_out(x_conv4) batch_dict.update({ 'encoded_spconv_tensor': out, 'encoded_spconv_tensor_stride': 8 }) batch_dict.update({ 'multi_scale_3d_features': { 'x_conv1': x_conv1, 'x_conv2': x_conv2, 'x_conv3': x_conv3, 'x_conv4': x_conv4, } }) return batch_dict
9,376
34.790076
118
py
SASA
SASA-main/pcdet/models/backbones_3d/pointnet2_backbone.py
import torch import torch.nn as nn from ...ops.pointnet2.pointnet2_batch import pointnet2_modules from ...ops.pointnet2.pointnet2_stack import pointnet2_modules as pointnet2_modules_stack from ...ops.pointnet2.pointnet2_stack import pointnet2_utils as pointnet2_utils_stack class PointNet2MSG(nn.Module): def __init__(self, model_cfg, input_channels, **kwargs): super().__init__() self.model_cfg = model_cfg self.SA_modules = nn.ModuleList() channel_in = input_channels - 3 self.num_points_each_layer = [] skip_channel_list = [input_channels - 3] for k in range(self.model_cfg.SA_CONFIG.NPOINTS.__len__()): mlps = self.model_cfg.SA_CONFIG.MLPS[k].copy() channel_out = 0 for idx in range(mlps.__len__()): mlps[idx] = [channel_in] + mlps[idx] channel_out += mlps[idx][-1] self.SA_modules.append( pointnet2_modules.PointnetSAModuleMSG( npoint=self.model_cfg.SA_CONFIG.NPOINTS[k], radii=self.model_cfg.SA_CONFIG.RADIUS[k], nsamples=self.model_cfg.SA_CONFIG.NSAMPLE[k], mlps=mlps, use_xyz=self.model_cfg.SA_CONFIG.get('USE_XYZ', True), ) ) skip_channel_list.append(channel_out) channel_in = channel_out self.FP_modules = nn.ModuleList() skip_sa_block = self.model_cfg.SA_CONFIG.NPOINTS.__len__() - self.model_cfg.FP_MLPS.__len__() for k in range(self.model_cfg.FP_MLPS.__len__()): pre_channel = self.model_cfg.FP_MLPS[k + 1][-1] if k + 1 < len(self.model_cfg.FP_MLPS) else channel_out self.FP_modules.append( pointnet2_modules.PointnetFPModule( mlp=[pre_channel + skip_channel_list[k + skip_sa_block]] + self.model_cfg.FP_MLPS[k] ) ) self.num_point_features = self.model_cfg.FP_MLPS[0][-1] def break_up_pc(self, pc): batch_idx = pc[:, 0] xyz = pc[:, 1:4].contiguous() features = (pc[:, 4:].contiguous() if pc.size(-1) > 4 else None) return batch_idx, xyz, features def forward(self, batch_dict): """ Args: batch_dict: batch_size: int vfe_features: (num_voxels, C) points: (num_points, 4 + C), [batch_idx, x, y, z, ...] Returns: batch_dict: encoded_spconv_tensor: sparse tensor point_features: (N, C) """ batch_size = batch_dict['batch_size'] points = batch_dict['points'] batch_idx, xyz, features = self.break_up_pc(points) xyz_batch_cnt = xyz.new_zeros(batch_size).int() for bs_idx in range(batch_size): xyz_batch_cnt[bs_idx] = (batch_idx == bs_idx).sum() assert xyz_batch_cnt.min() == xyz_batch_cnt.max() xyz = xyz.view(batch_size, -1, 3) features = features.view(batch_size, -1, features.shape[-1]).permute(0, 2, 1) if features is not None else None l_xyz, l_features = [xyz], [features] for i in range(len(self.SA_modules)): li_xyz, li_features = self.SA_modules[i](l_xyz[i], l_features[i]) l_xyz.append(li_xyz) l_features.append(li_features) for i in range(-1, -(len(self.FP_modules) + 1), -1): l_features[i - 1] = self.FP_modules[i]( l_xyz[i - 1], l_xyz[i], l_features[i - 1], l_features[i] ) # (B, C, N) point_features = l_features[0].permute(0, 2, 1).contiguous() # (B, N, C) batch_dict['point_features'] = point_features.view(-1, point_features.shape[-1]) batch_dict['point_coords'] = torch.cat((batch_idx[:, None].float(), l_xyz[0].view(-1, 3)), dim=1) return batch_dict class PointNet2FSMSG(nn.Module): def __init__(self, model_cfg, input_channels, **kwargs): super().__init__() self.model_cfg = model_cfg self.SA_modules = nn.ModuleList() channel_in = input_channels - 3 use_xyz = self.model_cfg.SA_CONFIG.get('USE_XYZ', True) dilated_group = self.model_cfg.SA_CONFIG.get('DILATED_RADIUS_GROUP', False) skip_connection = self.model_cfg.SA_CONFIG.get('SKIP_CONNECTION', False) weight_gamma = self.model_cfg.SA_CONFIG.get('WEIGHT_GAMMA', 1.0) self.aggregation_mlps = self.model_cfg.SA_CONFIG.get('AGGREGATION_MLPS', None) self.confidence_mlps = self.model_cfg.SA_CONFIG.get('CONFIDENCE_MLPS', None) self.num_points_each_layer = [] skip_channel_list = [input_channels - 3] for k in range(self.model_cfg.SA_CONFIG.NPOINT_LIST.__len__()): mlps = self.model_cfg.SA_CONFIG.MLPS[k].copy() channel_out = 0 for idx in range(mlps.__len__()): mlps[idx] = [channel_in] + mlps[idx] channel_out += mlps[idx][-1] if skip_connection: channel_out += channel_in if self.aggregation_mlps and self.aggregation_mlps[k]: aggregation_mlp = self.aggregation_mlps[k].copy() if aggregation_mlp.__len__() == 0: aggregation_mlp = None else: channel_out = aggregation_mlp[-1] else: aggregation_mlp = None if self.confidence_mlps and self.confidence_mlps[k]: confidence_mlp = self.confidence_mlps[k].copy() if confidence_mlp.__len__() == 0: confidence_mlp = None else: confidence_mlp = None self.SA_modules.append( pointnet2_modules.PointnetSAModuleFSMSG( npoint_list=self.model_cfg.SA_CONFIG.NPOINT_LIST[k], sample_range_list=self.model_cfg.SA_CONFIG.SAMPLE_RANGE_LIST[k], sample_method_list=self.model_cfg.SA_CONFIG.SAMPLE_METHOD_LIST[k], radii=self.model_cfg.SA_CONFIG.RADIUS[k], nsamples=self.model_cfg.SA_CONFIG.NSAMPLE[k], mlps=mlps, use_xyz=use_xyz, dilated_radius_group=dilated_group, skip_connection=skip_connection, weight_gamma=weight_gamma, aggregation_mlp=aggregation_mlp, confidence_mlp=confidence_mlp ) ) self.num_points_each_layer.append( sum(self.model_cfg.SA_CONFIG.NPOINT_LIST[k])) skip_channel_list.append(channel_out) channel_in = channel_out self.num_point_features = channel_out fp_mlps = self.model_cfg.get('FP_MLPS', None) if fp_mlps is not None: self.FP_modules = nn.ModuleList() l_skipped = self.model_cfg.SA_CONFIG.NPOINT_LIST.__len__() - self.model_cfg.FP_MLPS.__len__() for k in range(fp_mlps.__len__()): pre_channel = fp_mlps[k + 1][-1] if k + 1 < len(fp_mlps) else channel_out self.FP_modules.append( pointnet2_modules.PointnetFPModule( mlp=[pre_channel + skip_channel_list[k + l_skipped]] + fp_mlps[k] ) ) self.num_point_features = fp_mlps[0][-1] else: self.FP_modules = None def break_up_pc(self, pc): batch_idx = pc[:, 0] xyz = pc[:, 1:4].contiguous() features = (pc[:, 4:].contiguous() if pc.size(-1) > 4 else None) return batch_idx, xyz, features def forward(self, batch_dict): """ Args: batch_dict: batch_size: int points: (num_points, 4 + C), [batch_idx, x, y, z, ...] Returns: batch_dict: point_coords: (N, 3) point_features: (N, C) point_confidence_scores: (N, 1) """ batch_size = batch_dict['batch_size'] points = batch_dict['points'] batch_idx, xyz, features = self.break_up_pc(points) xyz_batch_cnt = xyz.new_zeros(batch_size).int() for bs_idx in range(batch_size): xyz_batch_cnt[bs_idx] = (batch_idx == bs_idx).sum() assert xyz_batch_cnt.min() == xyz_batch_cnt.max() xyz = xyz.view(batch_size, -1, 3).contiguous() features = features.view(batch_size, -1, features.shape[-1]) if features is not None else None features = features.permute(0, 2, 1).contiguous() if features is not None else None batch_idx = batch_idx.view(batch_size, -1).float() l_xyz, l_features, l_scores = [xyz], [features], [None] for i in range(len(self.SA_modules)): li_xyz, li_features, li_scores = self.SA_modules[i]( l_xyz[i], l_features[i], scores=l_scores[i]) l_xyz.append(li_xyz) l_features.append(li_features) l_scores.append(li_scores) # prepare for confidence loss l_xyz_flatten, l_scores_flatten = [], [] for i in range(1, len(l_xyz)): l_xyz_flatten.append(torch.cat([ batch_idx[:, :l_xyz[i].size(1)].reshape(-1, 1), l_xyz[i].reshape(-1, 3) ], dim=1)) # (N, 4) for i in range(1, len(l_scores)): if l_scores[i] is None: l_scores_flatten.append(None) else: l_scores_flatten.append(l_scores[i].reshape(-1, 1)) # (N, 1) batch_dict['point_coords_list'] = l_xyz_flatten batch_dict['point_scores_list'] = l_scores_flatten if self.FP_modules is not None: for i in range(-1, -(len(self.FP_modules) + 1), -1): l_features[i - 1] = self.FP_modules[i]( l_xyz[i - 1], l_xyz[i], l_features[i - 1], l_features[i] ) # (B, C, N) else: # take l_xyz[i - 1] and l_features[i - 1] i = 0 point_features = l_features[i - 1].permute(0, 2, 1).contiguous() # (B, N, C) batch_dict['point_features'] = point_features.view(-1, point_features.shape[-1]) batch_dict['point_coords'] = torch.cat(( batch_idx[:, :l_xyz[i - 1].size(1)].reshape(-1, 1).float(), l_xyz[i - 1].view(-1, 3)), dim=1) batch_dict['point_scores'] = l_scores[-1] # (B, N) return batch_dict class PointNet2Backbone(nn.Module): """ DO NOT USE THIS CURRENTLY SINCE IT MAY HAVE POTENTIAL BUGS, 20200723 """ def __init__(self, model_cfg, input_channels, **kwargs): assert False, 'DO NOT USE THIS CURRENTLY SINCE IT MAY HAVE POTENTIAL BUGS, 20200723' super().__init__() self.model_cfg = model_cfg self.SA_modules = nn.ModuleList() channel_in = input_channels - 3 self.num_points_each_layer = [] skip_channel_list = [input_channels] for k in range(self.model_cfg.SA_CONFIG.NPOINTS.__len__()): self.num_points_each_layer.append(self.model_cfg.SA_CONFIG.NPOINTS[k]) mlps = self.model_cfg.SA_CONFIG.MLPS[k].copy() channel_out = 0 for idx in range(mlps.__len__()): mlps[idx] = [channel_in] + mlps[idx] channel_out += mlps[idx][-1] self.SA_modules.append( pointnet2_modules_stack.StackSAModuleMSG( radii=self.model_cfg.SA_CONFIG.RADIUS[k], nsamples=self.model_cfg.SA_CONFIG.NSAMPLE[k], mlps=mlps, use_xyz=self.model_cfg.SA_CONFIG.get('USE_XYZ', True), ) ) skip_channel_list.append(channel_out) channel_in = channel_out self.FP_modules = nn.ModuleList() for k in range(self.model_cfg.FP_MLPS.__len__()): pre_channel = self.model_cfg.FP_MLPS[k + 1][-1] if k + 1 < len(self.model_cfg.FP_MLPS) else channel_out self.FP_modules.append( pointnet2_modules_stack.StackPointnetFPModule( mlp=[pre_channel + skip_channel_list[k]] + self.model_cfg.FP_MLPS[k] ) ) self.num_point_features = self.model_cfg.FP_MLPS[0][-1] def break_up_pc(self, pc): batch_idx = pc[:, 0] xyz = pc[:, 1:4].contiguous() features = (pc[:, 4:].contiguous() if pc.size(-1) > 4 else None) return batch_idx, xyz, features def forward(self, batch_dict): """ Args: batch_dict: batch_size: int vfe_features: (num_voxels, C) points: (num_points, 4 + C), [batch_idx, x, y, z, ...] Returns: batch_dict: encoded_spconv_tensor: sparse tensor point_features: (N, C) """ batch_size = batch_dict['batch_size'] points = batch_dict['points'] batch_idx, xyz, features = self.break_up_pc(points) xyz_batch_cnt = xyz.new_zeros(batch_size).int() for bs_idx in range(batch_size): xyz_batch_cnt[bs_idx] = (batch_idx == bs_idx).sum() l_xyz, l_features, l_batch_cnt = [xyz], [features], [xyz_batch_cnt] for i in range(len(self.SA_modules)): new_xyz_list = [] for k in range(batch_size): if len(l_xyz) == 1: cur_xyz = l_xyz[0][batch_idx == k] else: last_num_points = self.num_points_each_layer[i - 1] cur_xyz = l_xyz[-1][k * last_num_points: (k + 1) * last_num_points] cur_pt_idxs = pointnet2_utils_stack.furthest_point_sample( cur_xyz[None, :, :].contiguous(), self.num_points_each_layer[i] ).long()[0] if cur_xyz.shape[0] < self.num_points_each_layer[i]: empty_num = self.num_points_each_layer[i] - cur_xyz.shape[1] cur_pt_idxs[0, -empty_num:] = cur_pt_idxs[0, :empty_num] new_xyz_list.append(cur_xyz[cur_pt_idxs]) new_xyz = torch.cat(new_xyz_list, dim=0) new_xyz_batch_cnt = xyz.new_zeros(batch_size).int().fill_(self.num_points_each_layer[i]) li_xyz, li_features = self.SA_modules[i]( xyz=l_xyz[i], features=l_features[i], xyz_batch_cnt=l_batch_cnt[i], new_xyz=new_xyz, new_xyz_batch_cnt=new_xyz_batch_cnt ) l_xyz.append(li_xyz) l_features.append(li_features) l_batch_cnt.append(new_xyz_batch_cnt) l_features[0] = points[:, 1:] for i in range(-1, -(len(self.FP_modules) + 1), -1): l_features[i - 1] = self.FP_modules[i]( unknown=l_xyz[i - 1], unknown_batch_cnt=l_batch_cnt[i - 1], known=l_xyz[i], known_batch_cnt=l_batch_cnt[i], unknown_feats=l_features[i - 1], known_feats=l_features[i] ) batch_dict['point_features'] = l_features[0] batch_dict['point_coords'] = torch.cat((batch_idx[:, None].float(), l_xyz[0]), dim=1) return batch_dict
15,273
41.077135
119
py
SASA
SASA-main/pcdet/models/backbones_3d/pfe/voxel_set_abstraction.py
import torch import torch.nn as nn from ....ops.pointnet2.pointnet2_stack import pointnet2_modules as pointnet2_stack_modules from ....ops.pointnet2.pointnet2_stack import pointnet2_utils as pointnet2_stack_utils from ....utils import common_utils def bilinear_interpolate_torch(im, x, y): """ Args: im: (H, W, C) [y, x] x: (N) y: (N) Returns: """ x0 = torch.floor(x).long() x1 = x0 + 1 y0 = torch.floor(y).long() y1 = y0 + 1 x0 = torch.clamp(x0, 0, im.shape[1] - 1) x1 = torch.clamp(x1, 0, im.shape[1] - 1) y0 = torch.clamp(y0, 0, im.shape[0] - 1) y1 = torch.clamp(y1, 0, im.shape[0] - 1) Ia = im[y0, x0] Ib = im[y1, x0] Ic = im[y0, x1] Id = im[y1, x1] wa = (x1.type_as(x) - x) * (y1.type_as(y) - y) wb = (x1.type_as(x) - x) * (y - y0.type_as(y)) wc = (x - x0.type_as(x)) * (y1.type_as(y) - y) wd = (x - x0.type_as(x)) * (y - y0.type_as(y)) ans = torch.t((torch.t(Ia) * wa)) + torch.t(torch.t(Ib) * wb) + torch.t(torch.t(Ic) * wc) + torch.t(torch.t(Id) * wd) return ans class VoxelSetAbstraction(nn.Module): def __init__(self, model_cfg, voxel_size, point_cloud_range, num_bev_features=None, num_rawpoint_features=None, **kwargs): super().__init__() self.model_cfg = model_cfg self.voxel_size = voxel_size self.point_cloud_range = point_cloud_range SA_cfg = self.model_cfg.SA_LAYER self.SA_layers = nn.ModuleList() self.SA_layer_names = [] self.downsample_times_map = {} c_in = 0 for src_name in self.model_cfg.FEATURES_SOURCE: if src_name in ['bev', 'raw_points']: continue self.downsample_times_map[src_name] = SA_cfg[src_name].DOWNSAMPLE_FACTOR mlps = SA_cfg[src_name].MLPS for k in range(len(mlps)): mlps[k] = [mlps[k][0]] + mlps[k] cur_layer = pointnet2_stack_modules.StackSAModuleMSG( radii=SA_cfg[src_name].POOL_RADIUS, nsamples=SA_cfg[src_name].NSAMPLE, mlps=mlps, use_xyz=True, pool_method='max_pool', ) self.SA_layers.append(cur_layer) self.SA_layer_names.append(src_name) c_in += sum([x[-1] for x in mlps]) if 'bev' in self.model_cfg.FEATURES_SOURCE: c_bev = num_bev_features c_in += c_bev if 'raw_points' in self.model_cfg.FEATURES_SOURCE: mlps = SA_cfg['raw_points'].MLPS for k in range(len(mlps)): mlps[k] = [num_rawpoint_features - 3] + mlps[k] self.SA_rawpoints = pointnet2_stack_modules.StackSAModuleMSG( radii=SA_cfg['raw_points'].POOL_RADIUS, nsamples=SA_cfg['raw_points'].NSAMPLE, mlps=mlps, use_xyz=True, pool_method='max_pool' ) c_in += sum([x[-1] for x in mlps]) self.vsa_point_feature_fusion = nn.Sequential( nn.Linear(c_in, self.model_cfg.NUM_OUTPUT_FEATURES, bias=False), nn.BatchNorm1d(self.model_cfg.NUM_OUTPUT_FEATURES), nn.ReLU(), ) self.num_point_features = self.model_cfg.NUM_OUTPUT_FEATURES self.num_point_features_before_fusion = c_in def interpolate_from_bev_features(self, keypoints, bev_features, batch_size, bev_stride): x_idxs = (keypoints[:, :, 0] - self.point_cloud_range[0]) / self.voxel_size[0] y_idxs = (keypoints[:, :, 1] - self.point_cloud_range[1]) / self.voxel_size[1] x_idxs = x_idxs / bev_stride y_idxs = y_idxs / bev_stride point_bev_features_list = [] for k in range(batch_size): cur_x_idxs = x_idxs[k] cur_y_idxs = y_idxs[k] cur_bev_features = bev_features[k].permute(1, 2, 0) # (H, W, C) point_bev_features = bilinear_interpolate_torch(cur_bev_features, cur_x_idxs, cur_y_idxs) point_bev_features_list.append(point_bev_features.unsqueeze(dim=0)) point_bev_features = torch.cat(point_bev_features_list, dim=0) # (B, N, C0) return point_bev_features def get_sampled_points(self, batch_dict): batch_size = batch_dict['batch_size'] if self.model_cfg.POINT_SOURCE == 'raw_points': src_points = batch_dict['points'][:, 1:4] batch_indices = batch_dict['points'][:, 0].long() elif self.model_cfg.POINT_SOURCE == 'voxel_centers': src_points = common_utils.get_voxel_centers( batch_dict['voxel_coords'][:, 1:4], downsample_times=1, voxel_size=self.voxel_size, point_cloud_range=self.point_cloud_range ) batch_indices = batch_dict['voxel_coords'][:, 0].long() else: raise NotImplementedError keypoints_list = [] for bs_idx in range(batch_size): bs_mask = (batch_indices == bs_idx) sampled_points = src_points[bs_mask].unsqueeze(dim=0) # (1, N, 3) if self.model_cfg.SAMPLE_METHOD == 'FPS': cur_pt_idxs = pointnet2_stack_utils.furthest_point_sample( sampled_points[:, :, 0:3].contiguous(), self.model_cfg.NUM_KEYPOINTS ).long() if sampled_points.shape[1] < self.model_cfg.NUM_KEYPOINTS: empty_num = self.model_cfg.NUM_KEYPOINTS - sampled_points.shape[1] cur_pt_idxs[0, -empty_num:] = cur_pt_idxs[0, :empty_num] keypoints = sampled_points[0][cur_pt_idxs[0]].unsqueeze(dim=0) elif self.model_cfg.SAMPLE_METHOD == 'FastFPS': raise NotImplementedError else: raise NotImplementedError keypoints_list.append(keypoints) keypoints = torch.cat(keypoints_list, dim=0) # (B, M, 3) return keypoints def forward(self, batch_dict): """ Args: batch_dict: batch_size: keypoints: (B, num_keypoints, 3) multi_scale_3d_features: { 'x_conv4': ... } points: optional (N, 1 + 3 + C) [bs_idx, x, y, z, ...] spatial_features: optional spatial_features_stride: optional Returns: point_features: (N, C) point_coords: (N, 4) """ keypoints = self.get_sampled_points(batch_dict) point_features_list = [] if 'bev' in self.model_cfg.FEATURES_SOURCE: point_bev_features = self.interpolate_from_bev_features( keypoints, batch_dict['spatial_features'], batch_dict['batch_size'], bev_stride=batch_dict['spatial_features_stride'] ) point_features_list.append(point_bev_features) batch_size, num_keypoints, _ = keypoints.shape new_xyz = keypoints.view(-1, 3) new_xyz_batch_cnt = new_xyz.new_zeros(batch_size).int().fill_(num_keypoints) if 'raw_points' in self.model_cfg.FEATURES_SOURCE: raw_points = batch_dict['points'] xyz = raw_points[:, 1:4] xyz_batch_cnt = xyz.new_zeros(batch_size).int() for bs_idx in range(batch_size): xyz_batch_cnt[bs_idx] = (raw_points[:, 0] == bs_idx).sum() point_features = raw_points[:, 4:].contiguous() if raw_points.shape[1] > 4 else None pooled_points, pooled_features = self.SA_rawpoints( xyz=xyz.contiguous(), xyz_batch_cnt=xyz_batch_cnt, new_xyz=new_xyz, new_xyz_batch_cnt=new_xyz_batch_cnt, features=point_features, ) point_features_list.append(pooled_features.view(batch_size, num_keypoints, -1)) for k, src_name in enumerate(self.SA_layer_names): cur_coords = batch_dict['multi_scale_3d_features'][src_name].indices xyz = common_utils.get_voxel_centers( cur_coords[:, 1:4], downsample_times=self.downsample_times_map[src_name], voxel_size=self.voxel_size, point_cloud_range=self.point_cloud_range ) xyz_batch_cnt = xyz.new_zeros(batch_size).int() for bs_idx in range(batch_size): xyz_batch_cnt[bs_idx] = (cur_coords[:, 0] == bs_idx).sum() pooled_points, pooled_features = self.SA_layers[k]( xyz=xyz.contiguous(), xyz_batch_cnt=xyz_batch_cnt, new_xyz=new_xyz, new_xyz_batch_cnt=new_xyz_batch_cnt, features=batch_dict['multi_scale_3d_features'][src_name].features.contiguous(), ) point_features_list.append(pooled_features.view(batch_size, num_keypoints, -1)) point_features = torch.cat(point_features_list, dim=2) batch_idx = torch.arange(batch_size, device=keypoints.device).view(-1, 1).repeat(1, keypoints.shape[1]).view(-1) point_coords = torch.cat((batch_idx.view(-1, 1).float(), keypoints.view(-1, 3)), dim=1) batch_dict['point_features_before_fusion'] = point_features.view(-1, point_features.shape[-1]) point_features = self.vsa_point_feature_fusion(point_features.view(-1, point_features.shape[-1])) batch_dict['point_features'] = point_features # (BxN, C) batch_dict['point_coords'] = point_coords # (BxN, 4) return batch_dict
9,638
39.1625
121
py
SASA
SASA-main/pcdet/models/backbones_3d/vfe/vfe_template.py
import torch.nn as nn class VFETemplate(nn.Module): def __init__(self, model_cfg, **kwargs): super().__init__() self.model_cfg = model_cfg def get_output_feature_dim(self): raise NotImplementedError def forward(self, **kwargs): """ Args: **kwargs: Returns: batch_dict: ... vfe_features: (num_voxels, C) """ raise NotImplementedError
470
19.478261
45
py
SASA
SASA-main/pcdet/models/backbones_3d/vfe/mean_vfe.py
import torch from .vfe_template import VFETemplate class MeanVFE(VFETemplate): def __init__(self, model_cfg, num_point_features, **kwargs): super().__init__(model_cfg=model_cfg) self.num_point_features = num_point_features def get_output_feature_dim(self): return self.num_point_features def forward(self, batch_dict, **kwargs): """ Args: batch_dict: voxels: (num_voxels, max_points_per_voxel, C) voxel_num_points: optional (num_voxels) **kwargs: Returns: vfe_features: (num_voxels, C) """ voxel_features, voxel_num_points = batch_dict['voxels'], batch_dict['voxel_num_points'] points_mean = voxel_features[:, :, :].sum(dim=1, keepdim=False) normalizer = torch.clamp_min(voxel_num_points.view(-1, 1), min=1.0).type_as(voxel_features) points_mean = points_mean / normalizer batch_dict['voxel_features'] = points_mean.contiguous() return batch_dict
1,038
31.46875
99
py
SASA
SASA-main/pcdet/models/backbones_3d/vfe/pillar_vfe.py
import torch import torch.nn as nn import torch.nn.functional as F from .vfe_template import VFETemplate class PFNLayer(nn.Module): def __init__(self, in_channels, out_channels, use_norm=True, last_layer=False): super().__init__() self.last_vfe = last_layer self.use_norm = use_norm if not self.last_vfe: out_channels = out_channels // 2 if self.use_norm: self.linear = nn.Linear(in_channels, out_channels, bias=False) self.norm = nn.BatchNorm1d(out_channels, eps=1e-3, momentum=0.01) else: self.linear = nn.Linear(in_channels, out_channels, bias=True) self.part = 50000 def forward(self, inputs): if inputs.shape[0] > self.part: # nn.Linear performs randomly when batch size is too large num_parts = inputs.shape[0] // self.part part_linear_out = [self.linear(inputs[num_part*self.part:(num_part+1)*self.part]) for num_part in range(num_parts+1)] x = torch.cat(part_linear_out, dim=0) else: x = self.linear(inputs) torch.backends.cudnn.enabled = False x = self.norm(x.permute(0, 2, 1)).permute(0, 2, 1) if self.use_norm else x torch.backends.cudnn.enabled = True x = F.relu(x) x_max = torch.max(x, dim=1, keepdim=True)[0] if self.last_vfe: return x_max else: x_repeat = x_max.repeat(1, inputs.shape[1], 1) x_concatenated = torch.cat([x, x_repeat], dim=2) return x_concatenated class PillarVFE(VFETemplate): def __init__(self, model_cfg, num_point_features, voxel_size, point_cloud_range): super().__init__(model_cfg=model_cfg) self.use_norm = self.model_cfg.USE_NORM self.with_distance = self.model_cfg.WITH_DISTANCE self.use_absolute_xyz = self.model_cfg.USE_ABSLOTE_XYZ num_point_features += 6 if self.use_absolute_xyz else 3 if self.with_distance: num_point_features += 1 self.num_filters = self.model_cfg.NUM_FILTERS assert len(self.num_filters) > 0 num_filters = [num_point_features] + list(self.num_filters) pfn_layers = [] for i in range(len(num_filters) - 1): in_filters = num_filters[i] out_filters = num_filters[i + 1] pfn_layers.append( PFNLayer(in_filters, out_filters, self.use_norm, last_layer=(i >= len(num_filters) - 2)) ) self.pfn_layers = nn.ModuleList(pfn_layers) self.voxel_x = voxel_size[0] self.voxel_y = voxel_size[1] self.voxel_z = voxel_size[2] self.x_offset = self.voxel_x / 2 + point_cloud_range[0] self.y_offset = self.voxel_y / 2 + point_cloud_range[1] self.z_offset = self.voxel_z / 2 + point_cloud_range[2] def get_output_feature_dim(self): return self.num_filters[-1] def get_paddings_indicator(self, actual_num, max_num, axis=0): actual_num = torch.unsqueeze(actual_num, axis + 1) max_num_shape = [1] * len(actual_num.shape) max_num_shape[axis + 1] = -1 max_num = torch.arange(max_num, dtype=torch.int, device=actual_num.device).view(max_num_shape) paddings_indicator = actual_num.int() > max_num return paddings_indicator def forward(self, batch_dict, **kwargs): voxel_features, voxel_num_points, coords = batch_dict['voxels'], batch_dict['voxel_num_points'], batch_dict['voxel_coords'] points_mean = voxel_features[:, :, :3].sum(dim=1, keepdim=True) / voxel_num_points.type_as(voxel_features).view(-1, 1, 1) f_cluster = voxel_features[:, :, :3] - points_mean f_center = torch.zeros_like(voxel_features[:, :, :3]) f_center[:, :, 0] = voxel_features[:, :, 0] - (coords[:, 3].to(voxel_features.dtype).unsqueeze(1) * self.voxel_x + self.x_offset) f_center[:, :, 1] = voxel_features[:, :, 1] - (coords[:, 2].to(voxel_features.dtype).unsqueeze(1) * self.voxel_y + self.y_offset) f_center[:, :, 2] = voxel_features[:, :, 2] - (coords[:, 1].to(voxel_features.dtype).unsqueeze(1) * self.voxel_z + self.z_offset) if self.use_absolute_xyz: features = [voxel_features, f_cluster, f_center] else: features = [voxel_features[..., 3:], f_cluster, f_center] if self.with_distance: points_dist = torch.norm(voxel_features[:, :, :3], 2, 2, keepdim=True) features.append(points_dist) features = torch.cat(features, dim=-1) voxel_count = features.shape[1] mask = self.get_paddings_indicator(voxel_num_points, voxel_count, axis=0) mask = torch.unsqueeze(mask, -1).type_as(voxel_features) features *= mask for pfn in self.pfn_layers: features = pfn(features) features = features.squeeze() batch_dict['pillar_features'] = features return batch_dict
5,089
40.048387
137
py
SASA
SASA-main/pcdet/models/dense_heads/anchor_head_single.py
import numpy as np import torch.nn as nn from .anchor_head_template import AnchorHeadTemplate class AnchorHeadSingle(AnchorHeadTemplate): def __init__(self, model_cfg, input_channels, num_class, class_names, grid_size, point_cloud_range, predict_boxes_when_training=True, **kwargs): super().__init__( model_cfg=model_cfg, num_class=num_class, class_names=class_names, grid_size=grid_size, point_cloud_range=point_cloud_range, predict_boxes_when_training=predict_boxes_when_training ) self.num_anchors_per_location = sum(self.num_anchors_per_location) self.conv_cls = nn.Conv2d( input_channels, self.num_anchors_per_location * self.num_class, kernel_size=1 ) self.conv_box = nn.Conv2d( input_channels, self.num_anchors_per_location * self.box_coder.code_size, kernel_size=1 ) if self.model_cfg.get('USE_DIRECTION_CLASSIFIER', None) is not None: self.conv_dir_cls = nn.Conv2d( input_channels, self.num_anchors_per_location * self.model_cfg.NUM_DIR_BINS, kernel_size=1 ) else: self.conv_dir_cls = None self.init_weights() def init_weights(self): pi = 0.01 nn.init.constant_(self.conv_cls.bias, -np.log((1 - pi) / pi)) nn.init.normal_(self.conv_box.weight, mean=0, std=0.001) def forward(self, data_dict): spatial_features_2d = data_dict['spatial_features_2d'] cls_preds = self.conv_cls(spatial_features_2d) box_preds = self.conv_box(spatial_features_2d) cls_preds = cls_preds.permute(0, 2, 3, 1).contiguous() # [N, H, W, C] box_preds = box_preds.permute(0, 2, 3, 1).contiguous() # [N, H, W, C] self.forward_ret_dict['cls_preds'] = cls_preds self.forward_ret_dict['box_preds'] = box_preds if self.conv_dir_cls is not None: dir_cls_preds = self.conv_dir_cls(spatial_features_2d) dir_cls_preds = dir_cls_preds.permute(0, 2, 3, 1).contiguous() self.forward_ret_dict['dir_cls_preds'] = dir_cls_preds else: dir_cls_preds = None if self.training: targets_dict = self.assign_targets( gt_boxes=data_dict['gt_boxes'] ) self.forward_ret_dict.update(targets_dict) if not self.training or self.predict_boxes_when_training: batch_cls_preds, batch_box_preds = self.generate_predicted_boxes( batch_size=data_dict['batch_size'], cls_preds=cls_preds, box_preds=box_preds, dir_cls_preds=dir_cls_preds ) data_dict['batch_cls_preds'] = batch_cls_preds data_dict['batch_box_preds'] = batch_box_preds data_dict['cls_preds_normalized'] = False return data_dict
2,928
37.539474
136
py
SASA
SASA-main/pcdet/models/dense_heads/point_head_template.py
import torch import torch.nn as nn import torch.nn.functional as F from ...ops.roiaware_pool3d import roiaware_pool3d_utils from ...utils import common_utils, loss_utils class PointHeadTemplate(nn.Module): def __init__(self, model_cfg, num_class): super().__init__() self.model_cfg = model_cfg self.num_class = num_class self.build_losses(self.model_cfg.LOSS_CONFIG) self.forward_ret_dict = None def build_losses(self, losses_cfg): self.add_module( 'cls_loss_func', loss_utils.SigmoidFocalClassificationLoss(alpha=0.25, gamma=2.0) ) self.add_module( 'reg_loss_func', loss_utils.WeightedSmoothL1Loss( code_weights=losses_cfg.LOSS_WEIGHTS.get('code_weights', None) ) ) @staticmethod def make_fc_layers(fc_cfg, input_channels, output_channels): fc_layers = [] c_in = input_channels for k in range(0, fc_cfg.__len__()): fc_layers.extend([ nn.Linear(c_in, fc_cfg[k], bias=False), nn.BatchNorm1d(fc_cfg[k]), nn.ReLU(), ]) c_in = fc_cfg[k] fc_layers.append(nn.Linear(c_in, output_channels, bias=True)) return nn.Sequential(*fc_layers) def assign_stack_targets(self, points, gt_boxes, extend_gt_boxes=None, ret_box_labels=False, ret_part_labels=False, set_ignore_flag=True, use_ball_constraint=False, central_radius=2.0): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes: (B, M, 8) extend_gt_boxes: [B, M, 8] ret_box_labels: ret_part_labels: set_ignore_flag: use_ball_constraint: central_radius: Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_box_labels: (N1 + N2 + N3 + ..., code_size) """ assert len(points.shape) == 2 and points.shape[1] == 4, 'points.shape=%s' % str(points.shape) assert len(gt_boxes.shape) == 3 and gt_boxes.shape[2] == 8, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert extend_gt_boxes is None or len(extend_gt_boxes.shape) == 3 and extend_gt_boxes.shape[2] == 8, \ 'extend_gt_boxes.shape=%s' % str(extend_gt_boxes.shape) assert set_ignore_flag != use_ball_constraint, 'Choose one only!' batch_size = gt_boxes.shape[0] bs_idx = points[:, 0] point_cls_labels = points.new_zeros(points.shape[0]).long() point_box_labels = gt_boxes.new_zeros((points.shape[0], 8)) if ret_box_labels else None point_part_labels = gt_boxes.new_zeros((points.shape[0], 3)) if ret_part_labels else None for k in range(batch_size): bs_mask = (bs_idx == k) points_single = points[bs_mask][:, 1:4] point_cls_labels_single = point_cls_labels.new_zeros(bs_mask.sum()) box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) if set_ignore_flag: extend_box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), extend_gt_boxes[k:k+1, :, 0:7].contiguous() ).long().squeeze(dim=0) fg_flag = box_fg_flag ignore_flag = fg_flag ^ (extend_box_idxs_of_pts >= 0) point_cls_labels_single[ignore_flag] = -1 elif use_ball_constraint: box_centers = gt_boxes[k][box_idxs_of_pts][:, 0:3].clone() box_centers[:, 2] += gt_boxes[k][box_idxs_of_pts][:, 5] / 2 ball_flag = ((box_centers - points_single).norm(dim=1) < central_radius) fg_flag = box_fg_flag & ball_flag else: raise NotImplementedError gt_box_of_fg_points = gt_boxes[k][box_idxs_of_pts[fg_flag]] point_cls_labels_single[fg_flag] = 1 if self.num_class == 1 else gt_box_of_fg_points[:, -1].long() point_cls_labels[bs_mask] = point_cls_labels_single if ret_box_labels and gt_box_of_fg_points.shape[0] > 0: point_box_labels_single = point_box_labels.new_zeros((bs_mask.sum(), 8)) fg_point_box_labels = self.box_coder.encode_torch( gt_boxes=gt_box_of_fg_points[:, :-1], points=points_single[fg_flag], gt_classes=gt_box_of_fg_points[:, -1].long() ) point_box_labels_single[fg_flag] = fg_point_box_labels point_box_labels[bs_mask] = point_box_labels_single if ret_part_labels: point_part_labels_single = point_part_labels.new_zeros((bs_mask.sum(), 3)) transformed_points = points_single[fg_flag] - gt_box_of_fg_points[:, 0:3] transformed_points = common_utils.rotate_points_along_z( transformed_points.view(-1, 1, 3), -gt_box_of_fg_points[:, 6] ).view(-1, 3) offset = torch.tensor([0.5, 0.5, 0.5]).view(1, 3).type_as(transformed_points) point_part_labels_single[fg_flag] = (transformed_points / gt_box_of_fg_points[:, 3:6]) + offset point_part_labels[bs_mask] = point_part_labels_single targets_dict = { 'point_cls_labels': point_cls_labels, 'point_box_labels': point_box_labels, 'point_part_labels': point_part_labels } return targets_dict def get_cls_layer_loss(self, tb_dict=None): point_cls_labels = self.forward_ret_dict['point_cls_labels'].view(-1) point_cls_preds = self.forward_ret_dict['point_cls_preds'].view(-1, self.num_class) positives = (point_cls_labels > 0) negative_cls_weights = (point_cls_labels == 0) * 1.0 cls_weights = (negative_cls_weights + 15.0 * positives).float() pos_normalizer = positives.sum(dim=0).float() cls_weights /= torch.clamp(pos_normalizer, min=1.0) one_hot_targets = point_cls_preds.new_zeros(*list(point_cls_labels.shape), self.num_class + 1) one_hot_targets.scatter_(-1, (point_cls_labels * (point_cls_labels >= 0).long()).unsqueeze(dim=-1).long(), 1.0) one_hot_targets = one_hot_targets[..., 1:] cls_loss_src = self.cls_loss_func(point_cls_preds, one_hot_targets, weights=cls_weights) point_loss_cls = cls_loss_src.sum() loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS point_loss_cls = point_loss_cls * loss_weights_dict['point_cls_weight'] if tb_dict is None: tb_dict = {} tb_dict.update({ 'point_loss_cls': point_loss_cls.item(), 'point_pos_num': pos_normalizer.item() }) return point_loss_cls, tb_dict def get_part_layer_loss(self, tb_dict=None): pos_mask = self.forward_ret_dict['point_cls_labels'] > 0 pos_normalizer = max(1, (pos_mask > 0).sum().item()) point_part_labels = self.forward_ret_dict['point_part_labels'] point_part_preds = self.forward_ret_dict['point_part_preds'] point_loss_part = F.binary_cross_entropy(torch.sigmoid(point_part_preds), point_part_labels, reduction='none') point_loss_part = (point_loss_part.sum(dim=-1) * pos_mask.float()).sum() / (3 * pos_normalizer) loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS point_loss_part = point_loss_part * loss_weights_dict['point_part_weight'] if tb_dict is None: tb_dict = {} tb_dict.update({'point_loss_part': point_loss_part.item()}) return point_loss_part, tb_dict def get_box_layer_loss(self, tb_dict=None): pos_mask = self.forward_ret_dict['point_cls_labels'] > 0 point_box_labels = self.forward_ret_dict['point_box_labels'] point_box_preds = self.forward_ret_dict['point_box_preds'] reg_weights = pos_mask.float() pos_normalizer = pos_mask.sum().float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) point_loss_box_src = self.reg_loss_func( point_box_preds[None, ...], point_box_labels[None, ...], weights=reg_weights[None, ...] ) point_loss_box = point_loss_box_src.sum() loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS point_loss_box = point_loss_box * loss_weights_dict['point_box_weight'] if tb_dict is None: tb_dict = {} tb_dict.update({'point_loss_box': point_loss_box.item()}) return point_loss_box, tb_dict def generate_predicted_boxes(self, points, point_cls_preds, point_box_preds): """ Args: points: (N, 3) point_cls_preds: (N, num_class) point_box_preds: (N, box_code_size) Returns: point_cls_preds: (N, num_class) point_box_preds: (N, box_code_size) """ _, pred_classes = point_cls_preds.max(dim=-1) point_box_preds = self.box_coder.decode_torch(point_box_preds, points, pred_classes + 1) return point_cls_preds, point_box_preds def forward(self, **kwargs): raise NotImplementedError
9,475
45
119
py
SASA
SASA-main/pcdet/models/dense_heads/anchor_head_template.py
import numpy as np import torch import torch.nn as nn from ...utils import box_coder_utils, common_utils, loss_utils from .target_assigner.anchor_generator import AnchorGenerator from .target_assigner.atss_target_assigner import ATSSTargetAssigner from .target_assigner.axis_aligned_target_assigner import AxisAlignedTargetAssigner class AnchorHeadTemplate(nn.Module): def __init__(self, model_cfg, num_class, class_names, grid_size, point_cloud_range, predict_boxes_when_training): super().__init__() self.model_cfg = model_cfg self.num_class = num_class self.class_names = class_names self.predict_boxes_when_training = predict_boxes_when_training self.use_multihead = self.model_cfg.get('USE_MULTIHEAD', False) anchor_target_cfg = self.model_cfg.TARGET_ASSIGNER_CONFIG self.box_coder = getattr(box_coder_utils, anchor_target_cfg.BOX_CODER)( num_dir_bins=anchor_target_cfg.get('NUM_DIR_BINS', 6), **anchor_target_cfg.get('BOX_CODER_CONFIG', {}) ) anchor_generator_cfg = self.model_cfg.ANCHOR_GENERATOR_CONFIG anchors, self.num_anchors_per_location = self.generate_anchors( anchor_generator_cfg, grid_size=grid_size, point_cloud_range=point_cloud_range, anchor_ndim=self.box_coder.code_size ) self.anchors = [x.cuda() for x in anchors] self.target_assigner = self.get_target_assigner(anchor_target_cfg) self.forward_ret_dict = {} self.build_losses(self.model_cfg.LOSS_CONFIG) @staticmethod def generate_anchors(anchor_generator_cfg, grid_size, point_cloud_range, anchor_ndim=7): anchor_generator = AnchorGenerator( anchor_range=point_cloud_range, anchor_generator_config=anchor_generator_cfg ) feature_map_size = [grid_size[:2] // config['feature_map_stride'] for config in anchor_generator_cfg] anchors_list, num_anchors_per_location_list = anchor_generator.generate_anchors(feature_map_size) if anchor_ndim != 7: for idx, anchors in enumerate(anchors_list): pad_zeros = anchors.new_zeros([*anchors.shape[0:-1], anchor_ndim - 7]) new_anchors = torch.cat((anchors, pad_zeros), dim=-1) anchors_list[idx] = new_anchors return anchors_list, num_anchors_per_location_list def get_target_assigner(self, anchor_target_cfg): if anchor_target_cfg.NAME == 'ATSS': target_assigner = ATSSTargetAssigner( topk=anchor_target_cfg.TOPK, box_coder=self.box_coder, use_multihead=self.use_multihead, match_height=anchor_target_cfg.MATCH_HEIGHT ) elif anchor_target_cfg.NAME == 'AxisAlignedTargetAssigner': target_assigner = AxisAlignedTargetAssigner( model_cfg=self.model_cfg, class_names=self.class_names, box_coder=self.box_coder, match_height=anchor_target_cfg.MATCH_HEIGHT ) else: raise NotImplementedError return target_assigner def build_losses(self, losses_cfg): self.add_module( 'cls_loss_func', loss_utils.SigmoidFocalClassificationLoss(alpha=0.25, gamma=2.0) ) reg_loss_name = 'WeightedSmoothL1Loss' if losses_cfg.get('REG_LOSS_TYPE', None) is None \ else losses_cfg.REG_LOSS_TYPE self.add_module( 'reg_loss_func', getattr(loss_utils, reg_loss_name)(code_weights=losses_cfg.LOSS_WEIGHTS['code_weights']) ) self.add_module( 'dir_loss_func', loss_utils.WeightedCrossEntropyLoss() ) def assign_targets(self, gt_boxes): """ Args: gt_boxes: (B, M, 8) Returns: """ targets_dict = self.target_assigner.assign_targets( self.anchors, gt_boxes ) return targets_dict def get_cls_layer_loss(self): cls_preds = self.forward_ret_dict['cls_preds'] box_cls_labels = self.forward_ret_dict['box_cls_labels'] batch_size = int(cls_preds.shape[0]) cared = box_cls_labels >= 0 # [N, num_anchors] positives = box_cls_labels > 0 negatives = box_cls_labels == 0 negative_cls_weights = negatives * 1.0 cls_weights = (negative_cls_weights + 1.0 * positives).float() reg_weights = positives.float() if self.num_class == 1: # class agnostic box_cls_labels[positives] = 1 pos_normalizer = positives.sum(1, keepdim=True).float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) cls_weights /= torch.clamp(pos_normalizer, min=1.0) cls_targets = box_cls_labels * cared.type_as(box_cls_labels) cls_targets = cls_targets.unsqueeze(dim=-1) cls_targets = cls_targets.squeeze(dim=-1) one_hot_targets = torch.zeros( *list(cls_targets.shape), self.num_class + 1, dtype=cls_preds.dtype, device=cls_targets.device ) one_hot_targets.scatter_(-1, cls_targets.unsqueeze(dim=-1).long(), 1.0) cls_preds = cls_preds.view(batch_size, -1, self.num_class) one_hot_targets = one_hot_targets[..., 1:] cls_loss_src = self.cls_loss_func(cls_preds, one_hot_targets, weights=cls_weights) # [N, M] cls_loss = cls_loss_src.sum() / batch_size cls_loss = cls_loss * self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['cls_weight'] tb_dict = { 'rpn_loss_cls': cls_loss.item() } return cls_loss, tb_dict @staticmethod def add_sin_difference(boxes1, boxes2, dim=6): assert dim != -1 rad_pred_encoding = torch.sin(boxes1[..., dim:dim + 1]) * torch.cos(boxes2[..., dim:dim + 1]) rad_tg_encoding = torch.cos(boxes1[..., dim:dim + 1]) * torch.sin(boxes2[..., dim:dim + 1]) boxes1 = torch.cat([boxes1[..., :dim], rad_pred_encoding, boxes1[..., dim + 1:]], dim=-1) boxes2 = torch.cat([boxes2[..., :dim], rad_tg_encoding, boxes2[..., dim + 1:]], dim=-1) return boxes1, boxes2 @staticmethod def get_direction_target(anchors, reg_targets, one_hot=True, dir_offset=0, num_bins=2): batch_size = reg_targets.shape[0] anchors = anchors.view(batch_size, -1, anchors.shape[-1]) rot_gt = reg_targets[..., 6] + anchors[..., 6] offset_rot = common_utils.limit_period(rot_gt - dir_offset, 0, 2 * np.pi) dir_cls_targets = torch.floor(offset_rot / (2 * np.pi / num_bins)).long() dir_cls_targets = torch.clamp(dir_cls_targets, min=0, max=num_bins - 1) if one_hot: dir_targets = torch.zeros(*list(dir_cls_targets.shape), num_bins, dtype=anchors.dtype, device=dir_cls_targets.device) dir_targets.scatter_(-1, dir_cls_targets.unsqueeze(dim=-1).long(), 1.0) dir_cls_targets = dir_targets return dir_cls_targets def get_box_reg_layer_loss(self): box_preds = self.forward_ret_dict['box_preds'] box_dir_cls_preds = self.forward_ret_dict.get('dir_cls_preds', None) box_reg_targets = self.forward_ret_dict['box_reg_targets'] box_cls_labels = self.forward_ret_dict['box_cls_labels'] batch_size = int(box_preds.shape[0]) positives = box_cls_labels > 0 reg_weights = positives.float() pos_normalizer = positives.sum(1, keepdim=True).float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) if isinstance(self.anchors, list): if self.use_multihead: anchors = torch.cat( [anchor.permute(3, 4, 0, 1, 2, 5).contiguous().view(-1, anchor.shape[-1]) for anchor in self.anchors], dim=0) else: anchors = torch.cat(self.anchors, dim=-3) else: anchors = self.anchors anchors = anchors.view(1, -1, anchors.shape[-1]).repeat(batch_size, 1, 1) box_preds = box_preds.view(batch_size, -1, box_preds.shape[-1] // self.num_anchors_per_location if not self.use_multihead else box_preds.shape[-1]) # sin(a - b) = sinacosb-cosasinb box_preds_sin, reg_targets_sin = self.add_sin_difference(box_preds, box_reg_targets) loc_loss_src = self.reg_loss_func(box_preds_sin, reg_targets_sin, weights=reg_weights) # [N, M] loc_loss = loc_loss_src.sum() / batch_size loc_loss = loc_loss * self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['loc_weight'] box_loss = loc_loss tb_dict = { 'rpn_loss_loc': loc_loss.item() } if box_dir_cls_preds is not None: dir_targets = self.get_direction_target( anchors, box_reg_targets, dir_offset=self.model_cfg.DIR_OFFSET, num_bins=self.model_cfg.NUM_DIR_BINS ) dir_logits = box_dir_cls_preds.view(batch_size, -1, self.model_cfg.NUM_DIR_BINS) weights = positives.type_as(dir_logits) weights /= torch.clamp(weights.sum(-1, keepdim=True), min=1.0) dir_loss = self.dir_loss_func(dir_logits, dir_targets, weights=weights) dir_loss = dir_loss.sum() / batch_size dir_loss = dir_loss * self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['dir_weight'] box_loss += dir_loss tb_dict['rpn_loss_dir'] = dir_loss.item() return box_loss, tb_dict def get_loss(self): cls_loss, tb_dict = self.get_cls_layer_loss() box_loss, tb_dict_box = self.get_box_reg_layer_loss() tb_dict.update(tb_dict_box) rpn_loss = cls_loss + box_loss tb_dict['rpn_loss'] = rpn_loss.item() return rpn_loss, tb_dict def generate_predicted_boxes(self, batch_size, cls_preds, box_preds, dir_cls_preds=None): """ Args: batch_size: cls_preds: (N, H, W, C1) box_preds: (N, H, W, C2) dir_cls_preds: (N, H, W, C3) Returns: batch_cls_preds: (B, num_boxes, num_classes) batch_box_preds: (B, num_boxes, 7+C) """ if isinstance(self.anchors, list): if self.use_multihead: anchors = torch.cat([anchor.permute(3, 4, 0, 1, 2, 5).contiguous().view(-1, anchor.shape[-1]) for anchor in self.anchors], dim=0) else: anchors = torch.cat(self.anchors, dim=-3) else: anchors = self.anchors num_anchors = anchors.view(-1, anchors.shape[-1]).shape[0] batch_anchors = anchors.view(1, -1, anchors.shape[-1]).repeat(batch_size, 1, 1) batch_cls_preds = cls_preds.view(batch_size, num_anchors, -1).float() \ if not isinstance(cls_preds, list) else cls_preds batch_box_preds = box_preds.view(batch_size, num_anchors, -1) if not isinstance(box_preds, list) \ else torch.cat(box_preds, dim=1).view(batch_size, num_anchors, -1) batch_box_preds = self.box_coder.decode_torch(batch_box_preds, batch_anchors) if dir_cls_preds is not None: dir_offset = self.model_cfg.DIR_OFFSET dir_limit_offset = self.model_cfg.DIR_LIMIT_OFFSET dir_cls_preds = dir_cls_preds.view(batch_size, num_anchors, -1) if not isinstance(dir_cls_preds, list) \ else torch.cat(dir_cls_preds, dim=1).view(batch_size, num_anchors, -1) dir_labels = torch.max(dir_cls_preds, dim=-1)[1] period = (2 * np.pi / self.model_cfg.NUM_DIR_BINS) dir_rot = common_utils.limit_period( batch_box_preds[..., 6] - dir_offset, dir_limit_offset, period ) batch_box_preds[..., 6] = dir_rot + dir_offset + period * dir_labels.to(batch_box_preds.dtype) if isinstance(self.box_coder, box_coder_utils.PreviousResidualDecoder): batch_box_preds[..., 6] = common_utils.limit_period( -(batch_box_preds[..., 6] + np.pi / 2), offset=0.5, period=np.pi * 2 ) return batch_cls_preds, batch_box_preds def forward(self, **kwargs): raise NotImplementedError
12,364
43.800725
118
py
SASA
SASA-main/pcdet/models/dense_heads/anchor_head_multi.py
import numpy as np import torch import torch.nn as nn from ..backbones_2d import BaseBEVBackbone from .anchor_head_template import AnchorHeadTemplate class SingleHead(BaseBEVBackbone): def __init__(self, model_cfg, input_channels, num_class, num_anchors_per_location, code_size, rpn_head_cfg=None, head_label_indices=None, separate_reg_config=None): super().__init__(rpn_head_cfg, input_channels) self.num_anchors_per_location = num_anchors_per_location self.num_class = num_class self.code_size = code_size self.model_cfg = model_cfg self.separate_reg_config = separate_reg_config self.register_buffer('head_label_indices', head_label_indices) if self.separate_reg_config is not None: code_size_cnt = 0 self.conv_box = nn.ModuleDict() self.conv_box_names = [] num_middle_conv = self.separate_reg_config.NUM_MIDDLE_CONV num_middle_filter = self.separate_reg_config.NUM_MIDDLE_FILTER conv_cls_list = [] c_in = input_channels for k in range(num_middle_conv): conv_cls_list.extend([ nn.Conv2d( c_in, num_middle_filter, kernel_size=3, stride=1, padding=1, bias=False ), nn.BatchNorm2d(num_middle_filter), nn.ReLU() ]) c_in = num_middle_filter conv_cls_list.append(nn.Conv2d( c_in, self.num_anchors_per_location * self.num_class, kernel_size=3, stride=1, padding=1 )) self.conv_cls = nn.Sequential(*conv_cls_list) for reg_config in self.separate_reg_config.REG_LIST: reg_name, reg_channel = reg_config.split(':') reg_channel = int(reg_channel) cur_conv_list = [] c_in = input_channels for k in range(num_middle_conv): cur_conv_list.extend([ nn.Conv2d( c_in, num_middle_filter, kernel_size=3, stride=1, padding=1, bias=False ), nn.BatchNorm2d(num_middle_filter), nn.ReLU() ]) c_in = num_middle_filter cur_conv_list.append(nn.Conv2d( c_in, self.num_anchors_per_location * int(reg_channel), kernel_size=3, stride=1, padding=1, bias=True )) code_size_cnt += reg_channel self.conv_box[f'conv_{reg_name}'] = nn.Sequential(*cur_conv_list) self.conv_box_names.append(f'conv_{reg_name}') for m in self.conv_box.modules(): if isinstance(m, nn.Conv2d): nn.init.kaiming_normal_(m.weight, mode='fan_out', nonlinearity='relu') if m.bias is not None: nn.init.constant_(m.bias, 0) assert code_size_cnt == code_size, f'Code size does not match: {code_size_cnt}:{code_size}' else: self.conv_cls = nn.Conv2d( input_channels, self.num_anchors_per_location * self.num_class, kernel_size=1 ) self.conv_box = nn.Conv2d( input_channels, self.num_anchors_per_location * self.code_size, kernel_size=1 ) if self.model_cfg.get('USE_DIRECTION_CLASSIFIER', None) is not None: self.conv_dir_cls = nn.Conv2d( input_channels, self.num_anchors_per_location * self.model_cfg.NUM_DIR_BINS, kernel_size=1 ) else: self.conv_dir_cls = None self.use_multihead = self.model_cfg.get('USE_MULTIHEAD', False) self.init_weights() def init_weights(self): pi = 0.01 if isinstance(self.conv_cls, nn.Conv2d): nn.init.constant_(self.conv_cls.bias, -np.log((1 - pi) / pi)) else: nn.init.constant_(self.conv_cls[-1].bias, -np.log((1 - pi) / pi)) def forward(self, spatial_features_2d): ret_dict = {} spatial_features_2d = super().forward({'spatial_features': spatial_features_2d})['spatial_features_2d'] cls_preds = self.conv_cls(spatial_features_2d) if self.separate_reg_config is None: box_preds = self.conv_box(spatial_features_2d) else: box_preds_list = [] for reg_name in self.conv_box_names: box_preds_list.append(self.conv_box[reg_name](spatial_features_2d)) box_preds = torch.cat(box_preds_list, dim=1) if not self.use_multihead: box_preds = box_preds.permute(0, 2, 3, 1).contiguous() cls_preds = cls_preds.permute(0, 2, 3, 1).contiguous() else: H, W = box_preds.shape[2:] batch_size = box_preds.shape[0] box_preds = box_preds.view(-1, self.num_anchors_per_location, self.code_size, H, W).permute(0, 1, 3, 4, 2).contiguous() cls_preds = cls_preds.view(-1, self.num_anchors_per_location, self.num_class, H, W).permute(0, 1, 3, 4, 2).contiguous() box_preds = box_preds.view(batch_size, -1, self.code_size) cls_preds = cls_preds.view(batch_size, -1, self.num_class) if self.conv_dir_cls is not None: dir_cls_preds = self.conv_dir_cls(spatial_features_2d) if self.use_multihead: dir_cls_preds = dir_cls_preds.view( -1, self.num_anchors_per_location, self.model_cfg.NUM_DIR_BINS, H, W).permute(0, 1, 3, 4, 2).contiguous() dir_cls_preds = dir_cls_preds.view(batch_size, -1, self.model_cfg.NUM_DIR_BINS) else: dir_cls_preds = dir_cls_preds.permute(0, 2, 3, 1).contiguous() else: dir_cls_preds = None ret_dict['cls_preds'] = cls_preds ret_dict['box_preds'] = box_preds ret_dict['dir_cls_preds'] = dir_cls_preds return ret_dict class AnchorHeadMulti(AnchorHeadTemplate): def __init__(self, model_cfg, input_channels, num_class, class_names, grid_size, point_cloud_range, predict_boxes_when_training=True, **kwargs): super().__init__( model_cfg=model_cfg, num_class=num_class, class_names=class_names, grid_size=grid_size, point_cloud_range=point_cloud_range, predict_boxes_when_training=predict_boxes_when_training ) self.model_cfg = model_cfg self.separate_multihead = self.model_cfg.get('SEPARATE_MULTIHEAD', False) if self.model_cfg.get('SHARED_CONV_NUM_FILTER', None) is not None: shared_conv_num_filter = self.model_cfg.SHARED_CONV_NUM_FILTER self.shared_conv = nn.Sequential( nn.Conv2d(input_channels, shared_conv_num_filter, 3, stride=1, padding=1, bias=False), nn.BatchNorm2d(shared_conv_num_filter, eps=1e-3, momentum=0.01), nn.ReLU(), ) else: self.shared_conv = None shared_conv_num_filter = input_channels self.rpn_heads = None self.make_multihead(shared_conv_num_filter) def make_multihead(self, input_channels): rpn_head_cfgs = self.model_cfg.RPN_HEAD_CFGS rpn_heads = [] class_names = [] for rpn_head_cfg in rpn_head_cfgs: class_names.extend(rpn_head_cfg['HEAD_CLS_NAME']) for rpn_head_cfg in rpn_head_cfgs: num_anchors_per_location = sum([self.num_anchors_per_location[class_names.index(head_cls)] for head_cls in rpn_head_cfg['HEAD_CLS_NAME']]) head_label_indices = torch.from_numpy(np.array([ self.class_names.index(cur_name) + 1 for cur_name in rpn_head_cfg['HEAD_CLS_NAME'] ])) rpn_head = SingleHead( self.model_cfg, input_channels, len(rpn_head_cfg['HEAD_CLS_NAME']) if self.separate_multihead else self.num_class, num_anchors_per_location, self.box_coder.code_size, rpn_head_cfg, head_label_indices=head_label_indices, separate_reg_config=self.model_cfg.get('SEPARATE_REG_CONFIG', None) ) rpn_heads.append(rpn_head) self.rpn_heads = nn.ModuleList(rpn_heads) def forward(self, data_dict): spatial_features_2d = data_dict['spatial_features_2d'] if self.shared_conv is not None: spatial_features_2d = self.shared_conv(spatial_features_2d) ret_dicts = [] for rpn_head in self.rpn_heads: ret_dicts.append(rpn_head(spatial_features_2d)) cls_preds = [ret_dict['cls_preds'] for ret_dict in ret_dicts] box_preds = [ret_dict['box_preds'] for ret_dict in ret_dicts] ret = { 'cls_preds': cls_preds if self.separate_multihead else torch.cat(cls_preds, dim=1), 'box_preds': box_preds if self.separate_multihead else torch.cat(box_preds, dim=1), } if self.model_cfg.get('USE_DIRECTION_CLASSIFIER', False): dir_cls_preds = [ret_dict['dir_cls_preds'] for ret_dict in ret_dicts] ret['dir_cls_preds'] = dir_cls_preds if self.separate_multihead else torch.cat(dir_cls_preds, dim=1) self.forward_ret_dict.update(ret) if self.training: targets_dict = self.assign_targets( gt_boxes=data_dict['gt_boxes'] ) self.forward_ret_dict.update(targets_dict) if not self.training or self.predict_boxes_when_training: batch_cls_preds, batch_box_preds = self.generate_predicted_boxes( batch_size=data_dict['batch_size'], cls_preds=ret['cls_preds'], box_preds=ret['box_preds'], dir_cls_preds=ret.get('dir_cls_preds', None) ) if isinstance(batch_cls_preds, list): multihead_label_mapping = [] for idx in range(len(batch_cls_preds)): multihead_label_mapping.append(self.rpn_heads[idx].head_label_indices) data_dict['multihead_label_mapping'] = multihead_label_mapping data_dict['batch_cls_preds'] = batch_cls_preds data_dict['batch_box_preds'] = batch_box_preds data_dict['cls_preds_normalized'] = False return data_dict def get_cls_layer_loss(self): loss_weights = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS if 'pos_cls_weight' in loss_weights: pos_cls_weight = loss_weights['pos_cls_weight'] neg_cls_weight = loss_weights['neg_cls_weight'] else: pos_cls_weight = neg_cls_weight = 1.0 cls_preds = self.forward_ret_dict['cls_preds'] box_cls_labels = self.forward_ret_dict['box_cls_labels'] if not isinstance(cls_preds, list): cls_preds = [cls_preds] batch_size = int(cls_preds[0].shape[0]) cared = box_cls_labels >= 0 # [N, num_anchors] positives = box_cls_labels > 0 negatives = box_cls_labels == 0 negative_cls_weights = negatives * 1.0 * neg_cls_weight cls_weights = (negative_cls_weights + pos_cls_weight * positives).float() reg_weights = positives.float() if self.num_class == 1: # class agnostic box_cls_labels[positives] = 1 pos_normalizer = positives.sum(1, keepdim=True).float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) cls_weights /= torch.clamp(pos_normalizer, min=1.0) cls_targets = box_cls_labels * cared.type_as(box_cls_labels) one_hot_targets = torch.zeros( *list(cls_targets.shape), self.num_class + 1, dtype=cls_preds[0].dtype, device=cls_targets.device ) one_hot_targets.scatter_(-1, cls_targets.unsqueeze(dim=-1).long(), 1.0) one_hot_targets = one_hot_targets[..., 1:] start_idx = c_idx = 0 cls_losses = 0 for idx, cls_pred in enumerate(cls_preds): cur_num_class = self.rpn_heads[idx].num_class cls_pred = cls_pred.view(batch_size, -1, cur_num_class) if self.separate_multihead: one_hot_target = one_hot_targets[:, start_idx:start_idx + cls_pred.shape[1], c_idx:c_idx + cur_num_class] c_idx += cur_num_class else: one_hot_target = one_hot_targets[:, start_idx:start_idx + cls_pred.shape[1]] cls_weight = cls_weights[:, start_idx:start_idx + cls_pred.shape[1]] cls_loss_src = self.cls_loss_func(cls_pred, one_hot_target, weights=cls_weight) # [N, M] cls_loss = cls_loss_src.sum() / batch_size cls_loss = cls_loss * loss_weights['cls_weight'] cls_losses += cls_loss start_idx += cls_pred.shape[1] assert start_idx == one_hot_targets.shape[1] tb_dict = { 'rpn_loss_cls': cls_losses.item() } return cls_losses, tb_dict def get_box_reg_layer_loss(self): box_preds = self.forward_ret_dict['box_preds'] box_dir_cls_preds = self.forward_ret_dict.get('dir_cls_preds', None) box_reg_targets = self.forward_ret_dict['box_reg_targets'] box_cls_labels = self.forward_ret_dict['box_cls_labels'] positives = box_cls_labels > 0 reg_weights = positives.float() pos_normalizer = positives.sum(1, keepdim=True).float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) if not isinstance(box_preds, list): box_preds = [box_preds] batch_size = int(box_preds[0].shape[0]) if isinstance(self.anchors, list): if self.use_multihead: anchors = torch.cat( [anchor.permute(3, 4, 0, 1, 2, 5).contiguous().view(-1, anchor.shape[-1]) for anchor in self.anchors], dim=0 ) else: anchors = torch.cat(self.anchors, dim=-3) else: anchors = self.anchors anchors = anchors.view(1, -1, anchors.shape[-1]).repeat(batch_size, 1, 1) start_idx = 0 box_losses = 0 tb_dict = {} for idx, box_pred in enumerate(box_preds): box_pred = box_pred.view( batch_size, -1, box_pred.shape[-1] // self.num_anchors_per_location if not self.use_multihead else box_pred.shape[-1] ) box_reg_target = box_reg_targets[:, start_idx:start_idx + box_pred.shape[1]] reg_weight = reg_weights[:, start_idx:start_idx + box_pred.shape[1]] # sin(a - b) = sinacosb-cosasinb if box_dir_cls_preds is not None: box_pred_sin, reg_target_sin = self.add_sin_difference(box_pred, box_reg_target) loc_loss_src = self.reg_loss_func(box_pred_sin, reg_target_sin, weights=reg_weight) # [N, M] else: loc_loss_src = self.reg_loss_func(box_pred, box_reg_target, weights=reg_weight) # [N, M] loc_loss = loc_loss_src.sum() / batch_size loc_loss = loc_loss * self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['loc_weight'] box_losses += loc_loss tb_dict['rpn_loss_loc'] = tb_dict.get('rpn_loss_loc', 0) + loc_loss.item() if box_dir_cls_preds is not None: if not isinstance(box_dir_cls_preds, list): box_dir_cls_preds = [box_dir_cls_preds] dir_targets = self.get_direction_target( anchors, box_reg_targets, dir_offset=self.model_cfg.DIR_OFFSET, num_bins=self.model_cfg.NUM_DIR_BINS ) box_dir_cls_pred = box_dir_cls_preds[idx] dir_logit = box_dir_cls_pred.view(batch_size, -1, self.model_cfg.NUM_DIR_BINS) weights = positives.type_as(dir_logit) weights /= torch.clamp(weights.sum(-1, keepdim=True), min=1.0) weight = weights[:, start_idx:start_idx + box_pred.shape[1]] dir_target = dir_targets[:, start_idx:start_idx + box_pred.shape[1]] dir_loss = self.dir_loss_func(dir_logit, dir_target, weights=weight) dir_loss = dir_loss.sum() / batch_size dir_loss = dir_loss * self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['dir_weight'] box_losses += dir_loss tb_dict['rpn_loss_dir'] = tb_dict.get('rpn_loss_dir', 0) + dir_loss.item() start_idx += box_pred.shape[1] return box_losses, tb_dict
17,041
44.566845
117
py
SASA
SASA-main/pcdet/models/dense_heads/point_head_box.py
import torch from ...utils import box_coder_utils, box_utils from ...utils.loss_utils import PointSASALoss from .point_head_template import PointHeadTemplate class PointHeadBox(PointHeadTemplate): """ A simple point-based segmentation head, which are used for PointRCNN. Reference Paper: https://arxiv.org/abs/1812.04244 PointRCNN: 3D Object Proposal Generation and Detection from Point Cloud """ def __init__(self, num_class, input_channels, model_cfg, predict_boxes_when_training=False, **kwargs): super().__init__(model_cfg=model_cfg, num_class=num_class) self.predict_boxes_when_training = predict_boxes_when_training self.cls_layers = self.make_fc_layers( fc_cfg=self.model_cfg.CLS_FC, input_channels=input_channels, output_channels=num_class ) target_cfg = self.model_cfg.TARGET_CONFIG self.box_coder = getattr(box_coder_utils, target_cfg.BOX_CODER)( **target_cfg.BOX_CODER_CONFIG ) self.box_layers = self.make_fc_layers( fc_cfg=self.model_cfg.REG_FC, input_channels=input_channels, output_channels=self.box_coder.code_size ) # SASA Loss sasa_loss_cfg = model_cfg.LOSS_CONFIG.get('LOSS_SASA_CONFIG', None) if sasa_loss_cfg is not None: self.loss_point_sasa = PointSASALoss(**sasa_loss_cfg) else: self.loss_point_sasa = None def assign_targets(self, input_dict): """ Args: input_dict: point_features: (N1 + N2 + N3 + ..., C) batch_size: point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes (optional): (B, M, 8) Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_part_labels: (N1 + N2 + N3 + ..., 3) """ point_coords = input_dict['point_coords'] gt_boxes = input_dict['gt_boxes'] assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert point_coords.shape.__len__() in [2], 'points.shape=%s' % str(point_coords.shape) batch_size = gt_boxes.shape[0] extend_gt_boxes = box_utils.enlarge_box3d( gt_boxes.view(-1, gt_boxes.shape[-1]), extra_width=self.model_cfg.TARGET_CONFIG.GT_EXTRA_WIDTH ).view(batch_size, -1, gt_boxes.shape[-1]) targets_dict = self.assign_stack_targets( points=point_coords, gt_boxes=gt_boxes, extend_gt_boxes=extend_gt_boxes, set_ignore_flag=True, use_ball_constraint=False, ret_part_labels=False, ret_box_labels=True ) return targets_dict def get_sasa_layer_loss(self): point_loss_sasa_list = self.loss_point_sasa.loss_forward( self.forward_ret_dict['point_sasa_preds'], self.forward_ret_dict['point_sasa_labels'] ) point_loss_sasa = 0 tb_dict = dict() for i in range(len(point_loss_sasa_list)): cur_point_loss_sasa = point_loss_sasa_list[i] if cur_point_loss_sasa is None: continue point_loss_sasa = point_loss_sasa + cur_point_loss_sasa tb_dict['point_loss_sasa_layer_%d' % i] = point_loss_sasa_list[i].item() tb_dict['point_loss_sasa'] = point_loss_sasa.item() return point_loss_sasa, tb_dict def get_loss(self, tb_dict=None): tb_dict = {} if tb_dict is None else tb_dict point_loss_cls, tb_dict_1 = self.get_cls_layer_loss() point_loss_box, tb_dict_2 = self.get_box_layer_loss() point_loss = point_loss_cls + point_loss_box tb_dict.update(tb_dict_1) tb_dict.update(tb_dict_2) if self.loss_point_sasa is not None: point_loss_sasa, tb_dict_3 = self.get_sasa_layer_loss() point_loss = point_loss + point_loss_sasa tb_dict.update(tb_dict_3) return point_loss, tb_dict def forward(self, batch_dict): """ Args: batch_dict: batch_size: point_features: (N1 + N2 + N3 + ..., C) or (B, N, C) point_features_before_fusion: (N1 + N2 + N3 + ..., C) point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] point_labels (optional): (N1 + N2 + N3 + ...) gt_boxes (optional): (B, M, 8) Returns: batch_dict: point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) """ if self.model_cfg.get('USE_POINT_FEATURES_BEFORE_FUSION', False): point_features = batch_dict['point_features_before_fusion'] else: point_features = batch_dict['point_features'] point_cls_preds = self.cls_layers(point_features) # (total_points, num_class) point_box_preds = self.box_layers(point_features) # (total_points, box_code_size) point_cls_preds_max, _ = point_cls_preds.max(dim=-1) batch_dict['point_cls_scores'] = torch.sigmoid(point_cls_preds_max) ret_dict = {'point_cls_preds': point_cls_preds, 'point_box_preds': point_box_preds} if self.training: targets_dict = self.assign_targets(batch_dict) ret_dict['point_cls_labels'] = targets_dict['point_cls_labels'] ret_dict['point_box_labels'] = targets_dict['point_box_labels'] if self.loss_point_sasa is not None: point_sasa_labels = self.loss_point_sasa( batch_dict['point_coords_list'], batch_dict['point_scores_list'], batch_dict['gt_boxes'] ) ret_dict.update({ 'point_sasa_preds': batch_dict['point_scores_list'], 'point_sasa_labels': point_sasa_labels }) if not self.training or self.predict_boxes_when_training: point_cls_preds, point_box_preds = self.generate_predicted_boxes( points=batch_dict['point_coords'][:, 1:4], point_cls_preds=point_cls_preds, point_box_preds=point_box_preds ) batch_dict['batch_cls_preds'] = point_cls_preds batch_dict['batch_box_preds'] = point_box_preds batch_dict['batch_index'] = batch_dict['point_coords'][:, 0] batch_dict['cls_preds_normalized'] = False self.forward_ret_dict = ret_dict return batch_dict
6,616
41.146497
106
py
SASA
SASA-main/pcdet/models/dense_heads/point_head_simple.py
import torch from ...utils import box_utils from .point_head_template import PointHeadTemplate class PointHeadSimple(PointHeadTemplate): """ A simple point-based segmentation head, which are used for PV-RCNN keypoint segmentaion. Reference Paper: https://arxiv.org/abs/1912.13192 PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection """ def __init__(self, num_class, input_channels, model_cfg, **kwargs): super().__init__(model_cfg=model_cfg, num_class=num_class) self.cls_layers = self.make_fc_layers( fc_cfg=self.model_cfg.CLS_FC, input_channels=input_channels, output_channels=num_class ) def assign_targets(self, input_dict): """ Args: input_dict: point_features: (N1 + N2 + N3 + ..., C) batch_size: point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes (optional): (B, M, 8) Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_part_labels: (N1 + N2 + N3 + ..., 3) """ point_coords = input_dict['point_coords'] gt_boxes = input_dict['gt_boxes'] assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert point_coords.shape.__len__() in [2], 'points.shape=%s' % str(point_coords.shape) batch_size = gt_boxes.shape[0] extend_gt_boxes = box_utils.enlarge_box3d( gt_boxes.view(-1, gt_boxes.shape[-1]), extra_width=self.model_cfg.TARGET_CONFIG.GT_EXTRA_WIDTH ).view(batch_size, -1, gt_boxes.shape[-1]) targets_dict = self.assign_stack_targets( points=point_coords, gt_boxes=gt_boxes, extend_gt_boxes=extend_gt_boxes, set_ignore_flag=True, use_ball_constraint=False, ret_part_labels=False ) return targets_dict def get_loss(self, tb_dict=None): tb_dict = {} if tb_dict is None else tb_dict point_loss_cls, tb_dict_1 = self.get_cls_layer_loss() point_loss = point_loss_cls tb_dict.update(tb_dict_1) return point_loss, tb_dict def forward(self, batch_dict): """ Args: batch_dict: batch_size: point_features: (N1 + N2 + N3 + ..., C) or (B, N, C) point_features_before_fusion: (N1 + N2 + N3 + ..., C) point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] point_labels (optional): (N1 + N2 + N3 + ...) gt_boxes (optional): (B, M, 8) Returns: batch_dict: point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) """ if self.model_cfg.get('USE_POINT_FEATURES_BEFORE_FUSION', False): point_features = batch_dict['point_features_before_fusion'] else: point_features = batch_dict['point_features'] point_cls_preds = self.cls_layers(point_features) # (total_points, num_class) ret_dict = { 'point_cls_preds': point_cls_preds, } point_cls_scores = torch.sigmoid(point_cls_preds) batch_dict['point_cls_scores'], _ = point_cls_scores.max(dim=-1) if self.training: targets_dict = self.assign_targets(batch_dict) ret_dict['point_cls_labels'] = targets_dict['point_cls_labels'] self.forward_ret_dict = ret_dict return batch_dict
3,568
37.793478
106
py
SASA
SASA-main/pcdet/models/dense_heads/point_head_vote.py
import torch import torch.nn as nn import torch.nn.functional as F import numpy as np from ...ops.iou3d_nms import iou3d_nms_utils from ...ops.roiaware_pool3d import roiaware_pool3d_utils from ...ops.pointnet2.pointnet2_batch import pointnet2_modules from ...utils import box_coder_utils, box_utils, common_utils, loss_utils from .point_head_template import PointHeadTemplate class PointHeadVote(PointHeadTemplate): """ A simple vote-based detection head, which is used for 3DSSD. Reference Paper: https://arxiv.org/abs/2002.10187 3DSSD: Point-based 3D Single Stage Object Detector """ def __init__(self, num_class, input_channels, model_cfg, predict_boxes_when_training=False, **kwargs): super().__init__(model_cfg=model_cfg, num_class=num_class) use_bn = self.model_cfg.USE_BN self.predict_boxes_when_training = predict_boxes_when_training self.vote_cfg = self.model_cfg.VOTE_CONFIG self.vote_layers = self.make_fc_layers( input_channels=input_channels, output_channels=3, fc_list=self.vote_cfg.VOTE_FC ) self.sa_cfg = self.model_cfg.SA_CONFIG channel_in, channel_out = input_channels, 0 mlps = self.sa_cfg.MLPS.copy() for idx in range(mlps.__len__()): mlps[idx] = [channel_in] + mlps[idx] channel_out += mlps[idx][-1] self.SA_module = pointnet2_modules.PointnetSAModuleFSMSG( radii=self.sa_cfg.RADIUS, nsamples=self.sa_cfg.NSAMPLE, mlps=mlps, use_xyz=True, bn=use_bn ) channel_in = channel_out shared_fc_list = [] for k in range(0, self.model_cfg.SHARED_FC.__len__()): shared_fc_list.extend([ nn.Conv1d(channel_in, self.model_cfg.SHARED_FC[k], kernel_size=1, bias=False), nn.BatchNorm1d(self.model_cfg.SHARED_FC[k]), nn.ReLU() ]) channel_in = self.model_cfg.SHARED_FC[k] self.shared_fc_layer = nn.Sequential(*shared_fc_list) channel_in = self.model_cfg.SHARED_FC[-1] self.cls_layers = self.make_fc_layers( input_channels=channel_in, output_channels=num_class if not self.model_cfg.LOSS_CONFIG.LOSS_CLS == 'CrossEntropy' else num_class + 1, fc_list=self.model_cfg.CLS_FC ) target_cfg = self.model_cfg.TARGET_CONFIG self.box_coder = getattr(box_coder_utils, target_cfg.BOX_CODER)( **target_cfg.BOX_CODER_CONFIG ) self.reg_layers = self.make_fc_layers( input_channels=channel_in, output_channels=self.box_coder.code_size, fc_list=self.model_cfg.REG_FC ) self.init_weights(weight_init='xavier') def init_weights(self, weight_init='xavier'): if weight_init == 'kaiming': init_func = nn.init.kaiming_normal_ elif weight_init == 'xavier': init_func = nn.init.xavier_normal_ elif weight_init == 'normal': init_func = nn.init.normal_ else: raise NotImplementedError for m in self.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Conv1d): if weight_init == 'normal': init_func(m.weight, mean=0, std=0.001) else: init_func(m.weight) if m.bias is not None: nn.init.constant_(m.bias, 0) def build_losses(self, losses_cfg): # classification loss if losses_cfg.LOSS_CLS.startswith('WeightedBinaryCrossEntropy'): self.add_module( 'cls_loss_func', loss_utils.WeightedBinaryCrossEntropyLoss() ) elif losses_cfg.LOSS_CLS == 'WeightedCrossEntropy': self.add_module( 'cls_loss_func', loss_utils.WeightedCrossEntropyLoss() ) elif losses_cfg.LOSS_CLS == 'FocalLoss': self.add_module( 'cls_loss_func', loss_utils.SigmoidFocalClassificationLoss( **losses_cfg.get('LOSS_CLS_CONFIG', {}) ) ) else: raise NotImplementedError # regression loss if losses_cfg.LOSS_REG == 'WeightedSmoothL1Loss': self.add_module( 'reg_loss_func', loss_utils.WeightedSmoothL1Loss( code_weights=losses_cfg.LOSS_WEIGHTS.get('code_weights', None), **losses_cfg.get('LOSS_REG_CONFIG', {}) ) ) elif losses_cfg.LOSS_REG == 'WeightedL1Loss': self.add_module( 'reg_loss_func', loss_utils.WeightedL1Loss( code_weights=losses_cfg.LOSS_WEIGHTS.get('code_weights', None) ) ) else: raise NotImplementedError # sasa loss loss_sasa_cfg = losses_cfg.get('LOSS_SASA_CONFIG', None) if loss_sasa_cfg is not None: self.enable_sasa = True self.add_module( 'loss_point_sasa', loss_utils.PointSASALoss(**loss_sasa_cfg) ) else: self.enable_sasa = False def make_fc_layers(self, input_channels, output_channels, fc_list): fc_layers = [] pre_channel = input_channels for k in range(0, fc_list.__len__()): fc_layers.extend([ nn.Conv1d(pre_channel, fc_list[k], kernel_size=1, bias=False), nn.BatchNorm1d(fc_list[k]), nn.ReLU() ]) pre_channel = fc_list[k] fc_layers.append(nn.Conv1d(pre_channel, output_channels, kernel_size=1, bias=True)) fc_layers = nn.Sequential(*fc_layers) return fc_layers def assign_stack_targets_simple(self, points, gt_boxes, extend_gt_boxes=None, set_ignore_flag=True): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes: (B, M, 8) extend_gt_boxes: (B, M, 8), required if set ignore flag set_ignore_flag: Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignore point_reg_labels: (N1 + N2 + N3 + ..., 3), corresponding object centroid """ assert len(points.shape) == 2 and points.shape[1] == 4, 'points.shape=%s' % str(points.shape) assert len(gt_boxes.shape) == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert extend_gt_boxes is None or len(extend_gt_boxes.shape) == 3, \ 'extend_gt_boxes.shape=%s' % str(extend_gt_boxes.shape) assert not set_ignore_flag or extend_gt_boxes is not None batch_size = gt_boxes.shape[0] bs_idx = points[:, 0] point_cls_labels = points.new_zeros(points.shape[0]).long() point_reg_labels = gt_boxes.new_zeros((points.shape[0], 3)) for k in range(batch_size): bs_mask = (bs_idx == k) points_single = points[bs_mask][:, 1:4] point_cls_labels_single = point_cls_labels.new_zeros(bs_mask.sum()) box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) if extend_gt_boxes is not None: extend_box_idx_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), extend_gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) fg_flag = box_fg_flag ignore_flag = fg_flag ^ (extend_box_idx_of_pts >= 0) point_cls_labels_single[ignore_flag] = -1 gt_box_of_fg_points = gt_boxes[k][box_idxs_of_pts[box_fg_flag]] point_cls_labels_single[box_fg_flag] = 1 point_cls_labels[bs_mask] = point_cls_labels_single point_reg_labels_single = point_reg_labels.new_zeros((bs_mask.sum(), 3)) point_reg_labels_single[box_fg_flag] = gt_box_of_fg_points[:, 0:3] point_reg_labels[bs_mask] = point_reg_labels_single targets_dict = { 'point_cls_labels': point_cls_labels, 'point_reg_labels': point_reg_labels, } return targets_dict def assign_targets_simple(self, points, gt_boxes, extra_width=None, set_ignore_flag=True): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes: (B, M, 8) extra_width: (dx, dy, dz) extra width applied to gt boxes assign_method: binary or distance set_ignore_flag: Returns: point_vote_labels: (N1 + N2 + N3 + ..., 3) """ assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert points.shape.__len__() in [2], 'points.shape=%s' % str(points.shape) batch_size = gt_boxes.shape[0] extend_gt_boxes = box_utils.enlarge_box3d( gt_boxes.view(-1, gt_boxes.shape[-1]), extra_width=extra_width ).view(batch_size, -1, gt_boxes.shape[-1]) \ if extra_width is not None else gt_boxes if set_ignore_flag: targets_dict = self.assign_stack_targets_simple(points=points, gt_boxes=gt_boxes, extend_gt_boxes=extend_gt_boxes, set_ignore_flag=set_ignore_flag) else: targets_dict = self.assign_stack_targets_simple(points=points, gt_boxes=extend_gt_boxes, set_ignore_flag=set_ignore_flag) return targets_dict def assign_stack_targets_mask(self, points, gt_boxes, extend_gt_boxes=None, set_ignore_flag=True, use_ball_constraint=False, central_radius=2.0): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes: (B, M, 8) extend_gt_boxes: [B, M, 8] set_ignore_flag: use_ball_constraint: central_radius: Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_reg_labels: (N1 + N2 + N3 + ..., code_size) point_box_labels: (N1 + N2 + N3 + ..., 7) """ assert len(points.shape) == 2 and points.shape[1] == 4, 'points.shape=%s' % str(points.shape) assert len(gt_boxes.shape) == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert extend_gt_boxes is None or len(extend_gt_boxes.shape) == 3, \ 'extend_gt_boxes.shape=%s' % str(extend_gt_boxes.shape) assert set_ignore_flag != use_ball_constraint, 'Choose one only!' batch_size = gt_boxes.shape[0] bs_idx = points[:, 0] point_cls_labels = gt_boxes.new_zeros(points.shape[0]).long() point_reg_labels = gt_boxes.new_zeros((points.shape[0], self.box_coder.code_size)) point_box_labels = gt_boxes.new_zeros((points.shape[0], gt_boxes.size(2) - 1)) for k in range(batch_size): bs_mask = (bs_idx == k) points_single = points[bs_mask][:, 1:4] point_cls_labels_single = point_cls_labels.new_zeros(bs_mask.sum()) box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) if set_ignore_flag: extend_box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), extend_gt_boxes[k:k+1, :, 0:7].contiguous() ).long().squeeze(dim=0) fg_flag = box_fg_flag ignore_flag = fg_flag ^ (extend_box_idxs_of_pts >= 0) point_cls_labels_single[ignore_flag] = -1 elif use_ball_constraint: box_centers = gt_boxes[k][box_idxs_of_pts][:, 0:3].clone() ball_flag = ((box_centers - points_single).norm(dim=1) < central_radius) fg_flag = box_fg_flag & ball_flag ignore_flag = fg_flag ^ box_fg_flag point_cls_labels_single[ignore_flag] = -1 else: raise NotImplementedError gt_box_of_fg_points = gt_boxes[k][box_idxs_of_pts[fg_flag]] point_cls_labels_single[fg_flag] = 1 if self.num_class == 1 else gt_box_of_fg_points[:, -1].long() point_cls_labels[bs_mask] = point_cls_labels_single if gt_box_of_fg_points.shape[0] > 0: point_reg_labels_single = point_reg_labels.new_zeros((bs_mask.sum(), self.box_coder.code_size)) fg_point_box_labels = self.box_coder.encode_torch( gt_boxes=gt_box_of_fg_points[:, :-1], points=points_single[fg_flag], gt_classes=gt_box_of_fg_points[:, -1].long() ) point_reg_labels_single[fg_flag] = fg_point_box_labels point_reg_labels[bs_mask] = point_reg_labels_single point_box_labels_single = point_box_labels.new_zeros((bs_mask.sum(), gt_boxes.size(2) - 1)) point_box_labels_single[fg_flag] = gt_box_of_fg_points[:, :-1] point_box_labels[bs_mask] = point_box_labels_single targets_dict = { 'point_cls_labels': point_cls_labels, 'point_reg_labels': point_reg_labels, 'point_box_labels': point_box_labels } return targets_dict def assign_stack_targets_iou(self, points, pred_boxes, gt_boxes, pos_iou_threshold=0.5, neg_iou_threshold=0.35): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] pred_boxes: (N, 7/8) gt_boxes: (B, M, 8) pos_iou_threshold: neg_iou_threshold: Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_reg_labels: (N1 + N2 + N3 + ..., code_size) point_box_labels: (N1 + N2 + N3 + ..., 7) """ assert len(points.shape) == 2 and points.shape[1] == 4, 'points.shape=%s' % str(points.shape) assert len(pred_boxes.shape) == 2 and pred_boxes.shape[1] >= 7, 'pred_boxes.shape=%s' % str(pred_boxes.shape) assert len(gt_boxes.shape) == 3 and gt_boxes.shape[2] == 8, 'gt_boxes.shape=%s' % str(gt_boxes.shape) batch_size = gt_boxes.shape[0] bs_idx = points[:, 0] point_cls_labels = gt_boxes.new_zeros(pred_boxes.shape[0]).long() point_reg_labels = gt_boxes.new_zeros((pred_boxes.shape[0], self.box_coder.code_size)) point_box_labels = gt_boxes.new_zeros((pred_boxes.shape[0], 7)) for k in range(batch_size): bs_mask = (bs_idx == k) points_single = points[bs_mask][:, 1:4] pred_boxes_single = pred_boxes[bs_mask] point_cls_labels_single = point_cls_labels.new_zeros(bs_mask.sum()) pred_boxes_iou = iou3d_nms_utils.boxes_iou3d_gpu( pred_boxes_single, gt_boxes[k][:, :7] ) pred_boxes_iou, box_idxs_of_pts = torch.max(pred_boxes_iou, dim=-1) fg_flag = pred_boxes_iou > pos_iou_threshold ignore_flag = (pred_boxes_iou > neg_iou_threshold) ^ fg_flag gt_box_of_fg_points = gt_boxes[k][box_idxs_of_pts[fg_flag]] point_cls_labels_single[fg_flag] = 1 if self.num_class == 1 else gt_box_of_fg_points[:, -1].long() point_cls_labels_single[ignore_flag] = -1 point_cls_labels[bs_mask] = point_cls_labels_single if gt_box_of_fg_points.shape[0] > 0: point_reg_labels_single = point_reg_labels.new_zeros((bs_mask.sum(), self.box_coder.code_size)) fg_point_box_labels = self.box_coder.encode_torch( gt_boxes=gt_box_of_fg_points[:, :-1], points=points_single[fg_flag], gt_classes=gt_box_of_fg_points[:, -1].long() ) point_reg_labels_single[fg_flag] = fg_point_box_labels point_reg_labels[bs_mask] = point_reg_labels_single point_box_labels_single = point_box_labels.new_zeros((bs_mask.sum(), 7)) point_box_labels_single[fg_flag] = gt_box_of_fg_points[:, :-1] point_box_labels[bs_mask] = point_box_labels_single targets_dict = { 'point_cls_labels': point_cls_labels, 'point_reg_labels': point_reg_labels, 'point_box_labels': point_box_labels } return targets_dict def assign_targets(self, input_dict): """ Args: input_dict: batch_size: point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes (optional): (B, M, 8) Returns: point_part_labels: (N1 + N2 + N3 + ..., 3) """ assign_method = self.model_cfg.TARGET_CONFIG.ASSIGN_METHOD # mask or iou if assign_method == 'mask': points = input_dict['point_vote_coords'] gt_boxes = input_dict['gt_boxes'] assert points.shape.__len__() == 2, 'points.shape=%s' % str(points.shape) assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) central_radius = self.model_cfg.TARGET_CONFIG.get('GT_CENTRAL_RADIUS', 2.0) targets_dict = self.assign_stack_targets_mask( points=points, gt_boxes=gt_boxes, set_ignore_flag=False, use_ball_constraint=True, central_radius=central_radius ) elif assign_method == 'iou': points = input_dict['point_vote_coords'] pred_boxes = input_dict['point_box_preds'] gt_boxes = input_dict['gt_boxes'] assert points.shape.__len__() == 2, 'points.shape=%s' % str(points.shape) assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert pred_boxes.shape.__len__() == 2, 'pred_boxes.shape=%s' % str(pred_boxes.shape) pos_iou_threshold = self.model_cfg.TARGET_CONFIG.POS_IOU_THRESHOLD neg_iou_threshold = self.model_cfg.TARGET_CONFIG.NEG_IOU_THRESHOLD targets_dict = self.assign_stack_targets_iou( points=points, pred_boxes=pred_boxes, gt_boxes=gt_boxes, pos_iou_threshold=pos_iou_threshold, neg_iou_threshold=neg_iou_threshold ) else: raise NotImplementedError return targets_dict def get_vote_layer_loss(self, tb_dict=None): pos_mask = self.forward_ret_dict['vote_cls_labels'] > 0 vote_reg_labels = self.forward_ret_dict['vote_reg_labels'] vote_reg_preds = self.forward_ret_dict['point_vote_coords'] reg_weights = pos_mask.float() pos_normalizer = pos_mask.sum().float() reg_weights /= torch.clamp(pos_normalizer, min=1.0) vote_loss_reg_src = self.reg_loss_func( vote_reg_preds[None, ...], vote_reg_labels[None, ...], weights=reg_weights[None, ...]) vote_loss_reg = vote_loss_reg_src.sum() loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS vote_loss_reg = vote_loss_reg * loss_weights_dict['vote_reg_weight'] if tb_dict is None: tb_dict = {} tb_dict.update({'vote_loss_reg': vote_loss_reg.item()}) return vote_loss_reg, tb_dict @torch.no_grad() def generate_centerness_label(self, point_base, point_box_labels, pos_mask, epsilon=1e-6): """ Args: point_base: (N1 + N2 + N3 + ..., 3) point_box_labels: (N1 + N2 + N3 + ..., 7) pos_mask: (N1 + N2 + N3 + ...) epsilon: Returns: centerness_label: (N1 + N2 + N3 + ...) """ centerness = point_box_labels.new_zeros(pos_mask.shape) point_box_labels = point_box_labels[pos_mask, :] canonical_xyz = point_base[pos_mask, :] - point_box_labels[:, :3] rys = point_box_labels[:, -1] canonical_xyz = common_utils.rotate_points_along_z( canonical_xyz.unsqueeze(dim=1), -rys ).squeeze(dim=1) distance_front = point_box_labels[:, 3] / 2 - canonical_xyz[:, 0] distance_back = point_box_labels[:, 3] / 2 + canonical_xyz[:, 0] distance_left = point_box_labels[:, 4] / 2 - canonical_xyz[:, 1] distance_right = point_box_labels[:, 4] / 2 + canonical_xyz[:, 1] distance_top = point_box_labels[:, 5] / 2 - canonical_xyz[:, 2] distance_bottom = point_box_labels[:, 5] / 2 + canonical_xyz[:, 2] centerness_l = torch.min(distance_front, distance_back) / torch.max(distance_front, distance_back) centerness_w = torch.min(distance_left, distance_right) / torch.max(distance_left, distance_right) centerness_h = torch.min(distance_top, distance_bottom) / torch.max(distance_top, distance_bottom) centerness_pos = torch.clamp(centerness_l * centerness_w * centerness_h, min=epsilon) ** (1 / 3.0) centerness[pos_mask] = centerness_pos return centerness def get_axis_aligned_iou_loss_lidar(self, pred_boxes: torch.Tensor, gt_boxes: torch.Tensor): """ Args: pred_boxes: (N, 7) float Tensor. gt_boxes: (N, 7) float Tensor. Returns: iou_loss: (N) float Tensor. """ assert pred_boxes.shape[0] == gt_boxes.shape[0] pos_p, len_p, *cps = torch.split(pred_boxes, 3, dim=-1) pos_g, len_g, *cgs = torch.split(gt_boxes, 3, dim=-1) len_p = torch.clamp(len_p, min=1e-5) len_g = torch.clamp(len_g, min=1e-5) vol_p = len_p.prod(dim=-1) vol_g = len_g.prod(dim=-1) min_p, max_p = pos_p - len_p / 2, pos_p + len_p / 2 min_g, max_g = pos_g - len_g / 2, pos_g + len_g / 2 min_max = torch.min(max_p, max_g) max_min = torch.max(min_p, min_g) diff = torch.clamp(min_max - max_min, min=0) intersection = diff.prod(dim=-1) union = vol_p + vol_g - intersection iou_axis_aligned = intersection / torch.clamp(union, min=1e-5) iou_loss = 1 - iou_axis_aligned return iou_loss def get_corner_loss_lidar(self, pred_boxes: torch.Tensor, gt_boxes: torch.Tensor): """ Args: pred_boxes: (N, 7) float Tensor. gt_boxes: (N, 7) float Tensor. Returns: corner_loss: (N) float Tensor. """ assert pred_boxes.shape[0] == gt_boxes.shape[0] pred_box_corners = box_utils.boxes_to_corners_3d(pred_boxes) gt_box_corners = box_utils.boxes_to_corners_3d(gt_boxes) gt_boxes_flip = gt_boxes.clone() gt_boxes_flip[:, 6] += np.pi gt_box_corners_flip = box_utils.boxes_to_corners_3d(gt_boxes_flip) # (N, 8, 3) corner_loss = loss_utils.WeightedSmoothL1Loss.smooth_l1_loss(pred_box_corners - gt_box_corners, 1.0) corner_loss_flip = loss_utils.WeightedSmoothL1Loss.smooth_l1_loss(pred_box_corners - gt_box_corners_flip, 1.0) corner_loss = torch.min(corner_loss.sum(dim=2), corner_loss_flip.sum(dim=2)) return corner_loss.mean(dim=1) def get_cls_layer_loss(self, tb_dict=None): point_cls_labels = self.forward_ret_dict['point_cls_labels'].view(-1) point_cls_preds = self.forward_ret_dict['point_cls_preds'].view(-1, self.num_class) positives = point_cls_labels > 0 negatives = point_cls_labels == 0 cls_weights = positives * 1.0 + negatives * 1.0 one_hot_targets = point_cls_preds.new_zeros(*list(point_cls_labels.shape), self.num_class + 1) one_hot_targets.scatter_(-1, (point_cls_labels * (point_cls_labels >= 0).long()).unsqueeze(dim=-1).long(), 1.0) self.forward_ret_dict['point_cls_labels_onehot'] = one_hot_targets loss_cfgs = self.model_cfg.LOSS_CONFIG if 'WithCenterness' in loss_cfgs.LOSS_CLS: point_base = self.forward_ret_dict['point_vote_coords'] point_box_labels = self.forward_ret_dict['point_box_labels'] centerness_label = self.generate_centerness_label(point_base, point_box_labels, positives) loss_cls_cfg = loss_cfgs.get('LOSS_CLS_CONFIG', None) centerness_min = loss_cls_cfg['centerness_min'] if loss_cls_cfg is not None else 0.0 centerness_max = loss_cls_cfg['centerness_max'] if loss_cls_cfg is not None else 1.0 centerness_label = centerness_min + (centerness_max - centerness_min) * centerness_label one_hot_targets *= centerness_label.unsqueeze(dim=-1) point_loss_cls = self.cls_loss_func(point_cls_preds, one_hot_targets[..., 1:], weights=cls_weights) loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS point_loss_cls = point_loss_cls * loss_weights_dict['point_cls_weight'] if tb_dict is None: tb_dict = {} tb_dict.update({ 'point_pos_num': positives.sum().item() }) return point_loss_cls, cls_weights, tb_dict # point_loss_cls: (N) def get_box_layer_loss(self, tb_dict=None): pos_mask = self.forward_ret_dict['point_cls_labels'] > 0 point_reg_preds = self.forward_ret_dict['point_reg_preds'] point_reg_labels = self.forward_ret_dict['point_reg_labels'] reg_weights = pos_mask.float() loss_weights_dict = self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS if tb_dict is None: tb_dict = {} point_loss_offset_reg = self.reg_loss_func( point_reg_preds[None, :, :6], point_reg_labels[None, :, :6], weights=reg_weights[None, ...] ) point_loss_offset_reg = point_loss_offset_reg.sum(dim=-1).squeeze() if hasattr(self.box_coder, 'pred_velo') and self.box_coder.pred_velo: point_loss_velo_reg = self.reg_loss_func( point_reg_preds[None, :, 6 + 2 * self.box_coder.angle_bin_num:8 + 2 * self.box_coder.angle_bin_num], point_reg_labels[None, :, 6 + 2 * self.box_coder.angle_bin_num:8 + 2 * self.box_coder.angle_bin_num], weights=reg_weights[None, ...] ) point_loss_velo_reg = point_loss_velo_reg.sum(dim=-1).squeeze() point_loss_offset_reg = point_loss_offset_reg + point_loss_velo_reg point_loss_offset_reg *= loss_weights_dict['point_offset_reg_weight'] if isinstance(self.box_coder, box_coder_utils.PointBinResidualCoder): point_angle_cls_labels = \ point_reg_labels[:, 6:6 + self.box_coder.angle_bin_num] point_loss_angle_cls = F.cross_entropy( # angle bin cls point_reg_preds[:, 6:6 + self.box_coder.angle_bin_num], point_angle_cls_labels.argmax(dim=-1), reduction='none') * reg_weights point_angle_reg_preds = point_reg_preds[:, 6 + self.box_coder.angle_bin_num:6 + 2 * self.box_coder.angle_bin_num] point_angle_reg_labels = point_reg_labels[:, 6 + self.box_coder.angle_bin_num:6 + 2 * self.box_coder.angle_bin_num] point_angle_reg_preds = (point_angle_reg_preds * point_angle_cls_labels).sum(dim=-1, keepdim=True) point_angle_reg_labels = (point_angle_reg_labels * point_angle_cls_labels).sum(dim=-1, keepdim=True) point_loss_angle_reg = self.reg_loss_func( point_angle_reg_preds[None, ...], point_angle_reg_labels[None, ...], weights=reg_weights[None, ...] ) point_loss_angle_reg = point_loss_angle_reg.squeeze() point_loss_angle_cls *= loss_weights_dict['point_angle_cls_weight'] point_loss_angle_reg *= loss_weights_dict['point_angle_reg_weight'] point_loss_box = point_loss_offset_reg + point_loss_angle_cls + point_loss_angle_reg # (N) else: point_angle_reg_preds = point_reg_preds[:, 6:] point_angle_reg_labels = point_reg_labels[:, 6:] point_loss_angle_reg = self.reg_loss_func( point_angle_reg_preds[None, ...], point_angle_reg_labels[None, ...], weights=reg_weights[None, ...] ) point_loss_angle_reg *= loss_weights_dict['point_angle_reg_weight'] point_loss_box = point_loss_offset_reg + point_loss_angle_reg if reg_weights.sum() > 0: point_box_preds = self.forward_ret_dict['point_box_preds'] point_box_labels = self.forward_ret_dict['point_box_labels'] point_loss_box_aux = 0 if self.model_cfg.LOSS_CONFIG.get('AXIS_ALIGNED_IOU_LOSS_REGULARIZATION', False): point_loss_iou = self.get_axis_aligned_iou_loss_lidar( point_box_preds[pos_mask, :], point_box_labels[pos_mask, :] ) point_loss_iou *= self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['point_iou_weight'] point_loss_box_aux = point_loss_box_aux + point_loss_iou if self.model_cfg.LOSS_CONFIG.get('CORNER_LOSS_REGULARIZATION', False): point_loss_corner = self.get_corner_loss_lidar( point_box_preds[pos_mask, 0:7], point_box_labels[pos_mask, 0:7] ) point_loss_corner *= self.model_cfg.LOSS_CONFIG.LOSS_WEIGHTS['point_corner_weight'] point_loss_box_aux = point_loss_box_aux + point_loss_corner point_loss_box[pos_mask] = point_loss_box[pos_mask] + point_loss_box_aux return point_loss_box, reg_weights, tb_dict # point_loss_box: (N) def get_sasa_layer_loss(self, tb_dict=None): if self.enable_sasa: point_loss_sasa_list = self.loss_point_sasa.loss_forward( self.forward_ret_dict['point_sasa_preds'], self.forward_ret_dict['point_sasa_labels'] ) point_loss_sasa = 0 tb_dict = dict() for i in range(len(point_loss_sasa_list)): cur_point_loss_sasa = point_loss_sasa_list[i] if cur_point_loss_sasa is None: continue point_loss_sasa = point_loss_sasa + cur_point_loss_sasa tb_dict['point_loss_sasa_layer_%d' % i] = point_loss_sasa_list[i].item() tb_dict['point_loss_sasa'] = point_loss_sasa.item() return point_loss_sasa, tb_dict else: return None, None def get_loss(self, tb_dict=None): tb_dict = {} if tb_dict is None else tb_dict point_loss_vote, tb_dict_0 = self.get_vote_layer_loss() point_loss_cls, cls_weights, tb_dict_1 = self.get_cls_layer_loss() point_loss_box, box_weights, tb_dict_2 = self.get_box_layer_loss() point_loss_cls = point_loss_cls.sum() / torch.clamp(cls_weights.sum(), min=1.0) point_loss_box = point_loss_box.sum() / torch.clamp(box_weights.sum(), min=1.0) tb_dict.update({ 'point_loss_vote': point_loss_vote.item(), 'point_loss_cls': point_loss_cls.item(), 'point_loss_box': point_loss_box.item() }) point_loss = point_loss_vote + point_loss_cls + point_loss_box tb_dict.update(tb_dict_0) tb_dict.update(tb_dict_1) tb_dict.update(tb_dict_2) point_loss_sasa, tb_dict_3 = self.get_sasa_layer_loss() if point_loss_sasa is not None: tb_dict.update(tb_dict_3) point_loss += point_loss_sasa return point_loss, tb_dict def forward(self, batch_dict): """ Args: batch_dict: batch_size: point_features: (N1 + N2 + N3 + ..., C) point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] point_scores (optional): (B, N) gt_boxes (optional): (B, M, 8) Returns: batch_dict: point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) """ batch_size = batch_dict['batch_size'] point_coords = batch_dict['point_coords'] point_features = batch_dict['point_features'] batch_idx, point_coords = point_coords[:, 0], point_coords[:, 1:4] batch_idx = batch_idx.view(batch_size, -1, 1) point_coords = point_coords.view(batch_size, -1, 3).contiguous() point_features = point_features.reshape( batch_size, point_coords.size(1), -1 ).permute(0, 2, 1).contiguous() # candidate points sampling sample_range = self.model_cfg.SAMPLE_RANGE sample_batch_idx = batch_idx[:, sample_range[0]:sample_range[1], :].contiguous() candidate_coords = point_coords[:, sample_range[0]:sample_range[1], :].contiguous() candidate_features = point_features[:, :, sample_range[0]:sample_range[1]].contiguous() # generate vote points vote_offsets = self.vote_layers(candidate_features) # (B, 3, N) vote_translation_range = np.array(self.vote_cfg.MAX_TRANSLATION_RANGE, dtype=np.float32) vote_translation_range = torch.from_numpy(vote_translation_range).cuda().unsqueeze(dim=0).unsqueeze(dim=-1) vote_offsets = torch.max(vote_offsets, -vote_translation_range) vote_offsets = torch.min(vote_offsets, vote_translation_range) vote_coords = candidate_coords + vote_offsets.permute(0, 2, 1).contiguous() ret_dict = {'batch_size': batch_size, 'point_candidate_coords': candidate_coords.view(-1, 3).contiguous(), 'point_vote_coords': vote_coords.view(-1, 3).contiguous()} sample_batch_idx_flatten = sample_batch_idx.view(-1, 1).contiguous() # (N, 1) batch_dict['batch_index'] = sample_batch_idx_flatten.squeeze(-1) batch_dict['point_candidate_coords'] = torch.cat( # (N, 4) (sample_batch_idx_flatten, ret_dict['point_candidate_coords']), dim=-1) batch_dict['point_vote_coords'] = torch.cat( # (N, 4) (sample_batch_idx_flatten, ret_dict['point_vote_coords']), dim=-1) if self.training: # assign targets for vote loss extra_width = self.model_cfg.TARGET_CONFIG.get('VOTE_EXTRA_WIDTH', None) targets_dict = self.assign_targets_simple(batch_dict['point_candidate_coords'], batch_dict['gt_boxes'], extra_width=extra_width, set_ignore_flag=False) ret_dict['vote_cls_labels'] = targets_dict['point_cls_labels'] # (N) ret_dict['vote_reg_labels'] = targets_dict['point_reg_labels'] # (N, 3) _, point_features, _ = self.SA_module( point_coords, point_features, new_xyz=vote_coords ) point_features = self.shared_fc_layer(point_features) point_cls_preds = self.cls_layers(point_features) point_reg_preds = self.reg_layers(point_features) point_cls_preds = point_cls_preds.permute(0, 2, 1).contiguous() point_cls_preds = point_cls_preds.view(-1, point_cls_preds.shape[-1]).contiguous() point_reg_preds = point_reg_preds.permute(0, 2, 1).contiguous() point_reg_preds = point_reg_preds.view(-1, point_reg_preds.shape[-1]).contiguous() point_cls_scores = torch.sigmoid(point_cls_preds) batch_dict['point_cls_scores'] = point_cls_scores point_box_preds = self.box_coder.decode_torch(point_reg_preds, ret_dict['point_vote_coords']) batch_dict['point_box_preds'] = point_box_preds ret_dict.update({'point_cls_preds': point_cls_preds, 'point_reg_preds': point_reg_preds, 'point_box_preds': point_box_preds, 'point_cls_scores': point_cls_scores}) if self.training: targets_dict = self.assign_targets(batch_dict) ret_dict['point_cls_labels'] = targets_dict['point_cls_labels'] ret_dict['point_reg_labels'] = targets_dict['point_reg_labels'] ret_dict['point_box_labels'] = targets_dict['point_box_labels'] if self.enable_sasa: point_sasa_labels = self.loss_point_sasa( batch_dict['point_coords_list'], batch_dict['point_scores_list'], batch_dict['gt_boxes'] ) ret_dict.update({ 'point_sasa_preds': batch_dict['point_scores_list'], 'point_sasa_labels': point_sasa_labels }) if not self.training or self.predict_boxes_when_training: point_cls_preds, point_box_preds = self.generate_predicted_boxes( points=batch_dict['point_vote_coords'][:, 1:4], point_cls_preds=point_cls_preds, point_box_preds=point_reg_preds ) batch_dict['batch_cls_preds'] = point_cls_preds batch_dict['batch_box_preds'] = point_box_preds batch_dict['cls_preds_normalized'] = False self.forward_ret_dict = ret_dict return batch_dict
37,917
45.754624
127
py
SASA
SASA-main/pcdet/models/dense_heads/point_intra_part_head.py
import torch from ...utils import box_coder_utils, box_utils from .point_head_template import PointHeadTemplate class PointIntraPartOffsetHead(PointHeadTemplate): """ Point-based head for predicting the intra-object part locations. Reference Paper: https://arxiv.org/abs/1907.03670 From Points to Parts: 3D Object Detection from Point Cloud with Part-aware and Part-aggregation Network """ def __init__(self, num_class, input_channels, model_cfg, predict_boxes_when_training=False, **kwargs): super().__init__(model_cfg=model_cfg, num_class=num_class) self.predict_boxes_when_training = predict_boxes_when_training self.cls_layers = self.make_fc_layers( fc_cfg=self.model_cfg.CLS_FC, input_channels=input_channels, output_channels=num_class ) self.part_reg_layers = self.make_fc_layers( fc_cfg=self.model_cfg.PART_FC, input_channels=input_channels, output_channels=3 ) target_cfg = self.model_cfg.TARGET_CONFIG if target_cfg.get('BOX_CODER', None) is not None: self.box_coder = getattr(box_coder_utils, target_cfg.BOX_CODER)( **target_cfg.BOX_CODER_CONFIG ) self.box_layers = self.make_fc_layers( fc_cfg=self.model_cfg.REG_FC, input_channels=input_channels, output_channels=self.box_coder.code_size ) else: self.box_layers = None def assign_targets(self, input_dict): """ Args: input_dict: point_features: (N1 + N2 + N3 + ..., C) batch_size: point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes (optional): (B, M, 8) Returns: point_cls_labels: (N1 + N2 + N3 + ...), long type, 0:background, -1:ignored point_part_labels: (N1 + N2 + N3 + ..., 3) """ point_coords = input_dict['point_coords'] gt_boxes = input_dict['gt_boxes'] assert gt_boxes.shape.__len__() == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) assert point_coords.shape.__len__() in [2], 'points.shape=%s' % str(point_coords.shape) batch_size = gt_boxes.shape[0] extend_gt_boxes = box_utils.enlarge_box3d( gt_boxes.view(-1, gt_boxes.shape[-1]), extra_width=self.model_cfg.TARGET_CONFIG.GT_EXTRA_WIDTH ).view(batch_size, -1, gt_boxes.shape[-1]) targets_dict = self.assign_stack_targets( points=point_coords, gt_boxes=gt_boxes, extend_gt_boxes=extend_gt_boxes, set_ignore_flag=True, use_ball_constraint=False, ret_part_labels=True, ret_box_labels=(self.box_layers is not None) ) return targets_dict def get_loss(self, tb_dict=None): tb_dict = {} if tb_dict is None else tb_dict point_loss_cls, tb_dict = self.get_cls_layer_loss(tb_dict) point_loss_part, tb_dict = self.get_part_layer_loss(tb_dict) point_loss = point_loss_cls + point_loss_part if self.box_layers is not None: point_loss_box, tb_dict = self.get_box_layer_loss(tb_dict) point_loss += point_loss_box return point_loss, tb_dict def forward(self, batch_dict): """ Args: batch_dict: batch_size: point_features: (N1 + N2 + N3 + ..., C) or (B, N, C) point_coords: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] point_labels (optional): (N1 + N2 + N3 + ...) gt_boxes (optional): (B, M, 8) Returns: batch_dict: point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) """ point_features = batch_dict['point_features'] point_cls_preds = self.cls_layers(point_features) # (total_points, num_class) point_part_preds = self.part_reg_layers(point_features) ret_dict = { 'point_cls_preds': point_cls_preds, 'point_part_preds': point_part_preds, } if self.box_layers is not None: point_box_preds = self.box_layers(point_features) ret_dict['point_box_preds'] = point_box_preds point_cls_scores = torch.sigmoid(point_cls_preds) point_part_offset = torch.sigmoid(point_part_preds) batch_dict['point_cls_scores'], _ = point_cls_scores.max(dim=-1) batch_dict['point_part_offset'] = point_part_offset if self.training: targets_dict = self.assign_targets(batch_dict) ret_dict['point_cls_labels'] = targets_dict['point_cls_labels'] ret_dict['point_part_labels'] = targets_dict.get('point_part_labels') ret_dict['point_box_labels'] = targets_dict.get('point_box_labels') if self.box_layers is not None and (not self.training or self.predict_boxes_when_training): point_cls_preds, point_box_preds = self.generate_predicted_boxes( points=batch_dict['point_coords'][:, 1:4], point_cls_preds=point_cls_preds, point_box_preds=ret_dict['point_box_preds'] ) batch_dict['batch_cls_preds'] = point_cls_preds batch_dict['batch_box_preds'] = point_box_preds batch_dict['batch_index'] = batch_dict['point_coords'][:, 0] batch_dict['cls_preds_normalized'] = False self.forward_ret_dict = ret_dict return batch_dict
5,568
42.507813
107
py
SASA
SASA-main/pcdet/models/dense_heads/target_assigner/anchor_generator.py
import torch class AnchorGenerator(object): def __init__(self, anchor_range, anchor_generator_config): super().__init__() self.anchor_generator_cfg = anchor_generator_config self.anchor_range = anchor_range self.anchor_sizes = [config['anchor_sizes'] for config in anchor_generator_config] self.anchor_rotations = [config['anchor_rotations'] for config in anchor_generator_config] self.anchor_heights = [config['anchor_bottom_heights'] for config in anchor_generator_config] self.align_center = [config.get('align_center', False) for config in anchor_generator_config] assert len(self.anchor_sizes) == len(self.anchor_rotations) == len(self.anchor_heights) self.num_of_anchor_sets = len(self.anchor_sizes) def generate_anchors(self, grid_sizes): assert len(grid_sizes) == self.num_of_anchor_sets all_anchors = [] num_anchors_per_location = [] for grid_size, anchor_size, anchor_rotation, anchor_height, align_center in zip( grid_sizes, self.anchor_sizes, self.anchor_rotations, self.anchor_heights, self.align_center): num_anchors_per_location.append(len(anchor_rotation) * len(anchor_size) * len(anchor_height)) if align_center: x_stride = (self.anchor_range[3] - self.anchor_range[0]) / grid_size[0] y_stride = (self.anchor_range[4] - self.anchor_range[1]) / grid_size[1] x_offset, y_offset = x_stride / 2, y_stride / 2 else: x_stride = (self.anchor_range[3] - self.anchor_range[0]) / (grid_size[0] - 1) y_stride = (self.anchor_range[4] - self.anchor_range[1]) / (grid_size[1] - 1) x_offset, y_offset = 0, 0 x_shifts = torch.arange( self.anchor_range[0] + x_offset, self.anchor_range[3] + 1e-5, step=x_stride, dtype=torch.float32, ).cuda() y_shifts = torch.arange( self.anchor_range[1] + y_offset, self.anchor_range[4] + 1e-5, step=y_stride, dtype=torch.float32, ).cuda() z_shifts = x_shifts.new_tensor(anchor_height) num_anchor_size, num_anchor_rotation = anchor_size.__len__(), anchor_rotation.__len__() anchor_rotation = x_shifts.new_tensor(anchor_rotation) anchor_size = x_shifts.new_tensor(anchor_size) x_shifts, y_shifts, z_shifts = torch.meshgrid([ x_shifts, y_shifts, z_shifts ]) # [x_grid, y_grid, z_grid] anchors = torch.stack((x_shifts, y_shifts, z_shifts), dim=-1) # [x, y, z, 3] anchors = anchors[:, :, :, None, :].repeat(1, 1, 1, anchor_size.shape[0], 1) anchor_size = anchor_size.view(1, 1, 1, -1, 3).repeat([*anchors.shape[0:3], 1, 1]) anchors = torch.cat((anchors, anchor_size), dim=-1) anchors = anchors[:, :, :, :, None, :].repeat(1, 1, 1, 1, num_anchor_rotation, 1) anchor_rotation = anchor_rotation.view(1, 1, 1, 1, -1, 1).repeat([*anchors.shape[0:3], num_anchor_size, 1, 1]) anchors = torch.cat((anchors, anchor_rotation), dim=-1) # [x, y, z, num_size, num_rot, 7] anchors = anchors.permute(2, 1, 0, 3, 4, 5).contiguous() #anchors = anchors.view(-1, anchors.shape[-1]) anchors[..., 2] += anchors[..., 5] / 2 # shift to box centers all_anchors.append(anchors) return all_anchors, num_anchors_per_location if __name__ == '__main__': from easydict import EasyDict config = [ EasyDict({ 'anchor_sizes': [[2.1, 4.7, 1.7], [0.86, 0.91, 1.73], [0.84, 1.78, 1.78]], 'anchor_rotations': [0, 1.57], 'anchor_heights': [0, 0.5] }) ] A = AnchorGenerator( anchor_range=[-75.2, -75.2, -2, 75.2, 75.2, 4], anchor_generator_config=config ) import pdb pdb.set_trace() A.generate_anchors([[188, 188]])
3,990
48.8875
122
py
SASA
SASA-main/pcdet/models/dense_heads/target_assigner/axis_aligned_target_assigner.py
import numpy as np import torch from ....ops.iou3d_nms import iou3d_nms_utils from ....utils import box_utils class AxisAlignedTargetAssigner(object): def __init__(self, model_cfg, class_names, box_coder, match_height=False): super().__init__() anchor_generator_cfg = model_cfg.ANCHOR_GENERATOR_CONFIG anchor_target_cfg = model_cfg.TARGET_ASSIGNER_CONFIG self.box_coder = box_coder self.match_height = match_height self.class_names = np.array(class_names) self.anchor_class_names = [config['class_name'] for config in anchor_generator_cfg] self.pos_fraction = anchor_target_cfg.POS_FRACTION if anchor_target_cfg.POS_FRACTION >= 0 else None self.sample_size = anchor_target_cfg.SAMPLE_SIZE self.norm_by_num_examples = anchor_target_cfg.NORM_BY_NUM_EXAMPLES self.matched_thresholds = {} self.unmatched_thresholds = {} for config in anchor_generator_cfg: self.matched_thresholds[config['class_name']] = config['matched_threshold'] self.unmatched_thresholds[config['class_name']] = config['unmatched_threshold'] self.use_multihead = model_cfg.get('USE_MULTIHEAD', False) self.seperate_multihead = model_cfg.get('SEPARATE_MULTIHEAD', False) if self.seperate_multihead: rpn_head_cfgs = model_cfg.RPN_HEAD_CFGS self.gt_remapping = {} for rpn_head_cfg in rpn_head_cfgs: for idx, name in enumerate(rpn_head_cfg['HEAD_CLS_NAME']): self.gt_remapping[name] = idx + 1 def assign_targets(self, all_anchors, gt_boxes_with_classes): """ Args: all_anchors: [(N, 7), ...] gt_boxes: (B, M, 8) Returns: """ bbox_targets = [] cls_labels = [] reg_weights = [] batch_size = gt_boxes_with_classes.shape[0] gt_classes = gt_boxes_with_classes[:, :, -1] gt_boxes = gt_boxes_with_classes[:, :, :-1] for k in range(batch_size): cur_gt = gt_boxes[k] cnt = cur_gt.__len__() - 1 while cnt > 0 and cur_gt[cnt].sum() == 0: cnt -= 1 cur_gt = cur_gt[:cnt + 1] cur_gt_classes = gt_classes[k][:cnt + 1].int() target_list = [] for anchor_class_name, anchors in zip(self.anchor_class_names, all_anchors): if cur_gt_classes.shape[0] > 1: mask = torch.from_numpy(self.class_names[cur_gt_classes.cpu() - 1] == anchor_class_name) else: mask = torch.tensor([self.class_names[c - 1] == anchor_class_name for c in cur_gt_classes], dtype=torch.bool) if self.use_multihead: anchors = anchors.permute(3, 4, 0, 1, 2, 5).contiguous().view(-1, anchors.shape[-1]) if self.seperate_multihead: selected_classes = cur_gt_classes[mask].clone() if len(selected_classes) > 0: new_cls_id = self.gt_remapping[anchor_class_name] selected_classes[:] = new_cls_id else: selected_classes = cur_gt_classes[mask] else: feature_map_size = anchors.shape[:3] anchors = anchors.view(-1, anchors.shape[-1]) selected_classes = cur_gt_classes[mask] single_target = self.assign_targets_single( anchors, cur_gt[mask], gt_classes=selected_classes, matched_threshold=self.matched_thresholds[anchor_class_name], unmatched_threshold=self.unmatched_thresholds[anchor_class_name] ) target_list.append(single_target) if self.use_multihead: target_dict = { 'box_cls_labels': [t['box_cls_labels'].view(-1) for t in target_list], 'box_reg_targets': [t['box_reg_targets'].view(-1, self.box_coder.code_size) for t in target_list], 'reg_weights': [t['reg_weights'].view(-1) for t in target_list] } target_dict['box_reg_targets'] = torch.cat(target_dict['box_reg_targets'], dim=0) target_dict['box_cls_labels'] = torch.cat(target_dict['box_cls_labels'], dim=0).view(-1) target_dict['reg_weights'] = torch.cat(target_dict['reg_weights'], dim=0).view(-1) else: target_dict = { 'box_cls_labels': [t['box_cls_labels'].view(*feature_map_size, -1) for t in target_list], 'box_reg_targets': [t['box_reg_targets'].view(*feature_map_size, -1, self.box_coder.code_size) for t in target_list], 'reg_weights': [t['reg_weights'].view(*feature_map_size, -1) for t in target_list] } target_dict['box_reg_targets'] = torch.cat( target_dict['box_reg_targets'], dim=-2 ).view(-1, self.box_coder.code_size) target_dict['box_cls_labels'] = torch.cat(target_dict['box_cls_labels'], dim=-1).view(-1) target_dict['reg_weights'] = torch.cat(target_dict['reg_weights'], dim=-1).view(-1) bbox_targets.append(target_dict['box_reg_targets']) cls_labels.append(target_dict['box_cls_labels']) reg_weights.append(target_dict['reg_weights']) bbox_targets = torch.stack(bbox_targets, dim=0) cls_labels = torch.stack(cls_labels, dim=0) reg_weights = torch.stack(reg_weights, dim=0) all_targets_dict = { 'box_cls_labels': cls_labels, 'box_reg_targets': bbox_targets, 'reg_weights': reg_weights } return all_targets_dict def assign_targets_single(self, anchors, gt_boxes, gt_classes, matched_threshold=0.6, unmatched_threshold=0.45 ): num_anchors = anchors.shape[0] num_gt = gt_boxes.shape[0] labels = torch.ones((num_anchors,), dtype=torch.int32, device=anchors.device) * -1 gt_ids = torch.ones((num_anchors,), dtype=torch.int32, device=anchors.device) * -1 if len(gt_boxes) > 0 and anchors.shape[0] > 0: anchor_by_gt_overlap = iou3d_nms_utils.boxes_iou3d_gpu(anchors[:, 0:7], gt_boxes[:, 0:7]) \ if self.match_height else box_utils.boxes3d_nearest_bev_iou(anchors[:, 0:7], gt_boxes[:, 0:7]) anchor_to_gt_argmax = torch.from_numpy(anchor_by_gt_overlap.cpu().numpy().argmax(axis=1)).cuda() anchor_to_gt_max = anchor_by_gt_overlap[ torch.arange(num_anchors, device=anchors.device), anchor_to_gt_argmax ] gt_to_anchor_argmax = torch.from_numpy(anchor_by_gt_overlap.cpu().numpy().argmax(axis=0)).cuda() gt_to_anchor_max = anchor_by_gt_overlap[gt_to_anchor_argmax, torch.arange(num_gt, device=anchors.device)] empty_gt_mask = gt_to_anchor_max == 0 gt_to_anchor_max[empty_gt_mask] = -1 anchors_with_max_overlap = (anchor_by_gt_overlap == gt_to_anchor_max).nonzero()[:, 0] gt_inds_force = anchor_to_gt_argmax[anchors_with_max_overlap] labels[anchors_with_max_overlap] = gt_classes[gt_inds_force] gt_ids[anchors_with_max_overlap] = gt_inds_force.int() pos_inds = anchor_to_gt_max >= matched_threshold gt_inds_over_thresh = anchor_to_gt_argmax[pos_inds] labels[pos_inds] = gt_classes[gt_inds_over_thresh] gt_ids[pos_inds] = gt_inds_over_thresh.int() bg_inds = (anchor_to_gt_max < unmatched_threshold).nonzero()[:, 0] else: bg_inds = torch.arange(num_anchors, device=anchors.device) fg_inds = (labels > 0).nonzero()[:, 0] if self.pos_fraction is not None: num_fg = int(self.pos_fraction * self.sample_size) if len(fg_inds) > num_fg: num_disabled = len(fg_inds) - num_fg disable_inds = torch.randperm(len(fg_inds))[:num_disabled] labels[disable_inds] = -1 fg_inds = (labels > 0).nonzero()[:, 0] num_bg = self.sample_size - (labels > 0).sum() if len(bg_inds) > num_bg: enable_inds = bg_inds[torch.randint(0, len(bg_inds), size=(num_bg,))] labels[enable_inds] = 0 # bg_inds = torch.nonzero(labels == 0)[:, 0] else: if len(gt_boxes) == 0 or anchors.shape[0] == 0: labels[:] = 0 else: labels[bg_inds] = 0 labels[anchors_with_max_overlap] = gt_classes[gt_inds_force] bbox_targets = anchors.new_zeros((num_anchors, self.box_coder.code_size)) if len(gt_boxes) > 0 and anchors.shape[0] > 0: fg_gt_boxes = gt_boxes[anchor_to_gt_argmax[fg_inds], :] fg_anchors = anchors[fg_inds, :] bbox_targets[fg_inds, :] = self.box_coder.encode_torch(fg_gt_boxes, fg_anchors) reg_weights = anchors.new_zeros((num_anchors,)) if self.norm_by_num_examples: num_examples = (labels >= 0).sum() num_examples = num_examples if num_examples > 1.0 else 1.0 reg_weights[labels > 0] = 1.0 / num_examples else: reg_weights[labels > 0] = 1.0 ret_dict = { 'box_cls_labels': labels, 'box_reg_targets': bbox_targets, 'reg_weights': reg_weights, } return ret_dict
9,874
45.14486
118
py
SASA
SASA-main/pcdet/models/dense_heads/target_assigner/atss_target_assigner.py
import torch from ....ops.iou3d_nms import iou3d_nms_utils from ....utils import common_utils class ATSSTargetAssigner(object): """ Reference: https://arxiv.org/abs/1912.02424 """ def __init__(self, topk, box_coder, match_height=False): self.topk = topk self.box_coder = box_coder self.match_height = match_height def assign_targets(self, anchors_list, gt_boxes_with_classes, use_multihead=False): """ Args: anchors: [(N, 7), ...] gt_boxes: (B, M, 8) Returns: """ if not isinstance(anchors_list, list): anchors_list = [anchors_list] single_set_of_anchor = True else: single_set_of_anchor = len(anchors_list) == 1 cls_labels_list, reg_targets_list, reg_weights_list = [], [], [] for anchors in anchors_list: batch_size = gt_boxes_with_classes.shape[0] gt_classes = gt_boxes_with_classes[:, :, -1] gt_boxes = gt_boxes_with_classes[:, :, :-1] if use_multihead: anchors = anchors.permute(3, 4, 0, 1, 2, 5).contiguous().view(-1, anchors.shape[-1]) else: anchors = anchors.view(-1, anchors.shape[-1]) cls_labels, reg_targets, reg_weights = [], [], [] for k in range(batch_size): cur_gt = gt_boxes[k] cnt = cur_gt.__len__() - 1 while cnt > 0 and cur_gt[cnt].sum() == 0: cnt -= 1 cur_gt = cur_gt[:cnt + 1] cur_gt_classes = gt_classes[k][:cnt + 1] cur_cls_labels, cur_reg_targets, cur_reg_weights = self.assign_targets_single( anchors, cur_gt, cur_gt_classes ) cls_labels.append(cur_cls_labels) reg_targets.append(cur_reg_targets) reg_weights.append(cur_reg_weights) cls_labels = torch.stack(cls_labels, dim=0) reg_targets = torch.stack(reg_targets, dim=0) reg_weights = torch.stack(reg_weights, dim=0) cls_labels_list.append(cls_labels) reg_targets_list.append(reg_targets) reg_weights_list.append(reg_weights) if single_set_of_anchor: ret_dict = { 'box_cls_labels': cls_labels_list[0], 'box_reg_targets': reg_targets_list[0], 'reg_weights': reg_weights_list[0] } else: ret_dict = { 'box_cls_labels': torch.cat(cls_labels_list, dim=1), 'box_reg_targets': torch.cat(reg_targets_list, dim=1), 'reg_weights': torch.cat(reg_weights_list, dim=1) } return ret_dict def assign_targets_single(self, anchors, gt_boxes, gt_classes): """ Args: anchors: (N, 7) [x, y, z, dx, dy, dz, heading] gt_boxes: (M, 7) [x, y, z, dx, dy, dz, heading] gt_classes: (M) Returns: """ num_anchor = anchors.shape[0] num_gt = gt_boxes.shape[0] # select topk anchors for each gt_boxes if self.match_height: ious = iou3d_nms_utils.boxes_iou3d_gpu(anchors[:, 0:7], gt_boxes[:, 0:7]) # (N, M) else: ious = iou3d_nms_utils.boxes_iou_bev(anchors[:, 0:7], gt_boxes[:, 0:7]) distance = (anchors[:, None, 0:3] - gt_boxes[None, :, 0:3]).norm(dim=-1) # (N, M) _, topk_idxs = distance.topk(self.topk, dim=0, largest=False) # (K, M) candidate_ious = ious[topk_idxs, torch.arange(num_gt)] # (K, M) iou_mean_per_gt = candidate_ious.mean(dim=0) iou_std_per_gt = candidate_ious.std(dim=0) iou_thresh_per_gt = iou_mean_per_gt + iou_std_per_gt + 1e-6 is_pos = candidate_ious >= iou_thresh_per_gt[None, :] # (K, M) # check whether anchor_center in gt_boxes, only check BEV x-y axes candidate_anchors = anchors[topk_idxs.view(-1)] # (KxM, 7) gt_boxes_of_each_anchor = gt_boxes[:, :].repeat(self.topk, 1) # (KxM, 7) xyz_local = candidate_anchors[:, 0:3] - gt_boxes_of_each_anchor[:, 0:3] xyz_local = common_utils.rotate_points_along_z( xyz_local[:, None, :], -gt_boxes_of_each_anchor[:, 6] ).squeeze(dim=1) xy_local = xyz_local[:, 0:2] lw = gt_boxes_of_each_anchor[:, 3:5][:, [1, 0]] # bugfixed: w ==> y, l ==> x in local coords is_in_gt = ((xy_local <= lw / 2) & (xy_local >= -lw / 2)).all(dim=-1).view(-1, num_gt) # (K, M) is_pos = is_pos & is_in_gt # (K, M) for ng in range(num_gt): topk_idxs[:, ng] += ng * num_anchor # select the highest IoU if an anchor box is assigned with multiple gt_boxes INF = -0x7FFFFFFF ious_inf = torch.full_like(ious, INF).t().contiguous().view(-1) # (MxN) index = topk_idxs.view(-1)[is_pos.view(-1)] ious_inf[index] = ious.t().contiguous().view(-1)[index] ious_inf = ious_inf.view(num_gt, -1).t() # (N, M) anchors_to_gt_values, anchors_to_gt_indexs = ious_inf.max(dim=1) # match the gt_boxes to the anchors which have maximum iou with them max_iou_of_each_gt, argmax_iou_of_each_gt = ious.max(dim=0) anchors_to_gt_indexs[argmax_iou_of_each_gt] = torch.arange(0, num_gt, device=ious.device) anchors_to_gt_values[argmax_iou_of_each_gt] = max_iou_of_each_gt cls_labels = gt_classes[anchors_to_gt_indexs] cls_labels[anchors_to_gt_values == INF] = 0 matched_gts = gt_boxes[anchors_to_gt_indexs] pos_mask = cls_labels > 0 reg_targets = matched_gts.new_zeros((num_anchor, self.box_coder.code_size)) reg_weights = matched_gts.new_zeros(num_anchor) if pos_mask.sum() > 0: reg_targets[pos_mask > 0] = self.box_coder.encode_torch(matched_gts[pos_mask > 0], anchors[pos_mask > 0]) reg_weights[pos_mask] = 1.0 return cls_labels, reg_targets, reg_weights
6,050
41.612676
117
py
SASA
SASA-main/pcdet/models/roi_heads/roi_head_template.py
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from ...utils import box_coder_utils, common_utils, loss_utils from ..model_utils.model_nms_utils import class_agnostic_nms from .target_assigner.proposal_target_layer import ProposalTargetLayer class RoIHeadTemplate(nn.Module): def __init__(self, num_class, model_cfg): super().__init__() self.model_cfg = model_cfg self.num_class = num_class self.box_coder = getattr(box_coder_utils, self.model_cfg.TARGET_CONFIG.BOX_CODER)( **self.model_cfg.TARGET_CONFIG.get('BOX_CODER_CONFIG', {}) ) self.proposal_target_layer = ProposalTargetLayer(roi_sampler_cfg=self.model_cfg.TARGET_CONFIG) self.build_losses(self.model_cfg.LOSS_CONFIG) self.forward_ret_dict = None def build_losses(self, losses_cfg): self.add_module( 'reg_loss_func', loss_utils.WeightedSmoothL1Loss(code_weights=losses_cfg.LOSS_WEIGHTS['code_weights']) ) def make_fc_layers(self, input_channels, output_channels, fc_list): fc_layers = [] pre_channel = input_channels for k in range(0, fc_list.__len__()): fc_layers.extend([ nn.Conv1d(pre_channel, fc_list[k], kernel_size=1, bias=False), nn.BatchNorm1d(fc_list[k]), nn.ReLU() ]) pre_channel = fc_list[k] if self.model_cfg.DP_RATIO >= 0 and k == 0: fc_layers.append(nn.Dropout(self.model_cfg.DP_RATIO)) fc_layers.append(nn.Conv1d(pre_channel, output_channels, kernel_size=1, bias=True)) fc_layers = nn.Sequential(*fc_layers) return fc_layers @torch.no_grad() def proposal_layer(self, batch_dict, nms_config): """ Args: batch_dict: batch_size: batch_cls_preds: (B, num_boxes, num_classes | 1) or (N1+N2+..., num_classes | 1) batch_box_preds: (B, num_boxes, 7+C) or (N1+N2+..., 7+C) cls_preds_normalized: indicate whether batch_cls_preds is normalized batch_index: optional (N1+N2+...) nms_config: Returns: batch_dict: rois: (B, num_rois, 7+C) roi_scores: (B, num_rois) roi_labels: (B, num_rois) """ batch_size = batch_dict['batch_size'] batch_box_preds = batch_dict['batch_box_preds'] batch_cls_preds = batch_dict['batch_cls_preds'] rois = batch_box_preds.new_zeros((batch_size, nms_config.NMS_POST_MAXSIZE, batch_box_preds.shape[-1])) roi_scores = batch_box_preds.new_zeros((batch_size, nms_config.NMS_POST_MAXSIZE)) roi_labels = batch_box_preds.new_zeros((batch_size, nms_config.NMS_POST_MAXSIZE), dtype=torch.long) for index in range(batch_size): if batch_dict.get('batch_index', None) is not None: assert batch_cls_preds.shape.__len__() == 2 batch_mask = (batch_dict['batch_index'] == index) else: assert batch_dict['batch_cls_preds'].shape.__len__() == 3 batch_mask = index box_preds = batch_box_preds[batch_mask] cls_preds = batch_cls_preds[batch_mask] cur_roi_scores, cur_roi_labels = torch.max(cls_preds, dim=1) if nms_config.MULTI_CLASSES_NMS: raise NotImplementedError else: selected, selected_scores = class_agnostic_nms( box_scores=cur_roi_scores, box_preds=box_preds, nms_config=nms_config ) rois[index, :len(selected), :] = box_preds[selected] roi_scores[index, :len(selected)] = cur_roi_scores[selected] roi_labels[index, :len(selected)] = cur_roi_labels[selected] batch_dict['rois'] = rois batch_dict['roi_scores'] = roi_scores batch_dict['roi_labels'] = roi_labels + 1 batch_dict['has_class_labels'] = True if batch_cls_preds.shape[-1] > 1 else False batch_dict.pop('batch_index', None) return batch_dict def assign_targets(self, batch_dict): batch_size = batch_dict['batch_size'] with torch.no_grad(): targets_dict = self.proposal_target_layer.forward(batch_dict) rois = targets_dict['rois'] # (B, N, 7 + C) gt_of_rois = targets_dict['gt_of_rois'] # (B, N, 7 + C + 1) targets_dict['gt_of_rois_src'] = gt_of_rois.clone().detach() # canonical transformation roi_center = rois[:, :, 0:3] roi_ry = rois[:, :, 6] % (2 * np.pi) gt_of_rois[:, :, 0:3] = gt_of_rois[:, :, 0:3] - roi_center gt_of_rois[:, :, 6] = gt_of_rois[:, :, 6] - roi_ry # transfer LiDAR coords to local coords gt_of_rois = common_utils.rotate_points_along_z( points=gt_of_rois.view(-1, 1, gt_of_rois.shape[-1]), angle=-roi_ry.view(-1) ).view(batch_size, -1, gt_of_rois.shape[-1]) # flip orientation if rois have opposite orientation heading_label = gt_of_rois[:, :, 6] % (2 * np.pi) # 0 ~ 2pi opposite_flag = (heading_label > np.pi * 0.5) & (heading_label < np.pi * 1.5) heading_label[opposite_flag] = (heading_label[opposite_flag] + np.pi) % (2 * np.pi) # (0 ~ pi/2, 3pi/2 ~ 2pi) flag = heading_label > np.pi heading_label[flag] = heading_label[flag] - np.pi * 2 # (-pi/2, pi/2) heading_label = torch.clamp(heading_label, min=-np.pi / 2, max=np.pi / 2) gt_of_rois[:, :, 6] = heading_label targets_dict['gt_of_rois'] = gt_of_rois return targets_dict def get_box_reg_layer_loss(self, forward_ret_dict): loss_cfgs = self.model_cfg.LOSS_CONFIG code_size = self.box_coder.code_size reg_valid_mask = forward_ret_dict['reg_valid_mask'].view(-1) gt_boxes3d_ct = forward_ret_dict['gt_of_rois'][..., 0:code_size] gt_of_rois_src = forward_ret_dict['gt_of_rois_src'][..., 0:code_size].view(-1, code_size) rcnn_reg = forward_ret_dict['rcnn_reg'] # (rcnn_batch_size, C) roi_boxes3d = forward_ret_dict['rois'] rcnn_batch_size = gt_boxes3d_ct.view(-1, code_size).shape[0] fg_mask = (reg_valid_mask > 0) fg_sum = fg_mask.long().sum().item() tb_dict = {} if loss_cfgs.REG_LOSS == 'smooth-l1': rois_anchor = roi_boxes3d.clone().detach().view(-1, code_size) rois_anchor[:, 0:3] = 0 rois_anchor[:, 6] = 0 reg_targets = self.box_coder.encode_torch( gt_boxes3d_ct.view(rcnn_batch_size, code_size), rois_anchor ) rcnn_loss_reg = self.reg_loss_func( rcnn_reg.view(rcnn_batch_size, -1).unsqueeze(dim=0), reg_targets.unsqueeze(dim=0), ) # [B, M, 7] rcnn_loss_reg = (rcnn_loss_reg.view(rcnn_batch_size, -1) * fg_mask.unsqueeze(dim=-1).float()).sum() / max(fg_sum, 1) rcnn_loss_reg = rcnn_loss_reg * loss_cfgs.LOSS_WEIGHTS['rcnn_reg_weight'] tb_dict['rcnn_loss_reg'] = rcnn_loss_reg.item() if loss_cfgs.CORNER_LOSS_REGULARIZATION and fg_sum > 0: # TODO: NEED to BE CHECK fg_rcnn_reg = rcnn_reg.view(rcnn_batch_size, -1)[fg_mask] fg_roi_boxes3d = roi_boxes3d.view(-1, code_size)[fg_mask] fg_roi_boxes3d = fg_roi_boxes3d.view(1, -1, code_size) batch_anchors = fg_roi_boxes3d.clone().detach() roi_ry = fg_roi_boxes3d[:, :, 6].view(-1) roi_xyz = fg_roi_boxes3d[:, :, 0:3].view(-1, 3) batch_anchors[:, :, 0:3] = 0 rcnn_boxes3d = self.box_coder.decode_torch( fg_rcnn_reg.view(batch_anchors.shape[0], -1, code_size), batch_anchors ).view(-1, code_size) rcnn_boxes3d = common_utils.rotate_points_along_z( rcnn_boxes3d.unsqueeze(dim=1), roi_ry ).squeeze(dim=1) rcnn_boxes3d[:, 0:3] += roi_xyz loss_corner = loss_utils.get_corner_loss_lidar( rcnn_boxes3d[:, 0:7], gt_of_rois_src[fg_mask][:, 0:7] ) loss_corner = loss_corner.mean() loss_corner = loss_corner * loss_cfgs.LOSS_WEIGHTS['rcnn_corner_weight'] rcnn_loss_reg += loss_corner tb_dict['rcnn_loss_corner'] = loss_corner.item() else: raise NotImplementedError return rcnn_loss_reg, tb_dict def get_box_cls_layer_loss(self, forward_ret_dict): loss_cfgs = self.model_cfg.LOSS_CONFIG rcnn_cls = forward_ret_dict['rcnn_cls'] rcnn_cls_labels = forward_ret_dict['rcnn_cls_labels'].view(-1) if loss_cfgs.CLS_LOSS == 'BinaryCrossEntropy': rcnn_cls_flat = rcnn_cls.view(-1) batch_loss_cls = F.binary_cross_entropy(torch.sigmoid(rcnn_cls_flat), rcnn_cls_labels.float(), reduction='none') cls_valid_mask = (rcnn_cls_labels >= 0).float() rcnn_loss_cls = (batch_loss_cls * cls_valid_mask).sum() / torch.clamp(cls_valid_mask.sum(), min=1.0) elif loss_cfgs.CLS_LOSS == 'CrossEntropy': batch_loss_cls = F.cross_entropy(rcnn_cls, rcnn_cls_labels, reduction='none', ignore_index=-1) cls_valid_mask = (rcnn_cls_labels >= 0).float() rcnn_loss_cls = (batch_loss_cls * cls_valid_mask).sum() / torch.clamp(cls_valid_mask.sum(), min=1.0) else: raise NotImplementedError rcnn_loss_cls = rcnn_loss_cls * loss_cfgs.LOSS_WEIGHTS['rcnn_cls_weight'] tb_dict = {'rcnn_loss_cls': rcnn_loss_cls.item()} return rcnn_loss_cls, tb_dict def get_loss(self, tb_dict=None): tb_dict = {} if tb_dict is None else tb_dict rcnn_loss = 0 rcnn_loss_cls, cls_tb_dict = self.get_box_cls_layer_loss(self.forward_ret_dict) rcnn_loss += rcnn_loss_cls tb_dict.update(cls_tb_dict) rcnn_loss_reg, reg_tb_dict = self.get_box_reg_layer_loss(self.forward_ret_dict) rcnn_loss += rcnn_loss_reg tb_dict.update(reg_tb_dict) tb_dict['rcnn_loss'] = rcnn_loss.item() return rcnn_loss, tb_dict def generate_predicted_boxes(self, batch_size, rois, cls_preds, box_preds): """ Args: batch_size: rois: (B, N, 7) cls_preds: (BN, num_class) box_preds: (BN, code_size) Returns: """ code_size = self.box_coder.code_size # batch_cls_preds: (B, N, num_class or 1) batch_cls_preds = cls_preds.view(batch_size, -1, cls_preds.shape[-1]) batch_box_preds = box_preds.view(batch_size, -1, code_size) roi_ry = rois[:, :, 6].view(-1) roi_xyz = rois[:, :, 0:3].view(-1, 3) local_rois = rois.clone().detach() local_rois[:, :, 0:3] = 0 batch_box_preds = self.box_coder.decode_torch(batch_box_preds, local_rois).view(-1, code_size) batch_box_preds = common_utils.rotate_points_along_z( batch_box_preds.unsqueeze(dim=1), roi_ry ).squeeze(dim=1) batch_box_preds[:, 0:3] += roi_xyz batch_box_preds = batch_box_preds.view(batch_size, -1, code_size) return batch_cls_preds, batch_box_preds
11,451
43.216216
128
py
SASA
SASA-main/pcdet/models/roi_heads/partA2_head.py
import numpy as np import spconv import torch import torch.nn as nn from ...ops.roiaware_pool3d import roiaware_pool3d_utils from .roi_head_template import RoIHeadTemplate class PartA2FCHead(RoIHeadTemplate): def __init__(self, input_channels, model_cfg, num_class=1): super().__init__(num_class=num_class, model_cfg=model_cfg) self.model_cfg = model_cfg self.SA_modules = nn.ModuleList() block = self.post_act_block c0 = self.model_cfg.ROI_AWARE_POOL.NUM_FEATURES // 2 self.conv_part = spconv.SparseSequential( block(4, 64, 3, padding=1, indice_key='rcnn_subm1'), block(64, c0, 3, padding=1, indice_key='rcnn_subm1_1'), ) self.conv_rpn = spconv.SparseSequential( block(input_channels, 64, 3, padding=1, indice_key='rcnn_subm2'), block(64, c0, 3, padding=1, indice_key='rcnn_subm1_2'), ) shared_fc_list = [] pool_size = self.model_cfg.ROI_AWARE_POOL.POOL_SIZE pre_channel = self.model_cfg.ROI_AWARE_POOL.NUM_FEATURES * pool_size * pool_size * pool_size for k in range(0, self.model_cfg.SHARED_FC.__len__()): shared_fc_list.extend([ nn.Conv1d(pre_channel, self.model_cfg.SHARED_FC[k], kernel_size=1, bias=False), nn.BatchNorm1d(self.model_cfg.SHARED_FC[k]), nn.ReLU() ]) pre_channel = self.model_cfg.SHARED_FC[k] if k != self.model_cfg.SHARED_FC.__len__() - 1 and self.model_cfg.DP_RATIO > 0: shared_fc_list.append(nn.Dropout(self.model_cfg.DP_RATIO)) self.shared_fc_layer = nn.Sequential(*shared_fc_list) self.cls_layers = self.make_fc_layers( input_channels=pre_channel, output_channels=self.num_class, fc_list=self.model_cfg.CLS_FC ) self.reg_layers = self.make_fc_layers( input_channels=pre_channel, output_channels=self.box_coder.code_size * self.num_class, fc_list=self.model_cfg.REG_FC ) self.roiaware_pool3d_layer = roiaware_pool3d_utils.RoIAwarePool3d( out_size=self.model_cfg.ROI_AWARE_POOL.POOL_SIZE, max_pts_each_voxel=self.model_cfg.ROI_AWARE_POOL.MAX_POINTS_PER_VOXEL ) self.init_weights(weight_init='xavier') def init_weights(self, weight_init='xavier'): if weight_init == 'kaiming': init_func = nn.init.kaiming_normal_ elif weight_init == 'xavier': init_func = nn.init.xavier_normal_ elif weight_init == 'normal': init_func = nn.init.normal_ else: raise NotImplementedError for m in self.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Conv1d): if weight_init == 'normal': init_func(m.weight, mean=0, std=0.001) else: init_func(m.weight) if m.bias is not None: nn.init.constant_(m.bias, 0) nn.init.normal_(self.reg_layers[-1].weight, mean=0, std=0.001) def post_act_block(self, in_channels, out_channels, kernel_size, indice_key, stride=1, padding=0, conv_type='subm'): if conv_type == 'subm': m = spconv.SparseSequential( spconv.SubMConv3d(in_channels, out_channels, kernel_size, bias=False, indice_key=indice_key), nn.BatchNorm1d(out_channels, eps=1e-3, momentum=0.01), nn.ReLU(), ) elif conv_type == 'spconv': m = spconv.SparseSequential( spconv.SparseConv3d(in_channels, out_channels, kernel_size, stride=stride, padding=padding, bias=False, indice_key=indice_key), nn.BatchNorm1d(out_channels, eps=1e-3, momentum=0.01), nn.ReLU(), ) elif conv_type == 'inverseconv': m = spconv.SparseSequential( spconv.SparseInverseConv3d(in_channels, out_channels, kernel_size, indice_key=indice_key, bias=False), nn.BatchNorm1d(out_channels, eps=1e-3, momentum=0.01), nn.ReLU(), ) else: raise NotImplementedError return m def roiaware_pool(self, batch_dict): """ Args: batch_dict: batch_size: rois: (B, num_rois, 7 + C) point_coords: (num_points, 4) [bs_idx, x, y, z] point_features: (num_points, C) point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) Returns: """ batch_size = batch_dict['batch_size'] batch_idx = batch_dict['point_coords'][:, 0] point_coords = batch_dict['point_coords'][:, 1:4] point_features = batch_dict['point_features'] part_features = torch.cat(( batch_dict['point_part_offset'] if not self.model_cfg.get('DISABLE_PART', False) else point_coords, batch_dict['point_cls_scores'].view(-1, 1).detach() ), dim=1) part_features[part_features[:, -1] < self.model_cfg.SEG_MASK_SCORE_THRESH, 0:3] = 0 rois = batch_dict['rois'] pooled_part_features_list, pooled_rpn_features_list = [], [] for bs_idx in range(batch_size): bs_mask = (batch_idx == bs_idx) cur_point_coords = point_coords[bs_mask] cur_part_features = part_features[bs_mask] cur_rpn_features = point_features[bs_mask] cur_roi = rois[bs_idx][:, 0:7].contiguous() # (N, 7) pooled_part_features = self.roiaware_pool3d_layer.forward( cur_roi, cur_point_coords, cur_part_features, pool_method='avg' ) # (N, out_x, out_y, out_z, 4) pooled_rpn_features = self.roiaware_pool3d_layer.forward( cur_roi, cur_point_coords, cur_rpn_features, pool_method='max' ) # (N, out_x, out_y, out_z, C) pooled_part_features_list.append(pooled_part_features) pooled_rpn_features_list.append(pooled_rpn_features) pooled_part_features = torch.cat(pooled_part_features_list, dim=0) # (B * N, out_x, out_y, out_z, 4) pooled_rpn_features = torch.cat(pooled_rpn_features_list, dim=0) # (B * N, out_x, out_y, out_z, C) return pooled_part_features, pooled_rpn_features @staticmethod def fake_sparse_idx(sparse_idx, batch_size_rcnn): print('Warning: Sparse_Idx_Shape(%s) \r' % (str(sparse_idx.shape)), end='', flush=True) # at most one sample is non-empty, then fake the first voxels of each sample(BN needs at least # two values each channel) as non-empty for the below calculation sparse_idx = sparse_idx.new_zeros((batch_size_rcnn, 3)) bs_idxs = torch.arange(batch_size_rcnn).type_as(sparse_idx).view(-1, 1) sparse_idx = torch.cat((bs_idxs, sparse_idx), dim=1) return sparse_idx def forward(self, batch_dict): """ Args: batch_dict: Returns: """ targets_dict = self.proposal_layer( batch_dict, nms_config=self.model_cfg.NMS_CONFIG['TRAIN' if self.training else 'TEST'] ) if self.training: targets_dict = self.assign_targets(batch_dict) batch_dict['rois'] = targets_dict['rois'] batch_dict['roi_labels'] = targets_dict['roi_labels'] # RoI aware pooling pooled_part_features, pooled_rpn_features = self.roiaware_pool(batch_dict) batch_size_rcnn = pooled_part_features.shape[0] # (B * N, out_x, out_y, out_z, 4) # transform to sparse tensors sparse_shape = np.array(pooled_part_features.shape[1:4], dtype=np.int32) sparse_idx = pooled_part_features.sum(dim=-1).nonzero() # (non_empty_num, 4) ==> [bs_idx, x_idx, y_idx, z_idx] if sparse_idx.shape[0] < 3: sparse_idx = self.fake_sparse_idx(sparse_idx, batch_size_rcnn) if self.training: # these are invalid samples targets_dict['rcnn_cls_labels'].fill_(-1) targets_dict['reg_valid_mask'].fill_(-1) part_features = pooled_part_features[sparse_idx[:, 0], sparse_idx[:, 1], sparse_idx[:, 2], sparse_idx[:, 3]] rpn_features = pooled_rpn_features[sparse_idx[:, 0], sparse_idx[:, 1], sparse_idx[:, 2], sparse_idx[:, 3]] coords = sparse_idx.int() part_features = spconv.SparseConvTensor(part_features, coords, sparse_shape, batch_size_rcnn) rpn_features = spconv.SparseConvTensor(rpn_features, coords, sparse_shape, batch_size_rcnn) # forward rcnn network x_part = self.conv_part(part_features) x_rpn = self.conv_rpn(rpn_features) merged_feature = torch.cat((x_rpn.features, x_part.features), dim=1) # (N, C) shared_feature = spconv.SparseConvTensor(merged_feature, coords, sparse_shape, batch_size_rcnn) shared_feature = shared_feature.dense().view(batch_size_rcnn, -1, 1) shared_feature = self.shared_fc_layer(shared_feature) rcnn_cls = self.cls_layers(shared_feature).transpose(1, 2).contiguous().squeeze(dim=1) # (B, 1 or 2) rcnn_reg = self.reg_layers(shared_feature).transpose(1, 2).contiguous().squeeze(dim=1) # (B, C) if not self.training: batch_cls_preds, batch_box_preds = self.generate_predicted_boxes( batch_size=batch_dict['batch_size'], rois=batch_dict['rois'], cls_preds=rcnn_cls, box_preds=rcnn_reg ) batch_dict['batch_cls_preds'] = batch_cls_preds batch_dict['batch_box_preds'] = batch_box_preds batch_dict['cls_preds_normalized'] = False else: targets_dict['rcnn_cls'] = rcnn_cls targets_dict['rcnn_reg'] = rcnn_reg self.forward_ret_dict = targets_dict return batch_dict
10,039
43.622222
120
py
SASA
SASA-main/pcdet/models/roi_heads/pvrcnn_head.py
import torch.nn as nn from ...ops.pointnet2.pointnet2_stack import pointnet2_modules as pointnet2_stack_modules from ...utils import common_utils from .roi_head_template import RoIHeadTemplate class PVRCNNHead(RoIHeadTemplate): def __init__(self, input_channels, model_cfg, num_class=1): super().__init__(num_class=num_class, model_cfg=model_cfg) self.model_cfg = model_cfg mlps = self.model_cfg.ROI_GRID_POOL.MLPS for k in range(len(mlps)): mlps[k] = [input_channels] + mlps[k] self.roi_grid_pool_layer = pointnet2_stack_modules.StackSAModuleMSG( radii=self.model_cfg.ROI_GRID_POOL.POOL_RADIUS, nsamples=self.model_cfg.ROI_GRID_POOL.NSAMPLE, mlps=mlps, use_xyz=True, pool_method=self.model_cfg.ROI_GRID_POOL.POOL_METHOD, ) GRID_SIZE = self.model_cfg.ROI_GRID_POOL.GRID_SIZE c_out = sum([x[-1] for x in mlps]) pre_channel = GRID_SIZE * GRID_SIZE * GRID_SIZE * c_out shared_fc_list = [] for k in range(0, self.model_cfg.SHARED_FC.__len__()): shared_fc_list.extend([ nn.Conv1d(pre_channel, self.model_cfg.SHARED_FC[k], kernel_size=1, bias=False), nn.BatchNorm1d(self.model_cfg.SHARED_FC[k]), nn.ReLU() ]) pre_channel = self.model_cfg.SHARED_FC[k] if k != self.model_cfg.SHARED_FC.__len__() - 1 and self.model_cfg.DP_RATIO > 0: shared_fc_list.append(nn.Dropout(self.model_cfg.DP_RATIO)) self.shared_fc_layer = nn.Sequential(*shared_fc_list) self.cls_layers = self.make_fc_layers( input_channels=pre_channel, output_channels=self.num_class, fc_list=self.model_cfg.CLS_FC ) self.reg_layers = self.make_fc_layers( input_channels=pre_channel, output_channels=self.box_coder.code_size * self.num_class, fc_list=self.model_cfg.REG_FC ) self.init_weights(weight_init='xavier') def init_weights(self, weight_init='xavier'): if weight_init == 'kaiming': init_func = nn.init.kaiming_normal_ elif weight_init == 'xavier': init_func = nn.init.xavier_normal_ elif weight_init == 'normal': init_func = nn.init.normal_ else: raise NotImplementedError for m in self.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Conv1d): if weight_init == 'normal': init_func(m.weight, mean=0, std=0.001) else: init_func(m.weight) if m.bias is not None: nn.init.constant_(m.bias, 0) nn.init.normal_(self.reg_layers[-1].weight, mean=0, std=0.001) def roi_grid_pool(self, batch_dict): """ Args: batch_dict: batch_size: rois: (B, num_rois, 7 + C) point_coords: (num_points, 4) [bs_idx, x, y, z] point_features: (num_points, C) point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) Returns: """ batch_size = batch_dict['batch_size'] rois = batch_dict['rois'] point_coords = batch_dict['point_coords'] point_features = batch_dict['point_features'] point_features = point_features * batch_dict['point_cls_scores'].view(-1, 1) global_roi_grid_points, local_roi_grid_points = self.get_global_grid_points_of_roi( rois, grid_size=self.model_cfg.ROI_GRID_POOL.GRID_SIZE ) # (BxN, 6x6x6, 3) global_roi_grid_points = global_roi_grid_points.view(batch_size, -1, 3) # (B, Nx6x6x6, 3) xyz = point_coords[:, 1:4] xyz_batch_cnt = xyz.new_zeros(batch_size).int() batch_idx = point_coords[:, 0] for k in range(batch_size): xyz_batch_cnt[k] = (batch_idx == k).sum() new_xyz = global_roi_grid_points.view(-1, 3) new_xyz_batch_cnt = xyz.new_zeros(batch_size).int().fill_(global_roi_grid_points.shape[1]) pooled_points, pooled_features = self.roi_grid_pool_layer( xyz=xyz.contiguous(), xyz_batch_cnt=xyz_batch_cnt, new_xyz=new_xyz, new_xyz_batch_cnt=new_xyz_batch_cnt, features=point_features.contiguous(), ) # (M1 + M2 ..., C) pooled_features = pooled_features.view( -1, self.model_cfg.ROI_GRID_POOL.GRID_SIZE ** 3, pooled_features.shape[-1] ) # (BxN, 6x6x6, C) return pooled_features def get_global_grid_points_of_roi(self, rois, grid_size): rois = rois.view(-1, rois.shape[-1]) batch_size_rcnn = rois.shape[0] local_roi_grid_points = self.get_dense_grid_points(rois, batch_size_rcnn, grid_size) # (B, 6x6x6, 3) global_roi_grid_points = common_utils.rotate_points_along_z( local_roi_grid_points.clone(), rois[:, 6] ).squeeze(dim=1) global_center = rois[:, 0:3].clone() global_roi_grid_points += global_center.unsqueeze(dim=1) return global_roi_grid_points, local_roi_grid_points @staticmethod def get_dense_grid_points(rois, batch_size_rcnn, grid_size): faked_features = rois.new_ones((grid_size, grid_size, grid_size)) dense_idx = faked_features.nonzero() # (N, 3) [x_idx, y_idx, z_idx] dense_idx = dense_idx.repeat(batch_size_rcnn, 1, 1).float() # (B, 6x6x6, 3) local_roi_size = rois.view(batch_size_rcnn, -1)[:, 3:6] roi_grid_points = (dense_idx + 0.5) / grid_size * local_roi_size.unsqueeze(dim=1) \ - (local_roi_size.unsqueeze(dim=1) / 2) # (B, 6x6x6, 3) return roi_grid_points def forward(self, batch_dict): """ :param input_data: input dict :return: """ targets_dict = self.proposal_layer( batch_dict, nms_config=self.model_cfg.NMS_CONFIG['TRAIN' if self.training else 'TEST'] ) if self.training: targets_dict = self.assign_targets(batch_dict) batch_dict['rois'] = targets_dict['rois'] batch_dict['roi_labels'] = targets_dict['roi_labels'] # RoI aware pooling pooled_features = self.roi_grid_pool(batch_dict) # (BxN, 6x6x6, C) grid_size = self.model_cfg.ROI_GRID_POOL.GRID_SIZE batch_size_rcnn = pooled_features.shape[0] pooled_features = pooled_features.permute(0, 2, 1).\ contiguous().view(batch_size_rcnn, -1, grid_size, grid_size, grid_size) # (BxN, C, 6, 6, 6) shared_features = self.shared_fc_layer(pooled_features.view(batch_size_rcnn, -1, 1)) rcnn_cls = self.cls_layers(shared_features).transpose(1, 2).contiguous().squeeze(dim=1) # (B, 1 or 2) rcnn_reg = self.reg_layers(shared_features).transpose(1, 2).contiguous().squeeze(dim=1) # (B, C) if not self.training: batch_cls_preds, batch_box_preds = self.generate_predicted_boxes( batch_size=batch_dict['batch_size'], rois=batch_dict['rois'], cls_preds=rcnn_cls, box_preds=rcnn_reg ) batch_dict['batch_cls_preds'] = batch_cls_preds batch_dict['batch_box_preds'] = batch_box_preds batch_dict['cls_preds_normalized'] = False else: targets_dict['rcnn_cls'] = rcnn_cls targets_dict['rcnn_reg'] = rcnn_reg self.forward_ret_dict = targets_dict return batch_dict
7,628
40.688525
116
py
SASA
SASA-main/pcdet/models/roi_heads/pointrcnn_head.py
import torch import torch.nn as nn from ...ops.pointnet2.pointnet2_batch import pointnet2_modules from ...ops.roipoint_pool3d import roipoint_pool3d_utils from ...utils import common_utils from .roi_head_template import RoIHeadTemplate class PointRCNNHead(RoIHeadTemplate): def __init__(self, input_channels, model_cfg, num_class=1): super().__init__(num_class=num_class, model_cfg=model_cfg) self.model_cfg = model_cfg use_bn = self.model_cfg.USE_BN self.SA_modules = nn.ModuleList() channel_in = input_channels self.num_prefix_channels = 3 + 2 # xyz + point_scores + point_depth xyz_mlps = [self.num_prefix_channels] + self.model_cfg.XYZ_UP_LAYER shared_mlps = [] for k in range(len(xyz_mlps) - 1): shared_mlps.append(nn.Conv2d(xyz_mlps[k], xyz_mlps[k + 1], kernel_size=1, bias=not use_bn)) if use_bn: shared_mlps.append(nn.BatchNorm2d(xyz_mlps[k + 1])) shared_mlps.append(nn.ReLU()) self.xyz_up_layer = nn.Sequential(*shared_mlps) merge_channel_in = channel_in + self.model_cfg.XYZ_UP_LAYER[-1] self.merge_down_layer = nn.Sequential( nn.Conv2d(merge_channel_in, channel_in, kernel_size=1, bias=not use_bn), *[nn.BatchNorm2d(channel_in), nn.ReLU()] if use_bn else [nn.ReLU()] ) for k in range(self.model_cfg.SA_CONFIG.NPOINTS.__len__()): mlps = [channel_in] + self.model_cfg.SA_CONFIG.MLPS[k] npoint = self.model_cfg.SA_CONFIG.NPOINTS[k] if self.model_cfg.SA_CONFIG.NPOINTS[k] != -1 else None self.SA_modules.append( pointnet2_modules.PointnetSAModule( npoint=npoint, radius=self.model_cfg.SA_CONFIG.RADIUS[k], nsample=self.model_cfg.SA_CONFIG.NSAMPLE[k], mlp=mlps, use_xyz=True, bn=use_bn ) ) channel_in = mlps[-1] self.cls_layers = self.make_fc_layers( input_channels=channel_in, output_channels=self.num_class, fc_list=self.model_cfg.CLS_FC ) self.reg_layers = self.make_fc_layers( input_channels=channel_in, output_channels=self.box_coder.code_size * self.num_class, fc_list=self.model_cfg.REG_FC ) self.roipoint_pool3d_layer = roipoint_pool3d_utils.RoIPointPool3d( num_sampled_points=self.model_cfg.ROI_POINT_POOL.NUM_SAMPLED_POINTS, pool_extra_width=self.model_cfg.ROI_POINT_POOL.POOL_EXTRA_WIDTH ) self.init_weights(weight_init='xavier') def init_weights(self, weight_init='xavier'): if weight_init == 'kaiming': init_func = nn.init.kaiming_normal_ elif weight_init == 'xavier': init_func = nn.init.xavier_normal_ elif weight_init == 'normal': init_func = nn.init.normal_ else: raise NotImplementedError for m in self.modules(): if isinstance(m, nn.Conv2d) or isinstance(m, nn.Conv1d): if weight_init == 'normal': init_func(m.weight, mean=0, std=0.001) else: init_func(m.weight) if m.bias is not None: nn.init.constant_(m.bias, 0) nn.init.normal_(self.reg_layers[-1].weight, mean=0, std=0.001) def roipool3d_gpu(self, batch_dict): """ Args: batch_dict: batch_size: rois: (B, num_rois, 7 + C) point_coords: (num_points, 4) [bs_idx, x, y, z] point_features: (num_points, C) point_cls_scores: (N1 + N2 + N3 + ..., 1) point_part_offset: (N1 + N2 + N3 + ..., 3) Returns: """ batch_size = batch_dict['batch_size'] batch_idx = batch_dict['point_coords'][:, 0] point_coords = batch_dict['point_coords'][:, 1:4] point_features = batch_dict['point_features'] rois = batch_dict['rois'] # (B, num_rois, 7 + C) batch_cnt = point_coords.new_zeros(batch_size).int() for bs_idx in range(batch_size): batch_cnt[bs_idx] = (batch_idx == bs_idx).sum() assert batch_cnt.min() == batch_cnt.max() point_scores = batch_dict['point_cls_scores'].detach() point_depths = point_coords.norm(dim=1) / self.model_cfg.ROI_POINT_POOL.DEPTH_NORMALIZER - 0.5 point_features_list = [point_scores[:, None], point_depths[:, None], point_features] point_features_all = torch.cat(point_features_list, dim=1) batch_points = point_coords.view(batch_size, -1, 3) batch_point_features = point_features_all.view(batch_size, -1, point_features_all.shape[-1]) with torch.no_grad(): pooled_features, pooled_empty_flag = self.roipoint_pool3d_layer( batch_points, batch_point_features, rois ) # pooled_features: (B, num_rois, num_sampled_points, 3 + C), pooled_empty_flag: (B, num_rois) # canonical transformation roi_center = rois[:, :, 0:3] pooled_features[:, :, :, 0:3] -= roi_center.unsqueeze(dim=2) pooled_features = pooled_features.view(-1, pooled_features.shape[-2], pooled_features.shape[-1]) pooled_features[:, :, 0:3] = common_utils.rotate_points_along_z( pooled_features[:, :, 0:3], -rois.view(-1, rois.shape[-1])[:, 6] ) pooled_features[pooled_empty_flag.view(-1) > 0] = 0 return pooled_features def forward(self, batch_dict): """ Args: batch_dict: Returns: """ targets_dict = self.proposal_layer( batch_dict, nms_config=self.model_cfg.NMS_CONFIG['TRAIN' if self.training else 'TEST'] ) if self.training: targets_dict = self.assign_targets(batch_dict) batch_dict['rois'] = targets_dict['rois'] batch_dict['roi_labels'] = targets_dict['roi_labels'] pooled_features = self.roipool3d_gpu(batch_dict) # (total_rois, num_sampled_points, 3 + C) xyz_input = pooled_features[..., 0:self.num_prefix_channels].transpose(1, 2).unsqueeze(dim=3).contiguous() xyz_features = self.xyz_up_layer(xyz_input) point_features = pooled_features[..., self.num_prefix_channels:].transpose(1, 2).unsqueeze(dim=3) merged_features = torch.cat((xyz_features, point_features), dim=1) merged_features = self.merge_down_layer(merged_features) l_xyz, l_features = [pooled_features[..., 0:3].contiguous()], [merged_features.squeeze(dim=3).contiguous()] for i in range(len(self.SA_modules)): li_xyz, li_features = self.SA_modules[i](l_xyz[i], l_features[i]) l_xyz.append(li_xyz) l_features.append(li_features) shared_features = l_features[-1] # (total_rois, num_features, 1) rcnn_cls = self.cls_layers(shared_features).transpose(1, 2).contiguous().squeeze(dim=1) # (B, 1 or 2) rcnn_reg = self.reg_layers(shared_features).transpose(1, 2).contiguous().squeeze(dim=1) # (B, C) if not self.training: batch_cls_preds, batch_box_preds = self.generate_predicted_boxes( batch_size=batch_dict['batch_size'], rois=batch_dict['rois'], cls_preds=rcnn_cls, box_preds=rcnn_reg ) batch_dict['batch_cls_preds'] = batch_cls_preds batch_dict['batch_box_preds'] = batch_box_preds batch_dict['cls_preds_normalized'] = False else: targets_dict['rcnn_cls'] = rcnn_cls targets_dict['rcnn_reg'] = rcnn_reg self.forward_ret_dict = targets_dict return batch_dict
7,866
42.705556
116
py
SASA
SASA-main/pcdet/models/roi_heads/target_assigner/proposal_target_layer.py
import numpy as np import torch import torch.nn as nn from ....ops.iou3d_nms import iou3d_nms_utils class ProposalTargetLayer(nn.Module): def __init__(self, roi_sampler_cfg): super().__init__() self.roi_sampler_cfg = roi_sampler_cfg def forward(self, batch_dict): """ Args: batch_dict: batch_size: rois: (B, num_rois, 7 + C) roi_scores: (B, num_rois) gt_boxes: (B, N, 7 + C + 1) roi_labels: (B, num_rois) Returns: batch_dict: rois: (B, M, 7 + C) gt_of_rois: (B, M, 7 + C) gt_iou_of_rois: (B, M) roi_scores: (B, M) roi_labels: (B, M) reg_valid_mask: (B, M) rcnn_cls_labels: (B, M) """ batch_rois, batch_gt_of_rois, batch_roi_ious, batch_roi_scores, batch_roi_labels = self.sample_rois_for_rcnn( batch_dict=batch_dict ) # regression valid mask reg_valid_mask = (batch_roi_ious > self.roi_sampler_cfg.REG_FG_THRESH).long() # classification label if self.roi_sampler_cfg.CLS_SCORE_TYPE == 'cls': batch_cls_labels = (batch_roi_ious > self.roi_sampler_cfg.CLS_FG_THRESH).long() ignore_mask = (batch_roi_ious > self.roi_sampler_cfg.CLS_BG_THRESH) & \ (batch_roi_ious < self.roi_sampler_cfg.CLS_FG_THRESH) batch_cls_labels[ignore_mask > 0] = -1 elif self.roi_sampler_cfg.CLS_SCORE_TYPE == 'roi_iou': iou_bg_thresh = self.roi_sampler_cfg.CLS_BG_THRESH iou_fg_thresh = self.roi_sampler_cfg.CLS_FG_THRESH fg_mask = batch_roi_ious > iou_fg_thresh bg_mask = batch_roi_ious < iou_bg_thresh interval_mask = (fg_mask == 0) & (bg_mask == 0) batch_cls_labels = (fg_mask > 0).float() batch_cls_labels[interval_mask] = \ (batch_roi_ious[interval_mask] - iou_bg_thresh) / (iou_fg_thresh - iou_bg_thresh) else: raise NotImplementedError targets_dict = {'rois': batch_rois, 'gt_of_rois': batch_gt_of_rois, 'gt_iou_of_rois': batch_roi_ious, 'roi_scores': batch_roi_scores, 'roi_labels': batch_roi_labels, 'reg_valid_mask': reg_valid_mask, 'rcnn_cls_labels': batch_cls_labels} return targets_dict def sample_rois_for_rcnn(self, batch_dict): """ Args: batch_dict: batch_size: rois: (B, num_rois, 7 + C) roi_scores: (B, num_rois) gt_boxes: (B, N, 7 + C + 1) roi_labels: (B, num_rois) Returns: """ batch_size = batch_dict['batch_size'] rois = batch_dict['rois'] roi_scores = batch_dict['roi_scores'] roi_labels = batch_dict['roi_labels'] gt_boxes = batch_dict['gt_boxes'] code_size = rois.shape[-1] batch_rois = rois.new_zeros(batch_size, self.roi_sampler_cfg.ROI_PER_IMAGE, code_size) batch_gt_of_rois = rois.new_zeros(batch_size, self.roi_sampler_cfg.ROI_PER_IMAGE, code_size + 1) batch_roi_ious = rois.new_zeros(batch_size, self.roi_sampler_cfg.ROI_PER_IMAGE) batch_roi_scores = rois.new_zeros(batch_size, self.roi_sampler_cfg.ROI_PER_IMAGE) batch_roi_labels = rois.new_zeros((batch_size, self.roi_sampler_cfg.ROI_PER_IMAGE), dtype=torch.long) for index in range(batch_size): cur_roi, cur_gt, cur_roi_labels, cur_roi_scores = \ rois[index], gt_boxes[index], roi_labels[index], roi_scores[index] k = cur_gt.__len__() - 1 while k > 0 and cur_gt[k].sum() == 0: k -= 1 cur_gt = cur_gt[:k + 1] cur_gt = cur_gt.new_zeros((1, cur_gt.shape[1])) if len(cur_gt) == 0 else cur_gt if self.roi_sampler_cfg.get('SAMPLE_ROI_BY_EACH_CLASS', False): max_overlaps, gt_assignment = self.get_max_iou_with_same_class( rois=cur_roi, roi_labels=cur_roi_labels, gt_boxes=cur_gt[:, 0:7], gt_labels=cur_gt[:, -1].long() ) else: iou3d = iou3d_nms_utils.boxes_iou3d_gpu(cur_roi, cur_gt[:, 0:7]) # (M, N) max_overlaps, gt_assignment = torch.max(iou3d, dim=1) sampled_inds = self.subsample_rois(max_overlaps=max_overlaps) batch_rois[index] = cur_roi[sampled_inds] batch_roi_labels[index] = cur_roi_labels[sampled_inds] batch_roi_ious[index] = max_overlaps[sampled_inds] batch_roi_scores[index] = cur_roi_scores[sampled_inds] batch_gt_of_rois[index] = cur_gt[gt_assignment[sampled_inds]] return batch_rois, batch_gt_of_rois, batch_roi_ious, batch_roi_scores, batch_roi_labels def subsample_rois(self, max_overlaps): # sample fg, easy_bg, hard_bg fg_rois_per_image = int(np.round(self.roi_sampler_cfg.FG_RATIO * self.roi_sampler_cfg.ROI_PER_IMAGE)) fg_thresh = min(self.roi_sampler_cfg.REG_FG_THRESH, self.roi_sampler_cfg.CLS_FG_THRESH) fg_inds = ((max_overlaps >= fg_thresh)).nonzero().view(-1) easy_bg_inds = ((max_overlaps < self.roi_sampler_cfg.CLS_BG_THRESH_LO)).nonzero().view(-1) hard_bg_inds = ((max_overlaps < self.roi_sampler_cfg.REG_FG_THRESH) & (max_overlaps >= self.roi_sampler_cfg.CLS_BG_THRESH_LO)).nonzero().view(-1) fg_num_rois = fg_inds.numel() bg_num_rois = hard_bg_inds.numel() + easy_bg_inds.numel() if fg_num_rois > 0 and bg_num_rois > 0: # sampling fg fg_rois_per_this_image = min(fg_rois_per_image, fg_num_rois) rand_num = torch.from_numpy(np.random.permutation(fg_num_rois)).type_as(max_overlaps).long() fg_inds = fg_inds[rand_num[:fg_rois_per_this_image]] # sampling bg bg_rois_per_this_image = self.roi_sampler_cfg.ROI_PER_IMAGE - fg_rois_per_this_image bg_inds = self.sample_bg_inds( hard_bg_inds, easy_bg_inds, bg_rois_per_this_image, self.roi_sampler_cfg.HARD_BG_RATIO ) elif fg_num_rois > 0 and bg_num_rois == 0: # sampling fg rand_num = np.floor(np.random.rand(self.roi_sampler_cfg.ROI_PER_IMAGE) * fg_num_rois) rand_num = torch.from_numpy(rand_num).type_as(max_overlaps).long() fg_inds = fg_inds[rand_num] bg_inds = [] elif bg_num_rois > 0 and fg_num_rois == 0: # sampling bg bg_rois_per_this_image = self.roi_sampler_cfg.ROI_PER_IMAGE bg_inds = self.sample_bg_inds( hard_bg_inds, easy_bg_inds, bg_rois_per_this_image, self.roi_sampler_cfg.HARD_BG_RATIO ) else: print('maxoverlaps:(min=%f, max=%f)' % (max_overlaps.min().item(), max_overlaps.max().item())) print('ERROR: FG=%d, BG=%d' % (fg_num_rois, bg_num_rois)) raise NotImplementedError sampled_inds = torch.cat((fg_inds, bg_inds), dim=0) return sampled_inds @staticmethod def sample_bg_inds(hard_bg_inds, easy_bg_inds, bg_rois_per_this_image, hard_bg_ratio): if hard_bg_inds.numel() > 0 and easy_bg_inds.numel() > 0: hard_bg_rois_num = min(int(bg_rois_per_this_image * hard_bg_ratio), len(hard_bg_inds)) easy_bg_rois_num = bg_rois_per_this_image - hard_bg_rois_num # sampling hard bg rand_idx = torch.randint(low=0, high=hard_bg_inds.numel(), size=(hard_bg_rois_num,)).long() hard_bg_inds = hard_bg_inds[rand_idx] # sampling easy bg rand_idx = torch.randint(low=0, high=easy_bg_inds.numel(), size=(easy_bg_rois_num,)).long() easy_bg_inds = easy_bg_inds[rand_idx] bg_inds = torch.cat([hard_bg_inds, easy_bg_inds], dim=0) elif hard_bg_inds.numel() > 0 and easy_bg_inds.numel() == 0: hard_bg_rois_num = bg_rois_per_this_image # sampling hard bg rand_idx = torch.randint(low=0, high=hard_bg_inds.numel(), size=(hard_bg_rois_num,)).long() bg_inds = hard_bg_inds[rand_idx] elif hard_bg_inds.numel() == 0 and easy_bg_inds.numel() > 0: easy_bg_rois_num = bg_rois_per_this_image # sampling easy bg rand_idx = torch.randint(low=0, high=easy_bg_inds.numel(), size=(easy_bg_rois_num,)).long() bg_inds = easy_bg_inds[rand_idx] else: raise NotImplementedError return bg_inds @staticmethod def get_max_iou_with_same_class(rois, roi_labels, gt_boxes, gt_labels): """ Args: rois: (N, 7) roi_labels: (N) gt_boxes: (N, ) gt_labels: Returns: """ """ :param rois: (N, 7) :param roi_labels: (N) :param gt_boxes: (N, 8) :return: """ max_overlaps = rois.new_zeros(rois.shape[0]) gt_assignment = roi_labels.new_zeros(roi_labels.shape[0]) for k in range(gt_labels.min().item(), gt_labels.max().item() + 1): roi_mask = (roi_labels == k) gt_mask = (gt_labels == k) if roi_mask.sum() > 0 and gt_mask.sum() > 0: cur_roi = rois[roi_mask] cur_gt = gt_boxes[gt_mask] original_gt_assignment = gt_mask.nonzero().view(-1) iou3d = iou3d_nms_utils.boxes_iou3d_gpu(cur_roi, cur_gt) # (M, N) cur_max_overlaps, cur_gt_assignment = torch.max(iou3d, dim=1) max_overlaps[roi_mask] = cur_max_overlaps gt_assignment[roi_mask] = original_gt_assignment[cur_gt_assignment] return max_overlaps, gt_assignment
9,946
42.436681
117
py
SASA
SASA-main/pcdet/models/model_utils/model_nms_utils.py
import torch from ...ops.iou3d_nms import iou3d_nms_utils def class_agnostic_nms(box_scores, box_preds, nms_config, score_thresh=None): src_box_scores = box_scores if score_thresh is not None: scores_mask = (box_scores >= score_thresh) box_scores = box_scores[scores_mask] box_preds = box_preds[scores_mask] selected = [] if box_scores.shape[0] > 0: box_scores_nms, indices = torch.topk(box_scores, k=min(nms_config.NMS_PRE_MAXSIZE, box_scores.shape[0])) boxes_for_nms = box_preds[indices] keep_idx, selected_scores = getattr(iou3d_nms_utils, nms_config.NMS_TYPE)( boxes_for_nms[:, 0:7], box_scores_nms, nms_config.NMS_THRESH, **nms_config ) selected = indices[keep_idx[:nms_config.NMS_POST_MAXSIZE]] if score_thresh is not None: original_idxs = scores_mask.nonzero().view(-1) selected = original_idxs[selected] return selected, src_box_scores[selected] def multi_classes_nms(cls_scores, box_preds, nms_config, score_thresh=None): """ Args: cls_scores: (N, num_class) box_preds: (N, 7 + C) nms_config: score_thresh: Returns: """ pred_scores, pred_labels, pred_boxes = [], [], [] for k in range(cls_scores.shape[1]): if score_thresh is not None: scores_mask = (cls_scores[:, k] >= score_thresh) box_scores = cls_scores[scores_mask, k] cur_box_preds = box_preds[scores_mask] else: box_scores = cls_scores[:, k] selected = [] if box_scores.shape[0] > 0: box_scores_nms, indices = torch.topk(box_scores, k=min(nms_config.NMS_PRE_MAXSIZE, box_scores.shape[0])) boxes_for_nms = cur_box_preds[indices] keep_idx, selected_scores = getattr(iou3d_nms_utils, nms_config.NMS_TYPE)( boxes_for_nms[:, 0:7], box_scores_nms, nms_config.NMS_THRESH, **nms_config ) selected = indices[keep_idx[:nms_config.NMS_POST_MAXSIZE]] pred_scores.append(box_scores[selected]) pred_labels.append(box_scores.new_ones(len(selected)).long() * k) pred_boxes.append(cur_box_preds[selected]) pred_scores = torch.cat(pred_scores, dim=0) pred_labels = torch.cat(pred_labels, dim=0) pred_boxes = torch.cat(pred_boxes, dim=0) return pred_scores, pred_labels, pred_boxes
2,419
35.666667
116
py
SASA
SASA-main/pcdet/models/backbones_2d/base_bev_backbone.py
import numpy as np import torch import torch.nn as nn class BaseBEVBackbone(nn.Module): def __init__(self, model_cfg, input_channels): super().__init__() self.model_cfg = model_cfg if self.model_cfg.get('LAYER_NUMS', None) is not None: assert len(self.model_cfg.LAYER_NUMS) == len(self.model_cfg.LAYER_STRIDES) == len(self.model_cfg.NUM_FILTERS) layer_nums = self.model_cfg.LAYER_NUMS layer_strides = self.model_cfg.LAYER_STRIDES num_filters = self.model_cfg.NUM_FILTERS else: layer_nums = layer_strides = num_filters = [] if self.model_cfg.get('UPSAMPLE_STRIDES', None) is not None: assert len(self.model_cfg.UPSAMPLE_STRIDES) == len(self.model_cfg.NUM_UPSAMPLE_FILTERS) num_upsample_filters = self.model_cfg.NUM_UPSAMPLE_FILTERS upsample_strides = self.model_cfg.UPSAMPLE_STRIDES else: upsample_strides = num_upsample_filters = [] num_levels = len(layer_nums) c_in_list = [input_channels, *num_filters[:-1]] self.blocks = nn.ModuleList() self.deblocks = nn.ModuleList() for idx in range(num_levels): cur_layers = [ nn.ZeroPad2d(1), nn.Conv2d( c_in_list[idx], num_filters[idx], kernel_size=3, stride=layer_strides[idx], padding=0, bias=False ), nn.BatchNorm2d(num_filters[idx], eps=1e-3, momentum=0.01), nn.ReLU() ] for k in range(layer_nums[idx]): cur_layers.extend([ nn.Conv2d(num_filters[idx], num_filters[idx], kernel_size=3, padding=1, bias=False), nn.BatchNorm2d(num_filters[idx], eps=1e-3, momentum=0.01), nn.ReLU() ]) self.blocks.append(nn.Sequential(*cur_layers)) if len(upsample_strides) > 0: stride = upsample_strides[idx] if stride >= 1: self.deblocks.append(nn.Sequential( nn.ConvTranspose2d( num_filters[idx], num_upsample_filters[idx], upsample_strides[idx], stride=upsample_strides[idx], bias=False ), nn.BatchNorm2d(num_upsample_filters[idx], eps=1e-3, momentum=0.01), nn.ReLU() )) else: stride = np.round(1 / stride).astype(np.int) self.deblocks.append(nn.Sequential( nn.Conv2d( num_filters[idx], num_upsample_filters[idx], stride, stride=stride, bias=False ), nn.BatchNorm2d(num_upsample_filters[idx], eps=1e-3, momentum=0.01), nn.ReLU() )) c_in = sum(num_upsample_filters) if len(upsample_strides) > num_levels: self.deblocks.append(nn.Sequential( nn.ConvTranspose2d(c_in, c_in, upsample_strides[-1], stride=upsample_strides[-1], bias=False), nn.BatchNorm2d(c_in, eps=1e-3, momentum=0.01), nn.ReLU(), )) self.num_bev_features = c_in def forward(self, data_dict): """ Args: data_dict: spatial_features Returns: """ spatial_features = data_dict['spatial_features'] ups = [] ret_dict = {} x = spatial_features for i in range(len(self.blocks)): x = self.blocks[i](x) stride = int(spatial_features.shape[2] / x.shape[2]) ret_dict['spatial_features_%dx' % stride] = x if len(self.deblocks) > 0: ups.append(self.deblocks[i](x)) else: ups.append(x) if len(ups) > 1: x = torch.cat(ups, dim=1) elif len(ups) == 1: x = ups[0] if len(self.deblocks) > len(self.blocks): x = self.deblocks[-1](x) data_dict['spatial_features_2d'] = x return data_dict
4,318
37.221239
121
py
SASA
SASA-main/pcdet/models/backbones_2d/map_to_bev/pointpillar_scatter.py
import torch import torch.nn as nn class PointPillarScatter(nn.Module): def __init__(self, model_cfg, grid_size, **kwargs): super().__init__() self.model_cfg = model_cfg self.num_bev_features = self.model_cfg.NUM_BEV_FEATURES self.nx, self.ny, self.nz = grid_size assert self.nz == 1 def forward(self, batch_dict, **kwargs): pillar_features, coords = batch_dict['pillar_features'], batch_dict['voxel_coords'] batch_spatial_features = [] batch_size = coords[:, 0].max().int().item() + 1 for batch_idx in range(batch_size): spatial_feature = torch.zeros( self.num_bev_features, self.nz * self.nx * self.ny, dtype=pillar_features.dtype, device=pillar_features.device) batch_mask = coords[:, 0] == batch_idx this_coords = coords[batch_mask, :] indices = this_coords[:, 1] + this_coords[:, 2] * self.nx + this_coords[:, 3] indices = indices.type(torch.long) pillars = pillar_features[batch_mask, :] pillars = pillars.t() spatial_feature[:, indices] = pillars batch_spatial_features.append(spatial_feature) batch_spatial_features = torch.stack(batch_spatial_features, 0) batch_spatial_features = batch_spatial_features.view(batch_size, self.num_bev_features * self.nz, self.ny, self.nx) batch_dict['spatial_features'] = batch_spatial_features return batch_dict
1,545
39.684211
123
py
SASA
SASA-main/pcdet/models/backbones_2d/map_to_bev/height_compression.py
import torch.nn as nn class HeightCompression(nn.Module): def __init__(self, model_cfg, **kwargs): super().__init__() self.model_cfg = model_cfg self.num_bev_features = self.model_cfg.NUM_BEV_FEATURES def forward(self, batch_dict): """ Args: batch_dict: encoded_spconv_tensor: sparse tensor Returns: batch_dict: spatial_features: """ encoded_spconv_tensor = batch_dict['encoded_spconv_tensor'] spatial_features = encoded_spconv_tensor.dense() N, C, D, H, W = spatial_features.shape spatial_features = spatial_features.view(N, C * D, H, W) batch_dict['spatial_features'] = spatial_features batch_dict['spatial_features_stride'] = batch_dict['encoded_spconv_tensor_stride'] return batch_dict
870
31.259259
90
py
SASA
SASA-main/pcdet/datasets/dataset.py
from collections import defaultdict from pathlib import Path import numpy as np import torch.utils.data as torch_data from ..utils import common_utils from .augmentor.data_augmentor import DataAugmentor from .processor.data_processor import DataProcessor from .processor.point_feature_encoder import PointFeatureEncoder class DatasetTemplate(torch_data.Dataset): def __init__(self, dataset_cfg=None, class_names=None, training=True, root_path=None, logger=None): super().__init__() self.dataset_cfg = dataset_cfg self.training = training self.class_names = class_names self.logger = logger self.root_path = root_path if root_path is not None else Path(self.dataset_cfg.DATA_PATH) self.logger = logger if self.dataset_cfg is None or class_names is None: return self.point_cloud_range = np.array(self.dataset_cfg.POINT_CLOUD_RANGE, dtype=np.float32) self.point_feature_encoder = PointFeatureEncoder( self.dataset_cfg.POINT_FEATURE_ENCODING, point_cloud_range=self.point_cloud_range ) self.data_augmentor = DataAugmentor( self.root_path, self.dataset_cfg.DATA_AUGMENTOR, self.class_names, logger=self.logger ) if self.training else None self.data_processor = DataProcessor( self.dataset_cfg.DATA_PROCESSOR, point_cloud_range=self.point_cloud_range, training=self.training ) self.grid_size = self.data_processor.grid_size self.voxel_size = self.data_processor.voxel_size self.total_epochs = 0 self._merge_all_iters_to_one_epoch = False @property def mode(self): return 'train' if self.training else 'test' def __getstate__(self): d = dict(self.__dict__) del d['logger'] return d def __setstate__(self, d): self.__dict__.update(d) @staticmethod def generate_prediction_dicts(batch_dict, pred_dicts, class_names, output_path=None): """ To support a custom dataset, implement this function to receive the predicted results from the model, and then transform the unified normative coordinate to your required coordinate, and optionally save them to disk. Args: batch_dict: dict of original data from the dataloader pred_dicts: dict of predicted results from the model pred_boxes: (N, 7), Tensor pred_scores: (N), Tensor pred_labels: (N), Tensor class_names: output_path: if it is not None, save the results to this path Returns: """ def merge_all_iters_to_one_epoch(self, merge=True, epochs=None): if merge: self._merge_all_iters_to_one_epoch = True self.total_epochs = epochs else: self._merge_all_iters_to_one_epoch = False def __len__(self): raise NotImplementedError def __getitem__(self, index): """ To support a custom dataset, implement this function to load the raw data (and labels), then transform them to the unified normative coordinate and call the function self.prepare_data() to process the data and send them to the model. Args: index: Returns: """ raise NotImplementedError def prepare_data(self, data_dict): """ Args: data_dict: points: (N, 3 + C_in) gt_boxes: optional, (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] gt_names: optional, (N), string ... Returns: data_dict: frame_id: string points: (N, 3 + C_in) gt_boxes: optional, (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] gt_names: optional, (N), string use_lead_xyz: bool voxels: optional (num_voxels, max_points_per_voxel, 3 + C) voxel_coords: optional (num_voxels, 3) voxel_num_points: optional (num_voxels) ... """ if self.training: assert 'gt_boxes' in data_dict, 'gt_boxes should be provided for training' gt_boxes_mask = np.array([n in self.class_names for n in data_dict['gt_names']], dtype=np.bool_) data_dict = self.data_augmentor.forward( data_dict={ **data_dict, 'gt_boxes_mask': gt_boxes_mask } ) if data_dict.get('gt_boxes', None) is not None: selected = common_utils.keep_arrays_by_name(data_dict['gt_names'], self.class_names) data_dict['gt_boxes'] = data_dict['gt_boxes'][selected] data_dict['gt_names'] = data_dict['gt_names'][selected] gt_classes = np.array([self.class_names.index(n) + 1 for n in data_dict['gt_names']], dtype=np.int32) gt_boxes = np.concatenate((data_dict['gt_boxes'], gt_classes.reshape(-1, 1).astype(np.float32)), axis=1) data_dict['gt_boxes'] = gt_boxes data_dict = self.point_feature_encoder.forward(data_dict) data_dict = self.data_processor.forward( data_dict=data_dict ) if self.training and len(data_dict['gt_boxes']) == 0: new_index = np.random.randint(self.__len__()) return self.__getitem__(new_index) data_dict.pop('gt_names', None) return data_dict @staticmethod def collate_batch(batch_list, _unused=False): data_dict = defaultdict(list) for cur_sample in batch_list: for key, val in cur_sample.items(): data_dict[key].append(val) batch_size = len(batch_list) ret = {} for key, val in data_dict.items(): try: if key in ['voxels', 'voxel_num_points']: ret[key] = np.concatenate(val, axis=0) elif key in ['points', 'voxel_coords']: coors = [] for i, coor in enumerate(val): coor_pad = np.pad(coor, ((0, 0), (1, 0)), mode='constant', constant_values=i) coors.append(coor_pad) ret[key] = np.concatenate(coors, axis=0) elif key in ['gt_boxes']: max_gt = max([len(x) for x in val]) batch_gt_boxes3d = np.zeros((batch_size, max_gt, val[0].shape[-1]), dtype=np.float32) for k in range(batch_size): batch_gt_boxes3d[k, :val[k].__len__(), :] = val[k] ret[key] = batch_gt_boxes3d else: ret[key] = np.stack(val, axis=0) except: print('Error in collate_batch: key=%s' % key) raise TypeError ret['batch_size'] = batch_size return ret
6,966
37.071038
118
py
SASA
SASA-main/pcdet/datasets/__init__.py
import torch from torch.utils.data import DataLoader from torch.utils.data import DistributedSampler as _DistributedSampler from pcdet.utils import common_utils from .dataset import DatasetTemplate from .kitti.kitti_dataset import KittiDataset from .nuscenes.nuscenes_dataset import NuScenesDataset from .waymo.waymo_dataset import WaymoDataset __all__ = { 'DatasetTemplate': DatasetTemplate, 'KittiDataset': KittiDataset, 'NuScenesDataset': NuScenesDataset, 'WaymoDataset': WaymoDataset } class DistributedSampler(_DistributedSampler): def __init__(self, dataset, num_replicas=None, rank=None, shuffle=True): super().__init__(dataset, num_replicas=num_replicas, rank=rank) self.shuffle = shuffle def __iter__(self): if self.shuffle: g = torch.Generator() g.manual_seed(self.epoch) indices = torch.randperm(len(self.dataset), generator=g).tolist() else: indices = torch.arange(len(self.dataset)).tolist() indices += indices[:(self.total_size - len(indices))] assert len(indices) == self.total_size indices = indices[self.rank:self.total_size:self.num_replicas] assert len(indices) == self.num_samples return iter(indices) def build_dataloader(dataset_cfg, class_names, batch_size, dist, root_path=None, workers=4, logger=None, training=True, merge_all_iters_to_one_epoch=False, total_epochs=0): dataset = __all__[dataset_cfg.DATASET]( dataset_cfg=dataset_cfg, class_names=class_names, root_path=root_path, training=training, logger=logger, ) if merge_all_iters_to_one_epoch: assert hasattr(dataset, 'merge_all_iters_to_one_epoch') dataset.merge_all_iters_to_one_epoch(merge=True, epochs=total_epochs) if dist: if training: sampler = torch.utils.data.distributed.DistributedSampler(dataset) else: rank, world_size = common_utils.get_dist_info() sampler = DistributedSampler(dataset, world_size, rank, shuffle=False) else: sampler = None dataloader = DataLoader( dataset, batch_size=batch_size, pin_memory=True, num_workers=workers, shuffle=(sampler is None) and training, collate_fn=dataset.collate_batch, drop_last=False, sampler=sampler, timeout=0 ) return dataset, dataloader, sampler
2,440
32.438356
101
py
SASA
SASA-main/pcdet/datasets/waymo/waymo_dataset.py
# OpenPCDet PyTorch Dataloader and Evaluation Tools for Waymo Open Dataset # Reference https://github.com/open-mmlab/OpenPCDet # Written by Shaoshuai Shi, Chaoxu Guo # All Rights Reserved 2019-2020. import os import pickle import copy import numpy as np import torch import multiprocessing from tqdm import tqdm from pathlib import Path from ...ops.roiaware_pool3d import roiaware_pool3d_utils from ...utils import box_utils, common_utils from ..dataset import DatasetTemplate class WaymoDataset(DatasetTemplate): def __init__(self, dataset_cfg, class_names, training=True, root_path=None, logger=None): super().__init__( dataset_cfg=dataset_cfg, class_names=class_names, training=training, root_path=root_path, logger=logger ) self.data_path = self.root_path / self.dataset_cfg.PROCESSED_DATA_TAG self.split = self.dataset_cfg.DATA_SPLIT[self.mode] split_dir = self.root_path / 'ImageSets' / (self.split + '.txt') self.sample_sequence_list = [x.strip() for x in open(split_dir).readlines()] self.infos = [] self.include_waymo_data(self.mode) def set_split(self, split): super().__init__( dataset_cfg=self.dataset_cfg, class_names=self.class_names, training=self.training, root_path=self.root_path, logger=self.logger ) self.split = split split_dir = self.root_path / 'ImageSets' / (self.split + '.txt') self.sample_sequence_list = [x.strip() for x in open(split_dir).readlines()] self.infos = [] self.include_waymo_data(self.mode) def include_waymo_data(self, mode): self.logger.info('Loading Waymo dataset') waymo_infos = [] num_skipped_infos = 0 for k in range(len(self.sample_sequence_list)): sequence_name = os.path.splitext(self.sample_sequence_list[k])[0] info_path = self.data_path / sequence_name / ('%s.pkl' % sequence_name) info_path = self.check_sequence_name_with_all_version(info_path) if not info_path.exists(): num_skipped_infos += 1 continue with open(info_path, 'rb') as f: infos = pickle.load(f) waymo_infos.extend(infos) self.infos.extend(waymo_infos[:]) self.logger.info('Total skipped info %s' % num_skipped_infos) self.logger.info('Total samples for Waymo dataset: %d' % (len(waymo_infos))) if self.dataset_cfg.SAMPLED_INTERVAL[mode] > 1: sampled_waymo_infos = [] for k in range(0, len(self.infos), self.dataset_cfg.SAMPLED_INTERVAL[mode]): sampled_waymo_infos.append(self.infos[k]) self.infos = sampled_waymo_infos self.logger.info('Total sampled samples for Waymo dataset: %d' % len(self.infos)) @staticmethod def check_sequence_name_with_all_version(sequence_file): if '_with_camera_labels' not in str(sequence_file) and not sequence_file.exists(): sequence_file = Path(str(sequence_file[:-9]) + '_with_camera_labels.tfrecord') if '_with_camera_labels' in str(sequence_file) and not sequence_file.exists(): sequence_file = Path(str(sequence_file).replace('_with_camera_labels', '')) return sequence_file def get_infos(self, raw_data_path, save_path, num_workers=multiprocessing.cpu_count(), has_label=True, sampled_interval=1): import concurrent.futures as futures from functools import partial from . import waymo_utils print('---------------The waymo sample interval is %d, total sequecnes is %d-----------------' % (sampled_interval, len(self.sample_sequence_list))) process_single_sequence = partial( waymo_utils.process_single_sequence, save_path=save_path, sampled_interval=sampled_interval, has_label=has_label ) sample_sequence_file_list = [ self.check_sequence_name_with_all_version(raw_data_path / sequence_file) for sequence_file in self.sample_sequence_list ] # process_single_sequence(sample_sequence_file_list[0]) with futures.ThreadPoolExecutor(num_workers) as executor: sequence_infos = list(tqdm(executor.map(process_single_sequence, sample_sequence_file_list), total=len(sample_sequence_file_list))) all_sequences_infos = [item for infos in sequence_infos for item in infos] return all_sequences_infos def get_lidar(self, sequence_name, sample_idx): lidar_file = self.data_path / sequence_name / ('%04d.npy' % sample_idx) point_features = np.load(lidar_file) # (N, 7): [x, y, z, intensity, elongation, NLZ_flag] points_all, NLZ_flag = point_features[:, 0:5], point_features[:, 5] points_all = points_all[NLZ_flag == -1] points_all[:, 3] = np.tanh(points_all[:, 3]) return points_all def __len__(self): if self._merge_all_iters_to_one_epoch: return len(self.infos) * self.total_epochs return len(self.infos) def __getitem__(self, index): if self._merge_all_iters_to_one_epoch: index = index % len(self.infos) info = copy.deepcopy(self.infos[index]) pc_info = info['point_cloud'] sequence_name = pc_info['lidar_sequence'] sample_idx = pc_info['sample_idx'] points = self.get_lidar(sequence_name, sample_idx) input_dict = { 'points': points, 'frame_id': info['frame_id'], } if 'annos' in info: annos = info['annos'] annos = common_utils.drop_info_with_name(annos, name='unknown') if self.dataset_cfg.get('INFO_WITH_FAKELIDAR', False): gt_boxes_lidar = box_utils.boxes3d_kitti_fakelidar_to_lidar(annos['gt_boxes_lidar']) else: gt_boxes_lidar = annos['gt_boxes_lidar'] input_dict.update({ 'gt_names': annos['name'], 'gt_boxes': gt_boxes_lidar, 'num_points_in_gt': annos.get('num_points_in_gt', None) }) data_dict = self.prepare_data(data_dict=input_dict) data_dict['metadata'] = info.get('metadata', info['frame_id']) data_dict.pop('num_points_in_gt', None) return data_dict @staticmethod def generate_prediction_dicts(batch_dict, pred_dicts, class_names, output_path=None): """ Args: batch_dict: frame_id: pred_dicts: list of pred_dicts pred_boxes: (N, 7), Tensor pred_scores: (N), Tensor pred_labels: (N), Tensor class_names: output_path: Returns: """ def get_template_prediction(num_samples): ret_dict = { 'name': np.zeros(num_samples), 'score': np.zeros(num_samples), 'boxes_lidar': np.zeros([num_samples, 7]) } return ret_dict def generate_single_sample_dict(box_dict): pred_scores = box_dict['pred_scores'].cpu().numpy() pred_boxes = box_dict['pred_boxes'].cpu().numpy() pred_labels = box_dict['pred_labels'].cpu().numpy() pred_dict = get_template_prediction(pred_scores.shape[0]) if pred_scores.shape[0] == 0: return pred_dict pred_dict['name'] = np.array(class_names)[pred_labels - 1] pred_dict['score'] = pred_scores pred_dict['boxes_lidar'] = pred_boxes return pred_dict annos = [] for index, box_dict in enumerate(pred_dicts): single_pred_dict = generate_single_sample_dict(box_dict) single_pred_dict['frame_id'] = batch_dict['frame_id'][index] single_pred_dict['metadata'] = batch_dict['metadata'][index] annos.append(single_pred_dict) return annos def evaluation(self, det_annos, class_names, **kwargs): if 'annos' not in self.infos[0].keys(): return 'No ground-truth boxes for evaluation', {} def kitti_eval(eval_det_annos, eval_gt_annos): from ..kitti.kitti_object_eval_python import eval as kitti_eval from ..kitti import kitti_utils map_name_to_kitti = { 'Vehicle': 'Car', 'Pedestrian': 'Pedestrian', 'Cyclist': 'Cyclist', 'Sign': 'Sign', 'Car': 'Car' } kitti_utils.transform_annotations_to_kitti_format(eval_det_annos, map_name_to_kitti=map_name_to_kitti) kitti_utils.transform_annotations_to_kitti_format( eval_gt_annos, map_name_to_kitti=map_name_to_kitti, info_with_fakelidar=self.dataset_cfg.get('INFO_WITH_FAKELIDAR', False) ) kitti_class_names = [map_name_to_kitti[x] for x in class_names] ap_result_str, ap_dict = kitti_eval.get_official_eval_result( gt_annos=eval_gt_annos, dt_annos=eval_det_annos, current_classes=kitti_class_names ) return ap_result_str, ap_dict def waymo_eval(eval_det_annos, eval_gt_annos): from .waymo_eval import OpenPCDetWaymoDetectionMetricsEstimator eval = OpenPCDetWaymoDetectionMetricsEstimator() ap_dict = eval.waymo_evaluation( eval_det_annos, eval_gt_annos, class_name=class_names, distance_thresh=1000, fake_gt_infos=self.dataset_cfg.get('INFO_WITH_FAKELIDAR', False) ) ap_result_str = '\n' for key in ap_dict: ap_dict[key] = ap_dict[key][0] ap_result_str += '%s: %.4f \n' % (key, ap_dict[key]) return ap_result_str, ap_dict eval_det_annos = copy.deepcopy(det_annos) eval_gt_annos = [copy.deepcopy(info['annos']) for info in self.infos] if kwargs['eval_metric'] == 'kitti': ap_result_str, ap_dict = kitti_eval(eval_det_annos, eval_gt_annos) elif kwargs['eval_metric'] == 'waymo': ap_result_str, ap_dict = waymo_eval(eval_det_annos, eval_gt_annos) else: raise NotImplementedError return ap_result_str, ap_dict def create_groundtruth_database(self, info_path, save_path, used_classes=None, split='train', sampled_interval=10, processed_data_tag=None): database_save_path = save_path / ('pcdet_gt_database_%s_sampled_%d' % (split, sampled_interval)) db_info_save_path = save_path / ('pcdet_waymo_dbinfos_%s_sampled_%d.pkl' % (split, sampled_interval)) database_save_path.mkdir(parents=True, exist_ok=True) all_db_infos = {} with open(info_path, 'rb') as f: infos = pickle.load(f) for k in range(0, len(infos), sampled_interval): print('gt_database sample: %d/%d' % (k + 1, len(infos))) info = infos[k] pc_info = info['point_cloud'] sequence_name = pc_info['lidar_sequence'] sample_idx = pc_info['sample_idx'] points = self.get_lidar(sequence_name, sample_idx) annos = info['annos'] names = annos['name'] difficulty = annos['difficulty'] gt_boxes = annos['gt_boxes_lidar'] num_obj = gt_boxes.shape[0] box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( torch.from_numpy(points[:, 0:3]).unsqueeze(dim=0).float().cuda(), torch.from_numpy(gt_boxes[:, 0:7]).unsqueeze(dim=0).float().cuda() ).long().squeeze(dim=0).cpu().numpy() for i in range(num_obj): filename = '%s_%04d_%s_%d.bin' % (sequence_name, sample_idx, names[i], i) filepath = database_save_path / filename gt_points = points[box_idxs_of_pts == i] gt_points[:, :3] -= gt_boxes[i, :3] if (used_classes is None) or names[i] in used_classes: with open(filepath, 'w') as f: gt_points.tofile(f) db_path = str(filepath.relative_to(self.root_path)) # gt_database/xxxxx.bin db_info = {'name': names[i], 'path': db_path, 'sequence_name': sequence_name, 'sample_idx': sample_idx, 'gt_idx': i, 'box3d_lidar': gt_boxes[i], 'num_points_in_gt': gt_points.shape[0], 'difficulty': difficulty[i]} if names[i] in all_db_infos: all_db_infos[names[i]].append(db_info) else: all_db_infos[names[i]] = [db_info] for k, v in all_db_infos.items(): print('Database %s: %d' % (k, len(v))) with open(db_info_save_path, 'wb') as f: pickle.dump(all_db_infos, f) def create_waymo_infos(dataset_cfg, class_names, data_path, save_path, raw_data_tag='raw_data', processed_data_tag='waymo_processed_data', workers=multiprocessing.cpu_count()): dataset = WaymoDataset( dataset_cfg=dataset_cfg, class_names=class_names, root_path=data_path, training=False, logger=common_utils.create_logger() ) train_split, val_split = 'train', 'val' train_filename = save_path / ('waymo_infos_%s.pkl' % train_split) val_filename = save_path / ('waymo_infos_%s.pkl' % val_split) print('---------------Start to generate data infos---------------') dataset.set_split(train_split) waymo_infos_train = dataset.get_infos( raw_data_path=data_path / raw_data_tag, save_path=save_path / processed_data_tag, num_workers=workers, has_label=True, sampled_interval=1 ) with open(train_filename, 'wb') as f: pickle.dump(waymo_infos_train, f) print('----------------Waymo info train file is saved to %s----------------' % train_filename) dataset.set_split(val_split) waymo_infos_val = dataset.get_infos( raw_data_path=data_path / raw_data_tag, save_path=save_path / processed_data_tag, num_workers=workers, has_label=True, sampled_interval=1 ) with open(val_filename, 'wb') as f: pickle.dump(waymo_infos_val, f) print('----------------Waymo info val file is saved to %s----------------' % val_filename) print('---------------Start create groundtruth database for data augmentation---------------') dataset.set_split(train_split) dataset.create_groundtruth_database( info_path=train_filename, save_path=save_path, split='train', sampled_interval=10, used_classes=['Vehicle', 'Pedestrian', 'Cyclist'] ) print('---------------Data preparation Done---------------') if __name__ == '__main__': import argparse parser = argparse.ArgumentParser(description='arg parser') parser.add_argument('--cfg_file', type=str, default=None, help='specify the config of dataset') parser.add_argument('--func', type=str, default='create_waymo_infos', help='') args = parser.parse_args() if args.func == 'create_waymo_infos': import yaml from easydict import EasyDict dataset_cfg = EasyDict(yaml.load(open(args.cfg_file))) ROOT_DIR = (Path(__file__).resolve().parent / '../../../').resolve() create_waymo_infos( dataset_cfg=dataset_cfg, class_names=['Vehicle', 'Pedestrian', 'Cyclist'], data_path=ROOT_DIR / 'data' / 'waymo', save_path=ROOT_DIR / 'data' / 'waymo', raw_data_tag='raw_data', processed_data_tag=dataset_cfg.PROCESSED_DATA_TAG )
15,837
41.461126
127
py
SASA
SASA-main/pcdet/datasets/nuscenes/nuscenes_utils.py
""" The NuScenes data pre-processing and evaluation is modified from https://github.com/traveller59/second.pytorch and https://github.com/poodarchu/Det3D """ import operator from functools import reduce from pathlib import Path import numpy as np import tqdm from nuscenes.utils.data_classes import Box from nuscenes.utils.geometry_utils import transform_matrix from pyquaternion import Quaternion map_name_from_general_to_detection = { 'human.pedestrian.adult': 'pedestrian', 'human.pedestrian.child': 'pedestrian', 'human.pedestrian.wheelchair': 'ignore', 'human.pedestrian.stroller': 'ignore', 'human.pedestrian.personal_mobility': 'ignore', 'human.pedestrian.police_officer': 'pedestrian', 'human.pedestrian.construction_worker': 'pedestrian', 'animal': 'ignore', 'vehicle.car': 'car', 'vehicle.motorcycle': 'motorcycle', 'vehicle.bicycle': 'bicycle', 'vehicle.bus.bendy': 'bus', 'vehicle.bus.rigid': 'bus', 'vehicle.truck': 'truck', 'vehicle.construction': 'construction_vehicle', 'vehicle.emergency.ambulance': 'ignore', 'vehicle.emergency.police': 'ignore', 'vehicle.trailer': 'trailer', 'movable_object.barrier': 'barrier', 'movable_object.trafficcone': 'traffic_cone', 'movable_object.pushable_pullable': 'ignore', 'movable_object.debris': 'ignore', 'static_object.bicycle_rack': 'ignore', } cls_attr_dist = { 'barrier': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 0, 'vehicle.parked': 0, 'vehicle.stopped': 0, }, 'bicycle': { 'cycle.with_rider': 2791, 'cycle.without_rider': 8946, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 0, 'vehicle.parked': 0, 'vehicle.stopped': 0, }, 'bus': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 9092, 'vehicle.parked': 3294, 'vehicle.stopped': 3881, }, 'car': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 114304, 'vehicle.parked': 330133, 'vehicle.stopped': 46898, }, 'construction_vehicle': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 882, 'vehicle.parked': 11549, 'vehicle.stopped': 2102, }, 'ignore': { 'cycle.with_rider': 307, 'cycle.without_rider': 73, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 165, 'vehicle.parked': 400, 'vehicle.stopped': 102, }, 'motorcycle': { 'cycle.with_rider': 4233, 'cycle.without_rider': 8326, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 0, 'vehicle.parked': 0, 'vehicle.stopped': 0, }, 'pedestrian': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 157444, 'pedestrian.sitting_lying_down': 13939, 'pedestrian.standing': 46530, 'vehicle.moving': 0, 'vehicle.parked': 0, 'vehicle.stopped': 0, }, 'traffic_cone': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 0, 'vehicle.parked': 0, 'vehicle.stopped': 0, }, 'trailer': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 3421, 'vehicle.parked': 19224, 'vehicle.stopped': 1895, }, 'truck': { 'cycle.with_rider': 0, 'cycle.without_rider': 0, 'pedestrian.moving': 0, 'pedestrian.sitting_lying_down': 0, 'pedestrian.standing': 0, 'vehicle.moving': 21339, 'vehicle.parked': 55626, 'vehicle.stopped': 11097, }, } def get_available_scenes(nusc): available_scenes = [] print('total scene num:', len(nusc.scene)) for scene in nusc.scene: scene_token = scene['token'] scene_rec = nusc.get('scene', scene_token) sample_rec = nusc.get('sample', scene_rec['first_sample_token']) sd_rec = nusc.get('sample_data', sample_rec['data']['LIDAR_TOP']) has_more_frames = True scene_not_exist = False while has_more_frames: lidar_path, boxes, _ = nusc.get_sample_data(sd_rec['token']) if not Path(lidar_path).exists(): scene_not_exist = True break else: break # if not sd_rec['next'] == '': # sd_rec = nusc.get('sample_data', sd_rec['next']) # else: # has_more_frames = False if scene_not_exist: continue available_scenes.append(scene) print('exist scene num:', len(available_scenes)) return available_scenes def get_sample_data(nusc, sample_data_token, selected_anntokens=None): """ Returns the data path as well as all annotations related to that sample_data. Note that the boxes are transformed into the current sensor's coordinate frame. Args: nusc: sample_data_token: Sample_data token. selected_anntokens: If provided only return the selected annotation. Returns: """ # Retrieve sensor & pose records sd_record = nusc.get('sample_data', sample_data_token) cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token']) sensor_record = nusc.get('sensor', cs_record['sensor_token']) pose_record = nusc.get('ego_pose', sd_record['ego_pose_token']) data_path = nusc.get_sample_data_path(sample_data_token) if sensor_record['modality'] == 'camera': cam_intrinsic = np.array(cs_record['camera_intrinsic']) imsize = (sd_record['width'], sd_record['height']) else: cam_intrinsic = imsize = None # Retrieve all sample annotations and map to sensor coordinate system. if selected_anntokens is not None: boxes = list(map(nusc.get_box, selected_anntokens)) else: boxes = nusc.get_boxes(sample_data_token) # Make list of Box objects including coord system transforms. box_list = [] for box in boxes: box.velocity = nusc.box_velocity(box.token) # Move box to ego vehicle coord system box.translate(-np.array(pose_record['translation'])) box.rotate(Quaternion(pose_record['rotation']).inverse) # Move box to sensor coord system box.translate(-np.array(cs_record['translation'])) box.rotate(Quaternion(cs_record['rotation']).inverse) box_list.append(box) return data_path, box_list, cam_intrinsic def quaternion_yaw(q: Quaternion) -> float: """ Calculate the yaw angle from a quaternion. Note that this only works for a quaternion that represents a box in lidar or global coordinate frame. It does not work for a box in the camera frame. :param q: Quaternion of interest. :return: Yaw angle in radians. """ # Project into xy plane. v = np.dot(q.rotation_matrix, np.array([1, 0, 0])) # Measure yaw using arctan. yaw = np.arctan2(v[1], v[0]) return yaw def fill_trainval_infos(data_path, nusc, train_scenes, val_scenes, test=False, max_sweeps=10): train_nusc_infos = [] val_nusc_infos = [] progress_bar = tqdm.tqdm(total=len(nusc.sample), desc='create_info', dynamic_ncols=True) ref_chan = 'LIDAR_TOP' # The radar channel from which we track back n sweeps to aggregate the point cloud. chan = 'LIDAR_TOP' # The reference channel of the current sample_rec that the point clouds are mapped to. for index, sample in enumerate(nusc.sample): progress_bar.update() ref_sd_token = sample['data'][ref_chan] ref_sd_rec = nusc.get('sample_data', ref_sd_token) ref_cs_rec = nusc.get('calibrated_sensor', ref_sd_rec['calibrated_sensor_token']) ref_pose_rec = nusc.get('ego_pose', ref_sd_rec['ego_pose_token']) ref_time = 1e-6 * ref_sd_rec['timestamp'] ref_lidar_path, ref_boxes, _ = get_sample_data(nusc, ref_sd_token) ref_cam_front_token = sample['data']['CAM_FRONT'] ref_cam_path, _, ref_cam_intrinsic = nusc.get_sample_data(ref_cam_front_token) # Homogeneous transform from ego car frame to reference frame ref_from_car = transform_matrix( ref_cs_rec['translation'], Quaternion(ref_cs_rec['rotation']), inverse=True ) # Homogeneous transformation matrix from global to _current_ ego car frame car_from_global = transform_matrix( ref_pose_rec['translation'], Quaternion(ref_pose_rec['rotation']), inverse=True, ) info = { 'lidar_path': Path(ref_lidar_path).relative_to(data_path).__str__(), 'cam_front_path': Path(ref_cam_path).relative_to(data_path).__str__(), 'cam_intrinsic': ref_cam_intrinsic, 'token': sample['token'], 'sweeps': [], 'ref_from_car': ref_from_car, 'car_from_global': car_from_global, 'timestamp': ref_time, } sample_data_token = sample['data'][chan] curr_sd_rec = nusc.get('sample_data', sample_data_token) sweeps = [] while len(sweeps) < max_sweeps - 1: if curr_sd_rec['prev'] == '': if len(sweeps) == 0: sweep = { 'lidar_path': Path(ref_lidar_path).relative_to(data_path).__str__(), 'sample_data_token': curr_sd_rec['token'], 'transform_matrix': None, 'time_lag': curr_sd_rec['timestamp'] * 0, } sweeps.append(sweep) else: sweeps.append(sweeps[-1]) else: curr_sd_rec = nusc.get('sample_data', curr_sd_rec['prev']) # Get past pose current_pose_rec = nusc.get('ego_pose', curr_sd_rec['ego_pose_token']) global_from_car = transform_matrix( current_pose_rec['translation'], Quaternion(current_pose_rec['rotation']), inverse=False, ) # Homogeneous transformation matrix from sensor coordinate frame to ego car frame. current_cs_rec = nusc.get( 'calibrated_sensor', curr_sd_rec['calibrated_sensor_token'] ) car_from_current = transform_matrix( current_cs_rec['translation'], Quaternion(current_cs_rec['rotation']), inverse=False, ) tm = reduce(np.dot, [ref_from_car, car_from_global, global_from_car, car_from_current]) lidar_path = nusc.get_sample_data_path(curr_sd_rec['token']) time_lag = ref_time - 1e-6 * curr_sd_rec['timestamp'] sweep = { 'lidar_path': Path(lidar_path).relative_to(data_path).__str__(), 'sample_data_token': curr_sd_rec['token'], 'transform_matrix': tm, 'global_from_car': global_from_car, 'car_from_current': car_from_current, 'time_lag': time_lag, } sweeps.append(sweep) info['sweeps'] = sweeps assert len(info['sweeps']) == max_sweeps - 1, \ f"sweep {curr_sd_rec['token']} only has {len(info['sweeps'])} sweeps, " \ f"you should duplicate to sweep num {max_sweeps - 1}" if not test: annotations = [nusc.get('sample_annotation', token) for token in sample['anns']] # the filtering gives 0.5~1 map improvement num_lidar_pts = np.array([anno['num_lidar_pts'] for anno in annotations]) num_radar_pts = np.array([anno['num_radar_pts'] for anno in annotations]) mask = (num_lidar_pts + num_radar_pts > 0) locs = np.array([b.center for b in ref_boxes]).reshape(-1, 3) dims = np.array([b.wlh for b in ref_boxes]).reshape(-1, 3)[:, [1, 0, 2]] # wlh == > dxdydz (lwh) velocity = np.array([b.velocity for b in ref_boxes]).reshape(-1, 3) rots = np.array([quaternion_yaw(b.orientation) for b in ref_boxes]).reshape(-1, 1) names = np.array([b.name for b in ref_boxes]) tokens = np.array([b.token for b in ref_boxes]) gt_boxes = np.concatenate([locs, dims, rots, velocity[:, :2]], axis=1) assert len(annotations) == len(gt_boxes) == len(velocity) info['gt_boxes'] = gt_boxes[mask, :] info['gt_boxes_velocity'] = velocity[mask, :] info['gt_names'] = np.array([map_name_from_general_to_detection[name] for name in names])[mask] info['gt_boxes_token'] = tokens[mask] info['num_lidar_pts'] = num_lidar_pts[mask] info['num_radar_pts'] = num_radar_pts[mask] if sample['scene_token'] in train_scenes: train_nusc_infos.append(info) else: val_nusc_infos.append(info) progress_bar.close() return train_nusc_infos, val_nusc_infos def boxes_lidar_to_nusenes(det_info): boxes3d = det_info['boxes_lidar'] scores = det_info['score'] labels = det_info['pred_labels'] box_list = [] for k in range(boxes3d.shape[0]): quat = Quaternion(axis=[0, 0, 1], radians=boxes3d[k, 6]) velocity = (*boxes3d[k, 7:9], 0.0) if boxes3d.shape[1] == 9 else (0.0, 0.0, 0.0) box = Box( boxes3d[k, :3], boxes3d[k, [4, 3, 5]], # wlh quat, label=labels[k], score=scores[k], velocity=velocity, ) box_list.append(box) return box_list def lidar_nusc_box_to_global(nusc, boxes, sample_token): s_record = nusc.get('sample', sample_token) sample_data_token = s_record['data']['LIDAR_TOP'] sd_record = nusc.get('sample_data', sample_data_token) cs_record = nusc.get('calibrated_sensor', sd_record['calibrated_sensor_token']) sensor_record = nusc.get('sensor', cs_record['sensor_token']) pose_record = nusc.get('ego_pose', sd_record['ego_pose_token']) data_path = nusc.get_sample_data_path(sample_data_token) box_list = [] for box in boxes: # Move box to ego vehicle coord system box.rotate(Quaternion(cs_record['rotation'])) box.translate(np.array(cs_record['translation'])) # Move box to global coord system box.rotate(Quaternion(pose_record['rotation'])) box.translate(np.array(pose_record['translation'])) box_list.append(box) return box_list def transform_det_annos_to_nusc_annos(det_annos, nusc): nusc_annos = { 'results': {}, 'meta': None, } for det in det_annos: annos = [] box_list = boxes_lidar_to_nusenes(det) box_list = lidar_nusc_box_to_global( nusc=nusc, boxes=box_list, sample_token=det['metadata']['token'] ) for k, box in enumerate(box_list): name = det['name'][k] if np.sqrt(box.velocity[0] ** 2 + box.velocity[1] ** 2) > 0.2: if name in ['car', 'construction_vehicle', 'bus', 'truck', 'trailer']: attr = 'vehicle.moving' elif name in ['bicycle', 'motorcycle']: attr = 'cycle.with_rider' else: attr = None else: if name in ['pedestrian']: attr = 'pedestrian.standing' elif name in ['bus']: attr = 'vehicle.stopped' else: attr = None attr = attr if attr is not None else max( cls_attr_dist[name].items(), key=operator.itemgetter(1))[0] nusc_anno = { 'sample_token': det['metadata']['token'], 'translation': box.center.tolist(), 'size': box.wlh.tolist(), 'rotation': box.orientation.elements.tolist(), 'velocity': box.velocity[:2].tolist(), 'detection_name': name, 'detection_score': box.score, 'attribute_name': attr } annos.append(nusc_anno) nusc_annos['results'].update({det["metadata"]["token"]: annos}) return nusc_annos def format_nuscene_results(metrics, class_names, version='default'): result = '----------------Nuscene %s results-----------------\n' % version for name in class_names: threshs = ', '.join(list(metrics['label_aps'][name].keys())) ap_list = list(metrics['label_aps'][name].values()) err_name =', '.join([x.split('_')[0] for x in list(metrics['label_tp_errors'][name].keys())]) error_list = list(metrics['label_tp_errors'][name].values()) result += f'***{name} error@{err_name} | AP@{threshs}\n' result += ', '.join(['%.2f' % x for x in error_list]) + ' | ' result += ', '.join(['%.2f' % (x * 100) for x in ap_list]) result += f" | mean AP: {metrics['mean_dist_aps'][name]}" result += '\n' result += '--------------average performance-------------\n' details = {} for key, val in metrics['tp_errors'].items(): result += '%s:\t %.4f\n' % (key, val) details[key] = val result += 'mAP:\t %.4f\n' % metrics['mean_ap'] result += 'NDS:\t %.4f\n' % metrics['nd_score'] details.update({ 'mAP': metrics['mean_ap'], 'NDS': metrics['nd_score'], }) return result, details
18,474
35.876248
111
py
SASA
SASA-main/pcdet/datasets/nuscenes/nuscenes_dataset.py
import copy import pickle from pathlib import Path import numpy as np from tqdm import tqdm from ...ops.roiaware_pool3d import roiaware_pool3d_utils from ...utils import common_utils from ..dataset import DatasetTemplate class NuScenesDataset(DatasetTemplate): def __init__(self, dataset_cfg, class_names, training=True, root_path=None, logger=None): root_path = (root_path if root_path is not None else Path(dataset_cfg.DATA_PATH)) / dataset_cfg.VERSION super().__init__( dataset_cfg=dataset_cfg, class_names=class_names, training=training, root_path=root_path, logger=logger ) self.infos = [] self.include_nuscenes_data(self.mode) if self.training and self.dataset_cfg.get('BALANCED_RESAMPLING', False): self.infos = self.balanced_infos_resampling(self.infos) def include_nuscenes_data(self, mode): self.logger.info('Loading NuScenes dataset') nuscenes_infos = [] for info_path in self.dataset_cfg.INFO_PATH[mode]: info_path = self.root_path / info_path if not info_path.exists(): continue with open(info_path, 'rb') as f: infos = pickle.load(f) nuscenes_infos.extend(infos) self.infos.extend(nuscenes_infos) self.logger.info('Total samples for NuScenes dataset: %d' % (len(nuscenes_infos))) def balanced_infos_resampling(self, infos): """ Class-balanced sampling of nuScenes dataset from https://arxiv.org/abs/1908.09492 """ if self.class_names is None: return infos cls_infos = {name: [] for name in self.class_names} for info in infos: for name in set(info['gt_names']): if name in self.class_names: cls_infos[name].append(info) duplicated_samples = sum([len(v) for _, v in cls_infos.items()]) cls_dist = {k: len(v) / duplicated_samples for k, v in cls_infos.items()} sampled_infos = [] frac = 1.0 / len(self.class_names) ratios = [frac / v for v in cls_dist.values()] for cur_cls_infos, ratio in zip(list(cls_infos.values()), ratios): sampled_infos += np.random.choice( cur_cls_infos, int(len(cur_cls_infos) * ratio) ).tolist() self.logger.info('Total samples after balanced resampling: %s' % (len(sampled_infos))) cls_infos_new = {name: [] for name in self.class_names} for info in sampled_infos: for name in set(info['gt_names']): if name in self.class_names: cls_infos_new[name].append(info) cls_dist_new = {k: len(v) / len(sampled_infos) for k, v in cls_infos_new.items()} return sampled_infos def get_sweep(self, sweep_info): def remove_ego_points(points, center_radius=1.0): mask = ~((np.abs(points[:, 0]) < center_radius) & (np.abs(points[:, 1]) < center_radius)) return points[mask] lidar_path = self.root_path / sweep_info['lidar_path'] points_sweep = np.fromfile(str(lidar_path), dtype=np.float32, count=-1).reshape([-1, 5])[:, :4] points_sweep = remove_ego_points(points_sweep).T if sweep_info['transform_matrix'] is not None: num_points = points_sweep.shape[1] points_sweep[:3, :] = sweep_info['transform_matrix'].dot( np.vstack((points_sweep[:3, :], np.ones(num_points))))[:3, :] cur_times = sweep_info['time_lag'] * np.ones((1, points_sweep.shape[1])) return points_sweep.T, cur_times.T def get_lidar_with_sweeps(self, index, max_sweeps=1): info = self.infos[index] lidar_path = self.root_path / info['lidar_path'] points = np.fromfile(str(lidar_path), dtype=np.float32, count=-1).reshape([-1, 5])[:, :4] sweep_points_list = [points] sweep_times_list = [np.zeros((points.shape[0], 1))] for k in np.random.choice(len(info['sweeps']), max_sweeps - 1, replace=False): points_sweep, times_sweep = self.get_sweep(info['sweeps'][k]) sweep_points_list.append(points_sweep) sweep_times_list.append(times_sweep) points = np.concatenate(sweep_points_list, axis=0) times = np.concatenate(sweep_times_list, axis=0).astype(points.dtype) points = np.concatenate((points, times), axis=1) return points def __len__(self): if self._merge_all_iters_to_one_epoch: return len(self.infos) * self.total_epochs return len(self.infos) def __getitem__(self, index): if self._merge_all_iters_to_one_epoch: index = index % len(self.infos) info = copy.deepcopy(self.infos[index]) points = self.get_lidar_with_sweeps(index, max_sweeps=self.dataset_cfg.MAX_SWEEPS) input_dict = { 'points': points, 'frame_id': Path(info['lidar_path']).stem, 'metadata': {'token': info['token']} } if 'gt_boxes' in info: if self.dataset_cfg.get('FILTER_MIN_POINTS_IN_GT', False): mask = (info['num_lidar_pts'] > self.dataset_cfg.FILTER_MIN_POINTS_IN_GT - 1) else: mask = None input_dict.update({ 'gt_names': info['gt_names'] if mask is None else info['gt_names'][mask], 'gt_boxes': info['gt_boxes'] if mask is None else info['gt_boxes'][mask] }) data_dict = self.prepare_data(data_dict=input_dict) if self.dataset_cfg.get('SET_NAN_VELOCITY_TO_ZEROS', False): gt_boxes = data_dict['gt_boxes'] gt_boxes[np.isnan(gt_boxes)] = 0 data_dict['gt_boxes'] = gt_boxes if not self.dataset_cfg.PRED_VELOCITY and 'gt_boxes' in data_dict: data_dict['gt_boxes'] = data_dict['gt_boxes'][:, [0, 1, 2, 3, 4, 5, 6, -1]] return data_dict @staticmethod def generate_prediction_dicts(batch_dict, pred_dicts, class_names, output_path=None): """ Args: batch_dict: frame_id: pred_dicts: list of pred_dicts pred_boxes: (N, 7), Tensor pred_scores: (N), Tensor pred_labels: (N), Tensor class_names: output_path: Returns: """ def get_template_prediction(num_samples): ret_dict = { 'name': np.zeros(num_samples), 'score': np.zeros(num_samples), 'boxes_lidar': np.zeros([num_samples, 7]), 'pred_labels': np.zeros(num_samples) } return ret_dict def generate_single_sample_dict(box_dict): pred_scores = box_dict['pred_scores'].cpu().numpy() pred_boxes = box_dict['pred_boxes'].cpu().numpy() pred_labels = box_dict['pred_labels'].cpu().numpy() pred_dict = get_template_prediction(pred_scores.shape[0]) if pred_scores.shape[0] == 0: return pred_dict pred_dict['name'] = np.array(class_names)[pred_labels - 1] pred_dict['score'] = pred_scores pred_dict['boxes_lidar'] = pred_boxes pred_dict['pred_labels'] = pred_labels return pred_dict annos = [] for index, box_dict in enumerate(pred_dicts): single_pred_dict = generate_single_sample_dict(box_dict) single_pred_dict['frame_id'] = batch_dict['frame_id'][index] single_pred_dict['metadata'] = batch_dict['metadata'][index] annos.append(single_pred_dict) return annos def evaluation(self, det_annos, class_names, **kwargs): import json from nuscenes.nuscenes import NuScenes from . import nuscenes_utils nusc = NuScenes(version=self.dataset_cfg.VERSION, dataroot=str(self.root_path), verbose=True) nusc_annos = nuscenes_utils.transform_det_annos_to_nusc_annos(det_annos, nusc) nusc_annos['meta'] = { 'use_camera': False, 'use_lidar': True, 'use_radar': False, 'use_map': False, 'use_external': False, } output_path = Path(kwargs['output_path']) output_path.mkdir(exist_ok=True, parents=True) res_path = str(output_path / 'results_nusc.json') with open(res_path, 'w') as f: json.dump(nusc_annos, f) self.logger.info(f'The predictions of NuScenes have been saved to {res_path}') if self.dataset_cfg.VERSION == 'v1.0-test': return 'No ground-truth annotations for evaluation', {} from nuscenes.eval.detection.config import config_factory from nuscenes.eval.detection.evaluate import NuScenesEval eval_set_map = { 'v1.0-mini': 'mini_val', 'v1.0-trainval': 'val', 'v1.0-test': 'test' } try: eval_version = 'detection_cvpr_2019' eval_config = config_factory(eval_version) except: eval_version = 'cvpr_2019' eval_config = config_factory(eval_version) nusc_eval = NuScenesEval( nusc, config=eval_config, result_path=res_path, eval_set=eval_set_map[self.dataset_cfg.VERSION], output_dir=str(output_path), verbose=True, ) metrics_summary = nusc_eval.main(plot_examples=0, render_curves=False) with open(output_path / 'metrics_summary.json', 'r') as f: metrics = json.load(f) result_str, result_dict = nuscenes_utils.format_nuscene_results(metrics, self.class_names, version=eval_version) return result_str, result_dict def create_groundtruth_database(self, used_classes=None, max_sweeps=10): import torch database_save_path = self.root_path / f'gt_database_{max_sweeps}sweeps_withvelo' db_info_save_path = self.root_path / f'nuscenes_dbinfos_{max_sweeps}sweeps_withvelo.pkl' database_save_path.mkdir(parents=True, exist_ok=True) all_db_infos = {} for idx in tqdm(range(len(self.infos))): sample_idx = idx info = self.infos[idx] points = self.get_lidar_with_sweeps(idx, max_sweeps=max_sweeps) gt_boxes = info['gt_boxes'] gt_names = info['gt_names'] box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( torch.from_numpy(points[:, 0:3]).unsqueeze(dim=0).float().cuda(), torch.from_numpy(gt_boxes[:, 0:7]).unsqueeze(dim=0).float().cuda() ).long().squeeze(dim=0).cpu().numpy() for i in range(gt_boxes.shape[0]): filename = '%s_%s_%d.bin' % (sample_idx, gt_names[i], i) filepath = database_save_path / filename gt_points = points[box_idxs_of_pts == i] gt_points[:, :3] -= gt_boxes[i, :3] with open(filepath, 'w') as f: gt_points.tofile(f) if (used_classes is None) or gt_names[i] in used_classes: db_path = str(filepath.relative_to(self.root_path)) # gt_database/xxxxx.bin db_info = {'name': gt_names[i], 'path': db_path, 'image_idx': sample_idx, 'gt_idx': i, 'box3d_lidar': gt_boxes[i], 'num_points_in_gt': gt_points.shape[0]} if gt_names[i] in all_db_infos: all_db_infos[gt_names[i]].append(db_info) else: all_db_infos[gt_names[i]] = [db_info] for k, v in all_db_infos.items(): print('Database %s: %d' % (k, len(v))) with open(db_info_save_path, 'wb') as f: pickle.dump(all_db_infos, f) def create_nuscenes_info(version, data_path, save_path, max_sweeps=10): from nuscenes.nuscenes import NuScenes from nuscenes.utils import splits from . import nuscenes_utils data_path = data_path / version save_path = save_path / version assert version in ['v1.0-trainval', 'v1.0-test', 'v1.0-mini'] if version == 'v1.0-trainval': train_scenes = splits.train val_scenes = splits.val elif version == 'v1.0-test': train_scenes = splits.test val_scenes = [] elif version == 'v1.0-mini': train_scenes = splits.mini_train val_scenes = splits.mini_val else: raise NotImplementedError nusc = NuScenes(version=version, dataroot=data_path, verbose=True) available_scenes = nuscenes_utils.get_available_scenes(nusc) available_scene_names = [s['name'] for s in available_scenes] train_scenes = list(filter(lambda x: x in available_scene_names, train_scenes)) val_scenes = list(filter(lambda x: x in available_scene_names, val_scenes)) train_scenes = set([available_scenes[available_scene_names.index(s)]['token'] for s in train_scenes]) val_scenes = set([available_scenes[available_scene_names.index(s)]['token'] for s in val_scenes]) print('%s: train scene(%d), val scene(%d)' % (version, len(train_scenes), len(val_scenes))) train_nusc_infos, val_nusc_infos = nuscenes_utils.fill_trainval_infos( data_path=data_path, nusc=nusc, train_scenes=train_scenes, val_scenes=val_scenes, test='test' in version, max_sweeps=max_sweeps ) if version == 'v1.0-test': print('test sample: %d' % len(train_nusc_infos)) with open(save_path / f'nuscenes_infos_{max_sweeps}sweeps_test.pkl', 'wb') as f: pickle.dump(train_nusc_infos, f) else: print('train sample: %d, val sample: %d' % (len(train_nusc_infos), len(val_nusc_infos))) with open(save_path / f'nuscenes_infos_{max_sweeps}sweeps_train.pkl', 'wb') as f: pickle.dump(train_nusc_infos, f) with open(save_path / f'nuscenes_infos_{max_sweeps}sweeps_val.pkl', 'wb') as f: pickle.dump(val_nusc_infos, f) if __name__ == '__main__': import yaml import argparse from pathlib import Path from easydict import EasyDict parser = argparse.ArgumentParser(description='arg parser') parser.add_argument('--cfg_file', type=str, default=None, help='specify the config of dataset') parser.add_argument('--func', type=str, default='create_nuscenes_infos', help='') parser.add_argument('--version', type=str, default='v1.0-trainval', help='') args = parser.parse_args() if args.func == 'create_nuscenes_infos': dataset_cfg = EasyDict(yaml.load(open(args.cfg_file))) ROOT_DIR = (Path(__file__).resolve().parent / '../../../').resolve() dataset_cfg.VERSION = args.version create_nuscenes_info( version=dataset_cfg.VERSION, data_path=ROOT_DIR / 'data' / 'nuscenes', save_path=ROOT_DIR / 'data' / 'nuscenes', max_sweeps=dataset_cfg.MAX_SWEEPS, ) nuscenes_dataset = NuScenesDataset( dataset_cfg=dataset_cfg, class_names=None, root_path=ROOT_DIR / 'data' / 'nuscenes', logger=common_utils.create_logger(), training=True ) nuscenes_dataset.create_groundtruth_database(max_sweeps=dataset_cfg.MAX_SWEEPS)
15,322
39.861333
120
py
SASA
SASA-main/pcdet/datasets/kitti/kitti_dataset.py
import copy import pickle import numpy as np from skimage import io from ...ops.roiaware_pool3d import roiaware_pool3d_utils from ...utils import box_utils, calibration_kitti, common_utils, object3d_kitti from ..dataset import DatasetTemplate class KittiDataset(DatasetTemplate): def __init__(self, dataset_cfg, class_names, training=True, root_path=None, logger=None): """ Args: root_path: dataset_cfg: class_names: training: logger: """ super().__init__( dataset_cfg=dataset_cfg, class_names=class_names, training=training, root_path=root_path, logger=logger ) self.split = self.dataset_cfg.DATA_SPLIT[self.mode] self.root_split_path = self.root_path / ('training' if self.split != 'test' else 'testing') split_dir = self.root_path / 'ImageSets' / (self.split + '.txt') self.sample_id_list = [x.strip() for x in open(split_dir).readlines()] if split_dir.exists() else None self.kitti_infos = [] self.include_kitti_data(self.mode) def include_kitti_data(self, mode): if self.logger is not None: self.logger.info('Loading KITTI dataset') kitti_infos = [] for info_path in self.dataset_cfg.INFO_PATH[mode]: info_path = self.root_path / info_path if not info_path.exists(): continue with open(info_path, 'rb') as f: infos = pickle.load(f) kitti_infos.extend(infos) self.kitti_infos.extend(kitti_infos) if self.logger is not None: self.logger.info('Total samples for KITTI dataset: %d' % (len(kitti_infos))) def set_split(self, split): super().__init__( dataset_cfg=self.dataset_cfg, class_names=self.class_names, training=self.training, root_path=self.root_path, logger=self.logger ) self.split = split self.root_split_path = self.root_path / ('training' if self.split != 'test' else 'testing') split_dir = self.root_path / 'ImageSets' / (self.split + '.txt') self.sample_id_list = [x.strip() for x in open(split_dir).readlines()] if split_dir.exists() else None def get_lidar(self, idx): lidar_file = self.root_split_path / 'velodyne' / ('%s.bin' % idx) assert lidar_file.exists() return np.fromfile(str(lidar_file), dtype=np.float32).reshape(-1, 4) def get_image_shape(self, idx): img_file = self.root_split_path / 'image_2' / ('%s.png' % idx) assert img_file.exists() return np.array(io.imread(img_file).shape[:2], dtype=np.int32) def get_label(self, idx): label_file = self.root_split_path / 'label_2' / ('%s.txt' % idx) assert label_file.exists() return object3d_kitti.get_objects_from_label(label_file) def get_calib(self, idx): calib_file = self.root_split_path / 'calib' / ('%s.txt' % idx) assert calib_file.exists() return calibration_kitti.Calibration(calib_file) def get_road_plane(self, idx): plane_file = self.root_split_path / 'planes' / ('%s.txt' % idx) if not plane_file.exists(): return None with open(plane_file, 'r') as f: lines = f.readlines() lines = [float(i) for i in lines[3].split()] plane = np.asarray(lines) # Ensure normal is always facing up, this is in the rectified camera coordinate if plane[1] > 0: plane = -plane norm = np.linalg.norm(plane[0:3]) plane = plane / norm return plane @staticmethod def get_fov_flag(pts_rect, img_shape, calib): """ Args: pts_rect: img_shape: calib: Returns: """ pts_img, pts_rect_depth = calib.rect_to_img(pts_rect) val_flag_1 = np.logical_and(pts_img[:, 0] >= 0, pts_img[:, 0] < img_shape[1]) val_flag_2 = np.logical_and(pts_img[:, 1] >= 0, pts_img[:, 1] < img_shape[0]) val_flag_merge = np.logical_and(val_flag_1, val_flag_2) pts_valid_flag = np.logical_and(val_flag_merge, pts_rect_depth >= 0) return pts_valid_flag def get_infos(self, num_workers=4, has_label=True, count_inside_pts=True, sample_id_list=None): import concurrent.futures as futures def process_single_scene(sample_idx): print('%s sample_idx: %s' % (self.split, sample_idx)) info = {} pc_info = {'num_features': 4, 'lidar_idx': sample_idx} info['point_cloud'] = pc_info image_info = {'image_idx': sample_idx, 'image_shape': self.get_image_shape(sample_idx)} info['image'] = image_info calib = self.get_calib(sample_idx) P2 = np.concatenate([calib.P2, np.array([[0., 0., 0., 1.]])], axis=0) R0_4x4 = np.zeros([4, 4], dtype=calib.R0.dtype) R0_4x4[3, 3] = 1. R0_4x4[:3, :3] = calib.R0 V2C_4x4 = np.concatenate([calib.V2C, np.array([[0., 0., 0., 1.]])], axis=0) calib_info = {'P2': P2, 'R0_rect': R0_4x4, 'Tr_velo_to_cam': V2C_4x4} info['calib'] = calib_info if has_label: obj_list = self.get_label(sample_idx) annotations = {} annotations['name'] = np.array([obj.cls_type for obj in obj_list]) annotations['truncated'] = np.array([obj.truncation for obj in obj_list]) annotations['occluded'] = np.array([obj.occlusion for obj in obj_list]) annotations['alpha'] = np.array([obj.alpha for obj in obj_list]) annotations['bbox'] = np.concatenate([obj.box2d.reshape(1, 4) for obj in obj_list], axis=0) annotations['dimensions'] = np.array([[obj.l, obj.h, obj.w] for obj in obj_list]) # lhw(camera) format annotations['location'] = np.concatenate([obj.loc.reshape(1, 3) for obj in obj_list], axis=0) annotations['rotation_y'] = np.array([obj.ry for obj in obj_list]) annotations['score'] = np.array([obj.score for obj in obj_list]) annotations['difficulty'] = np.array([obj.level for obj in obj_list], np.int32) num_objects = len([obj.cls_type for obj in obj_list if obj.cls_type != 'DontCare']) num_gt = len(annotations['name']) index = list(range(num_objects)) + [-1] * (num_gt - num_objects) annotations['index'] = np.array(index, dtype=np.int32) loc = annotations['location'][:num_objects] dims = annotations['dimensions'][:num_objects] rots = annotations['rotation_y'][:num_objects] loc_lidar = calib.rect_to_lidar(loc) l, h, w = dims[:, 0:1], dims[:, 1:2], dims[:, 2:3] loc_lidar[:, 2] += h[:, 0] / 2 gt_boxes_lidar = np.concatenate([loc_lidar, l, w, h, -(np.pi / 2 + rots[..., np.newaxis])], axis=1) annotations['gt_boxes_lidar'] = gt_boxes_lidar info['annos'] = annotations if count_inside_pts: points = self.get_lidar(sample_idx) calib = self.get_calib(sample_idx) pts_rect = calib.lidar_to_rect(points[:, 0:3]) fov_flag = self.get_fov_flag(pts_rect, info['image']['image_shape'], calib) pts_fov = points[fov_flag] corners_lidar = box_utils.boxes_to_corners_3d(gt_boxes_lidar) num_points_in_gt = -np.ones(num_gt, dtype=np.int32) for k in range(num_objects): flag = box_utils.in_hull(pts_fov[:, 0:3], corners_lidar[k]) num_points_in_gt[k] = flag.sum() annotations['num_points_in_gt'] = num_points_in_gt return info sample_id_list = sample_id_list if sample_id_list is not None else self.sample_id_list with futures.ThreadPoolExecutor(num_workers) as executor: infos = executor.map(process_single_scene, sample_id_list) return list(infos) def create_groundtruth_database(self, info_path=None, used_classes=None, split='train'): import torch database_save_path = Path(self.root_path) / ('gt_database' if split == 'train' else ('gt_database_%s' % split)) db_info_save_path = Path(self.root_path) / ('kitti_dbinfos_%s.pkl' % split) database_save_path.mkdir(parents=True, exist_ok=True) all_db_infos = {} with open(info_path, 'rb') as f: infos = pickle.load(f) for k in range(len(infos)): print('gt_database sample: %d/%d' % (k + 1, len(infos))) info = infos[k] sample_idx = info['point_cloud']['lidar_idx'] points = self.get_lidar(sample_idx) annos = info['annos'] names = annos['name'] difficulty = annos['difficulty'] bbox = annos['bbox'] gt_boxes = annos['gt_boxes_lidar'] num_obj = gt_boxes.shape[0] point_indices = roiaware_pool3d_utils.points_in_boxes_cpu( torch.from_numpy(points[:, 0:3]), torch.from_numpy(gt_boxes) ).numpy() # (nboxes, npoints) for i in range(num_obj): filename = '%s_%s_%d.bin' % (sample_idx, names[i], i) filepath = database_save_path / filename gt_points = points[point_indices[i] > 0] gt_points[:, :3] -= gt_boxes[i, :3] with open(filepath, 'w') as f: gt_points.tofile(f) if (used_classes is None) or names[i] in used_classes: db_path = str(filepath.relative_to(self.root_path)) # gt_database/xxxxx.bin db_info = {'name': names[i], 'path': db_path, 'image_idx': sample_idx, 'gt_idx': i, 'box3d_lidar': gt_boxes[i], 'num_points_in_gt': gt_points.shape[0], 'difficulty': difficulty[i], 'bbox': bbox[i], 'score': annos['score'][i]} if names[i] in all_db_infos: all_db_infos[names[i]].append(db_info) else: all_db_infos[names[i]] = [db_info] for k, v in all_db_infos.items(): print('Database %s: %d' % (k, len(v))) with open(db_info_save_path, 'wb') as f: pickle.dump(all_db_infos, f) @staticmethod def generate_prediction_dicts(batch_dict, pred_dicts, class_names, output_path=None): """ Args: batch_dict: frame_id: pred_dicts: list of pred_dicts pred_boxes: (N, 7), Tensor pred_scores: (N), Tensor pred_labels: (N), Tensor class_names: output_path: Returns: """ def get_template_prediction(num_samples): ret_dict = { 'name': np.zeros(num_samples), 'truncated': np.zeros(num_samples), 'occluded': np.zeros(num_samples), 'alpha': np.zeros(num_samples), 'bbox': np.zeros([num_samples, 4]), 'dimensions': np.zeros([num_samples, 3]), 'location': np.zeros([num_samples, 3]), 'rotation_y': np.zeros(num_samples), 'score': np.zeros(num_samples), 'boxes_lidar': np.zeros([num_samples, 7]) } return ret_dict def generate_single_sample_dict(batch_index, box_dict): pred_scores = box_dict['pred_scores'].cpu().numpy() pred_boxes = box_dict['pred_boxes'].cpu().numpy() pred_labels = box_dict['pred_labels'].cpu().numpy() pred_dict = get_template_prediction(pred_scores.shape[0]) if pred_scores.shape[0] == 0: return pred_dict calib = batch_dict['calib'][batch_index] image_shape = batch_dict['image_shape'][batch_index] pred_boxes_camera = box_utils.boxes3d_lidar_to_kitti_camera(pred_boxes, calib) pred_boxes_img = box_utils.boxes3d_kitti_camera_to_imageboxes( pred_boxes_camera, calib, image_shape=image_shape ) pred_dict['name'] = np.array(class_names)[pred_labels - 1] pred_dict['alpha'] = -np.arctan2(-pred_boxes[:, 1], pred_boxes[:, 0]) + pred_boxes_camera[:, 6] pred_dict['bbox'] = pred_boxes_img pred_dict['dimensions'] = pred_boxes_camera[:, 3:6] pred_dict['location'] = pred_boxes_camera[:, 0:3] pred_dict['rotation_y'] = pred_boxes_camera[:, 6] pred_dict['score'] = pred_scores pred_dict['boxes_lidar'] = pred_boxes return pred_dict annos = [] for index, box_dict in enumerate(pred_dicts): frame_id = batch_dict['frame_id'][index] single_pred_dict = generate_single_sample_dict(index, box_dict) single_pred_dict['frame_id'] = frame_id annos.append(single_pred_dict) if output_path is not None: cur_det_file = output_path / ('%s.txt' % frame_id) with open(cur_det_file, 'w') as f: bbox = single_pred_dict['bbox'] loc = single_pred_dict['location'] dims = single_pred_dict['dimensions'] # lhw -> hwl for idx in range(len(bbox)): print('%s -1 -1 %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f %.4f' % (single_pred_dict['name'][idx], single_pred_dict['alpha'][idx], bbox[idx][0], bbox[idx][1], bbox[idx][2], bbox[idx][3], dims[idx][1], dims[idx][2], dims[idx][0], loc[idx][0], loc[idx][1], loc[idx][2], single_pred_dict['rotation_y'][idx], single_pred_dict['score'][idx]), file=f) return annos def evaluation(self, det_annos, class_names, **kwargs): if 'annos' not in self.kitti_infos[0].keys(): return None, {} from .kitti_object_eval_python import eval as kitti_eval eval_det_annos = copy.deepcopy(det_annos) eval_gt_annos = [copy.deepcopy(info['annos']) for info in self.kitti_infos] ap_result_str, ap_dict = kitti_eval.get_official_eval_result(eval_gt_annos, eval_det_annos, class_names) return ap_result_str, ap_dict def __len__(self): if self._merge_all_iters_to_one_epoch: return len(self.kitti_infos) * self.total_epochs return len(self.kitti_infos) def __getitem__(self, index): # index = 4 if self._merge_all_iters_to_one_epoch: index = index % len(self.kitti_infos) info = copy.deepcopy(self.kitti_infos[index]) sample_idx = info['point_cloud']['lidar_idx'] points = self.get_lidar(sample_idx) calib = self.get_calib(sample_idx) img_shape = info['image']['image_shape'] if self.dataset_cfg.FOV_POINTS_ONLY: pts_rect = calib.lidar_to_rect(points[:, 0:3]) fov_flag = self.get_fov_flag(pts_rect, img_shape, calib) points = points[fov_flag] input_dict = { 'points': points, 'frame_id': sample_idx, 'calib': calib, } if 'annos' in info: annos = info['annos'] annos = common_utils.drop_info_with_name(annos, name='DontCare') loc, dims, rots = annos['location'], annos['dimensions'], annos['rotation_y'] gt_names = annos['name'] gt_boxes_camera = np.concatenate([loc, dims, rots[..., np.newaxis]], axis=1).astype(np.float32) gt_boxes_lidar = box_utils.boxes3d_kitti_camera_to_lidar(gt_boxes_camera, calib) input_dict.update({ 'gt_names': gt_names, 'gt_boxes': gt_boxes_lidar }) road_plane = self.get_road_plane(sample_idx) if road_plane is not None: input_dict['road_plane'] = road_plane data_dict = self.prepare_data(data_dict=input_dict) data_dict['image_shape'] = img_shape return data_dict def create_kitti_infos(dataset_cfg, class_names, data_path, save_path, workers=4): dataset = KittiDataset(dataset_cfg=dataset_cfg, class_names=class_names, root_path=data_path, training=False) train_split, val_split, trainval_split = 'train', 'val', 'trainval' train_filename = save_path / ('kitti_infos_%s.pkl' % train_split) val_filename = save_path / ('kitti_infos_%s.pkl' % val_split) trainval_filename = save_path / 'kitti_infos_trainval.pkl' test_filename = save_path / 'kitti_infos_test.pkl' print('---------------Start to generate data infos---------------') dataset.set_split(train_split) kitti_infos_train = dataset.get_infos(num_workers=workers, has_label=True, count_inside_pts=True) with open(train_filename, 'wb') as f: pickle.dump(kitti_infos_train, f) print('Kitti info train file is saved to %s' % train_filename) dataset.set_split(val_split) kitti_infos_val = dataset.get_infos(num_workers=workers, has_label=True, count_inside_pts=True) with open(val_filename, 'wb') as f: pickle.dump(kitti_infos_val, f) print('Kitti info val file is saved to %s' % val_filename) with open(trainval_filename, 'wb') as f: pickle.dump(kitti_infos_train + kitti_infos_val, f) print('Kitti info trainval file is saved to %s' % trainval_filename) dataset.set_split('test') kitti_infos_test = dataset.get_infos(num_workers=workers, has_label=False, count_inside_pts=False) with open(test_filename, 'wb') as f: pickle.dump(kitti_infos_test, f) print('Kitti info test file is saved to %s' % test_filename) print('---------------Start create groundtruth database for data augmentation---------------') dataset.set_split(train_split) dataset.create_groundtruth_database(train_filename, split=train_split) dataset.set_split(trainval_split) dataset.create_groundtruth_database(trainval_filename, split=trainval_split) print('---------------Data preparation Done---------------') if __name__ == '__main__': import sys if sys.argv.__len__() > 1 and sys.argv[1] == 'create_kitti_infos': import yaml from pathlib import Path from easydict import EasyDict dataset_cfg = EasyDict(yaml.load(open(sys.argv[2]))) ROOT_DIR = (Path(__file__).resolve().parent / '../../../').resolve() create_kitti_infos( dataset_cfg=dataset_cfg, class_names=['Car', 'Pedestrian', 'Cyclist'], data_path=ROOT_DIR / 'data' / 'kitti', save_path=ROOT_DIR / 'data' / 'kitti' )
19,046
41.995485
140
py
SASA
SASA-main/pcdet/utils/box_utils.py
import numpy as np import scipy import torch import copy from scipy.spatial import Delaunay from ..ops.roiaware_pool3d import roiaware_pool3d_utils from . import common_utils def in_hull(p, hull): """ :param p: (N, K) test points :param hull: (M, K) M corners of a box :return (N) bool """ try: if not isinstance(hull, Delaunay): hull = Delaunay(hull) flag = hull.find_simplex(p) >= 0 except scipy.spatial.qhull.QhullError: print('Warning: not a hull %s' % str(hull)) flag = np.zeros(p.shape[0], dtype=np.bool) return flag def boxes_to_corners_3d(boxes3d): """ 7 -------- 4 /| /| 6 -------- 5 . | | | | . 3 -------- 0 |/ |/ 2 -------- 1 Args: boxes3d: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center Returns: """ boxes3d, is_numpy = common_utils.check_numpy_to_torch(boxes3d) template = boxes3d.new_tensor(( [1, 1, -1], [1, -1, -1], [-1, -1, -1], [-1, 1, -1], [1, 1, 1], [1, -1, 1], [-1, -1, 1], [-1, 1, 1], )) / 2 corners3d = boxes3d[:, None, 3:6].repeat(1, 8, 1) * template[None, :, :] corners3d = common_utils.rotate_points_along_z(corners3d.view(-1, 8, 3), boxes3d[:, 6]).view(-1, 8, 3) corners3d += boxes3d[:, None, 0:3] return corners3d.numpy() if is_numpy else corners3d def mask_boxes_outside_range_numpy(boxes, limit_range, min_num_corners=1): """ Args: boxes: (N, 7) [x, y, z, dx, dy, dz, heading, ...], (x, y, z) is the box center limit_range: [minx, miny, minz, maxx, maxy, maxz] min_num_corners: Returns: """ if boxes.shape[1] > 7: boxes = boxes[:, 0:7] corners = boxes_to_corners_3d(boxes) # (N, 8, 3) mask = ((corners >= limit_range[0:3]) & (corners <= limit_range[3:6])).all(axis=2) mask = mask.sum(axis=1) >= min_num_corners # (N) return mask def remove_points_in_boxes3d(points, boxes3d): """ Args: points: (num_points, 3 + C) boxes3d: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center, each box DO NOT overlaps Returns: """ boxes3d, is_numpy = common_utils.check_numpy_to_torch(boxes3d) points, is_numpy = common_utils.check_numpy_to_torch(points) point_masks = roiaware_pool3d_utils.points_in_boxes_cpu(points[:, 0:3], boxes3d) points = points[point_masks.sum(dim=0) == 0] return points.numpy() if is_numpy else points def boxes3d_kitti_camera_to_lidar(boxes3d_camera, calib): """ Args: boxes3d_camera: (N, 7) [x, y, z, l, h, w, r] in rect camera coords calib: Returns: boxes3d_lidar: [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center """ boxes3d_camera_copy = copy.deepcopy(boxes3d_camera) xyz_camera, r = boxes3d_camera_copy[:, 0:3], boxes3d_camera_copy[:, 6:7] l, h, w = boxes3d_camera_copy[:, 3:4], boxes3d_camera_copy[:, 4:5], boxes3d_camera_copy[:, 5:6] xyz_lidar = calib.rect_to_lidar(xyz_camera) xyz_lidar[:, 2] += h[:, 0] / 2 return np.concatenate([xyz_lidar, l, w, h, -(r + np.pi / 2)], axis=-1) def boxes3d_kitti_fakelidar_to_lidar(boxes3d_lidar): """ Args: boxes3d_fakelidar: (N, 7) [x, y, z, w, l, h, r] in old LiDAR coordinates, z is bottom center Returns: boxes3d_lidar: [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center """ boxes3d_lidar_copy = copy.deepcopy(boxes3d_lidar) w, l, h = boxes3d_lidar_copy[:, 3:4], boxes3d_lidar_copy[:, 4:5], boxes3d_lidar_copy[:, 5:6] r = boxes3d_lidar_copy[:, 6:7] boxes3d_lidar_copy[:, 2] += h[:, 0] / 2 return np.concatenate([boxes3d_lidar_copy[:, 0:3], l, w, h, -(r + np.pi / 2)], axis=-1) def boxes3d_kitti_lidar_to_fakelidar(boxes3d_lidar): """ Args: boxes3d_lidar: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center Returns: boxes3d_fakelidar: [x, y, z, w, l, h, r] in old LiDAR coordinates, z is bottom center """ boxes3d_lidar_copy = copy.deepcopy(boxes3d_lidar) dx, dy, dz = boxes3d_lidar_copy[:, 3:4], boxes3d_lidar_copy[:, 4:5], boxes3d_lidar_copy[:, 5:6] heading = boxes3d_lidar_copy[:, 6:7] boxes3d_lidar_copy[:, 2] -= dz[:, 0] / 2 return np.concatenate([boxes3d_lidar_copy[:, 0:3], dy, dx, dz, -heading - np.pi / 2], axis=-1) def enlarge_box3d(boxes3d, extra_width=(0, 0, 0)): """ Args: boxes3d: [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center extra_width: [extra_x, extra_y, extra_z] Returns: """ boxes3d, is_numpy = common_utils.check_numpy_to_torch(boxes3d) large_boxes3d = boxes3d.clone() large_boxes3d[:, 3:6] += boxes3d.new_tensor(extra_width)[None, :] return large_boxes3d def boxes3d_lidar_to_kitti_camera(boxes3d_lidar, calib): """ :param boxes3d_lidar: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center :param calib: :return: boxes3d_camera: (N, 7) [x, y, z, l, h, w, r] in rect camera coords """ boxes3d_lidar_copy = copy.deepcopy(boxes3d_lidar) xyz_lidar = boxes3d_lidar_copy[:, 0:3] l, w, h = boxes3d_lidar_copy[:, 3:4], boxes3d_lidar_copy[:, 4:5], boxes3d_lidar_copy[:, 5:6] r = boxes3d_lidar_copy[:, 6:7] xyz_lidar[:, 2] -= h.reshape(-1) / 2 xyz_cam = calib.lidar_to_rect(xyz_lidar) # xyz_cam[:, 1] += h.reshape(-1) / 2 r = -r - np.pi / 2 return np.concatenate([xyz_cam, l, h, w, r], axis=-1) def boxes3d_to_corners3d_kitti_camera(boxes3d, bottom_center=True): """ :param boxes3d: (N, 7) [x, y, z, l, h, w, ry] in camera coords, see the definition of ry in KITTI dataset :param bottom_center: whether y is on the bottom center of object :return: corners3d: (N, 8, 3) 7 -------- 4 /| /| 6 -------- 5 . | | | | . 3 -------- 0 |/ |/ 2 -------- 1 """ boxes_num = boxes3d.shape[0] l, h, w = boxes3d[:, 3], boxes3d[:, 4], boxes3d[:, 5] x_corners = np.array([l / 2., l / 2., -l / 2., -l / 2., l / 2., l / 2., -l / 2., -l / 2], dtype=np.float32).T z_corners = np.array([w / 2., -w / 2., -w / 2., w / 2., w / 2., -w / 2., -w / 2., w / 2.], dtype=np.float32).T if bottom_center: y_corners = np.zeros((boxes_num, 8), dtype=np.float32) y_corners[:, 4:8] = -h.reshape(boxes_num, 1).repeat(4, axis=1) # (N, 8) else: y_corners = np.array([h / 2., h / 2., h / 2., h / 2., -h / 2., -h / 2., -h / 2., -h / 2.], dtype=np.float32).T ry = boxes3d[:, 6] zeros, ones = np.zeros(ry.size, dtype=np.float32), np.ones(ry.size, dtype=np.float32) rot_list = np.array([[np.cos(ry), zeros, -np.sin(ry)], [zeros, ones, zeros], [np.sin(ry), zeros, np.cos(ry)]]) # (3, 3, N) R_list = np.transpose(rot_list, (2, 0, 1)) # (N, 3, 3) temp_corners = np.concatenate((x_corners.reshape(-1, 8, 1), y_corners.reshape(-1, 8, 1), z_corners.reshape(-1, 8, 1)), axis=2) # (N, 8, 3) rotated_corners = np.matmul(temp_corners, R_list) # (N, 8, 3) x_corners, y_corners, z_corners = rotated_corners[:, :, 0], rotated_corners[:, :, 1], rotated_corners[:, :, 2] x_loc, y_loc, z_loc = boxes3d[:, 0], boxes3d[:, 1], boxes3d[:, 2] x = x_loc.reshape(-1, 1) + x_corners.reshape(-1, 8) y = y_loc.reshape(-1, 1) + y_corners.reshape(-1, 8) z = z_loc.reshape(-1, 1) + z_corners.reshape(-1, 8) corners = np.concatenate((x.reshape(-1, 8, 1), y.reshape(-1, 8, 1), z.reshape(-1, 8, 1)), axis=2) return corners.astype(np.float32) def boxes3d_kitti_camera_to_imageboxes(boxes3d, calib, image_shape=None): """ :param boxes3d: (N, 7) [x, y, z, l, h, w, r] in rect camera coords :param calib: :return: box_2d_preds: (N, 4) [x1, y1, x2, y2] """ corners3d = boxes3d_to_corners3d_kitti_camera(boxes3d) pts_img, _ = calib.rect_to_img(corners3d.reshape(-1, 3)) corners_in_image = pts_img.reshape(-1, 8, 2) min_uv = np.min(corners_in_image, axis=1) # (N, 2) max_uv = np.max(corners_in_image, axis=1) # (N, 2) boxes2d_image = np.concatenate([min_uv, max_uv], axis=1) if image_shape is not None: boxes2d_image[:, 0] = np.clip(boxes2d_image[:, 0], a_min=0, a_max=image_shape[1] - 1) boxes2d_image[:, 1] = np.clip(boxes2d_image[:, 1], a_min=0, a_max=image_shape[0] - 1) boxes2d_image[:, 2] = np.clip(boxes2d_image[:, 2], a_min=0, a_max=image_shape[1] - 1) boxes2d_image[:, 3] = np.clip(boxes2d_image[:, 3], a_min=0, a_max=image_shape[0] - 1) return boxes2d_image def boxes_iou_normal(boxes_a, boxes_b): """ Args: boxes_a: (N, 4) [x1, y1, x2, y2] boxes_b: (M, 4) [x1, y1, x2, y2] Returns: """ assert boxes_a.shape[1] == boxes_b.shape[1] == 4 x_min = torch.max(boxes_a[:, 0, None], boxes_b[None, :, 0]) x_max = torch.min(boxes_a[:, 2, None], boxes_b[None, :, 2]) y_min = torch.max(boxes_a[:, 1, None], boxes_b[None, :, 1]) y_max = torch.min(boxes_a[:, 3, None], boxes_b[None, :, 3]) x_len = torch.clamp_min(x_max - x_min, min=0) y_len = torch.clamp_min(y_max - y_min, min=0) area_a = (boxes_a[:, 2] - boxes_a[:, 0]) * (boxes_a[:, 3] - boxes_a[:, 1]) area_b = (boxes_b[:, 2] - boxes_b[:, 0]) * (boxes_b[:, 3] - boxes_b[:, 1]) a_intersect_b = x_len * y_len iou = a_intersect_b / torch.clamp_min(area_a[:, None] + area_b[None, :] - a_intersect_b, min=1e-6) return iou def boxes3d_lidar_to_aligned_bev_boxes(boxes3d): """ Args: boxes3d: (N, 7 + C) [x, y, z, dx, dy, dz, heading] in lidar coordinate Returns: aligned_bev_boxes: (N, 4) [x1, y1, x2, y2] in the above lidar coordinate """ rot_angle = common_utils.limit_period(boxes3d[:, 6], offset=0.5, period=np.pi).abs() choose_dims = torch.where(rot_angle[:, None] < np.pi / 4, boxes3d[:, [3, 4]], boxes3d[:, [4, 3]]) aligned_bev_boxes = torch.cat((boxes3d[:, 0:2] - choose_dims / 2, boxes3d[:, 0:2] + choose_dims / 2), dim=1) return aligned_bev_boxes def boxes3d_nearest_bev_iou(boxes_a, boxes_b): """ Args: boxes_a: (N, 7) [x, y, z, dx, dy, dz, heading] boxes_b: (N, 7) [x, y, z, dx, dy, dz, heading] Returns: """ boxes_bev_a = boxes3d_lidar_to_aligned_bev_boxes(boxes_a) boxes_bev_b = boxes3d_lidar_to_aligned_bev_boxes(boxes_b) return boxes_iou_normal(boxes_bev_a, boxes_bev_b)
10,568
34.466443
118
py
SASA
SASA-main/pcdet/utils/loss_utils.py
import numpy as np import torch import torch.nn as nn import torch.nn.functional as F from . import box_utils from ..ops.roiaware_pool3d import roiaware_pool3d_utils class SigmoidFocalClassificationLoss(nn.Module): """ Sigmoid focal cross entropy loss. """ def __init__(self, gamma: float = 2.0, alpha: float = 0.25): """ Args: gamma: Weighting parameter to balance loss for hard and easy examples. alpha: Weighting parameter to balance loss for positive and negative examples. """ super(SigmoidFocalClassificationLoss, self).__init__() self.alpha = alpha self.gamma = gamma @staticmethod def sigmoid_cross_entropy_with_logits(input: torch.Tensor, target: torch.Tensor): """ PyTorch Implementation for tf.nn.sigmoid_cross_entropy_with_logits: max(x, 0) - x * z + log(1 + exp(-abs(x))) in https://www.tensorflow.org/api_docs/python/tf/nn/sigmoid_cross_entropy_with_logits Args: input: (B, #anchors, #classes) float tensor. Predicted logits for each class target: (B, #anchors, #classes) float tensor. One-hot encoded classification targets Returns: loss: (B, #anchors, #classes) float tensor. Sigmoid cross entropy loss without reduction """ loss = torch.clamp(input, min=0) - input * target + \ torch.log1p(torch.exp(-torch.abs(input))) return loss def forward(self, input: torch.Tensor, target: torch.Tensor, weights: torch.Tensor): """ Args: input: (B, #anchors, #classes) float tensor. Predicted logits for each class target: (B, #anchors, #classes) float tensor. One-hot encoded classification targets weights: (B, #anchors) float tensor. Anchor-wise weights. Returns: weighted_loss: (B, #anchors, #classes) float tensor after weighting. """ pred_sigmoid = torch.sigmoid(input) alpha_weight = target * self.alpha + (1 - target) * (1 - self.alpha) pt = target * (1.0 - pred_sigmoid) + (1.0 - target) * pred_sigmoid focal_weight = alpha_weight * torch.pow(pt, self.gamma) bce_loss = self.sigmoid_cross_entropy_with_logits(input, target) loss = focal_weight * bce_loss if weights.shape.__len__() == 2 or \ (weights.shape.__len__() == 1 and target.shape.__len__() == 2): weights = weights.unsqueeze(-1) assert weights.shape.__len__() == loss.shape.__len__() return loss * weights class WeightedSmoothL1Loss(nn.Module): """ Code-wise Weighted Smooth L1 Loss modified based on fvcore.nn.smooth_l1_loss https://github.com/facebookresearch/fvcore/blob/master/fvcore/nn/smooth_l1_loss.py | 0.5 * x ** 2 / beta if abs(x) < beta smoothl1(x) = | | abs(x) - 0.5 * beta otherwise, where x = input - target. """ def __init__(self, beta: float = 1.0 / 9.0, code_weights: list = None): """ Args: beta: Scalar float. L1 to L2 change point. For beta values < 1e-5, L1 loss is computed. code_weights: (#codes) float list if not None. Code-wise weights. """ super(WeightedSmoothL1Loss, self).__init__() self.beta = beta self.code_weights = code_weights if code_weights is not None: self.code_weights = np.array(code_weights, dtype=np.float32) self.code_weights = torch.from_numpy(self.code_weights).cuda() @staticmethod def smooth_l1_loss(diff, beta): if beta < 1e-5: loss = torch.abs(diff) else: n = torch.abs(diff) loss = torch.where(n < beta, 0.5 * n ** 2 / beta, n - 0.5 * beta) return loss def forward(self, input: torch.Tensor, target: torch.Tensor, weights: torch.Tensor = None): """ Args: input: (B, #anchors, #codes) float tensor. Ecoded predicted locations of objects. target: (B, #anchors, #codes) float tensor. Regression targets. weights: (B, #anchors) float tensor if not None. Returns: loss: (B, #anchors) float tensor. Weighted smooth l1 loss without reduction. """ target = torch.where(torch.isnan(target), input, target) # ignore nan targets diff = input - target # code-wise weighting if self.code_weights is not None: diff = diff * self.code_weights.view(1, 1, -1) loss = self.smooth_l1_loss(diff, self.beta) # anchor-wise weighting if weights is not None: assert weights.shape[0] == loss.shape[0] and weights.shape[1] == loss.shape[1] loss = loss * weights.unsqueeze(-1) return loss class WeightedL1Loss(nn.Module): def __init__(self, code_weights: list = None): """ Args: code_weights: (#codes) float list if not None. Code-wise weights. """ super(WeightedL1Loss, self).__init__() self.code_weights = code_weights if code_weights is not None: self.code_weights = np.array(code_weights, dtype=np.float32) self.code_weights = torch.from_numpy(self.code_weights).cuda() def forward(self, input: torch.Tensor, target: torch.Tensor, weights: torch.Tensor = None): """ Args: input: (B, #anchors, #codes) float tensor. Ecoded predicted locations of objects. target: (B, #anchors, #codes) float tensor. Regression targets. weights: (B, #anchors) float tensor if not None. Returns: loss: (B, #anchors) float tensor. Weighted smooth l1 loss without reduction. """ target = torch.where(torch.isnan(target), input, target) # ignore nan targets diff = input - target # code-wise weighting if self.code_weights is not None: diff = diff * self.code_weights.view(1, 1, -1) loss = torch.abs(diff) # anchor-wise weighting if weights is not None: assert weights.shape[0] == loss.shape[0] and weights.shape[1] == loss.shape[1] loss = loss * weights.unsqueeze(-1) return loss class WeightedBinaryCrossEntropyLoss(nn.Module): """ Transform input to fit the fomation of PyTorch offical cross entropy loss with anchor-wise weighting. """ def __init__(self): super(WeightedBinaryCrossEntropyLoss, self).__init__() def forward(self, input: torch.Tensor, target: torch.Tensor, weights: torch.Tensor): """ Args: input: (B, #anchors, #classes) float tensor. Predited logits for each class. target: (B, #anchors, #classes) float tensor. One-hot classification targets. weights: (B, #anchors) float tensor. Anchor-wise weights. Returns: loss: (B, #anchors) float tensor. Weighted cross entropy loss without reduction """ loss = F.binary_cross_entropy_with_logits(input, target, reduction='none').mean(dim=-1) * weights return loss class WeightedCrossEntropyLoss(nn.Module): """ Transform input to fit the fomation of PyTorch offical cross entropy loss with anchor-wise weighting. """ def __init__(self): super(WeightedCrossEntropyLoss, self).__init__() def forward(self, input: torch.Tensor, target: torch.Tensor, weights: torch.Tensor): """ Args: input: (B, #anchors, #classes) float tensor. Predited logits for each class. target: (B, #anchors, #classes) float tensor. One-hot classification targets. weights: (B, #anchors) float tensor. Anchor-wise weights. Returns: loss: (B, #anchors) float tensor. Weighted cross entropy loss without reduction """ input = input.permute(0, 2, 1) target = target.argmax(dim=-1) loss = F.cross_entropy(input, target, reduction='none') * weights return loss def get_corner_loss_lidar(pred_bbox3d: torch.Tensor, gt_bbox3d: torch.Tensor): """ Args: pred_bbox3d: (N, 7) float Tensor. gt_bbox3d: (N, 7) float Tensor. Returns: corner_loss: (N) float Tensor. """ assert pred_bbox3d.shape[0] == gt_bbox3d.shape[0] pred_box_corners = box_utils.boxes_to_corners_3d(pred_bbox3d) gt_box_corners = box_utils.boxes_to_corners_3d(gt_bbox3d) gt_bbox3d_flip = gt_bbox3d.clone() gt_bbox3d_flip[:, 6] += np.pi gt_box_corners_flip = box_utils.boxes_to_corners_3d(gt_bbox3d_flip) # (N, 8) corner_dist = torch.min(torch.norm(pred_box_corners - gt_box_corners, dim=2), torch.norm(pred_box_corners - gt_box_corners_flip, dim=2)) # (N, 8) corner_loss = WeightedSmoothL1Loss.smooth_l1_loss(corner_dist, beta=1.0) return corner_loss.mean(dim=1) class PointSASALoss(nn.Module): """ Layer-wise point segmentation loss, used for SASA. """ def __init__(self, func: str = 'BCE', layer_weights: list = None, extra_width: list = None, set_ignore_flag: bool = False): super(PointSASALoss, self).__init__() self.layer_weights = layer_weights if func == 'BCE': self.loss_func = WeightedBinaryCrossEntropyLoss() elif func == 'Focal': self.loss_func = SigmoidFocalClassificationLoss() else: raise NotImplementedError assert not set_ignore_flag or (set_ignore_flag and extra_width is not None) self.extra_width = extra_width self.set_ignore_flag = set_ignore_flag def assign_target(self, points, gt_boxes): """ Args: points: (N1 + N2 + N3 + ..., 4) [bs_idx, x, y, z] gt_boxes: (B, M, 8) Returns: point_cls_labels: (N1 + N2 + N3 + ...) """ assert len(points.shape) == 2 and points.shape[1] == 4, \ 'points.shape=%s' % str(points.shape) assert len(gt_boxes.shape) == 3, 'gt_boxes.shape=%s' % str(gt_boxes.shape) batch_size = gt_boxes.shape[0] extend_gt_boxes = box_utils.enlarge_box3d( gt_boxes.view(-1, gt_boxes.shape[-1]), extra_width=self.extra_width ).view(batch_size, -1, gt_boxes.shape[-1]) \ if self.extra_width is not None else gt_boxes bs_idx = points[:, 0] point_cls_labels = points.new_zeros(points.shape[0]).long() for k in range(batch_size): bs_mask = (bs_idx == k) points_single = points[bs_mask][:, 1:4] point_cls_labels_single = point_cls_labels.new_zeros(bs_mask.sum()) if not self.set_ignore_flag: box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), extend_gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) else: box_idxs_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) box_fg_flag = (box_idxs_of_pts >= 0) extend_box_idx_of_pts = roiaware_pool3d_utils.points_in_boxes_gpu( points_single.unsqueeze(dim=0), extend_gt_boxes[k:k + 1, :, 0:7].contiguous() ).long().squeeze(dim=0) ignore_flag = box_fg_flag ^ (extend_box_idx_of_pts >= 0) point_cls_labels_single[ignore_flag] = -1 point_cls_labels_single[box_fg_flag] = 1 point_cls_labels[bs_mask] = point_cls_labels_single return point_cls_labels # (N, ) 0: bg, 1: fg, -1: ignore def forward(self, l_points, l_scores, gt_boxes): """ Args: l_points: List of points, [(N, 4): bs_idx, x, y, z] l_scores: List of points, [(N, 1): predicted point scores] gt_boxes: (B, M, 8) Returns: l_labels: List of labels: [(N, 1): assigned segmentation labels] """ l_labels = [] for i in range(len(self.layer_weights)): li_scores = l_scores[i] if li_scores is None or self.layer_weights[i] == 0: l_labels.append(None) continue # binary segmentation labels: 0: bg, 1: fg, -1: ignore li_labels = self.assign_target(l_points[i], gt_boxes) l_labels.append(li_labels) return l_labels def loss_forward(self, l_scores, l_labels): """ Args: l_scores: List of points, [(N, 1): predicted point scores] l_labels: List of points, [(N, 1): assigned segmentation labels] Returns: l_loss: List of segmentation loss """ l_loss = [] for i in range(len(self.layer_weights)): li_scores, li_labels = l_scores[i], l_labels[i] if li_scores is None or li_labels is None: l_loss.append(None) continue positives, negatives = li_labels > 0, li_labels == 0 cls_weights = positives * 1.0 + negatives * 1.0 # (N, 1) pos_normalizer = cls_weights.sum(dim=0).float() one_hot_targets = li_scores.new_zeros( *list(li_labels.shape), 2 ) one_hot_targets.scatter_(-1, (li_labels > 0).long().unsqueeze(-1), 1.0) one_hot_targets = one_hot_targets[:, 1:] # (N, 1) li_loss = self.loss_func(li_scores[None], one_hot_targets[None], cls_weights.reshape(1, -1)) li_loss = self.layer_weights[i] * li_loss.sum() / torch.clamp( pos_normalizer, min=1.0) l_loss.append(li_loss) return l_loss
14,516
35.938931
105
py
SASA
SASA-main/pcdet/utils/box_coder_utils.py
import numpy as np import torch class ResidualCoder(object): def __init__(self, code_size=7, encode_angle_by_sincos=False, **kwargs): super().__init__() self.code_size = code_size self.encode_angle_by_sincos = encode_angle_by_sincos if self.encode_angle_by_sincos: self.code_size += 1 def encode_torch(self, boxes, anchors): """ Args: boxes: (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] anchors: (N, 7 + C) [x, y, z, dx, dy, dz, heading or *[cos, sin], ...] Returns: """ anchors[:, 3:6] = torch.clamp_min(anchors[:, 3:6], min=1e-5) boxes[:, 3:6] = torch.clamp_min(boxes[:, 3:6], min=1e-5) xa, ya, za, dxa, dya, dza, ra, *cas = torch.split(anchors, 1, dim=-1) xg, yg, zg, dxg, dyg, dzg, rg, *cgs = torch.split(boxes, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xt = (xg - xa) / diagonal yt = (yg - ya) / diagonal zt = (zg - za) / dza dxt = torch.log(dxg / dxa) dyt = torch.log(dyg / dya) dzt = torch.log(dzg / dza) if self.encode_angle_by_sincos: rt_cos = torch.cos(rg) - torch.cos(ra) rt_sin = torch.sin(rg) - torch.sin(ra) rts = [rt_cos, rt_sin] else: rts = [rg - ra] cts = [g - a for g, a in zip(cgs, cas)] return torch.cat([xt, yt, zt, dxt, dyt, dzt, *rts, *cts], dim=-1) def decode_torch(self, box_encodings, anchors): """ Args: box_encodings: (B, N, 7 + C) or (N, 7 + C) [x, y, z, dx, dy, dz, heading or *[cos, sin], ...] anchors: (B, N, 7 + C) or (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] Returns: """ xa, ya, za, dxa, dya, dza, ra, *cas = torch.split(anchors, 1, dim=-1) if not self.encode_angle_by_sincos: xt, yt, zt, dxt, dyt, dzt, rt, *cts = torch.split(box_encodings, 1, dim=-1) else: xt, yt, zt, dxt, dyt, dzt, cost, sint, *cts = torch.split(box_encodings, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xg = xt * diagonal + xa yg = yt * diagonal + ya zg = zt * dza + za dxg = torch.exp(dxt) * dxa dyg = torch.exp(dyt) * dya dzg = torch.exp(dzt) * dza if self.encode_angle_by_sincos: rg_cos = cost + torch.cos(ra) rg_sin = sint + torch.sin(ra) rg = torch.atan2(rg_sin, rg_cos) else: rg = rt + ra cgs = [t + a for t, a in zip(cts, cas)] return torch.cat([xg, yg, zg, dxg, dyg, dzg, rg, *cgs], dim=-1) class PreviousResidualDecoder(object): def __init__(self, code_size=7, **kwargs): super().__init__() self.code_size = code_size @staticmethod def decode_torch(box_encodings, anchors): """ Args: box_encodings: (B, N, 7 + ?) x, y, z, w, l, h, r, custom values anchors: (B, N, 7 + C) or (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] Returns: """ xa, ya, za, dxa, dya, dza, ra, *cas = torch.split(anchors, 1, dim=-1) xt, yt, zt, wt, lt, ht, rt, *cts = torch.split(box_encodings, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xg = xt * diagonal + xa yg = yt * diagonal + ya zg = zt * dza + za dxg = torch.exp(lt) * dxa dyg = torch.exp(wt) * dya dzg = torch.exp(ht) * dza rg = rt + ra cgs = [t + a for t, a in zip(cts, cas)] return torch.cat([xg, yg, zg, dxg, dyg, dzg, rg, *cgs], dim=-1) class PreviousResidualRoIDecoder(object): def __init__(self, code_size=7, **kwargs): super().__init__() self.code_size = code_size @staticmethod def decode_torch(box_encodings, anchors): """ Args: box_encodings: (B, N, 7 + ?) x, y, z, w, l, h, r, custom values anchors: (B, N, 7 + C) or (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] Returns: """ xa, ya, za, dxa, dya, dza, ra, *cas = torch.split(anchors, 1, dim=-1) xt, yt, zt, wt, lt, ht, rt, *cts = torch.split(box_encodings, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xg = xt * diagonal + xa yg = yt * diagonal + ya zg = zt * dza + za dxg = torch.exp(lt) * dxa dyg = torch.exp(wt) * dya dzg = torch.exp(ht) * dza rg = ra - rt cgs = [t + a for t, a in zip(cts, cas)] return torch.cat([xg, yg, zg, dxg, dyg, dzg, rg, *cgs], dim=-1) class PointResidualCoder(object): def __init__(self, code_size=8, use_mean_size=True, **kwargs): super().__init__() self.code_size = code_size self.use_mean_size = use_mean_size if self.use_mean_size: self.mean_size = torch.from_numpy(np.array(kwargs['mean_size'])).cuda().float() assert self.mean_size.min() > 0 def encode_torch(self, gt_boxes, points, gt_classes=None): """ Args: gt_boxes: (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] points: (N, 3) [x, y, z] gt_classes: (N) [1, num_classes] Returns: box_coding: (N, 8 + C) """ gt_boxes[:, 3:6] = torch.clamp_min(gt_boxes[:, 3:6], min=1e-5) xg, yg, zg, dxg, dyg, dzg, rg, *cgs = torch.split(gt_boxes, 1, dim=-1) xa, ya, za = torch.split(points, 1, dim=-1) if self.use_mean_size: assert gt_classes.max() <= self.mean_size.shape[0] point_anchor_size = self.mean_size[gt_classes - 1] dxa, dya, dza = torch.split(point_anchor_size, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xt = (xg - xa) / diagonal yt = (yg - ya) / diagonal zt = (zg - za) / dza dxt = torch.log(dxg / dxa) dyt = torch.log(dyg / dya) dzt = torch.log(dzg / dza) else: xt = (xg - xa) yt = (yg - ya) zt = (zg - za) dxt = torch.log(dxg) dyt = torch.log(dyg) dzt = torch.log(dzg) cts = [g for g in cgs] return torch.cat([xt, yt, zt, dxt, dyt, dzt, torch.cos(rg), torch.sin(rg), *cts], dim=-1) def decode_torch(self, box_encodings, points, pred_classes=None): """ Args: box_encodings: (N, 8 + C) [x, y, z, dx, dy, dz, cos, sin, ...] points: [x, y, z] pred_classes: (N) [1, num_classes] Returns: """ xt, yt, zt, dxt, dyt, dzt, cost, sint, *cts = torch.split(box_encodings, 1, dim=-1) xa, ya, za = torch.split(points, 1, dim=-1) if self.use_mean_size: assert pred_classes.max() <= self.mean_size.shape[0] point_anchor_size = self.mean_size[pred_classes - 1] dxa, dya, dza = torch.split(point_anchor_size, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xg = xt * diagonal + xa yg = yt * diagonal + ya zg = zt * dza + za dxg = torch.exp(dxt) * dxa dyg = torch.exp(dyt) * dya dzg = torch.exp(dzt) * dza else: xg = xt + xa yg = yt + ya zg = zt + za dxg, dyg, dzg = torch.split(torch.exp(box_encodings[..., 3:6]), 1, dim=-1) rg = torch.atan2(sint, cost) safe_rg = torch.where(torch.isnan(rg), torch.zeros_like(rg), rg) # avoid possible NaN problem cgs = [t for t in cts] return torch.cat([xg, yg, zg, dxg, dyg, dzg, safe_rg, *cgs], dim=-1) class PointBinResidualCoder(object): def __init__(self, code_size=30, use_mean_size=True, angle_bin_num=12, pred_velo=False, **kwargs): super().__init__() self.code_size = 6 + 2 * angle_bin_num self.angle_bin_num = angle_bin_num self.pred_velo = pred_velo if pred_velo: self.code_size += 2 self.use_mean_size = use_mean_size if self.use_mean_size: self.mean_size = torch.from_numpy(np.array(kwargs['mean_size'])).cuda().float() assert self.mean_size.min() > 0 def encode_angle_torch(self, angle): """ Args: angle: (N) Returns: angle_cls: (N, angle_bin_num) angle_res: (N, angle_bin_num) """ angle = torch.remainder(angle, np.pi * 2.0) angle_per_class = np.pi * 2.0 / float(self.angle_bin_num) shifted_angle = torch.remainder(angle + angle_per_class / 2.0, np.pi * 2.0) angle_cls_f = (shifted_angle / angle_per_class).floor() angle_cls = angle_cls_f.new_zeros(*list(angle_cls_f.shape), self.angle_bin_num) angle_cls.scatter_(-1, angle_cls_f.unsqueeze(-1).long(), 1.0) angle_res = shifted_angle - (angle_cls_f * angle_per_class + angle_per_class / 2.0) angle_res = angle_res / angle_per_class # normalize residual angle to [0, 1] angle_res = angle_cls * angle_res.unsqueeze(-1) return angle_cls, angle_res def decode_angle_torch(self, angle_cls, angle_res): """ Args: angle_cls: (N, angle_bin_num) angle_res: (N, angle_bin_num) Returns: angle: (N) """ angle_cls_idx = angle_cls.argmax(dim=-1) angle_cls_onehot = angle_cls.new_zeros(angle_cls.shape) angle_cls_onehot.scatter_(-1, angle_cls_idx.unsqueeze(-1), 1.0) angle_res = (angle_cls_onehot * angle_res).sum(dim=-1) angle = (angle_cls_idx.float() + angle_res) * (np.pi * 2.0 / float(self.angle_bin_num)) return angle def encode_torch(self, gt_boxes, points, gt_classes=None): """ Args: gt_boxes: (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...] points: (N, 3) [x, y, z] gt_classes: (N) [1, num_classes] Returns: box_coding: (N, 6 + 2 * B + C) """ gt_boxes[:, 3:6] = torch.clamp_min(gt_boxes[:, 3:6], min=1e-5) xg, yg, zg, dxg, dyg, dzg, rg, *cgs = torch.split(gt_boxes, 1, dim=-1) xa, ya, za = torch.split(points, 1, dim=-1) if self.use_mean_size: assert gt_classes.max() <= self.mean_size.shape[0] point_anchor_size = self.mean_size[gt_classes - 1] dxa, dya, dza = torch.split(point_anchor_size, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xt = (xg - xa) / diagonal yt = (yg - ya) / diagonal zt = (zg - za) / dza dxt = torch.log(dxg / dxa) dyt = torch.log(dyg / dya) dzt = torch.log(dzg / dza) else: xt = (xg - xa) yt = (yg - ya) zt = (zg - za) dxt = torch.log(dxg) dyt = torch.log(dyg) dzt = torch.log(dzg) rg_cls, rg_reg = self.encode_angle_torch(rg.squeeze(-1)) cts = [g for g in cgs] return torch.cat([xt, yt, zt, dxt, dyt, dzt, rg_cls, rg_reg, *cts], dim=-1) def decode_torch_kernel(self, box_offsets, box_angle_cls, box_angle_reg, points, pred_classes=None): """ Args: box_offsets: (N, 6) [x, y, z, dx, dy, dz] box_angle_cls: (N, angle_bin_num) box_angle_reg: (N, angle_bin_num) points: [x, y, z] pred_classes: (N) [1, num_classes] Returns: boxes3d: (N, 7) """ xt, yt, zt, dxt, dyt, dzt = torch.split(box_offsets, 1, dim=-1) xa, ya, za = torch.split(points, 1, dim=-1) if self.use_mean_size: assert pred_classes.max() <= self.mean_size.shape[0] point_anchor_size = self.mean_size[pred_classes - 1] dxa, dya, dza = torch.split(point_anchor_size, 1, dim=-1) diagonal = torch.sqrt(dxa ** 2 + dya ** 2) xg = xt * diagonal + xa yg = yt * diagonal + ya zg = zt * dza + za dxg = torch.exp(dxt) * dxa dyg = torch.exp(dyt) * dya dzg = torch.exp(dzt) * dza else: xg = xt + xa yg = yt + ya zg = zt + za dxg = torch.exp(dxt) dyg = torch.exp(dyt) dzg = torch.exp(dzt) rg = self.decode_angle_torch(box_angle_cls, box_angle_reg).unsqueeze(-1) return torch.cat([xg, yg, zg, dxg, dyg, dzg, rg], dim=-1) def decode_torch(self, box_encodings, points, pred_classes=None): """ Args: box_encodings: (N, 8 + C) [x, y, z, dx, dy, dz, cos, sin, ...] points: [x, y, z] pred_classes: (N) [1, num_classes] Returns: boxes3d: (N, 7) """ box_offsets = box_encodings[:, :6] box_angle_cls = box_encodings[:, 6:6 + self.angle_bin_num] box_angle_reg = box_encodings[:, 6 + self.angle_bin_num:6 + self.angle_bin_num * 2] cgs = box_encodings[:, 6 + self.angle_bin_num * 2:] boxes3d = self.decode_torch_kernel(box_offsets, box_angle_cls, box_angle_reg, points, pred_classes) return torch.cat([boxes3d, cgs], dim=-1)
13,258
35.226776
107
py
SASA
SASA-main/pcdet/utils/common_utils.py
import logging import os import pickle import random import shutil import subprocess import numpy as np import torch import torch.distributed as dist import torch.multiprocessing as mp def check_numpy_to_torch(x): if isinstance(x, np.ndarray): return torch.from_numpy(x).float(), True return x, False def limit_period(val, offset=0.5, period=np.pi): val, is_numpy = check_numpy_to_torch(val) ans = val - torch.floor(val / period + offset) * period return ans.numpy() if is_numpy else ans def drop_info_with_name(info, name): ret_info = {} keep_indices = [i for i, x in enumerate(info['name']) if x != name] for key in info.keys(): ret_info[key] = info[key][keep_indices] return ret_info def rotate_points_along_z(points, angle): """ Args: points: (B, N, 3 + C) angle: (B), angle along z-axis, angle increases x ==> y Returns: """ points, is_numpy = check_numpy_to_torch(points) angle, _ = check_numpy_to_torch(angle) cosa = torch.cos(angle) sina = torch.sin(angle) zeros = angle.new_zeros(points.shape[0]) ones = angle.new_ones(points.shape[0]) rot_matrix = torch.stack(( cosa, sina, zeros, -sina, cosa, zeros, zeros, zeros, ones ), dim=1).view(-1, 3, 3).float() points_rot = torch.matmul(points[:, :, 0:3], rot_matrix) points_rot = torch.cat((points_rot, points[:, :, 3:]), dim=-1) return points_rot.numpy() if is_numpy else points_rot def mask_points_by_range(points, limit_range): mask = (points[:, 0] >= limit_range[0]) & (points[:, 0] <= limit_range[3]) \ & (points[:, 1] >= limit_range[1]) & (points[:, 1] <= limit_range[4]) return mask def get_voxel_centers(voxel_coords, downsample_times, voxel_size, point_cloud_range): """ Args: voxel_coords: (N, 3) downsample_times: voxel_size: point_cloud_range: Returns: """ assert voxel_coords.shape[1] == 3 voxel_centers = voxel_coords[:, [2, 1, 0]].float() # (xyz) voxel_size = torch.tensor(voxel_size, device=voxel_centers.device).float() * downsample_times pc_range = torch.tensor(point_cloud_range[0:3], device=voxel_centers.device).float() voxel_centers = (voxel_centers + 0.5) * voxel_size + pc_range return voxel_centers def create_logger(log_file=None, rank=0, log_level=logging.INFO): logger = logging.getLogger(__name__) logger.setLevel(log_level if rank == 0 else 'ERROR') formatter = logging.Formatter('%(asctime)s %(levelname)5s %(message)s') console = logging.StreamHandler() console.setLevel(log_level if rank == 0 else 'ERROR') console.setFormatter(formatter) logger.addHandler(console) if log_file is not None: file_handler = logging.FileHandler(filename=log_file) file_handler.setLevel(log_level if rank == 0 else 'ERROR') file_handler.setFormatter(formatter) logger.addHandler(file_handler) return logger def set_random_seed(seed): random.seed(seed) np.random.seed(seed) torch.manual_seed(seed) torch.backends.cudnn.deterministic = True torch.backends.cudnn.benchmark = False def keep_arrays_by_name(gt_names, used_classes): inds = [i for i, x in enumerate(gt_names) if x in used_classes] inds = np.array(inds, dtype=np.int64) return inds def init_dist_slurm(tcp_port, local_rank, backend='nccl'): """ modified from https://github.com/open-mmlab/mmdetection Args: tcp_port: backend: Returns: """ proc_id = int(os.environ['SLURM_PROCID']) ntasks = int(os.environ['SLURM_NTASKS']) node_list = os.environ['SLURM_NODELIST'] num_gpus = torch.cuda.device_count() torch.cuda.set_device(proc_id % num_gpus) addr = subprocess.getoutput('scontrol show hostname {} | head -n1'.format(node_list)) os.environ['MASTER_PORT'] = str(tcp_port) os.environ['MASTER_ADDR'] = addr os.environ['WORLD_SIZE'] = str(ntasks) os.environ['RANK'] = str(proc_id) dist.init_process_group(backend=backend) total_gpus = dist.get_world_size() rank = dist.get_rank() return total_gpus, rank def init_dist_pytorch(tcp_port, local_rank, backend='nccl'): if mp.get_start_method(allow_none=True) is None: mp.set_start_method('spawn') num_gpus = torch.cuda.device_count() torch.cuda.set_device(local_rank % num_gpus) dist.init_process_group( backend=backend, init_method='tcp://127.0.0.1:%d' % tcp_port, rank=local_rank, world_size=num_gpus ) rank = dist.get_rank() return num_gpus, rank def get_dist_info(): if torch.__version__ < '1.0': initialized = dist._initialized else: if dist.is_available(): initialized = dist.is_initialized() else: initialized = False if initialized: rank = dist.get_rank() world_size = dist.get_world_size() else: rank = 0 world_size = 1 return rank, world_size def merge_results_dist(result_part, size, tmpdir): rank, world_size = get_dist_info() os.makedirs(tmpdir, exist_ok=True) dist.barrier() pickle.dump(result_part, open(os.path.join(tmpdir, 'result_part_{}.pkl'.format(rank)), 'wb')) dist.barrier() if rank != 0: return None part_list = [] for i in range(world_size): part_file = os.path.join(tmpdir, 'result_part_{}.pkl'.format(i)) part_list.append(pickle.load(open(part_file, 'rb'))) ordered_results = [] for res in zip(*part_list): ordered_results.extend(list(res)) ordered_results = ordered_results[:size] shutil.rmtree(tmpdir) return ordered_results
5,750
28.341837
97
py
SASA
SASA-main/pcdet/utils/visual_utils/visualize_utils.py
import mayavi.mlab as mlab import numpy as np import torch box_colormap = [ [1, 1, 1], [0, 1, 0], [0, 1, 1], [1, 1, 0], ] def check_numpy_to_torch(x): if isinstance(x, np.ndarray): return torch.from_numpy(x).float(), True return x, False def rotate_points_along_z(points, angle): """ Args: points: (B, N, 3 + C) angle: (B), angle along z-axis, angle increases x ==> y Returns: """ points, is_numpy = check_numpy_to_torch(points) angle, _ = check_numpy_to_torch(angle) cosa = torch.cos(angle) sina = torch.sin(angle) zeros = angle.new_zeros(points.shape[0]) ones = angle.new_ones(points.shape[0]) rot_matrix = torch.stack(( cosa, sina, zeros, -sina, cosa, zeros, zeros, zeros, ones ), dim=1).view(-1, 3, 3).float() points_rot = torch.matmul(points[:, :, 0:3], rot_matrix) points_rot = torch.cat((points_rot, points[:, :, 3:]), dim=-1) return points_rot.numpy() if is_numpy else points_rot def boxes_to_corners_3d(boxes3d): """ 7 -------- 4 /| /| 6 -------- 5 . | | | | . 3 -------- 0 |/ |/ 2 -------- 1 Args: boxes3d: (N, 7) [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center Returns: """ boxes3d, is_numpy = check_numpy_to_torch(boxes3d) template = boxes3d.new_tensor(( [1, 1, -1], [1, -1, -1], [-1, -1, -1], [-1, 1, -1], [1, 1, 1], [1, -1, 1], [-1, -1, 1], [-1, 1, 1], )) / 2 corners3d = boxes3d[:, None, 3:6].repeat(1, 8, 1) * template[None, :, :] corners3d = rotate_points_along_z(corners3d.view(-1, 8, 3), boxes3d[:, 6]).view(-1, 8, 3) corners3d += boxes3d[:, None, 0:3] return corners3d.numpy() if is_numpy else corners3d def visualize_pts(pts, fig=None, bgcolor=(0, 0, 0), fgcolor=(1.0, 1.0, 1.0), show_intensity=False, size=(800, 600), draw_origin=True): if not isinstance(pts, np.ndarray): pts = pts.cpu().numpy() if fig is None: fig = mlab.figure(figure=None, bgcolor=bgcolor, fgcolor=fgcolor, engine=None, size=size) if show_intensity: G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], pts[:, 3], mode='point', colormap='gnuplot', scale_factor=1, figure=fig) else: G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], mode='point', colormap='gnuplot', scale_factor=1, figure=fig) if draw_origin: mlab.points3d(0, 0, 0, color=(1, 1, 1), mode='cube', scale_factor=0.2) mlab.plot3d([0, 3], [0, 0], [0, 0], color=(0, 0, 1), tube_radius=0.1) mlab.plot3d([0, 0], [0, 3], [0, 0], color=(0, 1, 0), tube_radius=0.1) mlab.plot3d([0, 0], [0, 0], [0, 3], color=(1, 0, 0), tube_radius=0.1) return fig def draw_sphere_pts(pts, color=(0, 1, 0), fig=None, bgcolor=(0, 0, 0), scale_factor=0.2): if not isinstance(pts, np.ndarray): pts = pts.cpu().numpy() if fig is None: fig = mlab.figure(figure=None, bgcolor=bgcolor, fgcolor=None, engine=None, size=(600, 600)) if isinstance(color, np.ndarray) and color.shape[0] == 1: color = color[0] color = (color[0] / 255.0, color[1] / 255.0, color[2] / 255.0) if isinstance(color, np.ndarray): pts_color = np.zeros((pts.__len__(), 4), dtype=np.uint8) pts_color[:, 0:3] = color pts_color[:, 3] = 255 G = mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], np.arange(0, pts_color.__len__()), mode='sphere', scale_factor=scale_factor, figure=fig) G.glyph.color_mode = 'color_by_scalar' G.glyph.scale_mode = 'scale_by_vector' G.module_manager.scalar_lut_manager.lut.table = pts_color else: mlab.points3d(pts[:, 0], pts[:, 1], pts[:, 2], mode='sphere', color=color, colormap='gnuplot', scale_factor=scale_factor, figure=fig) mlab.points3d(0, 0, 0, color=(1, 1, 1), mode='cube', scale_factor=0.2) mlab.plot3d([0, 3], [0, 0], [0, 0], color=(0, 0, 1), line_width=3, tube_radius=None, figure=fig) mlab.plot3d([0, 0], [0, 3], [0, 0], color=(0, 1, 0), line_width=3, tube_radius=None, figure=fig) mlab.plot3d([0, 0], [0, 0], [0, 3], color=(1, 0, 0), line_width=3, tube_radius=None, figure=fig) return fig def draw_grid(x1, y1, x2, y2, fig, tube_radius=None, color=(0.5, 0.5, 0.5)): mlab.plot3d([x1, x1], [y1, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x2, x2], [y1, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x1, x2], [y1, y1], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) mlab.plot3d([x1, x2], [y2, y2], [0, 0], color=color, tube_radius=tube_radius, line_width=1, figure=fig) return fig def draw_multi_grid_range(fig, grid_size=20, bv_range=(-60, -60, 60, 60)): for x in range(bv_range[0], bv_range[2], grid_size): for y in range(bv_range[1], bv_range[3], grid_size): fig = draw_grid(x, y, x + grid_size, y + grid_size, fig) return fig def draw_scenes(points, gt_boxes=None, ref_boxes=None, ref_scores=None, ref_labels=None): if not isinstance(points, np.ndarray): points = points.cpu().numpy() if ref_boxes is not None and not isinstance(ref_boxes, np.ndarray): ref_boxes = ref_boxes.cpu().numpy() if gt_boxes is not None and not isinstance(gt_boxes, np.ndarray): gt_boxes = gt_boxes.cpu().numpy() if ref_scores is not None and not isinstance(ref_scores, np.ndarray): ref_scores = ref_scores.cpu().numpy() if ref_labels is not None and not isinstance(ref_labels, np.ndarray): ref_labels = ref_labels.cpu().numpy() fig = visualize_pts(points) fig = draw_multi_grid_range(fig, bv_range=(0, -40, 80, 40)) if gt_boxes is not None: corners3d = boxes_to_corners_3d(gt_boxes) fig = draw_corners3d(corners3d, fig=fig, color=(0, 0, 1), max_num=100) if ref_boxes is not None and len(ref_boxes) > 0: ref_corners3d = boxes_to_corners_3d(ref_boxes) if ref_labels is None: fig = draw_corners3d(ref_corners3d, fig=fig, color=(0, 1, 0), cls=ref_scores, max_num=100) else: for k in range(ref_labels.min(), ref_labels.max() + 1): cur_color = tuple(box_colormap[k % len(box_colormap)]) mask = (ref_labels == k) fig = draw_corners3d(ref_corners3d[mask], fig=fig, color=cur_color, cls=ref_scores[mask], max_num=100) mlab.view(azimuth=-179, elevation=54.0, distance=104.0, roll=90.0) return fig def draw_corners3d(corners3d, fig, color=(1, 1, 1), line_width=2, cls=None, tag='', max_num=500, tube_radius=None): """ :param corners3d: (N, 8, 3) :param fig: :param color: :param line_width: :param cls: :param tag: :param max_num: :return: """ import mayavi.mlab as mlab num = min(max_num, len(corners3d)) for n in range(num): b = corners3d[n] # (8, 3) if cls is not None: if isinstance(cls, np.ndarray): mlab.text3d(b[6, 0], b[6, 1], b[6, 2], '%.2f' % cls[n], scale=(0.3, 0.3, 0.3), color=color, figure=fig) else: mlab.text3d(b[6, 0], b[6, 1], b[6, 2], '%s' % cls[n], scale=(0.3, 0.3, 0.3), color=color, figure=fig) for k in range(0, 4): i, j = k, (k + 1) % 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = k + 4, (k + 1) % 4 + 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = k, k + 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = 0, 5 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) i, j = 1, 4 mlab.plot3d([b[i, 0], b[j, 0]], [b[i, 1], b[j, 1]], [b[i, 2], b[j, 2]], color=color, tube_radius=tube_radius, line_width=line_width, figure=fig) return fig
8,540
38.541667
121
py
SASA
SASA-main/pcdet/ops/roipoint_pool3d/roipoint_pool3d_utils.py
import torch import torch.nn as nn from torch.autograd import Function from ...utils import box_utils from . import roipoint_pool3d_cuda class RoIPointPool3d(nn.Module): def __init__(self, num_sampled_points=512, pool_extra_width=1.0): super().__init__() self.num_sampled_points = num_sampled_points self.pool_extra_width = pool_extra_width def forward(self, points, point_features, boxes3d): """ Args: points: (B, N, 3) point_features: (B, N, C) boxes3d: (B, M, 7), [x, y, z, dx, dy, dz, heading] Returns: pooled_features: (B, M, 512, 3 + C) pooled_empty_flag: (B, M) """ return RoIPointPool3dFunction.apply( points, point_features, boxes3d, self.pool_extra_width, self.num_sampled_points ) class RoIPointPool3dFunction(Function): @staticmethod def forward(ctx, points, point_features, boxes3d, pool_extra_width, num_sampled_points=512): """ Args: ctx: points: (B, N, 3) point_features: (B, N, C) boxes3d: (B, num_boxes, 7), [x, y, z, dx, dy, dz, heading] pool_extra_width: num_sampled_points: Returns: pooled_features: (B, num_boxes, 512, 3 + C) pooled_empty_flag: (B, num_boxes) """ assert points.shape.__len__() == 3 and points.shape[2] == 3 batch_size, boxes_num, feature_len = points.shape[0], boxes3d.shape[1], point_features.shape[2] pooled_boxes3d = box_utils.enlarge_box3d(boxes3d.view(-1, 7), pool_extra_width).view(batch_size, -1, 7) pooled_features = point_features.new_zeros((batch_size, boxes_num, num_sampled_points, 3 + feature_len)) pooled_empty_flag = point_features.new_zeros((batch_size, boxes_num)).int() roipoint_pool3d_cuda.forward( points.contiguous(), pooled_boxes3d.contiguous(), point_features.contiguous(), pooled_features, pooled_empty_flag ) return pooled_features, pooled_empty_flag @staticmethod def backward(ctx, grad_out): raise NotImplementedError if __name__ == '__main__': pass
2,226
31.75
112
py
SASA
SASA-main/pcdet/ops/pointnet2/pointnet2_stack/pointnet2_utils.py
import torch import torch.nn as nn from torch.autograd import Function, Variable from . import pointnet2_stack_cuda as pointnet2 class BallQuery(Function): @staticmethod def forward(ctx, radius: float, nsample: int, xyz: torch.Tensor, xyz_batch_cnt: torch.Tensor, new_xyz: torch.Tensor, new_xyz_batch_cnt): """ Args: ctx: radius: float, radius of the balls nsample: int, maximum number of features in the balls xyz: (N1 + N2 ..., 3) xyz coordinates of the features xyz_batch_cnt: (batch_size), [N1, N2, ...] new_xyz: (M1 + M2 ..., 3) centers of the ball query new_xyz_batch_cnt: (batch_size), [M1, M2, ...] Returns: idx: (M1 + M2, nsample) tensor with the indicies of the features that form the query balls """ assert new_xyz.is_contiguous() assert new_xyz_batch_cnt.is_contiguous() assert xyz.is_contiguous() assert xyz_batch_cnt.is_contiguous() B = xyz_batch_cnt.shape[0] M = new_xyz.shape[0] idx = torch.cuda.IntTensor(M, nsample).zero_() pointnet2.ball_query_wrapper(B, M, radius, nsample, new_xyz, new_xyz_batch_cnt, xyz, xyz_batch_cnt, idx) empty_ball_mask = (idx[:, 0] == -1) idx[empty_ball_mask] = 0 return idx, empty_ball_mask @staticmethod def backward(ctx, a=None): return None, None, None, None ball_query = BallQuery.apply class GroupingOperation(Function): @staticmethod def forward(ctx, features: torch.Tensor, features_batch_cnt: torch.Tensor, idx: torch.Tensor, idx_batch_cnt: torch.Tensor): """ Args: ctx: features: (N1 + N2 ..., C) tensor of features to group features_batch_cnt: (batch_size) [N1 + N2 ...] tensor containing the indicies of features to group with idx: (M1 + M2 ..., nsample) tensor containing the indicies of features to group with idx_batch_cnt: (batch_size) [M1 + M2 ...] tensor containing the indicies of features to group with Returns: output: (M1 + M2, C, nsample) tensor """ assert features.is_contiguous() assert features_batch_cnt.is_contiguous() assert idx.is_contiguous() assert idx_batch_cnt.is_contiguous() assert features.shape[0] == features_batch_cnt.sum(), \ 'features: %s, features_batch_cnt: %s' % (str(features.shape), str(features_batch_cnt)) assert idx.shape[0] == idx_batch_cnt.sum(), \ 'idx: %s, idx_batch_cnt: %s' % (str(idx.shape), str(idx_batch_cnt)) M, nsample = idx.size() N, C = features.size() B = idx_batch_cnt.shape[0] output = torch.cuda.FloatTensor(M, C, nsample) pointnet2.group_points_wrapper(B, M, C, nsample, features, features_batch_cnt, idx, idx_batch_cnt, output) ctx.for_backwards = (B, N, idx, features_batch_cnt, idx_batch_cnt) return output @staticmethod def backward(ctx, grad_out: torch.Tensor): """ Args: ctx: grad_out: (M1 + M2 ..., C, nsample) tensor of the gradients of the output from forward Returns: grad_features: (N1 + N2 ..., C) gradient of the features """ B, N, idx, features_batch_cnt, idx_batch_cnt = ctx.for_backwards M, C, nsample = grad_out.size() grad_features = Variable(torch.cuda.FloatTensor(N, C).zero_()) grad_out_data = grad_out.data.contiguous() pointnet2.group_points_grad_wrapper(B, M, C, N, nsample, grad_out_data, idx, idx_batch_cnt, features_batch_cnt, grad_features.data) return grad_features, None, None, None grouping_operation = GroupingOperation.apply class QueryAndGroup(nn.Module): def __init__(self, radius: float, nsample: int, use_xyz: bool = True): """ Args: radius: float, radius of ball nsample: int, maximum number of features to gather in the ball use_xyz: """ super().__init__() self.radius, self.nsample, self.use_xyz = radius, nsample, use_xyz def forward(self, xyz: torch.Tensor, xyz_batch_cnt: torch.Tensor, new_xyz: torch.Tensor, new_xyz_batch_cnt: torch.Tensor, features: torch.Tensor = None): """ Args: xyz: (N1 + N2 ..., 3) xyz coordinates of the features xyz_batch_cnt: (batch_size), [N1, N2, ...] new_xyz: (M1 + M2 ..., 3) centers of the ball query new_xyz_batch_cnt: (batch_size), [M1, M2, ...] features: (N1 + N2 ..., C) tensor of features to group Returns: new_features: (M1 + M2, C, nsample) tensor """ assert xyz.shape[0] == xyz_batch_cnt.sum(), 'xyz: %s, xyz_batch_cnt: %s' % (str(xyz.shape), str(new_xyz_batch_cnt)) assert new_xyz.shape[0] == new_xyz_batch_cnt.sum(), \ 'new_xyz: %s, new_xyz_batch_cnt: %s' % (str(new_xyz.shape), str(new_xyz_batch_cnt)) # idx: (M1 + M2 ..., nsample), empty_ball_mask: (M1 + M2 ...) idx, empty_ball_mask = ball_query(self.radius, self.nsample, xyz, xyz_batch_cnt, new_xyz, new_xyz_batch_cnt) grouped_xyz = grouping_operation(xyz, xyz_batch_cnt, idx, new_xyz_batch_cnt) # (M1 + M2, 3, nsample) grouped_xyz -= new_xyz.unsqueeze(-1) grouped_xyz[empty_ball_mask] = 0 if features is not None: grouped_features = grouping_operation(features, xyz_batch_cnt, idx, new_xyz_batch_cnt) # (M1 + M2, C, nsample) grouped_features[empty_ball_mask] = 0 if self.use_xyz: new_features = torch.cat([grouped_xyz, grouped_features], dim=1) # (M1 + M2 ..., C + 3, nsample) else: new_features = grouped_features else: assert self.use_xyz, "Cannot have not features and not use xyz as a feature!" new_features = grouped_xyz return new_features, idx class FurthestPointSampling(Function): @staticmethod def forward(ctx, xyz: torch.Tensor, npoint: int): """ Args: ctx: xyz: (B, N, 3) where N > npoint npoint: int, number of features in the sampled set Returns: output: (B, npoint) tensor containing the set """ assert xyz.is_contiguous() B, N, _ = xyz.size() output = torch.cuda.IntTensor(B, npoint) temp = torch.cuda.FloatTensor(B, N).fill_(1e10) pointnet2.furthest_point_sampling_wrapper(B, N, npoint, xyz, temp, output) return output @staticmethod def backward(xyz, a=None): return None, None furthest_point_sample = FurthestPointSampling.apply class ThreeNN(Function): @staticmethod def forward(ctx, unknown, unknown_batch_cnt, known, known_batch_cnt): """ Args: ctx: unknown: (N1 + N2..., 3) unknown_batch_cnt: (batch_size), [N1, N2, ...] known: (M1 + M2..., 3) known_batch_cnt: (batch_size), [M1, M2, ...] Returns: dist: (N1 + N2 ..., 3) l2 distance to the three nearest neighbors idx: (N1 + N2 ..., 3) index of the three nearest neighbors, range [0, M1+M2+...] """ assert unknown.shape.__len__() == 2 and unknown.shape[1] == 3 assert known.shape.__len__() == 2 and known.shape[1] == 3 assert unknown_batch_cnt.__len__() == known_batch_cnt.__len__() dist2 = unknown.new_zeros(unknown.shape) idx = unknown_batch_cnt.new_zeros(unknown.shape).int() pointnet2.three_nn_wrapper( unknown.contiguous(), unknown_batch_cnt.contiguous(), known.contiguous(), known_batch_cnt.contiguous(), dist2, idx ) return torch.sqrt(dist2), idx @staticmethod def backward(ctx, a=None, b=None): return None, None three_nn = ThreeNN.apply class ThreeInterpolate(Function): @staticmethod def forward(ctx, features: torch.Tensor, idx: torch.Tensor, weight: torch.Tensor): """ Args: ctx: features: (M1 + M2 ..., C) idx: [N1 + N2 ..., 3] weight: [N1 + N2 ..., 3] Returns: out_tensor: (N1 + N2 ..., C) """ assert idx.shape[0] == weight.shape[0] and idx.shape[1] == weight.shape[1] == 3 ctx.three_interpolate_for_backward = (idx, weight, features.shape[0]) output = features.new_zeros((idx.shape[0], features.shape[1])) pointnet2.three_interpolate_wrapper(features.contiguous(), idx.contiguous(), weight.contiguous(), output) return output @staticmethod def backward(ctx, grad_out: torch.Tensor): """ Args: ctx: grad_out: (N1 + N2 ..., C) Returns: grad_features: (M1 + M2 ..., C) """ idx, weight, M = ctx.three_interpolate_for_backward grad_features = grad_out.new_zeros((M, grad_out.shape[1])) pointnet2.three_interpolate_grad_wrapper( grad_out.contiguous(), idx.contiguous(), weight.contiguous(), grad_features ) return grad_features, None, None three_interpolate = ThreeInterpolate.apply if __name__ == '__main__': pass
9,462
34.441948
123
py
SASA
SASA-main/pcdet/ops/pointnet2/pointnet2_stack/pointnet2_modules.py
from typing import List import torch import torch.nn as nn import torch.nn.functional as F from . import pointnet2_utils class StackSAModuleMSG(nn.Module): def __init__(self, *, radii: List[float], nsamples: List[int], mlps: List[List[int]], use_xyz: bool = True, pool_method='max_pool'): """ Args: radii: list of float, list of radii to group with nsamples: list of int, number of samples in each ball query mlps: list of list of int, spec of the pointnet before the global pooling for each scale use_xyz: pool_method: max_pool / avg_pool """ super().__init__() assert len(radii) == len(nsamples) == len(mlps) self.groupers = nn.ModuleList() self.mlps = nn.ModuleList() for i in range(len(radii)): radius = radii[i] nsample = nsamples[i] self.groupers.append(pointnet2_utils.QueryAndGroup(radius, nsample, use_xyz=use_xyz)) mlp_spec = mlps[i] if use_xyz: mlp_spec[0] += 3 shared_mlps = [] for k in range(len(mlp_spec) - 1): shared_mlps.extend([ nn.Conv2d(mlp_spec[k], mlp_spec[k + 1], kernel_size=1, bias=False), nn.BatchNorm2d(mlp_spec[k + 1]), nn.ReLU() ]) self.mlps.append(nn.Sequential(*shared_mlps)) self.pool_method = pool_method self.init_weights() def init_weights(self): for m in self.modules(): if isinstance(m, nn.Conv2d): nn.init.kaiming_normal_(m.weight) if m.bias is not None: nn.init.constant_(m.bias, 0) if isinstance(m, nn.BatchNorm2d): nn.init.constant_(m.weight, 1.0) nn.init.constant_(m.bias, 0) def forward(self, xyz, xyz_batch_cnt, new_xyz, new_xyz_batch_cnt, features=None, empty_voxel_set_zeros=True): """ :param xyz: (N1 + N2 ..., 3) tensor of the xyz coordinates of the features :param xyz_batch_cnt: (batch_size), [N1, N2, ...] :param new_xyz: (M1 + M2 ..., 3) :param new_xyz_batch_cnt: (batch_size), [M1, M2, ...] :param features: (N1 + N2 ..., C) tensor of the descriptors of the the features :return: new_xyz: (M1 + M2 ..., 3) tensor of the new features' xyz new_features: (M1 + M2 ..., \sum_k(mlps[k][-1])) tensor of the new_features descriptors """ new_features_list = [] for k in range(len(self.groupers)): new_features, ball_idxs = self.groupers[k]( xyz, xyz_batch_cnt, new_xyz, new_xyz_batch_cnt, features ) # (M1 + M2, C, nsample) new_features = new_features.permute(1, 0, 2).unsqueeze(dim=0) # (1, C, M1 + M2 ..., nsample) new_features = self.mlps[k](new_features) # (1, C, M1 + M2 ..., nsample) if self.pool_method == 'max_pool': new_features = F.max_pool2d( new_features, kernel_size=[1, new_features.size(3)] ).squeeze(dim=-1) # (1, C, M1 + M2 ...) elif self.pool_method == 'avg_pool': new_features = F.avg_pool2d( new_features, kernel_size=[1, new_features.size(3)] ).squeeze(dim=-1) # (1, C, M1 + M2 ...) else: raise NotImplementedError new_features = new_features.squeeze(dim=0).permute(1, 0) # (M1 + M2 ..., C) new_features_list.append(new_features) new_features = torch.cat(new_features_list, dim=1) # (M1 + M2 ..., C) return new_xyz, new_features class StackPointnetFPModule(nn.Module): def __init__(self, *, mlp: List[int]): """ Args: mlp: list of int """ super().__init__() shared_mlps = [] for k in range(len(mlp) - 1): shared_mlps.extend([ nn.Conv2d(mlp[k], mlp[k + 1], kernel_size=1, bias=False), nn.BatchNorm2d(mlp[k + 1]), nn.ReLU() ]) self.mlp = nn.Sequential(*shared_mlps) def forward(self, unknown, unknown_batch_cnt, known, known_batch_cnt, unknown_feats=None, known_feats=None): """ Args: unknown: (N1 + N2 ..., 3) known: (M1 + M2 ..., 3) unknow_feats: (N1 + N2 ..., C1) known_feats: (M1 + M2 ..., C2) Returns: new_features: (N1 + N2 ..., C_out) """ dist, idx = pointnet2_utils.three_nn(unknown, unknown_batch_cnt, known, known_batch_cnt) dist_recip = 1.0 / (dist + 1e-8) norm = torch.sum(dist_recip, dim=-1, keepdim=True) weight = dist_recip / norm interpolated_feats = pointnet2_utils.three_interpolate(known_feats, idx, weight) if unknown_feats is not None: new_features = torch.cat([interpolated_feats, unknown_feats], dim=1) # (N1 + N2 ..., C2 + C1) else: new_features = interpolated_feats new_features = new_features.permute(1, 0)[None, :, :, None] # (1, C, N1 + N2 ..., 1) new_features = self.mlp(new_features) new_features = new_features.squeeze(dim=0).squeeze(dim=-1).permute(1, 0) # (N1 + N2 ..., C) return new_features
5,425
38.318841
113
py
SASA
SASA-main/pcdet/ops/pointnet2/pointnet2_batch/pointnet2_utils.py
from typing import List, Tuple import torch import torch.nn as nn import torch.nn.functional as F from torch.autograd import Function, Variable from . import pointnet2_batch_cuda as pointnet2 @torch.no_grad() def calc_dist_matrix_for_sampling(xyz: torch.Tensor, features: torch.Tensor = None, gamma: float = 1.0): dist = torch.cdist(xyz, xyz) if features is not None: dist += torch.cdist(features, features) * gamma return dist @torch.no_grad() def furthest_point_sample(xyz: torch.Tensor, npoint: int) -> torch.Tensor: """ Uses iterative furthest point sampling to select a set of npoint features that have the largest minimum distance :param ctx: :param xyz: (B, N, 3) where N > npoint :param npoint: int, number of features in the sampled set :return: output: (B, npoint) tensor containing the set """ assert xyz.is_contiguous() B, N, _ = xyz.size() output = torch.cuda.IntTensor(B, npoint) temp = torch.cuda.FloatTensor(B, N).fill_(1e10) pointnet2.furthest_point_sampling_wrapper(B, N, npoint, xyz, temp, output) return output @torch.no_grad() def furthest_point_sample_matrix(matrix: torch.Tensor, npoint: int) -> torch.Tensor: """ Uses iterative furthest point sampling to select a set of npoint features that have the largest minimum distance with a pairwise distance matrix :param matrix: (B, N, N) tensor of dist matrix :param npoint: int, number of features in the sampled set :return: output: (B, npoint) tensor containing the set """ assert matrix.is_contiguous() B, N, _ = matrix.size() output = torch.cuda.IntTensor(B, npoint) temp = torch.cuda.FloatTensor(B, N).fill_(1e10) pointnet2.furthest_point_sampling_matrix_wrapper(B, N, npoint, matrix, temp, output) return output @torch.no_grad() def furthest_point_sample_weights(xyz: torch.Tensor, weights: torch.Tensor, npoint: int) -> torch.Tensor: """ Uses iterative furthest point sampling to select a set of npoint features that have the largest minimum weighted distance Args: xyz: (B, N, 3), tensor of xyz coordinates weights: (B, N), tensor of point weights npoint: int, number of points in the sampled set Returns: output: (B, npoint) tensor containing the set """ assert xyz.is_contiguous() assert weights.is_contiguous() B, N, _ = xyz.size() output = torch.cuda.IntTensor(B, npoint) temp = torch.cuda.FloatTensor(B, N).fill_(1e10) pointnet2.furthest_point_sampling_weights_wrapper(B, N, npoint, xyz, weights, temp, output) return output class GatherOperation(Function): @staticmethod def forward(ctx, features: torch.Tensor, idx: torch.Tensor) -> torch.Tensor: """ :param ctx: :param features: (B, C, N) :param idx: (B, npoint) index tensor of the features to gather :return: output: (B, C, npoint) """ assert features.is_contiguous() assert idx.is_contiguous() B, npoint = idx.size() _, C, N = features.size() output = torch.cuda.FloatTensor(B, C, npoint) pointnet2.gather_points_wrapper(B, C, N, npoint, features, idx, output) ctx.for_backwards = (idx, C, N) return output @staticmethod def backward(ctx, grad_out): idx, C, N = ctx.for_backwards B, npoint = idx.size() grad_features = Variable(torch.cuda.FloatTensor(B, C, N).zero_()) grad_out_data = grad_out.data.contiguous() pointnet2.gather_points_grad_wrapper(B, C, N, npoint, grad_out_data, idx, grad_features.data) return grad_features, None gather_operation = GatherOperation.apply def three_nn(unknown: torch.Tensor, known: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: """ Find the three nearest neighbors of unknown in known :param unknown: (B, N, 3) :param known: (B, M, 3) :return: dist: (B, N, 3) l2 distance to the three nearest neighbors idx: (B, N, 3) index of 3 nearest neighbors """ assert unknown.is_contiguous() assert known.is_contiguous() B, N, _ = unknown.size() m = known.size(1) dist2 = torch.cuda.FloatTensor(B, N, 3) idx = torch.cuda.IntTensor(B, N, 3) pointnet2.three_nn_wrapper(B, N, m, unknown, known, dist2, idx) return torch.sqrt(dist2), idx class ThreeInterpolate(Function): @staticmethod def forward(ctx, features: torch.Tensor, idx: torch.Tensor, weight: torch.Tensor) -> torch.Tensor: """ Performs weight linear interpolation on 3 features :param ctx: :param features: (B, C, M) Features descriptors to be interpolated from :param idx: (B, n, 3) three nearest neighbors of the target features in features :param weight: (B, n, 3) weights :return: output: (B, C, N) tensor of the interpolated features """ assert features.is_contiguous() assert idx.is_contiguous() assert weight.is_contiguous() B, c, m = features.size() n = idx.size(1) ctx.three_interpolate_for_backward = (idx, weight, m) output = torch.cuda.FloatTensor(B, c, n) pointnet2.three_interpolate_wrapper(B, c, m, n, features, idx, weight, output) return output @staticmethod def backward(ctx, grad_out: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """ :param ctx: :param grad_out: (B, C, N) tensor with gradients of outputs :return: grad_features: (B, C, M) tensor with gradients of features None: None: """ idx, weight, m = ctx.three_interpolate_for_backward B, c, n = grad_out.size() grad_features = Variable(torch.cuda.FloatTensor(B, c, m).zero_()) grad_out_data = grad_out.data.contiguous() pointnet2.three_interpolate_grad_wrapper(B, c, n, m, grad_out_data, idx, weight, grad_features.data) return grad_features, None, None three_interpolate = ThreeInterpolate.apply class GroupingOperation(Function): @staticmethod def forward(ctx, features: torch.Tensor, idx: torch.Tensor) -> torch.Tensor: """ :param ctx: :param features: (B, C, N) tensor of features to group :param idx: (B, npoint, nsample) tensor containing the indicies of features to group with :return: output: (B, C, npoint, nsample) tensor """ assert features.is_contiguous() assert idx.is_contiguous() B, nfeatures, nsample = idx.size() _, C, N = features.size() output = torch.cuda.FloatTensor(B, C, nfeatures, nsample) pointnet2.group_points_wrapper(B, C, N, nfeatures, nsample, features, idx, output) ctx.for_backwards = (idx, N) return output @staticmethod def backward(ctx, grad_out: torch.Tensor) -> Tuple[torch.Tensor, torch.Tensor]: """ :param ctx: :param grad_out: (B, C, npoint, nsample) tensor of the gradients of the output from forward :return: grad_features: (B, C, N) gradient of the features """ idx, N = ctx.for_backwards B, C, npoint, nsample = grad_out.size() grad_features = Variable(torch.cuda.FloatTensor(B, C, N).zero_()) grad_out_data = grad_out.data.contiguous() pointnet2.group_points_grad_wrapper(B, C, N, npoint, nsample, grad_out_data, idx, grad_features.data) return grad_features, None grouping_operation = GroupingOperation.apply @torch.no_grad() def ball_query(radius: float, nsample: int, xyz: torch.Tensor, new_xyz: torch.Tensor): """ :param radius: float, radius of the balls :param nsample: int, maximum number of features in the balls :param xyz: (B, N, 3) xyz coordinates of the features :param new_xyz: (B, npoint, 3) centers of the ball query :return: idx_cnt: (B, npoint) tensor with the number of grouped points for each ball query idx: (B, npoint, nsample) tensor with the indicies of the features that form the query balls """ assert new_xyz.is_contiguous() assert xyz.is_contiguous() B, N, _ = xyz.size() npoint = new_xyz.size(1) idx = torch.cuda.IntTensor(B, npoint, nsample).zero_() idx_cnt = torch.cuda.IntTensor(B, npoint).zero_() pointnet2.ball_query_wrapper(B, N, npoint, radius, nsample, new_xyz, xyz, idx_cnt, idx) return idx_cnt, idx @torch.no_grad() def ball_query_dilated(radius_in: float, radius_out: float, nsample: int, xyz: torch.Tensor, new_xyz: torch.Tensor): """ :param radius_in: float, radius of the inner balls :param radius_out: float, radius of the outer balls :param nsample: int, maximum number of features in the balls :param xyz: (B, N, 3) xyz coordinates of the features :param new_xyz: (B, npoint, 3) centers of the ball query :return: idx_cnt: (B, npoint) tensor with the number of grouped points for each ball query idx: (B, npoint, nsample) tensor with the indicies of the features that form the query balls """ assert new_xyz.is_contiguous() assert xyz.is_contiguous() B, N, _ = xyz.size() npoint = new_xyz.size(1) idx_cnt = torch.cuda.IntTensor(B, npoint).zero_() idx = torch.cuda.IntTensor(B, npoint, nsample).zero_() pointnet2.ball_query_dilated_wrapper(B, N, npoint, radius_in, radius_out, nsample, new_xyz, xyz, idx_cnt, idx) return idx_cnt, idx class QueryAndGroup(nn.Module): def __init__(self, radius: float, nsample: int, use_xyz: bool = True): """ :param radius: float, radius of ball :param nsample: int, maximum number of features to gather in the ball :param use_xyz: """ super().__init__() self.radius, self.nsample, self.use_xyz = radius, nsample, use_xyz def forward(self, xyz: torch.Tensor, new_xyz: torch.Tensor, features: torch.Tensor = None): """ :param xyz: (B, N, 3) xyz coordinates of the features :param new_xyz: (B, npoint, 3) centroids :param features: (B, C, N) descriptors of the features :return: idx_cnt: (B, npoint) tensor with the number of grouped points for each ball query new_features: (B, 3 + C, npoint, nsample) """ idx_cnt, idx = ball_query(self.radius, self.nsample, xyz, new_xyz) xyz_trans = xyz.transpose(1, 2).contiguous() grouped_xyz = grouping_operation(xyz_trans, idx) # (B, 3, npoint, nsample) grouped_xyz -= new_xyz.transpose(1, 2).unsqueeze(-1) if features is not None: grouped_features = grouping_operation(features, idx) if self.use_xyz: new_features = torch.cat([grouped_xyz, grouped_features], dim=1) # (B, C + 3, npoint, nsample) else: new_features = grouped_features else: assert self.use_xyz, "Cannot have not features and not use xyz as a feature!" new_features = grouped_xyz return idx_cnt, new_features class QueryAndGroupDilated(nn.Module): def __init__(self, radius_in: float, radius_out: float, nsample: int, use_xyz: bool = True): """ :param radius_in: float, radius of inner ball :param radius_out: float, radius of outer ball :param nsample: int, maximum number of features to gather in the ball :param use_xyz: """ super().__init__() self.radius_in, self.radius_out, self.nsample, self.use_xyz = radius_in, radius_out, nsample, use_xyz def forward(self, xyz: torch.Tensor, new_xyz: torch.Tensor, features: torch.Tensor = None): """ :param xyz: (B, N, 3) xyz coordinates of the features :param new_xyz: (B, npoint, 3) centroids :param features: (B, C, N) descriptors of the features :return: new_features: (B, 3 + C, npoint, nsample) idx_cnt: (B, npoint) tensor with the number of grouped points for each ball query """ idx_cnt, idx = ball_query_dilated(self.radius_in, self.radius_out, self.nsample, xyz, new_xyz) xyz_trans = xyz.transpose(1, 2).contiguous() grouped_xyz = grouping_operation(xyz_trans, idx) # (B, 3, npoint, nsample) grouped_xyz -= new_xyz.transpose(1, 2).unsqueeze(-1) if features is not None: grouped_features = grouping_operation(features, idx) if self.use_xyz: new_features = torch.cat([grouped_xyz, grouped_features], dim=1) # (B, C + 3, npoint, nsample) else: new_features = grouped_features else: assert self.use_xyz, "Cannot have not features and not use xyz as a feature!" new_features = grouped_xyz return idx_cnt, new_features class GroupAll(nn.Module): def __init__(self, use_xyz: bool = True): super().__init__() self.use_xyz = use_xyz def forward(self, xyz: torch.Tensor, new_xyz: torch.Tensor, features: torch.Tensor = None): """ :param xyz: (B, N, 3) xyz coordinates of the features :param new_xyz: ignored :param features: (B, C, N) descriptors of the features :return: idx_cnt: (B, 1) new_features: (B, C + 3, 1, N) """ grouped_xyz = xyz.transpose(1, 2).unsqueeze(2) if features is not None: grouped_features = features.unsqueeze(2) if self.use_xyz: new_features = torch.cat([grouped_xyz, grouped_features], dim=1) # (B, 3 + C, 1, N) else: new_features = grouped_features else: new_features = grouped_xyz idx_cnt = new_features.new_ones(new_features.size(0), 1) return idx_cnt, new_features
14,049
35.588542
116
py
SASA
SASA-main/pcdet/ops/pointnet2/pointnet2_batch/pointnet2_modules.py
from typing import List import torch import torch.nn as nn import torch.nn.functional as F from . import pointnet2_utils class _PointnetSAModuleBase(nn.Module): def __init__(self): super().__init__() self.npoint = None self.groupers = None self.mlps = None self.pool_method = 'max_pool' def forward(self, xyz: torch.Tensor, features: torch.Tensor = None, new_xyz=None) -> (torch.Tensor, torch.Tensor): """ :param xyz: (B, N, 3) tensor of the xyz coordinates of the features :param features: (B, C, N) tensor of the descriptors of the the features :param new_xyz: (B, npoint, 3) tensor of the xyz coordinates of the grouping centers if specified :return: new_xyz: (B, npoint, 3) tensor of the new features' xyz new_features: (B, npoint, \sum_k(mlps[k][-1])) tensor of the new_features descriptors """ new_features_list = [] xyz_flipped = xyz.transpose(1, 2).contiguous() if new_xyz is None: new_xyz = pointnet2_utils.gather_operation( xyz_flipped, pointnet2_utils.furthest_point_sample(xyz, self.npoint) ).transpose(1, 2).contiguous() if self.npoint is not None else None for i in range(len(self.groupers)): idx_cnt, new_features = self.groupers[i](xyz, new_xyz, features) # (B, C, npoint, nsample) new_features = self.mlps[i](new_features) # (B, mlp[-1], npoint, nsample) idx_cnt_mask = (idx_cnt > 0).float() idx_cnt_mask = idx_cnt_mask.unsqueeze(dim=1).unsqueeze(dim=-1) new_features *= idx_cnt_mask if self.pool_method == 'max_pool': new_features = F.max_pool2d( new_features, kernel_size=[1, new_features.size(3)] ) # (B, mlp[-1], npoint, 1) elif self.pool_method == 'avg_pool': new_features = F.avg_pool2d( new_features, kernel_size=[1, new_features.size(3)] ) # (B, mlp[-1], npoint, 1) else: raise NotImplementedError new_features = new_features.squeeze(-1) # (B, mlp[-1], npoint) new_features_list.append(new_features) return new_xyz, torch.cat(new_features_list, dim=1) class PointnetSAModuleMSG(_PointnetSAModuleBase): """Pointnet set abstraction layer with multiscale grouping""" def __init__(self, *, npoint: int, radii: List[float], nsamples: List[int], mlps: List[List[int]], bn: bool = True, use_xyz: bool = True, pool_method='max_pool'): """ :param npoint: int :param radii: list of float, list of radii to group with :param nsamples: list of int, number of samples in each ball query :param mlps: list of list of int, spec of the pointnet before the global pooling for each scale :param bn: whether to use batchnorm :param use_xyz: :param pool_method: max_pool / avg_pool """ super().__init__() assert len(radii) == len(nsamples) == len(mlps) self.npoint = npoint self.groupers = nn.ModuleList() self.mlps = nn.ModuleList() for i in range(len(radii)): radius = radii[i] nsample = nsamples[i] self.groupers.append( pointnet2_utils.QueryAndGroup(radius, nsample, use_xyz=use_xyz) if npoint is not None else pointnet2_utils.GroupAll(use_xyz) ) mlp_spec = mlps[i] if use_xyz: mlp_spec[0] += 3 shared_mlps = [] for k in range(len(mlp_spec) - 1): shared_mlps.extend([ nn.Conv2d(mlp_spec[k], mlp_spec[k + 1], kernel_size=1, bias=False), nn.BatchNorm2d(mlp_spec[k + 1]), nn.ReLU() ]) self.mlps.append(nn.Sequential(*shared_mlps)) self.pool_method = pool_method class PointnetSAModule(PointnetSAModuleMSG): """Pointnet set abstraction layer""" def __init__(self, *, mlp: List[int], npoint: int = None, radius: float = None, nsample: int = None, bn: bool = True, use_xyz: bool = True, pool_method='max_pool'): """ :param mlp: list of int, spec of the pointnet before the global max_pool :param npoint: int, number of features :param radius: float, radius of ball :param nsample: int, number of samples in the ball query :param bn: whether to use batchnorm :param use_xyz: :param pool_method: max_pool / avg_pool """ super().__init__( mlps=[mlp], npoint=npoint, radii=[radius], nsamples=[nsample], bn=bn, use_xyz=use_xyz, pool_method=pool_method ) class _PointnetSAModuleFSBase(nn.Module): def __init__(self): super().__init__() self.groupers = None self.mlps = None self.npoint_list = [] self.sample_range_list = [[0, -1]] self.sample_method_list = ['d-fps'] self.radii = [] self.pool_method = 'max_pool' self.dilated_radius_group = False self.weight_gamma = 1.0 self.skip_connection = False self.aggregation_mlp = None self.confidence_mlp = None def forward(self, xyz: torch.Tensor, features: torch.Tensor = None, new_xyz=None, scores=None): """ :param xyz: (B, N, 3) tensor of the xyz coordinates of the features :param features: (B, C, N) tensor of the descriptors of the features :param new_xyz: :param scores: (B, N) tensor of confidence scores of points, required when using s-fps :return: new_xyz: (B, npoint, 3) tensor of the new features' xyz new_features: (B, npoint, \sum_k(mlps[k][-1])) tensor of the new_features descriptors """ new_features_list = [] xyz_flipped = xyz.transpose(1, 2).contiguous() if new_xyz is None: assert len(self.npoint_list) == len(self.sample_range_list) == len(self.sample_method_list) sample_idx_list = [] for i in range(len(self.sample_method_list)): xyz_slice = xyz[:, self.sample_range_list[i][0]:self.sample_range_list[i][1], :].contiguous() if self.sample_method_list[i] == 'd-fps': sample_idx = pointnet2_utils.furthest_point_sample(xyz_slice, self.npoint_list[i]) elif self.sample_method_list[i] == 'f-fps': features_slice = features[:, :, self.sample_range_list[i][0]:self.sample_range_list[i][1]] dist_matrix = pointnet2_utils.calc_dist_matrix_for_sampling(xyz_slice, features_slice.permute(0, 2, 1), self.weight_gamma) sample_idx = pointnet2_utils.furthest_point_sample_matrix(dist_matrix, self.npoint_list[i]) elif self.sample_method_list[i] == 's-fps': assert scores is not None scores_slice = \ scores[:, self.sample_range_list[i][0]:self.sample_range_list[i][1]].contiguous() scores_slice = scores_slice.sigmoid() ** self.weight_gamma sample_idx = pointnet2_utils.furthest_point_sample_weights( xyz_slice, scores_slice, self.npoint_list[i] ) else: raise NotImplementedError sample_idx_list.append(sample_idx + self.sample_range_list[i][0]) sample_idx = torch.cat(sample_idx_list, dim=-1) new_xyz = pointnet2_utils.gather_operation( xyz_flipped, sample_idx ).transpose(1, 2).contiguous() # (B, npoint, 3) if self.skip_connection: old_features = pointnet2_utils.gather_operation( features, sample_idx ) if features is not None else None # (B, C, npoint) for i in range(len(self.groupers)): idx_cnt, new_features = self.groupers[i](xyz, new_xyz, features) # (B, C, npoint, nsample) new_features = self.mlps[i](new_features) # (B, mlp[-1], npoint, nsample) idx_cnt_mask = (idx_cnt > 0).float() # (B, npoint) idx_cnt_mask = idx_cnt_mask.unsqueeze(1).unsqueeze(-1) # (B, 1, npoint, 1) new_features *= idx_cnt_mask if self.pool_method == 'max_pool': pooled_features = F.max_pool2d( new_features, kernel_size=[1, new_features.size(3)] ) # (B, mlp[-1], npoint, 1) elif self.pool_method == 'avg_pool': pooled_features = F.avg_pool2d( new_features, kernel_size=[1, new_features.size(3)] ) # (B, mlp[-1], npoint, 1) else: raise NotImplementedError new_features_list.append(pooled_features.squeeze(-1)) # (B, mlp[-1], npoint) if self.skip_connection and old_features is not None: new_features_list.append(old_features) new_features = torch.cat(new_features_list, dim=1) if self.aggregation_mlp is not None: new_features = self.aggregation_mlp(new_features) if self.confidence_mlp is not None: new_scores = self.confidence_mlp(new_features) new_scores = new_scores.squeeze(1) # (B, npoint) return new_xyz, new_features, new_scores return new_xyz, new_features, None class PointnetSAModuleFSMSG(_PointnetSAModuleFSBase): """Pointnet set abstraction layer with fusion sampling and multiscale grouping""" def __init__(self, *, npoint_list: List[int] = None, sample_range_list: List[List[int]] = None, sample_method_list: List[str] = None, radii: List[float], nsamples: List[int], mlps: List[List[int]], bn: bool = True, use_xyz: bool = True, pool_method='max_pool', dilated_radius_group: bool = False, skip_connection: bool = False, weight_gamma: float = 1.0, aggregation_mlp: List[int] = None, confidence_mlp: List[int] = None): """ :param npoint_list: list of int, number of samples for every sampling method :param sample_range_list: list of list of int, sample index range [left, right] for every sampling method :param sample_method_list: list of str, list of used sampling method, d-fps or f-fps :param radii: list of float, list of radii to group with :param nsamples: list of int, number of samples in each ball query :param mlps: list of list of int, spec of the pointnet before the global pooling for each scale :param bn: whether to use batchnorm :param use_xyz: :param pool_method: max_pool / avg_pool :param dilated_radius_group: whether to use radius dilated group :param skip_connection: whether to add skip connection :param weight_gamma: gamma for s-fps, default: 1.0 :param aggregation_mlp: list of int, spec aggregation mlp :param confidence_mlp: list of int, spec confidence mlp """ super().__init__() assert npoint_list is None or len(npoint_list) == len(sample_range_list) == len(sample_method_list) assert len(radii) == len(nsamples) == len(mlps) self.npoint_list = npoint_list self.sample_range_list = sample_range_list self.sample_method_list = sample_method_list self.radii = radii self.groupers = nn.ModuleList() self.mlps = nn.ModuleList() former_radius = 0.0 in_channels, out_channels = 0, 0 for i in range(len(radii)): radius = radii[i] nsample = nsamples[i] if dilated_radius_group: self.groupers.append( pointnet2_utils.QueryAndGroupDilated(former_radius, radius, nsample, use_xyz=use_xyz) ) else: self.groupers.append( pointnet2_utils.QueryAndGroup(radius, nsample, use_xyz=use_xyz) ) former_radius = radius mlp_spec = mlps[i] if use_xyz: mlp_spec[0] += 3 shared_mlp = [] for k in range(len(mlp_spec) - 1): shared_mlp.extend([ nn.Conv2d(mlp_spec[k], mlp_spec[k + 1], kernel_size=1, bias=False), nn.BatchNorm2d(mlp_spec[k + 1]), nn.ReLU() ]) self.mlps.append(nn.Sequential(*shared_mlp)) in_channels = mlp_spec[0] - 3 if use_xyz else mlp_spec[0] out_channels += mlp_spec[-1] self.pool_method = pool_method self.dilated_radius_group = dilated_radius_group self.skip_connection = skip_connection self.weight_gamma = weight_gamma if skip_connection: out_channels += in_channels if aggregation_mlp is not None: shared_mlp = [] for k in range(len(aggregation_mlp)): shared_mlp.extend([ nn.Conv1d(out_channels, aggregation_mlp[k], kernel_size=1, bias=False), nn.BatchNorm1d(aggregation_mlp[k]), nn.ReLU() ]) out_channels = aggregation_mlp[k] self.aggregation_mlp = nn.Sequential(*shared_mlp) else: self.aggregation_mlp = None if confidence_mlp is not None: shared_mlp = [] for k in range(len(confidence_mlp)): shared_mlp.extend([ nn.Conv1d(out_channels, confidence_mlp[k], kernel_size=1, bias=False), nn.BatchNorm1d(confidence_mlp[k]), nn.ReLU() ]) out_channels = confidence_mlp[k] shared_mlp.append( nn.Conv1d(out_channels, 1, kernel_size=1, bias=True), ) self.confidence_mlp = nn.Sequential(*shared_mlp) else: self.confidence_mlp = None class PointnetSAModuleFS(PointnetSAModuleFSMSG): """Pointnet set abstraction layer with fusion sampling""" def __init__(self, *, mlp: List[int], npoint_list: List[int] = None, sample_range_list: List[List[int]] = None, sample_method_list: List[str] = None, radius: float = None, nsample: int = None, bn: bool = True, use_xyz: bool = True, pool_method='max_pool', dilated_radius_group: bool = False, skip_connection: bool = False, weight_gamma: float = 1.0, aggregation_mlp: List[int] = None, confidence_mlp: List[int] = None): """ :param mlp: list of int, spec of the pointnet before the global max_pool :param npoint_list: list of int, number of samples for every sampling method :param sample_range_list: list of list of int, sample index range [left, right] for every sampling method :param sample_method_list: list of str, list of used sampling method, d-fps, f-fps or c-fps :param radius: float, radius of ball :param nsample: int, number of samples in the ball query :param bn: whether to use batchnorm :param use_xyz: :param pool_method: max_pool / avg_pool :param dilated_radius_group: whether to use radius dilated group :param skip_connection: whether to add skip connection :param weight_gamma: gamma for s-fps, default: 1.0 :param aggregation_mlp: list of int, spec aggregation mlp :param confidence_mlp: list of int, spec confidence mlp """ super().__init__( mlps=[mlp], npoint_list=npoint_list, sample_range_list=sample_range_list, sample_method_list=sample_method_list, radii=[radius], nsamples=[nsample], bn=bn, use_xyz=use_xyz, pool_method=pool_method, dilated_radius_group=dilated_radius_group, skip_connection=skip_connection, weight_gamma=weight_gamma, aggregation_mlp=aggregation_mlp, confidence_mlp=confidence_mlp ) class PointnetFPModule(nn.Module): r"""Propigates the features of one set to another""" def __init__(self, *, mlp: List[int], bn: bool = True): """ :param mlp: list of int :param bn: whether to use batchnorm """ super().__init__() shared_mlps = [] for k in range(len(mlp) - 1): shared_mlps.extend([ nn.Conv2d(mlp[k], mlp[k + 1], kernel_size=1, bias=False), nn.BatchNorm2d(mlp[k + 1]), nn.ReLU() ]) self.mlp = nn.Sequential(*shared_mlps) def forward( self, unknown: torch.Tensor, known: torch.Tensor, unknow_feats: torch.Tensor, known_feats: torch.Tensor ) -> torch.Tensor: """ :param unknown: (B, n, 3) tensor of the xyz positions of the unknown features :param known: (B, m, 3) tensor of the xyz positions of the known features :param unknow_feats: (B, C1, n) tensor of the features to be propigated to :param known_feats: (B, C2, m) tensor of features to be propigated :return: new_features: (B, mlp[-1], n) tensor of the features of the unknown features """ if known is not None: dist, idx = pointnet2_utils.three_nn(unknown, known) dist_recip = 1.0 / (dist + 1e-8) norm = torch.sum(dist_recip, dim=2, keepdim=True) weight = dist_recip / norm interpolated_feats = pointnet2_utils.three_interpolate(known_feats, idx, weight) else: interpolated_feats = known_feats.expand(*known_feats.size()[0:2], unknown.size(1)) if unknow_feats is not None: new_features = torch.cat([interpolated_feats, unknow_feats], dim=1) # (B, C2 + C1, n) else: new_features = interpolated_feats new_features = new_features.unsqueeze(-1) new_features = self.mlp(new_features) return new_features.squeeze(-1) if __name__ == "__main__": pass
18,910
41.688488
119
py
SASA
SASA-main/pcdet/ops/iou3d_nms/iou3d_nms_utils.py
""" 3D IoU Calculation and Rotated NMS Written by Shaoshuai Shi All Rights Reserved 2019-2020. """ import torch from ...utils import common_utils from . import iou3d_nms_cuda def boxes_bev_iou_cpu(boxes_a, boxes_b): """ Args: boxes_a: (N, 7) [x, y, z, dx, dy, dz, heading] boxes_b: (N, 7) [x, y, z, dx, dy, dz, heading] Returns: """ boxes_a, is_numpy = common_utils.check_numpy_to_torch(boxes_a) boxes_b, is_numpy = common_utils.check_numpy_to_torch(boxes_b) assert not (boxes_a.is_cuda or boxes_b.is_cuda), 'Only support CPU tensors' assert boxes_a.shape[1] == 7 and boxes_b.shape[1] == 7 ans_iou = boxes_a.new_zeros(torch.Size((boxes_a.shape[0], boxes_b.shape[0]))) iou3d_nms_cuda.boxes_iou_bev_cpu(boxes_a.contiguous(), boxes_b.contiguous(), ans_iou) return ans_iou.numpy() if is_numpy else ans_iou def boxes_iou_bev(boxes_a, boxes_b): """ Args: boxes_a: (N, 7) [x, y, z, dx, dy, dz, heading] boxes_b: (N, 7) [x, y, z, dx, dy, dz, heading] Returns: ans_iou: (N, M) """ assert boxes_a.shape[1] == boxes_b.shape[1] == 7 ans_iou = torch.cuda.FloatTensor(torch.Size((boxes_a.shape[0], boxes_b.shape[0]))).zero_() iou3d_nms_cuda.boxes_iou_bev_gpu(boxes_a.contiguous(), boxes_b.contiguous(), ans_iou) return ans_iou def boxes_iou3d_gpu(boxes_a, boxes_b): """ Args: boxes_a: (N, 7) [x, y, z, dx, dy, dz, heading] boxes_b: (N, 7) [x, y, z, dx, dy, dz, heading] Returns: ans_iou: (N, M) """ assert boxes_a.shape[1] == boxes_b.shape[1] == 7 # height overlap boxes_a_height_max = (boxes_a[:, 2] + boxes_a[:, 5] / 2).view(-1, 1) boxes_a_height_min = (boxes_a[:, 2] - boxes_a[:, 5] / 2).view(-1, 1) boxes_b_height_max = (boxes_b[:, 2] + boxes_b[:, 5] / 2).view(1, -1) boxes_b_height_min = (boxes_b[:, 2] - boxes_b[:, 5] / 2).view(1, -1) # bev overlap overlaps_bev = torch.cuda.FloatTensor(torch.Size((boxes_a.shape[0], boxes_b.shape[0]))).zero_() # (N, M) iou3d_nms_cuda.boxes_overlap_bev_gpu(boxes_a.contiguous(), boxes_b.contiguous(), overlaps_bev) max_of_min = torch.max(boxes_a_height_min, boxes_b_height_min) min_of_max = torch.min(boxes_a_height_max, boxes_b_height_max) overlaps_h = torch.clamp(min_of_max - max_of_min, min=0) # 3d iou overlaps_3d = overlaps_bev * overlaps_h vol_a = (boxes_a[:, 3] * boxes_a[:, 4] * boxes_a[:, 5]).view(-1, 1) vol_b = (boxes_b[:, 3] * boxes_b[:, 4] * boxes_b[:, 5]).view(1, -1) iou3d = overlaps_3d / torch.clamp(vol_a + vol_b - overlaps_3d, min=1e-6) return iou3d def nms_gpu(boxes, scores, thresh, pre_maxsize=None, **kwargs): """ :param boxes: (N, 7) [x, y, z, dx, dy, dz, heading] :param scores: (N) :param thresh: :return: """ assert boxes.shape[1] == 7 order = scores.sort(0, descending=True)[1] if pre_maxsize is not None: order = order[:pre_maxsize] boxes = boxes[order].contiguous() keep = torch.LongTensor(boxes.size(0)) num_out = iou3d_nms_cuda.nms_gpu(boxes, keep, thresh) return order[keep[:num_out].cuda()].contiguous(), None def nms_normal_gpu(boxes, scores, thresh, **kwargs): """ :param boxes: (N, 7) [x, y, z, dx, dy, dz, heading] :param scores: (N) :param thresh: :return: """ assert boxes.shape[1] == 7 order = scores.sort(0, descending=True)[1] boxes = boxes[order].contiguous() keep = torch.LongTensor(boxes.size(0)) num_out = iou3d_nms_cuda.nms_normal_gpu(boxes, keep, thresh) return order[keep[:num_out].cuda()].contiguous(), None
3,650
30.205128
109
py
SASA
SASA-main/pcdet/ops/roiaware_pool3d/roiaware_pool3d_utils.py
import torch import torch.nn as nn from torch.autograd import Function from ...utils import common_utils from . import roiaware_pool3d_cuda def points_in_boxes_cpu(points, boxes): """ Args: points: (num_points, 3) boxes: [x, y, z, dx, dy, dz, heading], (x, y, z) is the box center, each box DO NOT overlaps Returns: point_indices: (N, num_points) """ assert boxes.shape[1] == 7 assert points.shape[1] == 3 points, is_numpy = common_utils.check_numpy_to_torch(points) boxes, is_numpy = common_utils.check_numpy_to_torch(boxes) point_indices = points.new_zeros((boxes.shape[0], points.shape[0]), dtype=torch.int) roiaware_pool3d_cuda.points_in_boxes_cpu(boxes.float().contiguous(), points.float().contiguous(), point_indices) return point_indices.numpy() if is_numpy else point_indices def points_in_boxes_gpu(points, boxes): """ :param points: (B, M, 3) :param boxes: (B, T, 7), num_valid_boxes <= T :return box_idxs_of_pts: (B, M), default background = -1 """ assert boxes.shape[0] == points.shape[0] assert boxes.shape[2] == 7 and points.shape[2] == 3 batch_size, num_points, _ = points.shape box_idxs_of_pts = points.new_zeros((batch_size, num_points), dtype=torch.int).fill_(-1) roiaware_pool3d_cuda.points_in_boxes_gpu(boxes.contiguous(), points.contiguous(), box_idxs_of_pts) return box_idxs_of_pts def points_to_boxes_dist_gpu(points, boxes): """ Args: points: (B, M, 3) boxes: (B, T, 7), num_valid_boxes <= T Returns: pts_dist: (B, M), distance from pts to the nearest box """ assert boxes.shape[0] == points.shape[0] assert boxes.shape[2] == 7 and points.shape[2] == 3 batch_size, num_points, _ = points.shape num_boxes = boxes.shape[1] pts_dist = points.new_zeros((batch_size, num_points), dtype=torch.float32).fill_(1e5) roiaware_pool3d_cuda.points_to_boxes_dist_gpu(boxes.contiguous(), points.contiguous(), pts_dist) return pts_dist class RoIAwarePool3d(nn.Module): def __init__(self, out_size, max_pts_each_voxel=128): super().__init__() self.out_size = out_size self.max_pts_each_voxel = max_pts_each_voxel def forward(self, rois, pts, pts_feature, pool_method='max'): assert pool_method in ['max', 'avg'] return RoIAwarePool3dFunction.apply(rois, pts, pts_feature, self.out_size, self.max_pts_each_voxel, pool_method) class RoIAwarePool3dFunction(Function): @staticmethod def forward(ctx, rois, pts, pts_feature, out_size, max_pts_each_voxel, pool_method): """ Args: ctx: rois: (N, 7) [x, y, z, dx, dy, dz, heading] (x, y, z) is the box center pts: (npoints, 3) pts_feature: (npoints, C) out_size: int or tuple, like 7 or (7, 7, 7) max_pts_each_voxel: pool_method: 'max' or 'avg' Returns: pooled_features: (N, out_x, out_y, out_z, C) """ assert rois.shape[1] == 7 and pts.shape[1] == 3 if isinstance(out_size, int): out_x = out_y = out_z = out_size else: assert len(out_size) == 3 for k in range(3): assert isinstance(out_size[k], int) out_x, out_y, out_z = out_size num_rois = rois.shape[0] num_channels = pts_feature.shape[-1] num_pts = pts.shape[0] pooled_features = pts_feature.new_zeros((num_rois, out_x, out_y, out_z, num_channels)) argmax = pts_feature.new_zeros((num_rois, out_x, out_y, out_z, num_channels), dtype=torch.int) pts_idx_of_voxels = pts_feature.new_zeros((num_rois, out_x, out_y, out_z, max_pts_each_voxel), dtype=torch.int) pool_method_map = {'max': 0, 'avg': 1} pool_method = pool_method_map[pool_method] roiaware_pool3d_cuda.forward(rois, pts, pts_feature, argmax, pts_idx_of_voxels, pooled_features, pool_method) ctx.roiaware_pool3d_for_backward = (pts_idx_of_voxels, argmax, pool_method, num_pts, num_channels) return pooled_features @staticmethod def backward(ctx, grad_out): """ :param grad_out: (N, out_x, out_y, out_z, C) :return: grad_in: (npoints, C) """ pts_idx_of_voxels, argmax, pool_method, num_pts, num_channels = ctx.roiaware_pool3d_for_backward grad_in = grad_out.new_zeros((num_pts, num_channels)) roiaware_pool3d_cuda.backward(pts_idx_of_voxels, argmax, grad_out.contiguous(), grad_in, pool_method) return None, None, grad_in, None, None, None if __name__ == '__main__': pass
4,688
34.522727
120
py
chainer
chainer-master/setup.py
#!/usr/bin/env python import os import pkg_resources import sys from setuptools import setup import chainerx_build_helper if sys.version_info[:3] == (3, 5, 0): if not int(os.getenv('CHAINER_PYTHON_350_FORCE', '0')): msg = """ Chainer does not work with Python 3.5.0. We strongly recommend to use another version of Python. If you want to use Chainer with Python 3.5.0 at your own risk, set CHAINER_PYTHON_350_FORCE environment variable to 1.""" print(msg) sys.exit(1) requirements = { 'install': [ 'setuptools', 'typing_extensions', 'filelock', 'numpy>=1.9.0', 'protobuf>=3.0.0', 'six>=1.9.0', ], 'stylecheck': [ 'autopep8>=1.4.1,<1.5', 'flake8>=3.7,<3.8', 'pycodestyle>=2.5,<2.6', ], 'test': [ 'pytest<4.2.0', # 4.2.0 is slow collecting tests and times out on CI. 'attrs<19.2.0', # pytest 4.1.1 does not run with attrs==19.2.0 'mock', ], 'doctest': [ 'sphinx==1.8.2', 'matplotlib', 'theano', ], 'docs': [ 'sphinx==1.8.2', 'sphinx_rtd_theme', 'onnx<1.7.0', 'packaging', ], 'appveyor': [ '-r test', # pytest-timeout>=1.3.0 requires pytest>=3.6. # TODO(niboshi): Consider upgrading pytest to >=3.6 'pytest-timeout<1.3.0', ], 'jenkins': [ '-r test', # pytest-timeout>=1.3.0 requires pytest>=3.6. # TODO(niboshi): Consider upgrading pytest to >=3.6 'pytest-timeout<1.3.0', 'pytest-cov<2.10', # pytest-cov 2.10 requires pytest>=4.6 'nose', 'coveralls', 'codecov', 'coverage<5', # Otherwise, Python must be built with sqlite ], } def reduce_requirements(key): # Resolve recursive requirements notation (-r) reqs = requirements[key] resolved_reqs = [] for req in reqs: if req.startswith('-r'): depend_key = req[2:].lstrip() reduce_requirements(depend_key) resolved_reqs += requirements[depend_key] else: resolved_reqs.append(req) requirements[key] = resolved_reqs for k in requirements.keys(): reduce_requirements(k) extras_require = {k: v for k, v in requirements.items() if k != 'install'} setup_requires = [] install_requires = requirements['install'] tests_require = requirements['test'] def find_any_distribution(pkgs): for pkg in pkgs: try: return pkg_resources.get_distribution(pkg) except pkg_resources.DistributionNotFound: pass return None for pkg_name in ('ChainerMN', 'ONNX-Chainer'): distribution_name = pkg_name.lower().replace('-', '_') found_error = find_any_distribution([distribution_name]) if found_error is not None: msg = """ We detected that {name} is installed in your environment. {name} has been integrated to Chainer and no separate installation is necessary. Please uninstall the old {name} in advance. """ print(msg.format(name=pkg_name)) exit(1) here = os.path.abspath(os.path.dirname(__file__)) # Get __version__ variable exec(open(os.path.join(here, 'chainer', '_version.py')).read()) setup_kwargs = dict( name='chainer', version=__version__, # NOQA description='A flexible framework of neural networks', long_description=open('README.md').read(), long_description_content_type='text/markdown', author='Seiya Tokui', author_email='tokui@preferred.jp', url='https://chainer.org/', license='MIT License', packages=['chainer', 'chainer.backends', 'chainer.dataset', 'chainer.dataset.tabular', 'chainer.datasets', 'chainer.distributions', 'chainer.exporters', 'chainer.functions', 'chainer.functions.activation', 'chainer.functions.array', 'chainer.functions.connection', 'chainer.functions.evaluation', 'chainer.functions.loss', 'chainer.functions.math', 'chainer.functions.noise', 'chainer.functions.normalization', 'chainer.functions.pooling', 'chainer.functions.rnn', 'chainer.functions.theano', 'chainer.functions.util', 'chainer.function_hooks', 'chainer.iterators', 'chainer.initializers', 'chainer.links', 'chainer.links.activation', 'chainer.links.caffe', 'chainer.links.caffe.protobuf3', 'chainer.links.connection', 'chainer.links.loss', 'chainer.links.model', 'chainer.links.model.vision', 'chainer.links.normalization', 'chainer.links.rnn', 'chainer.links.theano', 'chainer.link_hooks', 'chainer.graph_optimizations', 'chainer.optimizers', 'chainer.optimizer_hooks', 'chainer.serializers', 'chainer.testing', 'chainer.training', 'chainer.training.extensions', 'chainer.training.triggers', 'chainer.training.updaters', 'chainer.utils', 'chainermn', 'chainermn.communicators', 'chainermn.datasets', 'chainermn.extensions', 'chainermn.functions', 'chainermn.iterators', 'chainermn.links', 'chainermn.testing', 'onnx_chainer', 'onnx_chainer.functions', 'onnx_chainer.testing'], package_data={ 'chainer': ['py.typed'], }, zip_safe=False, setup_requires=setup_requires, install_requires=install_requires, tests_require=tests_require, extras_require=extras_require, python_requires='>=3.5.0', ) build_chainerx = 0 != int(os.getenv('CHAINER_BUILD_CHAINERX', '0')) if (os.getenv('READTHEDOCS', None) == 'True' and os.getenv('READTHEDOCS_PROJECT', None) == 'chainer'): # ChainerX must be built in order to build the docs (on Read the Docs). build_chainerx = True # Try to prevent Read the Docs build timeouts. os.environ['MAKEFLAGS'] = '-j2' chainerx_build_helper.config_setup_kwargs(setup_kwargs, build_chainerx) setup(**setup_kwargs)
6,477
28.990741
78
py
chainer
chainer-master/examples/caffe_export/export.py
import argparse import os import numpy as np import chainer from chainer.exporters import caffe from chainer.links.model.vision import googlenet from chainer.links.model.vision import resnet from chainer.links.model.vision import vgg archs = { 'googlenet': googlenet.GoogLeNet, 'resnet50': resnet.ResNet50Layers, 'resnet101': resnet.ResNet101Layers, 'resnet152': resnet.ResNet152Layers, 'vgg16': vgg.VGG16Layers, } class DumpModel(chainer.Chain): def __init__(self, arch_name): super(DumpModel, self).__init__() with self.init_scope(): self.base_model = archs[arch_name]() def forward(self, img): return self.base_model(img, layers=['prob'])['prob'] def get_network_for_imagenet(arch_name): model = DumpModel(arch_name) input_image = np.ones((1, 3, 224, 224), dtype=np.float32) input = chainer.Variable(input_image) return model, input def main(): parser = argparse.ArgumentParser(description='Export') parser.add_argument( '--arch', '-a', type=str, required=True, choices=archs.keys(), help='Arch name. models: ' + ', '.join(archs.keys()) + '.') parser.add_argument( '--out-dir', '-o', type=str, required=True, help='Output directory name. ' 'chainer_model.prototxt, chainer_model.caffemodel' ' will be created in it') args = parser.parse_args() if not os.path.exists(args.out_dir): print('Created output directory: ' + args.out_dir) os.mkdir(args.out_dir) else: print('Overwriting the existing directory: ' + args.out_dir) if not os.path.isdir(args.out_dir): raise ValueError(args.out_dir + ' exists but not a directory!') print('load model') model, input = get_network_for_imagenet(args.arch) print('convert to caffe model') caffe.export(model, [input], args.out_dir, True) if __name__ == '__main__': main()
1,953
27.318841
71
py
chainer
chainer-master/examples/chainermn/cifar/models/VGG.py
from chainer.utils import argument import chainer import chainer.functions as F import chainer.links as L import warnings class Block(chainer.Chain): """A convolution, batch norm, ReLU block. A block in a feedforward network that performs a convolution followed by batch normalization followed by a ReLU activation. For the convolution operation, a square filter size is used. Args: out_channels (int): The number of output channels. ksize (int): The size of the filter is ksize x ksize. pad (int): The padding to use for the convolution. """ def __init__(self, out_channels, ksize, pad=1): super(Block, self).__init__() with self.init_scope(): self.conv = L.Convolution2D(None, out_channels, ksize, pad=pad, nobias=True) self.bn = L.BatchNormalization(out_channels) def forward(self, x): h = self.conv(x) h = self.bn(h) return F.relu(h) class VGG(chainer.Chain): """A VGG-style network for very small images. This model is based on the VGG-style model from http://torch.ch/blog/2015/07/30/cifar.html which is based on the network architecture from the paper: https://arxiv.org/pdf/1409.1556v6.pdf This model is intended to be used with either RGB or greyscale input images that are of size 32x32 pixels, such as those in the CIFAR10 and CIFAR100 datasets. On CIFAR10, it achieves approximately 89% accuracy on the test set with no data augmentation. On CIFAR100, it achieves approximately 63% accuracy on the test set with no data augmentation. Args: n_class_labels (int): The number of class labels. num_class_labels (int): The number of class labels. (deprecated) """ def __init__(self, n_class_labels=10, **kwargs): old_n_class_labels = argument.parse_kwargs( kwargs, ('num_class_labels', None)) if old_n_class_labels is not None: n_class_labels = old_n_class_labels warnings.warn( 'num_class_labels is deprecated.' 'Please consider using num_class_labels', DeprecationWarning) super(VGG, self).__init__() with self.init_scope(): self.block1_1 = Block(64, 3) self.block1_2 = Block(64, 3) self.block2_1 = Block(128, 3) self.block2_2 = Block(128, 3) self.block3_1 = Block(256, 3) self.block3_2 = Block(256, 3) self.block3_3 = Block(256, 3) self.block4_1 = Block(512, 3) self.block4_2 = Block(512, 3) self.block4_3 = Block(512, 3) self.block5_1 = Block(512, 3) self.block5_2 = Block(512, 3) self.block5_3 = Block(512, 3) self.fc1 = L.Linear(None, 512, nobias=True) self.bn_fc1 = L.BatchNormalization(512) self.fc2 = L.Linear(None, n_class_labels, nobias=True) def forward(self, x): # 64 channel blocks: h = self.block1_1(x) h = F.dropout(h, ratio=0.3) h = self.block1_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 128 channel blocks: h = self.block2_1(h) h = F.dropout(h, ratio=0.4) h = self.block2_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 256 channel blocks: h = self.block3_1(h) h = F.dropout(h, ratio=0.4) h = self.block3_2(h) h = F.dropout(h, ratio=0.4) h = self.block3_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block4_1(h) h = F.dropout(h, ratio=0.4) h = self.block4_2(h) h = F.dropout(h, ratio=0.4) h = self.block4_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block5_1(h) h = F.dropout(h, ratio=0.4) h = self.block5_2(h) h = F.dropout(h, ratio=0.4) h = self.block5_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) h = F.dropout(h, ratio=0.5) h = self.fc1(h) h = self.bn_fc1(h) h = F.relu(h) h = F.dropout(h, ratio=0.5) return self.fc2(h)
4,303
30.647059
76
py
chainer
chainer-master/examples/text_classification/text_datasets.py
import csv import glob import io import os import shutil import sys import tarfile import tempfile import numpy import chainer from nlp_utils import make_vocab from nlp_utils import normalize_text from nlp_utils import split_text from nlp_utils import transform_to_array URL_DBPEDIA = 'https://github.com/le-scientifique/torchDatasets/raw/master/dbpedia_csv.tar.gz' # NOQA URL_IMDB = 'https://ai.stanford.edu/~amaas/data/sentiment/aclImdb_v1.tar.gz' URL_OTHER_BASE = 'https://raw.githubusercontent.com/harvardnlp/sent-conv-torch/master/data/' # NOQA def download_dbpedia(): path = chainer.dataset.cached_download(URL_DBPEDIA) tf = tarfile.open(path, 'r') return tf def read_dbpedia(tf, split, shrink=1, char_based=False): dataset = [] f = tf.extractfile('dbpedia_csv/{}.csv'.format(split)) if sys.version_info > (3, 0): f = io.TextIOWrapper(f, encoding='utf-8') for i, (label, title, text) in enumerate(csv.reader(f)): if i % shrink != 0: continue label = int(label) - 1 # Index begins from 1 tokens = split_text(normalize_text(text), char_based) dataset.append((tokens, label)) return dataset def get_dbpedia(vocab=None, shrink=1, char_based=False): tf = download_dbpedia() print('read dbpedia') train = read_dbpedia(tf, 'train', shrink=shrink, char_based=char_based) test = read_dbpedia(tf, 'test', shrink=shrink, char_based=char_based) if vocab is None: print('constract vocabulary based on frequency') vocab = make_vocab(train) train = transform_to_array(train, vocab) test = transform_to_array(test, vocab) return train, test, vocab def download_imdb(): path = chainer.dataset.cached_download(URL_IMDB) tf = tarfile.open(path, 'r') # To read many files fast, tarfile is untared path = tempfile.mkdtemp() tf.extractall(path) return path def read_imdb(path, split, shrink=1, fine_grained=False, char_based=False): fg_label_dict = {'1': 0, '2': 0, '3': 1, '4': 1, '7': 2, '8': 2, '9': 3, '10': 3} def read_and_label(posneg, label): dataset = [] target = os.path.join(path, 'aclImdb', split, posneg, '*') for i, f_path in enumerate(glob.glob(target)): if i % shrink != 0: continue with io.open(f_path, encoding='utf-8', errors='ignore') as f: text = f.read().strip() tokens = split_text(normalize_text(text), char_based) if fine_grained: # extract from f_path. e.g. /pos/200_8.txt -> 8 label = fg_label_dict[f_path.split('_')[-1][:-4]] dataset.append((tokens, label)) else: dataset.append((tokens, label)) return dataset pos_dataset = read_and_label('pos', 0) neg_dataset = read_and_label('neg', 1) return pos_dataset + neg_dataset def get_imdb(vocab=None, shrink=1, fine_grained=False, char_based=False): tmp_path = download_imdb() print('read imdb') train = read_imdb(tmp_path, 'train', shrink=shrink, fine_grained=fine_grained, char_based=char_based) test = read_imdb(tmp_path, 'test', shrink=shrink, fine_grained=fine_grained, char_based=char_based) shutil.rmtree(tmp_path) if vocab is None: print('constract vocabulary based on frequency') vocab = make_vocab(train) train = transform_to_array(train, vocab) test = transform_to_array(test, vocab) return train, test, vocab def download_other_dataset(name): if name in ['custrev', 'mpqa', 'rt-polarity', 'subj']: files = [name + '.all'] elif name == 'TREC': files = [name + suff for suff in ['.train.all', '.test.all']] else: files = [name + suff for suff in ['.train', '.test']] file_paths = [] for f_name in files: url = os.path.join(URL_OTHER_BASE, f_name) path = chainer.dataset.cached_download(url) file_paths.append(path) return file_paths def read_other_dataset(path, shrink=1, char_based=False): dataset = [] with io.open(path, encoding='utf-8', errors='ignore') as f: for i, l in enumerate(f): if i % shrink != 0 or not len(l.strip()) >= 3: continue label, text = l.strip().split(None, 1) label = int(label) tokens = split_text(normalize_text(text), char_based) dataset.append((tokens, label)) return dataset def get_other_text_dataset(name, vocab=None, shrink=1, char_based=False, seed=777): assert(name in ['TREC', 'stsa.binary', 'stsa.fine', 'custrev', 'mpqa', 'rt-polarity', 'subj']) datasets = download_other_dataset(name) train = read_other_dataset( datasets[0], shrink=shrink, char_based=char_based) if len(datasets) == 2: test = read_other_dataset( datasets[1], shrink=shrink, char_based=char_based) else: numpy.random.seed(seed) alldata = numpy.random.permutation(train) train = alldata[:-len(alldata) // 10] test = alldata[-len(alldata) // 10:] if vocab is None: print('constract vocabulary based on frequency') vocab = make_vocab(train) train = transform_to_array(train, vocab) test = transform_to_array(test, vocab) return train, test, vocab
5,533
30.988439
102
py
chainer
chainer-master/examples/cifar/models/VGG.py
import chainer import chainer.functions as F import chainer.links as L class Block(chainer.Chain): """A convolution, batch norm, ReLU block. A block in a feedforward network that performs a convolution followed by batch normalization followed by a ReLU activation. For the convolution operation, a square filter size is used. Args: out_channels (int): The number of output channels. ksize (int): The size of the filter is ksize x ksize. pad (int): The padding to use for the convolution. """ def __init__(self, out_channels, ksize, pad=1): super(Block, self).__init__() with self.init_scope(): self.conv = L.Convolution2D(None, out_channels, ksize, pad=pad, nobias=True) self.bn = L.BatchNormalization(out_channels) def forward(self, x): h = self.conv(x) h = self.bn(h) return F.relu(h) class VGG(chainer.Chain): """A VGG-style network for very small images. This model is based on the VGG-style model from http://torch.ch/blog/2015/07/30/cifar.html which is based on the network architecture from the paper: https://arxiv.org/pdf/1409.1556v6.pdf This model is intended to be used with either RGB or greyscale input images that are of size 32x32 pixels, such as those in the CIFAR10 and CIFAR100 datasets. On CIFAR10, it achieves approximately 89% accuracy on the test set with no data augmentation. On CIFAR100, it achieves approximately 63% accuracy on the test set with no data augmentation. Args: class_labels (int): The number of class labels. """ def __init__(self, class_labels=10): super(VGG, self).__init__() with self.init_scope(): self.block1_1 = Block(64, 3) self.block1_2 = Block(64, 3) self.block2_1 = Block(128, 3) self.block2_2 = Block(128, 3) self.block3_1 = Block(256, 3) self.block3_2 = Block(256, 3) self.block3_3 = Block(256, 3) self.block4_1 = Block(512, 3) self.block4_2 = Block(512, 3) self.block4_3 = Block(512, 3) self.block5_1 = Block(512, 3) self.block5_2 = Block(512, 3) self.block5_3 = Block(512, 3) self.fc1 = L.Linear(None, 512, nobias=True) self.bn_fc1 = L.BatchNormalization(512) self.fc2 = L.Linear(None, class_labels, nobias=True) def forward(self, x): # 64 channel blocks: h = self.block1_1(x) h = F.dropout(h, ratio=0.3) h = self.block1_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 128 channel blocks: h = self.block2_1(h) h = F.dropout(h, ratio=0.4) h = self.block2_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 256 channel blocks: h = self.block3_1(h) h = F.dropout(h, ratio=0.4) h = self.block3_2(h) h = F.dropout(h, ratio=0.4) h = self.block3_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block4_1(h) h = F.dropout(h, ratio=0.4) h = self.block4_2(h) h = F.dropout(h, ratio=0.4) h = self.block4_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block5_1(h) h = F.dropout(h, ratio=0.4) h = self.block5_2(h) h = F.dropout(h, ratio=0.4) h = self.block5_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) h = F.dropout(h, ratio=0.5) h = self.fc1(h) h = self.bn_fc1(h) h = F.relu(h) h = F.dropout(h, ratio=0.5) return self.fc2(h)
3,799
29.894309
76
py
chainer
chainer-master/examples/modelzoo/download_model.py
#!/usr/bin/env python import argparse import zipfile import six parser = argparse.ArgumentParser( description='Download a Caffe reference model') parser.add_argument('model_type', choices=('alexnet', 'caffenet', 'googlenet', 'resnet'), help='Model type (alexnet, caffenet, googlenet)') args = parser.parse_args() if args.model_type == 'alexnet': url = 'http://dl.caffe.berkeleyvision.org/bvlc_alexnet.caffemodel' name = 'bvlc_alexnet.caffemodel' elif args.model_type == 'caffenet': url = 'http://dl.caffe.berkeleyvision.org/' \ 'bvlc_reference_caffenet.caffemodel' name = 'bvlc_reference_caffenet.caffemodel' elif args.model_type == 'googlenet': url = 'http://dl.caffe.berkeleyvision.org/bvlc_googlenet.caffemodel' name = 'bvlc_googlenet.caffemodel' elif args.model_type == 'resnet': raise RuntimeError('The resnet model file cannot be downloaded ' 'automatically. Please download manually: ' 'https://github.com/KaimingHe/deep-residual-networks' '#models') else: raise RuntimeError('Invalid model type. Choose from ' 'alexnet, caffenet and googlenet.') print('Downloading model file...') six.moves.urllib.request.urlretrieve(url, name) if args.model_type == 'resnet': with zipfile.ZipFile(name, 'r') as zf: zf.extractall('.') print('Done')
1,438
33.261905
76
py
chainer
chainer-master/examples/modelzoo/download_mean_file.py
#!/usr/bin/env python import six print('Downloading ILSVRC12 mean file for NumPy...') six.moves.urllib.request.urlretrieve( 'https://github.com/BVLC/caffe/raw/master/python/caffe/imagenet/' 'ilsvrc_2012_mean.npy', 'ilsvrc_2012_mean.npy') print('Done')
266
23.272727
69
py
chainer
chainer-master/examples/modelzoo/evaluate_caffe_net.py
#!/usr/bin/env python """Example code of evaluating a Caffe reference model for ILSVRC2012 task. Prerequisite: To run this example, crop the center of ILSVRC2012 validation images and scale them to 256x256, and make a list of space-separated CSV each column of which contains a full path to an image at the fist column and a zero- origin label at the second column (this format is same as that used by Caffe's ImageDataLayer). """ import argparse import os import sys import numpy as np from PIL import Image import chainer import chainer.functions as F from chainer.links import caffe parser = argparse.ArgumentParser( description='Evaluate a Caffe reference model on ILSVRC2012 dataset') parser.add_argument('dataset', help='Path to validation image-label list file') parser.add_argument('model_type', choices=('alexnet', 'caffenet', 'googlenet', 'resnet'), help='Model type (alexnet, caffenet, googlenet, resnet)') parser.add_argument('model', help='Path to the pretrained Caffe model') parser.add_argument('--basepath', '-b', default='/', help='Base path for images in the dataset') parser.add_argument('--mean', '-m', default='ilsvrc_2012_mean.npy', help='Path to the mean file') parser.add_argument('--batchsize', '-B', type=int, default=100, help='Minibatch size') parser.add_argument('--device', '-d', type=str, default='-1', help='Device specifier. Either ChainerX device ' 'specifier or an integer. If non-negative integer, ' 'CuPy arrays with specified device id are used. If ' 'negative integer, NumPy arrays are used') parser.set_defaults(test=False) group = parser.add_argument_group('deprecated arguments') group.add_argument('--gpu', '-g', dest='device', type=int, nargs='?', const=-1, help='GPU ID (negative value indicates CPU)') args = parser.parse_args() device = chainer.get_device(args.device) device.use() xp = device.xp assert args.batchsize > 0 chainer.config.train = False # All the codes will run in test mode dataset = [] with open(args.dataset) as list_file: for line in list_file: pair = line.strip().split() path = os.path.join(args.basepath, pair[0]) dataset.append((path, np.int32(pair[1]))) assert len(dataset) % args.batchsize == 0 print('Loading Caffe model file %s...' % args.model) func = caffe.CaffeFunction(args.model) print('Loaded') func.to_device(device) if args.model_type == 'alexnet' or args.model_type == 'caffenet': in_size = 227 mean_image = np.load(args.mean) def forward(x, t): y, = func(inputs={'data': x}, outputs=['fc8']) return F.softmax_cross_entropy(y, t), F.accuracy(y, t) elif args.model_type == 'googlenet': in_size = 224 # Constant mean over spatial pixels mean_image = np.ndarray((3, 256, 256), dtype=np.float32) mean_image[0] = 104 mean_image[1] = 117 mean_image[2] = 123 def forward(x, t): y, = func(inputs={'data': x}, outputs=['loss3/classifier'], disable=['loss1/ave_pool', 'loss2/ave_pool']) return F.softmax_cross_entropy(y, t), F.accuracy(y, t) elif args.model_type == 'resnet': in_size = 224 mean_image = np.load(args.mean) def forward(x, t): y, = func(inputs={'data': x}, outputs=['prob']) return F.softmax_cross_entropy(y, t), F.accuracy(y, t) cropwidth = 256 - in_size start = cropwidth // 2 stop = start + in_size mean_image = mean_image[:, start:stop, start:stop].copy() x_batch = np.ndarray((args.batchsize, 3, in_size, in_size), dtype=np.float32) y_batch = np.ndarray((args.batchsize,), dtype=np.int32) i = 0 count = 0 accum_loss = 0 accum_accuracy = 0 for path, label in dataset: image = np.asarray(Image.open(path)).transpose(2, 0, 1)[::-1] image = image[:, start:stop, start:stop].astype(np.float32) image -= mean_image x_batch[i] = image y_batch[i] = label i += 1 if i == args.batchsize: x = xp.asarray(x_batch) y = xp.asarray(y_batch) with chainer.no_backprop_mode(): loss, accuracy = forward(x, y) accum_loss += float(loss.array) * args.batchsize accum_accuracy += float(accuracy.array) * args.batchsize del x, y, loss, accuracy count += args.batchsize sys.stdout.write('{} / {}\r'.format(count, len(dataset))) sys.stdout.flush() i = 0 print('mean loss: {}'.format(accum_loss / count)) print('mean accuracy: {}'.format(accum_accuracy / count))
4,652
31.767606
79
py
chainer
chainer-master/examples/static_graph_optimizations/cifar/models/VGG.py
import chainer import chainer.functions as F import chainer.links as L from chainer import static_graph class Block(chainer.Chain): """A convolution, batch norm, ReLU block. A block in a feedforward network that performs a convolution followed by batch normalization followed by a ReLU activation. For the convolution operation, a square filter size is used. Args: out_channels (int): The number of output channels. ksize (int): The size of the filter is ksize x ksize. pad (int): The padding to use for the convolution. """ def __init__(self, out_channels, ksize, pad=1): super(Block, self).__init__() with self.init_scope(): self.conv = L.Convolution2D(None, out_channels, ksize, pad=pad, nobias=True) self.bn = L.BatchNormalization(out_channels) def __call__(self, x): h = self.conv(x) h = self.bn(h) return F.relu(h) class VGG(chainer.Chain): """A VGG-style network for very small images. This model is based on the VGG-style model from http://torch.ch/blog/2015/07/30/cifar.html which is based on the network architecture from the paper: https://arxiv.org/pdf/1409.1556v6.pdf This model is intended to be used with either RGB or greyscale input images that are of size 32x32 pixels, such as those in the CIFAR10 and CIFAR100 datasets. On CIFAR10, it achieves approximately 89% accuracy on the test set with no data augmentation. On CIFAR100, it achieves approximately 63% accuracy on the test set with no data augmentation. Args: class_labels (int): The number of class labels. """ def __init__(self, class_labels=10): super(VGG, self).__init__() with self.init_scope(): self.block1_1 = Block(64, 3) self.block1_2 = Block(64, 3) self.block2_1 = Block(128, 3) self.block2_2 = Block(128, 3) self.block3_1 = Block(256, 3) self.block3_2 = Block(256, 3) self.block3_3 = Block(256, 3) self.block4_1 = Block(512, 3) self.block4_2 = Block(512, 3) self.block4_3 = Block(512, 3) self.block5_1 = Block(512, 3) self.block5_2 = Block(512, 3) self.block5_3 = Block(512, 3) self.fc1 = L.Linear(None, 512, nobias=True) self.bn_fc1 = L.BatchNormalization(512) self.fc2 = L.Linear(None, class_labels, nobias=True) @static_graph def __call__(self, x): # 64 channel blocks: h = self.block1_1(x) h = F.dropout(h, ratio=0.3) h = self.block1_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 128 channel blocks: h = self.block2_1(h) h = F.dropout(h, ratio=0.4) h = self.block2_2(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 256 channel blocks: h = self.block3_1(h) h = F.dropout(h, ratio=0.4) h = self.block3_2(h) h = F.dropout(h, ratio=0.4) h = self.block3_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block4_1(h) h = F.dropout(h, ratio=0.4) h = self.block4_2(h) h = F.dropout(h, ratio=0.4) h = self.block4_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) # 512 channel blocks: h = self.block5_1(h) h = F.dropout(h, ratio=0.4) h = self.block5_2(h) h = F.dropout(h, ratio=0.4) h = self.block5_3(h) h = F.max_pooling_2d(h, ksize=2, stride=2) h = F.dropout(h, ratio=0.5) h = self.fc1(h) h = self.bn_fc1(h) h = F.relu(h) h = F.dropout(h, ratio=0.5) return self.fc2(h)
3,852
29.824
76
py
chainer
chainer-master/onnx_chainer/testing/test_mxnet.py
import collections import os import warnings import chainer import numpy as np from onnx_chainer.testing.test_onnxruntime import load_test_data try: import mxnet MXNET_AVAILABLE = True except ImportError: warnings.warn( 'MXNet is not installed. Please install mxnet to use ' 'testing utility for compatibility checking.', ImportWarning) MXNET_AVAILABLE = False def check_model_expect(test_path, input_names=None, rtol=1e-5, atol=1e-5): if not MXNET_AVAILABLE: raise ImportError('MXNet is not found on checking module.') model_path = os.path.join(test_path, 'model.onnx') sym, arg, aux = mxnet.contrib.onnx.import_model(model_path) mx_input_names = [graph_input for graph_input in sym.list_inputs() if graph_input not in arg and graph_input not in aux] if input_names is not None: assert list(sorted(input_names)) == list(sorted(mx_input_names)) test_data_sets = sorted([ p for p in os.listdir(test_path) if p.startswith('test_data_set_')]) for test_data in test_data_sets: test_data_path = os.path.join(test_path, test_data) assert os.path.isdir(test_data_path) inputs, outputs = load_test_data( test_data_path, mx_input_names, sym.list_outputs()) data_shapes = [(name, array.shape) for name, array in inputs.items()] mod = mxnet.mod.Module( symbol=sym, data_names=mx_input_names, context=mxnet.cpu(), label_names=None) mod.bind( for_training=chainer.config.train, data_shapes=data_shapes, label_shapes=None) mod.set_params( arg_params=arg, aux_params=aux, allow_missing=True, allow_extra=True) Batch = collections.namedtuple('Batch', ['data']) mx_input = [mxnet.nd.array(array) for array in inputs.values()] mod.forward(Batch(mx_input)) mx_outputs = [y.asnumpy() for y in mod.get_outputs()] for cy, my in zip(outputs.values(), mx_outputs): np.testing.assert_allclose(cy, my, rtol=rtol, atol=atol)
2,113
36.087719
77
py
chainer
chainer-master/tests/chainer_tests/exporters_tests/test_caffe.py
import os import unittest import warnings import numpy import chainer import chainer.functions as F import chainer.links as L from chainer import testing # The caffe submodule relies on protobuf which under protobuf==3.7.0 and # Python 3.7 raises a DeprecationWarning from the collections module. with warnings.catch_warnings(): warnings.filterwarnings(action='ignore', category=DeprecationWarning) from chainer.exporters import caffe # @testing.parameterize([ # {'layer': 'LinearFunction'}, # {'layer': 'Reshape'}, # {'layer': 'Convolution2DFunction'}, # {'layer': 'Deconvolution2DFunction'}, # {'layer': 'AveragePooling2D'}, # {'layer': 'MaxPooling2D'}, # {'layer': 'BatchNormalization'}, # {'layer': 'ReLU'}, # {'layer': 'LeakyReLU'}, # {'layer': 'Softmax'}, # {'layer': 'Sigmoid'}, # {'layer': 'Add'}, # ]) class TestCaffeExport(unittest.TestCase): def setUp(self): x = numpy.random.uniform(-1, 1, (1, 3, 7, 7)).astype(numpy.float32) self.x = chainer.Variable(x) def test_caffe_export_model(self): class Model(chainer.Chain): def __init__(self): super(Model, self).__init__() with self.init_scope(): self.l1 = L.Convolution2D(None, 1, 1, 1, 0, groups=1) self.b2 = L.BatchNormalization(1, eps=1e-2) self.l3 = L.Deconvolution2D(None, 1, 1, 1, 0, groups=1) self.l4 = L.Linear(None, 1) def forward(self, x): h = F.relu(self.l1(x)) h = self.b2(h) h = self.l3(h) return self.l4(h) assert_export_import_match(Model(), self.x) def test_Reshape(self): class Link(chainer.Chain): def forward(self, x): return F.reshape(x, (-1,)) assert_export_import_match(Link(), self.x) def test_LinearFunction(self): W = numpy.random.uniform(-1, 1, (1, numpy.prod(self.x.shape[1:]))) class Link(chainer.Chain): def __call__(self, x): return F.linear(x, W) assert_export_import_match(Link(), self.x) def test_AveragePooling2D(self): class Link(chainer.Chain): def forward(self, x): return F.average_pooling_2d(x, 1, 1, 0) assert_export_import_match(Link(), self.x) def test_MaxPooling2D(self): class Link(chainer.Chain): def forward(self, x): return F.max_pooling_2d(x, 1, 1, 0) assert_export_import_match(Link(), self.x) def test_LeakyReLU(self): class Link(chainer.Chain): def __call__(self, x): return F.leaky_relu(x, slope=0.1) assert_export_import_match(Link(), self.x) def test_Softmax(self): class Link(chainer.Chain): def forward(self, x): return F.softmax(x) assert_export_import_match(Link(), self.x) def test_Sigmoid(self): class Link(chainer.Chain): def forward(self, x): return F.sigmoid(x) assert_export_import_match(Link(), self.x) def test_Add(self): class Link(chainer.Chain): def forward(self, x): return x + x assert_export_import_match(Link(), self.x) def assert_export_import_match(l1, x): """Asserts that results from original Link and re-imported Link are close. """ l2 = export_and_import(l1, (x,), True) inputs = {'data': x} outputs = [l2.layers[-1][0]] with chainer.using_config('train', False): for v1, v2 in zip(l1(x), l2(inputs=inputs, outputs=outputs)[0]): testing.assert_allclose(v1.data, v2.data) def export_and_import(l, args, export_params=True, graph_name='test'): """Exports the given Link as Caffe model and returns the re-imported Link. """ with chainer.utils.tempdir() as tempdir: caffe.export(l, args, tempdir, export_params, graph_name) prototxt = os.path.join(tempdir, 'chainer_model.prototxt') caffemodel = os.path.join(tempdir, 'chainer_model.caffemodel') assert os.path.exists(prototxt) assert os.path.exists(caffemodel) # with open(prototxt) as f: print(f.read()) return L.caffe.CaffeFunction(caffemodel) testing.run_module(__name__, __file__)
4,411
28.218543
78
py
chainer
chainer-master/tests/chainer_tests/links_tests/caffe_tests/test_caffe_function.py
import os import tempfile import unittest import warnings import mock import numpy import six import chainer from chainer import links from chainer import testing # The caffe submodule relies on protobuf which under protobuf==3.7.0 and # Python 3.7 raises a DeprecationWarning from the collections module. with warnings.catch_warnings(): warnings.filterwarnings(action='ignore', category=DeprecationWarning) from chainer.links import caffe from chainer.links.caffe.caffe_function import caffe_pb def _iter_init(param, data): if isinstance(data, list): for d in data: if hasattr(param, 'add'): param.add() if isinstance(d, (list, dict)): _iter_init(param[-1], d) else: param[-1] = d else: param.append(d) elif isinstance(data, dict): for k, d in data.items(): if isinstance(d, (list, dict)): _iter_init(getattr(param, k), d) else: setattr(param, k, d) else: setattr(param, data) def _make_param(data): param = caffe_pb.NetParameter() _iter_init(param, data) return param class TestCaffeFunctionBase(unittest.TestCase): def setUp(self): param = _make_param(self.data) # The name can be used to open the file a second time, # while the named temporary file is still open on the Windows. with tempfile.NamedTemporaryFile(delete=False) as f: self.temp_file_path = f.name f.write(param.SerializeToString()) def tearDown(self): os.remove(self.temp_file_path) def init_func(self): self.func = caffe.CaffeFunction(self.temp_file_path) class TestCaffeFunctionBaseMock(TestCaffeFunctionBase): def setUp(self): outs = [] for shape in self.out_shapes: out_data = numpy.random.uniform(-1, 1, shape).astype(numpy.float32) outs.append(chainer.Variable(out_data)) self.outputs = tuple(outs) ret_value = outs[0] if len(outs) == 1 else tuple(outs) m = mock.MagicMock(name=self.func_name, return_value=ret_value) self.patch = mock.patch(self.func_name, m) self.mock = self.patch.start() super(TestCaffeFunctionBaseMock, self).setUp() def tearDown(self): super(TestCaffeFunctionBaseMock, self).tearDown() self.patch.stop() def call(self, inputs, outputs): invars = [] for shape in self.in_shapes: data = numpy.random.uniform(-1, 1, shape).astype(numpy.float32) invars.append(chainer.Variable(data)) self.inputs = invars with chainer.using_config('train', False): out = self.func(inputs=dict(zip(inputs, invars)), outputs=outputs) self.assertEqual(len(out), len(self.outputs)) for actual, expect in zip(out, self.outputs): self.assertIs(actual, expect) class TestConcat(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.concat' in_shapes = [(3, 2, 3), (3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Concat', 'bottom': ['x', 'y'], 'top': ['z'], 'concat_param': { 'axis': 2 } } ] } def test_concat(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x', 'y'], ['z']) self.mock.assert_called_once_with( (self.inputs[0], self.inputs[1]), axis=2) class TestConvolution(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Convolution2D.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Convolution', 'bottom': ['x'], 'top': ['y'], 'convolution_param': { 'kernel_size': [2], 'stride': [3], 'pad': [4], 'group': 3, 'bias_term': True, }, 'blobs': [ { 'num': 6, 'channels': 4, 'data': list(range(96)) }, { 'data': list(range(6)) } ] } ] } def test_convolution(self): self.init_func() self.assertEqual(len(self.func.layers), 1) f = self.func.l1 self.assertIsInstance(f, links.Convolution2D) for i in range(3): # 3 == group in_slice = slice(i * 4, (i + 1) * 4) # 4 == channels out_slice = slice(i * 2, (i + 1) * 2) # 2 == num / group w = f.W.data[out_slice, in_slice] numpy.testing.assert_array_equal( w.flatten(), range(i * 32, (i + 1) * 32)) numpy.testing.assert_array_equal( f.b.data, range(6)) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestDeconvolution(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Deconvolution2D.__call__' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Deconvolution', 'bottom': ['x'], 'top': ['y'], 'convolution_param': { 'kernel_size': [2], 'stride': [3], 'pad': [4], 'group': 3, 'bias_term': True, }, 'blobs': [ { 'num': 6, 'channels': 4, 'data': list(range(96)) }, { 'data': list(range(12)) } ] } ] } def test_deconvolution(self): self.init_func() self.assertEqual(len(self.func.layers), 1) f = self.func.l1 self.assertIsInstance(f, links.Deconvolution2D) for i in range(3): # 3 == group in_slice = slice(i * 4, (i + 1) * 4) # 4 == channels out_slice = slice(i * 2, (i + 1) * 2) # 2 == num / group w = f.W.data[out_slice, in_slice] numpy.testing.assert_array_equal( w.flatten(), range(i * 32, (i + 1) * 32)) numpy.testing.assert_array_equal( f.b.data, range(12)) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestData(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'Data', } ] } def test_data(self): self.init_func() self.assertEqual(len(self.func.layers), 0) class TestDropout(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.dropout' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Dropout', 'bottom': ['x'], 'top': ['y'], 'dropout_param': { 'dropout_ratio': 0.25 } } ] } def test_dropout(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with( self.inputs[0], ratio=0.25) class TestInnerProduct(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Linear.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'InnerProduct', 'bottom': ['x'], 'top': ['y'], 'inner_product_param': { 'bias_term': True, 'axis': 1 }, 'blobs': [ # weight { 'shape': { 'dim': [2, 3] }, 'data': list(range(6)), }, # bias { 'shape': { 'dim': [2] }, 'data': list(range(2)), } ] } ] } def test_linear(self): self.init_func() f = self.func.l1 self.assertIsInstance(f, links.Linear) numpy.testing.assert_array_equal( f.W.data, numpy.array([[0, 1, 2], [3, 4, 5]], dtype=numpy.float32)) numpy.testing.assert_array_equal( f.b.data, numpy.array([0, 1], dtype=numpy.float32)) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestInnerProductDim4(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Linear.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'InnerProduct', 'bottom': ['x'], 'top': ['y'], 'inner_product_param': { 'bias_term': False, 'axis': 1 }, 'blobs': [ # weight { 'shape': { 'dim': [4, 5, 2, 3] }, # when `ndim` == 4, `data` stored shape[2] x shape[3] # data 'data': list(range(6)), } ] } ] } def test_linear(self): self.init_func() f = self.func.l1 self.assertIsInstance(f, links.Linear) numpy.testing.assert_array_equal( f.W.data, numpy.array([[0, 1, 2], [3, 4, 5]], dtype=numpy.float32)) self.assertIsNone(f.b) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestInnerProductInvalidDim(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'InnerProduct', 'blobs': [ { 'shape': { 'dim': [2, 3, 4, 5, 6] # 5-dim is not supported }, }, ] } ] } def test_linear(self): with self.assertRaises(RuntimeError): self.init_func() class TestInnerProductNonDefaultAxis(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'InnerProduct', 'inner_product_param': { 'axis': 0 # non-default axis } } ] } def test_linear(self): with self.assertRaises(RuntimeError): self.init_func() class TestLRN(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.local_response_normalization' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'LRN', 'bottom': ['x'], 'top': ['y'], 'lrn_param': { 'local_size': 4, 'alpha': 0.5, 'beta': 0.25, 'norm_region': 0, # ACROSS_CHANNELS 'k': 0.5 }, } ] } def test_lrn(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with( self.inputs[0], n=4, k=0.5, alpha=0.5 / 4, beta=0.25) class TestLRNWithinChannel(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'LRN', 'lrn_param': { 'norm_region': 1, # WITHIN_CHANNELS is not supported }, } ] } def test_lrn(self): with self.assertRaises(RuntimeError): self.init_func() class TestMaxPooling(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.max_pooling_2d' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Pooling', 'bottom': ['x'], 'top': ['y'], 'pooling_param': { 'pool': 0, # MAX 'kernel_h': 2, 'kernel_w': 3, 'stride_h': 4, 'stride_w': 5, 'pad_h': 6, 'pad_w': 7, } } ] } def test_max_pooling(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with( self.inputs[0], (2, 3), stride=(4, 5), pad=(6, 7)) class TestAveragePooling(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.average_pooling_2d' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Pooling', 'bottom': ['x'], 'top': ['y'], 'pooling_param': { 'pool': 1, # AVE 'kernel_size': 2, 'stride': 4, 'pad': 6, } } ] } def test_max_pooling(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with( self.inputs[0], 2, stride=4, pad=6) class TestGlobalPooling(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.max_pooling_2d' in_shapes = [(3, 2, 3, 4)] out_shapes = [(3, 2, 3, 4)] data = { 'layer': [ { 'name': 'l1', 'type': 'Pooling', 'bottom': ['x'], 'top': ['y'], 'pooling_param': { 'pool': 0, # MAX 'global_pooling': True, } } ] } def test_global_pooling(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with( self.inputs[0], (3, 4), stride=1, pad=0) class TestStochasticPooling(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'Pooling', 'pooling_param': { 'pool': 2, # STOCHASTIC is not supported } } ] } def test_stochastic_pooling(self): with self.assertRaises(RuntimeError): self.init_func() class TestReLU(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.relu' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'ReLU', 'bottom': ['x'], 'top': ['y'], 'relu_param': { 'negative_slope': 0 } } ] } def test_lrn(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestLeakyReLU(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.leaky_relu' in_shapes = [(3, 2, 3)] out_shapes = [(3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'ReLU', 'bottom': ['x'], 'top': ['y'], 'relu_param': { 'negative_slope': 0.5 } } ] } def test_lrn(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0], slope=0.5) class TestReshape(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.reshape' in_shapes = [(3, 2, 3)] out_shapes = [(3, 6)] data = { 'layer': [ { 'name': 'l1', 'type': 'Reshape', 'bottom': ['x'], 'top': ['y'], 'reshape_param': { 'shape': { 'dim': [3, 6] } } } ] } def test_reshape(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0], shape=[3, 6]) class TestBatchNorm(TestCaffeFunctionBaseMock): func_name = 'chainer.links.BatchNormalization.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'BatchNorm', 'bottom': ['x'], 'top': ['y'], 'blobs': [ # For average mean. { 'shape': { 'dim': [3], }, 'data': list(six.moves.range(3)), }, # For average variance. { 'shape': { 'dim': [3], }, 'data': list(six.moves.range(3)), }, ], 'batch_norm_param': { 'use_global_stats': False, } } ] } def test_batchnorm(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0], finetune=False) class TestBatchNormUsingGlobalStats(TestCaffeFunctionBaseMock): func_name = 'chainer.links.BatchNormalization.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'BatchNorm', 'bottom': ['x'], 'top': ['y'], 'blobs': [ # For average mean. { 'shape': { 'dim': [3], }, 'data': list(six.moves.range(3)), }, # For average variance. { 'shape': { 'dim': [3], }, 'data': list(six.moves.range(3)), }, ], 'batch_norm_param': { 'use_global_stats': True, } } ] } def test_batchnorm(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0], finetune=False) class TestEltwiseProd(TestCaffeFunctionBaseMock): func_name = 'chainer.variable.Variable.__mul__' in_shapes = [(2, 3), (2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Eltwise', 'bottom': ['x1', 'x2', 'x3'], 'top': ['y'], 'eltwise_param': { 'operation': 0, # PROD }, } ] } def test_eltwise_prod(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x1', 'x2', 'x3'], ['y']) self.mock.assert_has_calls([mock.call(self.inputs[1]), mock.call(self.inputs[2])]) class TestEltwiseSum(TestCaffeFunctionBaseMock): func_name = 'chainer.variable.Variable.__add__' in_shapes = [(2, 3), (2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Eltwise', 'bottom': ['x1', 'x2', 'x3'], 'top': ['y'], 'eltwise_param': { 'operation': 1, # SUM }, } ] } def test_eltwise_sum(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x1', 'x2', 'x3'], ['y']) self.mock.assert_has_calls([mock.call(self.inputs[1]), mock.call(self.inputs[2])]) class TestEltwiseSumCoeff(TestCaffeFunctionBaseMock): func_name = 'chainer.variable.Variable.__add__' in_shapes = [(2, 3), (2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Eltwise', 'bottom': ['x1', 'x2', 'x3'], 'top': ['y'], 'eltwise_param': { 'operation': 1, # SUM 'coeff': list(six.moves.range(3)), }, } ] } def test_eltwise_sum(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x1', 'x2', 'x3'], ['y']) self.assertEqual(self.mock.call_count, 2) class TestEltwiseSumInvalidCoeff(TestCaffeFunctionBaseMock): func_name = 'chainer.variable.Variable.__add__' in_shapes = [(2, 3), (2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Eltwise', 'bottom': ['x1', 'x2', 'x3'], 'top': ['y'], 'eltwise_param': { 'operation': 1, # SUM # not same as number of bottoms 'coeff': list(six.moves.range(2)), }, } ] } def test_eltwise_sum(self): self.init_func() with self.assertRaises(AssertionError): self.call(['x1', 'x2', 'x3'], ['y']) class TestEltwiseMax(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.maximum' in_shapes = [(2, 3), (2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Eltwise', 'bottom': ['x1', 'x2', 'x3'], 'top': ['y'], 'eltwise_param': { 'operation': 2, # MAX }, } ] } def test_eltwise_max(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x1', 'x2', 'x3'], ['y']) self.mock.assert_has_calls( [mock.call(self.inputs[0], self.inputs[1]), mock.call(self.outputs[0], self.inputs[2])]) class TestScale(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Scale.forward' in_shapes = [(2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Scale', 'bottom': ['x', 'y'], 'top': ['z'], 'scale_param': { 'axis': 0, } } ] } def test_scale(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x', 'y'], ['z']) self.mock.assert_called_once_with(self.inputs[0], self.inputs[1]) class TestScaleOneBottom(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Scale.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Scale', 'bottom': ['x'], 'top': ['y'], 'blobs': [ { 'shape': { 'dim': [2, 3], }, 'data': list(six.moves.range(6)), } ], 'scale_param': { 'axis': 0, } } ] } def test_scale_one_bottom(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestScaleWithBias(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Scale.forward' in_shapes = [(2, 3), (2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Scale', 'bottom': ['x', 'y'], 'top': ['z'], 'blobs': [ { 'shape': { 'dim': [2, 3], }, 'data': list(six.moves.range(6)), } ], 'scale_param': { 'axis': 0, 'bias_term': True, } } ] } def test_scale_with_bias(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.assertTrue(hasattr(self.func.l1, 'bias')) self.call(['x', 'y'], ['z']) self.mock.assert_called_once_with(self.inputs[0], self.inputs[1]) class TestScaleOneBottomWithBias(TestCaffeFunctionBaseMock): func_name = 'chainer.links.Scale.forward' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Scale', 'bottom': ['x'], 'top': ['y'], 'blobs': [ # For W parameter. { 'shape': { 'dim': [2, 3], }, 'data': list(six.moves.range(6)), }, # For bias. { 'shape': { 'dim': [2, 3], }, 'data': list(six.moves.range(6)), } ], 'scale_param': { 'axis': 0, 'bias_term': True, } } ] } def test_scale_one_bottom_with_bias(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.assertTrue(hasattr(self.func.l1, 'bias')) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSlice(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.split_axis' in_shapes = [(3, 4, 3)] out_shapes = [(3, 2, 3), (3, 2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Slice', 'bottom': ['x'], 'top': ['y1', 'y2'], 'slice_param': { 'axis': 1 } } ] } def test_slice(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y1', 'y2']) self.mock.assert_called_once_with( self.inputs[0], indices_or_sections=2, axis=1 ) class TestSliceNoAxis(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.split_axis' in_shapes = [(4, 6, 4)] out_shapes = [(2, 6, 4), (2, 6, 4)] data = { 'layer': [ { 'name': 'l1', 'type': 'Slice', 'bottom': ['x'], 'top': ['y1', 'y2'], 'slice_param': { 'slice_dim': 0 } } ] } def test_slice_no_axis(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y1', 'y2']) self.mock.assert_called_once_with( self.inputs[0], indices_or_sections=2, axis=0 ) class TestSliceNoAxisNoSliceDim(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.split_axis' in_shapes = [(4, 6, 4)] out_shapes = [(4, 3, 4), (4, 3, 4)] data = { 'layer': [ { 'name': 'l1', 'type': 'Slice', 'bottom': ['x'], 'top': ['y1', 'y2'], } ] } def test_slice_no_axis_no_slice_dim(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y1', 'y2']) self.mock.assert_called_once_with( self.inputs[0], indices_or_sections=2, axis=1 ) class TestSliceSlicePoint(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.split_axis' in_shapes = [(4, 8, 6)] out_shapes = [(4, 3, 6), (4, 2, 6), (4, 3, 6)] data = { 'layer': [ { 'name': 'l1', 'type': 'Slice', 'bottom': ['x'], 'top': ['y1', 'y2', 'y3'], 'slice_param': { 'axis': 1, 'slice_point': [3, 5] } } ] } def test_slice_slice_point(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y1', 'y2', 'y3']) self.mock.assert_called_once_with( self.inputs[0], indices_or_sections=[3, 5], axis=1 ) class TestSigmoid(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.sigmoid' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Sigmoid', 'bottom': ['x'], 'top': ['y'], } ] } def test_sigmoid(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSoftmax(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.softmax' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Softmax', 'bottom': ['x'], 'top': ['y'], } ] } def test_softmax(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSoftmaxCaffeEngine(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.softmax' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Softmax', 'softmax_param': { 'engine': 1, # CAFFE }, 'bottom': ['x'], 'top': ['y'], } ] } def test_softmax_caffe_engine(self): # TODO(beam2d): Check if the mock is called with # chainer.config.use_cudnn == False self.init_func() self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSoftmaxcuDnnEngine(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.softmax' in_shapes = [(2, 3)] out_shapes = [(2, 3)] data = { 'layer': [ { 'name': 'l1', 'type': 'Softmax', 'softmax_param': { 'engine': 2, # CUDNN }, 'bottom': ['x'], 'top': ['y'], } ] } def test_softmax_cuDNN_engine(self): # TODO(beam2d): Check if the mock is called with # chainer.config.use_cudnn == True self.init_func() self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSoftmaxInvalidAxis(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'Softmax', 'softmax_param': { 'axis': 0, # invalid axis } } ] } def test_softmax_invalid_axis(self): with self.assertRaises(RuntimeError): self.init_func() class TestSoftmaxWithLoss(TestCaffeFunctionBaseMock): func_name = 'chainer.functions.softmax_cross_entropy' in_shapes = [(3, 2, 3)] out_shapes = [()] data = { 'layer': [ { 'name': 'l1', 'type': 'SoftmaxWithLoss', 'bottom': ['x'], 'top': ['y'], } ] } def test_softmax_with_loss(self): self.init_func() self.assertEqual(len(self.func.layers), 1) self.call(['x'], ['y']) self.mock.assert_called_once_with(self.inputs[0]) class TestSoftmaxWithLossInvalidAxis(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'SoftmaxWithLoss', 'softmax_param': { 'axis': 0, # invalid axis } } ] } def test_softmax_with_loss_invalid_axis(self): with self.assertRaises(RuntimeError): self.init_func() class TestSplit(TestCaffeFunctionBase): data = { 'layer': [ { 'name': 'l1', 'type': 'Split', 'bottom': ['x'], 'top': ['y', 'z'], } ] } def test_split(self): self.init_func() self.assertEqual(self.func.split_map, {'y': 'x', 'z': 'x'}) testing.run_module(__name__, __file__)
35,121
25.151899
79
py
chainer
chainer-master/tests/onnx_chainer_tests/conftest.py
import chainer import onnx import pytest import onnx_chainer def pytest_addoption(parser): parser.addoption( '--value-check-runtime', dest='value-check-runtime', default='onnxruntime', choices=['skip', 'onnxruntime', 'mxnet'], help='select test runtime') parser.addoption( '--opset-versions', dest='opset-versions', default=None, help='select opset versions, select from "min", "latest", ' 'or a list of numbers like "9,10"') @pytest.fixture(scope='function') def disable_experimental_warning(): org_config = chainer.disable_experimental_feature_warning chainer.disable_experimental_feature_warning = True try: yield finally: chainer.disable_experimental_feature_warning = org_config @pytest.fixture(scope='function') def check_model_expect(request): selected_runtime = request.config.getoption('value-check-runtime') if selected_runtime == 'onnxruntime': from onnx_chainer.testing.test_onnxruntime import check_model_expect # NOQA _checker = check_model_expect elif selected_runtime == 'mxnet': from onnx_chainer.testing.test_mxnet import check_model_expect _checker = check_model_expect else: def empty_func(*args, **kwargs): pass _checker = empty_func return _checker @pytest.fixture(scope='function') def target_opsets(request): opsets = request.config.getoption('opset-versions') min_version = onnx_chainer.MINIMUM_OPSET_VERSION max_version = min( onnx.defs.onnx_opset_version(), onnx_chainer.MAXIMUM_OPSET_VERSION) if opsets is None: return list(range(min_version, max_version + 1)) elif opsets == 'min': return [min_version] elif opsets == 'latest': return [max_version] else: try: versions = [int(i) for i in opsets.split(',')] except ValueError: raise ValueError('cannot convert {} to versions list'.format( opsets)) return versions
2,049
31.03125
84
py
chainer
chainer-master/docs/source/conf.py
# -*- coding: utf-8 -*- # # Chainer documentation build configuration file, created by # sphinx-quickstart on Sun May 10 12:22:10 2015. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import inspect import os import pkg_resources import sys sys.path.insert(0, os.path.abspath(os.path.dirname(__file__))) import _docstring_check import _autosummary_check __version__ = pkg_resources.get_distribution('chainer').version on_rtd = os.environ.get('READTHEDOCS', None) == 'True' rtd_version = os.environ.get('READTHEDOCS_VERSION') if rtd_version == 'latest': tag = 'master' else: tag = 'v{}'.format(__version__) extlinks = { 'blob': ('https://github.com/chainer/chainer/blob/{}/%s'.format(tag), ''), 'tree': ('https://github.com/chainer/chainer/tree/{}/%s'.format(tag), ''), } # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.autosummary', 'sphinx.ext.doctest', 'sphinx.ext.extlinks', 'sphinx.ext.intersphinx', 'sphinx.ext.mathjax', 'sphinx.ext.napoleon', 'sphinx.ext.linkcode', '_napoleon_patch', ] try: import sphinxcontrib.spelling # noqa extensions.append('sphinxcontrib.spelling') except ImportError: pass # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # source_suffix = ['.rst', '.md'] source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'Chainer' copyright = u'2015, Preferred Networks, inc. and Preferred Infrastructure, inc.' author = u'Preferred Networks, inc. and Preferred Infrastructure, inc.' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = __version__ # The full version, including alpha/beta/rc tags. release = __version__ # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = [] # The reST default role (used for this markup: `text`) to use for all # documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. #keep_warnings = False # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = False # Napoleon settings napoleon_use_ivar = True napoleon_include_special_with_doc = True # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. if not on_rtd: html_theme = 'sphinx_rtd_theme' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. #html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] html_style = 'css/modified_theme.css' if on_rtd: html_context = { 'css_files': [ 'https://media.readthedocs.org/css/sphinx_rtd_theme.css', 'https://media.readthedocs.org/css/readthedocs-doc-embed.css', '_static/css/modified_theme.css', ], } # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. #html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. html_show_sourcelink = False # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Language to be used for generating the HTML full-text search index. # Sphinx supports the following languages: # 'da', 'de', 'en', 'es', 'fi', 'fr', 'hu', 'it', 'ja' # 'nl', 'no', 'pt', 'ro', 'ru', 'sv', 'tr' #html_search_language = 'en' # A dictionary with options for the search language support, empty by default. # Now only 'ja' uses this config value #html_search_options = {'type': 'default'} # The name of a javascript file (relative to the configuration directory) that # implements a search results scorer. If empty, the default will be used. #html_search_scorer = 'scorer.js' # Output file base name for HTML help builder. htmlhelp_basename = 'Chainerdoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', # Latex figure (float) alignment #'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, 'Chainer.tex', u'Chainer Documentation', u'Preferred Networks, inc. and Preferred Infrastructure, inc.', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ (master_doc, 'chainer', u'Chainer Documentation', [author], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ (master_doc, 'Chainer', u'Chainer Documentation', author, 'Chainer', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. #texinfo_no_detailmenu = False autosummary_generate = True intersphinx_mapping = { 'python': ('https://docs.python.org/3/', None), 'numpy': ('https://docs.scipy.org/doc/numpy/', None), 'cupy': ('https://docs-cupy.chainer.org/en/latest/', None), 'chainercv': ('https://chainercv.readthedocs.io/en/latest/', None), } doctest_global_setup = ''' import os import numpy as np import chainer from chainer.backends import cuda from chainer.backends.cuda import cupy from chainer import Function, gradient_check, training, utils, Variable from chainer import datasets, iterators, optimizers, serializers from chainer import Link, Chain, ChainList import chainer.functions as F import chainer.links as L from chainer.testing import doctest_helper from chainer.training import extensions import chainerx import onnx_chainer np.random.seed(0) ''' spelling_lang = 'en_US' spelling_word_list_filename = 'spelling_wordlist.txt' def setup(app): app.connect('autodoc-process-docstring', _autodoc_process_docstring) app.connect('build-finished', _build_finished) def _autodoc_process_docstring(app, what, name, obj, options, lines): _docstring_check.check(app, what, name, obj, options, lines) def _build_finished(app, exception): if exception is None: _autosummary_check.check(app, exception) def _import_object_from_name(module_name, fullname): obj = sys.modules.get(module_name) if obj is None: return None for comp in fullname.split('.'): obj = getattr(obj, comp) return obj def _is_egg_directory(path): return (path.endswith('.egg') and os.path.isdir(os.path.join(path, 'EGG-INFO'))) def _is_git_root(path): return os.path.isdir(os.path.join(path, '.git')) _source_root = None def _find_source_root(source_abs_path): # Note that READTHEDOCS* environment variable cannot be used, because they # are not set under docker environment. global _source_root if _source_root is None: dir = os.path.dirname(source_abs_path) while True: if _is_egg_directory(dir) or _is_git_root(dir): # Reached the root directory _source_root = dir break dir_ = os.path.dirname(dir) if len(dir_) == len(dir): raise RuntimeError('Couldn\'t parse root directory from ' 'source file: {}'.format(source_abs_path)) dir = dir_ return _source_root def _get_source_relative_path(source_abs_path): return os.path.relpath(source_abs_path, _find_source_root(source_abs_path)) def _get_sourcefile_and_linenumber(obj): # Retrieve the original function wrapped by contextlib.contextmanager if callable(obj): closure = getattr(obj, '__closure__', None) if closure is not None: obj = closure[0].cell_contents # Get the source file name and line number at which obj is defined. try: filename = inspect.getsourcefile(obj) except TypeError: # obj is not a module, class, function, ..etc. return None, None # inspect can return None for cython objects if filename is None: return None, None # Get the source line number _, linenum = inspect.getsourcelines(obj) return filename, linenum def linkcode_resolve(domain, info): if domain != 'py' or not info['module']: return None if 1 == int(os.environ.get('CHAINER_DOCS_SKIP_LINKCODE', 0)): return None # Import the object from module path obj = _import_object_from_name(info['module'], info['fullname']) # If it's not defined in the internal module, return None. mod = inspect.getmodule(obj) if mod is None: return None if not (mod.__name__ == 'chainer' or mod.__name__.startswith('chainer.')): return None # Retrieve source file name and line number filename, linenum = _get_sourcefile_and_linenumber(obj) if filename is None or linenum is None: return None filename = os.path.realpath(filename) relpath = _get_source_relative_path(filename) return 'https://github.com/chainer/chainer/blob/{}/{}#L{}'.format( tag, relpath, linenum)
14,917
30.209205
80
py
chainer
chainer-master/chainer/initializer.py
import typing as tp # NOQA from chainer import types # NOQA from chainer import utils class Initializer(object): """Initializes array. It initializes the given array. Attributes: dtype: Data type specifier. It is for type check in ``__call__`` function. """ def __init__(self, dtype: tp.Optional[types.DTypeSpec] = None) -> None: self.dtype = dtype # type: types.DTypeSpec def __call__(self, array: types.NdArray) -> None: """Initializes given array. This method destructively changes the value of array. The derived class is required to implement this method. The algorithms used to make the new values depend on the concrete derived classes. Args: array (:ref:`ndarray`): An array to be initialized by this initializer. """ raise NotImplementedError() # Original code forked from MIT licensed keras project # https://github.com/fchollet/keras/blob/master/keras/initializations.py def get_fans(shape): if not isinstance(shape, tuple): raise ValueError( 'shape must be tuple. Actual type: {}'.format(type(shape))) if len(shape) < 2: raise ValueError( 'shape must be of length >= 2. Actual shape: {}'.format(shape)) receptive_field_size = utils.size_of_shape(shape[2:]) fan_in = shape[1] * receptive_field_size fan_out = shape[0] * receptive_field_size return fan_in, fan_out
1,502
27.358491
75
py
chainer
chainer-master/chainer/functions/array/as_strided.py
import numpy as np import six from chainer import backend from chainer.backends import cuda from chainer import function_node from chainer.utils import type_check index_dtype = {t().itemsize: t for t in np.sctypes['int']} def _byte2step(iterable, itemsize): for i in iterable: assert i % itemsize == 0 return tuple([i // itemsize for i in iterable]) def _step2byte(iterable, itemsize): return tuple([i * itemsize for i in iterable]) def _maybe_overlapping_memory(shape, strides): """Returns bool value indicating the array with such shape and strides might have overlapping memory. Args: shape (tuple of int): The shape of output. strides (tuple of int): The strides of output, given in the unit of steps. storage_offset (int): The offset between the head of allocated memory and the pointer of first element, given in the unit of steps. Returns: bool: Existence of the overlapping memory """ max_ptr_in_slice = 0 for stride, size in sorted(zip([abs(s) for s in strides], shape)): if stride <= max_ptr_in_slice: return True max_ptr_in_slice += stride * (size - 1) return False def _min_index(shape, strides, storage_offset): """Returns the leftest index in the array (in the unit-steps) Args: shape (tuple of int): The shape of output. strides (tuple of int): The strides of output, given in the unit of steps. storage_offset (int): The offset between the head of allocated memory and the pointer of first element, given in the unit of steps. Returns: int: The leftest pointer in the array """ sh_st_neg = [sh_st for sh_st in zip(shape, strides) if sh_st[1] < 0] if not sh_st_neg: return storage_offset else: return storage_offset + six.moves.reduce( lambda base, sh_st: base + (sh_st[0] - 1) * sh_st[1], sh_st_neg, 0) def _max_index(shape, strides, storage_offset): """Returns the rightest index in the array Args: shape (tuple of int): The shape of output. strides (tuple of int): The strides of output, given in unit-steps. storage_offset (int): The offset between the head of allocated memory and the pointer of first element, given in the unit of steps. Returns: int: The rightest pointer in the array """ sh_st_pos = [sh_st for sh_st in zip(shape, strides) if sh_st[1] > 0] if not sh_st_pos: return storage_offset else: return storage_offset + six.moves.reduce( lambda base, sh_st: base + (sh_st[0] - 1) * sh_st[1], sh_st_pos, 0) def _index_add(augend, indices, addend): """Wrapper of :func:`cupyx.scatter_add` and :func:`numpy.add.at` Args: augend (:class:`numpy.ndarray` or :class:`cupy.ndarray`): The array modified in-place. indices (:class:`numpy.ndarray` or :class:`cupy.ndarray`): The indices of ``augend``. The shape is the same to the ``addend``. addend (:class:`numpy.ndarray` or :class:`cupy.ndarray`): The array to be added. Returns: None """ if isinstance(augend, cuda.ndarray): cuda.cupyx.scatter_add(augend, indices, addend) elif isinstance(augend, np.ndarray): np.add.at(augend, indices, addend) def _get_base_array(array): """Get the founder of :class:`numpy.ndarray`. Args: array (:class:`numpy.ndarray`): The view of the base array. Returns: :class:`numpy.ndarray`: The base array. """ base_array_candidate = array while base_array_candidate.base is not None: base_array_candidate = base_array_candidate.base return base_array_candidate def _stride_array(array, shape, strides, storage_offset): """Wrapper of :func:`numpy.lib.stride_tricks.as_strided`. .. note: ``strides`` and ``storage_offset`` is given in the unit of steps instead the unit of bytes. This specification differs from that of :func:`numpy.lib.stride_tricks.as_strided`. Args: array (:class:`numpy.ndarray` of :class:`cupy.ndarray`): The base array for the returned view. shape (tuple of int): The shape of the returned view. strides (tuple of int): The strides of the returned view, given in the unit of steps. storage_offset (int): The offset from the leftest pointer of allocated memory to the first element of returned view, given in the unit of steps. Returns: :class:`numpy.ndarray` or :class:`cupy.ndarray`: The new view for the base array. """ min_index = _min_index(shape, strides, storage_offset) max_index = _max_index(shape, strides, storage_offset) strides = _step2byte(strides, array.itemsize) storage_offset, = _step2byte((storage_offset,), array.itemsize) if min_index < 0: raise ValueError('Out of buffer: too small index was specified') if isinstance(array, cuda.ndarray): pooled_memory = array.data.mem if (max_index + 1) * array.itemsize > pooled_memory.size: raise ValueError('Out of buffer: too large index was specified') memptr = cuda.cupy.cuda.memory.MemoryPointer(pooled_memory, storage_offset) return cuda.cupy.ndarray(shape, array.dtype, memptr, strides) elif isinstance(array, np.ndarray): base_array = _get_base_array(array) if (max_index + 1) * base_array.itemsize > base_array.nbytes: raise ValueError('Out of buffer: too large index was specified') return np.ndarray(shape, base_array.dtype, base_array.data, storage_offset, strides) else: raise TypeError('Only (np|cp).ndarray is accepted') class TensorGeometry(object): def __init__(self, array): self.shape = array.shape self.strides = _byte2step(array.strides, array.itemsize) if isinstance(array, np.ndarray): base_array = _get_base_array(array) array_ptr = array.__array_interface__['data'][0] base_array_ptr = base_array.__array_interface__['data'][0] offset_bytes = array_ptr - base_array_ptr elif isinstance(array, cuda.ndarray): offset_bytes = array.data.ptr - array.data.mem.ptr else: raise ValueError('only (np|cp).ndarray is supported') self.storage_offset, = _byte2step((offset_bytes,), array.itemsize) self.itemsize = array.itemsize @property def ndim(self): return len(self.shape) class AsStrided(function_node.FunctionNode): """Transportation of :func:`torch.Tensor.as_strided`. While :func:`torch.Tensor.as_strided` does not support nagative strides, this implementation does support it. """ def __init__(self, shape, strides, storage_offset=None): self.shape = shape self.strides = strides self.storage_offset = storage_offset self.input_geometry = None def check_type_forward(self, in_types): type_check.expect(in_types.size() == 1) def forward(self, inputs): assert len(inputs) > 0 x = inputs[0] self.input_geometry = TensorGeometry(x) if self.storage_offset is None: self.storage_offset = self.input_geometry.storage_offset return _stride_array(x, self.shape, self.strides, self.storage_offset), def backward(self, _, grad_outputs): """Backward computation which calls :class:`AsStridedGrad`. .. note: While this implementation is based on *New-Style Function Implementation*, the backward computation does not support double-backpropagation due to *layout agnostic* algorithm ( originally named in the note of pytorch). """ return AsStridedGrad(self.input_geometry, self.shape, self.strides, self.storage_offset).apply(grad_outputs) class AsStridedGrad(function_node.FunctionNode): """Backward of :func:`~chainer.functions.as_strided`. """ def __init__(self, input_geometry, shape, strides, storage_offset): self.input_geometry = input_geometry self.shape = shape self.strides = strides self.storage_offset = storage_offset def forward(self, grads): assert len(grads) > 0 gy = grads[0] if gy.dtype not in np.sctypes['float']: raise TypeError('Only float is supported for back propagation') xp = backend.get_array_module(gy) input_geometry = self.input_geometry itemsize = input_geometry.itemsize if 0 in input_geometry.shape: return xp.zeros(input_geometry.shape) # 1. remove redundant axis from input/output # [redundant axis] # axis with shape==0, shape==1 or strides==0 if 0 in gy.shape: return backend.get_array_module(gy).zeros(input_geometry.shape) else: out_shape = tuple([ self.shape[i] for i in six.moves.range(gy.ndim) if self.shape[i] != 1 and self.strides[i] != 0]) out_strides = tuple([ self.strides[i] for i in six.moves.range(gy.ndim) if self.shape[i] != 1 and self.strides[i] != 0]) gy = gy.sum( tuple([i for i in six.moves.range(gy.ndim) if self.strides[i] == 0])) gy = gy.squeeze() out_storage_offset = self.storage_offset inp_shape = tuple([input_geometry.shape[i] for i in six.moves.range(input_geometry.ndim) if input_geometry.shape[i] != 1]) inp_strides = tuple([input_geometry.strides[i] for i in six.moves.range(input_geometry.ndim) if input_geometry.shape[i] != 1]) inp_storage_offset = input_geometry.storage_offset # 2. calculate minimum required storage for gradient computation inp_min_ptr = _min_index(inp_shape, inp_strides, input_geometry.storage_offset) out_min_ptr = _min_index(out_shape, out_strides, self.storage_offset) common_min_ptr = min(inp_min_ptr, out_min_ptr) inp_max_ptr = _max_index(inp_shape, inp_strides, input_geometry.storage_offset) out_max_ptr = _max_index(out_shape, out_strides, self.storage_offset) common_max_ptr = max(inp_max_ptr, out_max_ptr) base_size = (common_max_ptr - common_min_ptr) + 1 storage = xp.zeros(base_size, dtype=gy.dtype) flatten_full_indices = xp.arange(base_size, dtype=index_dtype[itemsize]) out_maybe_overlap = _maybe_overlapping_memory(out_shape, out_strides) if out_maybe_overlap: out_indices = _stride_array(flatten_full_indices, out_shape, out_strides, out_storage_offset - common_min_ptr) _index_add(storage, out_indices, gy) else: storage_view = _stride_array(storage, out_shape, out_strides, out_storage_offset - common_min_ptr) storage_view[:] = gy[:] inp_maybe_overlap = _maybe_overlapping_memory(inp_shape, inp_strides) if inp_maybe_overlap: count = xp.zeros_like(storage) inp_indices = _stride_array(flatten_full_indices, inp_shape, inp_strides, inp_storage_offset - common_min_ptr) _index_add(count, inp_indices, xp.ones(1)) with np.errstate(divide='ignore', invalid='ignore'): storage /= count return _stride_array(storage, inp_shape, inp_strides, inp_storage_offset - common_min_ptr), def backward(self, target_input_indexes, grad_outputs): raise NotImplementedError def as_strided(x, shape, strides, storage_offset=None): """Create a new view of array with the given shape, strides, and offset. Args: x (tuple of :class:`~chainer.Variable` or :class:`numpy.ndarray` or \ :class:`cupy.ndarray`): The array pointing a memory buffer. Its view is totally ignored. shape (tuple of int): The shape of output. strides (tuple of int): The strides of output, given in the unit of steps. storage_offset (int): The offset between the head of allocated memory and the pointer of first element, given in the unit of steps. Returns: ~chainer.Variable: The strided variable. .. warning:: Users should be aware that this function potentially causes unintended side effects. See `numpy.lib.stride_tricks.as_strided`_ for the detail. .. note:: The backward algorithm is borrowed from `torch.Tensor.as_strided`. Therefore, the returned gradient of ``backward`` is *layout-agnostic* when ``x`` contains memory overlap. See notes in pytorch's source code (as_strided Backward and layout-aware/agnostic autograd) too. .. note:: In this function ``strides`` and ``storage_offset`` are given in the unit of steps instead of bytes. This specification differs from :func:`numpy.lib.stride_tricks.as_strided`. .. admonition:: Example >>> from chainer import functions as F, Variable >>> x = Variable(np.arange(4, dtype=np.float32)) >>> x variable([0., 1., 2., 3.]) >>> y = F.as_strided(x, (3, 2), (1, 1), 0) >>> y variable([[0., 1.], [1., 2.], [2., 3.]]) >>> y.grad = np.ones((3, 2), dtype=np.float32) >>> y.backward() >>> x.grad array([1., 2., 2., 1.], dtype=float32) .. _numpy.lib.stride_tricks.as_strided: https://docs.scipy.org/doc/numpy/reference/generated/\ numpy.lib.stride_tricks.as_strided.html """ return AsStrided(shape, strides, storage_offset).apply((x,))[0]
14,427
35.994872
79
py
chainer
chainer-master/chainer/links/caffe/caffe_function.py
import warnings import numpy import six from chainer import configuration from chainer import functions from chainer import initializer from chainer import link from chainer.links.caffe.protobuf3 import caffe_pb2 as caffe_pb from chainer.links.connection import convolution_2d from chainer.links.connection import deconvolution_2d from chainer.links.connection import linear from chainer.links.connection import scale from chainer.links.normalization import batch_normalization from chainer.utils import argument try: # This method is undocumented, but is required to read large size of # model files when a user uses cpp-implementation. from google.protobuf.pyext import _message _message.SetAllowOversizeProtos(True) except ImportError: pass _type_to_method = {} _oldname_to_method = {} def _layer(typ, oldname): def decorator(meth): global _type_to_method _type_to_method[typ] = meth if oldname is not None: typevalue = getattr(caffe_pb.V1LayerParameter, oldname) _oldname_to_method[typevalue] = meth return meth return decorator class _Blob(initializer.Initializer): chunk_size = 1024 * 1024 def __init__(self, blob): super(_Blob, self).__init__() self.data = blob.data def __call__(self, array): array = array.ravel() size = len(array) indices = list(range(0, size, self.chunk_size)) # Rather than accessing Protobuf's RepeatedScalar fields directly, # creating a intermediate list by indexing is more efficient due to # the implementation of the Python extension of Protobuf. # To avoid allocating excessively large lists, we limit the length # of lists by `chunk_size`. for start, end in zip(indices, indices[1:] + [size]): array[start:end] = self.data[start:end] class _ConvolutionBlob(_Blob): def __init__(self, blob, group): super(_ConvolutionBlob, self).__init__(blob) self.group = group def __call__(self, array): n_out, n_in = array.shape[:2] part_out = n_out // self.group part_in = n_in // self.group array[...] = 0 part_size = len(self.data) // self.group for i in six.moves.range(self.group): out_slice = slice(i * part_out, (i + 1) * part_out) in_slice = slice(i * part_in, (i + 1) * part_in) w = array[out_slice, in_slice] data = numpy.array(self.data[i * part_size:(i + 1) * part_size]) w[:] = data.reshape(w.shape) class CaffeFunction(link.Chain): """Caffe emulator based on the model file of Caffe. Given a protocol buffers file of a Caffe model, this class loads and emulates it on :class:`~chainer.Variable` objects. It supports the official reference models provided by BVLC. .. note:: CaffeFunction ignores the following layers: - Layers that CaffeFunction does not support (including data layers) - Layers that have no top blobs - Layers whose bottom blobs are incomplete (i.e., some or all of them are not given nor computed) .. warning:: It does not support full compatibility against Caffe. Some layers and configurations are not implemented in Chainer yet, though the reference models provided by the BVLC team are supported except data layers. .. admonition:: Example Consider we want to extract the (unnormalized) log class probability of given images using BVLC reference CaffeNet. The model can be downloaded from: http://dl.caffe.berkeleyvision.org/bvlc_reference_caffenet.caffemodel We want to compute the ``fc8`` blob from the ``data`` blob. It is simply written as follows:: # Load the model func = CaffeFunction('path/to/bvlc_reference_caffenet.caffemodel') # Minibatch of size 10 x_data = numpy.ndarray((10, 3, 227, 227), dtype=numpy.float32) ... # (Fill the minibatch here) # Forward the pre-trained net x = Variable(x_data) y, = func(inputs={'data': x}, outputs=['fc8']) The result ``y`` contains the Variable corresponding to the ``fc8`` blob. The computational graph is memorized as a usual forward computation in Chainer, so we can run backprop through this pre-trained net. Args: model_path (str): Path to the binary-proto model file of Caffe. Attributes: forwards (dict): A mapping from layer names to corresponding functions. """ def __init__(self, model_path): super(CaffeFunction, self).__init__() net = caffe_pb.NetParameter() with open(model_path, 'rb') as model_file: net.MergeFromString(model_file.read()) self.forwards = {} self.split_map = {} self.layers = [] if net.layer: for layer in net.layer: meth = _type_to_method.get(layer.type) if meth: meth(self, layer) else: warnings.warn( 'Skip the layer "%s", since CaffeFunction does not ' 'support %s layer' % (layer.name, layer.type)) else: # v1 format for layer in net.layers: meth = _oldname_to_method.get(layer.type) if meth: meth(self, layer) else: warnings.warn( 'Skip the layer "%s", since CaffeFunction does not ' 'support it' % layer.name) def forward(self, inputs, outputs, disable=(), **kwargs): """forward(self, inputs, outputs, disable=()) Executes a sub-network of the network. This function acts as an interpreter of the network definition for Caffe. On execution, it interprets each layer one by one, and if the bottom blobs are already computed, then emulates the layer and stores output blobs as :class:`~chainer.Variable` objects. Args: inputs (dict): A dictionary whose key-value pairs indicate initial correspondences between blob names and :class:`~chainer.Variable` objects. outputs (Iterable): A list of blob names whose corresponding :class:`~chainer.Variable` objects are returned. disable (Iterable): A list of layer names that will be ignored during the forward computation. Returns: tuple: A tuple of output :class:`~chainer.Variable` objects corresponding to elements of the `outputs` argument. """ if kwargs: argument.check_unexpected_kwargs( kwargs, train='train argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) variables = dict(inputs) disable = set(disable) for func_name, bottom, top in self.layers: if (func_name in disable or func_name not in self.forwards or any(blob not in variables for blob in bottom)): continue func = self.forwards[func_name] input_vars = tuple(variables[blob] for blob in bottom) output_vars = func(*input_vars) if not isinstance(output_vars, (tuple, list)): output_vars = output_vars, for var, name in zip(output_vars, top): variables[name] = var self.variables = variables return tuple(variables[blob] for blob in outputs) def _add_layer(self, layer): bottom = [] for blob_name in layer.bottom: bottom.append(self.split_map.get(blob_name, blob_name)) self.layers.append((layer.name, bottom, list(layer.top))) @_layer('Concat', 'CONCAT') def _setup_concat(self, layer): param = layer.concat_param axis = param.axis if axis == 1 and param.concat_dim != 1: axis = param.concat_dim self.forwards[layer.name] = _ListArgumentFcuntion( functions.concat, axis=axis) self._add_layer(layer) @_layer('Convolution', 'CONVOLUTION') def _setup_convolution(self, layer): blobs = layer.blobs param = layer.convolution_param ksize = _get_ksize(param) stride = _get_stride(param) pad = _get_pad(param) num = _get_num(blobs[0]) channels = _get_channels(blobs[0]) bias_term = param.bias_term n_in = channels * param.group n_out = num func = convolution_2d.Convolution2D( n_in, n_out, ksize, stride, pad, nobias=not bias_term, initialW=_ConvolutionBlob(blobs[0], param.group), initial_bias=_Blob(blobs[1]) if bias_term else None) with self.init_scope(): setattr(self, layer.name, func) self.forwards[layer.name] = _CallChildLink(self, layer.name) self._add_layer(layer) @_layer('Deconvolution', 'DECONVOLUTION') def _setup_deconvolution(self, layer): blobs = layer.blobs param = layer.convolution_param ksize = _get_ksize(param) stride = _get_stride(param) pad = _get_pad(param) num = _get_num(blobs[0]) channels = _get_channels(blobs[0]) bias_term = param.bias_term n_in = num n_out = channels * param.group func = deconvolution_2d.Deconvolution2D( n_in, n_out, ksize, stride, pad, nobias=not bias_term, initialW=_ConvolutionBlob(blobs[0], param.group), initial_bias=_Blob(blobs[1]) if bias_term else None) with self.init_scope(): setattr(self, layer.name, func) self.forwards[layer.name] = _CallChildLink(self, layer.name) self._add_layer(layer) @_layer('Data', 'DATA') def _setup_data(self, layer): # We silently skip the data layer. pass @_layer('Dropout', 'DROPOUT') def _setup_dropout(self, layer): param = layer.dropout_param self.forwards[layer.name] = _SingleArgumentFunction( functions.dropout, ratio=param.dropout_ratio) self._add_layer(layer) @_layer('InnerProduct', 'INNER_PRODUCT') def _setup_inner_product(self, layer): param = layer.inner_product_param bias_term = param.bias_term if param.axis != 1: raise RuntimeError( 'Non-default axis in InnerProduct is not supported') blobs = layer.blobs width, height = _get_width(blobs[0]), _get_height(blobs[0]) func = linear.Linear( width, height, nobias=not bias_term, initialW=_Blob(blobs[0]), initial_bias=_Blob(blobs[1]) if bias_term else None) with self.init_scope(): setattr(self, layer.name, func) self.forwards[layer.name] = _CallChildLink(self, layer.name) self._add_layer(layer) @_layer('LRN', 'LRN') def _setup_lrn(self, layer): param = layer.lrn_param if param.norm_region != param.ACROSS_CHANNELS: raise RuntimeError('Within-channel LRN is not supported') fwd = _SingleArgumentFunction( functions.local_response_normalization, n=param.local_size, k=param.k, alpha=param.alpha / param.local_size, beta=param.beta) self.forwards[layer.name] = fwd self._add_layer(layer) @_layer('Pooling', 'POOLING') def _setup_pooling(self, layer): param = layer.pooling_param ksize = _get_ksize(param) stride = _get_stride(param) pad = _get_pad(param) if param.pool == param.MAX: func = functions.max_pooling_2d elif param.pool == param.AVE: func = functions.average_pooling_2d else: raise RuntimeError('Stochastic pooling is not supported') if param.global_pooling and not ksize: # if global_pooling is set but no kernel size, the kernel size # is computed dynamically to cover the whole input feature map def _func(x, stride, pad): return func(x, x.shape[2:], stride=stride, pad=pad) fw = _SingleArgumentFunction(_func, stride=stride, pad=pad) else: fw = _SingleArgumentFunction(func, ksize, stride=stride, pad=pad) self.forwards[layer.name] = fw self._add_layer(layer) @_layer('ReLU', 'RELU') def _setup_relu(self, layer): slope = layer.relu_param.negative_slope if slope != 0: fw = _SingleArgumentFunction(functions.leaky_relu, slope=slope) else: fw = functions.relu self.forwards[layer.name] = fw self._add_layer(layer) @_layer('Reshape', None) def _setup_reshape(self, layer): shape = layer.reshape_param.shape.dim fw = _SingleArgumentFunction(functions.reshape, shape=shape) self.forwards[layer.name] = fw self._add_layer(layer) @_layer('BatchNorm', None) def _setup_batchnorm(self, layer): # Get layer parameters. blobs = layer.blobs param = layer.batch_norm_param use_global_stats = param.use_global_stats decay = param.moving_average_fraction eps = param.eps size = int(blobs[0].shape.dim[0]) # Get channel dim from mean blob. # Make BatchNormalization link. func = batch_normalization.BatchNormalization( size, decay=decay, eps=eps, use_gamma=False, use_beta=False) _Blob(blobs[0])(func.avg_mean) _Blob(blobs[1])(func.avg_var) # Scale the means and variances if a scaling factor is appended to the # blobs to correctly mimic to the behavior of Caffe. See # https://github.com/BVLC/caffe/issues/4885 if len(blobs) >= 3: scaling_factor = blobs[2].data func.avg_mean /= scaling_factor[0] func.avg_var /= scaling_factor[0] with self.init_scope(): setattr(self, layer.name, func) # Add layer. if use_global_stats: func_class = _SingleArgumentFunctionTestMode else: func_class = _SingleArgumentFunction fwd = func_class(_CallChildLink(self, layer.name), finetune=False) self.forwards[layer.name] = fwd self._add_layer(layer) @_layer('Eltwise', 'ELTWISE') def _setup_eltwise(self, layer): # stable_prod_grad parameter is not supported now. operation = layer.eltwise_param.operation coeffs = layer.eltwise_param.coeff or None self.forwards[layer.name] = _EltwiseFunction(operation, coeffs) self._add_layer(layer) @_layer('Scale', None) def _setup_scale(self, layer): # Following parameters are not supported now: # - negative axis # - num_axes # - filler # - bias_filler # Get layer parameters. bottom = layer.bottom blobs = layer.blobs axis = layer.scale_param.axis bias_term = layer.scale_param.bias_term # Case of only one bottom where W is learnt parameter. if len(bottom) == 1: W_shape = blobs[0].shape.dim func = scale.Scale(axis, W_shape, bias_term) _Blob(blobs[0])(func.W.data) if bias_term: _Blob(blobs[1])(func.bias.b.data) # Case of two bottoms where W is given as a bottom. else: shape = blobs[0].shape.dim if bias_term else None func = scale.Scale( axis, bias_term=bias_term, bias_shape=shape) if bias_term: _Blob(blobs[0])(func.bias.b.data) # Add layer. with self.init_scope(): setattr(self, layer.name, func) self.forwards[layer.name] = _CallChildLink(self, layer.name) self._add_layer(layer) @_layer('Slice', 'SLICE') def _setup_slice(self, layer): if layer.slice_param.HasField('axis'): axis = layer.slice_param.axis elif layer.slice_param.HasField('slice_dim'): axis = layer.slice_param.slice_dim else: axis = 1 if layer.slice_param.slice_point: indices_or_sections = list(layer.slice_param.slice_point) else: indices_or_sections = len(list(layer.top)) self.forwards[layer.name] = _SingleArgumentFunction( functions.split_axis, indices_or_sections=indices_or_sections, axis=axis ) self._add_layer(layer) @_layer('Softmax', 'SOFTMAX') def _setup_softmax(self, layer): if layer.softmax_param.axis != 1: raise RuntimeError( 'Softmax along non-channel axis is not supported') if layer.softmax_param.engine == 0: # DEFAULT fw = functions.softmax elif layer.softmax_param.engine == 1: # CAFFE fw = _SingleArgumentFunctionWithCudnn(False, functions.softmax) elif layer.softmax_param.engine == 2: # CUDNN fw = _SingleArgumentFunctionWithCudnn(True, functions.softmax) self.forwards[layer.name] = fw self._add_layer(layer) @_layer('Sigmoid', 'SIGMOID') def _setup_sigmoid(self, layer): if layer.sigmoid_param.engine == 0: # DEFAULT fw = functions.sigmoid elif layer.sigmoid_param.engine == 1: # CAFFE fw = _SingleArgumentFunctionWithCudnn(False, functions.sigmoid) elif layer.sigmoid_param.engine == 2: # CUDNN fw = _SingleArgumentFunctionWithCudnn(True, functions.sigmoid) self.forwards[layer.name] = fw self._add_layer(layer) @_layer('SoftmaxWithLoss', 'SOFTMAX_LOSS') def _setup_softmax_with_loss(self, layer): if layer.softmax_param.axis != 1: raise RuntimeError( 'Softmax along non-channel axis is not supported') self.forwards[layer.name] = functions.softmax_cross_entropy self._add_layer(layer) @_layer('Split', 'SPLIT') def _setup_split(self, layer): for top in layer.top: self.split_map[top] = layer.bottom[0] # Internal functions def _get_ksize(param): if param.kernel_h > 0: return param.kernel_h, param.kernel_w elif type(param.kernel_size) == int: return param.kernel_size elif len(param.kernel_size) == 1: return param.kernel_size[0] else: return param.kernel_size def _get_stride(param): if param.stride_h > 0: return param.stride_h, param.stride_w elif type(param.stride) == int: return param.stride elif len(param.stride) == 0: return 1 elif len(param.stride) == 1: return param.stride[0] else: return param.stride def _get_pad(param): if param.pad_h > 0 or param.pad_w > 0: return param.pad_h, param.pad_w elif type(param.pad) == int: return param.pad elif len(param.pad) == 0: return 0 elif len(param.pad) == 1: return param.pad[0] else: return param.pad def _get_num(blob): if blob.num > 0: return blob.num else: return blob.shape.dim[0] def _get_channels(blob): if blob.channels > 0: return blob.channels else: return blob.shape.dim[1] def _get_height(blob): if blob.height > 0: return blob.height elif len(blob.shape.dim) == 2: return blob.shape.dim[0] elif len(blob.shape.dim) == 4: return blob.shape.dim[2] else: raise RuntimeError( '{}-dimensional array is not supported'.format( len(blob.shape.dim))) def _get_width(blob): if blob.width > 0: return blob.width elif len(blob.shape.dim) == 2: return blob.shape.dim[1] elif len(blob.shape.dim) == 4: return blob.shape.dim[3] else: raise RuntimeError( '{}-dimensional array is not supported'.format( len(blob.shape.dim))) # Internal class # __call__ must return Variable or tuple class _SingleArgumentFunction(object): def __init__(self, func, *args, **kwargs): self.func = func self.args = args self.kwargs = kwargs def __call__(self, x): return self.func(x, *self.args, **self.kwargs) class _SingleArgumentFunctionTestMode(_SingleArgumentFunction): def __call__(self, x): with configuration.using_config('train', False): return super(_SingleArgumentFunctionTestMode, self).__call__(x) class _ListArgumentFcuntion(object): def __init__(self, func, **kwargs): self.func = func self.kwargs = kwargs def __call__(self, *xs): return self.func(xs, **self.kwargs) class _SingleArgumentFunctionWithCudnn(_SingleArgumentFunction): def __init__(self, use_cudnn, func, *args, **kwargs): super(_SingleArgumentFunctionWithCudnn, self).__init__( func, *args, **kwargs) self.use_cudnn = use_cudnn def __call__(self, x): with configuration.using_config('use_cudnn', self.use_cudnn): return super(_SingleArgumentFunctionWithCudnn, self).__call__(x) class _CallChildLink(object): def __init__(self, caffe_func, name): self.name = name self.caffe_func = caffe_func def __call__(self, *xs, **kwargs): return self.caffe_func[self.name](*xs, **kwargs) class _EltwiseFunction(object): def __init__(self, operation, coeffs=None): if coeffs is not None: assert len(coeffs) > 0 self.operation = operation self.coeffs = coeffs def __call__(self, *xs): operation = self.operation if operation == 0: # PROD return six.moves.reduce(lambda x, y: x * y, xs), elif operation == 1: # SUM coeffs = self.coeffs if coeffs is not None: assert len(xs) == len(coeffs) xs = [x * coeff for x, coeff in zip(xs, coeffs)] return six.moves.reduce(lambda x, y: x + y, xs), elif operation == 2: # MAX return six.moves.reduce(lambda x, y: functions.maximum(x, y), xs), else: raise ValueError('Invalid EltwiseParameter.EltwiseOp value.')
22,528
31.841108
79
py
chainer
chainer-master/chainer/links/caffe/__init__.py
from chainer.links.caffe.caffe_function import CaffeFunction # NOQA
69
34
68
py
chainer
chainer-master/chainer/links/caffe/protobuf3/caffe_pb2.py
# Generated by the protocol buffer compiler. DO NOT EDIT! # source: caffe.proto import sys _b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1')) from google.protobuf.internal import enum_type_wrapper from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database from google.protobuf import descriptor_pb2 # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() DESCRIPTOR = _descriptor.FileDescriptor( name='caffe.proto', package='caffe', syntax='proto2', serialized_pb=_b('\n\x0b\x63\x61\x66\x66\x65.proto\x12\x05\x63\x61\x66\x66\x65\"\x1c\n\tBlobShape\x12\x0f\n\x03\x64im\x18\x01 \x03(\x03\x42\x02\x10\x01\"\xcc\x01\n\tBlobProto\x12\x1f\n\x05shape\x18\x07 \x01(\x0b\x32\x10.caffe.BlobShape\x12\x10\n\x04\x64\x61ta\x18\x05 \x03(\x02\x42\x02\x10\x01\x12\x10\n\x04\x64iff\x18\x06 \x03(\x02\x42\x02\x10\x01\x12\x17\n\x0b\x64ouble_data\x18\x08 \x03(\x01\x42\x02\x10\x01\x12\x17\n\x0b\x64ouble_diff\x18\t \x03(\x01\x42\x02\x10\x01\x12\x0e\n\x03num\x18\x01 \x01(\x05:\x01\x30\x12\x13\n\x08\x63hannels\x18\x02 \x01(\x05:\x01\x30\x12\x11\n\x06height\x18\x03 \x01(\x05:\x01\x30\x12\x10\n\x05width\x18\x04 \x01(\x05:\x01\x30\"2\n\x0f\x42lobProtoVector\x12\x1f\n\x05\x62lobs\x18\x01 \x03(\x0b\x32\x10.caffe.BlobProto\"\x81\x01\n\x05\x44\x61tum\x12\x10\n\x08\x63hannels\x18\x01 \x01(\x05\x12\x0e\n\x06height\x18\x02 \x01(\x05\x12\r\n\x05width\x18\x03 \x01(\x05\x12\x0c\n\x04\x64\x61ta\x18\x04 \x01(\x0c\x12\r\n\x05label\x18\x05 \x01(\x05\x12\x12\n\nfloat_data\x18\x06 \x03(\x02\x12\x16\n\x07\x65ncoded\x18\x07 \x01(\x08:\x05\x66\x61lse\"\x8a\x02\n\x0f\x46illerParameter\x12\x16\n\x04type\x18\x01 \x01(\t:\x08\x63onstant\x12\x10\n\x05value\x18\x02 \x01(\x02:\x01\x30\x12\x0e\n\x03min\x18\x03 \x01(\x02:\x01\x30\x12\x0e\n\x03max\x18\x04 \x01(\x02:\x01\x31\x12\x0f\n\x04mean\x18\x05 \x01(\x02:\x01\x30\x12\x0e\n\x03std\x18\x06 \x01(\x02:\x01\x31\x12\x12\n\x06sparse\x18\x07 \x01(\x05:\x02-1\x12\x42\n\rvariance_norm\x18\x08 \x01(\x0e\x32#.caffe.FillerParameter.VarianceNorm:\x06\x46\x41N_IN\"4\n\x0cVarianceNorm\x12\n\n\x06\x46\x41N_IN\x10\x00\x12\x0b\n\x07\x46\x41N_OUT\x10\x01\x12\x0b\n\x07\x41VERAGE\x10\x02\"\x8e\x02\n\x0cNetParameter\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\r\n\x05input\x18\x03 \x03(\t\x12%\n\x0binput_shape\x18\x08 \x03(\x0b\x32\x10.caffe.BlobShape\x12\x11\n\tinput_dim\x18\x04 \x03(\x05\x12\x1d\n\x0e\x66orce_backward\x18\x05 \x01(\x08:\x05\x66\x61lse\x12\x1e\n\x05state\x18\x06 \x01(\x0b\x32\x0f.caffe.NetState\x12\x19\n\ndebug_info\x18\x07 \x01(\x08:\x05\x66\x61lse\x12$\n\x05layer\x18\x64 \x03(\x0b\x32\x15.caffe.LayerParameter\x12\'\n\x06layers\x18\x02 \x03(\x0b\x32\x17.caffe.V1LayerParameter\"\x9c\n\n\x0fSolverParameter\x12\x0b\n\x03net\x18\x18 \x01(\t\x12&\n\tnet_param\x18\x19 \x01(\x0b\x32\x13.caffe.NetParameter\x12\x11\n\ttrain_net\x18\x01 \x01(\t\x12\x10\n\x08test_net\x18\x02 \x03(\t\x12,\n\x0ftrain_net_param\x18\x15 \x01(\x0b\x32\x13.caffe.NetParameter\x12+\n\x0etest_net_param\x18\x16 \x03(\x0b\x32\x13.caffe.NetParameter\x12$\n\x0btrain_state\x18\x1a \x01(\x0b\x32\x0f.caffe.NetState\x12#\n\ntest_state\x18\x1b \x03(\x0b\x32\x0f.caffe.NetState\x12\x11\n\ttest_iter\x18\x03 \x03(\x05\x12\x18\n\rtest_interval\x18\x04 \x01(\x05:\x01\x30\x12 \n\x11test_compute_loss\x18\x13 \x01(\x08:\x05\x66\x61lse\x12!\n\x13test_initialization\x18 \x01(\x08:\x04true\x12\x0f\n\x07\x62\x61se_lr\x18\x05 \x01(\x02\x12\x0f\n\x07\x64isplay\x18\x06 \x01(\x05\x12\x17\n\x0c\x61verage_loss\x18! \x01(\x05:\x01\x31\x12\x10\n\x08max_iter\x18\x07 \x01(\x05\x12\x14\n\titer_size\x18$ \x01(\x05:\x01\x31\x12\x11\n\tlr_policy\x18\x08 \x01(\t\x12\r\n\x05gamma\x18\t \x01(\x02\x12\r\n\x05power\x18\n \x01(\x02\x12\x10\n\x08momentum\x18\x0b \x01(\x02\x12\x14\n\x0cweight_decay\x18\x0c \x01(\x02\x12\x1f\n\x13regularization_type\x18\x1d \x01(\t:\x02L2\x12\x10\n\x08stepsize\x18\r \x01(\x05\x12\x11\n\tstepvalue\x18\" \x03(\x05\x12\x1a\n\x0e\x63lip_gradients\x18# \x01(\x02:\x02-1\x12\x13\n\x08snapshot\x18\x0e \x01(\x05:\x01\x30\x12\x17\n\x0fsnapshot_prefix\x18\x0f \x01(\t\x12\x1c\n\rsnapshot_diff\x18\x10 \x01(\x08:\x05\x66\x61lse\x12K\n\x0fsnapshot_format\x18% \x01(\x0e\x32%.caffe.SolverParameter.SnapshotFormat:\x0b\x42INARYPROTO\x12;\n\x0bsolver_mode\x18\x11 \x01(\x0e\x32!.caffe.SolverParameter.SolverMode:\x03GPU\x12\x14\n\tdevice_id\x18\x12 \x01(\x05:\x01\x30\x12\x17\n\x0brandom_seed\x18\x14 \x01(\x03:\x02-1\x12\x11\n\x04type\x18( \x01(\t:\x03SGD\x12\x14\n\x05\x64\x65lta\x18\x1f \x01(\x02:\x05\x31\x65-08\x12\x18\n\tmomentum2\x18\' \x01(\x02:\x05\x30.999\x12\x11\n\trms_decay\x18& \x01(\x02\x12\x19\n\ndebug_info\x18\x17 \x01(\x08:\x05\x66\x61lse\x12\"\n\x14snapshot_after_train\x18\x1c \x01(\x08:\x04true\x12;\n\x0bsolver_type\x18\x1e \x01(\x0e\x32!.caffe.SolverParameter.SolverType:\x03SGD\"+\n\x0eSnapshotFormat\x12\x08\n\x04HDF5\x10\x00\x12\x0f\n\x0b\x42INARYPROTO\x10\x01\"\x1e\n\nSolverMode\x12\x07\n\x03\x43PU\x10\x00\x12\x07\n\x03GPU\x10\x01\"U\n\nSolverType\x12\x07\n\x03SGD\x10\x00\x12\x0c\n\x08NESTEROV\x10\x01\x12\x0b\n\x07\x41\x44\x41GRAD\x10\x02\x12\x0b\n\x07RMSPROP\x10\x03\x12\x0c\n\x08\x41\x44\x41\x44\x45LTA\x10\x04\x12\x08\n\x04\x41\x44\x41M\x10\x05\"l\n\x0bSolverState\x12\x0c\n\x04iter\x18\x01 \x01(\x05\x12\x13\n\x0blearned_net\x18\x02 \x01(\t\x12!\n\x07history\x18\x03 \x03(\x0b\x32\x10.caffe.BlobProto\x12\x17\n\x0c\x63urrent_step\x18\x04 \x01(\x05:\x01\x30\"N\n\x08NetState\x12!\n\x05phase\x18\x01 \x01(\x0e\x32\x0c.caffe.Phase:\x04TEST\x12\x10\n\x05level\x18\x02 \x01(\x05:\x01\x30\x12\r\n\x05stage\x18\x03 \x03(\t\"s\n\x0cNetStateRule\x12\x1b\n\x05phase\x18\x01 \x01(\x0e\x32\x0c.caffe.Phase\x12\x11\n\tmin_level\x18\x02 \x01(\x05\x12\x11\n\tmax_level\x18\x03 \x01(\x05\x12\r\n\x05stage\x18\x04 \x03(\t\x12\x11\n\tnot_stage\x18\x05 \x03(\t\"\xa3\x01\n\tParamSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x31\n\nshare_mode\x18\x02 \x01(\x0e\x32\x1d.caffe.ParamSpec.DimCheckMode\x12\x12\n\x07lr_mult\x18\x03 \x01(\x02:\x01\x31\x12\x15\n\ndecay_mult\x18\x04 \x01(\x02:\x01\x31\"*\n\x0c\x44imCheckMode\x12\n\n\x06STRICT\x10\x00\x12\x0e\n\nPERMISSIVE\x10\x01\"\xc0\x12\n\x0eLayerParameter\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x0c\n\x04type\x18\x02 \x01(\t\x12\x0e\n\x06\x62ottom\x18\x03 \x03(\t\x12\x0b\n\x03top\x18\x04 \x03(\t\x12\x1b\n\x05phase\x18\n \x01(\x0e\x32\x0c.caffe.Phase\x12\x13\n\x0bloss_weight\x18\x05 \x03(\x02\x12\x1f\n\x05param\x18\x06 \x03(\x0b\x32\x10.caffe.ParamSpec\x12\x1f\n\x05\x62lobs\x18\x07 \x03(\x0b\x32\x10.caffe.BlobProto\x12\x16\n\x0epropagate_down\x18\x0b \x03(\x08\x12$\n\x07include\x18\x08 \x03(\x0b\x32\x13.caffe.NetStateRule\x12$\n\x07\x65xclude\x18\t \x03(\x0b\x32\x13.caffe.NetStateRule\x12\x37\n\x0ftransform_param\x18\x64 \x01(\x0b\x32\x1e.caffe.TransformationParameter\x12(\n\nloss_param\x18\x65 \x01(\x0b\x32\x14.caffe.LossParameter\x12\x30\n\x0e\x61\x63\x63uracy_param\x18\x66 \x01(\x0b\x32\x18.caffe.AccuracyParameter\x12,\n\x0c\x61rgmax_param\x18g \x01(\x0b\x32\x16.caffe.ArgMaxParameter\x12\x34\n\x10\x62\x61tch_norm_param\x18\x8b\x01 \x01(\x0b\x32\x19.caffe.BatchNormParameter\x12)\n\nbias_param\x18\x8d\x01 \x01(\x0b\x32\x14.caffe.BiasParameter\x12,\n\x0c\x63oncat_param\x18h \x01(\x0b\x32\x16.caffe.ConcatParameter\x12?\n\x16\x63ontrastive_loss_param\x18i \x01(\x0b\x32\x1f.caffe.ContrastiveLossParameter\x12\x36\n\x11\x63onvolution_param\x18j \x01(\x0b\x32\x1b.caffe.ConvolutionParameter\x12(\n\ndata_param\x18k \x01(\x0b\x32\x14.caffe.DataParameter\x12.\n\rdropout_param\x18l \x01(\x0b\x32\x17.caffe.DropoutParameter\x12\x33\n\x10\x64ummy_data_param\x18m \x01(\x0b\x32\x19.caffe.DummyDataParameter\x12.\n\reltwise_param\x18n \x01(\x0b\x32\x17.caffe.EltwiseParameter\x12\'\n\telu_param\x18\x8c\x01 \x01(\x0b\x32\x13.caffe.ELUParameter\x12+\n\x0b\x65mbed_param\x18\x89\x01 \x01(\x0b\x32\x15.caffe.EmbedParameter\x12&\n\texp_param\x18o \x01(\x0b\x32\x13.caffe.ExpParameter\x12/\n\rflatten_param\x18\x87\x01 \x01(\x0b\x32\x17.caffe.FlattenParameter\x12\x31\n\x0fhdf5_data_param\x18p \x01(\x0b\x32\x18.caffe.HDF5DataParameter\x12\x35\n\x11hdf5_output_param\x18q \x01(\x0b\x32\x1a.caffe.HDF5OutputParameter\x12\x33\n\x10hinge_loss_param\x18r \x01(\x0b\x32\x19.caffe.HingeLossParameter\x12\x33\n\x10image_data_param\x18s \x01(\x0b\x32\x19.caffe.ImageDataParameter\x12\x39\n\x13infogain_loss_param\x18t \x01(\x0b\x32\x1c.caffe.InfogainLossParameter\x12\x39\n\x13inner_product_param\x18u \x01(\x0b\x32\x1c.caffe.InnerProductParameter\x12\'\n\tlog_param\x18\x86\x01 \x01(\x0b\x32\x13.caffe.LogParameter\x12&\n\tlrn_param\x18v \x01(\x0b\x32\x13.caffe.LRNParameter\x12\x35\n\x11memory_data_param\x18w \x01(\x0b\x32\x1a.caffe.MemoryDataParameter\x12&\n\tmvn_param\x18x \x01(\x0b\x32\x13.caffe.MVNParameter\x12.\n\rpooling_param\x18y \x01(\x0b\x32\x17.caffe.PoolingParameter\x12*\n\x0bpower_param\x18z \x01(\x0b\x32\x15.caffe.PowerParameter\x12+\n\x0bprelu_param\x18\x83\x01 \x01(\x0b\x32\x15.caffe.PReLUParameter\x12-\n\x0cpython_param\x18\x82\x01 \x01(\x0b\x32\x16.caffe.PythonParameter\x12\x33\n\x0freduction_param\x18\x88\x01 \x01(\x0b\x32\x19.caffe.ReductionParameter\x12(\n\nrelu_param\x18{ \x01(\x0b\x32\x14.caffe.ReLUParameter\x12/\n\rreshape_param\x18\x85\x01 \x01(\x0b\x32\x17.caffe.ReshapeParameter\x12+\n\x0bscale_param\x18\x8e\x01 \x01(\x0b\x32\x15.caffe.ScaleParameter\x12.\n\rsigmoid_param\x18| \x01(\x0b\x32\x17.caffe.SigmoidParameter\x12.\n\rsoftmax_param\x18} \x01(\x0b\x32\x17.caffe.SoftmaxParameter\x12\'\n\tspp_param\x18\x84\x01 \x01(\x0b\x32\x13.caffe.SPPParameter\x12*\n\x0bslice_param\x18~ \x01(\x0b\x32\x15.caffe.SliceParameter\x12(\n\ntanh_param\x18\x7f \x01(\x0b\x32\x14.caffe.TanHParameter\x12\x33\n\x0fthreshold_param\x18\x80\x01 \x01(\x0b\x32\x19.caffe.ThresholdParameter\x12)\n\ntile_param\x18\x8a\x01 \x01(\x0b\x32\x14.caffe.TileParameter\x12\x36\n\x11window_data_param\x18\x81\x01 \x01(\x0b\x32\x1a.caffe.WindowDataParameter\"\xb6\x01\n\x17TransformationParameter\x12\x10\n\x05scale\x18\x01 \x01(\x02:\x01\x31\x12\x15\n\x06mirror\x18\x02 \x01(\x08:\x05\x66\x61lse\x12\x14\n\tcrop_size\x18\x03 \x01(\r:\x01\x30\x12\x11\n\tmean_file\x18\x04 \x01(\t\x12\x12\n\nmean_value\x18\x05 \x03(\x02\x12\x1a\n\x0b\x66orce_color\x18\x06 \x01(\x08:\x05\x66\x61lse\x12\x19\n\nforce_gray\x18\x07 \x01(\x08:\x05\x66\x61lse\"\xc2\x01\n\rLossParameter\x12\x14\n\x0cignore_label\x18\x01 \x01(\x05\x12\x44\n\rnormalization\x18\x03 \x01(\x0e\x32&.caffe.LossParameter.NormalizationMode:\x05VALID\x12\x11\n\tnormalize\x18\x02 \x01(\x08\"B\n\x11NormalizationMode\x12\x08\n\x04\x46ULL\x10\x00\x12\t\n\x05VALID\x10\x01\x12\x0e\n\nBATCH_SIZE\x10\x02\x12\x08\n\x04NONE\x10\x03\"L\n\x11\x41\x63\x63uracyParameter\x12\x10\n\x05top_k\x18\x01 \x01(\r:\x01\x31\x12\x0f\n\x04\x61xis\x18\x02 \x01(\x05:\x01\x31\x12\x14\n\x0cignore_label\x18\x03 \x01(\x05\"M\n\x0f\x41rgMaxParameter\x12\x1a\n\x0bout_max_val\x18\x01 \x01(\x08:\x05\x66\x61lse\x12\x10\n\x05top_k\x18\x02 \x01(\r:\x01\x31\x12\x0c\n\x04\x61xis\x18\x03 \x01(\x05\"9\n\x0f\x43oncatParameter\x12\x0f\n\x04\x61xis\x18\x02 \x01(\x05:\x01\x31\x12\x15\n\nconcat_dim\x18\x01 \x01(\r:\x01\x31\"j\n\x12\x42\x61tchNormParameter\x12\x18\n\x10use_global_stats\x18\x01 \x01(\x08\x12&\n\x17moving_average_fraction\x18\x02 \x01(\x02:\x05\x30.999\x12\x12\n\x03\x65ps\x18\x03 \x01(\x02:\x05\x31\x65-05\"]\n\rBiasParameter\x12\x0f\n\x04\x61xis\x18\x01 \x01(\x05:\x01\x31\x12\x13\n\x08num_axes\x18\x02 \x01(\x05:\x01\x31\x12&\n\x06\x66iller\x18\x03 \x01(\x0b\x32\x16.caffe.FillerParameter\"L\n\x18\x43ontrastiveLossParameter\x12\x11\n\x06margin\x18\x01 \x01(\x02:\x01\x31\x12\x1d\n\x0elegacy_version\x18\x02 \x01(\x08:\x05\x66\x61lse\"\xfc\x03\n\x14\x43onvolutionParameter\x12\x12\n\nnum_output\x18\x01 \x01(\r\x12\x17\n\tbias_term\x18\x02 \x01(\x08:\x04true\x12\x0b\n\x03pad\x18\x03 \x03(\r\x12\x13\n\x0bkernel_size\x18\x04 \x03(\r\x12\x0e\n\x06stride\x18\x06 \x03(\r\x12\x10\n\x08\x64ilation\x18\x12 \x03(\r\x12\x10\n\x05pad_h\x18\t \x01(\r:\x01\x30\x12\x10\n\x05pad_w\x18\n \x01(\r:\x01\x30\x12\x10\n\x08kernel_h\x18\x0b \x01(\r\x12\x10\n\x08kernel_w\x18\x0c \x01(\r\x12\x10\n\x08stride_h\x18\r \x01(\r\x12\x10\n\x08stride_w\x18\x0e \x01(\r\x12\x10\n\x05group\x18\x05 \x01(\r:\x01\x31\x12-\n\rweight_filler\x18\x07 \x01(\x0b\x32\x16.caffe.FillerParameter\x12+\n\x0b\x62ias_filler\x18\x08 \x01(\x0b\x32\x16.caffe.FillerParameter\x12;\n\x06\x65ngine\x18\x0f \x01(\x0e\x32\".caffe.ConvolutionParameter.Engine:\x07\x44\x45\x46\x41ULT\x12\x0f\n\x04\x61xis\x18\x10 \x01(\x05:\x01\x31\x12\x1e\n\x0f\x66orce_nd_im2col\x18\x11 \x01(\x08:\x05\x66\x61lse\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"\xa4\x02\n\rDataParameter\x12\x0e\n\x06source\x18\x01 \x01(\t\x12\x12\n\nbatch_size\x18\x04 \x01(\r\x12\x14\n\trand_skip\x18\x07 \x01(\r:\x01\x30\x12\x31\n\x07\x62\x61\x63kend\x18\x08 \x01(\x0e\x32\x17.caffe.DataParameter.DB:\x07LEVELDB\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x11\n\tmean_file\x18\x03 \x01(\t\x12\x14\n\tcrop_size\x18\x05 \x01(\r:\x01\x30\x12\x15\n\x06mirror\x18\x06 \x01(\x08:\x05\x66\x61lse\x12\"\n\x13\x66orce_encoded_color\x18\t \x01(\x08:\x05\x66\x61lse\x12\x13\n\x08prefetch\x18\n \x01(\r:\x01\x34\"\x1b\n\x02\x44\x42\x12\x0b\n\x07LEVELDB\x10\x00\x12\x08\n\x04LMDB\x10\x01\".\n\x10\x44ropoutParameter\x12\x1a\n\rdropout_ratio\x18\x01 \x01(\x02:\x03\x30.5\"\xa0\x01\n\x12\x44ummyDataParameter\x12+\n\x0b\x64\x61ta_filler\x18\x01 \x03(\x0b\x32\x16.caffe.FillerParameter\x12\x1f\n\x05shape\x18\x06 \x03(\x0b\x32\x10.caffe.BlobShape\x12\x0b\n\x03num\x18\x02 \x03(\r\x12\x10\n\x08\x63hannels\x18\x03 \x03(\r\x12\x0e\n\x06height\x18\x04 \x03(\r\x12\r\n\x05width\x18\x05 \x03(\r\"\xa5\x01\n\x10\x45ltwiseParameter\x12\x39\n\toperation\x18\x01 \x01(\x0e\x32!.caffe.EltwiseParameter.EltwiseOp:\x03SUM\x12\r\n\x05\x63oeff\x18\x02 \x03(\x02\x12\x1e\n\x10stable_prod_grad\x18\x03 \x01(\x08:\x04true\"\'\n\tEltwiseOp\x12\x08\n\x04PROD\x10\x00\x12\x07\n\x03SUM\x10\x01\x12\x07\n\x03MAX\x10\x02\" \n\x0c\x45LUParameter\x12\x10\n\x05\x61lpha\x18\x01 \x01(\x02:\x01\x31\"\xac\x01\n\x0e\x45mbedParameter\x12\x12\n\nnum_output\x18\x01 \x01(\r\x12\x11\n\tinput_dim\x18\x02 \x01(\r\x12\x17\n\tbias_term\x18\x03 \x01(\x08:\x04true\x12-\n\rweight_filler\x18\x04 \x01(\x0b\x32\x16.caffe.FillerParameter\x12+\n\x0b\x62ias_filler\x18\x05 \x01(\x0b\x32\x16.caffe.FillerParameter\"D\n\x0c\x45xpParameter\x12\x10\n\x04\x62\x61se\x18\x01 \x01(\x02:\x02-1\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x10\n\x05shift\x18\x03 \x01(\x02:\x01\x30\"9\n\x10\x46lattenParameter\x12\x0f\n\x04\x61xis\x18\x01 \x01(\x05:\x01\x31\x12\x14\n\x08\x65nd_axis\x18\x02 \x01(\x05:\x02-1\"O\n\x11HDF5DataParameter\x12\x0e\n\x06source\x18\x01 \x01(\t\x12\x12\n\nbatch_size\x18\x02 \x01(\r\x12\x16\n\x07shuffle\x18\x03 \x01(\x08:\x05\x66\x61lse\"(\n\x13HDF5OutputParameter\x12\x11\n\tfile_name\x18\x01 \x01(\t\"^\n\x12HingeLossParameter\x12\x30\n\x04norm\x18\x01 \x01(\x0e\x32\x1e.caffe.HingeLossParameter.Norm:\x02L1\"\x16\n\x04Norm\x12\x06\n\x02L1\x10\x01\x12\x06\n\x02L2\x10\x02\"\x97\x02\n\x12ImageDataParameter\x12\x0e\n\x06source\x18\x01 \x01(\t\x12\x15\n\nbatch_size\x18\x04 \x01(\r:\x01\x31\x12\x14\n\trand_skip\x18\x07 \x01(\r:\x01\x30\x12\x16\n\x07shuffle\x18\x08 \x01(\x08:\x05\x66\x61lse\x12\x15\n\nnew_height\x18\t \x01(\r:\x01\x30\x12\x14\n\tnew_width\x18\n \x01(\r:\x01\x30\x12\x16\n\x08is_color\x18\x0b \x01(\x08:\x04true\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x11\n\tmean_file\x18\x03 \x01(\t\x12\x14\n\tcrop_size\x18\x05 \x01(\r:\x01\x30\x12\x15\n\x06mirror\x18\x06 \x01(\x08:\x05\x66\x61lse\x12\x15\n\x0broot_folder\x18\x0c \x01(\t:\x00\"\'\n\x15InfogainLossParameter\x12\x0e\n\x06source\x18\x01 \x01(\t\"\xb1\x01\n\x15InnerProductParameter\x12\x12\n\nnum_output\x18\x01 \x01(\r\x12\x17\n\tbias_term\x18\x02 \x01(\x08:\x04true\x12-\n\rweight_filler\x18\x03 \x01(\x0b\x32\x16.caffe.FillerParameter\x12+\n\x0b\x62ias_filler\x18\x04 \x01(\x0b\x32\x16.caffe.FillerParameter\x12\x0f\n\x04\x61xis\x18\x05 \x01(\x05:\x01\x31\"D\n\x0cLogParameter\x12\x10\n\x04\x62\x61se\x18\x01 \x01(\x02:\x02-1\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x10\n\x05shift\x18\x03 \x01(\x02:\x01\x30\"\xb8\x02\n\x0cLRNParameter\x12\x15\n\nlocal_size\x18\x01 \x01(\r:\x01\x35\x12\x10\n\x05\x61lpha\x18\x02 \x01(\x02:\x01\x31\x12\x12\n\x04\x62\x65ta\x18\x03 \x01(\x02:\x04\x30.75\x12\x44\n\x0bnorm_region\x18\x04 \x01(\x0e\x32\x1e.caffe.LRNParameter.NormRegion:\x0f\x41\x43ROSS_CHANNELS\x12\x0c\n\x01k\x18\x05 \x01(\x02:\x01\x31\x12\x33\n\x06\x65ngine\x18\x06 \x01(\x0e\x32\x1a.caffe.LRNParameter.Engine:\x07\x44\x45\x46\x41ULT\"5\n\nNormRegion\x12\x13\n\x0f\x41\x43ROSS_CHANNELS\x10\x00\x12\x12\n\x0eWITHIN_CHANNEL\x10\x01\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"Z\n\x13MemoryDataParameter\x12\x12\n\nbatch_size\x18\x01 \x01(\r\x12\x10\n\x08\x63hannels\x18\x02 \x01(\r\x12\x0e\n\x06height\x18\x03 \x01(\r\x12\r\n\x05width\x18\x04 \x01(\r\"d\n\x0cMVNParameter\x12 \n\x12normalize_variance\x18\x01 \x01(\x08:\x04true\x12\x1e\n\x0f\x61\x63ross_channels\x18\x02 \x01(\x08:\x05\x66\x61lse\x12\x12\n\x03\x65ps\x18\x03 \x01(\x02:\x05\x31\x65-09\"\xa2\x03\n\x10PoolingParameter\x12\x35\n\x04pool\x18\x01 \x01(\x0e\x32\".caffe.PoolingParameter.PoolMethod:\x03MAX\x12\x0e\n\x03pad\x18\x04 \x01(\r:\x01\x30\x12\x10\n\x05pad_h\x18\t \x01(\r:\x01\x30\x12\x10\n\x05pad_w\x18\n \x01(\r:\x01\x30\x12\x13\n\x0bkernel_size\x18\x02 \x01(\r\x12\x10\n\x08kernel_h\x18\x05 \x01(\r\x12\x10\n\x08kernel_w\x18\x06 \x01(\r\x12\x11\n\x06stride\x18\x03 \x01(\r:\x01\x31\x12\x10\n\x08stride_h\x18\x07 \x01(\r\x12\x10\n\x08stride_w\x18\x08 \x01(\r\x12\x37\n\x06\x65ngine\x18\x0b \x01(\x0e\x32\x1e.caffe.PoolingParameter.Engine:\x07\x44\x45\x46\x41ULT\x12\x1d\n\x0eglobal_pooling\x18\x0c \x01(\x08:\x05\x66\x61lse\".\n\nPoolMethod\x12\x07\n\x03MAX\x10\x00\x12\x07\n\x03\x41VE\x10\x01\x12\x0e\n\nSTOCHASTIC\x10\x02\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"F\n\x0ePowerParameter\x12\x10\n\x05power\x18\x01 \x01(\x02:\x01\x31\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x10\n\x05shift\x18\x03 \x01(\x02:\x01\x30\"g\n\x0fPythonParameter\x12\x0e\n\x06module\x18\x01 \x01(\t\x12\r\n\x05layer\x18\x02 \x01(\t\x12\x13\n\tparam_str\x18\x03 \x01(\t:\x00\x12 \n\x11share_in_parallel\x18\x04 \x01(\x08:\x05\x66\x61lse\"\xad\x01\n\x12ReductionParameter\x12=\n\toperation\x18\x01 \x01(\x0e\x32%.caffe.ReductionParameter.ReductionOp:\x03SUM\x12\x0f\n\x04\x61xis\x18\x02 \x01(\x05:\x01\x30\x12\x10\n\x05\x63oeff\x18\x03 \x01(\x02:\x01\x31\"5\n\x0bReductionOp\x12\x07\n\x03SUM\x10\x01\x12\x08\n\x04\x41SUM\x10\x02\x12\t\n\x05SUMSQ\x10\x03\x12\x08\n\x04MEAN\x10\x04\"\x8d\x01\n\rReLUParameter\x12\x19\n\x0enegative_slope\x18\x01 \x01(\x02:\x01\x30\x12\x34\n\x06\x65ngine\x18\x02 \x01(\x0e\x32\x1b.caffe.ReLUParameter.Engine:\x07\x44\x45\x46\x41ULT\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"Z\n\x10ReshapeParameter\x12\x1f\n\x05shape\x18\x01 \x01(\x0b\x32\x10.caffe.BlobShape\x12\x0f\n\x04\x61xis\x18\x02 \x01(\x05:\x01\x30\x12\x14\n\x08num_axes\x18\x03 \x01(\x05:\x02-1\"\xa5\x01\n\x0eScaleParameter\x12\x0f\n\x04\x61xis\x18\x01 \x01(\x05:\x01\x31\x12\x13\n\x08num_axes\x18\x02 \x01(\x05:\x01\x31\x12&\n\x06\x66iller\x18\x03 \x01(\x0b\x32\x16.caffe.FillerParameter\x12\x18\n\tbias_term\x18\x04 \x01(\x08:\x05\x66\x61lse\x12+\n\x0b\x62ias_filler\x18\x05 \x01(\x0b\x32\x16.caffe.FillerParameter\"x\n\x10SigmoidParameter\x12\x37\n\x06\x65ngine\x18\x01 \x01(\x0e\x32\x1e.caffe.SigmoidParameter.Engine:\x07\x44\x45\x46\x41ULT\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"L\n\x0eSliceParameter\x12\x0f\n\x04\x61xis\x18\x03 \x01(\x05:\x01\x31\x12\x13\n\x0bslice_point\x18\x02 \x03(\r\x12\x14\n\tslice_dim\x18\x01 \x01(\r:\x01\x31\"\x89\x01\n\x10SoftmaxParameter\x12\x37\n\x06\x65ngine\x18\x01 \x01(\x0e\x32\x1e.caffe.SoftmaxParameter.Engine:\x07\x44\x45\x46\x41ULT\x12\x0f\n\x04\x61xis\x18\x02 \x01(\x05:\x01\x31\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"r\n\rTanHParameter\x12\x34\n\x06\x65ngine\x18\x01 \x01(\x0e\x32\x1b.caffe.TanHParameter.Engine:\x07\x44\x45\x46\x41ULT\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"/\n\rTileParameter\x12\x0f\n\x04\x61xis\x18\x01 \x01(\x05:\x01\x31\x12\r\n\x05tiles\x18\x02 \x01(\x05\"*\n\x12ThresholdParameter\x12\x14\n\tthreshold\x18\x01 \x01(\x02:\x01\x30\"\xc1\x02\n\x13WindowDataParameter\x12\x0e\n\x06source\x18\x01 \x01(\t\x12\x10\n\x05scale\x18\x02 \x01(\x02:\x01\x31\x12\x11\n\tmean_file\x18\x03 \x01(\t\x12\x12\n\nbatch_size\x18\x04 \x01(\r\x12\x14\n\tcrop_size\x18\x05 \x01(\r:\x01\x30\x12\x15\n\x06mirror\x18\x06 \x01(\x08:\x05\x66\x61lse\x12\x19\n\x0c\x66g_threshold\x18\x07 \x01(\x02:\x03\x30.5\x12\x19\n\x0c\x62g_threshold\x18\x08 \x01(\x02:\x03\x30.5\x12\x19\n\x0b\x66g_fraction\x18\t \x01(\x02:\x04\x30.25\x12\x16\n\x0b\x63ontext_pad\x18\n \x01(\r:\x01\x30\x12\x17\n\tcrop_mode\x18\x0b \x01(\t:\x04warp\x12\x1b\n\x0c\x63\x61\x63he_images\x18\x0c \x01(\x08:\x05\x66\x61lse\x12\x15\n\x0broot_folder\x18\r \x01(\t:\x00\"\xeb\x01\n\x0cSPPParameter\x12\x16\n\x0epyramid_height\x18\x01 \x01(\r\x12\x31\n\x04pool\x18\x02 \x01(\x0e\x32\x1e.caffe.SPPParameter.PoolMethod:\x03MAX\x12\x33\n\x06\x65ngine\x18\x06 \x01(\x0e\x32\x1a.caffe.SPPParameter.Engine:\x07\x44\x45\x46\x41ULT\".\n\nPoolMethod\x12\x07\n\x03MAX\x10\x00\x12\x07\n\x03\x41VE\x10\x01\x12\x0e\n\nSTOCHASTIC\x10\x02\"+\n\x06\x45ngine\x12\x0b\n\x07\x44\x45\x46\x41ULT\x10\x00\x12\t\n\x05\x43\x41\x46\x46\x45\x10\x01\x12\t\n\x05\x43UDNN\x10\x02\"\xe0\x13\n\x10V1LayerParameter\x12\x0e\n\x06\x62ottom\x18\x02 \x03(\t\x12\x0b\n\x03top\x18\x03 \x03(\t\x12\x0c\n\x04name\x18\x04 \x01(\t\x12$\n\x07include\x18 \x03(\x0b\x32\x13.caffe.NetStateRule\x12$\n\x07\x65xclude\x18! \x03(\x0b\x32\x13.caffe.NetStateRule\x12/\n\x04type\x18\x05 \x01(\x0e\x32!.caffe.V1LayerParameter.LayerType\x12\x1f\n\x05\x62lobs\x18\x06 \x03(\x0b\x32\x10.caffe.BlobProto\x12\x0e\n\x05param\x18\xe9\x07 \x03(\t\x12>\n\x0f\x62lob_share_mode\x18\xea\x07 \x03(\x0e\x32$.caffe.V1LayerParameter.DimCheckMode\x12\x10\n\x08\x62lobs_lr\x18\x07 \x03(\x02\x12\x14\n\x0cweight_decay\x18\x08 \x03(\x02\x12\x13\n\x0bloss_weight\x18# \x03(\x02\x12\x30\n\x0e\x61\x63\x63uracy_param\x18\x1b \x01(\x0b\x32\x18.caffe.AccuracyParameter\x12,\n\x0c\x61rgmax_param\x18\x17 \x01(\x0b\x32\x16.caffe.ArgMaxParameter\x12,\n\x0c\x63oncat_param\x18\t \x01(\x0b\x32\x16.caffe.ConcatParameter\x12?\n\x16\x63ontrastive_loss_param\x18( \x01(\x0b\x32\x1f.caffe.ContrastiveLossParameter\x12\x36\n\x11\x63onvolution_param\x18\n \x01(\x0b\x32\x1b.caffe.ConvolutionParameter\x12(\n\ndata_param\x18\x0b \x01(\x0b\x32\x14.caffe.DataParameter\x12.\n\rdropout_param\x18\x0c \x01(\x0b\x32\x17.caffe.DropoutParameter\x12\x33\n\x10\x64ummy_data_param\x18\x1a \x01(\x0b\x32\x19.caffe.DummyDataParameter\x12.\n\reltwise_param\x18\x18 \x01(\x0b\x32\x17.caffe.EltwiseParameter\x12&\n\texp_param\x18) \x01(\x0b\x32\x13.caffe.ExpParameter\x12\x31\n\x0fhdf5_data_param\x18\r \x01(\x0b\x32\x18.caffe.HDF5DataParameter\x12\x35\n\x11hdf5_output_param\x18\x0e \x01(\x0b\x32\x1a.caffe.HDF5OutputParameter\x12\x33\n\x10hinge_loss_param\x18\x1d \x01(\x0b\x32\x19.caffe.HingeLossParameter\x12\x33\n\x10image_data_param\x18\x0f \x01(\x0b\x32\x19.caffe.ImageDataParameter\x12\x39\n\x13infogain_loss_param\x18\x10 \x01(\x0b\x32\x1c.caffe.InfogainLossParameter\x12\x39\n\x13inner_product_param\x18\x11 \x01(\x0b\x32\x1c.caffe.InnerProductParameter\x12&\n\tlrn_param\x18\x12 \x01(\x0b\x32\x13.caffe.LRNParameter\x12\x35\n\x11memory_data_param\x18\x16 \x01(\x0b\x32\x1a.caffe.MemoryDataParameter\x12&\n\tmvn_param\x18\" \x01(\x0b\x32\x13.caffe.MVNParameter\x12.\n\rpooling_param\x18\x13 \x01(\x0b\x32\x17.caffe.PoolingParameter\x12*\n\x0bpower_param\x18\x15 \x01(\x0b\x32\x15.caffe.PowerParameter\x12(\n\nrelu_param\x18\x1e \x01(\x0b\x32\x14.caffe.ReLUParameter\x12.\n\rsigmoid_param\x18& \x01(\x0b\x32\x17.caffe.SigmoidParameter\x12.\n\rsoftmax_param\x18\' \x01(\x0b\x32\x17.caffe.SoftmaxParameter\x12*\n\x0bslice_param\x18\x1f \x01(\x0b\x32\x15.caffe.SliceParameter\x12(\n\ntanh_param\x18% \x01(\x0b\x32\x14.caffe.TanHParameter\x12\x32\n\x0fthreshold_param\x18\x19 \x01(\x0b\x32\x19.caffe.ThresholdParameter\x12\x35\n\x11window_data_param\x18\x14 \x01(\x0b\x32\x1a.caffe.WindowDataParameter\x12\x37\n\x0ftransform_param\x18$ \x01(\x0b\x32\x1e.caffe.TransformationParameter\x12(\n\nloss_param\x18* \x01(\x0b\x32\x14.caffe.LossParameter\x12&\n\x05layer\x18\x01 \x01(\x0b\x32\x17.caffe.V0LayerParameter\"\xd8\x04\n\tLayerType\x12\x08\n\x04NONE\x10\x00\x12\n\n\x06\x41\x42SVAL\x10#\x12\x0c\n\x08\x41\x43\x43URACY\x10\x01\x12\n\n\x06\x41RGMAX\x10\x1e\x12\x08\n\x04\x42NLL\x10\x02\x12\n\n\x06\x43ONCAT\x10\x03\x12\x14\n\x10\x43ONTRASTIVE_LOSS\x10%\x12\x0f\n\x0b\x43ONVOLUTION\x10\x04\x12\x08\n\x04\x44\x41TA\x10\x05\x12\x11\n\rDECONVOLUTION\x10\'\x12\x0b\n\x07\x44ROPOUT\x10\x06\x12\x0e\n\nDUMMY_DATA\x10 \x12\x12\n\x0e\x45UCLIDEAN_LOSS\x10\x07\x12\x0b\n\x07\x45LTWISE\x10\x19\x12\x07\n\x03\x45XP\x10&\x12\x0b\n\x07\x46LATTEN\x10\x08\x12\r\n\tHDF5_DATA\x10\t\x12\x0f\n\x0bHDF5_OUTPUT\x10\n\x12\x0e\n\nHINGE_LOSS\x10\x1c\x12\n\n\x06IM2COL\x10\x0b\x12\x0e\n\nIMAGE_DATA\x10\x0c\x12\x11\n\rINFOGAIN_LOSS\x10\r\x12\x11\n\rINNER_PRODUCT\x10\x0e\x12\x07\n\x03LRN\x10\x0f\x12\x0f\n\x0bMEMORY_DATA\x10\x1d\x12\x1d\n\x19MULTINOMIAL_LOGISTIC_LOSS\x10\x10\x12\x07\n\x03MVN\x10\"\x12\x0b\n\x07POOLING\x10\x11\x12\t\n\x05POWER\x10\x1a\x12\x08\n\x04RELU\x10\x12\x12\x0b\n\x07SIGMOID\x10\x13\x12\x1e\n\x1aSIGMOID_CROSS_ENTROPY_LOSS\x10\x1b\x12\x0b\n\x07SILENCE\x10$\x12\x0b\n\x07SOFTMAX\x10\x14\x12\x10\n\x0cSOFTMAX_LOSS\x10\x15\x12\t\n\x05SPLIT\x10\x16\x12\t\n\x05SLICE\x10!\x12\x08\n\x04TANH\x10\x17\x12\x0f\n\x0bWINDOW_DATA\x10\x18\x12\r\n\tTHRESHOLD\x10\x1f\"*\n\x0c\x44imCheckMode\x12\n\n\x06STRICT\x10\x00\x12\x0e\n\nPERMISSIVE\x10\x01\"\xfd\x07\n\x10V0LayerParameter\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x0c\n\x04type\x18\x02 \x01(\t\x12\x12\n\nnum_output\x18\x03 \x01(\r\x12\x16\n\x08\x62iasterm\x18\x04 \x01(\x08:\x04true\x12-\n\rweight_filler\x18\x05 \x01(\x0b\x32\x16.caffe.FillerParameter\x12+\n\x0b\x62ias_filler\x18\x06 \x01(\x0b\x32\x16.caffe.FillerParameter\x12\x0e\n\x03pad\x18\x07 \x01(\r:\x01\x30\x12\x12\n\nkernelsize\x18\x08 \x01(\r\x12\x10\n\x05group\x18\t \x01(\r:\x01\x31\x12\x11\n\x06stride\x18\n \x01(\r:\x01\x31\x12\x35\n\x04pool\x18\x0b \x01(\x0e\x32\".caffe.V0LayerParameter.PoolMethod:\x03MAX\x12\x1a\n\rdropout_ratio\x18\x0c \x01(\x02:\x03\x30.5\x12\x15\n\nlocal_size\x18\r \x01(\r:\x01\x35\x12\x10\n\x05\x61lpha\x18\x0e \x01(\x02:\x01\x31\x12\x12\n\x04\x62\x65ta\x18\x0f \x01(\x02:\x04\x30.75\x12\x0c\n\x01k\x18\x16 \x01(\x02:\x01\x31\x12\x0e\n\x06source\x18\x10 \x01(\t\x12\x10\n\x05scale\x18\x11 \x01(\x02:\x01\x31\x12\x10\n\x08meanfile\x18\x12 \x01(\t\x12\x11\n\tbatchsize\x18\x13 \x01(\r\x12\x13\n\x08\x63ropsize\x18\x14 \x01(\r:\x01\x30\x12\x15\n\x06mirror\x18\x15 \x01(\x08:\x05\x66\x61lse\x12\x1f\n\x05\x62lobs\x18\x32 \x03(\x0b\x32\x10.caffe.BlobProto\x12\x10\n\x08\x62lobs_lr\x18\x33 \x03(\x02\x12\x14\n\x0cweight_decay\x18\x34 \x03(\x02\x12\x14\n\trand_skip\x18\x35 \x01(\r:\x01\x30\x12\x1d\n\x10\x64\x65t_fg_threshold\x18\x36 \x01(\x02:\x03\x30.5\x12\x1d\n\x10\x64\x65t_bg_threshold\x18\x37 \x01(\x02:\x03\x30.5\x12\x1d\n\x0f\x64\x65t_fg_fraction\x18\x38 \x01(\x02:\x04\x30.25\x12\x1a\n\x0f\x64\x65t_context_pad\x18: \x01(\r:\x01\x30\x12\x1b\n\rdet_crop_mode\x18; \x01(\t:\x04warp\x12\x12\n\x07new_num\x18< \x01(\x05:\x01\x30\x12\x17\n\x0cnew_channels\x18= \x01(\x05:\x01\x30\x12\x15\n\nnew_height\x18> \x01(\x05:\x01\x30\x12\x14\n\tnew_width\x18? \x01(\x05:\x01\x30\x12\x1d\n\x0eshuffle_images\x18@ \x01(\x08:\x05\x66\x61lse\x12\x15\n\nconcat_dim\x18\x41 \x01(\r:\x01\x31\x12\x36\n\x11hdf5_output_param\x18\xe9\x07 \x01(\x0b\x32\x1a.caffe.HDF5OutputParameter\".\n\nPoolMethod\x12\x07\n\x03MAX\x10\x00\x12\x07\n\x03\x41VE\x10\x01\x12\x0e\n\nSTOCHASTIC\x10\x02\"W\n\x0ePReLUParameter\x12&\n\x06\x66iller\x18\x01 \x01(\x0b\x32\x16.caffe.FillerParameter\x12\x1d\n\x0e\x63hannel_shared\x18\x02 \x01(\x08:\x05\x66\x61lse*\x1c\n\x05Phase\x12\t\n\x05TRAIN\x10\x00\x12\x08\n\x04TEST\x10\x01') ) _sym_db.RegisterFileDescriptor(DESCRIPTOR) _PHASE = _descriptor.EnumDescriptor( name='Phase', full_name='caffe.Phase', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='TRAIN', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='TEST', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=14776, serialized_end=14804, ) _sym_db.RegisterEnumDescriptor(_PHASE) Phase = enum_type_wrapper.EnumTypeWrapper(_PHASE) TRAIN = 0 TEST = 1 _FILLERPARAMETER_VARIANCENORM = _descriptor.EnumDescriptor( name='VarianceNorm', full_name='caffe.FillerParameter.VarianceNorm', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='FAN_IN', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='FAN_OUT', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='AVERAGE', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=658, serialized_end=710, ) _sym_db.RegisterEnumDescriptor(_FILLERPARAMETER_VARIANCENORM) _SOLVERPARAMETER_SNAPSHOTFORMAT = _descriptor.EnumDescriptor( name='SnapshotFormat', full_name='caffe.SolverParameter.SnapshotFormat', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='HDF5', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='BINARYPROTO', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=2132, serialized_end=2175, ) _sym_db.RegisterEnumDescriptor(_SOLVERPARAMETER_SNAPSHOTFORMAT) _SOLVERPARAMETER_SOLVERMODE = _descriptor.EnumDescriptor( name='SolverMode', full_name='caffe.SolverParameter.SolverMode', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='CPU', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='GPU', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=2177, serialized_end=2207, ) _sym_db.RegisterEnumDescriptor(_SOLVERPARAMETER_SOLVERMODE) _SOLVERPARAMETER_SOLVERTYPE = _descriptor.EnumDescriptor( name='SolverType', full_name='caffe.SolverParameter.SolverType', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='SGD', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='NESTEROV', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='ADAGRAD', index=2, number=2, options=None, type=None), _descriptor.EnumValueDescriptor( name='RMSPROP', index=3, number=3, options=None, type=None), _descriptor.EnumValueDescriptor( name='ADADELTA', index=4, number=4, options=None, type=None), _descriptor.EnumValueDescriptor( name='ADAM', index=5, number=5, options=None, type=None), ], containing_type=None, options=None, serialized_start=2209, serialized_end=2294, ) _sym_db.RegisterEnumDescriptor(_SOLVERPARAMETER_SOLVERTYPE) _PARAMSPEC_DIMCHECKMODE = _descriptor.EnumDescriptor( name='DimCheckMode', full_name='caffe.ParamSpec.DimCheckMode', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='STRICT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='PERMISSIVE', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=2725, serialized_end=2767, ) _sym_db.RegisterEnumDescriptor(_PARAMSPEC_DIMCHECKMODE) _LOSSPARAMETER_NORMALIZATIONMODE = _descriptor.EnumDescriptor( name='NormalizationMode', full_name='caffe.LossParameter.NormalizationMode', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='FULL', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='VALID', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='BATCH_SIZE', index=2, number=2, options=None, type=None), _descriptor.EnumValueDescriptor( name='NONE', index=3, number=3, options=None, type=None), ], containing_type=None, options=None, serialized_start=5454, serialized_end=5520, ) _sym_db.RegisterEnumDescriptor(_LOSSPARAMETER_NORMALIZATIONMODE) _CONVOLUTIONPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.ConvolutionParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_CONVOLUTIONPARAMETER_ENGINE) _DATAPARAMETER_DB = _descriptor.EnumDescriptor( name='DB', full_name='caffe.DataParameter.DB', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='LEVELDB', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='LMDB', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=6796, serialized_end=6823, ) _sym_db.RegisterEnumDescriptor(_DATAPARAMETER_DB) _ELTWISEPARAMETER_ELTWISEOP = _descriptor.EnumDescriptor( name='EltwiseOp', full_name='caffe.EltwiseParameter.EltwiseOp', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='PROD', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='SUM', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='MAX', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=7163, serialized_end=7202, ) _sym_db.RegisterEnumDescriptor(_ELTWISEPARAMETER_ELTWISEOP) _HINGELOSSPARAMETER_NORM = _descriptor.EnumDescriptor( name='Norm', full_name='caffe.HingeLossParameter.Norm', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='L1', index=0, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='L2', index=1, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=7737, serialized_end=7759, ) _sym_db.RegisterEnumDescriptor(_HINGELOSSPARAMETER_NORM) _LRNPARAMETER_NORMREGION = _descriptor.EnumDescriptor( name='NormRegion', full_name='caffe.LRNParameter.NormRegion', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='ACROSS_CHANNELS', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='WITHIN_CHANNEL', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=8549, serialized_end=8602, ) _sym_db.RegisterEnumDescriptor(_LRNPARAMETER_NORMREGION) _LRNPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.LRNParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_LRNPARAMETER_ENGINE) _POOLINGPARAMETER_POOLMETHOD = _descriptor.EnumDescriptor( name='PoolMethod', full_name='caffe.PoolingParameter.PoolMethod', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='MAX', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='AVE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='STOCHASTIC', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=9171, serialized_end=9217, ) _sym_db.RegisterEnumDescriptor(_POOLINGPARAMETER_POOLMETHOD) _POOLINGPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.PoolingParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_POOLINGPARAMETER_ENGINE) _REDUCTIONPARAMETER_REDUCTIONOP = _descriptor.EnumDescriptor( name='ReductionOp', full_name='caffe.ReductionParameter.ReductionOp', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='SUM', index=0, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='ASUM', index=1, number=2, options=None, type=None), _descriptor.EnumValueDescriptor( name='SUMSQ', index=2, number=3, options=None, type=None), _descriptor.EnumValueDescriptor( name='MEAN', index=3, number=4, options=None, type=None), ], containing_type=None, options=None, serialized_start=9562, serialized_end=9615, ) _sym_db.RegisterEnumDescriptor(_REDUCTIONPARAMETER_REDUCTIONOP) _RELUPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.ReLUParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_RELUPARAMETER_ENGINE) _SIGMOIDPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.SigmoidParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_SIGMOIDPARAMETER_ENGINE) _SOFTMAXPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.SoftmaxParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_SOFTMAXPARAMETER_ENGINE) _TANHPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.TanHParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_TANHPARAMETER_ENGINE) _SPPPARAMETER_POOLMETHOD = _descriptor.EnumDescriptor( name='PoolMethod', full_name='caffe.SPPParameter.PoolMethod', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='MAX', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='AVE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='STOCHASTIC', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=9171, serialized_end=9217, ) _sym_db.RegisterEnumDescriptor(_SPPPARAMETER_POOLMETHOD) _SPPPARAMETER_ENGINE = _descriptor.EnumDescriptor( name='Engine', full_name='caffe.SPPParameter.Engine', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='DEFAULT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='CAFFE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='CUDNN', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=6485, serialized_end=6528, ) _sym_db.RegisterEnumDescriptor(_SPPPARAMETER_ENGINE) _V1LAYERPARAMETER_LAYERTYPE = _descriptor.EnumDescriptor( name='LayerType', full_name='caffe.V1LayerParameter.LayerType', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='NONE', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='ABSVAL', index=1, number=35, options=None, type=None), _descriptor.EnumValueDescriptor( name='ACCURACY', index=2, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='ARGMAX', index=3, number=30, options=None, type=None), _descriptor.EnumValueDescriptor( name='BNLL', index=4, number=2, options=None, type=None), _descriptor.EnumValueDescriptor( name='CONCAT', index=5, number=3, options=None, type=None), _descriptor.EnumValueDescriptor( name='CONTRASTIVE_LOSS', index=6, number=37, options=None, type=None), _descriptor.EnumValueDescriptor( name='CONVOLUTION', index=7, number=4, options=None, type=None), _descriptor.EnumValueDescriptor( name='DATA', index=8, number=5, options=None, type=None), _descriptor.EnumValueDescriptor( name='DECONVOLUTION', index=9, number=39, options=None, type=None), _descriptor.EnumValueDescriptor( name='DROPOUT', index=10, number=6, options=None, type=None), _descriptor.EnumValueDescriptor( name='DUMMY_DATA', index=11, number=32, options=None, type=None), _descriptor.EnumValueDescriptor( name='EUCLIDEAN_LOSS', index=12, number=7, options=None, type=None), _descriptor.EnumValueDescriptor( name='ELTWISE', index=13, number=25, options=None, type=None), _descriptor.EnumValueDescriptor( name='EXP', index=14, number=38, options=None, type=None), _descriptor.EnumValueDescriptor( name='FLATTEN', index=15, number=8, options=None, type=None), _descriptor.EnumValueDescriptor( name='HDF5_DATA', index=16, number=9, options=None, type=None), _descriptor.EnumValueDescriptor( name='HDF5_OUTPUT', index=17, number=10, options=None, type=None), _descriptor.EnumValueDescriptor( name='HINGE_LOSS', index=18, number=28, options=None, type=None), _descriptor.EnumValueDescriptor( name='IM2COL', index=19, number=11, options=None, type=None), _descriptor.EnumValueDescriptor( name='IMAGE_DATA', index=20, number=12, options=None, type=None), _descriptor.EnumValueDescriptor( name='INFOGAIN_LOSS', index=21, number=13, options=None, type=None), _descriptor.EnumValueDescriptor( name='INNER_PRODUCT', index=22, number=14, options=None, type=None), _descriptor.EnumValueDescriptor( name='LRN', index=23, number=15, options=None, type=None), _descriptor.EnumValueDescriptor( name='MEMORY_DATA', index=24, number=29, options=None, type=None), _descriptor.EnumValueDescriptor( name='MULTINOMIAL_LOGISTIC_LOSS', index=25, number=16, options=None, type=None), _descriptor.EnumValueDescriptor( name='MVN', index=26, number=34, options=None, type=None), _descriptor.EnumValueDescriptor( name='POOLING', index=27, number=17, options=None, type=None), _descriptor.EnumValueDescriptor( name='POWER', index=28, number=26, options=None, type=None), _descriptor.EnumValueDescriptor( name='RELU', index=29, number=18, options=None, type=None), _descriptor.EnumValueDescriptor( name='SIGMOID', index=30, number=19, options=None, type=None), _descriptor.EnumValueDescriptor( name='SIGMOID_CROSS_ENTROPY_LOSS', index=31, number=27, options=None, type=None), _descriptor.EnumValueDescriptor( name='SILENCE', index=32, number=36, options=None, type=None), _descriptor.EnumValueDescriptor( name='SOFTMAX', index=33, number=20, options=None, type=None), _descriptor.EnumValueDescriptor( name='SOFTMAX_LOSS', index=34, number=21, options=None, type=None), _descriptor.EnumValueDescriptor( name='SPLIT', index=35, number=22, options=None, type=None), _descriptor.EnumValueDescriptor( name='SLICE', index=36, number=33, options=None, type=None), _descriptor.EnumValueDescriptor( name='TANH', index=37, number=23, options=None, type=None), _descriptor.EnumValueDescriptor( name='WINDOW_DATA', index=38, number=24, options=None, type=None), _descriptor.EnumValueDescriptor( name='THRESHOLD', index=39, number=31, options=None, type=None), ], containing_type=None, options=None, serialized_start=13017, serialized_end=13617, ) _sym_db.RegisterEnumDescriptor(_V1LAYERPARAMETER_LAYERTYPE) _V1LAYERPARAMETER_DIMCHECKMODE = _descriptor.EnumDescriptor( name='DimCheckMode', full_name='caffe.V1LayerParameter.DimCheckMode', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='STRICT', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='PERMISSIVE', index=1, number=1, options=None, type=None), ], containing_type=None, options=None, serialized_start=2725, serialized_end=2767, ) _sym_db.RegisterEnumDescriptor(_V1LAYERPARAMETER_DIMCHECKMODE) _V0LAYERPARAMETER_POOLMETHOD = _descriptor.EnumDescriptor( name='PoolMethod', full_name='caffe.V0LayerParameter.PoolMethod', filename=None, file=DESCRIPTOR, values=[ _descriptor.EnumValueDescriptor( name='MAX', index=0, number=0, options=None, type=None), _descriptor.EnumValueDescriptor( name='AVE', index=1, number=1, options=None, type=None), _descriptor.EnumValueDescriptor( name='STOCHASTIC', index=2, number=2, options=None, type=None), ], containing_type=None, options=None, serialized_start=9171, serialized_end=9217, ) _sym_db.RegisterEnumDescriptor(_V0LAYERPARAMETER_POOLMETHOD) _BLOBSHAPE = _descriptor.Descriptor( name='BlobShape', full_name='caffe.BlobShape', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='dim', full_name='caffe.BlobShape.dim', index=0, number=1, type=3, cpp_type=2, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=_descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001'))), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=22, serialized_end=50, ) _BLOBPROTO = _descriptor.Descriptor( name='BlobProto', full_name='caffe.BlobProto', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='shape', full_name='caffe.BlobProto.shape', index=0, number=7, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='data', full_name='caffe.BlobProto.data', index=1, number=5, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=_descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001'))), _descriptor.FieldDescriptor( name='diff', full_name='caffe.BlobProto.diff', index=2, number=6, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=_descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001'))), _descriptor.FieldDescriptor( name='double_data', full_name='caffe.BlobProto.double_data', index=3, number=8, type=1, cpp_type=5, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=_descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001'))), _descriptor.FieldDescriptor( name='double_diff', full_name='caffe.BlobProto.double_diff', index=4, number=9, type=1, cpp_type=5, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=_descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001'))), _descriptor.FieldDescriptor( name='num', full_name='caffe.BlobProto.num', index=5, number=1, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='channels', full_name='caffe.BlobProto.channels', index=6, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='height', full_name='caffe.BlobProto.height', index=7, number=3, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='width', full_name='caffe.BlobProto.width', index=8, number=4, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=53, serialized_end=257, ) _BLOBPROTOVECTOR = _descriptor.Descriptor( name='BlobProtoVector', full_name='caffe.BlobProtoVector', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='blobs', full_name='caffe.BlobProtoVector.blobs', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=259, serialized_end=309, ) _DATUM = _descriptor.Descriptor( name='Datum', full_name='caffe.Datum', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='channels', full_name='caffe.Datum.channels', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='height', full_name='caffe.Datum.height', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='width', full_name='caffe.Datum.width', index=2, number=3, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='data', full_name='caffe.Datum.data', index=3, number=4, type=12, cpp_type=9, label=1, has_default_value=False, default_value=_b(""), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='label', full_name='caffe.Datum.label', index=4, number=5, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='float_data', full_name='caffe.Datum.float_data', index=5, number=6, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='encoded', full_name='caffe.Datum.encoded', index=6, number=7, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=312, serialized_end=441, ) _FILLERPARAMETER = _descriptor.Descriptor( name='FillerParameter', full_name='caffe.FillerParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='type', full_name='caffe.FillerParameter.type', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("constant").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='value', full_name='caffe.FillerParameter.value', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='min', full_name='caffe.FillerParameter.min', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='max', full_name='caffe.FillerParameter.max', index=3, number=4, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean', full_name='caffe.FillerParameter.mean', index=4, number=5, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='std', full_name='caffe.FillerParameter.std', index=5, number=6, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='sparse', full_name='caffe.FillerParameter.sparse', index=6, number=7, type=5, cpp_type=1, label=1, has_default_value=True, default_value=-1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='variance_norm', full_name='caffe.FillerParameter.variance_norm', index=7, number=8, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _FILLERPARAMETER_VARIANCENORM, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=444, serialized_end=710, ) _NETPARAMETER = _descriptor.Descriptor( name='NetParameter', full_name='caffe.NetParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='name', full_name='caffe.NetParameter.name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='input', full_name='caffe.NetParameter.input', index=1, number=3, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='input_shape', full_name='caffe.NetParameter.input_shape', index=2, number=8, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='input_dim', full_name='caffe.NetParameter.input_dim', index=3, number=4, type=5, cpp_type=1, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='force_backward', full_name='caffe.NetParameter.force_backward', index=4, number=5, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='state', full_name='caffe.NetParameter.state', index=5, number=6, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='debug_info', full_name='caffe.NetParameter.debug_info', index=6, number=7, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='layer', full_name='caffe.NetParameter.layer', index=7, number=100, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='layers', full_name='caffe.NetParameter.layers', index=8, number=2, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=713, serialized_end=983, ) _SOLVERPARAMETER = _descriptor.Descriptor( name='SolverParameter', full_name='caffe.SolverParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='net', full_name='caffe.SolverParameter.net', index=0, number=24, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='net_param', full_name='caffe.SolverParameter.net_param', index=1, number=25, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='train_net', full_name='caffe.SolverParameter.train_net', index=2, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_net', full_name='caffe.SolverParameter.test_net', index=3, number=2, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='train_net_param', full_name='caffe.SolverParameter.train_net_param', index=4, number=21, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_net_param', full_name='caffe.SolverParameter.test_net_param', index=5, number=22, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='train_state', full_name='caffe.SolverParameter.train_state', index=6, number=26, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_state', full_name='caffe.SolverParameter.test_state', index=7, number=27, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_iter', full_name='caffe.SolverParameter.test_iter', index=8, number=3, type=5, cpp_type=1, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_interval', full_name='caffe.SolverParameter.test_interval', index=9, number=4, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_compute_loss', full_name='caffe.SolverParameter.test_compute_loss', index=10, number=19, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='test_initialization', full_name='caffe.SolverParameter.test_initialization', index=11, number=32, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='base_lr', full_name='caffe.SolverParameter.base_lr', index=12, number=5, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='display', full_name='caffe.SolverParameter.display', index=13, number=6, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='average_loss', full_name='caffe.SolverParameter.average_loss', index=14, number=33, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='max_iter', full_name='caffe.SolverParameter.max_iter', index=15, number=7, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='iter_size', full_name='caffe.SolverParameter.iter_size', index=16, number=36, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='lr_policy', full_name='caffe.SolverParameter.lr_policy', index=17, number=8, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='gamma', full_name='caffe.SolverParameter.gamma', index=18, number=9, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='power', full_name='caffe.SolverParameter.power', index=19, number=10, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='momentum', full_name='caffe.SolverParameter.momentum', index=20, number=11, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_decay', full_name='caffe.SolverParameter.weight_decay', index=21, number=12, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='regularization_type', full_name='caffe.SolverParameter.regularization_type', index=22, number=29, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("L2").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stepsize', full_name='caffe.SolverParameter.stepsize', index=23, number=13, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stepvalue', full_name='caffe.SolverParameter.stepvalue', index=24, number=34, type=5, cpp_type=1, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='clip_gradients', full_name='caffe.SolverParameter.clip_gradients', index=25, number=35, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(-1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='snapshot', full_name='caffe.SolverParameter.snapshot', index=26, number=14, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='snapshot_prefix', full_name='caffe.SolverParameter.snapshot_prefix', index=27, number=15, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='snapshot_diff', full_name='caffe.SolverParameter.snapshot_diff', index=28, number=16, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='snapshot_format', full_name='caffe.SolverParameter.snapshot_format', index=29, number=37, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='solver_mode', full_name='caffe.SolverParameter.solver_mode', index=30, number=17, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='device_id', full_name='caffe.SolverParameter.device_id', index=31, number=18, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='random_seed', full_name='caffe.SolverParameter.random_seed', index=32, number=20, type=3, cpp_type=2, label=1, has_default_value=True, default_value=-1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='type', full_name='caffe.SolverParameter.type', index=33, number=40, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("SGD").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='delta', full_name='caffe.SolverParameter.delta', index=34, number=31, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1e-08), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='momentum2', full_name='caffe.SolverParameter.momentum2', index=35, number=39, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.999), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='rms_decay', full_name='caffe.SolverParameter.rms_decay', index=36, number=38, type=2, cpp_type=6, label=1, has_default_value=False, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='debug_info', full_name='caffe.SolverParameter.debug_info', index=37, number=23, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='snapshot_after_train', full_name='caffe.SolverParameter.snapshot_after_train', index=38, number=28, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='solver_type', full_name='caffe.SolverParameter.solver_type', index=39, number=30, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _SOLVERPARAMETER_SNAPSHOTFORMAT, _SOLVERPARAMETER_SOLVERMODE, _SOLVERPARAMETER_SOLVERTYPE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=986, serialized_end=2294, ) _SOLVERSTATE = _descriptor.Descriptor( name='SolverState', full_name='caffe.SolverState', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='iter', full_name='caffe.SolverState.iter', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='learned_net', full_name='caffe.SolverState.learned_net', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='history', full_name='caffe.SolverState.history', index=2, number=3, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='current_step', full_name='caffe.SolverState.current_step', index=3, number=4, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=2296, serialized_end=2404, ) _NETSTATE = _descriptor.Descriptor( name='NetState', full_name='caffe.NetState', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='phase', full_name='caffe.NetState.phase', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='level', full_name='caffe.NetState.level', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stage', full_name='caffe.NetState.stage', index=2, number=3, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=2406, serialized_end=2484, ) _NETSTATERULE = _descriptor.Descriptor( name='NetStateRule', full_name='caffe.NetStateRule', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='phase', full_name='caffe.NetStateRule.phase', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='min_level', full_name='caffe.NetStateRule.min_level', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='max_level', full_name='caffe.NetStateRule.max_level', index=2, number=3, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stage', full_name='caffe.NetStateRule.stage', index=3, number=4, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='not_stage', full_name='caffe.NetStateRule.not_stage', index=4, number=5, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=2486, serialized_end=2601, ) _PARAMSPEC = _descriptor.Descriptor( name='ParamSpec', full_name='caffe.ParamSpec', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='name', full_name='caffe.ParamSpec.name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='share_mode', full_name='caffe.ParamSpec.share_mode', index=1, number=2, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='lr_mult', full_name='caffe.ParamSpec.lr_mult', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='decay_mult', full_name='caffe.ParamSpec.decay_mult', index=3, number=4, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _PARAMSPEC_DIMCHECKMODE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=2604, serialized_end=2767, ) _LAYERPARAMETER = _descriptor.Descriptor( name='LayerParameter', full_name='caffe.LayerParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='name', full_name='caffe.LayerParameter.name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='type', full_name='caffe.LayerParameter.type', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bottom', full_name='caffe.LayerParameter.bottom', index=2, number=3, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='top', full_name='caffe.LayerParameter.top', index=3, number=4, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='phase', full_name='caffe.LayerParameter.phase', index=4, number=10, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='loss_weight', full_name='caffe.LayerParameter.loss_weight', index=5, number=5, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='param', full_name='caffe.LayerParameter.param', index=6, number=6, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blobs', full_name='caffe.LayerParameter.blobs', index=7, number=7, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='propagate_down', full_name='caffe.LayerParameter.propagate_down', index=8, number=11, type=8, cpp_type=7, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='include', full_name='caffe.LayerParameter.include', index=9, number=8, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='exclude', full_name='caffe.LayerParameter.exclude', index=10, number=9, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='transform_param', full_name='caffe.LayerParameter.transform_param', index=11, number=100, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='loss_param', full_name='caffe.LayerParameter.loss_param', index=12, number=101, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='accuracy_param', full_name='caffe.LayerParameter.accuracy_param', index=13, number=102, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='argmax_param', full_name='caffe.LayerParameter.argmax_param', index=14, number=103, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batch_norm_param', full_name='caffe.LayerParameter.batch_norm_param', index=15, number=139, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_param', full_name='caffe.LayerParameter.bias_param', index=16, number=141, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='concat_param', full_name='caffe.LayerParameter.concat_param', index=17, number=104, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='contrastive_loss_param', full_name='caffe.LayerParameter.contrastive_loss_param', index=18, number=105, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='convolution_param', full_name='caffe.LayerParameter.convolution_param', index=19, number=106, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='data_param', full_name='caffe.LayerParameter.data_param', index=20, number=107, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dropout_param', full_name='caffe.LayerParameter.dropout_param', index=21, number=108, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dummy_data_param', full_name='caffe.LayerParameter.dummy_data_param', index=22, number=109, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='eltwise_param', full_name='caffe.LayerParameter.eltwise_param', index=23, number=110, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='elu_param', full_name='caffe.LayerParameter.elu_param', index=24, number=140, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='embed_param', full_name='caffe.LayerParameter.embed_param', index=25, number=137, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='exp_param', full_name='caffe.LayerParameter.exp_param', index=26, number=111, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='flatten_param', full_name='caffe.LayerParameter.flatten_param', index=27, number=135, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hdf5_data_param', full_name='caffe.LayerParameter.hdf5_data_param', index=28, number=112, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hdf5_output_param', full_name='caffe.LayerParameter.hdf5_output_param', index=29, number=113, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hinge_loss_param', full_name='caffe.LayerParameter.hinge_loss_param', index=30, number=114, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='image_data_param', full_name='caffe.LayerParameter.image_data_param', index=31, number=115, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='infogain_loss_param', full_name='caffe.LayerParameter.infogain_loss_param', index=32, number=116, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='inner_product_param', full_name='caffe.LayerParameter.inner_product_param', index=33, number=117, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='log_param', full_name='caffe.LayerParameter.log_param', index=34, number=134, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='lrn_param', full_name='caffe.LayerParameter.lrn_param', index=35, number=118, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='memory_data_param', full_name='caffe.LayerParameter.memory_data_param', index=36, number=119, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mvn_param', full_name='caffe.LayerParameter.mvn_param', index=37, number=120, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pooling_param', full_name='caffe.LayerParameter.pooling_param', index=38, number=121, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='power_param', full_name='caffe.LayerParameter.power_param', index=39, number=122, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='prelu_param', full_name='caffe.LayerParameter.prelu_param', index=40, number=131, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='python_param', full_name='caffe.LayerParameter.python_param', index=41, number=130, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='reduction_param', full_name='caffe.LayerParameter.reduction_param', index=42, number=136, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='relu_param', full_name='caffe.LayerParameter.relu_param', index=43, number=123, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='reshape_param', full_name='caffe.LayerParameter.reshape_param', index=44, number=133, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale_param', full_name='caffe.LayerParameter.scale_param', index=45, number=142, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='sigmoid_param', full_name='caffe.LayerParameter.sigmoid_param', index=46, number=124, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='softmax_param', full_name='caffe.LayerParameter.softmax_param', index=47, number=125, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='spp_param', full_name='caffe.LayerParameter.spp_param', index=48, number=132, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='slice_param', full_name='caffe.LayerParameter.slice_param', index=49, number=126, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='tanh_param', full_name='caffe.LayerParameter.tanh_param', index=50, number=127, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='threshold_param', full_name='caffe.LayerParameter.threshold_param', index=51, number=128, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='tile_param', full_name='caffe.LayerParameter.tile_param', index=52, number=138, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='window_data_param', full_name='caffe.LayerParameter.window_data_param', index=53, number=129, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=2770, serialized_end=5138, ) _TRANSFORMATIONPARAMETER = _descriptor.Descriptor( name='TransformationParameter', full_name='caffe.TransformationParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='scale', full_name='caffe.TransformationParameter.scale', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mirror', full_name='caffe.TransformationParameter.mirror', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='crop_size', full_name='caffe.TransformationParameter.crop_size', index=2, number=3, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean_file', full_name='caffe.TransformationParameter.mean_file', index=3, number=4, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean_value', full_name='caffe.TransformationParameter.mean_value', index=4, number=5, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='force_color', full_name='caffe.TransformationParameter.force_color', index=5, number=6, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='force_gray', full_name='caffe.TransformationParameter.force_gray', index=6, number=7, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5141, serialized_end=5323, ) _LOSSPARAMETER = _descriptor.Descriptor( name='LossParameter', full_name='caffe.LossParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='ignore_label', full_name='caffe.LossParameter.ignore_label', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='normalization', full_name='caffe.LossParameter.normalization', index=1, number=3, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='normalize', full_name='caffe.LossParameter.normalize', index=2, number=2, type=8, cpp_type=7, label=1, has_default_value=False, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _LOSSPARAMETER_NORMALIZATIONMODE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5326, serialized_end=5520, ) _ACCURACYPARAMETER = _descriptor.Descriptor( name='AccuracyParameter', full_name='caffe.AccuracyParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='top_k', full_name='caffe.AccuracyParameter.top_k', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.AccuracyParameter.axis', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='ignore_label', full_name='caffe.AccuracyParameter.ignore_label', index=2, number=3, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5522, serialized_end=5598, ) _ARGMAXPARAMETER = _descriptor.Descriptor( name='ArgMaxParameter', full_name='caffe.ArgMaxParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='out_max_val', full_name='caffe.ArgMaxParameter.out_max_val', index=0, number=1, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='top_k', full_name='caffe.ArgMaxParameter.top_k', index=1, number=2, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.ArgMaxParameter.axis', index=2, number=3, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5600, serialized_end=5677, ) _CONCATPARAMETER = _descriptor.Descriptor( name='ConcatParameter', full_name='caffe.ConcatParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.ConcatParameter.axis', index=0, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='concat_dim', full_name='caffe.ConcatParameter.concat_dim', index=1, number=1, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5679, serialized_end=5736, ) _BATCHNORMPARAMETER = _descriptor.Descriptor( name='BatchNormParameter', full_name='caffe.BatchNormParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='use_global_stats', full_name='caffe.BatchNormParameter.use_global_stats', index=0, number=1, type=8, cpp_type=7, label=1, has_default_value=False, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='moving_average_fraction', full_name='caffe.BatchNormParameter.moving_average_fraction', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.999), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='eps', full_name='caffe.BatchNormParameter.eps', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1e-05), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5738, serialized_end=5844, ) _BIASPARAMETER = _descriptor.Descriptor( name='BiasParameter', full_name='caffe.BiasParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.BiasParameter.axis', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='num_axes', full_name='caffe.BiasParameter.num_axes', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='filler', full_name='caffe.BiasParameter.filler', index=2, number=3, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5846, serialized_end=5939, ) _CONTRASTIVELOSSPARAMETER = _descriptor.Descriptor( name='ContrastiveLossParameter', full_name='caffe.ContrastiveLossParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='margin', full_name='caffe.ContrastiveLossParameter.margin', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='legacy_version', full_name='caffe.ContrastiveLossParameter.legacy_version', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=5941, serialized_end=6017, ) _CONVOLUTIONPARAMETER = _descriptor.Descriptor( name='ConvolutionParameter', full_name='caffe.ConvolutionParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='num_output', full_name='caffe.ConvolutionParameter.num_output', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_term', full_name='caffe.ConvolutionParameter.bias_term', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad', full_name='caffe.ConvolutionParameter.pad', index=2, number=3, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_size', full_name='caffe.ConvolutionParameter.kernel_size', index=3, number=4, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride', full_name='caffe.ConvolutionParameter.stride', index=4, number=6, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dilation', full_name='caffe.ConvolutionParameter.dilation', index=5, number=18, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad_h', full_name='caffe.ConvolutionParameter.pad_h', index=6, number=9, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad_w', full_name='caffe.ConvolutionParameter.pad_w', index=7, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_h', full_name='caffe.ConvolutionParameter.kernel_h', index=8, number=11, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_w', full_name='caffe.ConvolutionParameter.kernel_w', index=9, number=12, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride_h', full_name='caffe.ConvolutionParameter.stride_h', index=10, number=13, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride_w', full_name='caffe.ConvolutionParameter.stride_w', index=11, number=14, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='group', full_name='caffe.ConvolutionParameter.group', index=12, number=5, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_filler', full_name='caffe.ConvolutionParameter.weight_filler', index=13, number=7, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_filler', full_name='caffe.ConvolutionParameter.bias_filler', index=14, number=8, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='engine', full_name='caffe.ConvolutionParameter.engine', index=15, number=15, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.ConvolutionParameter.axis', index=16, number=16, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='force_nd_im2col', full_name='caffe.ConvolutionParameter.force_nd_im2col', index=17, number=17, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _CONVOLUTIONPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=6020, serialized_end=6528, ) _DATAPARAMETER = _descriptor.Descriptor( name='DataParameter', full_name='caffe.DataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='source', full_name='caffe.DataParameter.source', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batch_size', full_name='caffe.DataParameter.batch_size', index=1, number=4, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='rand_skip', full_name='caffe.DataParameter.rand_skip', index=2, number=7, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='backend', full_name='caffe.DataParameter.backend', index=3, number=8, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.DataParameter.scale', index=4, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean_file', full_name='caffe.DataParameter.mean_file', index=5, number=3, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='crop_size', full_name='caffe.DataParameter.crop_size', index=6, number=5, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mirror', full_name='caffe.DataParameter.mirror', index=7, number=6, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='force_encoded_color', full_name='caffe.DataParameter.force_encoded_color', index=8, number=9, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='prefetch', full_name='caffe.DataParameter.prefetch', index=9, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=4, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _DATAPARAMETER_DB, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=6531, serialized_end=6823, ) _DROPOUTPARAMETER = _descriptor.Descriptor( name='DropoutParameter', full_name='caffe.DropoutParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='dropout_ratio', full_name='caffe.DropoutParameter.dropout_ratio', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=6825, serialized_end=6871, ) _DUMMYDATAPARAMETER = _descriptor.Descriptor( name='DummyDataParameter', full_name='caffe.DummyDataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='data_filler', full_name='caffe.DummyDataParameter.data_filler', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shape', full_name='caffe.DummyDataParameter.shape', index=1, number=6, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='num', full_name='caffe.DummyDataParameter.num', index=2, number=2, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='channels', full_name='caffe.DummyDataParameter.channels', index=3, number=3, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='height', full_name='caffe.DummyDataParameter.height', index=4, number=4, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='width', full_name='caffe.DummyDataParameter.width', index=5, number=5, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=6874, serialized_end=7034, ) _ELTWISEPARAMETER = _descriptor.Descriptor( name='EltwiseParameter', full_name='caffe.EltwiseParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='operation', full_name='caffe.EltwiseParameter.operation', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='coeff', full_name='caffe.EltwiseParameter.coeff', index=1, number=2, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stable_prod_grad', full_name='caffe.EltwiseParameter.stable_prod_grad', index=2, number=3, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _ELTWISEPARAMETER_ELTWISEOP, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7037, serialized_end=7202, ) _ELUPARAMETER = _descriptor.Descriptor( name='ELUParameter', full_name='caffe.ELUParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='alpha', full_name='caffe.ELUParameter.alpha', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7204, serialized_end=7236, ) _EMBEDPARAMETER = _descriptor.Descriptor( name='EmbedParameter', full_name='caffe.EmbedParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='num_output', full_name='caffe.EmbedParameter.num_output', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='input_dim', full_name='caffe.EmbedParameter.input_dim', index=1, number=2, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_term', full_name='caffe.EmbedParameter.bias_term', index=2, number=3, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_filler', full_name='caffe.EmbedParameter.weight_filler', index=3, number=4, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_filler', full_name='caffe.EmbedParameter.bias_filler', index=4, number=5, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7239, serialized_end=7411, ) _EXPPARAMETER = _descriptor.Descriptor( name='ExpParameter', full_name='caffe.ExpParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='base', full_name='caffe.ExpParameter.base', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(-1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.ExpParameter.scale', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shift', full_name='caffe.ExpParameter.shift', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7413, serialized_end=7481, ) _FLATTENPARAMETER = _descriptor.Descriptor( name='FlattenParameter', full_name='caffe.FlattenParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.FlattenParameter.axis', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='end_axis', full_name='caffe.FlattenParameter.end_axis', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=-1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7483, serialized_end=7540, ) _HDF5DATAPARAMETER = _descriptor.Descriptor( name='HDF5DataParameter', full_name='caffe.HDF5DataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='source', full_name='caffe.HDF5DataParameter.source', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batch_size', full_name='caffe.HDF5DataParameter.batch_size', index=1, number=2, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shuffle', full_name='caffe.HDF5DataParameter.shuffle', index=2, number=3, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7542, serialized_end=7621, ) _HDF5OUTPUTPARAMETER = _descriptor.Descriptor( name='HDF5OutputParameter', full_name='caffe.HDF5OutputParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='file_name', full_name='caffe.HDF5OutputParameter.file_name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7623, serialized_end=7663, ) _HINGELOSSPARAMETER = _descriptor.Descriptor( name='HingeLossParameter', full_name='caffe.HingeLossParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='norm', full_name='caffe.HingeLossParameter.norm', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _HINGELOSSPARAMETER_NORM, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7665, serialized_end=7759, ) _IMAGEDATAPARAMETER = _descriptor.Descriptor( name='ImageDataParameter', full_name='caffe.ImageDataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='source', full_name='caffe.ImageDataParameter.source', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batch_size', full_name='caffe.ImageDataParameter.batch_size', index=1, number=4, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='rand_skip', full_name='caffe.ImageDataParameter.rand_skip', index=2, number=7, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shuffle', full_name='caffe.ImageDataParameter.shuffle', index=3, number=8, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_height', full_name='caffe.ImageDataParameter.new_height', index=4, number=9, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_width', full_name='caffe.ImageDataParameter.new_width', index=5, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='is_color', full_name='caffe.ImageDataParameter.is_color', index=6, number=11, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.ImageDataParameter.scale', index=7, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean_file', full_name='caffe.ImageDataParameter.mean_file', index=8, number=3, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='crop_size', full_name='caffe.ImageDataParameter.crop_size', index=9, number=5, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mirror', full_name='caffe.ImageDataParameter.mirror', index=10, number=6, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='root_folder', full_name='caffe.ImageDataParameter.root_folder', index=11, number=12, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=7762, serialized_end=8041, ) _INFOGAINLOSSPARAMETER = _descriptor.Descriptor( name='InfogainLossParameter', full_name='caffe.InfogainLossParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='source', full_name='caffe.InfogainLossParameter.source', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8043, serialized_end=8082, ) _INNERPRODUCTPARAMETER = _descriptor.Descriptor( name='InnerProductParameter', full_name='caffe.InnerProductParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='num_output', full_name='caffe.InnerProductParameter.num_output', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_term', full_name='caffe.InnerProductParameter.bias_term', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_filler', full_name='caffe.InnerProductParameter.weight_filler', index=2, number=3, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_filler', full_name='caffe.InnerProductParameter.bias_filler', index=3, number=4, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.InnerProductParameter.axis', index=4, number=5, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8085, serialized_end=8262, ) _LOGPARAMETER = _descriptor.Descriptor( name='LogParameter', full_name='caffe.LogParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='base', full_name='caffe.LogParameter.base', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(-1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.LogParameter.scale', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shift', full_name='caffe.LogParameter.shift', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8264, serialized_end=8332, ) _LRNPARAMETER = _descriptor.Descriptor( name='LRNParameter', full_name='caffe.LRNParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='local_size', full_name='caffe.LRNParameter.local_size', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=True, default_value=5, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='alpha', full_name='caffe.LRNParameter.alpha', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='beta', full_name='caffe.LRNParameter.beta', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.75), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='norm_region', full_name='caffe.LRNParameter.norm_region', index=3, number=4, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='k', full_name='caffe.LRNParameter.k', index=4, number=5, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='engine', full_name='caffe.LRNParameter.engine', index=5, number=6, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _LRNPARAMETER_NORMREGION, _LRNPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8335, serialized_end=8647, ) _MEMORYDATAPARAMETER = _descriptor.Descriptor( name='MemoryDataParameter', full_name='caffe.MemoryDataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='batch_size', full_name='caffe.MemoryDataParameter.batch_size', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='channels', full_name='caffe.MemoryDataParameter.channels', index=1, number=2, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='height', full_name='caffe.MemoryDataParameter.height', index=2, number=3, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='width', full_name='caffe.MemoryDataParameter.width', index=3, number=4, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8649, serialized_end=8739, ) _MVNPARAMETER = _descriptor.Descriptor( name='MVNParameter', full_name='caffe.MVNParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='normalize_variance', full_name='caffe.MVNParameter.normalize_variance', index=0, number=1, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='across_channels', full_name='caffe.MVNParameter.across_channels', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='eps', full_name='caffe.MVNParameter.eps', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1e-09), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8741, serialized_end=8841, ) _POOLINGPARAMETER = _descriptor.Descriptor( name='PoolingParameter', full_name='caffe.PoolingParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='pool', full_name='caffe.PoolingParameter.pool', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad', full_name='caffe.PoolingParameter.pad', index=1, number=4, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad_h', full_name='caffe.PoolingParameter.pad_h', index=2, number=9, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad_w', full_name='caffe.PoolingParameter.pad_w', index=3, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_size', full_name='caffe.PoolingParameter.kernel_size', index=4, number=2, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_h', full_name='caffe.PoolingParameter.kernel_h', index=5, number=5, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernel_w', full_name='caffe.PoolingParameter.kernel_w', index=6, number=6, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride', full_name='caffe.PoolingParameter.stride', index=7, number=3, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride_h', full_name='caffe.PoolingParameter.stride_h', index=8, number=7, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride_w', full_name='caffe.PoolingParameter.stride_w', index=9, number=8, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='engine', full_name='caffe.PoolingParameter.engine', index=10, number=11, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='global_pooling', full_name='caffe.PoolingParameter.global_pooling', index=11, number=12, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _POOLINGPARAMETER_POOLMETHOD, _POOLINGPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=8844, serialized_end=9262, ) _POWERPARAMETER = _descriptor.Descriptor( name='PowerParameter', full_name='caffe.PowerParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='power', full_name='caffe.PowerParameter.power', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.PowerParameter.scale', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shift', full_name='caffe.PowerParameter.shift', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9264, serialized_end=9334, ) _PYTHONPARAMETER = _descriptor.Descriptor( name='PythonParameter', full_name='caffe.PythonParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='module', full_name='caffe.PythonParameter.module', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='layer', full_name='caffe.PythonParameter.layer', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='param_str', full_name='caffe.PythonParameter.param_str', index=2, number=3, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='share_in_parallel', full_name='caffe.PythonParameter.share_in_parallel', index=3, number=4, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9336, serialized_end=9439, ) _REDUCTIONPARAMETER = _descriptor.Descriptor( name='ReductionParameter', full_name='caffe.ReductionParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='operation', full_name='caffe.ReductionParameter.operation', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.ReductionParameter.axis', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='coeff', full_name='caffe.ReductionParameter.coeff', index=2, number=3, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _REDUCTIONPARAMETER_REDUCTIONOP, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9442, serialized_end=9615, ) _RELUPARAMETER = _descriptor.Descriptor( name='ReLUParameter', full_name='caffe.ReLUParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='negative_slope', full_name='caffe.ReLUParameter.negative_slope', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='engine', full_name='caffe.ReLUParameter.engine', index=1, number=2, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _RELUPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9618, serialized_end=9759, ) _RESHAPEPARAMETER = _descriptor.Descriptor( name='ReshapeParameter', full_name='caffe.ReshapeParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='shape', full_name='caffe.ReshapeParameter.shape', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.ReshapeParameter.axis', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='num_axes', full_name='caffe.ReshapeParameter.num_axes', index=2, number=3, type=5, cpp_type=1, label=1, has_default_value=True, default_value=-1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9761, serialized_end=9851, ) _SCALEPARAMETER = _descriptor.Descriptor( name='ScaleParameter', full_name='caffe.ScaleParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.ScaleParameter.axis', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='num_axes', full_name='caffe.ScaleParameter.num_axes', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='filler', full_name='caffe.ScaleParameter.filler', index=2, number=3, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_term', full_name='caffe.ScaleParameter.bias_term', index=3, number=4, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_filler', full_name='caffe.ScaleParameter.bias_filler', index=4, number=5, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=9854, serialized_end=10019, ) _SIGMOIDPARAMETER = _descriptor.Descriptor( name='SigmoidParameter', full_name='caffe.SigmoidParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='engine', full_name='caffe.SigmoidParameter.engine', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _SIGMOIDPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10021, serialized_end=10141, ) _SLICEPARAMETER = _descriptor.Descriptor( name='SliceParameter', full_name='caffe.SliceParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.SliceParameter.axis', index=0, number=3, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='slice_point', full_name='caffe.SliceParameter.slice_point', index=1, number=2, type=13, cpp_type=3, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='slice_dim', full_name='caffe.SliceParameter.slice_dim', index=2, number=1, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10143, serialized_end=10219, ) _SOFTMAXPARAMETER = _descriptor.Descriptor( name='SoftmaxParameter', full_name='caffe.SoftmaxParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='engine', full_name='caffe.SoftmaxParameter.engine', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='axis', full_name='caffe.SoftmaxParameter.axis', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _SOFTMAXPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10222, serialized_end=10359, ) _TANHPARAMETER = _descriptor.Descriptor( name='TanHParameter', full_name='caffe.TanHParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='engine', full_name='caffe.TanHParameter.engine', index=0, number=1, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _TANHPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10361, serialized_end=10475, ) _TILEPARAMETER = _descriptor.Descriptor( name='TileParameter', full_name='caffe.TileParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='axis', full_name='caffe.TileParameter.axis', index=0, number=1, type=5, cpp_type=1, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='tiles', full_name='caffe.TileParameter.tiles', index=1, number=2, type=5, cpp_type=1, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10477, serialized_end=10524, ) _THRESHOLDPARAMETER = _descriptor.Descriptor( name='ThresholdParameter', full_name='caffe.ThresholdParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='threshold', full_name='caffe.ThresholdParameter.threshold', index=0, number=1, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10526, serialized_end=10568, ) _WINDOWDATAPARAMETER = _descriptor.Descriptor( name='WindowDataParameter', full_name='caffe.WindowDataParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='source', full_name='caffe.WindowDataParameter.source', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.WindowDataParameter.scale', index=1, number=2, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mean_file', full_name='caffe.WindowDataParameter.mean_file', index=2, number=3, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batch_size', full_name='caffe.WindowDataParameter.batch_size', index=3, number=4, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='crop_size', full_name='caffe.WindowDataParameter.crop_size', index=4, number=5, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mirror', full_name='caffe.WindowDataParameter.mirror', index=5, number=6, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='fg_threshold', full_name='caffe.WindowDataParameter.fg_threshold', index=6, number=7, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bg_threshold', full_name='caffe.WindowDataParameter.bg_threshold', index=7, number=8, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='fg_fraction', full_name='caffe.WindowDataParameter.fg_fraction', index=8, number=9, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.25), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='context_pad', full_name='caffe.WindowDataParameter.context_pad', index=9, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='crop_mode', full_name='caffe.WindowDataParameter.crop_mode', index=10, number=11, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("warp").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='cache_images', full_name='caffe.WindowDataParameter.cache_images', index=11, number=12, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='root_folder', full_name='caffe.WindowDataParameter.root_folder', index=12, number=13, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10571, serialized_end=10892, ) _SPPPARAMETER = _descriptor.Descriptor( name='SPPParameter', full_name='caffe.SPPParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='pyramid_height', full_name='caffe.SPPParameter.pyramid_height', index=0, number=1, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pool', full_name='caffe.SPPParameter.pool', index=1, number=2, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='engine', full_name='caffe.SPPParameter.engine', index=2, number=6, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _SPPPARAMETER_POOLMETHOD, _SPPPARAMETER_ENGINE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=10895, serialized_end=11130, ) _V1LAYERPARAMETER = _descriptor.Descriptor( name='V1LayerParameter', full_name='caffe.V1LayerParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='bottom', full_name='caffe.V1LayerParameter.bottom', index=0, number=2, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='top', full_name='caffe.V1LayerParameter.top', index=1, number=3, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='name', full_name='caffe.V1LayerParameter.name', index=2, number=4, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='include', full_name='caffe.V1LayerParameter.include', index=3, number=32, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='exclude', full_name='caffe.V1LayerParameter.exclude', index=4, number=33, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='type', full_name='caffe.V1LayerParameter.type', index=5, number=5, type=14, cpp_type=8, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blobs', full_name='caffe.V1LayerParameter.blobs', index=6, number=6, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='param', full_name='caffe.V1LayerParameter.param', index=7, number=1001, type=9, cpp_type=9, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blob_share_mode', full_name='caffe.V1LayerParameter.blob_share_mode', index=8, number=1002, type=14, cpp_type=8, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blobs_lr', full_name='caffe.V1LayerParameter.blobs_lr', index=9, number=7, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_decay', full_name='caffe.V1LayerParameter.weight_decay', index=10, number=8, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='loss_weight', full_name='caffe.V1LayerParameter.loss_weight', index=11, number=35, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='accuracy_param', full_name='caffe.V1LayerParameter.accuracy_param', index=12, number=27, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='argmax_param', full_name='caffe.V1LayerParameter.argmax_param', index=13, number=23, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='concat_param', full_name='caffe.V1LayerParameter.concat_param', index=14, number=9, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='contrastive_loss_param', full_name='caffe.V1LayerParameter.contrastive_loss_param', index=15, number=40, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='convolution_param', full_name='caffe.V1LayerParameter.convolution_param', index=16, number=10, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='data_param', full_name='caffe.V1LayerParameter.data_param', index=17, number=11, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dropout_param', full_name='caffe.V1LayerParameter.dropout_param', index=18, number=12, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dummy_data_param', full_name='caffe.V1LayerParameter.dummy_data_param', index=19, number=26, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='eltwise_param', full_name='caffe.V1LayerParameter.eltwise_param', index=20, number=24, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='exp_param', full_name='caffe.V1LayerParameter.exp_param', index=21, number=41, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hdf5_data_param', full_name='caffe.V1LayerParameter.hdf5_data_param', index=22, number=13, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hdf5_output_param', full_name='caffe.V1LayerParameter.hdf5_output_param', index=23, number=14, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hinge_loss_param', full_name='caffe.V1LayerParameter.hinge_loss_param', index=24, number=29, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='image_data_param', full_name='caffe.V1LayerParameter.image_data_param', index=25, number=15, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='infogain_loss_param', full_name='caffe.V1LayerParameter.infogain_loss_param', index=26, number=16, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='inner_product_param', full_name='caffe.V1LayerParameter.inner_product_param', index=27, number=17, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='lrn_param', full_name='caffe.V1LayerParameter.lrn_param', index=28, number=18, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='memory_data_param', full_name='caffe.V1LayerParameter.memory_data_param', index=29, number=22, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mvn_param', full_name='caffe.V1LayerParameter.mvn_param', index=30, number=34, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pooling_param', full_name='caffe.V1LayerParameter.pooling_param', index=31, number=19, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='power_param', full_name='caffe.V1LayerParameter.power_param', index=32, number=21, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='relu_param', full_name='caffe.V1LayerParameter.relu_param', index=33, number=30, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='sigmoid_param', full_name='caffe.V1LayerParameter.sigmoid_param', index=34, number=38, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='softmax_param', full_name='caffe.V1LayerParameter.softmax_param', index=35, number=39, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='slice_param', full_name='caffe.V1LayerParameter.slice_param', index=36, number=31, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='tanh_param', full_name='caffe.V1LayerParameter.tanh_param', index=37, number=37, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='threshold_param', full_name='caffe.V1LayerParameter.threshold_param', index=38, number=25, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='window_data_param', full_name='caffe.V1LayerParameter.window_data_param', index=39, number=20, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='transform_param', full_name='caffe.V1LayerParameter.transform_param', index=40, number=36, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='loss_param', full_name='caffe.V1LayerParameter.loss_param', index=41, number=42, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='layer', full_name='caffe.V1LayerParameter.layer', index=42, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _V1LAYERPARAMETER_LAYERTYPE, _V1LAYERPARAMETER_DIMCHECKMODE, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=11133, serialized_end=13661, ) _V0LAYERPARAMETER = _descriptor.Descriptor( name='V0LayerParameter', full_name='caffe.V0LayerParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='name', full_name='caffe.V0LayerParameter.name', index=0, number=1, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='type', full_name='caffe.V0LayerParameter.type', index=1, number=2, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='num_output', full_name='caffe.V0LayerParameter.num_output', index=2, number=3, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='biasterm', full_name='caffe.V0LayerParameter.biasterm', index=3, number=4, type=8, cpp_type=7, label=1, has_default_value=True, default_value=True, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_filler', full_name='caffe.V0LayerParameter.weight_filler', index=4, number=5, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='bias_filler', full_name='caffe.V0LayerParameter.bias_filler', index=5, number=6, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pad', full_name='caffe.V0LayerParameter.pad', index=6, number=7, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='kernelsize', full_name='caffe.V0LayerParameter.kernelsize', index=7, number=8, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='group', full_name='caffe.V0LayerParameter.group', index=8, number=9, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='stride', full_name='caffe.V0LayerParameter.stride', index=9, number=10, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='pool', full_name='caffe.V0LayerParameter.pool', index=10, number=11, type=14, cpp_type=8, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='dropout_ratio', full_name='caffe.V0LayerParameter.dropout_ratio', index=11, number=12, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='local_size', full_name='caffe.V0LayerParameter.local_size', index=12, number=13, type=13, cpp_type=3, label=1, has_default_value=True, default_value=5, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='alpha', full_name='caffe.V0LayerParameter.alpha', index=13, number=14, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='beta', full_name='caffe.V0LayerParameter.beta', index=14, number=15, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.75), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='k', full_name='caffe.V0LayerParameter.k', index=15, number=22, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='source', full_name='caffe.V0LayerParameter.source', index=16, number=16, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='scale', full_name='caffe.V0LayerParameter.scale', index=17, number=17, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(1), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='meanfile', full_name='caffe.V0LayerParameter.meanfile', index=18, number=18, type=9, cpp_type=9, label=1, has_default_value=False, default_value=_b("").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='batchsize', full_name='caffe.V0LayerParameter.batchsize', index=19, number=19, type=13, cpp_type=3, label=1, has_default_value=False, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='cropsize', full_name='caffe.V0LayerParameter.cropsize', index=20, number=20, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='mirror', full_name='caffe.V0LayerParameter.mirror', index=21, number=21, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blobs', full_name='caffe.V0LayerParameter.blobs', index=22, number=50, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='blobs_lr', full_name='caffe.V0LayerParameter.blobs_lr', index=23, number=51, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='weight_decay', full_name='caffe.V0LayerParameter.weight_decay', index=24, number=52, type=2, cpp_type=6, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='rand_skip', full_name='caffe.V0LayerParameter.rand_skip', index=25, number=53, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='det_fg_threshold', full_name='caffe.V0LayerParameter.det_fg_threshold', index=26, number=54, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='det_bg_threshold', full_name='caffe.V0LayerParameter.det_bg_threshold', index=27, number=55, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.5), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='det_fg_fraction', full_name='caffe.V0LayerParameter.det_fg_fraction', index=28, number=56, type=2, cpp_type=6, label=1, has_default_value=True, default_value=float(0.25), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='det_context_pad', full_name='caffe.V0LayerParameter.det_context_pad', index=29, number=58, type=13, cpp_type=3, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='det_crop_mode', full_name='caffe.V0LayerParameter.det_crop_mode', index=30, number=59, type=9, cpp_type=9, label=1, has_default_value=True, default_value=_b("warp").decode('utf-8'), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_num', full_name='caffe.V0LayerParameter.new_num', index=31, number=60, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_channels', full_name='caffe.V0LayerParameter.new_channels', index=32, number=61, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_height', full_name='caffe.V0LayerParameter.new_height', index=33, number=62, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='new_width', full_name='caffe.V0LayerParameter.new_width', index=34, number=63, type=5, cpp_type=1, label=1, has_default_value=True, default_value=0, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='shuffle_images', full_name='caffe.V0LayerParameter.shuffle_images', index=35, number=64, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='concat_dim', full_name='caffe.V0LayerParameter.concat_dim', index=36, number=65, type=13, cpp_type=3, label=1, has_default_value=True, default_value=1, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='hdf5_output_param', full_name='caffe.V0LayerParameter.hdf5_output_param', index=37, number=1001, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ _V0LAYERPARAMETER_POOLMETHOD, ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=13664, serialized_end=14685, ) _PRELUPARAMETER = _descriptor.Descriptor( name='PReLUParameter', full_name='caffe.PReLUParameter', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='filler', full_name='caffe.PReLUParameter.filler', index=0, number=1, type=11, cpp_type=10, label=1, has_default_value=False, default_value=None, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), _descriptor.FieldDescriptor( name='channel_shared', full_name='caffe.PReLUParameter.channel_shared', index=1, number=2, type=8, cpp_type=7, label=1, has_default_value=True, default_value=False, message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, options=None), ], extensions=[ ], nested_types=[], enum_types=[ ], options=None, is_extendable=False, syntax='proto2', extension_ranges=[], oneofs=[ ], serialized_start=14687, serialized_end=14774, ) _BLOBPROTO.fields_by_name['shape'].message_type = _BLOBSHAPE _BLOBPROTOVECTOR.fields_by_name['blobs'].message_type = _BLOBPROTO _FILLERPARAMETER.fields_by_name['variance_norm'].enum_type = _FILLERPARAMETER_VARIANCENORM _FILLERPARAMETER_VARIANCENORM.containing_type = _FILLERPARAMETER _NETPARAMETER.fields_by_name['input_shape'].message_type = _BLOBSHAPE _NETPARAMETER.fields_by_name['state'].message_type = _NETSTATE _NETPARAMETER.fields_by_name['layer'].message_type = _LAYERPARAMETER _NETPARAMETER.fields_by_name['layers'].message_type = _V1LAYERPARAMETER _SOLVERPARAMETER.fields_by_name['net_param'].message_type = _NETPARAMETER _SOLVERPARAMETER.fields_by_name['train_net_param'].message_type = _NETPARAMETER _SOLVERPARAMETER.fields_by_name['test_net_param'].message_type = _NETPARAMETER _SOLVERPARAMETER.fields_by_name['train_state'].message_type = _NETSTATE _SOLVERPARAMETER.fields_by_name['test_state'].message_type = _NETSTATE _SOLVERPARAMETER.fields_by_name['snapshot_format'].enum_type = _SOLVERPARAMETER_SNAPSHOTFORMAT _SOLVERPARAMETER.fields_by_name['solver_mode'].enum_type = _SOLVERPARAMETER_SOLVERMODE _SOLVERPARAMETER.fields_by_name['solver_type'].enum_type = _SOLVERPARAMETER_SOLVERTYPE _SOLVERPARAMETER_SNAPSHOTFORMAT.containing_type = _SOLVERPARAMETER _SOLVERPARAMETER_SOLVERMODE.containing_type = _SOLVERPARAMETER _SOLVERPARAMETER_SOLVERTYPE.containing_type = _SOLVERPARAMETER _SOLVERSTATE.fields_by_name['history'].message_type = _BLOBPROTO _NETSTATE.fields_by_name['phase'].enum_type = _PHASE _NETSTATERULE.fields_by_name['phase'].enum_type = _PHASE _PARAMSPEC.fields_by_name['share_mode'].enum_type = _PARAMSPEC_DIMCHECKMODE _PARAMSPEC_DIMCHECKMODE.containing_type = _PARAMSPEC _LAYERPARAMETER.fields_by_name['phase'].enum_type = _PHASE _LAYERPARAMETER.fields_by_name['param'].message_type = _PARAMSPEC _LAYERPARAMETER.fields_by_name['blobs'].message_type = _BLOBPROTO _LAYERPARAMETER.fields_by_name['include'].message_type = _NETSTATERULE _LAYERPARAMETER.fields_by_name['exclude'].message_type = _NETSTATERULE _LAYERPARAMETER.fields_by_name['transform_param'].message_type = _TRANSFORMATIONPARAMETER _LAYERPARAMETER.fields_by_name['loss_param'].message_type = _LOSSPARAMETER _LAYERPARAMETER.fields_by_name['accuracy_param'].message_type = _ACCURACYPARAMETER _LAYERPARAMETER.fields_by_name['argmax_param'].message_type = _ARGMAXPARAMETER _LAYERPARAMETER.fields_by_name['batch_norm_param'].message_type = _BATCHNORMPARAMETER _LAYERPARAMETER.fields_by_name['bias_param'].message_type = _BIASPARAMETER _LAYERPARAMETER.fields_by_name['concat_param'].message_type = _CONCATPARAMETER _LAYERPARAMETER.fields_by_name['contrastive_loss_param'].message_type = _CONTRASTIVELOSSPARAMETER _LAYERPARAMETER.fields_by_name['convolution_param'].message_type = _CONVOLUTIONPARAMETER _LAYERPARAMETER.fields_by_name['data_param'].message_type = _DATAPARAMETER _LAYERPARAMETER.fields_by_name['dropout_param'].message_type = _DROPOUTPARAMETER _LAYERPARAMETER.fields_by_name['dummy_data_param'].message_type = _DUMMYDATAPARAMETER _LAYERPARAMETER.fields_by_name['eltwise_param'].message_type = _ELTWISEPARAMETER _LAYERPARAMETER.fields_by_name['elu_param'].message_type = _ELUPARAMETER _LAYERPARAMETER.fields_by_name['embed_param'].message_type = _EMBEDPARAMETER _LAYERPARAMETER.fields_by_name['exp_param'].message_type = _EXPPARAMETER _LAYERPARAMETER.fields_by_name['flatten_param'].message_type = _FLATTENPARAMETER _LAYERPARAMETER.fields_by_name['hdf5_data_param'].message_type = _HDF5DATAPARAMETER _LAYERPARAMETER.fields_by_name['hdf5_output_param'].message_type = _HDF5OUTPUTPARAMETER _LAYERPARAMETER.fields_by_name['hinge_loss_param'].message_type = _HINGELOSSPARAMETER _LAYERPARAMETER.fields_by_name['image_data_param'].message_type = _IMAGEDATAPARAMETER _LAYERPARAMETER.fields_by_name['infogain_loss_param'].message_type = _INFOGAINLOSSPARAMETER _LAYERPARAMETER.fields_by_name['inner_product_param'].message_type = _INNERPRODUCTPARAMETER _LAYERPARAMETER.fields_by_name['log_param'].message_type = _LOGPARAMETER _LAYERPARAMETER.fields_by_name['lrn_param'].message_type = _LRNPARAMETER _LAYERPARAMETER.fields_by_name['memory_data_param'].message_type = _MEMORYDATAPARAMETER _LAYERPARAMETER.fields_by_name['mvn_param'].message_type = _MVNPARAMETER _LAYERPARAMETER.fields_by_name['pooling_param'].message_type = _POOLINGPARAMETER _LAYERPARAMETER.fields_by_name['power_param'].message_type = _POWERPARAMETER _LAYERPARAMETER.fields_by_name['prelu_param'].message_type = _PRELUPARAMETER _LAYERPARAMETER.fields_by_name['python_param'].message_type = _PYTHONPARAMETER _LAYERPARAMETER.fields_by_name['reduction_param'].message_type = _REDUCTIONPARAMETER _LAYERPARAMETER.fields_by_name['relu_param'].message_type = _RELUPARAMETER _LAYERPARAMETER.fields_by_name['reshape_param'].message_type = _RESHAPEPARAMETER _LAYERPARAMETER.fields_by_name['scale_param'].message_type = _SCALEPARAMETER _LAYERPARAMETER.fields_by_name['sigmoid_param'].message_type = _SIGMOIDPARAMETER _LAYERPARAMETER.fields_by_name['softmax_param'].message_type = _SOFTMAXPARAMETER _LAYERPARAMETER.fields_by_name['spp_param'].message_type = _SPPPARAMETER _LAYERPARAMETER.fields_by_name['slice_param'].message_type = _SLICEPARAMETER _LAYERPARAMETER.fields_by_name['tanh_param'].message_type = _TANHPARAMETER _LAYERPARAMETER.fields_by_name['threshold_param'].message_type = _THRESHOLDPARAMETER _LAYERPARAMETER.fields_by_name['tile_param'].message_type = _TILEPARAMETER _LAYERPARAMETER.fields_by_name['window_data_param'].message_type = _WINDOWDATAPARAMETER _LOSSPARAMETER.fields_by_name['normalization'].enum_type = _LOSSPARAMETER_NORMALIZATIONMODE _LOSSPARAMETER_NORMALIZATIONMODE.containing_type = _LOSSPARAMETER _BIASPARAMETER.fields_by_name['filler'].message_type = _FILLERPARAMETER _CONVOLUTIONPARAMETER.fields_by_name['weight_filler'].message_type = _FILLERPARAMETER _CONVOLUTIONPARAMETER.fields_by_name['bias_filler'].message_type = _FILLERPARAMETER _CONVOLUTIONPARAMETER.fields_by_name['engine'].enum_type = _CONVOLUTIONPARAMETER_ENGINE _CONVOLUTIONPARAMETER_ENGINE.containing_type = _CONVOLUTIONPARAMETER _DATAPARAMETER.fields_by_name['backend'].enum_type = _DATAPARAMETER_DB _DATAPARAMETER_DB.containing_type = _DATAPARAMETER _DUMMYDATAPARAMETER.fields_by_name['data_filler'].message_type = _FILLERPARAMETER _DUMMYDATAPARAMETER.fields_by_name['shape'].message_type = _BLOBSHAPE _ELTWISEPARAMETER.fields_by_name['operation'].enum_type = _ELTWISEPARAMETER_ELTWISEOP _ELTWISEPARAMETER_ELTWISEOP.containing_type = _ELTWISEPARAMETER _EMBEDPARAMETER.fields_by_name['weight_filler'].message_type = _FILLERPARAMETER _EMBEDPARAMETER.fields_by_name['bias_filler'].message_type = _FILLERPARAMETER _HINGELOSSPARAMETER.fields_by_name['norm'].enum_type = _HINGELOSSPARAMETER_NORM _HINGELOSSPARAMETER_NORM.containing_type = _HINGELOSSPARAMETER _INNERPRODUCTPARAMETER.fields_by_name['weight_filler'].message_type = _FILLERPARAMETER _INNERPRODUCTPARAMETER.fields_by_name['bias_filler'].message_type = _FILLERPARAMETER _LRNPARAMETER.fields_by_name['norm_region'].enum_type = _LRNPARAMETER_NORMREGION _LRNPARAMETER.fields_by_name['engine'].enum_type = _LRNPARAMETER_ENGINE _LRNPARAMETER_NORMREGION.containing_type = _LRNPARAMETER _LRNPARAMETER_ENGINE.containing_type = _LRNPARAMETER _POOLINGPARAMETER.fields_by_name['pool'].enum_type = _POOLINGPARAMETER_POOLMETHOD _POOLINGPARAMETER.fields_by_name['engine'].enum_type = _POOLINGPARAMETER_ENGINE _POOLINGPARAMETER_POOLMETHOD.containing_type = _POOLINGPARAMETER _POOLINGPARAMETER_ENGINE.containing_type = _POOLINGPARAMETER _REDUCTIONPARAMETER.fields_by_name['operation'].enum_type = _REDUCTIONPARAMETER_REDUCTIONOP _REDUCTIONPARAMETER_REDUCTIONOP.containing_type = _REDUCTIONPARAMETER _RELUPARAMETER.fields_by_name['engine'].enum_type = _RELUPARAMETER_ENGINE _RELUPARAMETER_ENGINE.containing_type = _RELUPARAMETER _RESHAPEPARAMETER.fields_by_name['shape'].message_type = _BLOBSHAPE _SCALEPARAMETER.fields_by_name['filler'].message_type = _FILLERPARAMETER _SCALEPARAMETER.fields_by_name['bias_filler'].message_type = _FILLERPARAMETER _SIGMOIDPARAMETER.fields_by_name['engine'].enum_type = _SIGMOIDPARAMETER_ENGINE _SIGMOIDPARAMETER_ENGINE.containing_type = _SIGMOIDPARAMETER _SOFTMAXPARAMETER.fields_by_name['engine'].enum_type = _SOFTMAXPARAMETER_ENGINE _SOFTMAXPARAMETER_ENGINE.containing_type = _SOFTMAXPARAMETER _TANHPARAMETER.fields_by_name['engine'].enum_type = _TANHPARAMETER_ENGINE _TANHPARAMETER_ENGINE.containing_type = _TANHPARAMETER _SPPPARAMETER.fields_by_name['pool'].enum_type = _SPPPARAMETER_POOLMETHOD _SPPPARAMETER.fields_by_name['engine'].enum_type = _SPPPARAMETER_ENGINE _SPPPARAMETER_POOLMETHOD.containing_type = _SPPPARAMETER _SPPPARAMETER_ENGINE.containing_type = _SPPPARAMETER _V1LAYERPARAMETER.fields_by_name['include'].message_type = _NETSTATERULE _V1LAYERPARAMETER.fields_by_name['exclude'].message_type = _NETSTATERULE _V1LAYERPARAMETER.fields_by_name['type'].enum_type = _V1LAYERPARAMETER_LAYERTYPE _V1LAYERPARAMETER.fields_by_name['blobs'].message_type = _BLOBPROTO _V1LAYERPARAMETER.fields_by_name['blob_share_mode'].enum_type = _V1LAYERPARAMETER_DIMCHECKMODE _V1LAYERPARAMETER.fields_by_name['accuracy_param'].message_type = _ACCURACYPARAMETER _V1LAYERPARAMETER.fields_by_name['argmax_param'].message_type = _ARGMAXPARAMETER _V1LAYERPARAMETER.fields_by_name['concat_param'].message_type = _CONCATPARAMETER _V1LAYERPARAMETER.fields_by_name['contrastive_loss_param'].message_type = _CONTRASTIVELOSSPARAMETER _V1LAYERPARAMETER.fields_by_name['convolution_param'].message_type = _CONVOLUTIONPARAMETER _V1LAYERPARAMETER.fields_by_name['data_param'].message_type = _DATAPARAMETER _V1LAYERPARAMETER.fields_by_name['dropout_param'].message_type = _DROPOUTPARAMETER _V1LAYERPARAMETER.fields_by_name['dummy_data_param'].message_type = _DUMMYDATAPARAMETER _V1LAYERPARAMETER.fields_by_name['eltwise_param'].message_type = _ELTWISEPARAMETER _V1LAYERPARAMETER.fields_by_name['exp_param'].message_type = _EXPPARAMETER _V1LAYERPARAMETER.fields_by_name['hdf5_data_param'].message_type = _HDF5DATAPARAMETER _V1LAYERPARAMETER.fields_by_name['hdf5_output_param'].message_type = _HDF5OUTPUTPARAMETER _V1LAYERPARAMETER.fields_by_name['hinge_loss_param'].message_type = _HINGELOSSPARAMETER _V1LAYERPARAMETER.fields_by_name['image_data_param'].message_type = _IMAGEDATAPARAMETER _V1LAYERPARAMETER.fields_by_name['infogain_loss_param'].message_type = _INFOGAINLOSSPARAMETER _V1LAYERPARAMETER.fields_by_name['inner_product_param'].message_type = _INNERPRODUCTPARAMETER _V1LAYERPARAMETER.fields_by_name['lrn_param'].message_type = _LRNPARAMETER _V1LAYERPARAMETER.fields_by_name['memory_data_param'].message_type = _MEMORYDATAPARAMETER _V1LAYERPARAMETER.fields_by_name['mvn_param'].message_type = _MVNPARAMETER _V1LAYERPARAMETER.fields_by_name['pooling_param'].message_type = _POOLINGPARAMETER _V1LAYERPARAMETER.fields_by_name['power_param'].message_type = _POWERPARAMETER _V1LAYERPARAMETER.fields_by_name['relu_param'].message_type = _RELUPARAMETER _V1LAYERPARAMETER.fields_by_name['sigmoid_param'].message_type = _SIGMOIDPARAMETER _V1LAYERPARAMETER.fields_by_name['softmax_param'].message_type = _SOFTMAXPARAMETER _V1LAYERPARAMETER.fields_by_name['slice_param'].message_type = _SLICEPARAMETER _V1LAYERPARAMETER.fields_by_name['tanh_param'].message_type = _TANHPARAMETER _V1LAYERPARAMETER.fields_by_name['threshold_param'].message_type = _THRESHOLDPARAMETER _V1LAYERPARAMETER.fields_by_name['window_data_param'].message_type = _WINDOWDATAPARAMETER _V1LAYERPARAMETER.fields_by_name['transform_param'].message_type = _TRANSFORMATIONPARAMETER _V1LAYERPARAMETER.fields_by_name['loss_param'].message_type = _LOSSPARAMETER _V1LAYERPARAMETER.fields_by_name['layer'].message_type = _V0LAYERPARAMETER _V1LAYERPARAMETER_LAYERTYPE.containing_type = _V1LAYERPARAMETER _V1LAYERPARAMETER_DIMCHECKMODE.containing_type = _V1LAYERPARAMETER _V0LAYERPARAMETER.fields_by_name['weight_filler'].message_type = _FILLERPARAMETER _V0LAYERPARAMETER.fields_by_name['bias_filler'].message_type = _FILLERPARAMETER _V0LAYERPARAMETER.fields_by_name['pool'].enum_type = _V0LAYERPARAMETER_POOLMETHOD _V0LAYERPARAMETER.fields_by_name['blobs'].message_type = _BLOBPROTO _V0LAYERPARAMETER.fields_by_name['hdf5_output_param'].message_type = _HDF5OUTPUTPARAMETER _V0LAYERPARAMETER_POOLMETHOD.containing_type = _V0LAYERPARAMETER _PRELUPARAMETER.fields_by_name['filler'].message_type = _FILLERPARAMETER DESCRIPTOR.message_types_by_name['BlobShape'] = _BLOBSHAPE DESCRIPTOR.message_types_by_name['BlobProto'] = _BLOBPROTO DESCRIPTOR.message_types_by_name['BlobProtoVector'] = _BLOBPROTOVECTOR DESCRIPTOR.message_types_by_name['Datum'] = _DATUM DESCRIPTOR.message_types_by_name['FillerParameter'] = _FILLERPARAMETER DESCRIPTOR.message_types_by_name['NetParameter'] = _NETPARAMETER DESCRIPTOR.message_types_by_name['SolverParameter'] = _SOLVERPARAMETER DESCRIPTOR.message_types_by_name['SolverState'] = _SOLVERSTATE DESCRIPTOR.message_types_by_name['NetState'] = _NETSTATE DESCRIPTOR.message_types_by_name['NetStateRule'] = _NETSTATERULE DESCRIPTOR.message_types_by_name['ParamSpec'] = _PARAMSPEC DESCRIPTOR.message_types_by_name['LayerParameter'] = _LAYERPARAMETER DESCRIPTOR.message_types_by_name['TransformationParameter'] = _TRANSFORMATIONPARAMETER DESCRIPTOR.message_types_by_name['LossParameter'] = _LOSSPARAMETER DESCRIPTOR.message_types_by_name['AccuracyParameter'] = _ACCURACYPARAMETER DESCRIPTOR.message_types_by_name['ArgMaxParameter'] = _ARGMAXPARAMETER DESCRIPTOR.message_types_by_name['ConcatParameter'] = _CONCATPARAMETER DESCRIPTOR.message_types_by_name['BatchNormParameter'] = _BATCHNORMPARAMETER DESCRIPTOR.message_types_by_name['BiasParameter'] = _BIASPARAMETER DESCRIPTOR.message_types_by_name['ContrastiveLossParameter'] = _CONTRASTIVELOSSPARAMETER DESCRIPTOR.message_types_by_name['ConvolutionParameter'] = _CONVOLUTIONPARAMETER DESCRIPTOR.message_types_by_name['DataParameter'] = _DATAPARAMETER DESCRIPTOR.message_types_by_name['DropoutParameter'] = _DROPOUTPARAMETER DESCRIPTOR.message_types_by_name['DummyDataParameter'] = _DUMMYDATAPARAMETER DESCRIPTOR.message_types_by_name['EltwiseParameter'] = _ELTWISEPARAMETER DESCRIPTOR.message_types_by_name['ELUParameter'] = _ELUPARAMETER DESCRIPTOR.message_types_by_name['EmbedParameter'] = _EMBEDPARAMETER DESCRIPTOR.message_types_by_name['ExpParameter'] = _EXPPARAMETER DESCRIPTOR.message_types_by_name['FlattenParameter'] = _FLATTENPARAMETER DESCRIPTOR.message_types_by_name['HDF5DataParameter'] = _HDF5DATAPARAMETER DESCRIPTOR.message_types_by_name['HDF5OutputParameter'] = _HDF5OUTPUTPARAMETER DESCRIPTOR.message_types_by_name['HingeLossParameter'] = _HINGELOSSPARAMETER DESCRIPTOR.message_types_by_name['ImageDataParameter'] = _IMAGEDATAPARAMETER DESCRIPTOR.message_types_by_name['InfogainLossParameter'] = _INFOGAINLOSSPARAMETER DESCRIPTOR.message_types_by_name['InnerProductParameter'] = _INNERPRODUCTPARAMETER DESCRIPTOR.message_types_by_name['LogParameter'] = _LOGPARAMETER DESCRIPTOR.message_types_by_name['LRNParameter'] = _LRNPARAMETER DESCRIPTOR.message_types_by_name['MemoryDataParameter'] = _MEMORYDATAPARAMETER DESCRIPTOR.message_types_by_name['MVNParameter'] = _MVNPARAMETER DESCRIPTOR.message_types_by_name['PoolingParameter'] = _POOLINGPARAMETER DESCRIPTOR.message_types_by_name['PowerParameter'] = _POWERPARAMETER DESCRIPTOR.message_types_by_name['PythonParameter'] = _PYTHONPARAMETER DESCRIPTOR.message_types_by_name['ReductionParameter'] = _REDUCTIONPARAMETER DESCRIPTOR.message_types_by_name['ReLUParameter'] = _RELUPARAMETER DESCRIPTOR.message_types_by_name['ReshapeParameter'] = _RESHAPEPARAMETER DESCRIPTOR.message_types_by_name['ScaleParameter'] = _SCALEPARAMETER DESCRIPTOR.message_types_by_name['SigmoidParameter'] = _SIGMOIDPARAMETER DESCRIPTOR.message_types_by_name['SliceParameter'] = _SLICEPARAMETER DESCRIPTOR.message_types_by_name['SoftmaxParameter'] = _SOFTMAXPARAMETER DESCRIPTOR.message_types_by_name['TanHParameter'] = _TANHPARAMETER DESCRIPTOR.message_types_by_name['TileParameter'] = _TILEPARAMETER DESCRIPTOR.message_types_by_name['ThresholdParameter'] = _THRESHOLDPARAMETER DESCRIPTOR.message_types_by_name['WindowDataParameter'] = _WINDOWDATAPARAMETER DESCRIPTOR.message_types_by_name['SPPParameter'] = _SPPPARAMETER DESCRIPTOR.message_types_by_name['V1LayerParameter'] = _V1LAYERPARAMETER DESCRIPTOR.message_types_by_name['V0LayerParameter'] = _V0LAYERPARAMETER DESCRIPTOR.message_types_by_name['PReLUParameter'] = _PRELUPARAMETER DESCRIPTOR.enum_types_by_name['Phase'] = _PHASE BlobShape = _reflection.GeneratedProtocolMessageType('BlobShape', (_message.Message,), dict( DESCRIPTOR = _BLOBSHAPE, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.BlobShape) )) _sym_db.RegisterMessage(BlobShape) BlobProto = _reflection.GeneratedProtocolMessageType('BlobProto', (_message.Message,), dict( DESCRIPTOR = _BLOBPROTO, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.BlobProto) )) _sym_db.RegisterMessage(BlobProto) BlobProtoVector = _reflection.GeneratedProtocolMessageType('BlobProtoVector', (_message.Message,), dict( DESCRIPTOR = _BLOBPROTOVECTOR, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.BlobProtoVector) )) _sym_db.RegisterMessage(BlobProtoVector) Datum = _reflection.GeneratedProtocolMessageType('Datum', (_message.Message,), dict( DESCRIPTOR = _DATUM, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.Datum) )) _sym_db.RegisterMessage(Datum) FillerParameter = _reflection.GeneratedProtocolMessageType('FillerParameter', (_message.Message,), dict( DESCRIPTOR = _FILLERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.FillerParameter) )) _sym_db.RegisterMessage(FillerParameter) NetParameter = _reflection.GeneratedProtocolMessageType('NetParameter', (_message.Message,), dict( DESCRIPTOR = _NETPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.NetParameter) )) _sym_db.RegisterMessage(NetParameter) SolverParameter = _reflection.GeneratedProtocolMessageType('SolverParameter', (_message.Message,), dict( DESCRIPTOR = _SOLVERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SolverParameter) )) _sym_db.RegisterMessage(SolverParameter) SolverState = _reflection.GeneratedProtocolMessageType('SolverState', (_message.Message,), dict( DESCRIPTOR = _SOLVERSTATE, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SolverState) )) _sym_db.RegisterMessage(SolverState) NetState = _reflection.GeneratedProtocolMessageType('NetState', (_message.Message,), dict( DESCRIPTOR = _NETSTATE, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.NetState) )) _sym_db.RegisterMessage(NetState) NetStateRule = _reflection.GeneratedProtocolMessageType('NetStateRule', (_message.Message,), dict( DESCRIPTOR = _NETSTATERULE, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.NetStateRule) )) _sym_db.RegisterMessage(NetStateRule) ParamSpec = _reflection.GeneratedProtocolMessageType('ParamSpec', (_message.Message,), dict( DESCRIPTOR = _PARAMSPEC, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ParamSpec) )) _sym_db.RegisterMessage(ParamSpec) LayerParameter = _reflection.GeneratedProtocolMessageType('LayerParameter', (_message.Message,), dict( DESCRIPTOR = _LAYERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.LayerParameter) )) _sym_db.RegisterMessage(LayerParameter) TransformationParameter = _reflection.GeneratedProtocolMessageType('TransformationParameter', (_message.Message,), dict( DESCRIPTOR = _TRANSFORMATIONPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.TransformationParameter) )) _sym_db.RegisterMessage(TransformationParameter) LossParameter = _reflection.GeneratedProtocolMessageType('LossParameter', (_message.Message,), dict( DESCRIPTOR = _LOSSPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.LossParameter) )) _sym_db.RegisterMessage(LossParameter) AccuracyParameter = _reflection.GeneratedProtocolMessageType('AccuracyParameter', (_message.Message,), dict( DESCRIPTOR = _ACCURACYPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.AccuracyParameter) )) _sym_db.RegisterMessage(AccuracyParameter) ArgMaxParameter = _reflection.GeneratedProtocolMessageType('ArgMaxParameter', (_message.Message,), dict( DESCRIPTOR = _ARGMAXPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ArgMaxParameter) )) _sym_db.RegisterMessage(ArgMaxParameter) ConcatParameter = _reflection.GeneratedProtocolMessageType('ConcatParameter', (_message.Message,), dict( DESCRIPTOR = _CONCATPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ConcatParameter) )) _sym_db.RegisterMessage(ConcatParameter) BatchNormParameter = _reflection.GeneratedProtocolMessageType('BatchNormParameter', (_message.Message,), dict( DESCRIPTOR = _BATCHNORMPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.BatchNormParameter) )) _sym_db.RegisterMessage(BatchNormParameter) BiasParameter = _reflection.GeneratedProtocolMessageType('BiasParameter', (_message.Message,), dict( DESCRIPTOR = _BIASPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.BiasParameter) )) _sym_db.RegisterMessage(BiasParameter) ContrastiveLossParameter = _reflection.GeneratedProtocolMessageType('ContrastiveLossParameter', (_message.Message,), dict( DESCRIPTOR = _CONTRASTIVELOSSPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ContrastiveLossParameter) )) _sym_db.RegisterMessage(ContrastiveLossParameter) ConvolutionParameter = _reflection.GeneratedProtocolMessageType('ConvolutionParameter', (_message.Message,), dict( DESCRIPTOR = _CONVOLUTIONPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ConvolutionParameter) )) _sym_db.RegisterMessage(ConvolutionParameter) DataParameter = _reflection.GeneratedProtocolMessageType('DataParameter', (_message.Message,), dict( DESCRIPTOR = _DATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.DataParameter) )) _sym_db.RegisterMessage(DataParameter) DropoutParameter = _reflection.GeneratedProtocolMessageType('DropoutParameter', (_message.Message,), dict( DESCRIPTOR = _DROPOUTPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.DropoutParameter) )) _sym_db.RegisterMessage(DropoutParameter) DummyDataParameter = _reflection.GeneratedProtocolMessageType('DummyDataParameter', (_message.Message,), dict( DESCRIPTOR = _DUMMYDATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.DummyDataParameter) )) _sym_db.RegisterMessage(DummyDataParameter) EltwiseParameter = _reflection.GeneratedProtocolMessageType('EltwiseParameter', (_message.Message,), dict( DESCRIPTOR = _ELTWISEPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.EltwiseParameter) )) _sym_db.RegisterMessage(EltwiseParameter) ELUParameter = _reflection.GeneratedProtocolMessageType('ELUParameter', (_message.Message,), dict( DESCRIPTOR = _ELUPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ELUParameter) )) _sym_db.RegisterMessage(ELUParameter) EmbedParameter = _reflection.GeneratedProtocolMessageType('EmbedParameter', (_message.Message,), dict( DESCRIPTOR = _EMBEDPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.EmbedParameter) )) _sym_db.RegisterMessage(EmbedParameter) ExpParameter = _reflection.GeneratedProtocolMessageType('ExpParameter', (_message.Message,), dict( DESCRIPTOR = _EXPPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ExpParameter) )) _sym_db.RegisterMessage(ExpParameter) FlattenParameter = _reflection.GeneratedProtocolMessageType('FlattenParameter', (_message.Message,), dict( DESCRIPTOR = _FLATTENPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.FlattenParameter) )) _sym_db.RegisterMessage(FlattenParameter) HDF5DataParameter = _reflection.GeneratedProtocolMessageType('HDF5DataParameter', (_message.Message,), dict( DESCRIPTOR = _HDF5DATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.HDF5DataParameter) )) _sym_db.RegisterMessage(HDF5DataParameter) HDF5OutputParameter = _reflection.GeneratedProtocolMessageType('HDF5OutputParameter', (_message.Message,), dict( DESCRIPTOR = _HDF5OUTPUTPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.HDF5OutputParameter) )) _sym_db.RegisterMessage(HDF5OutputParameter) HingeLossParameter = _reflection.GeneratedProtocolMessageType('HingeLossParameter', (_message.Message,), dict( DESCRIPTOR = _HINGELOSSPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.HingeLossParameter) )) _sym_db.RegisterMessage(HingeLossParameter) ImageDataParameter = _reflection.GeneratedProtocolMessageType('ImageDataParameter', (_message.Message,), dict( DESCRIPTOR = _IMAGEDATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ImageDataParameter) )) _sym_db.RegisterMessage(ImageDataParameter) InfogainLossParameter = _reflection.GeneratedProtocolMessageType('InfogainLossParameter', (_message.Message,), dict( DESCRIPTOR = _INFOGAINLOSSPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.InfogainLossParameter) )) _sym_db.RegisterMessage(InfogainLossParameter) InnerProductParameter = _reflection.GeneratedProtocolMessageType('InnerProductParameter', (_message.Message,), dict( DESCRIPTOR = _INNERPRODUCTPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.InnerProductParameter) )) _sym_db.RegisterMessage(InnerProductParameter) LogParameter = _reflection.GeneratedProtocolMessageType('LogParameter', (_message.Message,), dict( DESCRIPTOR = _LOGPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.LogParameter) )) _sym_db.RegisterMessage(LogParameter) LRNParameter = _reflection.GeneratedProtocolMessageType('LRNParameter', (_message.Message,), dict( DESCRIPTOR = _LRNPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.LRNParameter) )) _sym_db.RegisterMessage(LRNParameter) MemoryDataParameter = _reflection.GeneratedProtocolMessageType('MemoryDataParameter', (_message.Message,), dict( DESCRIPTOR = _MEMORYDATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.MemoryDataParameter) )) _sym_db.RegisterMessage(MemoryDataParameter) MVNParameter = _reflection.GeneratedProtocolMessageType('MVNParameter', (_message.Message,), dict( DESCRIPTOR = _MVNPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.MVNParameter) )) _sym_db.RegisterMessage(MVNParameter) PoolingParameter = _reflection.GeneratedProtocolMessageType('PoolingParameter', (_message.Message,), dict( DESCRIPTOR = _POOLINGPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.PoolingParameter) )) _sym_db.RegisterMessage(PoolingParameter) PowerParameter = _reflection.GeneratedProtocolMessageType('PowerParameter', (_message.Message,), dict( DESCRIPTOR = _POWERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.PowerParameter) )) _sym_db.RegisterMessage(PowerParameter) PythonParameter = _reflection.GeneratedProtocolMessageType('PythonParameter', (_message.Message,), dict( DESCRIPTOR = _PYTHONPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.PythonParameter) )) _sym_db.RegisterMessage(PythonParameter) ReductionParameter = _reflection.GeneratedProtocolMessageType('ReductionParameter', (_message.Message,), dict( DESCRIPTOR = _REDUCTIONPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ReductionParameter) )) _sym_db.RegisterMessage(ReductionParameter) ReLUParameter = _reflection.GeneratedProtocolMessageType('ReLUParameter', (_message.Message,), dict( DESCRIPTOR = _RELUPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ReLUParameter) )) _sym_db.RegisterMessage(ReLUParameter) ReshapeParameter = _reflection.GeneratedProtocolMessageType('ReshapeParameter', (_message.Message,), dict( DESCRIPTOR = _RESHAPEPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ReshapeParameter) )) _sym_db.RegisterMessage(ReshapeParameter) ScaleParameter = _reflection.GeneratedProtocolMessageType('ScaleParameter', (_message.Message,), dict( DESCRIPTOR = _SCALEPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ScaleParameter) )) _sym_db.RegisterMessage(ScaleParameter) SigmoidParameter = _reflection.GeneratedProtocolMessageType('SigmoidParameter', (_message.Message,), dict( DESCRIPTOR = _SIGMOIDPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SigmoidParameter) )) _sym_db.RegisterMessage(SigmoidParameter) SliceParameter = _reflection.GeneratedProtocolMessageType('SliceParameter', (_message.Message,), dict( DESCRIPTOR = _SLICEPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SliceParameter) )) _sym_db.RegisterMessage(SliceParameter) SoftmaxParameter = _reflection.GeneratedProtocolMessageType('SoftmaxParameter', (_message.Message,), dict( DESCRIPTOR = _SOFTMAXPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SoftmaxParameter) )) _sym_db.RegisterMessage(SoftmaxParameter) TanHParameter = _reflection.GeneratedProtocolMessageType('TanHParameter', (_message.Message,), dict( DESCRIPTOR = _TANHPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.TanHParameter) )) _sym_db.RegisterMessage(TanHParameter) TileParameter = _reflection.GeneratedProtocolMessageType('TileParameter', (_message.Message,), dict( DESCRIPTOR = _TILEPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.TileParameter) )) _sym_db.RegisterMessage(TileParameter) ThresholdParameter = _reflection.GeneratedProtocolMessageType('ThresholdParameter', (_message.Message,), dict( DESCRIPTOR = _THRESHOLDPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.ThresholdParameter) )) _sym_db.RegisterMessage(ThresholdParameter) WindowDataParameter = _reflection.GeneratedProtocolMessageType('WindowDataParameter', (_message.Message,), dict( DESCRIPTOR = _WINDOWDATAPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.WindowDataParameter) )) _sym_db.RegisterMessage(WindowDataParameter) SPPParameter = _reflection.GeneratedProtocolMessageType('SPPParameter', (_message.Message,), dict( DESCRIPTOR = _SPPPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.SPPParameter) )) _sym_db.RegisterMessage(SPPParameter) V1LayerParameter = _reflection.GeneratedProtocolMessageType('V1LayerParameter', (_message.Message,), dict( DESCRIPTOR = _V1LAYERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.V1LayerParameter) )) _sym_db.RegisterMessage(V1LayerParameter) V0LayerParameter = _reflection.GeneratedProtocolMessageType('V0LayerParameter', (_message.Message,), dict( DESCRIPTOR = _V0LAYERPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.V0LayerParameter) )) _sym_db.RegisterMessage(V0LayerParameter) PReLUParameter = _reflection.GeneratedProtocolMessageType('PReLUParameter', (_message.Message,), dict( DESCRIPTOR = _PRELUPARAMETER, __module__ = 'caffe_pb2' # @@protoc_insertion_point(class_scope:caffe.PReLUParameter) )) _sym_db.RegisterMessage(PReLUParameter) _BLOBSHAPE.fields_by_name['dim'].has_options = True _BLOBSHAPE.fields_by_name['dim']._options = _descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001')) _BLOBPROTO.fields_by_name['data'].has_options = True _BLOBPROTO.fields_by_name['data']._options = _descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001')) _BLOBPROTO.fields_by_name['diff'].has_options = True _BLOBPROTO.fields_by_name['diff']._options = _descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001')) _BLOBPROTO.fields_by_name['double_data'].has_options = True _BLOBPROTO.fields_by_name['double_data']._options = _descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001')) _BLOBPROTO.fields_by_name['double_diff'].has_options = True _BLOBPROTO.fields_by_name['double_diff']._options = _descriptor._ParseOptions(descriptor_pb2.FieldOptions(), _b('\020\001')) # @@protoc_insertion_point(module_scope)
242,354
41.963127
27,796
py
chainer
chainer-master/chainer/links/model/vision/resnet.py
import collections import os import sys import warnings import numpy try: from PIL import Image available = True except ImportError as e: available = False _import_error = e import chainer from chainer.dataset.convert import concat_examples from chainer.dataset import download from chainer import function from chainer.functions.activation.relu import relu from chainer.functions.activation.softmax import softmax from chainer.functions.array.reshape import reshape from chainer.functions.math.sum import sum from chainer.functions.pooling.average_pooling_2d import average_pooling_2d from chainer.functions.pooling.max_pooling_nd import max_pooling_2d from chainer.initializers import constant from chainer.initializers import normal from chainer import link from chainer.links.connection.convolution_2d import Convolution2D from chainer.links.connection.linear import Linear from chainer.links.normalization.batch_normalization import BatchNormalization from chainer.serializers import npz from chainer.utils import argument from chainer.utils import imgproc from chainer.variable import Variable class ResNetLayers(link.Chain): """A pre-trained CNN model provided by MSRA. When you specify the path of the pre-trained chainer model serialized as a ``.npz`` file in the constructor, this chain model automatically initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector per image, or fine-tune the model on a different dataset. Note that unlike ``VGG16Layers``, it does not automatically download a pre-trained caffemodel. This caffemodel can be downloaded at `GitHub <https://github.com/KaimingHe/deep-residual-networks>`_. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. See: K. He et. al., `Deep Residual Learning for Image Recognition <https://arxiv.org/abs/1512.03385>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically loads and converts the caffemodel from ``$CHAINER_DATASET_ROOT/pfnet/chainer/models/ResNet-{n-layers}-model.caffemodel``, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value by modifying the environment variable and {n_layers} is replaced with the specified number of layers given as the first argument to this constructor. Note that in this case the converted chainer model is stored on the same directory and automatically used from the next time. If this argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.HeNormal(scale=1.0)``. n_layers (int): The number of layers of this model. It should be either 50, 101, or 152. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model, n_layers, downsample_fb=False): super(ResNetLayers, self).__init__() if pretrained_model: # As a sampling process is time-consuming, # we employ a zero initializer for faster computation. conv_kwargs = {'initialW': constant.Zero()} else: # employ default initializers used in the original paper conv_kwargs = {'initialW': normal.HeNormal(scale=1.0)} kwargs = conv_kwargs.copy() kwargs['downsample_fb'] = downsample_fb if n_layers == 50: block = [3, 4, 6, 3] elif n_layers == 101: block = [3, 4, 23, 3] elif n_layers == 152: block = [3, 8, 36, 3] else: raise ValueError('The n_layers argument should be either 50, 101,' ' or 152, but {} was given.'.format(n_layers)) with self.init_scope(): self.conv1 = Convolution2D(3, 64, 7, 2, 3, **conv_kwargs) self.bn1 = BatchNormalization(64) self.res2 = BuildingBlock(block[0], 64, 64, 256, 1, **kwargs) self.res3 = BuildingBlock(block[1], 256, 128, 512, 2, **kwargs) self.res4 = BuildingBlock(block[2], 512, 256, 1024, 2, **kwargs) self.res5 = BuildingBlock(block[3], 1024, 512, 2048, 2, **kwargs) self.fc6 = Linear(2048, 1000) if pretrained_model and pretrained_model.endswith('.caffemodel'): _retrieve(n_layers, 'ResNet-{}-model.npz'.format(n_layers), pretrained_model, self) elif pretrained_model: npz.load_npz(pretrained_model, self) @property def functions(self): return collections.OrderedDict([ ('conv1', [self.conv1, self.bn1, relu]), ('pool1', [lambda x: max_pooling_2d(x, ksize=3, stride=2)]), ('res2', [self.res2]), ('res3', [self.res3]), ('res4', [self.res4]), ('res5', [self.res5]), ('pool5', [_global_average_pooling_2d]), ('fc6', [self.fc6]), ('prob', [softmax]), ]) @property def available_layers(self): return list(self.functions.keys()) @classmethod def convert_caffemodel_to_npz(cls, path_caffemodel, path_npz, n_layers=50): """Converts a pre-trained caffemodel to a chainer model. Args: path_caffemodel (str): Path of the pre-trained caffemodel. path_npz (str): Path of the converted chainer model. """ # As CaffeFunction uses shortcut symbols, # we import CaffeFunction here. from chainer.links.caffe.caffe_function import CaffeFunction caffemodel = CaffeFunction(path_caffemodel) chainermodel = cls(pretrained_model=None, n_layers=n_layers) if n_layers == 50: _transfer_resnet50(caffemodel, chainermodel) elif n_layers == 101: _transfer_resnet101(caffemodel, chainermodel) elif n_layers == 152: _transfer_resnet152(caffemodel, chainermodel) else: raise ValueError('The n_layers argument should be either 50, 101,' ' or 152, but {} was given.'.format(n_layers)) npz.save_npz(path_npz, chainermodel, compression=False) def forward(self, x, layers=None, **kwargs): """forward(self, x, layers=['prob']) Computes all the feature maps specified by ``layers``. Args: x (~chainer.Variable): Input variable. It should be prepared by ``prepare`` function. layers (list of str): The list of layer names you want to extract. Returns: Dictionary of ~chainer.Variable: A directory in which the key contains the layer name and the value contains the corresponding feature map variable. """ if layers is None: layers = ['prob'] if kwargs: argument.check_unexpected_kwargs( kwargs, test='test argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) h = x activations = {} target_layers = set(layers) for key, funcs in self.functions.items(): if not target_layers: break for func in funcs: h = func(h) if key in target_layers: activations[key] = h target_layers.remove(key) return activations def extract(self, images, layers=None, size=(224, 224), **kwargs): """extract(self, images, layers=['pool5'], size=(224, 224)) Extracts all the feature maps of given images. The difference of directly executing ``forward`` is that it directly accepts images as an input and automatically transforms them to a proper variable. That is, it is also interpreted as a shortcut method that implicitly calls ``prepare`` and ``forward`` functions. Unlike ``predict`` method, this method does not override ``chainer.config.train`` and ``chainer.config.enable_backprop`` configuration. If you want to extract features without updating model parameters, you need to manually set configuration when calling this method as follows: .. code-block:: python # model is an instance of ResNetLayers (50 or 101 or 152 layers) with chainer.using_config('train', False): with chainer.using_config('enable_backprop', False): feature = model.extract([image]) Args: images (iterable of PIL.Image or numpy.ndarray): Input images. layers (list of str): The list of layer names you want to extract. size (pair of ints): The resolution of resized images used as an input of CNN. All the given images are not resized if this argument is ``None``, but the resolutions of all the images should be the same. Returns: Dictionary of ~chainer.Variable: A directory in which the key contains the layer name and the value contains the corresponding feature map variable. """ if layers is None: layers = ['pool5'] if kwargs: argument.check_unexpected_kwargs( kwargs, test='test argument is not supported anymore. ' 'Use chainer.using_config', volatile='volatile argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) x = concat_examples([prepare(img, size=size) for img in images]) x = Variable(self.xp.asarray(x)) return self(x, layers=layers) def predict(self, images, oversample=True): """Computes all the probabilities of given images. Args: images (iterable of PIL.Image or numpy.ndarray): Input images. When you specify a color image as a :class:`numpy.ndarray`, make sure that color order is RGB. oversample (bool): If ``True``, it averages results across center, corners, and mirrors. Otherwise, it uses only the center. Returns: ~chainer.Variable: Output that contains the class probabilities of given images. """ x = concat_examples([prepare(img, size=(256, 256)) for img in images]) if oversample: x = imgproc.oversample(x, crop_dims=(224, 224)) else: x = x[:, :, 16:240, 16:240] # Use no_backprop_mode to reduce memory consumption with function.no_backprop_mode(), chainer.using_config('train', False): x = Variable(self.xp.asarray(x)) y = self(x, layers=['prob'])['prob'] if oversample: n = len(y) // 10 y_shape = y.shape[1:] y = reshape(y, (n, 10) + y_shape) y = sum(y, axis=1) / 10 return y class ResNet50Layers(ResNetLayers): """A pre-trained CNN model with 50 layers provided by MSRA. When you specify the path of the pre-trained chainer model serialized as a ``.npz`` file in the constructor, this chain model automatically initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector per image, or fine-tune the model on a different dataset. Note that unlike ``VGG16Layers``, it does not automatically download a pre-trained caffemodel. This caffemodel can be downloaded at `GitHub <https://github.com/KaimingHe/deep-residual-networks>`_. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. ResNet50 has 25,557,096 trainable parameters, and it's 58% and 43% fewer than ResNet101 and ResNet152, respectively. On the other hand, the top-5 classification accuracy on ImageNet dataset drops only 0.7% and 1.1% from ResNet101 and ResNet152, respectively. Therefore, ResNet50 may have the best balance between the accuracy and the model size. It would be basically just enough for many cases, but some advanced models for object detection or semantic segmentation use deeper ones as their building blocks, so these deeper ResNets are here for making reproduction work easier. See: K. He et. al., `Deep Residual Learning for Image Recognition <https://arxiv.org/abs/1512.03385>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically loads and converts the caffemodel from ``$CHAINER_DATASET_ROOT/pfnet/chainer/models/ResNet-50-model.caffemodel``, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value by modifying the environment variable. Note that in this case the converted chainer model is stored on the same directory and automatically used from the next time. If this argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.HeNormal(scale=1.0)``. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto', downsample_fb=False): if pretrained_model == 'auto': pretrained_model = 'ResNet-50-model.caffemodel' super(ResNet50Layers, self).__init__( pretrained_model, 50, downsample_fb) class ResNet101Layers(ResNetLayers): """A pre-trained CNN model with 101 layers provided by MSRA. When you specify the path of the pre-trained chainer model serialized as a ``.npz`` file in the constructor, this chain model automatically initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector per image, or fine-tune the model on a different dataset. Note that unlike ``VGG16Layers``, it does not automatically download a pre-trained caffemodel. This caffemodel can be downloaded at `GitHub <https://github.com/KaimingHe/deep-residual-networks>`_. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. ResNet101 has 44,549,224 trainable parameters, and it's 43% fewer than ResNet152 model, while the top-5 classification accuracy on ImageNet dataset drops 1.1% from ResNet152. For many cases, ResNet50 may have the best balance between the accuracy and the model size. See: K. He et. al., `Deep Residual Learning for Image Recognition <https://arxiv.org/abs/1512.03385>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically loads and converts the caffemodel from ``$CHAINER_DATASET_ROOT/pfnet/chainer/models/ResNet-101-model.caffemodel``, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value by modifying the environment variable. Note that in this case the converted chainer model is stored on the same directory and automatically used from the next time. If this argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.HeNormal(scale=1.0)``. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto', downsample_fb=False): if pretrained_model == 'auto': pretrained_model = 'ResNet-101-model.caffemodel' super(ResNet101Layers, self).__init__( pretrained_model, 101, downsample_fb) class ResNet152Layers(ResNetLayers): """A pre-trained CNN model with 152 layers provided by MSRA. When you specify the path of the pre-trained chainer model serialized as a ``.npz`` file in the constructor, this chain model automatically initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector per image, or fine-tune the model on a different dataset. Note that unlike ``VGG16Layers``, it does not automatically download a pre-trained caffemodel. This caffemodel can be downloaded at `GitHub <https://github.com/KaimingHe/deep-residual-networks>`_. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. ResNet152 has 60,192,872 trainable parameters, and it's the deepest ResNet model and it achieves the best result on ImageNet classification task in `ILSVRC 2015 <http://image-net.org/challenges/LSVRC/2015/results#loc>`_. See: K. He et. al., `Deep Residual Learning for Image Recognition <https://arxiv.org/abs/1512.03385>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically loads and converts the caffemodel from ``$CHAINER_DATASET_ROOT/pfnet/chainer/models/ResNet-152-model.caffemodel``, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value by modifying the environment variable. Note that in this case the converted chainer model is stored on the same directory and automatically used from the next time. If this argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.HeNormal(scale=1.0)``. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto', downsample_fb=False): if pretrained_model == 'auto': pretrained_model = 'ResNet-152-model.caffemodel' super(ResNet152Layers, self).__init__( pretrained_model, 152, downsample_fb) def prepare(image, size=(224, 224)): """Converts the given image to a numpy array for ResNet. Note that this method must be called before calling ``forward``, because the pre-trained resnet model will resize the given image, convert from RGB to BGR, subtract the mean, and permute the dimensions before calling. Args: image (PIL.Image or numpy.ndarray): Input image. If an input is ``numpy.ndarray``, its shape must be ``(height, width)``, ``(height, width, channels)``, or ``(channels, height, width)``, and the order of the channels must be RGB. size (pair of ints): Size of converted images. If ``None``, the given image is not resized. Returns: numpy.ndarray: The converted output array. """ if not available: raise ImportError('PIL cannot be loaded. Install Pillow!\n' 'The actual import error is as follows:\n' + str(_import_error)) dtype = chainer.get_dtype() if isinstance(image, numpy.ndarray): if image.ndim == 3: if image.shape[0] == 1: image = image[0, :, :] elif image.shape[0] == 3: image = image.transpose((1, 2, 0)) image = Image.fromarray(image.astype(numpy.uint8)) image = image.convert('RGB') if size: image = image.resize(size) image = numpy.asarray(image, dtype=dtype) image = image[:, :, ::-1] # NOTE: in the original paper they subtract a fixed mean image, # however, in order to support arbitrary size we instead use the # mean pixel (rather than mean image) as with VGG team. The mean # value used in ResNet is slightly different from that of VGG16. image -= numpy.array( [103.063, 115.903, 123.152], dtype=dtype) image = image.transpose((2, 0, 1)) return image class BuildingBlock(link.Chain): """A building block that consists of several Bottleneck layers. Args: n_layer (int): *(deprecated since v7.0.0)* `n_layer` is now deprecated for consistency of naming choice. Please use `n_layers` instead. n_layers (int): Number of layers used in the building block. in_channels (int): Number of channels of input arrays. mid_channels (int): Number of channels of intermediate arrays. out_channels (int): Number of channels of output arrays. stride (int or tuple of ints): Stride of filter application. initialW (4-D array): Initial weight value used in the convolutional layers. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). """ def __init__(self, n_layers=None, in_channels=None, mid_channels=None, out_channels=None, stride=None, initialW=None, downsample_fb=None, **kwargs): super(BuildingBlock, self).__init__() if 'n_layer' in kwargs: warnings.warn( 'Argument `n_layer` is deprecated. ' 'Please use `n_layers` instead', DeprecationWarning) n_layers = kwargs['n_layer'] with self.init_scope(): self.a = BottleneckA( in_channels, mid_channels, out_channels, stride, initialW, downsample_fb) self._forward = ['a'] for i in range(n_layers - 1): name = 'b{}'.format(i + 1) bottleneck = BottleneckB(out_channels, mid_channels, initialW) setattr(self, name, bottleneck) self._forward.append(name) def forward(self, x): for name in self._forward: l = getattr(self, name) x = l(x) return x class BottleneckA(link.Chain): """A bottleneck layer that reduces the resolution of the feature map. Args: in_channels (int): Number of channels of input arrays. mid_channels (int): Number of channels of intermediate arrays. out_channels (int): Number of channels of output arrays. stride (int or tuple of ints): Stride of filter application. initialW (4-D array): Initial weight value used in the convolutional layers. downsample_fb (bool): If this argument is specified as ``False``, it performs downsampling by placing stride 2 on the 1x1 convolutional layers (the original MSRA ResNet). If this argument is specified as ``True``, it performs downsampling by placing stride 2 on the 3x3 convolutional layers (Facebook ResNet). """ def __init__(self, in_channels, mid_channels, out_channels, stride=2, initialW=None, downsample_fb=False): super(BottleneckA, self).__init__() # In the original MSRA ResNet, stride=2 is on 1x1 convolution. # In Facebook ResNet, stride=2 is on 3x3 convolution. stride_1x1, stride_3x3 = (1, stride) if downsample_fb else (stride, 1) with self.init_scope(): self.conv1 = Convolution2D( in_channels, mid_channels, 1, stride_1x1, 0, initialW=initialW, nobias=True) self.bn1 = BatchNormalization(mid_channels) self.conv2 = Convolution2D( mid_channels, mid_channels, 3, stride_3x3, 1, initialW=initialW, nobias=True) self.bn2 = BatchNormalization(mid_channels) self.conv3 = Convolution2D( mid_channels, out_channels, 1, 1, 0, initialW=initialW, nobias=True) self.bn3 = BatchNormalization(out_channels) self.conv4 = Convolution2D( in_channels, out_channels, 1, stride, 0, initialW=initialW, nobias=True) self.bn4 = BatchNormalization(out_channels) def forward(self, x): h1 = relu(self.bn1(self.conv1(x))) h1 = relu(self.bn2(self.conv2(h1))) h1 = self.bn3(self.conv3(h1)) h2 = self.bn4(self.conv4(x)) return relu(h1 + h2) class BottleneckB(link.Chain): """A bottleneck layer that maintains the resolution of the feature map. Args: in_channels (int): Number of channels of input and output arrays. mid_channels (int): Number of channels of intermediate arrays. initialW (4-D array): Initial weight value used in the convolutional layers. """ def __init__(self, in_channels, mid_channels, initialW=None): super(BottleneckB, self).__init__() with self.init_scope(): self.conv1 = Convolution2D( in_channels, mid_channels, 1, 1, 0, initialW=initialW, nobias=True) self.bn1 = BatchNormalization(mid_channels) self.conv2 = Convolution2D( mid_channels, mid_channels, 3, 1, 1, initialW=initialW, nobias=True) self.bn2 = BatchNormalization(mid_channels) self.conv3 = Convolution2D( mid_channels, in_channels, 1, 1, 0, initialW=initialW, nobias=True) self.bn3 = BatchNormalization(in_channels) def forward(self, x): h = relu(self.bn1(self.conv1(x))) h = relu(self.bn2(self.conv2(h))) h = self.bn3(self.conv3(h)) return relu(h + x) def _global_average_pooling_2d(x): n, channel, rows, cols = x.shape h = average_pooling_2d(x, (rows, cols), stride=1) h = reshape(h, (n, channel)) return h def _transfer_components(src, dst_conv, dst_bn, bname, cname): src_conv = getattr(src, 'res{}_branch{}'.format(bname, cname)) src_bn = getattr(src, 'bn{}_branch{}'.format(bname, cname)) src_scale = getattr(src, 'scale{}_branch{}'.format(bname, cname)) dst_conv.W.array[:] = src_conv.W.array dst_bn.avg_mean[:] = src_bn.avg_mean dst_bn.avg_var[:] = src_bn.avg_var dst_bn.gamma.array[:] = src_scale.W.array dst_bn.beta.array[:] = src_scale.bias.b.array def _transfer_bottleneckA(src, dst, name): _transfer_components(src, dst.conv1, dst.bn1, name, '2a') _transfer_components(src, dst.conv2, dst.bn2, name, '2b') _transfer_components(src, dst.conv3, dst.bn3, name, '2c') _transfer_components(src, dst.conv4, dst.bn4, name, '1') def _transfer_bottleneckB(src, dst, name): _transfer_components(src, dst.conv1, dst.bn1, name, '2a') _transfer_components(src, dst.conv2, dst.bn2, name, '2b') _transfer_components(src, dst.conv3, dst.bn3, name, '2c') def _transfer_block(src, dst, names): _transfer_bottleneckA(src, dst.a, names[0]) for i, name in enumerate(names[1:]): dst_bottleneckB = getattr(dst, 'b{}'.format(i + 1)) _transfer_bottleneckB(src, dst_bottleneckB, name) def _transfer_resnet50(src, dst): dst.conv1.W.array[:] = src.conv1.W.array dst.conv1.b.array[:] = src.conv1.b.array dst.bn1.avg_mean[:] = src.bn_conv1.avg_mean dst.bn1.avg_var[:] = src.bn_conv1.avg_var dst.bn1.gamma.array[:] = src.scale_conv1.W.array dst.bn1.beta.array[:] = src.scale_conv1.bias.b.array _transfer_block(src, dst.res2, ['2a', '2b', '2c']) _transfer_block(src, dst.res3, ['3a', '3b', '3c', '3d']) _transfer_block(src, dst.res4, ['4a', '4b', '4c', '4d', '4e', '4f']) _transfer_block(src, dst.res5, ['5a', '5b', '5c']) dst.fc6.W.array[:] = src.fc1000.W.array dst.fc6.b.array[:] = src.fc1000.b.array def _transfer_resnet101(src, dst): dst.conv1.W.array[:] = src.conv1.W.array dst.bn1.avg_mean[:] = src.bn_conv1.avg_mean dst.bn1.avg_var[:] = src.bn_conv1.avg_var dst.bn1.gamma.array[:] = src.scale_conv1.W.array dst.bn1.beta.array[:] = src.scale_conv1.bias.b.array _transfer_block(src, dst.res2, ['2a', '2b', '2c']) _transfer_block(src, dst.res3, ['3a', '3b1', '3b2', '3b3']) _transfer_block(src, dst.res4, ['4a'] + ['4b{}'.format(i) for i in range(1, 23)]) _transfer_block(src, dst.res5, ['5a', '5b', '5c']) dst.fc6.W.array[:] = src.fc1000.W.array dst.fc6.b.array[:] = src.fc1000.b.array def _transfer_resnet152(src, dst): dst.conv1.W.array[:] = src.conv1.W.array dst.bn1.avg_mean[:] = src.bn_conv1.avg_mean dst.bn1.avg_var[:] = src.bn_conv1.avg_var dst.bn1.gamma.array[:] = src.scale_conv1.W.array dst.bn1.beta.array[:] = src.scale_conv1.bias.b.array _transfer_block(src, dst.res2, ['2a', '2b', '2c']) _transfer_block(src, dst.res3, ['3a'] + ['3b{}'.format(i) for i in range(1, 8)]) _transfer_block(src, dst.res4, ['4a'] + ['4b{}'.format(i) for i in range(1, 36)]) _transfer_block(src, dst.res5, ['5a', '5b', '5c']) dst.fc6.W.array[:] = src.fc1000.W.array dst.fc6.b.array[:] = src.fc1000.b.array def _make_npz(path_npz, path_caffemodel, model, n_layers): sys.stderr.write( 'Now loading caffemodel (usually it may take few minutes)\n') sys.stderr.flush() if not os.path.exists(path_caffemodel): raise IOError( 'The pre-trained caffemodel does not exist. Please download it ' 'from \'https://github.com/KaimingHe/deep-residual-networks\', ' 'and place it on {}'.format(path_caffemodel)) ResNetLayers.convert_caffemodel_to_npz(path_caffemodel, path_npz, n_layers) npz.load_npz(path_npz, model) return model def _retrieve(n_layers, name_npz, name_caffemodel, model): root = download.get_dataset_directory('pfnet/chainer/models/') path = os.path.join(root, name_npz) path_caffemodel = os.path.join(root, name_caffemodel) return download.cache_or_load_file( path, lambda path: _make_npz(path, path_caffemodel, model, n_layers), lambda path: npz.load_npz(path, model))
33,339
41.74359
94
py
chainer
chainer-master/chainer/links/model/vision/vgg.py
import collections import os import sys import numpy try: from PIL import Image available = True except ImportError as e: available = False _import_error = e import chainer from chainer.dataset.convert import concat_examples from chainer.dataset import download from chainer import function from chainer.functions.activation.relu import relu from chainer.functions.activation.softmax import softmax from chainer.functions.array.reshape import reshape from chainer.functions.math.sum import sum from chainer.functions.noise.dropout import dropout from chainer.functions.pooling.max_pooling_nd import max_pooling_2d from chainer.initializers import constant from chainer.initializers import normal from chainer import link from chainer.links.connection.convolution_2d import Convolution2D from chainer.links.connection.linear import Linear from chainer.serializers import npz from chainer.utils import argument from chainer.utils import imgproc from chainer.variable import Variable class VGGLayers(link.Chain): """A pre-trained CNN model provided by VGG team. You can use ``VGG16Layers`` or ``VGG19Layers`` for concrete implementations. During initialization, this chain model automatically downloads the pre-trained caffemodel, convert to another chainer model, stores it on your local directory, and initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector from a given image, or fine-tune the model on a different dataset. Note that these pre-trained models are released under Creative Commons Attribution License. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. See: K. Simonyan and A. Zisserman, `Very Deep Convolutional Networks for Large-Scale Image Recognition <https://arxiv.org/abs/1409.1556>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically downloads the caffemodel from the internet. Note that in this case the converted chainer model is stored on ``$CHAINER_DATASET_ROOT/pfnet/chainer/models`` directory, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value as a environment variable. The converted chainer model is automatically used from the second time. If the argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.Normal(scale=0.01)``. n_layers (int): The number of layers of this model. It should be either 16 or 19. Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto', n_layers=16): super(VGGLayers, self).__init__() if pretrained_model: # As a sampling process is time-consuming, # we employ a zero initializer for faster computation. init = constant.Zero() kwargs = {'initialW': init, 'initial_bias': init} else: # employ default initializers used in the original paper kwargs = { 'initialW': normal.Normal(0.01), 'initial_bias': constant.Zero(), } if n_layers not in [16, 19]: raise ValueError( 'The n_layers argument should be either 16 or 19, ' 'but {} was given.'.format(n_layers) ) with self.init_scope(): self.conv1_1 = Convolution2D(3, 64, 3, 1, 1, **kwargs) self.conv1_2 = Convolution2D(64, 64, 3, 1, 1, **kwargs) self.conv2_1 = Convolution2D(64, 128, 3, 1, 1, **kwargs) self.conv2_2 = Convolution2D(128, 128, 3, 1, 1, **kwargs) self.conv3_1 = Convolution2D(128, 256, 3, 1, 1, **kwargs) self.conv3_2 = Convolution2D(256, 256, 3, 1, 1, **kwargs) self.conv3_3 = Convolution2D(256, 256, 3, 1, 1, **kwargs) self.conv4_1 = Convolution2D(256, 512, 3, 1, 1, **kwargs) self.conv4_2 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.conv4_3 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.conv5_1 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.conv5_2 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.conv5_3 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.fc6 = Linear(512 * 7 * 7, 4096, **kwargs) self.fc7 = Linear(4096, 4096, **kwargs) self.fc8 = Linear(4096, 1000, **kwargs) if n_layers == 19: self.conv3_4 = Convolution2D(256, 256, 3, 1, 1, **kwargs) self.conv4_4 = Convolution2D(512, 512, 3, 1, 1, **kwargs) self.conv5_4 = Convolution2D(512, 512, 3, 1, 1, **kwargs) if pretrained_model == 'auto': if n_layers == 16: _retrieve( 'VGG_ILSVRC_16_layers.npz', 'https://www.robots.ox.ac.uk/%7Evgg/software/very_deep/' 'caffe/VGG_ILSVRC_16_layers.caffemodel', self) else: _retrieve( 'VGG_ILSVRC_19_layers.npz', 'http://www.robots.ox.ac.uk/%7Evgg/software/very_deep/' 'caffe/VGG_ILSVRC_19_layers.caffemodel', self) elif pretrained_model: npz.load_npz(pretrained_model, self) @property def functions(self): # This class will not be used directly. raise NotImplementedError @property def available_layers(self): return list(self.functions.keys()) @classmethod def convert_caffemodel_to_npz(cls, path_caffemodel, path_npz): """Converts a pre-trained caffemodel to a chainer model. Args: path_caffemodel (str): Path of the pre-trained caffemodel. path_npz (str): Path of the converted chainer model. """ # As CaffeFunction uses shortcut symbols, # we import CaffeFunction here. from chainer.links.caffe.caffe_function import CaffeFunction caffemodel = CaffeFunction(path_caffemodel) npz.save_npz(path_npz, caffemodel, compression=False) def forward(self, x, layers=None, **kwargs): """forward(self, x, layers=['prob']) Computes all the feature maps specified by ``layers``. Args: x (~chainer.Variable): Input variable. It should be prepared by ``prepare`` function. layers (list of str): The list of layer names you want to extract. If ``None``, 'prob' will be used as layers. Returns: Dictionary of ~chainer.Variable: A dictionary in which the key contains the layer and the value contains the corresponding feature map variable. """ if layers is None: layers = ['prob'] if kwargs: argument.check_unexpected_kwargs( kwargs, test='test argument is not supported anymore. ' 'Use chainer.using_config' ) argument.assert_kwargs_empty(kwargs) h = x activations = {} target_layers = set(layers) for key, funcs in self.functions.items(): if not target_layers: break for func in funcs: h = func(h) if key in target_layers: activations[key] = h target_layers.remove(key) return activations def extract(self, images, layers=None, size=(224, 224), **kwargs): """extract(self, images, layers=['fc7'], size=(224, 224)) Extracts all the feature maps of given images. The difference of directly executing ``forward`` is that it directly accepts images as an input and automatically transforms them to a proper variable. That is, it is also interpreted as a shortcut method that implicitly calls ``prepare`` and ``forward`` functions. Unlike ``predict`` method, this method does not override ``chainer.config.train`` and ``chainer.config.enable_backprop`` configuration. If you want to extract features without updating model parameters, you need to manually set configuration when calling this method as follows: .. code-block:: python # model is an instance of VGGLayers (16 or 19 layers) with chainer.using_config('train', False): with chainer.using_config('enable_backprop', False): feature = model.extract([image]) Args: images (iterable of PIL.Image or numpy.ndarray): Input images. layers (list of str): The list of layer names you want to extract. size (pair of ints): The resolution of resized images used as an input of CNN. All the given images are not resized if this argument is ``None``, but the resolutions of all the images should be the same. Returns: Dictionary of ~chainer.Variable: A directory in which the key contains the layer name and the value contains the corresponding feature map variable. """ if layers is None: layers = ['fc7'] if kwargs: argument.check_unexpected_kwargs( kwargs, test='test argument is not supported anymore. ' 'Use chainer.using_config', volatile='volatile argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) x = concat_examples([prepare(img, size=size) for img in images]) x = Variable(self.xp.asarray(x)) return self(x, layers=layers) def predict(self, images, oversample=True): """Computes all the probabilities of given images. Args: images (iterable of PIL.Image or numpy.ndarray): Input images. When you specify a color image as a :class:`numpy.ndarray`, make sure that color order is RGB. oversample (bool): If ``True``, it averages results across center, corners, and mirrors. Otherwise, it uses only the center. Returns: ~chainer.Variable: Output that contains the class probabilities of given images. """ x = concat_examples([prepare(img, size=(256, 256)) for img in images]) if oversample: x = imgproc.oversample(x, crop_dims=(224, 224)) else: x = x[:, :, 16:240, 16:240] # Use no_backprop_mode to reduce memory consumption with function.no_backprop_mode(), chainer.using_config('train', False): x = Variable(self.xp.asarray(x)) y = self(x, layers=['prob'])['prob'] if oversample: n = len(y) // 10 y_shape = y.shape[1:] y = reshape(y, (n, 10) + y_shape) y = sum(y, axis=1) / 10 return y class VGG16Layers(VGGLayers): """A pre-trained CNN model with 16 layers provided by VGG team. During initialization, this chain model automatically downloads the pre-trained caffemodel, convert to another chainer model, stores it on your local directory, and initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector from a given image, or fine-tune the model on a different dataset. Note that this pre-trained model is released under Creative Commons Attribution License. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. See: K. Simonyan and A. Zisserman, `Very Deep Convolutional Networks for Large-Scale Image Recognition <https://arxiv.org/abs/1409.1556>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically downloads the caffemodel from the internet. Note that in this case the converted chainer model is stored on ``$CHAINER_DATASET_ROOT/pfnet/chainer/models`` directory, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value as a environment variable. The converted chainer model is automatically used from the second time. If the argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.Normal(scale=0.01)``. Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto'): super(VGG16Layers, self).__init__(pretrained_model, 16) @property def functions(self): return collections.OrderedDict([ ('conv1_1', [self.conv1_1, relu]), ('conv1_2', [self.conv1_2, relu]), ('pool1', [_max_pooling_2d]), ('conv2_1', [self.conv2_1, relu]), ('conv2_2', [self.conv2_2, relu]), ('pool2', [_max_pooling_2d]), ('conv3_1', [self.conv3_1, relu]), ('conv3_2', [self.conv3_2, relu]), ('conv3_3', [self.conv3_3, relu]), ('pool3', [_max_pooling_2d]), ('conv4_1', [self.conv4_1, relu]), ('conv4_2', [self.conv4_2, relu]), ('conv4_3', [self.conv4_3, relu]), ('pool4', [_max_pooling_2d]), ('conv5_1', [self.conv5_1, relu]), ('conv5_2', [self.conv5_2, relu]), ('conv5_3', [self.conv5_3, relu]), ('pool5', [_max_pooling_2d]), ('fc6', [self.fc6, relu, dropout]), ('fc7', [self.fc7, relu, dropout]), ('fc8', [self.fc8]), ('prob', [softmax]), ]) class VGG19Layers(VGGLayers): """A pre-trained CNN model with 19 layers provided by VGG team. During initialization, this chain model automatically downloads the pre-trained caffemodel, convert to another chainer model, stores it on your local directory, and initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector from a given image, or fine-tune the model on a different dataset. Note that this pre-trained model is released under Creative Commons Attribution License. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. See: K. Simonyan and A. Zisserman, `Very Deep Convolutional Networks for Large-Scale Image Recognition <https://arxiv.org/abs/1409.1556>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically downloads the caffemodel from the internet. Note that in this case the converted chainer model is stored on ``$CHAINER_DATASET_ROOT/pfnet/chainer/models`` directory, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value as a environment variable. The converted chainer model is automatically used from the second time. If the argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in the original paper, i.e., ``chainer.initializers.Normal(scale=0.01)``. Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto'): super(VGG19Layers, self).__init__(pretrained_model, 19) @property def functions(self): return collections.OrderedDict([ ('conv1_1', [self.conv1_1, relu]), ('conv1_2', [self.conv1_2, relu]), ('pool1', [_max_pooling_2d]), ('conv2_1', [self.conv2_1, relu]), ('conv2_2', [self.conv2_2, relu]), ('pool2', [_max_pooling_2d]), ('conv3_1', [self.conv3_1, relu]), ('conv3_2', [self.conv3_2, relu]), ('conv3_3', [self.conv3_3, relu]), ('conv3_4', [self.conv3_4, relu]), ('pool3', [_max_pooling_2d]), ('conv4_1', [self.conv4_1, relu]), ('conv4_2', [self.conv4_2, relu]), ('conv4_3', [self.conv4_3, relu]), ('conv4_4', [self.conv4_4, relu]), ('pool4', [_max_pooling_2d]), ('conv5_1', [self.conv5_1, relu]), ('conv5_2', [self.conv5_2, relu]), ('conv5_3', [self.conv5_3, relu]), ('conv5_4', [self.conv5_4, relu]), ('pool5', [_max_pooling_2d]), ('fc6', [self.fc6, relu, dropout]), ('fc7', [self.fc7, relu, dropout]), ('fc8', [self.fc8]), ('prob', [softmax]), ]) def prepare(image, size=(224, 224)): """Converts the given image to the numpy array for VGG models. Note that you have to call this method before ``forward`` because the pre-trained vgg model requires to resize the given image, covert the RGB to the BGR, subtract the mean, and permute the dimensions before calling. Args: image (PIL.Image or numpy.ndarray): Input image. If an input is ``numpy.ndarray``, its shape must be ``(height, width)``, ``(height, width, channels)``, or ``(channels, height, width)``, and the order of the channels must be RGB. size (pair of ints): Size of converted images. If ``None``, the given image is not resized. Returns: numpy.ndarray: The converted output array. """ if not available: raise ImportError('PIL cannot be loaded. Install Pillow!\n' 'The actual import error is as follows:\n' + str(_import_error)) dtype = chainer.get_dtype() if isinstance(image, numpy.ndarray): if image.ndim == 3: if image.shape[0] == 1: image = image[0, :, :] elif image.shape[0] == 3: image = image.transpose((1, 2, 0)) image = Image.fromarray(image.astype(numpy.uint8)) image = image.convert('RGB') if size: image = image.resize(size) image = numpy.asarray(image, dtype=dtype) image = image[:, :, ::-1] image -= numpy.array( [103.939, 116.779, 123.68], dtype=dtype) image = image.transpose((2, 0, 1)) return image def _max_pooling_2d(x): return max_pooling_2d(x, ksize=2) def _make_npz(path_npz, url, model): path_caffemodel = download.cached_download(url) sys.stderr.write( 'Now loading caffemodel (usually it may take few minutes)\n') sys.stderr.flush() VGGLayers.convert_caffemodel_to_npz(path_caffemodel, path_npz) npz.load_npz(path_npz, model) return model def _retrieve(name, url, model): root = download.get_dataset_directory('pfnet/chainer/models/') path = os.path.join(root, name) return download.cache_or_load_file( path, lambda path: _make_npz(path, url, model), lambda path: npz.load_npz(path, model))
20,475
39.546535
79
py
chainer
chainer-master/chainer/links/model/vision/googlenet.py
import collections import os import sys import numpy try: from PIL import Image available = True except ImportError as e: available = False _import_error = e import chainer from chainer.dataset.convert import concat_examples from chainer.dataset import download from chainer import function from chainer.functions.activation.relu import relu from chainer.functions.activation.softmax import softmax from chainer.functions.array.reshape import reshape from chainer.functions.math.average import average from chainer.functions.noise.dropout import dropout from chainer.functions.normalization.local_response_normalization import ( local_response_normalization) from chainer.functions.pooling.average_pooling_2d import average_pooling_2d from chainer.functions.pooling.max_pooling_nd import max_pooling_2d from chainer.initializers import constant from chainer.initializers import uniform from chainer import link from chainer.links.connection.convolution_2d import Convolution2D from chainer.links.connection.inception import Inception from chainer.links.connection.linear import Linear from chainer.serializers import npz from chainer.utils import argument from chainer.utils import imgproc from chainer.variable import Variable class GoogLeNet(link.Chain): """A pre-trained GoogLeNet model provided by BVLC. When you specify the path of the pre-trained chainer model serialized as a ``.npz`` file in the constructor, this chain model automatically initializes all the parameters with it. This model would be useful when you want to extract a semantic feature vector per image, or fine-tune the model on a different dataset. If you want to manually convert the pre-trained caffemodel to a chainer model that can be specified in the constructor, please use ``convert_caffemodel_to_npz`` classmethod instead. GoogLeNet, which is also called Inception-v1, is an architecture of convolutional neural network proposed in 2014. This model is relatively lightweight and requires small memory footprint during training compared with modern architectures such as ResNet. Therefore, if you fine-tune your network based on a model pre-trained by Imagenet and need to train it with large batch size, GoogLeNet may be useful. On the other hand, if you just want an off-the-shelf classifier, we recommend that you use ResNet50 or other models since they are more accurate than GoogLeNet. The original model is provided here: `<https://github.com/BVLC/caffe/tree/master/models/bvlc_googlenet>`_ Args: pretrained_model (str): the destination of the pre-trained chainer model serialized as a ``.npz`` file. If this argument is specified as ``auto``, it automatically downloads the caffemodel from the internet. Note that in this case the converted chainer model is stored on ``$CHAINER_DATASET_ROOT/pfnet/chainer/models`` directory, where ``$CHAINER_DATASET_ROOT`` is set as ``$HOME/.chainer/dataset`` unless you specify another value as a environment variable. The converted chainer model is automatically used from the second time. If the argument is specified as ``None``, all the parameters are not initialized by the pre-trained model, but the default initializer used in BVLC, i.e., ``chainer.initializers.LeCunUniform(scale=1.0)``. Note that, in Caffe, when weight_filler is specified as "xavier" type without variance_norm parameter, the weights are initialized by Uniform(-s, s), where :math:`s = \\sqrt{\\frac{3}{fan_{in}}}` and :math:`fan_{in}` is the number of input units. This corresponds to LeCunUniform in Chainer but not GlorotUniform. Attributes: available_layers (list of str): The list of available layer names used by ``forward`` and ``extract`` methods. """ def __init__(self, pretrained_model='auto'): super(GoogLeNet, self).__init__() if pretrained_model: # As a sampling process is time-consuming, # we employ a zero initializer for faster computation. kwargs = {'initialW': constant.Zero()} else: # employ default initializers used in BVLC. For more detail, see # https://github.com/chainer/chainer/pull/2424#discussion_r109642209 kwargs = {'initialW': uniform.LeCunUniform(scale=1.0)} with self.init_scope(): self.conv1 = Convolution2D(3, 64, 7, stride=2, pad=3, **kwargs) self.conv2_reduce = Convolution2D(64, 64, 1, **kwargs) self.conv2 = Convolution2D(64, 192, 3, stride=1, pad=1, **kwargs) self.inc3a = Inception(192, 64, 96, 128, 16, 32, 32) self.inc3b = Inception(256, 128, 128, 192, 32, 96, 64) self.inc4a = Inception(480, 192, 96, 208, 16, 48, 64) self.inc4b = Inception(512, 160, 112, 224, 24, 64, 64) self.inc4c = Inception(512, 128, 128, 256, 24, 64, 64) self.inc4d = Inception(512, 112, 144, 288, 32, 64, 64) self.inc4e = Inception(528, 256, 160, 320, 32, 128, 128) self.inc5a = Inception(832, 256, 160, 320, 32, 128, 128) self.inc5b = Inception(832, 384, 192, 384, 48, 128, 128) self.loss3_fc = Linear(1024, 1000, **kwargs) self.loss1_conv = Convolution2D(512, 128, 1, **kwargs) self.loss1_fc1 = Linear(2048, 1024, **kwargs) self.loss1_fc2 = Linear(1024, 1000, **kwargs) self.loss2_conv = Convolution2D(528, 128, 1, **kwargs) self.loss2_fc1 = Linear(2048, 1024, **kwargs) self.loss2_fc2 = Linear(1024, 1000, **kwargs) if pretrained_model == 'auto': _retrieve( 'bvlc_googlenet.npz', 'http://dl.caffe.berkeleyvision.org/bvlc_googlenet.caffemodel', self) elif pretrained_model: npz.load_npz(pretrained_model, self) @property def functions(self): return collections.OrderedDict([ ('conv1', [self.conv1, relu]), ('pool1', [_max_pooling_2d, _local_response_normalization]), ('conv2_reduce', [self.conv2_reduce, relu]), ('conv2', [self.conv2, relu, _local_response_normalization]), ('pool2', [_max_pooling_2d]), ('inception_3a', [self.inc3a]), ('inception_3b', [self.inc3b]), ('pool3', [_max_pooling_2d]), ('inception_4a', [self.inc4a]), ('inception_4b', [self.inc4b]), ('inception_4c', [self.inc4c]), ('inception_4d', [self.inc4d]), ('inception_4e', [self.inc4e]), ('pool4', [_max_pooling_2d]), ('inception_5a', [self.inc5a]), ('inception_5b', [self.inc5b]), ('pool5', [_average_pooling_2d_k7]), ('loss3_fc', [_dropout, self.loss3_fc]), ('prob', [softmax]), # Since usually the following outputs are not used, they are put # after 'prob' to be skipped for efficiency. ('loss1_fc2', [_average_pooling_2d_k5, self.loss1_conv, relu, self.loss1_fc1, relu, self.loss1_fc2]), ('loss2_fc2', [_average_pooling_2d_k5, self.loss2_conv, relu, self.loss2_fc1, relu, self.loss2_fc2]) ]) @property def available_layers(self): return list(self.functions.keys()) @classmethod def convert_caffemodel_to_npz(cls, path_caffemodel, path_npz): """Converts a pre-trained caffemodel to a chainer model. Args: path_caffemodel (str): Path of the pre-trained caffemodel. path_npz (str): Path of the converted chainer model. """ # As CaffeFunction uses shortcut symbols, # we import CaffeFunction here. from chainer.links.caffe.caffe_function import CaffeFunction caffemodel = CaffeFunction(path_caffemodel) chainermodel = cls(pretrained_model=None) _transfer_googlenet(caffemodel, chainermodel) npz.save_npz(path_npz, chainermodel, compression=False) def forward(self, x, layers=None, **kwargs): """forward(self, x, layers=['prob']) Computes all the feature maps specified by ``layers``. Args: x (~chainer.Variable): Input variable. It should be prepared by ``prepare`` function. layers (list of str): The list of layer names you want to extract. Returns: Dictionary of ~chainer.Variable: A directory in which the key contains the layer name and the value contains the corresponding feature map variable. """ if layers is None: layers = ['prob'] if kwargs: argument.check_unexpected_kwargs( kwargs, train='train argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) h = x activations = {} inception_4a_cache = None inception_4d_cache = None target_layers = set(layers) for key, funcs in self.functions.items(): if not target_layers: break if key == 'loss1_fc2': h = inception_4a_cache elif key == 'loss2_fc2': h = inception_4d_cache for func in funcs: h = func(h) if key in target_layers: activations[key] = h target_layers.remove(key) if key == 'inception_4a': inception_4a_cache = h elif key == 'inception_4d': inception_4d_cache = h return activations def extract(self, images, layers=None, size=(224, 224), **kwargs): """extract(self, images, layers=['pool5'], size=(224, 224)) Extracts all the feature maps of given images. The difference of directly executing ``forward`` is that it directly accepts images as an input and automatically transforms them to a proper variable. That is, it is also interpreted as a shortcut method that implicitly calls ``prepare`` and ``forward`` functions. Unlike ``predict`` method, this method does not override ``chainer.config.train`` and ``chainer.config.enable_backprop`` configuration. If you want to extract features without updating model parameters, you need to manually set configuration when calling this method as follows: .. code-block:: python # model is an instance of `GoogLeNet` with chainer.using_config('train', False): with chainer.using_config('enable_backprop', False): feature = model.extract([image]) Args: images (iterable of PIL.Image or numpy.ndarray): Input images. layers (list of str): The list of layer names you want to extract. size (pair of ints): The resolution of resized images used as an input of CNN. All the given images are not resized if this argument is ``None``, but the resolutions of all the images should be the same. Returns: Dictionary of ~chainer.Variable: A directory in which the key contains the layer name and the value contains the corresponding feature map variable. """ if layers is None: layers = ['pool5'] if kwargs: argument.check_unexpected_kwargs( kwargs, train='train argument is not supported anymore. ' 'Use chainer.using_config', volatile='volatile argument is not supported anymore. ' 'Use chainer.using_config') argument.assert_kwargs_empty(kwargs) x = concat_examples([prepare(img, size=size) for img in images]) x = Variable(self.xp.asarray(x)) return self(x, layers=layers) def predict(self, images, oversample=True): """Computes all the probabilities of given images. Args: images (iterable of PIL.Image or numpy.ndarray): Input images. When you specify a color image as a :class:`numpy.ndarray`, make sure that color order is RGB. oversample (bool): If ``True``, it averages results across center, corners, and mirrors. Otherwise, it uses only the center. Returns: ~chainer.Variable: Output that contains the class probabilities of given images. """ x = concat_examples([prepare(img, size=(256, 256)) for img in images]) if oversample: x = imgproc.oversample(x, crop_dims=(224, 224)) else: x = x[:, :, 16:240, 16:240] # Use no_backprop_mode to reduce memory consumption with function.no_backprop_mode(), chainer.using_config('train', False): x = Variable(self.xp.asarray(x)) y = self(x, layers=['prob'])['prob'] if oversample: n = len(y) // 10 y_shape = y.shape[1:] y = reshape(y, (n, 10) + y_shape) y = average(y, axis=1) return y def prepare(image, size=(224, 224)): """Converts the given image to the numpy array for GoogLeNet. Note that you have to call this method before ``forward`` because the pre-trained GoogLeNet model requires to resize the given image, covert the RGB to the BGR, subtract the mean, and permute the dimensions before calling. Args: image (PIL.Image or numpy.ndarray): Input image. If an input is ``numpy.ndarray``, its shape must be ``(height, width)``, ``(height, width, channels)``, or ``(channels, height, width)``, and the order of the channels must be RGB. size (pair of ints): Size of converted images. If ``None``, the given image is not resized. Returns: numpy.ndarray: The converted output array. """ if not available: raise ImportError('PIL cannot be loaded. Install Pillow!\n' 'The actual import error is as follows:\n' + str(_import_error)) dtype = chainer.get_dtype() if isinstance(image, numpy.ndarray): if image.ndim == 3: if image.shape[0] == 1: image = image[0, :, :] elif image.shape[0] == 3: image = image.transpose((1, 2, 0)) image = Image.fromarray(image.astype(numpy.uint8)) image = image.convert('RGB') if size: image = image.resize(size) image = numpy.asarray(image, dtype=dtype) image = image[:, :, ::-1] image -= numpy.array([104.0, 117.0, 123.0], dtype=dtype) # BGR image = image.transpose((2, 0, 1)) return image def _transfer_inception(src, dst, names): for name in names: chain = getattr(dst, 'inc{}'.format(name)) src_prefix = 'inception_{}/'.format(name) chain.conv1.W.array[:] = src[src_prefix + '1x1'].W.array chain.conv1.b.array[:] = src[src_prefix + '1x1'].b.array chain.proj3.W.array[:] = src[src_prefix + '3x3_reduce'].W.array chain.proj3.b.array[:] = src[src_prefix + '3x3_reduce'].b.array chain.conv3.W.array[:] = src[src_prefix + '3x3'].W.array chain.conv3.b.array[:] = src[src_prefix + '3x3'].b.array chain.proj5.W.array[:] = src[src_prefix + '5x5_reduce'].W.array chain.proj5.b.array[:] = src[src_prefix + '5x5_reduce'].b.array chain.conv5.W.array[:] = src[src_prefix + '5x5'].W.array chain.conv5.b.array[:] = src[src_prefix + '5x5'].b.array chain.projp.W.array[:] = src[src_prefix + 'pool_proj'].W.array chain.projp.b.array[:] = src[src_prefix + 'pool_proj'].b.array def _transfer_googlenet(src, dst): # 1 ################################################################# dst.conv1.W.array[:] = src['conv1/7x7_s2'].W.array dst.conv1.b.array[:] = src['conv1/7x7_s2'].b.array # 2 ################################################################# dst.conv2_reduce.W.array[:] = src['conv2/3x3_reduce'].W.array dst.conv2_reduce.b.array[:] = src['conv2/3x3_reduce'].b.array dst.conv2.W.array[:] = src['conv2/3x3'].W.array dst.conv2.b.array[:] = src['conv2/3x3'].b.array # 3, 4, 5 ########################################################### _transfer_inception(src, dst, ['3a', '3b', '4a', '4b', '4c', '4d', '4e', '5a', '5b']) # outputs ############################################################ dst.loss1_conv.W.array[:] = src['loss1/conv'].W.array dst.loss1_conv.b.array[:] = src['loss1/conv'].b.array dst.loss1_fc1.W.array[:] = src['loss1/fc'].W.array dst.loss1_fc1.b.array[:] = src['loss1/fc'].b.array dst.loss1_fc2.W.array[:] = src['loss1/classifier'].W.array dst.loss1_fc2.b.array[:] = src['loss1/classifier'].b.array dst.loss2_conv.W.array[:] = src['loss2/conv'].W.array dst.loss2_conv.b.array[:] = src['loss2/conv'].b.array dst.loss2_fc1.W.array[:] = src['loss2/fc'].W.array dst.loss2_fc1.b.array[:] = src['loss2/fc'].b.array dst.loss2_fc2.W.array[:] = src['loss2/classifier'].W.array dst.loss2_fc2.b.array[:] = src['loss2/classifier'].b.array dst.loss3_fc.W.array[:] = src['loss3/classifier'].W.array dst.loss3_fc.b.array[:] = src['loss3/classifier'].b.array def _max_pooling_2d(x): return max_pooling_2d(x, ksize=3, stride=2) def _local_response_normalization(x): return local_response_normalization(x, n=5, k=1, alpha=1e-4 / 5) def _average_pooling_2d_k5(x): return average_pooling_2d(x, ksize=5, stride=3) def _average_pooling_2d_k7(x): return average_pooling_2d(x, ksize=7, stride=1) def _dropout(x): return dropout(x, ratio=0.4) def _make_npz(path_npz, url, model): path_caffemodel = download.cached_download(url) sys.stderr.write( 'Now loading caffemodel (usually it may take few minutes)\n') sys.stderr.flush() GoogLeNet.convert_caffemodel_to_npz(path_caffemodel, path_npz) npz.load_npz(path_npz, model) return model def _retrieve(name_npz, url, model): root = download.get_dataset_directory('pfnet/chainer/models/') path = os.path.join(root, name_npz) return download.cache_or_load_file( path, lambda path: _make_npz(path, url, model), lambda path: npz.load_npz(path, model))
18,874
40.032609
80
py
chainer
chainer-master/chainer/initializers/uniform.py
import numpy from chainer import backend from chainer import initializer from chainer.utils import argument # Original code forked from MIT licensed keras project # https://github.com/fchollet/keras/blob/master/keras/initializations.py class Uniform(initializer.Initializer): """Initializes array with a scaled uniform distribution. Each element of the array is initialized by the value drawn independently from uniform distribution :math:`[-scale, scale]`. Attributes: scale (float): A constant that determines the scale of the uniform distribution. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=0.05, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(Uniform, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) if self.rng is None: device = backend.get_device_from_array(array) array[...] = device.xp.random.uniform( low=-self.scale, high=self.scale, size=array.shape) else: backend.copyto(array, self.rng.uniform( low=-self.scale, high=self.scale, size=array.shape).astype(array.dtype, copy=False)) class LeCunUniform(initializer.Initializer): """Initializes array with a scaled uniform distribution. Each element of the array is initialized by the value drawn independently from uniform distribution :math:`[-s, s]` where :math:`s = scale \\times \\sqrt{\\frac{3}{fan_{in}}}`. Here :math:`fan_{in}` is the number of input units. Reference: LeCun 98, Efficient Backprop http://yann.lecun.com/exdb/publis/pdf/lecun-98b.pdf Attributes: scale (float): A constant that determines the scale of the uniform distribution. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(LeCunUniform, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) s = self.scale * numpy.sqrt(3. / fan_in) Uniform(s, rng=self.rng)(array) class GlorotUniform(initializer.Initializer): """Initializes array with a scaled uniform distribution. Each element of the array is initialized by the value drawn independently from uniform distribution :math:`[-s, s]` where :math:`s = scale \\times \\sqrt{\\frac{6}{fan_{in} + fan_{out}}}`. Here, :math:`fan_{in}` and :math:`fan_{out}` are the number of input and output units, respectively. Attributes: scale (float): A constant that determines the scale of the uniform distribution. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(GlorotUniform, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) s = self.scale * numpy.sqrt(6. / (fan_in + fan_out)) Uniform(s, rng=self.rng)(array) class HeUniform(initializer.Initializer): """Initializes array with scaled uniform distribution. Each element of the array is initialized by the value drawn independently from uniform distribution :math:`[-s, s]` where :math:`s = scale \\times \\sqrt{\\frac{6}{fan_{in}}}`. Here, :math:`fan_{in}` is the number of input units. Attributes: scale (float): A constant that determines the scale of the uniform distribution. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(HeUniform, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) s = self.scale * numpy.sqrt(6. / fan_in) Uniform(s, rng=self.rng)(array)
5,202
33.230263
76
py
chainer
chainer-master/chainer/initializers/normal.py
import numpy from chainer import backend from chainer.backends import cuda from chainer import initializer from chainer.utils import argument # Original code forked from MIT licensed keras project # https://github.com/fchollet/keras/blob/master/keras/initializations.py class Normal(initializer.Initializer): """Initializes array with a normal distribution. Each element of the array is initialized by the value drawn independently from Gaussian distribution whose mean is 0, and standard deviation is ``scale``. Args: scale (float): Standard deviation of Gaussian distribution. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=0.05, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(Normal, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) if self.rng is None: device = backend.get_device_from_array(array) args = {'loc': 0.0, 'scale': self.scale, 'size': array.shape} if device.xp is cuda.cupy: # Only CuPy supports dtype option if self.dtype == numpy.float32 or self.dtype == numpy.float16: # float16 is not supported in cuRAND args['dtype'] = numpy.float32 array[...] = device.xp.random.normal(**args) else: backend.copyto(array, self.rng.normal( loc=0.0, scale=self.scale, size=array.shape).astype(array.dtype, copy=False)) class LeCunNormal(initializer.Initializer): """Initializes array with scaled Gaussian distribution. Each element of the array is initialized by the value drawn independently from Gaussian distribution whose mean is 0, and standard deviation is :math:`scale \\times \\sqrt{\\frac{1}{fan_{in}}}`, where :math:`fan_{in}` is the number of input units. Reference: LeCun 98, Efficient Backprop http://yann.lecun.com/exdb/publis/pdf/lecun-98b.pdf Args: scale (float): A constant that determines the scale of the standard deviation. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(LeCunNormal, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) s = self.scale * numpy.sqrt(1. / fan_in) Normal(s, rng=self.rng)(array) class GlorotNormal(initializer.Initializer): """Initializes array with scaled Gaussian distribution. Each element of the array is initialized by the value drawn independently from Gaussian distribution whose mean is 0, and standard deviation is :math:`scale \\times \\sqrt{\\frac{2}{fan_{in} + fan_{out}}}`, where :math:`fan_{in}` and :math:`fan_{out}` are the number of input and output units, respectively. Reference: Glorot & Bengio, AISTATS 2010 Args: scale (float): A constant that determines the scale of the standard deviation. dtype: Data type specifier. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, **kwargs): self.scale = scale rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(GlorotNormal, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) s = self.scale * numpy.sqrt(2. / (fan_in + fan_out)) Normal(s, rng=self.rng)(array) class HeNormal(initializer.Initializer): """Initializes array with scaled Gaussian distribution. Each element of the array is initialized by the value drawn independently from Gaussian distribution whose mean is 0, and standard deviation is :math:`scale \\times \\sqrt{\\frac{2}{fan}}`. If ``fan_option == 'fan_in'``, :math:`fan` is the number of input units. If ``fan_option == 'fan_out'``, :math:`fan` is the number of output units. Reference: He et al., https://arxiv.org/abs/1502.01852 Args: scale (float): A constant that determines the scale of the standard deviation. dtype: Data type specifier. fan_option ({'fan_in', 'fan_out'}): Decides how to compute the standard deviation. The default value is ``'fan_in'``. rng (xp.random.RandomState): Pseudo-random number generator. """ def __init__(self, scale=1.0, dtype=None, fan_option='fan_in', **kwargs): self.scale = scale self.fan_option = fan_option rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng super(HeNormal, self).__init__(dtype) def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) fan_in, fan_out = initializer.get_fans(array.shape) if self.fan_option == 'fan_in': s = self.scale * numpy.sqrt(2. / fan_in) elif self.fan_option == 'fan_out': s = self.scale * numpy.sqrt(2. / fan_out) else: raise ValueError( 'fan_option should be either \'fan_in\' or \'fan_out\'.') Normal(s, rng=self.rng)(array)
6,197
34.016949
78
py
chainer
chainer-master/chainer/initializers/orthogonal.py
import numpy from chainer import backend from chainer import initializer from chainer import utils from chainer.utils import argument _orthogonal_constraints = { # (assert emb., assert proj.) 'auto': (False, False), 'projection': (False, True), 'embedding': (True, False), 'basis': (True, True), } # Original code forked from MIT licensed keras project # https://github.com/fchollet/keras/blob/master/keras/initializations.py class Orthogonal(initializer.Initializer): """Initializes array with an orthogonal system. This initializer first makes a matrix of the same shape as the array to be initialized whose elements are drawn independently from standard Gaussian distribution. Next, it applies QR decomposition to (the transpose of) the matrix. To make the decomposition (almost surely) unique, we require the diagonal of the triangular matrix R to be non-negative (see e.g. Edelman & Rao, https://web.eecs.umich.edu/~rajnrao/Acta05rmt.pdf). Then, it initializes the array with the (semi-)orthogonal matrix Q. Finally, the array is multiplied by the constant ``scale``. If the ``ndim`` of the input array is more than 2, we consider the array to be a matrix by concatenating all axes except the first one. The number of vectors consisting of the orthogonal system (i.e. first element of the shape of the array) must be equal to or smaller than the dimension of each vector (i.e. second element of the shape of the array). Attributes: scale (float): A constant to be multiplied by. dtype: Data type specifier. mode (str): Assertion on the initialized shape. ``'auto'`` (default), ``'projection'`` (before v7), ``'embedding'``, or ``'basis'``. rng (xp.random.RandomState): Pseudo-random number generator. Reference: Saxe et al., https://arxiv.org/abs/1312.6120 """ def __init__(self, scale=1.1, dtype=None, mode='auto', **kwargs): self.scale = scale self.mode = mode rng = None if kwargs: rng, = argument.parse_kwargs(kwargs, ('rng', rng)) self.rng = rng try: self._checks = _orthogonal_constraints[mode] except KeyError: raise ValueError( 'Invalid mode: {}. Choose from {}.'.format( repr(mode), ', '.join(repr(m) for m in _orthogonal_constraints))) super(Orthogonal, self).__init__(dtype) # TODO(Kenta Oono) # How do we treat overcomplete base-system case? def __call__(self, array): if self.dtype is not None: assert array.dtype == self.dtype,\ '{} != {}'.format(array.dtype, self.dtype) if not array.shape: # 0-dim case if self.rng is None: a = numpy.random.randint(2) else: a = self.rng.randint(2) a = int(a) array[...] = self.scale * (2 * a - 1) elif not array.size: raise ValueError('Array to be initialized must be non-empty.') else: # numpy.prod returns float value when the argument is empty. out_dim = len(array) in_dim = utils.size_of_shape(array.shape[1:]) if (in_dim > out_dim and self._checks[0]) or ( in_dim < out_dim and self._checks[1]): raise ValueError( 'Cannot make orthogonal {}. ' 'shape = {}, interpreted as ' '{}-dim input and {}-dim output.'.format( self.mode, array.shape, in_dim, out_dim)) transpose = in_dim > out_dim if self.rng is None: a = numpy.random.normal(size=(out_dim, in_dim)) else: a_tmp = self.rng.normal(size=(out_dim, in_dim)) a = numpy.empty(a_tmp.shape, dtype=a_tmp.dtype) backend.copyto(a, a_tmp) if transpose: a = a.T # cupy.linalg.qr requires cusolver in CUDA 8+ q, r = numpy.linalg.qr(a) q *= numpy.copysign(self.scale, numpy.diag(r)) if transpose: q = q.T backend.copyto(array, q.reshape(array.shape).astype( array.dtype, copy=False))
4,361
38.297297
78
py
chainer
chainer-master/chainer/exporters/__init__.py
from chainer.exporters import caffe # NOQA
44
21.5
43
py
chainer
chainer-master/chainer/exporters/caffe.py
import collections import heapq import os import numpy import six import chainer from chainer import function from chainer import function_node from chainer.links.caffe.protobuf3 import caffe_pb2 as caffe_pb from chainer import variable _function_types = (function.Function, function_node.FunctionNode) def _add_blob(layer, shape, data): # The following part is ridiculously slow!! # TODO(okuta): Replace with C++ extension call blob = layer.blobs.add() blob.shape.dim[:] = shape blob.data[:] = data.flatten() def _dump_graph(outputs): fan_out = collections.defaultdict(int) cand_funcs = [] def add_cand_to_check(cands): for cand in cands: x = cand.creator if x is None: continue if x not in fan_out: # `len(fan_out)` is in order to avoid comparing `x` heapq.heappush(cand_funcs, (-x.rank, len(fan_out), x)) fan_out[x] += 1 add_cand_to_check(outputs) while cand_funcs: _, _, func = heapq.heappop(cand_funcs) assert isinstance(func, _function_types) add_cand_to_check(func.inputs) ret = [] cand_funcs = [] seen_set = set() def add_cand(cands): cands = [cand.creator for cand in cands if cand.creator is not None] for x in cands: if x in seen_set: continue order = 1 if fan_out[x] == 1 and len(cands) == 1: order = -len(seen_set) # Negate since heapq is min-heap # `len(seen_set)` is in order to avoid comparing `x` heapq.heappush(cand_funcs, (order, -x.rank, -len(seen_set), x)) seen_set.add(x) add_cand(outputs) while cand_funcs: _, _, _, func = heapq.heappop(cand_funcs) ret.append(func) add_cand(func.inputs) return ret[::-1] class _RetrieveAsCaffeModel(object): debug = False def __init__(self, prototxt, caffemodel=None): self.caffemodel = caffemodel self.prototxt = prototxt # key:string, val:dict(key: func, val: index) self.naming_map = collections.defaultdict(dict) def _get_layer_name(self, layer): """Generate layer name like "Convolution2DFunction-10-2". The first number means rank of the layer (depth from the top), and the second number is for preventing duplication (different layer objects can have same rank) Args: layer (~chainer.Function_node): Function object Returns: str: A string to be used for the ``name`` field of the graph in the exported Caffe model. """ label = '{}-{}'.format(layer.label, layer.rank) d = self.naming_map[label] if layer not in d.keys(): d[layer] = len(d) + 1 return '{}-{}'.format(label, d[layer]) def _get_parent_name(self, parent_): if parent_ is None: return 'data' return self._get_layer_name(parent_) def _gen_layer_prototxt(self, layer_params, name='layer', depth=0, indent=2): if isinstance(layer_params, (dict, collections.OrderedDict)): s = name + ' {\n' indent_s = ' ' * ((depth + 1) * indent) for key, val in layer_params.items(): s += indent_s + \ self._gen_layer_prototxt(val, name=key, depth=depth + 1) s += ' ' * (depth * indent) s += '}\n' return s elif isinstance(layer_params, bool): return '{}: {}\n'.format(name, 'true' if layer_params else 'false') elif isinstance(layer_params, six.integer_types + (float,)): return '{}: {}\n'.format(name, layer_params) elif isinstance(layer_params, str): return '{}: "{}"\n'.format(name, layer_params) elif isinstance(layer_params, list): s = '' indent_s = ' ' * depth * indent for i, t in enumerate(layer_params): if i != 0: s += indent_s s += self._gen_layer_prototxt(t, name=name, depth=depth + 1) return s else: raise ValueError('Unsupported type: ' + str(type(layer_params))) def dump_function_object(self, func, prototxt, net): assert isinstance(func, _function_types) layer_name = self._get_layer_name(func) parent_layer_names = [self._get_parent_name(input_.creator) for input_ in func.inputs] params = collections.OrderedDict() params['type'] = None params['name'] = layer_name params['bottom'] = parent_layer_names params['top'] = [layer_name] layer = None if net is not None: layer = net.layer.add() if func.label == 'LinearFunction': if len(func.inputs) == 2: _, W = func.inputs b = None else: _, W, b = func.inputs n_out, n_in = W.shape inner_product_param = { 'num_output': n_out, 'bias_term': b is not None, } params['type'] = 'InnerProduct' params['inner_product_param'] = inner_product_param params['bottom'] = params['bottom'][:1] if net is not None: for k, v in six.iteritems(inner_product_param): setattr(layer.inner_product_param, k, v) _add_blob(layer, list(W.shape), W.data) if b is not None: b.retain_data() _add_blob(layer, list(b.shape), b.data) elif func.label in ('Convolution2DFunction', 'Deconvolution2DFunction'): if len(func.inputs) == 2: _, W = func.inputs b = None else: _, W, b = func.inputs n_out, n_in, kw, kh = W.shape convolution_param = { 'num_output': n_out, 'bias_term': b is not None, 'pad_w': func.pw, 'pad_h': func.ph, 'stride_w': func.sx, 'stride_h': func.sy, 'kernel_w': kw, 'kernel_h': kh, 'group': func.groups } params['bottom'] = params['bottom'][:1] if func.label == 'Convolution2DFunction': params['type'] = 'Convolution' else: params['type'] = 'Deconvolution' convolution_param['num_output'] = n_in params['convolution_param'] = convolution_param if net is not None: for k, v in six.iteritems(convolution_param): setattr(layer.convolution_param, k, v) _add_blob(layer, [n_out, n_in, kh, kw], W.data) if b is not None: b.retain_data() _add_blob(layer, [n_out], b.data) elif func.label == 'AveragePooling2D': kw = func.kw kh = func.kh pooling_param = { 'pool': 1, 'pad_w': func.pw, 'pad_h': func.ph, 'stride_w': func.sx, 'stride_h': func.sy, 'kernel_w': kw, 'kernel_h': kh, } params['type'] = 'Pooling' params['pooling_param'] = pooling_param if net is not None: for k, v in six.iteritems(pooling_param): setattr(layer.pooling_param, k, v) elif func.label == 'MaxPoolingND' and func.ndim == 2: kh, kw = func.ksize sy, sx = func.stride ph, pw = func.pad pooling_param = { 'pool': 0, 'pad_w': pw, 'pad_h': ph, 'stride_w': sx, 'stride_h': sy, 'kernel_w': kw, 'kernel_h': kh, } params['type'] = 'Pooling' params['pooling_param'] = pooling_param if net is not None: for k, v in six.iteritems(pooling_param): setattr(layer.pooling_param, k, v) elif func.label == 'LocalResponseNormalization': lrn_param = { 'norm_region': 0, # ACROSS_CHANNELS 'local_size': func.n, 'k': func.k, 'alpha': func.alpha * func.n, 'beta': func.beta, } params['type'] = 'LRN' params['lrn_param'] = lrn_param if net is not None: for k, v in six.iteritems(lrn_param): setattr(layer.lrn_param, k, v) elif func.label == 'FixedBatchNormalization': _, gamma, beta, mean, var = func.inputs batch_norm_param = {'use_global_stats': True, 'eps': func.eps} params['type'] = 'BatchNorm' params['bottom'] = params['bottom'][:1] params['batch_norm_param'] = batch_norm_param if net is not None: for k, v in six.iteritems(batch_norm_param): setattr(layer.batch_norm_param, k, v) _add_blob(layer, [mean.data.size], mean.data) _add_blob(layer, [var.data.size], var.data) _add_blob(layer, [1], numpy.ones((1,), dtype=numpy.float32)) if gamma.data is None and beta.data is None: pass else: bn_name = layer_name + '_bn' params['name'] = bn_name params['top'] = [bn_name] if prototxt is not None: prototxt.write(self._gen_layer_prototxt(params)) if net is not None: layer.name = params['name'] layer.type = params['type'] layer.bottom[:] = params['bottom'] layer.top[:] = params['top'] layer.phase = caffe_pb.TEST del params, layer params = collections.OrderedDict() params['type'] = 'Scale' params['name'] = layer_name params['bottom'] = [bn_name] params['top'] = [layer_name] if net is not None: layer = net.layer.add() beta.retain_data() bias_term = beta.data is not None scale_param = { 'axis': 1, 'bias_term': bias_term, } params['scale_param'] = scale_param if net is not None: for k, v in six.iteritems(scale_param): setattr(layer.scale_param, k, v) _add_blob(layer, [gamma.data.size], gamma.data) if bias_term: _add_blob(layer, [beta.data.size], beta.data) elif func.label == 'ReLU': params['type'] = 'ReLU' elif func.label == 'LeakyReLU': relu_param = {'negative_slope': func.slope} params['type'] = 'ReLU' params['relu_param'] = relu_param if net is not None: for k, v in six.iteritems(relu_param): setattr(layer.relu_param, k, v) elif func.label == 'Concat': axis = func.axis concat_param = {'axis': axis} params['type'] = 'Concat' params['concat_param'] = concat_param if net is not None: for k, v in six.iteritems(concat_param): setattr(layer.concat_param, k, v) elif func.label == 'Softmax': params['type'] = 'Softmax' elif func.label == 'Sigmoid': params['type'] = 'Sigmoid' elif func.label == 'Reshape': input_ = func.inputs[0] parent = input_.creator parent_layer_name = parent_layer_names[0] if 'Reshape' in parent_layer_name: grandparent = parent.inputs[0].creator parent_layer_name = self._get_parent_name(grandparent) reshape_param = {'shape': {'dim': list(func.shape)}} params['type'] = 'Reshape' params['bottom'] = [parent_layer_name] params['reshape_param'] = reshape_param if layer is not None: dim = reshape_param['shape']['dim'] layer.reshape_param.shape.dim[:] = dim elif func.label == '_ + _': params['type'] = 'Eltwise' else: raise Exception( 'Cannot convert, name={}, rank={}, label={}, inputs={}'.format( layer_name, func.rank, func.label, parent_layer_names)) if prototxt is not None: prototxt.write(self._gen_layer_prototxt(params)) if net is not None: layer.name = params['name'] layer.type = params['type'] layer.bottom[:] = params['bottom'] layer.top[:] = params['top'] layer.phase = caffe_pb.TEST def __call__(self, name, inputs, outputs): dumped_list = _dump_graph(outputs) f = None net = None if self.caffemodel is not None: net = caffe_pb.NetParameter() try: if self.prototxt is not None: f = open(self.prototxt, 'wt') f.write('name: "{}"\n'.format(name)) assert len(inputs) == 1 f.write('layer {\n' ' name: "data"\n' ' type: "Input"\n' ' top: "data"\n' ' input_param { shape: {') for i in inputs[0].shape: f.write(' dim: ' + str(i)) f.write(' } }\n' '}\n') for i in dumped_list: self.dump_function_object(i, f, net) finally: if f is not None: f.close() if net is not None: with open(self.caffemodel, 'wb') as f: f.write(net.SerializeToString()) if self.debug: import google.protobuf.text_format with open(self.caffemodel + '.txt', 'w') as f: f.write(google.protobuf.text_format.MessageToString(net)) def export(model, args, directory=None, export_params=True, graph_name='Graph'): """(Experimental) Export a computational graph as Caffe format. Args: model (~chainer.Chain): The model object you want to export in Caffe format. It should have :meth:`__call__` method because the second argument ``args`` is directly given to the model by the ``()`` accessor. args (list of ~chainer.Variable): The arguments which are given to the model directly. directory (str): The directory used for saving the resulting Caffe model. If None, nothing is saved to the disk. export_params (bool): If True, this function exports all the parameters included in the given model at the same time. If False, the exported Caffe model doesn't include any parameter values. graph_name (str): A string to be used for the ``name`` field of the graph in the exported Caffe model. .. note:: Currently, this function supports networks that created by following layer functions. - :func:`~chainer.functions.linear` - :func:`~chainer.functions.convolution_2d` - :func:`~chainer.functions.deconvolution_2d` - :func:`~chainer.functions.max_pooling_2d` - :func:`~chainer.functions.average_pooling_2d` - :func:`~chainer.functions.batch_normalization` - :func:`~chainer.functions.local_response_normalization` - :func:`~chainer.functions.relu` - :func:`~chainer.functions.leaky_relu` - :func:`~chainer.functions.concat` - :func:`~chainer.functions.softmax` - :func:`~chainer.functions.reshape` - :func:`~chainer.functions.add` This function can export at least following networks. - GoogLeNet - ResNet - VGG And, this function use testing (evaluation) mode. .. admonition:: Example >>> from chainer.exporters import caffe >>> >>> class Model(chainer.Chain): ... def __init__(self): ... super(Model, self).__init__() ... with self.init_scope(): ... self.l1 = L.Convolution2D(None, 1, 1, 1, 0) ... self.b2 = L.BatchNormalization(1) ... self.l3 = L.Linear(None, 1) ... ... def __call__(self, x): ... h = F.relu(self.l1(x)) ... h = self.b2(h) ... return self.l3(h) ... >>> x = chainer.Variable(np.zeros((1, 10, 10, 10), np.float32)) >>> caffe.export(Model(), [x], None, True, 'test') """ assert isinstance(args, (tuple, list)) if len(args) != 1: raise NotImplementedError() for i in args: assert isinstance(i, variable.Variable) with function.force_backprop_mode(), chainer.using_config('train', False): output = model(*args) if isinstance(output, variable.Variable): output = [output] assert isinstance(output, (tuple, list)) for i in output: assert isinstance(i, variable.Variable) prototxt = None caffemodel = None if directory is not None: prototxt = os.path.join(directory, 'chainer_model.prototxt') if export_params: caffemodel = os.path.join(directory, 'chainer_model.caffemodel') retriever = _RetrieveAsCaffeModel(prototxt, caffemodel) retriever(graph_name, args, output)
18,050
35.763747
79
py
DeSpaWN
DeSpaWN-main/Script_DeSpaWN.py
# -*- coding: utf-8 -*- """ Title: Fully Learnable Deep Wavelet Transform for Unsupervised Monitoring of High-Frequency Time Series ------ (DeSpaWN) Description: -------------- Toy script to showcase the deep neural network DeSpaWN. Please cite the corresponding paper: Michau, G., Frusque, G., & Fink, O. (2022). Fully learnable deep wavelet transform for unsupervised monitoring of high-frequency time series. Proceedings of the National Academy of Sciences, 119(8). Version: 1.0 -------- @author: Dr. Gabriel Michau, -------- Chair of Intelligent Maintenance Systems ETH Zürich Created on 15.01.2022 Licence: ---------- MIT License Copyright (c) 2022 Dr. Gabriel Michau Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ # Usual packages import numpy as np import pandas as pd import matplotlib.pyplot as plt import tensorflow as tf # designed with 2.1.0 /!\ models output changes with tf>2.2.0 import tensorflow.keras as keras from lib import despawn # Load a toy time series data to run DeSPAWN signal = pd.read_csv("monthly-sunspots.csv") lTrain = 2000 # length of the training section signalT = ((signal['Sunspots']-signal['Sunspots'].mean())/signal['Sunspots'].std()).values[np.newaxis,:,np.newaxis,np.newaxis] signal = signalT[:,:lTrain,:,:] # Number of decomposition level is max log2 of input TS level = np.floor(np.log2(signal.shape[1])).astype(int) # Train hard thresholding (HT) coefficient? trainHT = True # Initialise HT value initHT = 0.3 # Which loss to consider for wavelet coeffs ('l1' or None) lossCoeff='l1' # Weight for sparsity loss versus residual? lossFactor = 1.0 # Train wavelets? (Trainable kernels) kernTrainable = True # Which training mode? # cf (https://arxiv.org/pdf/2105.00899.pdf -- https://doi.org/10.1073/pnas.2106598119) [Section 4.4 Ablation Study] # CQF => learn wavelet 0 infer all other kernels from the network # PerLayer => learn one wavelet per level, infer others # PerFilter => learn wavelet + scaling function per level + infer other # Free => learn everything mode = 'PerLayer' # QMF PerLayer PerFilter Free # Initialise wavelet kernel (here db-4) kernelInit = np.array([-0.010597401785069032, 0.0328830116668852, 0.030841381835560764, -0.18703481171909309, -0.027983769416859854, 0.6308807679298589, 0.7148465705529157, 0.2303778133088965]) epochs = 1000 verbose = 2 # Set sparsity (dummy) loss: def coeffLoss(yTrue,yPred): return lossFactor*tf.reduce_mean(yPred,keepdims=True) # Set residual loss: def recLoss(yTrue,yPred): return tf.math.abs(yTrue-yPred) keras.backend.clear_session() # generates two models: # model1 outputs the reconstructed signals and the loss on the wavelet coefficients # model2 outputs the reconstructed signals and wavelet coefficients model1,model2 = despawn.createDeSpaWN(inputSize=None, kernelInit=kernelInit, kernTrainable=kernTrainable, level=level, lossCoeff=lossCoeff, kernelsConstraint=mode, initHT=initHT, trainHT=trainHT) opt = keras.optimizers.Nadam(learning_rate=0.001, beta_1=0.9, beta_2=0.999, epsilon=1e-07, name="Nadam") # For the training we only use model1 model1.compile(optimizer=opt, loss=[recLoss, coeffLoss]) # the sparsity term has no ground truth => just input an empty numpy array as ground truth (anything would do, in coeffLoss, yTrue is not called) H = model1.fit(signal,[signal,np.empty((signal.shape[0]))], epochs=epochs, verbose=verbose) # Examples for plotting the model outputs and learnings indPlot = 0 out = model1.predict(signal) outC = model2.predict(signal) # Test part of the signal outTe = model1.predict(signalT[:,lTrain:,:,:]) outCTe = model2.predict(signalT[:,lTrain:,:,:]) fig = plt.figure(1) fig.clf() ax = fig.add_subplot(2,1,1) ax.plot(np.arange(signal.shape[1]),signal[indPlot,:,0,0]) ax.plot(np.arange(signal.shape[1]),out[0][indPlot,:,0,0]) ax.plot(np.arange(signal.shape[1],signalT.shape[1]),signalT[indPlot,lTrain:,0,0]) ax.plot(np.arange(signal.shape[1],signalT.shape[1]),outTe[0][indPlot,:,0,0]) ax.legend(['Train Original','Train Reconstructed','Test Original', 'Test Reconstructed']) ax = fig.add_subplot(2,2,3) idpl = 0 for e,o in enumerate(outC[1:]): ax.boxplot(np.abs(np.squeeze(o[indPlot,:,:,:])), positions=[e], widths=0.8) ax.set_xlabel('Decomposition Level') ax.set_ylabel('Coefficient Distribution') trainYLim = ax.get_ylim() trainXLim = ax.get_xlim() ax = fig.add_subplot(2,2,4) idpl = 0 for e,o in enumerate(outCTe[1:]): print(o.shape[1]) if o.shape[1]>1: ax.boxplot(np.abs(np.squeeze(o[indPlot,:,:,:])), positions=[e], widths=0.8) else: ax.plot(e,np.abs(np.squeeze(o[indPlot,:,:,:])),'o',color='k') ax.set_xlabel('Decomposition Level') ax.set_ylabel('Coefficient Distribution') ax.set_ylim(trainYLim) ax.set_xlim(trainXLim)
5,817
38.849315
195
py
DeSpaWN
DeSpaWN-main/lib/despawnLayers.py
# -*- coding: utf-8 -*- """ Title: Fully Learnable Deep Wavelet Transform for Unsupervised Monitoring of High-Frequency Time Series ------ (DeSpaWN) Description: -------------- Function to generate the layers used in DeSpaWN TF model. Please cite the corresponding paper: Michau, G., Frusque, G., & Fink, O. (2022). Fully learnable deep wavelet transform for unsupervised monitoring of high-frequency time series. Proceedings of the National Academy of Sciences, 119(8). Version: 1.0 -------- @author: Dr. Gabriel Michau, -------- Chair of Intelligent Maintenance Systems ETH Zürich Created on 15.01.2022 Licence: ---------- MIT License Copyright (c) 2022 Dr. Gabriel Michau Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ import tensorflow as tf class Kernel(tf.keras.layers.Layer): def __init__(self, kernelInit=8, trainKern=True, **kwargs): self.trainKern = trainKern if isinstance(kernelInit,int): self.kernelSize = kernelInit self.kernelInit = 'random_normal' else: self.kernelSize = kernelInit.__len__() self.kernelInit = tf.constant_initializer(kernelInit) super(Kernel, self).__init__(**kwargs) def build(self, input_shape): self.kernel = self.add_weight(shape = (self.kernelSize,1,1,1), initializer = self.kernelInit, trainable = self.trainKern, name='kernel') super(Kernel, self).build(input_shape) def call(self, inputs): return self.kernel class LowPassWave(tf.keras.layers.Layer): """ Layer that performs a convolution between its two inputs with stride (2,1) """ def __init__(self, **kwargs): super(LowPassWave, self).__init__(**kwargs) def build(self, input_shape): super(LowPassWave, self).build(input_shape) def call(self, inputs): return tf.nn.conv2d(inputs[0], inputs[1], padding="SAME", strides=(2, 1)) class HighPassWave(tf.keras.layers.Layer): """ Layer that performs a convolution between its two inputs with stride (2,1). Performs first the reverse alternative flip on the second inputs """ def __init__(self, **kwargs): super(HighPassWave, self).__init__(**kwargs) def build(self, input_shape): self.qmfFlip = tf.reshape(tf.Variable([(-1)**(i) for i in range(input_shape[1][0])], dtype='float32', name='mask', trainable=False),(-1,1,1,1)) super(HighPassWave, self).build(input_shape) def call(self, inputs): # print(self.qmfFlip) return tf.nn.conv2d(inputs[0], tf.math.multiply(tf.reverse(inputs[1],[0]),self.qmfFlip), padding="SAME", strides=(2, 1)) class LowPassTrans(tf.keras.layers.Layer): """ Layer that performs a convolution transpose between its two inputs with stride (2,1). The third input specifies the size of the reconstructed signal (to make sure it matches the decomposed one) """ def __init__(self, **kwargs): super(LowPassTrans, self).__init__(**kwargs) def build(self, input_shape): super(LowPassTrans, self).build(input_shape) def call(self, inputs): return tf.nn.conv2d_transpose(inputs[0], inputs[1], inputs[2], padding="SAME", strides=(2, 1)) class HighPassTrans(tf.keras.layers.Layer): """ Layer that performs a convolution transpose between its two inputs with stride (2,1). Performs first the reverse alternative flip on the second inputs The third input specifies the size of the reconstructed signal (to make sure it matches the decomposed one) """ def __init__(self, **kwargs): super(HighPassTrans, self).__init__(**kwargs) def build(self, input_shape): self.qmfFlip = tf.reshape(tf.Variable([(-1)**(i) for i in range(input_shape[1][0])], dtype='float32', name='mask', trainable=False),(-1,1,1,1)) super(HighPassTrans, self).build(input_shape) def call(self, inputs): return tf.nn.conv2d_transpose(inputs[0], tf.math.multiply(tf.reverse(inputs[1],[0]),self.qmfFlip), inputs[2], padding="SAME", strides=(2, 1)) class HardThresholdAssym(tf.keras.layers.Layer): """ Learnable Hard-thresholding layers """ def __init__(self, init=None, trainBias=True, **kwargs): if isinstance(init,float) or isinstance(init,int): self.init = tf.constant_initializer(init) else: self.init = 'ones' self.trainBias = trainBias super(HardThresholdAssym, self).__init__(**kwargs) def build(self, input_shape): self.thrP = self.add_weight(shape = (1,1,1,1), initializer=self.init, trainable = self.trainBias, name='threshold+') self.thrN = self.add_weight(shape = (1,1,1,1), initializer=self.init, trainable = self.trainBias, name='threshold-') super(HardThresholdAssym, self).build(input_shape) def call(self, inputs): return tf.math.multiply(inputs,tf.math.sigmoid(10*(inputs-self.thrP))+\ tf.math.sigmoid(-10*(inputs+self.thrN)))
6,326
40.084416
111
py