code
stringlengths
101
5.91M
.parametrize('statement_type,value', [(stmt.IntPrimitiveStatement, 42), (stmt.FloatPrimitiveStatement, 42.23), (stmt.StringPrimitiveStatement, 'foo'), (stmt.BytesPrimitiveStatement, b'foo'), (stmt.BooleanPrimitiveStatement, True), (stmt.ComplexPrimitiveStatement, (4 + 1j)), (stmt.ClassPrimitiveStatement, 0)]) def test_primitive_statement_equals_same(statement_type, default_test_case, value): statement = statement_type(default_test_case, value) assert statement.__eq__(statement)
.skipif((not _ti_core.GGUI_AVAILABLE), reason='GGUI Not Available') _utils.test(arch=[ti.vulkan]) def test_multi_windows(): window = ti.ui.Window('x', (128, 128), vsync=True, show_window=False) window2 = ti.ui.Window('x2', (128, 128), vsync=True, show_window=False)
class VideoDataset(Dataset): def __init__(self, train, query, gallery, seq_len=15, sample_method='evenly', **kwargs): super(VideoDataset, self).__init__(train, query, gallery, **kwargs) self.seq_len = seq_len self.sample_method = sample_method if (self.transform is None): raise RuntimeError('transform must not be None') def __getitem__(self, index): (img_paths, pid, camid, dsetid) = self.data[index] num_imgs = len(img_paths) if (self.sample_method == 'random'): indices = np.arange(num_imgs) replace = (False if (num_imgs >= self.seq_len) else True) indices = np.random.choice(indices, size=self.seq_len, replace=replace) indices = np.sort(indices) elif (self.sample_method == 'evenly'): if (num_imgs >= self.seq_len): num_imgs -= (num_imgs % self.seq_len) indices = np.arange(0, num_imgs, (num_imgs / self.seq_len)) else: indices = np.arange(0, num_imgs) num_pads = (self.seq_len - num_imgs) indices = np.concatenate([indices, (np.ones(num_pads).astype(np.int32) * (num_imgs - 1))]) assert (len(indices) == self.seq_len) elif (self.sample_method == 'all'): indices = np.arange(num_imgs) else: raise ValueError('Unknown sample method: {}'.format(self.sample_method)) imgs = [] for index in indices: img_path = img_paths[int(index)] img = read_image(img_path) if (self.transform is not None): img = self.transform(img) img = img.unsqueeze(0) imgs.append(img) imgs = torch.cat(imgs, dim=0) item = {'img': imgs, 'pid': pid, 'camid': camid, 'dsetid': dsetid} return item def show_summary(self): num_train_pids = self.get_num_pids(self.train) num_train_cams = self.get_num_cams(self.train) num_query_pids = self.get_num_pids(self.query) num_query_cams = self.get_num_cams(self.query) num_gallery_pids = self.get_num_pids(self.gallery) num_gallery_cams = self.get_num_cams(self.gallery) print('=> Loaded {}'.format(self.__class__.__name__)) print(' ') print(' subset | # ids | # tracklets | # cameras') print(' ') print(' train | {:5d} | {:11d} | {:9d}'.format(num_train_pids, len(self.train), num_train_cams)) print(' query | {:5d} | {:11d} | {:9d}'.format(num_query_pids, len(self.query), num_query_cams)) print(' gallery | {:5d} | {:11d} | {:9d}'.format(num_gallery_pids, len(self.gallery), num_gallery_cams)) print(' ')
def success_probability(F: float) -> float: return (((F ** 2) + (((2 * F) * (1 - F)) / 3)) + (5 * (((1 - F) / 3) ** 2)))
def is_interactive_execution(): is_interactive = None try: get_ipython().__class__.__name__ is_interactive = True except: is_interactive = False return is_interactive
def create_stats_ordered_dict(name, data, stat_prefix=None, always_show_all_stats=True, exclude_max_min=False): if (stat_prefix is not None): name = '{}{}'.format(stat_prefix, name) if isinstance(data, Number): return OrderedDict({name: data}) if (len(data) == 0): return OrderedDict() if isinstance(data, tuple): ordered_dict = OrderedDict() for (number, d) in enumerate(data): sub_dict = create_stats_ordered_dict('{0}_{1}'.format(name, number), d) ordered_dict.update(sub_dict) return ordered_dict if isinstance(data, list): try: iter(data[0]) except TypeError: pass else: data = np.concatenate(data) if (isinstance(data, np.ndarray) and (data.size == 1) and (not always_show_all_stats)): return OrderedDict({name: float(data)}) stats = OrderedDict([((name + ' Mean'), np.mean(data)), ((name + ' Std'), np.std(data))]) if (not exclude_max_min): stats[(name + ' Max')] = np.max(data) stats[(name + ' Min')] = np.min(data) return stats
class PythonCreoConnection(): def __init__(self, creo_exe_path, no_graphic_op=True, no_input_op=True): pythoncom.CoInitialize() self.model = None self.models = [] self.creo_exe_path = creo_exe_path self.no_graphic_op = (' -g:no_graphics' if no_graphic_op else '') self.no_input_op = (' -i:rpc_input' if no_input_op else '') print('Trying to form connection') self.asyncconn = Dispatch('pfcls.pfcAsyncConnection') print('Asyncconn worked') self.conn = self.asyncconn.Start(f'{self.creo_exe_path}{self.no_graphic_op}{self.no_input_op}', '.') print('Connection formed') self.session = self.conn.Session print('Session started') self.session.SetConfigOption('mass_property_calculate', 'automatic') print('Connection formed') def __enter__(self): return self def open_file(self, path): options = Dispatch('pfcls.pfcRetrieveModelOptions') o = options.Create() file = Dispatch('pfcls.pfcModelDescriptor') path = re.sub('\\.prt(\\.[0-9]+)', '.prt', path) f = file.CreateFromFilename(path) self.model = self.session.RetrieveModelWithOpts(f, o) self.models.append(self.model) self.session.SetConfigOption('regen_failure_handling', 'no_resolve_mode') self.session.OpenFile(f) return self.model def activate_window(self, model_id): self.window = self.session.GetModelWindow(self.models[model_id]) self.window.Activate() def close_window(self): self.window.Close() def set_parameter(self, mdl, param_name, value): param = self.get_param_obj(mdl, param_name) modelitem = Dispatch('pfcls.MpfcModelItem') if isinstance(value, bool): val = modelitem.CreateBoolParamValue(value) elif isinstance(value, (float, int)): val = modelitem.CreateDoubleParamValue(value) else: raise CreoWrapperError('Invalid value type') param.SetScaledValue(val, None) def assign_paramset(self, mdl, paramset): for (param, value) in paramset.items(): self.set_parameter(mdl, param, value) def regenerate(self, mdl): try: self.session.SetConfigOption('regen_failure_handling', 'resolve_mode') instrs = Dispatch('pfcls.pfcRegenInstructions') regen_instructions = instrs.Create(False, True, None) mdl.Regenerate(regen_instructions) except Exception: raise CreoWrapperError('Model failed to regenerate') def get_param_obj(mdl, param_name): param_obj = mdl.GetParam(param_name) try: paramvalue = param_obj.value return param_obj except AttributeError: raise CreoWrapperError(f'Parameter {param_name} not found') def get_param(self, mdl, param_name): param_obj = self.get_param_obj(mdl, param_name) return self._export_param(param_obj) def _export_param(param_obj): param_data_type = param_obj.Value.discr if (param_data_type == 0): param_value = param_obj.Value.StringValue elif (param_data_type == 1): param_value = param_obj.Value.IntValue elif (param_data_type == 2): param_value = param_obj.Value.BoolValue elif (param_data_type == 3): param_value = param_obj.Value.DoubleValue else: raise ValueError return {param_obj.Name: {'description': param_obj.Description, 'value': param_value}} def get_param_list(self, mdl, skip_default=True): param_list = {} param_list_obj = mdl.ListParams() no_params = param_list_obj.Count for idx in range(no_params): param_item = self._export_param(param_list_obj.Item(idx)) if (skip_default and (list(param_item.keys())[0] in ['DESCRIPTION', 'MODELED_BY', 'PTC_MATERIAL_NAME'])): continue param_list.update(param_item) return param_list def save_model(mdl, filepath): instrs = Dispatch('pfcls.pfcSTLASCIIExportInstructions') export_instructions = instrs.Create(None) mdl.Export(filepath, export_instructions) def close(self): self.conn.End() def __exit__(self, exc_type, exc_value, traceback): for i in range(len(self.models)): try: window = self.session.GetModelWindow(self.models[i]) window.Activate() window.Close() except: pass self.close() print('Connection closed')
class PC(BaseTabularAlgo, BaseTabularAlgoFull): def __init__(self, data: TabularData, prior_knowledge: Optional[PriorKnowledge]=None, CI_test: Union[(PartialCorrelation, KCI, DiscreteCI_tests)]=PartialCorrelation(), use_multiprocessing: Optional[bool]=False, **kargs): BaseTabularAlgo.__init__(self, data=data, prior_knowledge=prior_knowledge, CI_test=CI_test, use_multiprocessing=use_multiprocessing) BaseTabularAlgoFull.__init__(self, **kargs) def run(self, pvalue_thres: float=0.05, max_condition_set_size: Optional[int]=None) -> Tuple[(Dict[(Union[(int, str)], ResultInfoTabularFull)], List[Tuple[(Union[(int, str)], Union[(int, str)])]])]: self.start() separation_sets = {} graph = {key: [] for key in self.data.var_names} for name in self.data.var_names: pc_d = PCSingle(self.data, prior_knowledge=self.prior_knowledge, CI_test=self.CI_test) result = pc_d.run(target_var=name, pvalue_thres=pvalue_thres, max_condition_set_size=max_condition_set_size, full_cd=True) graph[name] = [p for p in result['undirected_edges']] separation_sets[name] = pc_d.separation_set_dict del result['undirected_edges'] self.result[name] = result self.stop() self.separation_sets = separation_sets self.skeleton = deepcopy(graph) graph = orient_edges(graph, separation_sets, self.data.var_names) graph = self.prior_knowledge.post_process_tabular(graph) for name in self.data.var_names: remove_node_list = [p for p in self.result[name]['value_dict'].keys() if (p not in graph[name])] for node in remove_node_list: del self.result[name]['value_dict'][node] del self.result[name]['pvalue_dict'][node] for name in self.data.var_names: self.result[name]['parents'] = list(set(graph[name])) return self.result def get_parents(self, pvalue_thres: float=0.05, target_var: Optional[Union[(int, str)]]=None) -> Dict[(Union[(int, str)], Tuple[Union[(int, str)]])]: parents = BaseTabularAlgoFull.get_parents(self, pvalue_thres=pvalue_thres, target_var=target_var) return parents
class VGG(nn.Module): def __init__(self, features, num_class=100): super().__init__() self.features = features self.classifier = nn.Sequential(nn.Linear(512, 4096), nn.ReLU(inplace=True), nn.Dropout(), nn.Linear(4096, 4096), nn.ReLU(inplace=True), nn.Dropout(), nn.Linear(4096, num_class)) def forward(self, x): output = self.features(x) output = output.view(output.size()[0], (- 1)) output = self.classifier(output) return output
def data_process(config): mode = config['mode'] assert (mode in ('term', 'category')) base_path = config['base_path'] raw_train_path = os.path.join(base_path, 'raw/train.xml') raw_val_path = os.path.join(base_path, 'raw/val.xml') raw_test_path = os.path.join(base_path, 'raw/test.xml') lowercase = config['lowercase'] if (mode == 'term'): train_data = parse_sentence_term(raw_train_path, lowercase=lowercase) val_data = parse_sentence_term(raw_val_path, lowercase=lowercase) test_data = parse_sentence_term(raw_test_path, lowercase=lowercase) else: train_data = parse_sentence_category(raw_train_path, lowercase=lowercase) val_data = parse_sentence_category(raw_val_path, lowercase=lowercase) test_data = parse_sentence_category(raw_test_path, lowercase=lowercase) remove_list = ['conflict'] train_data = category_filter(train_data, remove_list) val_data = category_filter(val_data, remove_list) test_data = category_filter(test_data, remove_list) (word2index, index2word) = build_vocab(train_data, max_size=config['max_vocab_size'], min_freq=config['min_vocab_freq']) if (not os.path.exists(os.path.join(base_path, 'processed'))): os.makedirs(os.path.join(base_path, 'processed')) if (mode == 'term'): save_term_data(train_data, word2index, os.path.join(base_path, 'processed/train.npz')) save_term_data(val_data, word2index, os.path.join(base_path, 'processed/val.npz')) save_term_data(test_data, word2index, os.path.join(base_path, 'processed/test.npz')) else: save_category_data(train_data, word2index, os.path.join(base_path, 'processed/train.npz')) save_category_data(val_data, word2index, os.path.join(base_path, 'processed/val.npz')) save_category_data(test_data, word2index, os.path.join(base_path, 'processed/test.npz')) glove = load_glove(config['glove_path'], len(index2word), word2index) sentiment_matrix = load_sentiment_matrix(config['glove_path'], config['sentiment_path']) np.save(os.path.join(base_path, 'processed/glove.npy'), glove) np.save(os.path.join(base_path, 'processed/sentiment_matrix.npy'), sentiment_matrix) with open(os.path.join(base_path, 'processed/word2index.pickle'), 'wb') as handle: pickle.dump(word2index, handle) with open(os.path.join(base_path, 'processed/index2word.pickle'), 'wb') as handle: pickle.dump(index2word, handle) analyze = (analyze_term if (mode == 'term') else analyze_category) log = {'vocab_size': len(index2word), 'oov_size': (len(word2index) - len(index2word)), 'train_data': analyze(train_data), 'val_data': analyze(val_data), 'test_data': analyze(test_data), 'num_categories': 3} if (not os.path.exists(os.path.join(base_path, 'log'))): os.makedirs(os.path.join(base_path, 'log')) with open(os.path.join(base_path, 'log/log.yml'), 'w') as handle: yaml.safe_dump(log, handle, encoding='utf-8', allow_unicode=True, default_flow_style=False)
class ColumnPlaceholder(): def __init__(self, id_in_pattern, attributes): self.id_in_pattern = id_in_pattern self.attributes = attributes self.column = None def attach_to_column(self, column): self.column = column
.utils.register_keras_serializable() class FactorizationMachineLayer(tf.keras.layers.Layer): def __init__(self, field_dims, factors, kernel_initializer: Union[(Text, tf.keras.initializers.Initializer)]='truncated_normal', kernel_regularizer: Union[(Text, None, tf.keras.regularizers.Regularizer)]=None, **kwargs): super().__init__(**kwargs) self.embedding = Embedding(field_dims, factors, kernel_initializer, kernel_regularizer) self.linear = Linear(field_dims, tf.initializers.zeros()) self._supports_masking = True def call(self, inputs: tf.Tensor, training=False) -> tf.Tensor: linear = self.linear(inputs, training) second_order = tf.expand_dims(self.embedding(inputs, training), axis=(- 1)) return (linear + second_order) def get_config(self): config = {'use_bias': self._use_bias, 'kernel_initializer': tf.keras.initializers.serialize(self._kernel_initializer), 'bias_initializer': tf.keras.initializers.serialize(self._bias_initializer), 'kernel_regularizer': tf.keras.regularizers.serialize(self._kernel_regularizer), 'bias_regularizer': tf.keras.regularizers.serialize(self._bias_regularizer)} base_config = super().get_config() return dict((list(base_config.items()) + list(config.items())))
class GoogleSearchGetSearchHistory(VirtualFunctionTool): name = 'GoogleSearchGetSearchHistory' summary = "Retrieve the user's search history." parameters: List[ArgParameter] = [{'name': 'time_range', 'type': 'string', 'description': "The time range to get the search history, in the format of 'YYYY-MM-DD..YYYY-MM-DD'. Default is the past 7 days.", 'required': False}] returns: List[ArgReturn] = [{'name': 'history', 'type': 'array', 'description': "The user's search history, each includes 'keyword', 'time' (in the format of 'YYYY-MM-DD:HH:MM:SS'), and 'type' (one of ['web', 'image', 'video', 'news'])."}] exceptions: List[ArgException] = [{'name': 'InvalidRequestException', 'description': "The 'time_range' is invalid."}]
def sample_initial_states(rng): state = zika.State().values() state *= rng.uniform(low=0.95, high=1.05, size=state.shape) return zika.State(*state)
.parametrize('exposed_cases, exposed_total, control_cases, control_total, expected_rr', [(1, 4, 3, 8, (0.25 / 0.375)), (0, 10, 5, 20, 0), (0, 10, 0, 20, np.nan), (5, 15, 0, 20, np.inf)]) def test_relative_risk(exposed_cases, exposed_total, control_cases, control_total, expected_rr): result = relative_risk(exposed_cases, exposed_total, control_cases, control_total) assert_allclose(result.relative_risk, expected_rr, rtol=1e-13)
('just_spaces') class JustSpacesWordSplitter(WordSplitter): def split_words(self, sentence: str) -> List[Token]: return [Token(t) for t in sentence.split()] def from_params(cls, params: Params) -> 'WordSplitter': params.assert_empty(cls.__name__) return cls()
class GaussianBlur(object): 'Gaussian blur augmentation in SimCLR def __init__(self, sigma=[0.1, 2.0]): self.sigma = sigma def __call__(self, x): sigma = random.uniform(self.sigma[0], self.sigma[1]) x = x.filter(ImageFilter.GaussianBlur(radius=sigma)) return x
class Schema(object): thisown = _swig_property((lambda x: x.this.own()), (lambda x, v: x.this.own(v)), doc='The membership flag') __repr__ = _swig_repr def __init__(self, *args): _snap.Schema_swiginit(self, _snap.new_Schema(*args)) __swig_destroy__ = _snap.delete_Schema def __add__(self, Val): return _snap.Schema___add__(self, Val) def __eq__(self, Vec): return _snap.Schema___eq__(self, Vec) def __lt__(self, Vec): return _snap.Schema___lt__(self, Vec) def GetMemUsed(self): return _snap.Schema_GetMemUsed(self) def GetMemSize(self): return _snap.Schema_GetMemSize(self) def Gen(self, *args): return _snap.Schema_Gen(self, *args) def GenExt(self, _ValT, _Vals): return _snap.Schema_GenExt(self, _ValT, _Vals) def IsExt(self): return _snap.Schema_IsExt(self) def Reserve(self, *args): return _snap.Schema_Reserve(self, *args) def Clr(self, DoDel=True, NoDelLim=(- 1)): return _snap.Schema_Clr(self, DoDel, NoDelLim) def Trunc(self, _Vals=(- 1)): return _snap.Schema_Trunc(self, _Vals) def Reduce(self, _Vals=(- 1)): return _snap.Schema_Reduce(self, _Vals) def Pack(self): return _snap.Schema_Pack(self) def MoveFrom(self, Vec): return _snap.Schema_MoveFrom(self, Vec) def CopyUniqueFrom(self, Vec, Offset, Sz): return _snap.Schema_CopyUniqueFrom(self, Vec, Offset, Sz) def Empty(self): return _snap.Schema_Empty(self) def Len(self): return _snap.Schema_Len(self) def Reserved(self): return _snap.Schema_Reserved(self) def Last(self, *args): return _snap.Schema_Last(self, *args) def LastValN(self): return _snap.Schema_LastValN(self) def LastLast(self, *args): return _snap.Schema_LastLast(self, *args) def GetRndVal(self, *args): return _snap.Schema_GetRndVal(self, *args) def BegI(self): return _snap.Schema_BegI(self) def EndI(self): return _snap.Schema_EndI(self) def GetI(self, ValN): return _snap.Schema_GetI(self, ValN) def Add(self, *args): return _snap.Schema_Add(self, *args) def AddMP(self, Val): return _snap.Schema_AddMP(self, Val) def MoveLastMP(self, Val, Inc): return _snap.Schema_MoveLastMP(self, Val, Inc) def AddV(self, ValV): return _snap.Schema_AddV(self, ValV) def AddSorted(self, Val, Asc=True, _MxVals=(- 1)): return _snap.Schema_AddSorted(self, Val, Asc, _MxVals) def AddBackSorted(self, Val, Asc): return _snap.Schema_AddBackSorted(self, Val, Asc) def AddMerged(self, Val): return _snap.Schema_AddMerged(self, Val) def AddVMerged(self, ValV): return _snap.Schema_AddVMerged(self, ValV) def AddUnique(self, Val): return _snap.Schema_AddUnique(self, Val) def GetVal(self, *args): return _snap.Schema_GetVal(self, *args) def SetVal(self, ValN, Val): return _snap.Schema_SetVal(self, ValN, Val) def GetSubValV(self, BValN, EValN, ValV): return _snap.Schema_GetSubValV(self, BValN, EValN, ValV) def Ins(self, ValN, Val): return _snap.Schema_Ins(self, ValN, Val) def Del(self, *args): return _snap.Schema_Del(self, *args) def DelLast(self): return _snap.Schema_DelLast(self) def DelIfIn(self, Val): return _snap.Schema_DelIfIn(self, Val) def DelAll(self, Val): return _snap.Schema_DelAll(self, Val) def PutAll(self, Val): return _snap.Schema_PutAll(self, Val) def Swap(self, *args): return _snap.Schema_Swap(self, *args) def SwapI(LVal, RVal): return _snap.Schema_SwapI(LVal, RVal) SwapI = staticmethod(SwapI) def NextPerm(self): return _snap.Schema_NextPerm(self) def PrevPerm(self): return _snap.Schema_PrevPerm(self) def GetPivotValN(self, LValN, RValN): return _snap.Schema_GetPivotValN(self, LValN, RValN) def BSort(self, MnLValN, MxRValN, Asc): return _snap.Schema_BSort(self, MnLValN, MxRValN, Asc) def ISort(self, MnLValN, MxRValN, Asc): return _snap.Schema_ISort(self, MnLValN, MxRValN, Asc) def Partition(self, MnLValN, MxRValN, Asc): return _snap.Schema_Partition(self, MnLValN, MxRValN, Asc) def QSort(self, MnLValN, MxRValN, Asc): return _snap.Schema_QSort(self, MnLValN, MxRValN, Asc) def Sort(self, Asc=True): return _snap.Schema_Sort(self, Asc) def IsSorted(self, Asc=True): return _snap.Schema_IsSorted(self, Asc) def Shuffle(self, Rnd): return _snap.Schema_Shuffle(self, Rnd) def Reverse(self, *args): return _snap.Schema_Reverse(self, *args) def Merge(self): return _snap.Schema_Merge(self) def Intrs(self, *args): return _snap.Schema_Intrs(self, *args) def Union(self, *args): return _snap.Schema_Union(self, *args) def Diff(self, *args): return _snap.Schema_Diff(self, *args) def IntrsLen(self, ValV): return _snap.Schema_IntrsLen(self, ValV) def UnionLen(self, ValV): return _snap.Schema_UnionLen(self, ValV) def Count(self, Val): return _snap.Schema_Count(self, Val) def SearchBin(self, *args): return _snap.Schema_SearchBin(self, *args) def SearchBinLeft(self, Val, InsValN): return _snap.Schema_SearchBinLeft(self, Val, InsValN) def SearchForw(self, Val, BValN=0): return _snap.Schema_SearchForw(self, Val, BValN) def SearchBack(self, Val): return _snap.Schema_SearchBack(self, Val) def SearchVForw(self, ValV, BValN=0): return _snap.Schema_SearchVForw(self, ValV, BValN) def IsIn(self, *args): return _snap.Schema_IsIn(self, *args) def IsInBin(self, Val): return _snap.Schema_IsInBin(self, Val) def GetDat(self, Val): return _snap.Schema_GetDat(self, Val) def GetAddDat(self, Val): return _snap.Schema_GetAddDat(self, Val) def GetMxValN(self): return _snap.Schema_GetMxValN(self) def GetV(*args): return _snap.Schema_GetV(*args) GetV = staticmethod(GetV)
def copy_dlls(outdir): libdir = None libnames = None libneeded = ('libgfortran', 'libgcc_s', 'libquadmath') liboptional = ('libwinpthread',) for p in os.environ['PATH'].split(';'): if (not os.path.isdir(p)): continue libnames = [f for f in os.listdir(p) if (f.lower().endswith('.dll') and f.startswith((libneeded + liboptional)))] if (len([f for f in libnames if f.startswith(libneeded)]) == len(libneeded)): libdir = p break if (libdir is None): raise RuntimeError("Can't locate runtime libraries.") if (not os.path.exists(outdir)): os.makedirs(outdir) for f in libnames: shutil.copy(os.path.join(libdir, f), outdir) return [os.path.join(outdir, f) for f in libnames]
def _combining_class(cp): v = unicodedata.combining(unichr(cp)) if (v == 0): if (not unicodedata.name(unichr(cp))): raise ValueError('Unknown character in unicodedata') return v
class NeveuSchwarzLieConformalAlgebra(GradedLieConformalAlgebra): def __init__(self, R): nsdict = {('L', 'L'): {0: {('L', 1): 1}, 1: {('L', 0): 2}, 3: {('C', 0): R(2).inverse_of_unit()}}, ('L', 'G'): {0: {('G', 1): 1}, 1: {('G', 0): (R(3) * R(2).inverse_of_unit())}}, ('G', 'G'): {0: {('L', 0): 2}, 2: {('C', 0): (R(2) * R(3).inverse_of_unit())}}} from sage.rings.rational_field import QQ weights = (2, QQ((3, 2))) parity = (0, 1) GradedLieConformalAlgebra.__init__(self, R, nsdict, names=('L', 'G'), central_elements=('C',), weights=weights, parity=parity) def _repr_(self): return 'The Neveu-Schwarz super Lie conformal algebra over {}'.format(self.base_ring())
class QAFactEvalEvaluator(): def __init__(self, device=0): os.makedirs('models/answering', exist_ok=True) if (not os.listdir('models/answering')): model_url = ' output_path = 'models/answering/model.zip' gdown.download(model_url, output_path, quiet=False) shutil.unpack_archive(output_path, 'models/answering') os.remove(output_path) if ((not os.path.exists('models/quip-512-mocha')) or (not os.listdir('models/quip-512-mocha'))): model_url = ' output_path = 'models/quip-512-mocha.tar.gz' try: _ = wget.download(model_url, output_path) except: raise Exception("Error while downloading 'quip-512-mocha'") shutil.unpack_archive(output_path, 'models/') os.remove(output_path) self.qafacteval = QAFactEval(cuda_device=device, use_lerc_quip=True, verbose=False, generation_batch_size=32, answering_batch_size=32, lerc_batch_size=8, lerc_quip_path=f'models/quip-512-mocha', answering_model_dir=f'models/answering') def evaluate(self, data, global_knowledge): sources = [] generateds = [] for sample in data: source = '\n'.join([get_instruction(sample), sample['knowledge'], retrieve_global_knowledge(sample['output'], global_knowledge)]) sources.append(source) generateds.append([sample['output']]) results = self.qafacteval.score_batch_qafacteval(sources, generateds, return_qa_pairs=True) scores = [(x[0]['qa-eval']['lerc_quip'] if x[1][0] else 0.5) for x in results] meta = [] for x in results: answer_list = [y for y in x[1][0] if (y['prediction']['f1'] <= 0.6)] meta.append(answer_list) return (scores, meta)
class SPDocVQA(Dataset): def __init__(self, imbd_dir, images_dir, split, kwargs): data = np.load(os.path.join(imbd_dir, 'new_imdb_{:s}.npy'.format(split)), allow_pickle=True) self.header = data[0] self.imdb = data[1:] self.hierarchical_method = kwargs.get('hierarchical_method', False) self.max_answers = 2 self.images_dir = images_dir self.use_images = kwargs.get('use_images', False) self.get_raw_ocr_data = kwargs.get('get_raw_ocr_data', False) def __len__(self): return len(self.imdb) def __getitem__(self, idx): record = self.imdb[idx] question = record['question'] context = ' '.join([word.lower() for word in record['ocr_tokens']]) context_page_corresp = [0 for ix in range(len(context))] answers = list(set((answer.lower() for answer in record['answers']))) if self.use_images: image_name = os.path.join(self.images_dir, '{:s}.png'.format(record['image_name'])) image = Image.open(image_name).convert('RGB') if self.get_raw_ocr_data: words = [word.lower() for word in record['ocr_tokens']] boxes = np.array([bbox for bbox in record['ocr_normalized_boxes']]) if self.hierarchical_method: words = [words] boxes = [boxes] image_name = [image_name] image = [image] (start_idxs, end_idxs) = self._get_start_end_idx(context, answers) sample_info = {'question_id': record['question_id'], 'questions': question, 'contexts': context, 'answers': answers, 'start_indxs': start_idxs, 'end_indxs': end_idxs} if self.use_images: sample_info['image_names'] = image_name sample_info['images'] = image if self.get_raw_ocr_data: sample_info['words'] = words sample_info['boxes'] = boxes sample_info['num_pages'] = 1 sample_info['answer_page_idx'] = 0 else: sample_info['context_page_corresp'] = context_page_corresp sample_info['start_indxs'] = start_idxs sample_info['end_indxs'] = end_idxs return sample_info def _get_start_end_idx(self, context, answers): answer_positions = [] for answer in answers: start_idx = context.find(answer) if (start_idx != (- 1)): end_idx = (start_idx + len(answer)) answer_positions.append([start_idx, end_idx]) if (len(answer_positions) > 0): (start_idx, end_idx) = random.choice(answer_positions) else: (start_idx, end_idx) = (0, 0) return (start_idx, end_idx)
def depthwise_conv2d(_input, kernel_size, stride=1, padding='SAME', param_initializer=None): in_features = int(_input.get_shape()[3]) if (not param_initializer): param_initializer = {} output = _input with tf.variable_scope('conv'): init_key = ('%s/weight' % tf.get_variable_scope().name) initializer = param_initializer.get(init_key, tf.contrib.layers.variance_scaling_initializer()) weight = tf.get_variable(name='weight', shape=[kernel_size, kernel_size, in_features, 1], initializer=initializer) if (padding == 'SAME'): pad = (kernel_size // 2) paddings = tf.constant([[0, 0], [pad, pad], [pad, pad], [0, 0]]) output = tf.pad(output, paddings, 'CONSTANT') output = tf.nn.depthwise_conv2d(output, weight, [1, stride, stride, 1], 'VALID', data_format='NHWC') return output
def metric_max_over_ground_truths(metric_fn, prediction, ground_truths): scores_for_ground_truths = [] for ground_truth in ground_truths: score = metric_fn(prediction, ground_truth) scores_for_ground_truths.append(score) return max(scores_for_ground_truths)
class ICNet(nn.Module): def __init__(self, num_classes): super(ICNet, self).__init__() self.conv_sub1 = nn.Sequential(Conv3x3BNReLU(3, 32, 2), Conv3x3BNReLU(32, 32, 2), Conv3x3BNReLU(32, 64, 2)) self.backbone = Backbone() self.ppm = PyramidPoolingModule() self.cff_12 = CascadeFeatureFusion(128, 64, 128, num_classes) self.cff_24 = CascadeFeatureFusion(2048, 512, 128, num_classes) self.conv_cls = nn.Conv2d(128, num_classes, 1, bias=False) def forward(self, x): x_sub1 = self.conv_sub1(x) x_sub2 = F.interpolate(x, scale_factor=0.5, mode='bilinear') (_, x_sub2, _, _) = self.backbone(x_sub2) x_sub4 = F.interpolate(x, scale_factor=0.25, mode='bilinear') (_, _, _, x_sub4) = self.backbone(x_sub4) x_sub4 = self.ppm(x_sub4) outs = list() (x_cff_24, x_24_cls) = self.cff_24(x_sub4, x_sub2) outs.append(x_24_cls) (x_cff_12, x_12_cls) = self.cff_12(x_cff_24, x_sub1) outs.append(x_12_cls) up_x2 = F.interpolate(x_cff_12, scale_factor=2, mode='bilinear') up_x2 = self.conv_cls(up_x2) outs.append(up_x2) up_x8 = F.interpolate(up_x2, scale_factor=4, mode='bilinear') outs.append(up_x8) outs.reverse() return outs
def normalize_words(target_words_batch, predicted_words_batch, glm_alternatives=None): excluded_words = ['<UNK>', 'UH', 'UM', 'EH', 'MM', 'HM', 'AH', 'HUH', 'HA', 'ER', 'OOF', 'HEE', 'ACH', 'EEE', 'EW'] target_words_batch = expand_contractions_batch(target_words_batch) predicted_words_batch = expand_contractions_batch(predicted_words_batch) alternative2tgt_word_batch = [] for tgt_utterance in target_words_batch: alternative2tgt_word = defaultdict(str) if (glm_alternatives is not None): for tgt_wrd in tgt_utterance: alts = glm_alternatives[tgt_wrd] for alt in alts: if ((alt != tgt_wrd) and (len(alt) > 0)): alternative2tgt_word[alt] = tgt_wrd alternative2tgt_word_batch.append(alternative2tgt_word) checked_predicted_words_batch = [] for (i, pred_utterance) in enumerate(predicted_words_batch): alternative2tgt_word = alternative2tgt_word_batch[i] checked_predicted_words = [] for pred_wrd in pred_utterance: pred_wrd = re.sub('\\[.*?\\]', '', pred_wrd) pred_wrd = pred_wrd.translate(str.maketrans('', '', string.punctuation)) if ((pred_wrd != 'LAUGHTER') and pred_wrd.endswith('LAUGHTER')): pred_wrd = pred_wrd.replace('LAUGHTER', '') if ((pred_wrd != 'NOISE') and pred_wrd.endswith('NOISE')): pred_wrd = pred_wrd.replace('NOISE', '') if pred_wrd.endswith('VOCALIZED'): pred_wrd = pred_wrd.replace('VOCALIZED', '') if (pred_wrd in excluded_words): continue tgt_wrd = alternative2tgt_word[pred_wrd] if (len(tgt_wrd) > 0): pred_wrd = tgt_wrd if (len(pred_wrd) > 0): checked_predicted_words.append(pred_wrd) checked_predicted_words_batch.append(checked_predicted_words) return (target_words_batch, checked_predicted_words_batch)
class LossOutput(): loss: LossRecord reconstruction_loss: (LossRecord | None) = None kl_local: (LossRecord | None) = None kl_global: (LossRecord | None) = None classification_loss: (LossRecord | None) = None logits: (Tensor | None) = None true_labels: (Tensor | None) = None extra_metrics: (dict[(str, Tensor)] | None) = field(default_factory=dict) n_obs_minibatch: (int | None) = None reconstruction_loss_sum: Tensor = field(default=None, init=False) kl_local_sum: Tensor = field(default=None, init=False) kl_global_sum: Tensor = field(default=None, init=False) def __post_init__(self): self.loss = self.dict_sum(self.loss) if ((self.n_obs_minibatch is None) and (self.reconstruction_loss is None)): raise ValueError('Must provide either n_obs_minibatch or reconstruction_loss') default = (0 * self.loss) if (self.reconstruction_loss is None): self.reconstruction_loss = default if (self.kl_local is None): self.kl_local = default if (self.kl_global is None): self.kl_global = default self.reconstruction_loss = self._as_dict('reconstruction_loss') self.kl_local = self._as_dict('kl_local') self.kl_global = self._as_dict('kl_global') self.reconstruction_loss_sum = self.dict_sum(self.reconstruction_loss).sum() self.kl_local_sum = self.dict_sum(self.kl_local).sum() self.kl_global_sum = self.dict_sum(self.kl_global) if ((self.reconstruction_loss is not None) and (self.n_obs_minibatch is None)): rec_loss = self.reconstruction_loss self.n_obs_minibatch = list(rec_loss.values())[0].shape[0] if ((self.classification_loss is not None) and ((self.logits is None) or (self.true_labels is None))): raise ValueError('Must provide `logits` and `true_labels` if `classification_loss` is provided.') def dict_sum(dictionary: (dict[(str, Tensor)] | Tensor)): if isinstance(dictionary, dict): return sum(dictionary.values()) else: return dictionary def extra_metrics_keys(self) -> Iterable[str]: return self.extra_metrics.keys() def _as_dict(self, attr_name: str): attr = getattr(self, attr_name) if isinstance(attr, dict): return attr else: return {attr_name: attr}
def mathematica_console(readline=True): from sage.repl.rich_output.display_manager import get_display_manager if (not get_display_manager().is_in_terminal()): raise RuntimeError('Can use the console only in the terminal. Try %%mathematica magics instead.') if (not readline): os.system('math') return else: os.system('math-readline') return
_function def fully_qualified_name(filename): module = os.path.splitext(os.path.basename(filename))[0] return '.'.join((package(filename) + (module,)))
def get_split(split_name, dataset_dir, file_pattern=None, reader=None): if (split_name not in _SPLITS_TO_SIZES): raise ValueError(('split name %s was not recognized.' % split_name)) if (not file_pattern): file_pattern = _FILE_PATTERN file_pattern = os.path.join(dataset_dir, (file_pattern % split_name)) if (reader is None): reader = tf.TFRecordReader keys_to_features = {'image/encoded': tf.FixedLenFeature((), tf.string, default_value=''), 'image/format': tf.FixedLenFeature((), tf.string, default_value='raw'), 'image/class/label': tf.FixedLenFeature([1], tf.int64, default_value=tf.zeros([1], dtype=tf.int64))} items_to_handlers = {'image': slim.tfexample_decoder.Image(shape=[28, 28, 1], channels=1), 'label': slim.tfexample_decoder.Tensor('image/class/label', shape=[])} decoder = slim.tfexample_decoder.TFExampleDecoder(keys_to_features, items_to_handlers) labels_to_names = None if dataset_utils.has_labels(dataset_dir): labels_to_names = dataset_utils.read_label_file(dataset_dir) return slim.dataset.Dataset(data_sources=file_pattern, reader=reader, decoder=decoder, num_samples=_SPLITS_TO_SIZES[split_name], num_classes=_NUM_CLASSES, items_to_descriptions=_ITEMS_TO_DESCRIPTIONS, labels_to_names=labels_to_names)
def binary_search_tree_shape(w, left_to_right=True): if (not w): return BinaryTree() if left_to_right: root = w[0] else: root = w[(- 1)] left = [x for x in w if (x < root)] right = [x for x in w if (x > root)] return BinaryTree([binary_search_tree_shape(left, left_to_right), binary_search_tree_shape(right, left_to_right)])
class TestMiniAlexNet(test_util.TestCase): def _MiniAlexNetNoDropout(self, order): model = model_helper.ModelHelper(name='alexnet') conv1 = brew.conv(model, 'data', 'conv1', 3, 16, 11, ('XavierFill', {}), ('ConstantFill', {}), stride=4, pad=0) relu1 = brew.relu(model, conv1, 'relu1') norm1 = brew.lrn(model, relu1, 'norm1', size=5, alpha=0.0001, beta=0.75) pool1 = brew.max_pool(model, norm1, 'pool1', kernel=3, stride=2) conv2 = brew.group_conv(model, pool1, 'conv2', 16, 32, 5, ('XavierFill', {}), ('ConstantFill', {'value': 0.1}), group=2, stride=1, pad=2) relu2 = brew.relu(model, conv2, 'relu2') norm2 = brew.lrn(model, relu2, 'norm2', size=5, alpha=0.0001, beta=0.75) pool2 = brew.max_pool(model, norm2, 'pool2', kernel=3, stride=2) conv3 = brew.conv(model, pool2, 'conv3', 32, 64, 3, ('XavierFill', {'std': 0.01}), ('ConstantFill', {}), pad=1) relu3 = brew.relu(model, conv3, 'relu3') conv4 = brew.group_conv(model, relu3, 'conv4', 64, 64, 3, ('XavierFill', {}), ('ConstantFill', {'value': 0.1}), group=2, pad=1) relu4 = brew.relu(model, conv4, 'relu4') conv5 = brew.group_conv(model, relu4, 'conv5', 64, 32, 3, ('XavierFill', {}), ('ConstantFill', {'value': 0.1}), group=2, pad=1) relu5 = brew.relu(model, conv5, 'relu5') pool5 = brew.max_pool(model, relu5, 'pool5', kernel=3, stride=2) fc6 = brew.fc(model, pool5, 'fc6', 1152, 1024, ('XavierFill', {}), ('ConstantFill', {'value': 0.1})) relu6 = brew.relu(model, fc6, 'relu6') fc7 = brew.fc(model, relu6, 'fc7', 1024, 1024, ('XavierFill', {}), ('ConstantFill', {'value': 0.1})) relu7 = brew.relu(model, fc7, 'relu7') fc8 = brew.fc(model, relu7, 'fc8', 1024, 5, ('XavierFill', {}), ('ConstantFill', {'value': 0.0})) pred = brew.softmax(model, fc8, 'pred') xent = model.LabelCrossEntropy([pred, 'label'], 'xent') loss = model.AveragedLoss([xent], ['loss']) model.AddGradientOperators([loss]) return model def _testMiniAlexNet(self, order): model = self._MiniAlexNetNoDropout(order) workspace.ResetWorkspace() workspace.RunNetOnce(model.param_init_net) inputs = dict([(str(name), workspace.FetchBlob(str(name))) for name in model.params]) if (order == 'NCHW'): inputs['data'] = np.random.rand(4, 3, 227, 227).astype(np.float32) else: inputs['data'] = np.random.rand(4, 227, 227, 3).astype(np.float32) inputs['label'] = np.array([1, 2, 3, 4]).astype(np.int32) cpu_device = caffe2_pb2.DeviceOption() cpu_device.device_type = caffe2_pb2.CPU gpu_device = caffe2_pb2.DeviceOption() gpu_device.device_type = workspace.GpuDeviceType checker = device_checker.DeviceChecker(0.05, [cpu_device, gpu_device]) ret = checker.CheckNet(model.net.Proto(), inputs, ignore=['_pool1_idx', '_pool2_idx', '_pool5_idx']) self.assertEqual(ret, True) ((not workspace.has_gpu_support), 'No GPU support. Skipping test.') def testMiniAlexNetNCHW(self): self._testMiniAlexNet('NCHW')
def plot_pit_histogram(predicted, observed, **kwargs): plt.bar(x=predicted[1:], height=np.diff(observed), width=(- np.diff(predicted)), align='edge', fill=False, edgecolor='black', **kwargs) plt.xlim((0, 1)) plt.xlabel('Probability Integral Transform') plt.ylabel('Density') plt.axhline((1.0 / (len(predicted) - 1)), linestyle='--', color='grey') plt.title('PIT Histogram')
.parametrize('seed', [313]) .parametrize('ctx, func_name', ctxs) def test_log_forward_backward(seed, ctx, func_name): from nbla_test_utils import cap_ignore_region, function_tester rng = np.random.RandomState(seed) inputs = [np.clip((np.abs(rng.randn(2, 3, 4).astype(np.float32)) * 10000.0), 0.01, 10000.0)] function_tester(rng, F.log, np.log, inputs, atol_f=1e-06, atol_b=0.01, ctx=ctx, func_name=func_name)
_utils.test(arch=get_host_arch_list()) def test_python_scope_matrix_operations(): for ops in matrix_operation_types: (a, b) = test_matrix_arrays[:2] (m1, m2) = (ti.Matrix(a), ti.Matrix(b)) c = ops(m1, m2) assert np.allclose(c.to_numpy(), ops(a, b))
class Networks3D(object): def __init__(self): self.Conv3D_Block = Layers2D().Conv2D_Block self.Conv3DTranspose_Block = Layers2D().Conv2DTranspose_Block self.Residual3D_Block = Layers2D().Residual3D_Block def build_patch_discriminator3D(self, model_shape, filters=32, k_size=4, drop=True, rate=0.5, summary=False, model_file=None, name='gan_d_'): if model_file: model = self.build_pretrained_model(model_file) if summary: model.summary() return model else: init = RandomNormal(stddev=0.02) n_rows = model_shape[0] n_cols = model_shape[1] n_layers = model_shape[2] c_dims = model_shape[3] input_shape = (n_rows, n_cols, n_layers, c_dims) input_layer = Input(shape=input_shape, name=(name + 'input')) d = self.Conv3D_Block(input_layer, filters, name=(name + '1'), bn=False) d = self.Conv3D_Block(d, (2 * filters), name=(name + '2')) d = self.Conv3D_Block(d, (4 * filters), name=(name + '3')) d = self.Conv3D_Block(d, (8 * filters), strides=1, name=(name + '4')) feat = Flatten(name=(name + 'flatten'))(d) if drop: d = Dropout(rate=0.5, name=(name + '_dropout'))(d, training=True) logits = Conv3D(1, 4, strides=1, padding='same', kernel_initializer=init, name=(name + 'logits'))(d) out = Activation('sigmoid', name=(name + 'sigmoid'))(logits) model = Model(inputs=[input_layer], outputs=[out, logits, feat], name=('Discriminator_' + name[(- 3):])) if summary: model.summary() return model def build_resnet_generator3D(self, model_shape, filters=32, k_size=3, last_act='tanh', summary=False, model_file=None, name='gan_g_'): if model_file: model = self.utils.build_pretrained_model(model_file) if summary: model.summary() return model else: init = RandomNormal(stddev=0.02) n_rows = model_shape[0] n_cols = model_shape[1] n_layers = model_shape[2] in_c_dims = model_shape[3] out_c_dims = model_shape[4] (n_rows_e1, n_rows_e2, n_rows_e4, n_rows_e8) = ((n_rows // 1), (n_rows // 2), (n_rows // 4), (n_rows // 8)) rows_matching = np.equal([(2 * n_rows_e2), (2 * n_rows_e4), (2 * n_rows_e8)], [n_rows_e1, n_rows_e2, n_rows_e4]) index_rows = np.where(np.logical_not(rows_matching))[0] (n_cols_e1, n_cols_e2, n_cols_e4, n_cols_e8) = ((n_cols // 1), (n_cols // 2), (n_cols // 4), (n_cols // 8)) cols_matching = np.equal([(2 * n_cols_e2), (2 * n_cols_e4), (2 * n_cols_e8)], [n_cols_e1, n_cols_e2, n_cols_e4]) index_cols = np.where(np.logical_not(cols_matching))[0] input_shape = (n_rows, n_cols, in_c_dims) input_layer = Input(shape=input_shape, name=(name + '_input')) e1 = self.Conv3D_Block(input_layer, n_kernels=filters, k_size=7, strides=1, bn=False, name=(name + 'e1')) e2 = self.Conv3D_Block(e1, (2 * filters), k_size=k_size, bn_training=True, name=(name + 'e2')) e3 = self.Conv3D_Block(e2, (4 * filters), k_size=k_size, bn_training=True, name=(name + 'e3')) e4 = self.Conv3D_Block(e3, (8 * filters), k_size=k_size, bn=False, name=(name + 'e4')) rb1 = self.Residual3D_Block(e4, n_kernels=(8 * filters), k_size=k_size, bn_training=True, name=(name + '1_')) rb2 = self.Residual3D_Block(rb1, n_kernels=(8 * filters), k_size=k_size, bn_training=True, name=(name + '2_')) rb3 = self.Residual3D_Block(rb2, n_kernels=(8 * filters), k_size=k_size, bn_training=True, name=(name + '3_')) rb3 = Dropout(rate=0.5, name=(name + 'drop_1'))(rb3, training=True) rb4 = self.Residual3D_Block(rb3, n_kernels=(8 * filters), k_size=k_size, bn_training=True, name=(name + '4_')) rb4 = Dropout(rate=0.5, name=(name + 'drop_2'))(rb4, training=True) rb5 = self.Residual3D_Block(rb4, n_kernels=(8 * filters), k_size=k_size, bn_training=True, name=(name + '5_')) rb5 = Dropout(rate=0.5, name=(name + 'drop_3'))(rb5, training=True) d1 = self.Conv3DTranspose_Block(rb5, (4 * filters), k_size=k_size, activation='linear', name=(name + 'd1')) if ((index_rows == 2) or (index_cols == 2)): d1 = BilinearUpsampling(output_size=((n_rows // 4), (n_cols // 4)), name=(name + '_bilinear'))(d1) d1 = Concatenate(name=(name + 'conc_1'))([d1, e3]) d1 = Activation('relu', name=(name + '_act_1'))(d1) d2 = self.Conv3DTranspose_Block(d1, (2 * filters), k_size=k_size, activation='linear', name=(name + 'd2')) if ((index_rows == 1) or (index_cols == 1)): d2 = BilinearUpsampling(output_size=((n_rows // 2), (n_cols // 2)), name=(name + '_bilinear'))(d2) d2 = Concatenate(name=(name + 'conc_2'))([d2, e2]) d2 = Activation('relu', name=(name + '_act_2'))(d2) d3 = self.Conv3DTranspose_Block(d2, (1 * filters), k_size=k_size, activation='linear', name=(name + 'd3')) if ((index_rows == 0) or (index_cols == 0)): d3 = BilinearUpsampling(output_size=(n_rows, n_cols), name=(name + '_bilinear'))(d2) d3 = Concatenate(name=(name + 'conc_3'))([d3, e1]) d3 = Activation('relu', name=(name + 'act_3'))(d3) output = Conv3DTranspose(out_c_dims, 7, strides=1, padding='same', kernel_initializer=init, name=(name + 'd_out'))(d3) output = Activation(last_act, name=(name + last_act))(output) model = Model(inputs=[input_layer], outputs=[output], name=('Generator' + name[(- 3):])) if summary: model.summary() return model
def parse_line(line, domain): tokens = line.split(' ')[1:] context = tokens[:(2 * domain.input_length())] choice_str = tokens[(- domain.selection_length()):] (cnts, vals) = domain.parse_context(context) picks = [] for (i, c) in enumerate(choice_str[:(domain.selection_length() // 2)]): if (c in ('<disconnect>', '<no_agreement>')): picks.append((- 1)) else: (idx, pick) = domain.parse_choice(c) assert (idx == i) picks.append(pick) return (cnts, vals, picks)
def rmse_loss(outputs, targets): return tf.sqrt(tf.reduce_mean(tf.square(tf.sub(targets, outputs))))
def test_multiple_in_connectors(): sdfg = dace.SDFG('mctest') sdfg.add_array('A', [1], dace.float64) state = sdfg.add_state() a = state.add_read('A') b = state.add_read('A') tasklet = state.add_tasklet('dosomething', {'a'}, {}, 'a * a') state.add_edge(a, None, tasklet, 'a', dace.Memlet('A[0]')) state.add_edge(b, None, tasklet, 'a', dace.Memlet('A[0]')) try: sdfg.validate() raise AssertionError('SDFG validates successfully, test failed!') except InvalidSDFGError as ex: print('Exception caught, test passed')
class FineModel(abc.ABC): controls: List[fenics.Function] cost_functional_value: float def __init__(self) -> None: pass def solve_and_evaluate(self) -> None: pass
def main(fuzzer, seed, output, group, program, argument, thread, command, cgroup_path='', scale_num=1): controller_class = str_to_class(f'{str.upper(fuzzer)}Controller') if (controller_class is None): print(f"{fuzzer} controller doesn't exist.") controller = controller_class(seed=os.path.realpath(seed), output=os.path.realpath(output), group=group, program=program, argument=argument, thread=thread, cgroup_path=cgroup_path) controller.init() command = command if (command == 'start'): controller.start() elif (command == 'stop'): controller.stop() elif (command == 'pause'): controller.pause() elif (command == 'resume'): controller.resume() elif (command == 'scale'): controller.scale(scale_num)
def test(): pred_sequences = [['O', 'S-LOC', 'O', 'O', 'B-PER', 'E-PER'], ['O', 'S-MISC', 'O', 'E-ORG', 'O', 'B-PER', 'I-PER', 'E-PER']] gold_sequences = [['O', 'B-LOC', 'E-LOC', 'O', 'B-PER', 'E-PER'], ['O', 'S-MISC', 'B-ORG', 'E-ORG', 'O', 'B-PER', 'E-PER', 'S-LOC']] print(score_by_token(pred_sequences, gold_sequences)) print(score_by_entity(pred_sequences, gold_sequences))
_test() def test_vector_reduce(): N.set(24) X = np.random.rand(N.get()).astype(dace.float32.type) s = dace.scalar(dace.float32) sdfg = vector_reduce.to_sdfg() sdfg.apply_transformations(FPGATransformSDFG) sdfg(x=X, s=s, N=N) s_exp = 0.0 for x in X: s_exp += x diff = (np.linalg.norm((s_exp - s)) / N.get()) assert (diff <= 1e-05) return sdfg
class TestQuantizeDynamicJitPasses(QuantizationTestCase): def test_prepare_dynamic(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc = torch.nn.Linear(5, 5) def forward(self, x): return self.fc(x) model = torch.jit.script(M()) for qconfig in [float16_dynamic_qconfig, default_dynamic_qconfig]: m = prepare_dynamic_jit(model, {'': qconfig}) assert (len(attrs_with_prefix(m.fc, '_observer_')) == 1) if (qconfig == float16_dynamic_qconfig): observer_name = 'PlaceholderObserver = prim::GetAttr[name="_observer_' FileCheck().check(observer_name).run(m.fc.graph) else: assert (len(attrs_with_prefix(m, '_observer_')) == 1) observer_name = 'DynamicQuantObserver = prim::GetAttr[name="_observer_' FileCheck().check(observer_name).check('prim::GetAttr[name="fc"]').check('prim::CallMethod').check_not('Observer = prim::GetAttr[name="_observer_').run(m.graph) def test_prepare_dynamic_child_qconfig(self): class Sub(torch.nn.Module): def __init__(self): super(Sub, self).__init__() self.fc = torch.nn.Linear(5, 5) def forward(self, x): return self.fc(x) class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.conv = torch.nn.Conv2d(3, 5, 3) self.sub = Sub() def forward(self, x): return self.sub(self.conv(x)) m = torch.jit.script(M()) m = prepare_dynamic_jit(m, {'sub.fc': default_dynamic_qconfig}) assert (len(attrs_with_prefix(m, '_observer_')) == 1) assert (len(attrs_with_prefix(m.conv, '_observer_')) == 0) assert (len(attrs_with_prefix(m.sub, '_observer_')) == 0) assert (len(attrs_with_prefix(m.sub.fc, '_observer_')) == 1) FileCheck().check('prim::GetAttr[name="sub').check('prim::CallMethod').check('DynamicQuantObserver = prim::GetAttr[name="_observer_').check('prim::CallMethod').check_not('Observer = prim::GetAttr[name="_observer_').run(m.graph) def test_insert_quant_dequant_linear_dynamic(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc1 = torch.nn.Linear(5, 5).float() self.fc2 = torch.nn.Linear(5, 5).float() def forward(self, x): x = self.fc1(x) return self.fc2(x) for is_per_channel in [True, False]: m = torch.jit.script(M()) qconfig = (per_channel_dynamic_qconfig if (is_per_channel is True) else default_dynamic_qconfig) m = quantize_dynamic_jit(m, {'': qconfig}, debug=True) assert (len(m._modules._c.items()) == 2), 'Expected to have two submodule of linear' wt_quant_func = ('aten::quantize_per_channel' if is_per_channel else 'aten::quantize_per_tensor') act_quant_func = 'aten::quantize_per_tensor' FileCheck().check('aten::_choose_qparams_per_tensor').check_next(act_quant_func).check_next('aten::dequantize').check('aten::_choose_qparams_per_tensor').check_next(act_quant_func).check_next('aten::dequantize').check(wt_quant_func).check_next('aten::dequantize').check_not(wt_quant_func).check('return').run(m.graph) _qengines def test_dynamic_multi_op(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc1 = torch.nn.Linear(5, 5).to(dtype=torch.float) def forward(self, x): x = (x + 5) return self.fc1(x) x = torch.randn(5, 5) for tracing in [True, False]: model = self.checkGraphModeOp(M(), x, 'quantized::linear_dynamic', tracing=tracing, dynamic=True) FileCheck().check('aten::add').run(model.graph) _qengines def test_dynamic_quant_multi_uses(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc = torch.nn.Linear(5, 5).float() def forward(self, x): size1 = x.size() size2 = x.size() return (self.fc(x), size1, size2) x = torch.randn(5, 5) for tracing in [True, False]: model = self.checkGraphModeOp(M(), x, 'quantized::linear_dynamic', tracing=tracing, dynamic=True) FileCheck().check_not('aten::_choose_qparams_per_tensor').run(model.graph) _qengines def test_dynamic_shared_weights(self): class myMod(torch.nn.Module): def __init__(self, weight): super().__init__() self.linear = nn.Linear(5, 5) self.linear.weight = weight def forward(self, x): return self.linear(x) class DynamicModel(torch.nn.Module): def __init__(self): super(DynamicModel, self).__init__() self.weight = torch.nn.Parameter(torch.ones(5, 5)) self.mod1 = myMod(self.weight) def forward(self, x): y = self.mod1(x) z = torch.nn.functional.linear(y, self.weight) return z model = torch.jit.script(DynamicModel()).eval() data = torch.randn(5, 5, dtype=torch.float) quant_ops = ['mod1', ''] counts = [1, 2] for (op, count) in zip(quant_ops, counts): qconfig_dict = {op: default_dynamic_qconfig} m1 = quantize_dynamic_jit(model, qconfig_dict) out_graph = m1(data) FileCheck().check_count('quantized::linear_dynamic(', count, exactly=True).check_not('aten::_choose_qparams_per_tensor').run(m1.graph) m2 = prepare_dynamic_jit(model, qconfig_dict) m2(data) m2 = convert_dynamic_jit(m2, debug=False) out_ref = m2(data) self.assertEqual(out_graph, out_ref) _qengines def test_dynamic_with_if(self): class Res(torch.nn.Module): def __init__(self): super(Res, self).__init__() self.weight = torch.nn.Parameter(torch.ones(5, 5)) def forward(self, x, cond): if cond: return torch.nn.functional.linear(x, self.weight) else: return torch.nn.functional.linear(x, self.weight) class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.res1 = Res() self.res2 = Res() def forward(self, x): x = self.res1(x, True) x = self.res2(x, False) return x model = torch.jit.script(M()).eval() data = torch.randn(5, 5, dtype=torch.float) qconfig_dict = {'': default_dynamic_qconfig} for tracing in [True, False]: m1 = self.checkGraphModeOp(M(), data, 'quantized::linear_dynamic', tracing=tracing, dynamic=True) FileCheck().check_count('quantized::linear_dynamic(', 2, exactly=True).check_not('aten::_choose_qparams_per_tensor').run(m1.graph) ref_qparams = [] qconfig = script_qconfig(default_dynamic_qconfig) wt_module = wrap_cpp_module(qconfig.weight) for wt in [model.res1.weight, model.res2.weight]: wt_module(wt) qparams = wt_module.calculate_qparams() ref_qparams.append((qparams[0].item(), qparams[1].item())) m2 = quantize_dynamic_jit(model, qconfig_dict, debug=True) graph_params = [] for (x, obs) in m2._modules._c.items(): if (x == 'res1'): graph_params.append((obs.getattr('6_scale_0'), obs.getattr('6_zero_point_0'))) elif (x == 'res2'): graph_params.append((obs.getattr('10_scale_0'), obs.getattr('10_zero_point_0'))) self.assertEqual(ref_qparams, graph_params) def test_dynamic_weight_observer(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc = torch.nn.Linear(5, 5).float() self.fc2 = torch.nn.Linear(5, 5).float() def forward(self, x): x = self.fc(x) return self.fc2(x) qconfig_dict = {'': default_dynamic_qconfig} eager_model = M().eval() for tracing in [True, False]: x = torch.rand(5, 5) model = get_script_module(eager_model, tracing, x) ref_qparams = [] for wt in [model.fc.weight, model.fc2.weight]: wt_module = default_dynamic_qconfig.weight() wt_module(wt) qparams = wt_module.calculate_qparams() ref_qparams.append((qparams[0].item(), qparams[1].item())) model = quantize_dynamic_jit(model, qconfig_dict, debug=True) graph_qparams = [] for (x, obs) in model._modules._c.items(): graph_qparams.append((obs.getattr('3_scale_0'), obs.getattr('3_zero_point_0'))) self.assertEqual(ref_qparams, graph_qparams) def test_convert_dynamic_fp16(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc = torch.nn.Linear(5, 5) def forward(self, x): return self.fc(x) m = torch.jit.script(M()) m = quantize_dynamic_jit(m, {'': float16_dynamic_qconfig}, debug=True) FileCheck().check('aten::_saturate_weight_to_fp16').check('aten::linear').check_not('aten::dequantize').check_not('aten::quantize').run(m.graph) def test_quantize_dynamic_fp16(self): class M(torch.nn.Module): def __init__(self): super(M, self).__init__() self.fc = torch.nn.Linear(5, 5) def forward(self, x): return self.fc(x) m = torch.jit.script(M()) m = quantize_dynamic_jit(m, {'': float16_dynamic_qconfig}) FileCheck().check('quantized::linear_prepack_fp16').check_next('quantized::linear_dynamic_fp16').check_not('aten::linear').check_not('aten::dequantize').check_not('aten::quantize').run(m.graph)
class TestSummaryEncoder(unittest.TestCase): def setUp(self): self.df = pd.DataFrame({'categories': ['a', 'b', 'c', 'a', 'b', 'c', 'a', 'b']}) self.target = np.array([1, 2, 0, 4, 5, 0, 6, 7]) self.col = 'categories' def assert_same_quantile(self, quantile): quantile_results = encoders.QuantileEncoder(cols=[self.col], quantile=quantile).fit_transform(self.df, self.target) summary_results = encoders.SummaryEncoder(cols=[self.col], quantiles=[quantile]).fit_transform(self.df, self.target) percentile = round((quantile * 100)) col_name = ((str(self.col) + '_') + str(percentile)) np.testing.assert_allclose(quantile_results[self.col].values, summary_results[col_name].values) def test_several_quantiles(self): for quantile in [0.1, 0.5, 0.9]: self.assert_same_quantile(quantile) def test_several_quantiles_reverse(self): quantile_list = [0.2, 0.1, 0.8] summary_results = encoders.SummaryEncoder(cols=[self.col], quantiles=quantile_list).fit_transform(self.df, self.target) for quantile in quantile_list: quantile_results = encoders.QuantileEncoder(cols=[self.col], quantile=quantile).fit_transform(self.df, self.target) percentile = round((quantile * 100)) col_name = ((str(self.col) + '_') + str(percentile)) np.testing.assert_allclose(quantile_results[self.col].values, summary_results[col_name].values)
def _split_channels(num_feat, num_groups): split = [(num_feat // num_groups) for _ in range(num_groups)] split[0] += (num_feat - sum(split)) return split
class FmaDatasetInfo(DatasetInfo): name = 'fma' def preprocess_id_col(self, df: pd.DataFrame) -> pd.DataFrame: df[self.id_col] = df[self.id_col].apply((lambda x: f'{x:06}')) return df def id_to_filename(self, track_id: Union[(str, int)], dirname: Optional[str]=None): track_id = int(track_id) filename = f'{track_id:06}.wav' if dirname: filename = os.path.join(dirname, filename) return filename
class CTRLModelTest(CommonTestCases.CommonModelTester): all_model_classes = ((CTRLModel, CTRLLMHeadModel) if is_torch_available() else ()) test_pruning = False test_torchscript = False test_resize_embeddings = False test_head_masking = False class CTRLModelTester(object): def __init__(self, parent, batch_size=13, seq_length=7, is_training=True, use_token_type_ids=True, use_input_mask=True, use_labels=True, use_mc_token_ids=True, vocab_size=99, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, intermediate_size=37, hidden_act='gelu', hidden_dropout_prob=0.1, attention_probs_dropout_prob=0.1, max_position_embeddings=512, type_vocab_size=16, type_sequence_label_size=2, initializer_range=0.02, num_labels=3, num_choices=4, scope=None): self.parent = parent self.batch_size = batch_size self.seq_length = seq_length self.is_training = is_training self.use_token_type_ids = use_token_type_ids self.use_input_mask = use_input_mask self.use_labels = use_labels self.use_mc_token_ids = use_mc_token_ids self.vocab_size = vocab_size self.hidden_size = hidden_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.intermediate_size = intermediate_size self.hidden_act = hidden_act self.hidden_dropout_prob = hidden_dropout_prob self.attention_probs_dropout_prob = attention_probs_dropout_prob self.max_position_embeddings = max_position_embeddings self.type_vocab_size = type_vocab_size self.type_sequence_label_size = type_sequence_label_size self.initializer_range = initializer_range self.num_labels = num_labels self.num_choices = num_choices self.scope = scope def prepare_config_and_inputs(self): input_ids = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) input_mask = None if self.use_input_mask: input_mask = ids_tensor([self.batch_size, self.seq_length], vocab_size=2) token_type_ids = None if self.use_token_type_ids: token_type_ids = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size) mc_token_ids = None if self.use_mc_token_ids: mc_token_ids = ids_tensor([self.batch_size, self.num_choices], self.seq_length) sequence_labels = None token_labels = None choice_labels = None if self.use_labels: sequence_labels = ids_tensor([self.batch_size], self.type_sequence_label_size) token_labels = ids_tensor([self.batch_size, self.seq_length], self.num_labels) choice_labels = ids_tensor([self.batch_size], self.num_choices) config = CTRLConfig(vocab_size_or_config_json_file=self.vocab_size, n_embd=self.hidden_size, n_layer=self.num_hidden_layers, n_head=self.num_attention_heads, n_positions=self.max_position_embeddings, n_ctx=self.max_position_embeddings) head_mask = ids_tensor([self.num_hidden_layers, self.num_attention_heads], 2) return (config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels) def check_loss_output(self, result): self.parent.assertListEqual(list(result['loss'].size()), []) def create_and_check_ctrl_model(self, config, input_ids, input_mask, head_mask, token_type_ids, *args): model = CTRLModel(config=config) model.eval() model(input_ids, token_type_ids=token_type_ids, head_mask=head_mask) model(input_ids, token_type_ids=token_type_ids) (sequence_output, presents) = model(input_ids) result = {'sequence_output': sequence_output, 'presents': presents} self.parent.assertListEqual(list(result['sequence_output'].size()), [self.batch_size, self.seq_length, self.hidden_size]) self.parent.assertEqual(len(result['presents']), config.n_layer) def create_and_check_lm_head_model(self, config, input_ids, input_mask, head_mask, token_type_ids, *args): model = CTRLLMHeadModel(config) model.eval() (loss, lm_logits, _) = model(input_ids, token_type_ids=token_type_ids, labels=input_ids) result = {'loss': loss, 'lm_logits': lm_logits} self.parent.assertListEqual(list(result['loss'].size()), []) self.parent.assertListEqual(list(result['lm_logits'].size()), [self.batch_size, self.seq_length, self.vocab_size]) def prepare_config_and_inputs_for_common(self): config_and_inputs = self.prepare_config_and_inputs() (config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels) = config_and_inputs inputs_dict = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask} return (config, inputs_dict) def setUp(self): self.model_tester = CTRLModelTest.CTRLModelTester(self) self.config_tester = ConfigTester(self, config_class=CTRLConfig, n_embd=37) def test_config(self): self.config_tester.run_common_tests() def test_ctrl_model(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*config_and_inputs) def test_ctrl_lm_head_model(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*config_and_inputs) .slow def test_model_from_pretrained(self): cache_dir = '/tmp/transformers_test/' for model_name in list(CTRL_PRETRAINED_MODEL_ARCHIVE_MAP.keys())[:1]: model = CTRLModel.from_pretrained(model_name, cache_dir=cache_dir) shutil.rmtree(cache_dir) self.assertIsNotNone(model)
class TarballTestCase(unittest.TestCase): def test_tarball(self): pkg = Package('configure') tarball = Tarball(pkg.tarball_filename) self.assertEqual(tarball, pkg.tarball) self.assertEqual(pkg, tarball.package) with CapturedOutput() as (stdout, stderr): with CapturedLog() as _: tarball.download() self.assertEqual(stdout.getvalue(), '') self.assertTrue(tarball.checksum_verifies()) def test_checksum(self): pkg = Package('configure') tarball = pkg.tarball with CapturedOutput() as (stdout, stderr): with CapturedLog() as log: tarball.download() self.assertTrue(tarball.checksum_verifies()) with open(tarball.upstream_fqn, 'w') as f: f.write('foobar') self.assertFalse(tarball.checksum_verifies()) with CapturedOutput() as (stdout, stderr): with CapturedLog() as log: tarball.download() msg = log.messages() self.assertTrue((('INFO', 'Attempting to download package {0} from mirrors'.format(pkg.tarball_filename)) in msg)) self.assertEqual(stdout.getvalue(), '') self.assertEqual(stderr.getvalue(), '[]\n') self.assertTrue(tarball.checksum_verifies())
def to_pianoroll(music, colors): multitrack = music.to_pypianoroll() stacked = (multitrack.stack() > 0) colormatrix = np.array(colors[:len(music)]) reshaped = stacked.reshape(len(music), (- 1)) recolored = (255 - np.matmul((255 - colormatrix.T), reshaped)) clipped = np.round(np.clip(recolored, 0, 255)).astype(np.uint8) reshaped = clipped.reshape((3, stacked.shape[1], stacked.shape[2])) transposed = np.moveaxis(reshaped, 0, (- 1)) return np.flip(transposed, axis=1).transpose(1, 0, 2)
def build_image_encoder(config, direct_features=False, **kwargs): from mmf.modules.encoders import ImageEncoderFactory, ImageFeatureEncoderFactory if direct_features: module = ImageFeatureEncoderFactory(config) else: module = ImageEncoderFactory(config) return module.module
def do_analyse_snli(file_path, dev=True, use_loss=False, stop=None): results = [] with open(file_path, 'r', encoding='utf-8') as file: find_entry = False output = [0, 0.0, 0.0, 0.0, 0.0] for line in file: if (not find_entry): if line.startswith('data round'): output[0] = int(line.split(' ')[(- 4)].split(':')[(- 1)]) if ((stop is not None) and (output[0] > stop)): break if line.startswith('==> for dev'): output[1] = float(line.split(' ')[(- 1)]) output[2] = float(line.split(' ')[(- 3)][:(- 1)]) find_entry = True elif line.startswith('~~> for test'): output[3] = float(line.split(' ')[(- 1)]) output[4] = float(line.split(' ')[(- 3)][:(- 1)]) results.append(output) find_entry = False output = [0, 0.0, 0.0, 0.0, 0.0] if (len(results) > 0): print('max step:', results[(- 1)][0]) sort = (1 if dev else 3) if use_loss: sort += 1 output = list(sorted(results, key=(lambda elem: elem[sort]), reverse=(not use_loss))) for elem in output[:20]: print(('step: %d, dev: %.4f, dev_loss: %.4f, test: %.4f, test_loss: %.4f' % (elem[0], elem[1], elem[2], elem[3], elem[4])))
def save_pkl(data, file_path): with open(file_path, 'wb') as handle: pickle.dump(data, handle)
def BVRedAnd(a): if z3_debug(): _z3_assert(is_bv(a), 'First argument must be a Z3 bit-vector expression') return BitVecRef(Z3_mk_bvredand(a.ctx_ref(), a.as_ast()), a.ctx)
class ModelArguments(): model_name_or_path: Optional[str] = field(default=None, metadata={'help': "The model checkpoint for weights initialization.Don't set if you want to train a model from scratch."}) config_name: Optional[str] = field(default=None, metadata={'help': 'Pretrained config name or path if not the same as model_name'}) config_overrides: Optional[str] = field(default=None, metadata={'help': 'Override some existing default config settings when a model is trained from scratch. Example: n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index'}) tokenizer_name: Optional[str] = field(default=None, metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'}) cache_dir: Optional[str] = field(default=None, metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'}) use_fast_tokenizer: bool = field(default=True, metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'}) model_revision: str = field(default='main', metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'}) use_auth_token: bool = field(default=False, metadata={'help': 'Will use the token generated when running `transformers-cli login` (necessary to use this script with private models).'}) def __post_init__(self): if ((self.config_overrides is not None) and ((self.config_name is not None) or (self.model_name_or_path is not None))): raise ValueError("--config_overrides can't be used in combination with --config_name or --model_name_or_path")
class LayerPQC(LayeredPQC): def __init__(self, encoding_circuit: LayeredPQC): super().__init__(encoding_circuit.num_qubits, encoding_circuit.variable_groups) def add_operation(self, operation: _operation): self.operation_list.append(operation)
def load_data_to_gpu(batch_dict): for (key, val) in batch_dict.items(): try: if isinstance(val, np.ndarray): batch_dict[key] = torch.from_numpy(val).float().cuda() elif isinstance(val, torch.Tensor): batch_dict[key] = val.float().cuda() except: pass
.spark .parametrize('column', ['user_id']) .usefixtures('simple_dataframe') def test_label_encoder_spark(column, simple_dataframe): rule = LabelEncodingRule(column) encoder = LabelEncoder([rule]).fit(simple_dataframe) mapped_data = encoder.transform(simple_dataframe) rebuild_original_cols = encoder.inverse_transform(mapped_data).withColumn(column, F.col(column)) columns_order = ['user_id', 'item_id', 'timestamp'] df1 = simple_dataframe.orderBy(*columns_order).toPandas()[columns_order] df2 = rebuild_original_cols.orderBy(*columns_order).toPandas()[columns_order] pd.testing.assert_frame_equal(df1, df2)
.experimental def test_predict_pairs_k(log): model = ADMMSLIM(seed=SEED) model.fit(log) pairs_pred_k = model.predict_pairs(pairs=log.select('user_idx', 'item_idx'), log=log, k=1) pairs_pred = model.predict_pairs(pairs=log.select('user_idx', 'item_idx'), log=log, k=None) assert (pairs_pred_k.groupBy('user_idx').count().filter((sf.col('count') > 1)).count() == 0) assert (pairs_pred.groupBy('user_idx').count().filter((sf.col('count') > 1)).count() > 0)
def test_should_call_integrated_gradients_callback(random_data, convolutional_model, output_dir, mocker): mock_explainer = mocker.MagicMock(explain=mocker.MagicMock(return_value=np.zeros((28, 28)))) mocker.patch('tf_explain.callbacks.integrated_gradients.IntegratedGradients', return_value=mock_explainer) (images, labels) = random_data callbacks = [IntegratedGradientsCallback(validation_data=random_data, class_index=0, output_dir=output_dir, n_steps=3)] convolutional_model.fit(images, labels, batch_size=2, epochs=1, callbacks=callbacks) mock_explainer.explain.assert_called_once_with(random_data, convolutional_model, 0, 3) assert (len([_ for _ in output_dir.iterdir()]) == 1)
def _worker_run_collect(all_args): try: (collect_once, counter, lock, threshold, args) = all_args collected = [] while True: with lock: if (counter.value >= threshold): return collected (result, inc) = collect_once(singleton_pool.G, *args) collected.append(result) with lock: counter.value += inc if (counter.value >= threshold): return collected except Exception: raise Exception(''.join(traceback.format_exception(*sys.exc_info())))
class Up(nn.Module): def __init__(self, in_ch, out_ch): super(Up, self).__init__() self.up = nn.Sequential(nn.ConvTranspose2d(in_ch, out_ch, kernel_size=3, stride=2, padding=1, output_padding=1), nn.ReLU()) def forward(self, x): x = self.up(x) return x
class states_t(object): __slots__ = ['p'] def __init__(self, p=None, _skip_initialize=False): if _skip_initialize: return self.p = (Vector2d._default() if (p is None) else p) def from_all_fields(p): return states_t(p=p) def _skytype_meta(): return dict(type='struct', package='codegen_cpp_test', name='states_t') def _default(cls): return cls() def __repr__(self): return 'states_t({})'.format(', '.join(('{}={}'.format(name, repr(getattr(self, name))) for name in self.__slots__))) def __eq__(self, other): if (not isinstance(other, states_t)): return NotImplemented return (self.p == other.p) __hash__ = None def encode(self): buf = BytesIO() buf.write(states_t._get_packed_fingerprint()) self._encode_one(buf) return buf.getvalue() def _encode_one(self, buf): if hasattr(self.p, '_get_packed_fingerprint'): assert (self.p._get_packed_fingerprint() == Vector2d._get_packed_fingerprint()) else: assert (self.p._get_hash_recursive([]) == Vector2d._get_hash_recursive([])) self.p._encode_one(buf) def decode(data): if hasattr(data, 'read'): buf = T.cast(T.BinaryIO, data) else: buf = BytesIO(T.cast(bytes, data)) if (buf.read(8) != states_t._get_packed_fingerprint()): raise ValueError('Decode error') return states_t._decode_one(buf) def _decode_one(buf): self = states_t(_skip_initialize=True) self.p = Vector2d._decode_one(buf) return self def _get_hash_recursive(parents): if (states_t in parents): return 0 newparents = (parents + [states_t]) tmphash = (( + Vector2d._get_hash_recursive(newparents)) & ) tmphash = ((((tmphash << 1) & ) + (tmphash >> 63)) & ) return tmphash _packed_fingerprint = None def _get_packed_fingerprint(): if (states_t._packed_fingerprint is None): states_t._packed_fingerprint = struct.pack('>Q', states_t._get_hash_recursive([])) return states_t._packed_fingerprint def deepcopy(self, **kwargs): result = copy.deepcopy(self) for key in kwargs: if (not hasattr(result, key)): raise KeyError(('Type states_t does not have attribute: ' + str(key))) setattr(result, key, kwargs[key]) return result
class RingBuffer(object): def __init__(self, maxlen, shape, dtype=np.float32, data=None): self.maxlen = maxlen self.start = RawValue('L') self.length = RawValue('L') self.shape = shape if (data is None): self.data = np.zeros(((maxlen,) + shape), dtype=dtype) else: assert (data.shape == ((maxlen,) + shape)) assert (data.dtype == dtype) self.data = data def _get_state(self): end_idx = (self.start.value + self.length.value) indices = range(self.start.value, end_idx) return (self.start.value, self.length.value, self.data.take(indices, axis=0, mode='wrap')) def _set_state(self, start, length, data): self.start.value = start self.length.value = length self.data[:length] = data self.data = np.roll(self.data, start, axis=0) def __len__(self): return self.length.value def __getitem__(self, idx): if ((idx < 0) or (idx >= self.length.value)): raise KeyError() return self.data[((self.start.value + idx) % self.maxlen)] def get_batch(self, idxs): return self.data[((self.start.value + idxs) % self.length.value)] def append(self, var): if (self.length.value < self.maxlen): self.length.value += 1 elif (self.length.value == self.maxlen): self.start.value = ((self.start.value + 1) % self.maxlen) else: raise RuntimeError() self.data[(((self.start.value + self.length.value) - 1) % self.maxlen)] = var def _append_batch_with_space(self, var): len_batch = len(var) start_pos = ((self.start.value + self.length.value) % self.maxlen) self.data[start_pos:(start_pos + len_batch)] = var if (self.length.value < self.maxlen): self.length.value += len_batch assert (self.length.value <= self.maxlen), 'this should never happen!' else: self.start.value = ((self.start.value + len_batch) % self.maxlen) return np.arange(start_pos, (start_pos + len_batch)) def append_batch(self, var): len_batch = len(var) assert (len_batch < self.maxlen), 'trying to add a batch that is too big!' start_pos = ((self.start.value + self.length.value) % self.maxlen) if ((start_pos + len_batch) <= self.maxlen): idxs = self._append_batch_with_space(var) else: (first_batch, second_batch) = np.split(var, [(self.maxlen - start_pos)]) idxs1 = self._append_batch_with_space(first_batch) idxs2 = self._append_batch_with_space(second_batch) idxs = np.concatenate((idxs1, idxs2)) return idxs
class NonBottleneck1d(BaseModule): def __init__(self, channels, drop_rate=0, dilation=1, num_conv_layer=2, conv_cfg=None, norm_cfg=dict(type='BN', eps=0.001), act_cfg=dict(type='ReLU'), init_cfg=None): super(NonBottleneck1d, self).__init__(init_cfg=init_cfg) self.conv_cfg = conv_cfg self.norm_cfg = norm_cfg self.act_cfg = act_cfg self.act = build_activation_layer(self.act_cfg) self.convs_layers = nn.ModuleList() for conv_layer in range(num_conv_layer): first_conv_padding = ((1, 0) if (conv_layer == 0) else (dilation, 0)) first_conv_dilation = (1 if (conv_layer == 0) else (dilation, 1)) second_conv_padding = ((0, 1) if (conv_layer == 0) else (0, dilation)) second_conv_dilation = (1 if (conv_layer == 0) else (1, dilation)) self.convs_layers.append(build_conv_layer(self.conv_cfg, channels, channels, kernel_size=(3, 1), stride=1, padding=first_conv_padding, bias=True, dilation=first_conv_dilation)) self.convs_layers.append(self.act) self.convs_layers.append(build_conv_layer(self.conv_cfg, channels, channels, kernel_size=(1, 3), stride=1, padding=second_conv_padding, bias=True, dilation=second_conv_dilation)) self.convs_layers.append(build_norm_layer(self.norm_cfg, channels)[1]) if (conv_layer == 0): self.convs_layers.append(self.act) else: self.convs_layers.append(nn.Dropout(p=drop_rate)) def forward(self, input): output = input for conv in self.convs_layers: output = conv(output) output = self.act((output + input)) return output
class ComplexDoubleVectorSpace_class(FreeModule_ambient_field): def __init__(self, n): FreeModule_ambient_field.__init__(self, sage.rings.complex_double.CDF, n) def coordinates(self, v): return v
class SuperbASR(ASR): def default_config(self) -> dict: return dict(start=0, stop=None, target_dir=MISSING, cache_dir=None, remove_all_cache=False, prepare_data=dict(dataset_root=MISSING, train_sets=['train-clean-100'], valid_sets=['dev-clean'], test_sets=['test-clean']), prepare_tokenizer_data=dict(), build_tokenizer=dict(vocab_type='character'), build_dataset=dict(), build_batch_sampler=dict(train=dict(batch_size=32, max_length=2000, shuffle=True), valid=dict(batch_size=1), test=dict(batch_size=1)), build_upstream=dict(name=MISSING), build_featurizer=dict(layer_selections=None, normalize=False), build_downstream=dict(model_conf=dict(module='LSTM', proj_size=1024, hidden_size=[1024, 1024], dropout=[0.2, 0.2], layer_norm=[False, False], proj=[False, False], sample_rate=[1, 1], sample_style='concat', bidirectional=True), specaug_conf=dict(freq_mask_width_range=(0, 50), num_freq_mask=4, time_mask_width_range=(0, 40), num_time_mask=2)), build_model=dict(upstream_trainable=False), build_task=dict(log_metrics=['cer', 'wer']), build_optimizer=dict(name='Adam', conf=dict(lr=0.0001)), build_scheduler=dict(name='ExponentialLR', gamma=0.9), save_model=dict(extra_conf=dict(build_downstream_conf='${build_downstream}')), save_task=dict(), train=dict(total_steps=200000, log_step=100, eval_step=2000, save_step=500, gradient_clipping=1.0, gradient_accumulate=1, valid_metric='wer', valid_higher_better=False, auto_resume=True, resume_ckpt_dir=None)) def prepare_data(self, prepare_data: dict, target_dir: str, cache_dir: str, get_path_only: bool=False): return prepare_librispeech(**self._get_current_arguments(flatten_dict='prepare_data')) def prepare_tokenizer_data(self, prepare_tokenizer_data: dict, target_dir: str, cache_dir: str, train_csv: str, valid_csv: str, test_csvs: List[str], get_path_only: bool=False): tokenizer_data_name = f'{Path(train_csv).stem}.tokenizer_data' tokenizer_data_path = (Path(target_dir) / f'{tokenizer_data_name}.txt') if get_path_only: return tokenizer_data_path all_text = pd.read_csv(train_csv)['transcription'] with tokenizer_data_path.open('w') as f: f.writelines([f'''{line} ''' for line in all_text]) return tokenizer_data_path def build_tokenizer(self, build_tokenizer: dict, target_dir: str, cache_dir: str, tokenizer_data_path: str, get_path_only: bool=False): return prepare_common_tokenizer(**self._get_current_arguments(flatten_dict='build_tokenizer')) def build_dataset(self, build_dataset: dict, target_dir: str, cache_dir: str, mode: str, data_csv: str, tokenizer_path: str): csv = pd.read_csv(data_csv) audio_loader = LoadAudio(csv['wav_path'].tolist()) with open(tokenizer_path, 'rb') as f: tokenizer = pickle.load(f) text_encoder = EncodeText(csv['transcription'].tolist(), tokenizer) ids = csv['id'].tolist() class Speech2TextDataset(): def __len__(self): return len(audio_loader) def __getitem__(self, index: int): audio = audio_loader[index] text = text_encoder[index] return {'x': audio['wav'], 'x_len': audio['wav_len'], 'class_ids': text['class_ids'], 'labels': text['labels'], 'unique_name': ids[index]} dataset = Speech2TextDataset() return dataset def build_batch_sampler(self, build_batch_sampler: dict, target_dir: str, cache_dir: str, mode: str, data_csv: str, dataset: Dataset): class Config(): train: dict = None valid: dict = None test: dict = None conf = Config(**build_batch_sampler) if (mode == 'train'): wav_lens = get_info(dataset, ['x_len'], (Path(target_dir) / 'train_stats')) sampler = SortedBucketingSampler(wav_lens, **(conf.train or {})) elif (mode == 'valid'): sampler = FixedBatchSizeBatchSampler(dataset, **(conf.valid or {})) elif (mode == 'test'): sampler = FixedBatchSizeBatchSampler(dataset, **(conf.test or {})) return sampler def build_downstream(self, build_downstream: dict, downstream_input_size: int, downstream_output_size: int, downstream_input_stride: int): class Config(): model_conf: dict = None specaug_conf: dict = None conf = Config(**build_downstream) model = RNNEncoder(downstream_input_size, downstream_output_size, **(conf.model_conf or {})) downstream = ModelWithSpecaug(model, **(conf.specaug_conf or {})) return downstream
class Checkpoints(): def __init__(self, logger=None, checkpoint_dir=None, experiment_id=None): self.logger = logger super(Checkpoints, self).__init__() self.checkpoint_dir = checkpoint_dir self.experiment_id = experiment_id self.checkpoint_path = self.get_checkpoint() def get_checkpoint(self): checkpoint_path = os.path.join(self.checkpoint_dir, self.experiment_id) if (not os.path.exists(checkpoint_path)): os.makedirs(checkpoint_path) checkpoint_path = os.path.join(checkpoint_path, '{type}.pth') return checkpoint_path def load_checkpoint(self, save_path, model, optimizer=None): error_str = "Checkpoint '{}' not found" assert os.path.exists(save_path), error_str.format(save_path) checkpoint_state = torch.load(save_path, map_location='cpu') model.load_state_dict(checkpoint_state['model'], strict=False) if optimizer: optimizer.load_state_dict(checkpoint_state['optimizer']) self.logger.info('Checkpoint loaded from {}'.format(save_path)) return checkpoint_state['epoch'] def resume_checkpoint(self, model, optimizer): checkpoint_path = self.checkpoint_path.format('last') chkpnt = torch.load(checkpoint_path) model.load_state_dict(chkpnt['model'], strict=False) start_epoch = chkpnt['epoch'] optimizer.load_state_dict(chkpnt['optimizer']) self.logger.info('Resumed checkpoint: {}'.format('xxx')) return start_epoch def save_checkpoint(self, model, checkpoint_path, epoch=(- 1), type='last', optimizer=None): checkpoint_state = {'epoch': epoch, 'model': model.state_dict(), 'optimizer': optimizer.state_dict()} torch.save(checkpoint_state, checkpoint_path) if (type == 'best'): torch.save(model.state_dict(), checkpoint_path.replace('best', 'deploy')) self.logger.info('Checkpoint saved to {}'.format(checkpoint_path)) def autosave_checkpoint(self, model, epoch, type, optimizer): if is_main_process(): checkpoint_path = self.checkpoint_path.format(type=type) self.save_checkpoint(checkpoint_path=checkpoint_path, epoch=epoch, type=type, model=model, optimizer=optimizer)
class MixedPercisionActivationSearch2Bit(MixedPercisionActivationBaseTest): def __init__(self, unit_test): super().__init__(unit_test) self.expected_config = [2, 8, 2, 2] def get_kpi(self): return KPI(96, 768) def compare(self, quantized_models, float_model, input_x=None, quantization_info=None): self.verify_config(quantization_info.mixed_precision_cfg, self.expected_config)
.xfail def test_exception(): def func(): foo = 1 bar = 0 try: result = (0 / 0) except ZeroDivisionError: result = (foo + bar) return result assert (func() == 1) dummy_block = BasicBlock([]) return_block = BasicBlock([Instr('LOAD_FAST', arg='result'), Instr('RETURN_VALUE')]) try_block = BasicBlock([Instr('LOAD_FAST', arg='foo'), Instr('LOAD_FAST', arg='bar'), Instr('BINARY_TRUE_DIVIDE'), Instr('DUP_TOP'), Instr('LOAD_GLOBAL', arg='ZeroDivisionError'), Instr('COMPARE_OP', arg=Compare.EXC_MATCH), Instr('POP_JUMP_IF_FALSE', arg=dummy_block)]) except_block = BasicBlock([Instr('LOAD_FAST', arg='foo'), Instr('LOAD_FAST', arg='bar'), Instr('BINARY_ADD'), Instr('STORE_FAST', arg='result'), Instr('JUMP_FORWARD', arg=dummy_block)]) function_block = BasicBlock([Instr('LOAD_CONST', arg=1), Instr('STORE_FAST', arg='foo'), Instr('LOAD_CONST', arg=0), Instr('STORE_FAST', arg='bar')]) expected_instructions = [] expected_instructions.extend(return_block) expected_instructions.extend(except_block) expected_instructions.extend(function_block) expected_instructions.extend(try_block) sliced_instructions = slice_function_at_return(func) assert (len(sliced_instructions) == len(expected_instructions))
def test_jim21(): text = str(ak.with_parameter(ak.Array([1, 2, 3, None, [1], [1, 2], [1, 2, 3]]), 'wonky', 'string').type) print(text) parsedtype = deduce_type(text) assert (str(parsedtype) == text)
class DegenerateMetric(TensorField): def __init__(self, vector_field_module, name, signature=None, latex_name=None): TensorField.__init__(self, vector_field_module, (0, 2), name=name, latex_name=latex_name, sym=(0, 1)) ndim = self._ambient_domain.dimension() if (signature is None): signature = ((ndim - 1), 0, 1) else: try: for elt in signature: if ((elt < 0) or (not isinstance(elt, (int, Integer)))): raise ValueError('{} must be a positive integer'.format(elt)) if (elt > ndim): raise ValueError('{} must be less than {}'.format(elt, ndim)) sign = ((signature[0] + signature[1]) + signature[2]) if (sign != ndim): raise ValueError('{} is different from {}'.format(sign, ndim)) except TypeError: raise TypeError('signature must be an iterable') self._signature = (signature[0], signature[1], signature[2]) self._signature_pm = self._signature def _repr_(self): return self._final_repr((('degenerate metric ' + self._name) + ' ')) def _new_instance(self): return type(self)(self._vmodule, 'unnamed metric', signature=self._signature, latex_name='\\text{unnamed metric}') def signature(self): return self._signature def set(self, symbiform): if (not isinstance(symbiform, TensorField)): raise TypeError('the argument must be a tensor field') if (symbiform._tensor_type != (0, 2)): raise TypeError('the argument must be of tensor type (0,2)') if (symbiform._sym != ((0, 1),)): raise TypeError('the argument must be symmetric') if (not symbiform._domain.is_subset(self._domain)): raise TypeError(('the symmetric bilinear form is not defined ' + 'on the metric domain')) self._restrictions.clear() if isinstance(symbiform, TensorFieldParal): rst = self.restrict(symbiform._domain) rst.set(symbiform) else: for (dom, symbiform_rst) in symbiform._restrictions.items(): rst = self.restrict(dom) rst.set(symbiform_rst) def restrict(self, subdomain, dest_map=None): if (subdomain == self._domain): return self if (subdomain not in self._restrictions): resu = TensorField.restrict(self, subdomain, dest_map=dest_map) resu._signature = self._signature resu._signature_pm = self._signature_pm self._restrictions[subdomain] = resu return self._restrictions[subdomain] def determinant(self): return self._domain.zero_scalar_field() det = determinant
def setImgWidthElement(data, img_key, width_key, resize=True): try: img = data[img_key] except KeyError: logger.error(('No image data (key: %s)' % img_key)) return try: width = data[width_key] if (resize and (width < img.shape[1])): height = int(((width * img.shape[0]) / img.shape[1])) logger.debug(('Resize to (%d, %d)' % (width, height))) data[img_key] = cv2.resize(img, (width, height)) except KeyError: width = img.shape[1] data[width_key] = width
def log_gaussian(x, mu, sigma, reduce_sum=True): res = tfd.Normal(mu, sigma).log_prob(x) if reduce_sum: return tf.reduce_sum(res) else: return res
def render_dist(df: pd.DataFrame, x: str, typ: str, plot_width: int, plot_height: int) -> Figure: assert (typ in ['pdf', 'cdf']) tooltips = [(x, ''), (typ.upper(), f'{{{typ}}}'), ('Label', f'{{{typ}}}_label')] y_range = Range1d(0, (df[typ].max() * 1.01)) x_range = Range1d(0, (df[f'x_{typ}'].max() * 1.01)) fig = tweak_figure(Figure(x_range=x_range, y_range=y_range, plot_width=plot_width, plot_height=plot_height, title=' ', tools='hover', toolbar_location=None, tooltips=tooltips)) for (idx, label) in enumerate(LABELS): group = df[(df[f'{typ}_label'] == label)] fig.line(x=f'x_{typ}', y=typ, source=group, color=CATEGORY10[idx], legend_label=label) relocate_legend(fig, 'left') return fig
def OA_10_205(): pplane_size = (((16 ** 2) + 16) + 1) baer_subplane_size = (((4 ** 2) + 4) + 1) B = [0, 1, 22, 33, 83, 122, 135, 141, 145, 159, 175, 200, 226, 229, 231, 238, 246] pplane = [[((xx + i) % pplane_size) for xx in B] for i in range(pplane_size)] baer_subplane = set([((i * pplane_size) / baer_subplane_size) for i in range(baer_subplane_size)]) p = list(baer_subplane)[0] lines_through_p = [B for B in pplane if (p in B)] lines_through_p.sort(key=(lambda s: len(baer_subplane.intersection(s)))) groups = [[xx for xx in l if (xx not in baer_subplane)] for l in lines_through_p[(((4 ** 2) - 4) - 9):]] relabel = {v: i for (i, v) in enumerate(sum(groups, []))} GDD = [[relabel[xx] for xx in B if (xx in relabel)] for B in pplane if (p not in B)] GDD.extend([([relabel[xx] for xx in G] + [204]) for G in groups]) blocks_of_size_9 = [B for B in GDD if (len(B) == 9)] blocks_of_size_9_union = sum(blocks_of_size_9, []) OA = OA_from_PBD(10, 205, [B for B in GDD if (len(B) != 9)], check=False)[:(- 205)] OA.extend([[B[xx] for xx in R] for R in orthogonal_array(10, 9) for B in blocks_of_size_9]) OA.extend([([i] * 10) for i in set(range(205)).difference(blocks_of_size_9_union)]) return OA
def _load_sample_hash() -> Tuple[(Array, Array)]: return (jnp.array([[, , , ], [, , , ], [, 6709002, , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [7139188, , , ], [2653767, , , ], [, , 6716808, ], [, 49709, , ], [, 3340423, , 7686579], [, , , 9605779], [, , , ], [, , , ], [7195005, , , ], [, , 7001617, ], [, , , ], [, , , 535365], [, , , ], [, 2462075, , ], [, , , 3695004], [, , , ], [, , 2460251, ], [, , , ], [, , , ], [, , , ], [, , 6039937, ], [, , , ], [, 5520088, , ], [, , , ], [910719, , , ], [, 1098476, , ], [, 8624220, , ], [, , , ], [, , , ], [, , , ], [4708326, , , ], [, , , ], [, , 9109075, ], [, , 224865, ], [, , , ], [, , , 8532577], [, , 5027048, ], [, , , ], [, , 7928914, ], [, , , ], [, 1743692, , ], [, , , ], [2632453, 7269809, , ], [1996183, , , ], [6233580, , , ], [, , 8374039, ], [, , , ], [, , , 2978505], [1664472, , , ], [, 3228834, 6198166, ], [, , , ], [, 3386355, , ], [, 9420273, , ], [, , , 6772527], [, , , 1630009], [9004247, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, 6845803, , ], [, 5193469, , ], [, , , ], [, , 7912677, ], [, , , ], [6760779, , , 5480498], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [, , , ], [9724895, , , ], [, , , ], [3961481, , , ], [, , , ], [, , , 8920985], [, , 6446892, ], [, , , ], [, , , ], [, , , ], [1270809, , , ], [4862706, , , 6219870], [, , , ], [, , , ], [, , , ], [, , , ]], dtype=jnp.int32), jnp.array([[71233, 771721, 71505, 706185], [289177, 484147, 358809, 484147], [359355, 549137, 359096, 549137], [350631, 558133, 350630, 554037], [370087, 538677, 370087, 538677], [4432, 899725, 4432, 904077], [678487, 229987, 678487, 229987], [423799, 480614, 423799, 480870], [549958, 284804, 549958, 280708], [423848, 480565, 423848, 480549], [489129, 283940, 554921, 283940], [86641, 822120, 86641, 822120], [206370, 702394, 206370, 567209], [500533, 407959, 500533, 407959], [759723, 79137, 759723, 79137], [563305, 345460, 559209, 345460], [231733, 611478, 231733, 611478], [502682, 406082, 498585, 406082], [554567, 288662, 554567, 288662], [476823, 427846, 476823, 427846], [488823, 415846, 488823, 415846], [431687, 477078, 431687, 477078], [419159, 424070, 415062, 424070], [493399, 345734, 493143, 345718], [678295, 230451, 678295, 230451], [496520, 342596, 496520, 346709], [567109, 276116, 567109, 276116], [624005, 284758, 624005, 284758], [420249, 484420, 420248, 484420], [217715, 621418, 217715, 621418], [344884, 493977, 344884, 493977], [550841, 292132, 550841, 292132], [284262, 558967, 284006, 558967], [152146, 756616, 152146, 756616], [144466, 698763, 144466, 694667], [284261, 624504, 284261, 624504], [288406, 620102, 288405, 620358], [301366, 607383, 301366, 607382], [468771, 435882, 468771, 435882], [555688, 283444, 555688, 283444], [485497, 414820, 485497, 414820], [633754, 275010, 633754, 275010], [419141, 489608, 419157, 489608], [694121, 214387, 694121, 214387], [480869, 427639, 481125, 427639], [489317, 419447, 489301, 419447], [152900, 747672, 152900, 747672], [348516, 494457, 348516, 494457], [534562, 370088, 534562, 370088], [371272, 537475, 371274, 537475], [144194, 760473, 144194, 760473], [567962, 275011, 567962, 275011], [493161, 350052, 493161, 350052], [490138, 348979, 490138, 348979], [328450, 506552, 328450, 506552], [148882, 759593, 148626, 755497], [642171, 266593, 642171, 266593], [685894, 218774, 685894, 218774], [674182, 234548, 674214, 234548], [756347, 152146, 690811, 86353], [612758, 291894, 612758, 291894], [296550, 612214, 296550, 612214], [363130, 475730, 363130, 475730], [691559, 16496, 691559, 16496], [340755, 502202, 336659, 502218], [632473, 210499, 628377, 210483], [564410, 266513, 564410, 266513], [427366, 481399, 427366, 481399], [493159, 349797, 493159, 415605], [331793, 576972, 331793, 576972], [416681, 492084, 416681, 492084], [813496, 95265, 813496, 91153], [695194, 213571, 695194, 213571], [436105, 407124, 436105, 407124], [836970, 6243, 902506, 6243], [160882, 747882, 160882, 747882], [493977, 414788, 489624, 414788], [29184, 551096, 29184, 616888], [903629, 4880, 899517, 4880], [351419, 553250, 351419, 553250], [75554, 767671, 75554, 767671], [279909, 563304, 279909, 563304], [215174, 628054, 215174, 628054], [361365, 481864, 361365, 481864], [424022, 484743, 358486, 484725], [271650, 633018, 271650, 633018], [681896, 226867, 616088, 226867], [222580, 686184, 222564, 686184], [144451, 698778, 209987, 698778], [532883, 310086, 532883, 310086], [628872, 279893, 628872, 279893], [533797, 374951, 533797, 374951], [91713, 817036, 91713, 817036], [427605, 477046, 431718, 477046], [145490, 689529, 145490, 689529], [551098, 291875, 551098, 291875], [349781, 558984, 349781, 558983], [205378, 703115, 205378, 703115], [362053, 546456, 362053, 546456], [612248, 226371, 678040, 226371]], dtype=jnp.int32))
class PickleableMagicMock(mock.MagicMock): def __reduce__(self): return (mock.MagicMock, ())
def log_scalar_sum(key: str, value: float, priority: int=10, round: Optional[int]=None): for agg in get_active_aggregators(): if (key not in agg): agg.add_meter(key, SumMeter(round=round), priority) agg[key].update(value)
class TFMobileBertPreTrainedModel(metaclass=DummyObject): _backends = ['tf'] def __init__(self, *args, **kwargs): requires_backends(self, ['tf'])
def load_scone(dirname): dfs = [] for filename in glob.glob((dirname + '/*.csv')): df = pd.read_csv(filename, index_col=0) df['category'] = os.path.basename(filename).replace('.csv', '') dfs.append(df) data_df = pd.concat(dfs) def as_example(row): suffix = ('' if (row['category'] == 'one_scoped') else '_edited') hkey = ('sentence2' + suffix) question = (row[hkey][0].lower() + row[hkey][1:].strip('.')) question = f'Can we logically conclude for sure that {question}?' label = ('Yes' if (row[('gold_label' + suffix)] == 'entailment') else 'No') return dspy.Example({'context': row[('sentence1' + suffix)], 'question': question, 'answer': label, 'category': row['category']}).with_inputs('context', 'question') return list(data_df.apply(as_example, axis=1).values)
def test_string_literal_in_callback_2(): success = False _inhibitor def cb(a): nonlocal success if (a == "b'Hello World!'"): success = True def tester(a): cb("b'Hello World!'") a = np.random.rand(1) tester(a) assert (success is True)
def sample_neigh(graphs, size): ps = np.array([len(g) for g in graphs], dtype=np.float) ps /= np.sum(ps) dist = stats.rv_discrete(values=(np.arange(len(graphs)), ps)) while True: idx = dist.rvs() graph = graphs[idx] start_node = random.choice(list(graph.nodes)) neigh = [start_node] frontier = list((set(graph.neighbors(start_node)) - set(neigh))) visited = set([start_node]) while ((len(neigh) < size) and frontier): new_node = random.choice(list(frontier)) assert (new_node not in neigh) neigh.append(new_node) visited.add(new_node) frontier += list(graph.neighbors(new_node)) frontier = [x for x in frontier if (x not in visited)] if (len(neigh) == size): return (graph, neigh)
def interactions_rating_pandas_dataset(): events = pd.DataFrame({'user_id': [0, 0, 1, 1, 1, 2], 'item_id': [0, 1, 0, 2, 3, 1], 'rating': [1.1, 1.2, 1.3, 2, 3, 4]}) return {'interactions': events, 'user_col': 'user_id', 'item_col': 'item_id', 'ratings_col': 'rating', 'users_cardinality': 3, 'items_cardinality': 4}
class DummyTask(FairseqTask): def __init__(self, args): super().__init__(args) self.dictionary = get_dummy_dictionary() if getattr(self.args, 'ctc', False): self.dictionary.add_symbol('<ctc_blank>') self.src_dict = self.dictionary self.tgt_dict = self.dictionary def source_dictionary(self): return self.src_dict def target_dictionary(self): return self.dictionary
def re_train(mode): if (mode == 0): dbh.create_table() dbh.populate_db() synth.synthesize(20) alphabet_model.train_alphabets() if (mode == 1): dbh.create_pose_table() dbh.populate_words() word_model.train_words()
class AutonomousSystem(Printable, Graphable, Configurable): __asn: int __subnets: List[IPv4Network] __routers: Dict[(str, Node)] __hosts: Dict[(str, Node)] __nets: Dict[(str, Network)] __name_servers: List[str] def __init__(self, asn: int, subnetTemplate: str='10.{}.0.0/16'): super().__init__() self.__hosts = {} self.__routers = {} self.__nets = {} self.__asn = asn self.__subnets = (None if (asn > 255) else list(IPv4Network(subnetTemplate.format(asn)).subnets(new_prefix=24))) self.__name_servers = [] def setNameServers(self, servers: List[str]) -> AutonomousSystem: self.__name_servers = servers return self def getNameServers(self) -> List[str]: return self.__name_servers def getPrefixList(self) -> List[str]: rslt = requests.get(RIS_PREFIXLIST_URL, {'resource': self.__asn}) assert (rslt.status_code == 200), 'RIPEstat API returned non-200' json = rslt.json() assert (json['status'] == 'ok'), 'RIPEstat API returned not-OK' return [p['prefix'] for p in json['data']['prefixes'] if (':' not in p['prefix'])] def registerNodes(self, emulator: Emulator): reg = emulator.getRegistry() for val in list(self.__nets.values()): net: Network = val if (net.getRemoteAccessProvider() != None): rap = net.getRemoteAccessProvider() brNode = self.createRouter('br-{}'.format(net.getName())) brNet = emulator.getServiceNetwork() rap.configureRemoteAccess(emulator, net, brNode, brNet) for router in list(self.__routers.values()): if issubclass(router.__class__, RealWorldRouter): router.joinNetwork(emulator.getServiceNetwork().getName()) for (key, val) in self.__nets.items(): reg.register(str(self.__asn), 'net', key, val) for (key, val) in self.__hosts.items(): reg.register(str(self.__asn), 'hnode', key, val) for (key, val) in self.__routers.items(): reg.register(str(self.__asn), 'rnode', key, val) def configure(self, emulator: Emulator): for host in self.__hosts.values(): if (len(host.getNameServers()) == 0): host.setNameServers(self.__name_servers) host.configure(emulator) for router in self.__routers.values(): if (len(router.getNameServers()) == 0): router.setNameServers(self.__name_servers) router.configure(emulator) def getAsn(self) -> int: return self.__asn def createNetwork(self, name: str, prefix: str='auto', direct: bool=True, aac: AddressAssignmentConstraint=None) -> Network: assert ((prefix != 'auto') or (self.__asn <= 255)), "can't use auto: asn > 255" network = (IPv4Network(prefix) if (prefix != 'auto') else self.__subnets.pop(0)) assert (name not in self.__nets), 'Network with name {} already exist.'.format(name) self.__nets[name] = Network(name, NetworkType.Local, network, aac, direct) return self.__nets[name] def getNetwork(self, name: str) -> Network: return self.__nets[name] def getNetworks(self) -> List[str]: return list(self.__nets.keys()) def createRouter(self, name: str) -> Node: assert (name not in self.__routers), 'Router with name {} already exists.'.format(name) self.__routers[name] = Node(name, NodeRole.Router, self.__asn) return self.__routers[name] def createRealWorldRouter(self, name: str, hideHops: bool=True, prefixes: List[str]=None) -> Node: assert (name not in self.__routers), 'Router with name {} already exists.'.format(name) router: RealWorldRouter = Node(name, NodeRole.Router, self.__asn) router.__class__ = RealWorldRouter router.initRealWorld(hideHops) if (prefixes == None): prefixes = self.getPrefixList() for prefix in prefixes: router.addRealWorldRoute(prefix) self.__routers[name] = router return router def getRouters(self) -> List[str]: return list(self.__routers.keys()) def getRouter(self, name: str) -> Node: return self.__routers[name] def createHost(self, name: str) -> Node: assert (name not in self.__hosts), 'Host with name {} already exists.'.format(name) self.__hosts[name] = Node(name, NodeRole.Host, self.__asn) return self.__hosts[name] def getHost(self, name: str) -> Node: return self.__hosts[name] def getHosts(self) -> List[str]: return list(self.__hosts.keys()) def _doCreateGraphs(self, emulator: Emulator): l2graph = self._addGraph('AS{}: Layer 2 Connections'.format(self.__asn), False) for obj in self.__nets.values(): net: Network = obj l2graph.addVertex('Network: {}'.format(net.getName()), shape='rectangle', group='AS{}'.format(self.__asn)) for obj in self.__routers.values(): router: Node = obj rtrname = 'Router: {}'.format(router.getName(), group='AS{}'.format(self.__asn)) l2graph.addVertex(rtrname, group='AS{}'.format(self.__asn), shape='diamond') for iface in router.getInterfaces(): net = iface.getNet() netname = 'Network: {}'.format(net.getName()) if (net.getType() == NetworkType.InternetExchange): netname = 'Exchange: {}...'.format(net.getName()) l2graph.addVertex(netname, shape='rectangle') if (net.getType() == NetworkType.CrossConnect): netname = 'CrossConnect: {}...'.format(net.getName()) l2graph.addVertex(netname, shape='rectangle') l2graph.addEdge(rtrname, netname) for obj in self.__hosts.values(): router: Node = obj rtrname = 'Host: {}'.format(router.getName(), group='AS{}'.format(self.__asn)) l2graph.addVertex(rtrname, group='AS{}'.format(self.__asn)) for iface in router.getInterfaces(): net = iface.getNet() netname = 'Network: {}'.format(net.getName()) l2graph.addEdge(rtrname, netname) def print(self, indent: int) -> str: out = (' ' * indent) out += 'AutonomousSystem {}:\n'.format(self.__asn) indent += 4 out += (' ' * indent) out += 'Networks:\n' for net in self.__nets.values(): out += net.print((indent + 4)) out += (' ' * indent) out += 'Routers:\n' for node in self.__routers.values(): out += node.print((indent + 4)) out += (' ' * indent) out += 'Hosts:\n' for host in self.__hosts.values(): out += host.print((indent + 4)) return out
def Simulator_customizations(module): Simulator = module['ns3::Simulator'] Simulator.add_custom_method_wrapper('Schedule', '_wrap_Simulator_Schedule', flags=['METH_VARARGS', 'METH_KEYWORDS', 'METH_STATIC']) Simulator.add_custom_method_wrapper('ScheduleNow', '_wrap_Simulator_ScheduleNow', flags=['METH_VARARGS', 'METH_KEYWORDS', 'METH_STATIC']) Simulator.add_custom_method_wrapper('ScheduleDestroy', '_wrap_Simulator_ScheduleDestroy', flags=['METH_VARARGS', 'METH_KEYWORDS', 'METH_STATIC']) Simulator.add_custom_method_wrapper('Run', '_wrap_Simulator_Run', flags=['METH_VARARGS', 'METH_KEYWORDS', 'METH_STATIC'])
class XLMModelTest(CommonTestCases.CommonModelTester): all_model_classes = ((XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple) if is_torch_available() else ()) class XLMModelTester(object): def __init__(self, parent, batch_size=13, seq_length=7, is_training=True, use_input_lengths=True, use_token_type_ids=True, use_labels=True, gelu_activation=True, sinusoidal_embeddings=False, causal=False, asm=False, n_langs=2, vocab_size=99, n_special=0, hidden_size=32, num_hidden_layers=5, num_attention_heads=4, hidden_dropout_prob=0.1, attention_probs_dropout_prob=0.1, max_position_embeddings=512, type_vocab_size=16, type_sequence_label_size=2, initializer_range=0.02, num_labels=3, num_choices=4, summary_type='last', use_proj=True, scope=None): self.parent = parent self.batch_size = batch_size self.seq_length = seq_length self.is_training = is_training self.use_input_lengths = use_input_lengths self.use_token_type_ids = use_token_type_ids self.use_labels = use_labels self.gelu_activation = gelu_activation self.sinusoidal_embeddings = sinusoidal_embeddings self.asm = asm self.n_langs = n_langs self.vocab_size = vocab_size self.n_special = n_special self.summary_type = summary_type self.causal = causal self.use_proj = use_proj self.hidden_size = hidden_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.hidden_dropout_prob = hidden_dropout_prob self.attention_probs_dropout_prob = attention_probs_dropout_prob self.max_position_embeddings = max_position_embeddings self.n_langs = n_langs self.type_sequence_label_size = type_sequence_label_size self.initializer_range = initializer_range self.summary_type = summary_type self.num_labels = num_labels self.num_choices = num_choices self.scope = scope def prepare_config_and_inputs(self): input_ids = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) input_mask = ids_tensor([self.batch_size, self.seq_length], 2).float() input_lengths = None if self.use_input_lengths: input_lengths = ((ids_tensor([self.batch_size], vocab_size=2) + self.seq_length) - 2) token_type_ids = None if self.use_token_type_ids: token_type_ids = ids_tensor([self.batch_size, self.seq_length], self.n_langs) sequence_labels = None token_labels = None is_impossible_labels = None if self.use_labels: sequence_labels = ids_tensor([self.batch_size], self.type_sequence_label_size) token_labels = ids_tensor([self.batch_size, self.seq_length], self.num_labels) is_impossible_labels = ids_tensor([self.batch_size], 2).float() config = XLMConfig(vocab_size_or_config_json_file=self.vocab_size, n_special=self.n_special, emb_dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, gelu_activation=self.gelu_activation, sinusoidal_embeddings=self.sinusoidal_embeddings, asm=self.asm, causal=self.causal, n_langs=self.n_langs, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, summary_type=self.summary_type, use_proj=self.use_proj) return (config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask) def check_loss_output(self, result): self.parent.assertListEqual(list(result['loss'].size()), []) def create_and_check_xlm_model(self, config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask): model = XLMModel(config=config) model.eval() outputs = model(input_ids, lengths=input_lengths, langs=token_type_ids) outputs = model(input_ids, langs=token_type_ids) outputs = model(input_ids) sequence_output = outputs[0] result = {'sequence_output': sequence_output} self.parent.assertListEqual(list(result['sequence_output'].size()), [self.batch_size, self.seq_length, self.hidden_size]) def create_and_check_xlm_lm_head(self, config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask): model = XLMWithLMHeadModel(config) model.eval() (loss, logits) = model(input_ids, token_type_ids=token_type_ids, labels=token_labels) result = {'loss': loss, 'logits': logits} self.parent.assertListEqual(list(result['loss'].size()), []) self.parent.assertListEqual(list(result['logits'].size()), [self.batch_size, self.seq_length, self.vocab_size]) def create_and_check_xlm_simple_qa(self, config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask): model = XLMForQuestionAnsweringSimple(config) model.eval() outputs = model(input_ids) outputs = model(input_ids, start_positions=sequence_labels, end_positions=sequence_labels) (loss, start_logits, end_logits) = outputs result = {'loss': loss, 'start_logits': start_logits, 'end_logits': end_logits} self.parent.assertListEqual(list(result['start_logits'].size()), [self.batch_size, self.seq_length]) self.parent.assertListEqual(list(result['end_logits'].size()), [self.batch_size, self.seq_length]) self.check_loss_output(result) def create_and_check_xlm_qa(self, config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask): model = XLMForQuestionAnswering(config) model.eval() outputs = model(input_ids) (start_top_log_probs, start_top_index, end_top_log_probs, end_top_index, cls_logits) = outputs outputs = model(input_ids, start_positions=sequence_labels, end_positions=sequence_labels, cls_index=sequence_labels, is_impossible=is_impossible_labels, p_mask=input_mask) outputs = model(input_ids, start_positions=sequence_labels, end_positions=sequence_labels, cls_index=sequence_labels, is_impossible=is_impossible_labels) (total_loss,) = outputs outputs = model(input_ids, start_positions=sequence_labels, end_positions=sequence_labels) (total_loss,) = outputs result = {'loss': total_loss, 'start_top_log_probs': start_top_log_probs, 'start_top_index': start_top_index, 'end_top_log_probs': end_top_log_probs, 'end_top_index': end_top_index, 'cls_logits': cls_logits} self.parent.assertListEqual(list(result['loss'].size()), []) self.parent.assertListEqual(list(result['start_top_log_probs'].size()), [self.batch_size, model.config.start_n_top]) self.parent.assertListEqual(list(result['start_top_index'].size()), [self.batch_size, model.config.start_n_top]) self.parent.assertListEqual(list(result['end_top_log_probs'].size()), [self.batch_size, (model.config.start_n_top * model.config.end_n_top)]) self.parent.assertListEqual(list(result['end_top_index'].size()), [self.batch_size, (model.config.start_n_top * model.config.end_n_top)]) self.parent.assertListEqual(list(result['cls_logits'].size()), [self.batch_size]) def create_and_check_xlm_sequence_classif(self, config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask): model = XLMForSequenceClassification(config) model.eval() (logits,) = model(input_ids) (loss, logits) = model(input_ids, labels=sequence_labels) result = {'loss': loss, 'logits': logits} self.parent.assertListEqual(list(result['loss'].size()), []) self.parent.assertListEqual(list(result['logits'].size()), [self.batch_size, self.type_sequence_label_size]) def prepare_config_and_inputs_for_common(self): config_and_inputs = self.prepare_config_and_inputs() (config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, input_mask) = config_and_inputs inputs_dict = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return (config, inputs_dict) def setUp(self): self.model_tester = XLMModelTest.XLMModelTester(self) self.config_tester = ConfigTester(self, config_class=XLMConfig, emb_dim=37) def test_config(self): self.config_tester.run_common_tests() def test_xlm_model(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*config_and_inputs) def test_xlm_lm_head(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*config_and_inputs) def test_xlm_simple_qa(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*config_and_inputs) def test_xlm_qa(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*config_and_inputs) def test_xlm_sequence_classif(self): config_and_inputs = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*config_and_inputs) .slow def test_model_from_pretrained(self): cache_dir = '/tmp/transformers_test/' for model_name in list(XLM_PRETRAINED_MODEL_ARCHIVE_MAP.keys())[:1]: model = XLMModel.from_pretrained(model_name, cache_dir=cache_dir) shutil.rmtree(cache_dir) self.assertIsNotNone(model)
class ReducedContextAggModel(MeanPoolModel): def init_graph(self, X, batch_size, ckpt=None, device='gpu', **model_params): self.batch_size = batch_size n_dim_x = len(X[0].values[0]) n_dim_q = len(X[1].values[0]) n_dim_p = len(X[2].values[0]) n_dim_c = len(X[3].values[0]) for (key, val) in model_params.items(): if key.startswith('n_hidden'): model_params[key] = int(model_params[key]) self.model = self.create_graph(None, n_dim_x=n_dim_x, n_dim_q=n_dim_q, n_dim_p=n_dim_p, n_dim_c=n_dim_c, device=device, **model_params) self.init = tf.initializers.global_variables() TF_CONFIG = tf.ConfigProto(gpu_options=tf.GPUOptions(per_process_gpu_memory_fraction=0.1), allow_soft_placement=True) self.sess = sess = tf.Session(config=TF_CONFIG) self.saver = tf.train.Saver() if (ckpt is not None): self.saver.restore(self.sess, 'tmp/{}.ckpt'.format(ckpt)) def fit(self, X, y=None, X_val=None, y_val=None, batch_size=32, n_epochs=50, patience=10, verbose=1, **model_params): sess = self.sess sess.run(self.init) train_dl = DataLoader(X, batch_size, shuffle=True, seed=21) test_dl = DataLoader(X_val, batch_size, shuffle=False, seed=21) model = self.model best_score = np.inf best_probs = None since_best = 0 for epoch in range(n_epochs): pbar = (tqdm(desc='Trn', total=(len(X) + len(X_val))) if verbose else contextlib.suppress()) with pbar: loss = [] for (idx, (batch_x, batch_q, batch_p, batch_c, batch_y, seq_lens)) in enumerate(train_dl): (loss_, _) = sess.run([model.loss, model.train_op], feed_dict={model.d_X: batch_x, model.d_Q: batch_q, model.d_P: batch_p, model.d_C: batch_c, model.d_y: batch_y, model.d_seq_lens: seq_lens}) loss.append(loss_) if verbose: pbar.set_description('Trn {:2d}, Loss={:.3f}, Val-Loss={:.3f}'.format(epoch, np.mean(loss), np.inf)) pbar.update(len(batch_x)) trn_loss = np.mean(loss) (loss, y_true, probs) = self.predict(test_dl, epoch=epoch, trn_loss=trn_loss, pbar=pbar) score = log_loss(np.array(y_true), np.array(probs)) if (score < best_score): best_score = score best_probs = probs since_best = 0 self.saver.save(sess, 'tmp/best_model.ckpt') else: since_best += 1 if verbose: pbar.set_description('Trn {:2d}, Loss={:.3f}, Val-Loss={:.3f}'.format(epoch, trn_loss, score)) pbar.update(len(X_val)) if (since_best > patience): break if verbose: print('Best score on validation set: ', best_score) self.best_score = best_score tf.reset_default_graph() return self def predict(self, X, y=None, epoch=None, trn_loss=None, pbar=None, verbose=1, batch_size=32, **parameters): if (trn_loss is None): self.init_graph(X, batch_size, ckpt='best_model', device='cpu', **parameters) if (trn_loss is None): test_dl = DataLoader(X, self.batch_size) else: test_dl = X (loss, y_true, probs) = ([], [], []) for (idx, (batch_x, batch_q, batch_p, batch_c, batch_y, seq_lens)) in enumerate(test_dl): (loss_, y_true_, probs_) = self.sess.run([self.model.loss, self.model.d_y, self.model.probs], feed_dict={self.model.d_X: batch_x, self.model.d_Q: batch_q, self.model.d_P: batch_p, self.model.d_C: batch_c, self.model.d_y: batch_y, self.model.d_seq_lens: seq_lens}) loss.append(loss_) y_true += y_true_.tolist() probs += probs_.tolist() if (trn_loss is None): tf.reset_default_graph() return (loss, np.array(y_true), probs) def create_graph(self, batch_size=None, seq_len=None, n_dim_x=1024, n_dim_q=1024, n_dim_p=1024, n_dim_c=1024, n_hidden_x=1024, n_hidden_q=1024, n_hidden_p=1024, n_hidden_c=1024, dropout_rate_x=0.5, dropout_rate_p=0.5, dropout_rate_q=0.5, dropout_rate_c=0.5, dropout_rate_fc=0.5, reg_x=0.01, reg_q=0.01, reg_p=0.01, reg_c=0.01, reg_fc=0.01, label_smoothing=0.1, data_type=tf.float32, activation='tanh', use_pretrained=False, use_plus_features=False, use_context=False, seed=None, device='gpu'): def gelu_fast(_x): return ((0.5 * _x) * (1 + tf.tanh((tf.sqrt((2 / np.pi)) * (_x + (0.044715 * tf.pow(_x, 3))))))) activ = tf.nn.relu tf.reset_default_graph() with tf.device('/{}:0'.format(device)): d_X = tf.placeholder(data_type, [batch_size, n_dim_x]) d_Q = tf.placeholder(data_type, [batch_size, n_dim_q]) d_P = tf.placeholder(data_type, [batch_size, n_dim_p]) d_C = tf.placeholder(data_type, [batch_size, n_dim_c]) d_y = tf.placeholder(tf.int32, [batch_size, 3]) d_seq_lens = tf.placeholder(tf.int32, [batch_size]) with tf.name_scope('dense_concat_layers'): dense_init = tf.keras.initializers.glorot_normal(seed=21) X = tf.keras.layers.Dropout(dropout_rate_x, seed=7)(d_X) X = tf.keras.layers.Dense(n_hidden_x, activation=None, kernel_initializer=dense_init, kernel_regularizer=tf.contrib.layers.l1_regularizer(reg_x))(X) X = tf.layers.batch_normalization(X) X = activ(X) Q = tf.keras.layers.Dropout(dropout_rate_q, seed=7)(d_Q) Q = tf.keras.layers.Dense(n_hidden_q, activation=None, kernel_initializer=dense_init, kernel_regularizer=tf.contrib.layers.l1_regularizer(reg_q))(Q) Q = tf.layers.batch_normalization(Q) Q = activ(Q) P = tf.keras.layers.Dropout(dropout_rate_p, seed=7)(d_P) P = tf.keras.layers.Dense(n_hidden_p, activation=None, kernel_initializer=dense_init, kernel_regularizer=tf.contrib.layers.l1_regularizer(reg_p))(P) P = tf.layers.batch_normalization(P) P = activ(P) C = tf.keras.layers.Dropout(dropout_rate_c, seed=7)(d_C) C = tf.keras.layers.Dense(n_hidden_c, activation=None, kernel_initializer=dense_init, kernel_regularizer=tf.contrib.layers.l1_regularizer(reg_c))(C) C = tf.layers.batch_normalization(C) C = activ(C) feats = [X] if use_pretrained: feats.append(P) if use_plus_features: feats.append(Q) if use_context: feats.append(C) X = tf.concat(feats, axis=(- 1)) X = tf.keras.layers.Dropout(dropout_rate_fc, seed=7)(X) y_hat = tf.keras.layers.Dense(3, name='output', kernel_initializer=dense_init, kernel_regularizer=tf.contrib.layers.l2_regularizer(reg_fc))(X) with tf.name_scope('loss'): probs = tf.nn.softmax(y_hat, axis=(- 1)) loss = tf.losses.softmax_cross_entropy(d_y, logits=y_hat, label_smoothing=label_smoothing) loss = tf.reduce_mean(loss) with tf.name_scope('optimizer'): global_step = tf.Variable(0, trainable=False, dtype=tf.int32) learning_rate = 0.005 learning_rate = tf.train.cosine_decay_restarts(learning_rate, global_step, 500, t_mul=1, m_mul=1, alpha=0.01, name=None) optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate) train_op = optimizer.minimize(loss, global_step=global_step) return AttrDict(locals()) def mean_featurizer(X_bert, X_plus, X_pretrained, y_true): batch_x = [] batch_q = [] batch_p = [] batch_c = [] seq_lens = [] batch_y = [] max_len = 512 max_len_tok = 20 for (idx, row) in tqdm(X_bert.iterrows(), total=len(X_bert)): plus_row = X_plus.loc[idx] x = np.array(row.bert) q = np.array(plus_row.plus) p = X_pretrained.loc[idx].pretrained c = np.array(row.cls).reshape((- 1)) mention_token = 0 if (row.pronoun_offset == min(row.pronoun_offset, row.a_offset, row.b_offset)): mention_token = 0 pronoun_vec = x[row.pronoun_offset_token:(row.pronoun_offset_token + 1)] a_vec = x[row.a_span[mention_token]:(row.a_span[mention_token] + 1)] b_vec = x[row.b_span[mention_token]:(row.b_span[mention_token] + 1)] x = np.hstack((np.mean(pronoun_vec, axis=0), np.mean(a_vec, axis=0), np.mean(b_vec, axis=0))).reshape((- 1)) pronoun_vec = q[plus_row.pronoun_offset_token:(plus_row.pronoun_offset_token + 1)] a_vec = q[plus_row.a_span[mention_token]:(plus_row.a_span[mention_token] + 1)] b_vec = q[plus_row.b_span[mention_token]:(plus_row.b_span[mention_token] + 1)] q = np.hstack((np.mean(pronoun_vec, axis=0), np.mean(a_vec, axis=0), np.mean(b_vec, axis=0))).reshape((- 1)) y = y_true.loc[idx].values seq_len = len(row.tokens) batch_x.append(x) batch_q.append(q) batch_p.append(p) batch_c.append(c) batch_y.append(y) seq_lens.append(seq_len) return pd.DataFrame([batch_x, batch_q, batch_p, batch_c, batch_y, seq_lens]).T
class IndPool(): def __init__(self, pool_size): self.pool_size = pool_size if (self.pool_size > 0): self.num = 0 self.inds = [] def fetch(self): return self.inds def is_full(self): full = (self.num == self.pool_size) return full def update(self, inds): for ind in inds: if (self.num < self.pool_size): self.num = (self.num + 1) else: self.inds.pop(0) self.inds.append(ind) return self.inds
def build_stamp_source(root_dir=None, stamp_fname='pyx-stamps', include_dirs=None): if (root_dir is None): root_dir = os.getcwd() if (include_dirs is None): include_dirs = [pjoin(root_dir, 'src')] pyxes = find_pyx(root_dir) pyx_defs = build_stamp(pyxes, include_dirs=include_dirs) write_stamps(pyx_defs, stamp_fname)
def init_logging(args): handlers = [logging.StreamHandler()] if ((not args.no_log) and (args.log_file is not None)): mode = ('a' if os.path.isfile(args.resume_training) else 'w') handlers.append(logging.FileHandler(args.log_file, mode=mode)) logging.basicConfig(handlers=handlers, format='[%(asctime)s] %(message)s', datefmt='%Y-%m-%d %H:%M:%S', level=logging.INFO) logging.info(('COMMAND: %s' % ' '.join(sys.argv))) logging.info('Arguments: {}'.format(vars(args)))
(('ScriptModule' not in core._REGISTERED_OPERATORS), 'Script module integration in Caffe2 is not enabled') class TestScriptModule(test_util.TestCase): def _createFeedModule(self): workspace.FeedBlob('m', MyModule()) def testCreation(self): m = MyModule() workspace.FeedBlob('module', m) m2 = workspace.FetchBlob('module') self.assertTrue((m2 is not None)) def testForward(self): self._createFeedModule() val = np.random.rand(5, 5).astype(np.float32) param = np.array([[1, 2, 3, 4, 5]]).astype(np.float32) workspace.FeedBlob('w', val) workspace.RunOperatorOnce(core.CreateOperator('ScriptModule', ['m', 'w'], ['y'])) np.testing.assert_almost_equal(workspace.FetchBlob('y'), np.matmul(param, val), decimal=5) def testMultiInputOutput(self): self._createFeedModule() val = np.random.rand(5, 5).astype(np.float32) workspace.FeedBlob('w', val) val2 = np.random.rand(5, 5).astype(np.float32) workspace.FeedBlob('w2', val2) workspace.RunOperatorOnce(core.CreateOperator('ScriptModule', ['m', 'w', 'w2'], ['y'], method='multi_input')) workspace.RunOperatorOnce(core.CreateOperator('ScriptModule', ['m', 'w'], ['y1', 'y2'], method='multi_output')) np.testing.assert_almost_equal(workspace.FetchBlob('y'), ((val + val2) + 2), decimal=5) np.testing.assert_almost_equal(workspace.FetchBlob('y1'), val, decimal=5) np.testing.assert_almost_equal(workspace.FetchBlob('y2'), (val + 1), decimal=5) def testSerialization(self): tmpdir = tempfile.mkdtemp() try: self._createFeedModule() workspace.RunOperatorOnce(core.CreateOperator('Save', ['m'], [], absolute_path=1, db=os.path.join(tmpdir, 'db'), db_type='minidb')) workspace.ResetWorkspace() self.assertFalse(workspace.HasBlob('m')) workspace.RunOperatorOnce(core.CreateOperator('Load', [], [], absolute_path=1, db=os.path.join(tmpdir, 'db'), db_type='minidb', load_all=1)) self.assertTrue(workspace.HasBlob('m')) val = np.random.rand(5, 5).astype(np.float32) param = np.array([[1, 2, 3, 4, 5]]).astype(np.float32) workspace.FeedBlob('w', val) workspace.RunOperatorOnce(core.CreateOperator('ScriptModule', ['m', 'w'], ['y'])) np.testing.assert_almost_equal(workspace.FetchBlob('y'), np.matmul(param, val), decimal=5) finally: try: shutil.rmtree(tmpdir) except OSError as e: if (e.errno != errno.ENOENT): raise
class AttentionUNet(nn.Module): def __init__(self, img_ch=3, output_ch=1): super(AttentionUNet, self).__init__() self.MaxPool = nn.MaxPool2d(kernel_size=2, stride=2) self.Conv1 = ConvBlock(img_ch, 64) self.Conv2 = ConvBlock(64, 128) self.Conv3 = ConvBlock(128, 256) self.Conv4 = ConvBlock(256, 512) self.Conv5 = ConvBlock(512, 1024) self.Up5 = UpConv(1024, 512) self.Att5 = AttentionBlock(F_g=512, F_l=512, n_coefficients=256) self.UpConv5 = ConvBlock(1024, 512) self.Up4 = UpConv(512, 256) self.Att4 = AttentionBlock(F_g=256, F_l=256, n_coefficients=128) self.UpConv4 = ConvBlock(512, 256) self.Up3 = UpConv(256, 128) self.Att3 = AttentionBlock(F_g=128, F_l=128, n_coefficients=64) self.UpConv3 = ConvBlock(256, 128) self.Up2 = UpConv(128, 64) self.Att2 = AttentionBlock(F_g=64, F_l=64, n_coefficients=32) self.UpConv2 = ConvBlock(128, 64) self.Conv = nn.Conv2d(64, output_ch, kernel_size=1, stride=1, padding=0) def forward(self, x): e1 = self.Conv1(x) e2 = self.MaxPool(e1) e2 = self.Conv2(e2) e3 = self.MaxPool(e2) e3 = self.Conv3(e3) e4 = self.MaxPool(e3) e4 = self.Conv4(e4) e5 = self.MaxPool(e4) e5 = self.Conv5(e5) d5 = self.Up5(e5) s4 = self.Att5(gate=d5, skip_connection=e4) d5 = torch.cat((s4, d5), dim=1) d5 = self.UpConv5(d5) d4 = self.Up4(d5) s3 = self.Att4(gate=d4, skip_connection=e3) d4 = torch.cat((s3, d4), dim=1) d4 = self.UpConv4(d4) d3 = self.Up3(d4) s2 = self.Att3(gate=d3, skip_connection=e2) d3 = torch.cat((s2, d3), dim=1) d3 = self.UpConv3(d3) d2 = self.Up2(d3) s1 = self.Att2(gate=d2, skip_connection=e1) d2 = torch.cat((s1, d2), dim=1) d2 = self.UpConv2(d2) out = self.Conv(d2) return out
def train(args, train_dataset, model, tokenizer): if (args.local_rank in [(- 1), 0]): tb_writer = SummaryWriter() args.train_batch_size = (args.per_gpu_train_batch_size * max(1, args.n_gpu)) train_sampler = (RandomSampler(train_dataset) if (args.local_rank == (- 1)) else DistributedSampler(train_dataset)) train_dataloader = DataLoader(train_dataset, sampler=train_sampler, batch_size=args.train_batch_size) if (args.max_steps > 0): t_total = args.max_steps args.num_train_epochs = ((args.max_steps // (len(train_dataloader) // args.gradient_accumulation_steps)) + 1) else: t_total = ((len(train_dataloader) // args.gradient_accumulation_steps) * args.num_train_epochs) no_decay = ['bias', 'LayerNorm.weight'] optimizer_grouped_parameters = [{'params': [p for (n, p) in model.named_parameters() if (not any(((nd in n) for nd in no_decay)))], 'weight_decay': args.weight_decay}, {'params': [p for (n, p) in model.named_parameters() if any(((nd in n) for nd in no_decay))], 'weight_decay': 0.0}] optimizer = AdamW(optimizer_grouped_parameters, lr=args.learning_rate, eps=args.adam_epsilon) scheduler = get_linear_schedule_with_warmup(optimizer, num_warmup_steps=args.warmup_steps, num_training_steps=t_total) if args.fp16: try: from apex import amp except ImportError: raise ImportError('Please install apex from to use fp16 training.') (model, optimizer) = amp.initialize(model, optimizer, opt_level=args.fp16_opt_level) if (args.n_gpu > 1): model = torch.nn.DataParallel(model) if (args.local_rank != (- 1)): model = torch.nn.parallel.DistributedDataParallel(model, device_ids=[args.local_rank], output_device=args.local_rank, find_unused_parameters=True) logger.info('***** Running training *****') logger.info(' Num examples = %d', len(train_dataset)) logger.info(' Num Epochs = %d', args.num_train_epochs) logger.info(' Instantaneous batch size per GPU = %d', args.per_gpu_train_batch_size) logger.info(' Total train batch size (w. parallel, distributed & accumulation) = %d', ((args.train_batch_size * args.gradient_accumulation_steps) * (torch.distributed.get_world_size() if (args.local_rank != (- 1)) else 1))) logger.info(' Gradient Accumulation steps = %d', args.gradient_accumulation_steps) logger.info(' Total optimization steps = %d', t_total) global_step = 0 (tr_loss, logging_loss) = (0.0, 0.0) model.zero_grad() train_iterator = trange(int(args.num_train_epochs), desc='Epoch', disable=(args.local_rank not in [(- 1), 0])) set_seed(args) for _ in train_iterator: epoch_iterator = tqdm(train_dataloader, desc='Iteration', disable=(args.local_rank not in [(- 1), 0])) for (step, batch) in enumerate(epoch_iterator): model.train() batch = tuple((t.to(args.device) for t in batch)) inputs = {'input_ids': batch[0], 'attention_mask': batch[1], 'token_type_ids': batch[2], 'labels': batch[3]} outputs = model(**inputs) loss = outputs[0] if (args.n_gpu > 1): loss = loss.mean() if (args.gradient_accumulation_steps > 1): loss = (loss / args.gradient_accumulation_steps) if args.fp16: with amp.scale_loss(loss, optimizer) as scaled_loss: scaled_loss.backward() torch.nn.utils.clip_grad_norm_(amp.master_params(optimizer), args.max_grad_norm) else: loss.backward() torch.nn.utils.clip_grad_norm_(model.parameters(), args.max_grad_norm) tr_loss += loss.item() if (((step + 1) % args.gradient_accumulation_steps) == 0): optimizer.step() scheduler.step() model.zero_grad() global_step += 1 if ((args.local_rank in [(- 1), 0]) and (args.logging_steps > 0) and ((global_step % args.logging_steps) == 0)): if ((args.local_rank == (- 1)) and args.evaluate_during_training): results = evaluate(args, model, tokenizer) for (key, value) in results.items(): tb_writer.add_scalar('eval_{}'.format(key), value, global_step) tb_writer.add_scalar('lr', scheduler.get_lr()[0], global_step) tb_writer.add_scalar('loss', ((tr_loss - logging_loss) / args.logging_steps), global_step) logging_loss = tr_loss if ((args.local_rank in [(- 1), 0]) and (args.save_steps > 0) and ((global_step % args.save_steps) == 0)): output_dir = os.path.join(args.output_dir, 'checkpoint-{}'.format(global_step)) if (not os.path.exists(output_dir)): os.makedirs(output_dir) model_to_save = (model.module if hasattr(model, 'module') else model) model_to_save.save_pretrained(output_dir) tokenizer.save_vocabulary(output_dir) torch.save(args, os.path.join(output_dir, 'training_args.bin')) logger.info('Saving model checkpoint to %s', output_dir) if ((args.max_steps > 0) and (global_step > args.max_steps)): epoch_iterator.close() break if ((args.max_steps > 0) and (global_step > args.max_steps)): train_iterator.close() break if (args.local_rank in [(- 1), 0]): tb_writer.close() return (global_step, (tr_loss / global_step))
.parametrize('a,b', [(10, 3), ((- 10), 3), (10, (- 3)), ((- 10), (- 3))]) _utils.test() def test_c_style_mod(a, b): z = ti.field(ti.i32, shape=()) def func(x: ti.i32, y: ti.i32): z[None] = ti.raw_mod(x, y) func(a, b) assert (z[None] == _c_mod(a, b))
def _converId(img_id): img_id = img_id.split('-') if ('train' in img_id[0]): new_id = int(img_id[1]) elif ('val' in img_id[0]): new_id = (int(img_id[1]) + 1000000) elif ('test' in img_id[0]): new_id = (int(img_id[1]) + 2000000) else: pdb.set_trace() return new_id
def warmup_lr_scheduler(optimizer, warmup_iters, warmup_factor): def f(x): if (x >= warmup_iters): return 1 alpha = (float(x) / warmup_iters) return ((warmup_factor * (1 - alpha)) + alpha) return torch.optim.lr_scheduler.LambdaLR(optimizer, f)