content
stringlengths
27
928k
path
stringlengths
4
230
size
int64
27
928k
nl_text
stringlengths
21
396k
nl_size
int64
21
396k
nl_language
stringlengths
2
3
nl_language_score
float64
0.04
1
from rest_framework import serializers from .models import Canteen from accounts.serializers import UserForSerializer from model_location.serializers import CityViewSerializer from model_media.serializers import MediaViewSerializer # Canteen model serializer class CanteenSerializer(serializers.ModelSerializer): class Meta: model = Canteen fields = '__all__' # Canteen model serializer to view class CanteenViewSerializer(serializers.ModelSerializer): user = UserForSerializer(read_only=True) city = CityViewSerializer(read_only=True) images = MediaViewSerializer(read_only=True, many=True) class Meta: model = Canteen fields = '__all__'
models/model_canteen/serializers.py
696
Canteen model serializer Canteen model serializer to view
57
en
0.829794
# MIT License # # Copyright (c) 2019 Red Hat, Inc. # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE.
packit/cli/__init__.py
1,107
MIT License Copyright (c) 2019 Red Hat, Inc. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
1,065
en
0.863681
import os import random from riscv_definitions import * NONE = 0 CF_J = 1 CF_BR = 2 CF_RET = 3 MEM_R = 4 MEM_W = 5 CSR = 6 PREFIX = '_p' MAIN = '_l' SUFFIX = '_s' class Word(): def __init__(self, label: int, insts: list, tpe=NONE, xregs=[], fregs=[], imms=[], symbols=[], populated=False): self.label = label self.tpe = tpe self.insts = insts self.len_insts = len(insts) self.xregs = xregs self.fregs = fregs self.imms = imms self.symbols = symbols self.operands = xregs + fregs + [ imm[0] for imm in imms ] + symbols self.populated = populated self.ret_insts = [] def pop_inst(self, inst, opvals): for (op, val) in opvals.items(): inst = inst.replace(op, val) return inst def populate(self, opvals, part=MAIN): for op in self.operands: assert op in opvals.keys(), \ '{} is not in label {} Word opvals'.format(op, self.label) pop_insts = [] for inst in self.insts: p_inst = self.pop_inst(inst, opvals) pop_insts.append(p_inst) ret_insts = [ '{:<8}{:<42}'.format(part + str(self.label) + ':', pop_insts.pop(0)) ] for i in range(len(pop_insts)): ret_insts.append('{:8}{:<42}'.format('', pop_insts.pop(0))) self.populated = True self.ret_insts = ret_insts def reset_label(self, new_label, part): old_label = self.label self.label = new_label if self.populated: self.ret_insts[0] = '{:8}{:<42}'.format(part + str(self.label) + ':', self.ret_insts[0][8:]) return (old_label, new_label) else: return None def repop_label(self, label_map, max_label, part): if self.populated: for i in range(len(self.ret_insts)): inst = self.ret_insts[i] tmps = inst.split(', ' + part) if len(tmps) > 1: label = tmps[1].split(' ')[0] old = int(label) new = label_map.get(old, random.randint(self.label + 1, max_label)) new_inst = inst[8:].replace(part + '{}'.format(old), part + '{}'.format(new)) inst = '{:<8}{:<50}'.format(inst[0:8], new_inst) self.ret_insts[i] = inst else: return def get_insts(self): assert self.populated, \ 'Word is not populated' return self.ret_insts def word_jal(opcode, syntax, xregs, fregs, imms, symbols): tpe = CF_J insts = [ syntax ] return (tpe, insts) def word_jalr(opcode, syntax, xregs, fregs, imms, symbols): tpe = CF_J insts = [ 'la xreg1, symbol', syntax ] symbols.append('symbol') return (tpe, insts) # Need to update def word_branch(opcode, syntax, xregs, fregs, imms, symbols): tpe = CF_BR insts = [ syntax ] return (tpe, insts) def word_ret(opcode, syntax, xregs, fregs, imms, symbols): tpe = CF_RET if syntax == 'mret': epc = 'mepc' elif syntax == 'sret': epc = 'sepc' else: epc = 'uepc' insts = [ 'la xreg0, symbol', 'csrrw zero, {}, xreg0'.format(epc), syntax ] xregs.append('xreg0') symbols.append('symbol') return (tpe, insts) def word_mem_r(opcode, syntax, xregs, fregs, imms, symbols): tpe = MEM_R rand = random.random() if rand < 0.1: mask_addr = [ 'lui xreg2, 0xffe00', 'xor xreg1, xreg1, xreg2' ] xregs.append('xreg2') else: mask_addr = [] insts = [ 'la xreg1, symbol' ] + mask_addr + [ syntax ] symbols.append('symbol') return (tpe, insts) def word_mem_w(opcode, syntax, xregs, fregs, imms, symbols): tpe = MEM_W rand = random.random() if rand < 0.1: mask_addr = [ 'lui xreg2, 0xffe00', 'xor xreg1, xreg1, xreg2' ] xregs.append('xreg2') else: mask_addr = [] insts = [ 'la xreg1, symbol' ] + mask_addr + [ syntax ] symbols.append('symbol') return (tpe, insts) def word_atomic(opcode, syntax, xregs, fregs, imms, symbols): tpe = MEM_W rand = random.random() if rand < 0.1: mask_addr = [ 'lui xreg2, 0xffe00', 'xor xreg1, xreg1, xreg2' ] xregs.append('xreg2') else: mask_addr = [] insts = [ 'la xreg1, symbol', 'addi xreg1, xreg1, imm6' ] + \ mask_addr + \ [ syntax ] if opcode in rv64.keys(): imms.append(('imm6', 8)) else: imms.append(('imm6', 4)) symbols.append('symbol') return (tpe, insts) def word_csr_r(opcode, syntax, xregs, fregs, imms, symbols): csr = random.choice(csr_names) if 'pmpaddr' in csr: tpe = MEM_R insts = [ 'la xreg1, symbol', 'srai xreg1, xreg1, 1', syntax.format(csr) ] symbols.append('symbol') else: tpe = CSR insts = [ 'xor xreg1, xreg1, xreg1'] for i in range(random.randint(0, 3)): set_bits = random.choice([1, 3]) offset = random.randint(0, 31) insts = insts + \ ['addi xreg{}, zero, {}'.format(i+2, set_bits), 'slli xreg{}, xreg{}, {}'.format(i+2, i+2, offset), 'add xreg1, xreg1, xreg{}'.format(i+2) ] xregs.append('xreg{}'.format(i+2)) insts.append(syntax.format(csr)) return (tpe, insts) def word_csr_i(opcode, syntax, xregs, fregs, imms, symbols): tpe = CSR csr = random.choice(csr_names) insts = [ syntax.format(csr) ] return (tpe, insts) def word_sfence(opcode, syntax, xregs, fregs, imms, symbols): tpe = NONE pt_symbol = random.choice([ 'pt0', 'pt1', 'pt2', 'pt3' ]) imms += [ ('uimm1', 1), ('uimm6', 8) ] insts = [ 'li xreg0, uimm1', 'la xreg1, {}'.format(pt_symbol), 'addi xreg1, xreg1, uimm6' ] + \ [ syntax ] return (tpe, insts) def word_fp(opcode, syntax, xregs, fregs, imms, symbols): tpe = NONE # rm = random.choice([ 'rne', 'rtz', 'rdn', # 'rup', 'rmm', 'dyn']) # Unset rounding mode testing rm = 'rne' insts = [ syntax.format(rm) ] return (tpe, insts) """ Opcodes_words Dictionary of opcodes - word generation functions to handle opcodes which need special instructions """ opcodes_words = { 'jal': (['jal'], word_jal), 'jalr': (['jalr'], word_jalr), 'branch': (list(rv32i_btype.keys()), word_branch), 'ret': (['mret', 'sret', 'uret'], word_ret), 'mem_r': (['lb', 'lh', 'lw', 'ld', 'lbu', 'lhu', 'lwu', \ 'flw', 'fld', 'flq'], word_mem_r), 'mem_w': (['sb', 'sh', 'sw', 'sd', 'fsw', 'fsd', 'fsq'], word_mem_w), 'atomic': (list(rv32a.keys()) + list(rv64a.keys()), word_atomic), 'csr_r': (['csrrw', 'csrrs', 'csrrc'], word_csr_r), 'csr_i': (['csrrwi', 'csrrsi', 'csrrci'], word_csr_i), 'sfence': (['sfence.vma'], word_sfence), 'fp': (list(rv32f.keys()) + list(rv64f.keys()) + list(rv32d.keys()) + \ list(rv64d.keys()) + list(rv32q.keys()) + list(rv64q.keys()), word_fp) }
Fuzzer/src/word.py
7,388
Need to update rm = random.choice([ 'rne', 'rtz', 'rdn', 'rup', 'rmm', 'dyn']) Unset rounding mode testing
127
en
0.280481
"""URLs for the ``django-frequently`` application.""" from django.conf.urls import url from . import views urlpatterns = [ url(r'^$', views.EntryCategoryListView.as_view(), name='frequently_list'), url(r'^your-question/$', views.EntryCreateView.as_view(), name='frequently_submit_question'), url(r'^(?P<slug>[a-z-0-9]+)/$', views.EntryDetailView.as_view(), name='frequently_entry_detail'), ]
frequently/urls.py
457
URLs for the ``django-frequently`` application.
47
en
0.894827
# Support various prediction methods for predicting cluster membership # of new or unseen points. There are several ways to interpret how # to do this correctly, so we provide several methods for # the different use cases that may arise. import numpy as np from sklearn.neighbors import KDTree, BallTree from .dist_metrics import DistanceMetric from ._hdbscan_tree import compute_stability, labelling_at_cut, recurse_leaf_dfs from ._prediction_utils import (get_tree_row_with_child, dist_membership_vector, outlier_membership_vector, prob_in_some_cluster, all_points_dist_membership_vector, all_points_outlier_membership_vector, all_points_prob_in_some_cluster) from warnings import warn class PredictionData(object): """ Extra data that allows for faster prediction if cached. Parameters ---------- data : array (n_samples, n_features) The original data set that was clustered condensed_tree : CondensedTree The condensed tree object created by a clustering min_samples : int The min_samples value used in clustering tree_type : string, optional Which type of space tree to use for core distance computation. One of: * ``kdtree`` * ``balltree`` metric : string, optional The metric used to determine distance for the clustering. This is the metric that will be used for the space tree to determine core distances etc. **kwargs : Any further arguments to the metric. Attributes ---------- raw_data : array (n_samples, n_features) The original data set that was clustered tree : KDTree or BallTree A space partitioning tree that can be queried for nearest neighbors. core_distances : array (n_samples,) The core distances for every point in the original data set. cluster_map : dict A dictionary mapping cluster numbers in the condensed tree to labels in the final selected clustering. cluster_tree : structured array A version of the condensed tree that only contains clusters, not individual points. max_lambdas : dict A dictionary mapping cluster numbers in the condensed tree to the maximum lambda value seen in that cluster. """ _tree_type_map = {'kdtree': KDTree, 'balltree': BallTree} def _clusters_below(self, cluster): result = [] to_process = [cluster] while to_process: result.extend(to_process) to_process = \ self.cluster_tree['child'][np.in1d(self.cluster_tree['parent'], to_process)] to_process = to_process.tolist() return result def _recurse_leaf_dfs(self, current_node): children = self.cluster_tree[self.cluster_tree['parent'] == current_node]['child'] if len(children) == 0: return [current_node, ] else: return sum( [recurse_leaf_dfs(self.cluster_tree, child) for child in children], []) def __init__(self, data, condensed_tree, min_samples, tree_type='kdtree', metric='euclidean', **kwargs): self.raw_data = data self.tree = self._tree_type_map[tree_type](self.raw_data, metric=metric, **kwargs) self.core_distances = self.tree.query(data, k=min_samples)[0][:, -1] self.dist_metric = DistanceMetric.get_metric(metric, **kwargs) selected_clusters = condensed_tree._select_clusters() # raw_condensed_tree = condensed_tree.to_numpy() raw_condensed_tree = condensed_tree._raw_tree self.cluster_map = {c: n for n, c in enumerate(sorted(list(selected_clusters)))} self.reverse_cluster_map = {n: c for c, n in self.cluster_map.items()} self.cluster_tree = raw_condensed_tree[raw_condensed_tree['child_size'] > 1] self.max_lambdas = {} self.leaf_max_lambdas = {} self.exemplars = [] all_clusters = set(np.hstack([self.cluster_tree['parent'], self.cluster_tree['child']])) for cluster in all_clusters: self.leaf_max_lambdas[cluster] = raw_condensed_tree['lambda_val'][ raw_condensed_tree['parent'] == cluster].max() for cluster in selected_clusters: self.max_lambdas[cluster] = \ raw_condensed_tree['lambda_val'][raw_condensed_tree['parent'] == cluster].max() for sub_cluster in self._clusters_below(cluster): self.cluster_map[sub_cluster] = self.cluster_map[cluster] self.max_lambdas[sub_cluster] = self.max_lambdas[cluster] cluster_exemplars = np.array([], dtype=np.int64) for leaf in self._recurse_leaf_dfs(cluster): leaf_max_lambda = raw_condensed_tree['lambda_val'][ raw_condensed_tree['parent'] == leaf].max() points = raw_condensed_tree['child'][ (raw_condensed_tree['parent'] == leaf) & (raw_condensed_tree['lambda_val'] == leaf_max_lambda)] cluster_exemplars = np.hstack([cluster_exemplars, points]) self.exemplars.append(self.raw_data[cluster_exemplars]) def _find_neighbor_and_lambda(neighbor_indices, neighbor_distances, core_distances, min_samples): """ Find the nearest mutual reachability neighbor of a point, and compute the associated lambda value for the point, given the mutual reachability distance to a nearest neighbor. Parameters ---------- neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points min_samples : int The min_samples value used to generate core distances. Returns ------- neighbor : int The index into the full raw data set of the nearest mutual reachability distance neighbor of the point. lambda_ : float The lambda value at which this point joins/merges with `neighbor`. """ neighbor_core_distances = core_distances[neighbor_indices] point_core_distances = neighbor_distances[min_samples] * np.ones( neighbor_indices.shape[0]) mr_distances = np.vstack(( neighbor_core_distances, point_core_distances, neighbor_distances )).max(axis=0) nn_index = mr_distances.argmin() nearest_neighbor = neighbor_indices[nn_index] if mr_distances[nn_index] > 0.0: lambda_ = 1. / mr_distances[nn_index] else: lambda_ = np.finfo(np.double).max return nearest_neighbor, lambda_ def _extend_condensed_tree(tree, neighbor_indices, neighbor_distances, core_distances, min_samples): """ Create a new condensed tree with an additional point added, allowing for computations as if this point had been part of the original tree. Note that this makes as little change to the tree as possible, with no re-optimizing/re-condensing so that the selected clusters remain effectively unchanged. Parameters ---------- tree : structured array The raw format condensed tree to update. neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points min_samples : int The min_samples value used to generate core distances. Returns ------- new_tree : structured array The original tree with an extra row providing the parent cluster and lambda information for a new point given index -1. """ tree_root = tree['parent'].min() nearest_neighbor, lambda_ = _find_neighbor_and_lambda(neighbor_indices, neighbor_distances, core_distances, min_samples ) neighbor_tree_row = get_tree_row_with_child(tree, nearest_neighbor) potential_cluster = neighbor_tree_row['parent'] if neighbor_tree_row['lambda_val'] <= lambda_: # New point departs with the old new_tree_row = (potential_cluster, -1, 1, neighbor_tree_row['lambda_val']) else: # Find appropriate cluster based on lambda of new point while potential_cluster > tree_root and \ tree[tree['child'] == potential_cluster]['lambda_val'] >= lambda_: potential_cluster = tree['parent'][tree['child'] == potential_cluster][0] new_tree_row = (potential_cluster, -1, 1, lambda_) return np.append(tree, new_tree_row) def _find_cluster_and_probability(tree, cluster_tree, neighbor_indices, neighbor_distances, core_distances, cluster_map, max_lambdas, min_samples): """ Return the cluster label (of the original clustering) and membership probability of a new data point. Parameters ---------- tree : CondensedTree The condensed tree associated with the clustering. cluster_tree : structured_array The raw form of the condensed tree with only cluster information (no data on individual points). This is significantly more compact. neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points cluster_map : dict A dictionary mapping cluster numbers in the condensed tree to labels in the final selected clustering. max_lambdas : dict A dictionary mapping cluster numbers in the condensed tree to the maximum lambda value seen in that cluster. min_samples : int The min_samples value used to generate core distances. """ raw_tree = tree._raw_tree tree_root = cluster_tree['parent'].min() nearest_neighbor, lambda_ = _find_neighbor_and_lambda(neighbor_indices, neighbor_distances, core_distances, min_samples ) neighbor_tree_row = get_tree_row_with_child(raw_tree, nearest_neighbor) potential_cluster = neighbor_tree_row['parent'] if neighbor_tree_row['lambda_val'] > lambda_: # Find appropriate cluster based on lambda of new point while potential_cluster > tree_root and \ cluster_tree['lambda_val'][cluster_tree['child'] == potential_cluster] >= lambda_: potential_cluster = cluster_tree['parent'][cluster_tree['child'] == potential_cluster][0] if potential_cluster in cluster_map: cluster_label = cluster_map[potential_cluster] else: cluster_label = -1 if cluster_label >= 0: max_lambda = max_lambdas[potential_cluster] if max_lambda > 0.0: lambda_ = min(max_lambda, lambda_) prob = (lambda_ / max_lambda) else: prob = 1.0 else: prob = 0.0 return cluster_label, prob def approximate_predict(clusterer, points_to_predict): """Predict the cluster label of new points. The returned labels will be those of the original clustering found by ``clusterer``, and therefore are not (necessarily) the cluster labels that would be found by clustering the original data combined with ``points_to_predict``, hence the 'approximate' label. If you simply wish to assign new points to an existing clustering in the 'best' way possible, this is the function to use. If you want to predict how ``points_to_predict`` would cluster with the original data under HDBSCAN the most efficient existing approach is to simply recluster with the new point(s) added to the original dataset. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. points_to_predict : array, or array-like (n_samples, n_features) The new data points to predict cluster labels for. They should have the same dimensionality as the original dataset over which clusterer was fit. Returns ------- labels : array (n_samples,) The predicted labels of the ``points_to_predict`` probabilities : array (n_samples,) The soft cluster scores for each of the ``points_to_predict`` See Also -------- :py:func:`hdbscan.predict.membership_vector` :py:func:`hdbscan.predict.all_points_membership_vectors` """ if clusterer.prediction_data_ is None: raise ValueError('Clusterer does not have prediction data!' ' Try fitting with prediction_data=True set,' ' or run generate_prediction_data on the clusterer') points_to_predict = np.asarray(points_to_predict) if points_to_predict.shape[1] != \ clusterer.prediction_data_.raw_data.shape[1]: raise ValueError('New points dimension does not match fit data!') if clusterer.prediction_data_.cluster_tree.shape[0] == 0: warn('Clusterer does not have any defined clusters, new data' ' will be automatically predicted as noise.') labels = -1 * np.ones(points_to_predict.shape[0], dtype=np.int32) probabilities = np.zeros(points_to_predict.shape[0], dtype=np.float32) return labels, probabilities labels = np.empty(points_to_predict.shape[0], dtype=np.int) probabilities = np.empty(points_to_predict.shape[0], dtype=np.float64) min_samples = clusterer.min_samples or clusterer.min_cluster_size neighbor_distances, neighbor_indices = \ clusterer.prediction_data_.tree.query(points_to_predict, k=2 * min_samples) for i in range(points_to_predict.shape[0]): label, prob = _find_cluster_and_probability( clusterer.condensed_tree_, clusterer.prediction_data_.cluster_tree, neighbor_indices[i], neighbor_distances[i], clusterer.prediction_data_.core_distances, clusterer.prediction_data_.cluster_map, clusterer.prediction_data_.max_lambdas, min_samples ) labels[i] = label probabilities[i] = prob return labels, probabilities def membership_vector(clusterer, points_to_predict): """Predict soft cluster membership. The result produces a vector for each point in ``points_to_predict`` that gives a probability that the given point is a member of a cluster for each of the selected clusters of the ``clusterer``. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. points_to_predict : array, or array-like (n_samples, n_features) The new data points to predict cluster labels for. They should have the same dimensionality as the original dataset over which clusterer was fit. Returns ------- membership_vectors : array (n_samples, n_clusters) The probability that point ``i`` is a member of cluster ``j`` is in ``membership_vectors[i, j]``. See Also -------- :py:func:`hdbscan.predict.predict` :py:func:`hdbscan.predict.all_points_membership_vectors` """ clusters = np.array( sorted(list(clusterer.condensed_tree_._select_clusters()))).astype(np.intp) result = np.empty((points_to_predict.shape[0], clusters.shape[0]), dtype=np.float64) min_samples = clusterer.min_samples or clusterer.min_cluster_size neighbor_distances, neighbor_indices = \ clusterer.prediction_data_.tree.query(points_to_predict, k=2 * min_samples) for i in range(points_to_predict.shape[0]): # We need to find where in the tree the new point would go # for the purposes of outlier membership approximation nearest_neighbor, lambda_ = \ _find_neighbor_and_lambda( neighbor_indices[i], neighbor_distances[i], clusterer.prediction_data_.core_distances, min_samples) neighbor_tree_row = get_tree_row_with_child( clusterer.condensed_tree_._raw_tree, nearest_neighbor) if neighbor_tree_row['lambda_val'] <= lambda_: lambda_ = neighbor_tree_row['lambda_val'] distance_vec = dist_membership_vector( points_to_predict[i], clusterer.prediction_data_.exemplars, clusterer.prediction_data_.dist_metric) outlier_vec = outlier_membership_vector( nearest_neighbor, lambda_, clusters, clusterer.condensed_tree_._raw_tree, clusterer.prediction_data_.leaf_max_lambdas, clusterer.prediction_data_.cluster_tree) result[i] = distance_vec ** 0.5 * outlier_vec ** 2.0 result[i] /= result[i].sum() result[i] *= prob_in_some_cluster( nearest_neighbor, lambda_, clusters, clusterer.condensed_tree_._raw_tree, clusterer.prediction_data_.leaf_max_lambdas, clusterer.prediction_data_.cluster_tree) return result def all_points_membership_vectors(clusterer): """Predict soft cluster membership vectors for all points in the original dataset the clusterer was trained on. This function is more efficient by making use of the fact that all points are already in the condensed tree, and processing in bulk. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. This method does not work if the clusterer was trained with ``metric='precomputed'``. Returns ------- membership_vectors : array (n_samples, n_clusters) The probability that point ``i`` of the original dataset is a member of cluster ``j`` is in ``membership_vectors[i, j]``. See Also -------- :py:func:`hdbscan.predict.predict` :py:func:`hdbscan.predict.all_points_membership_vectors` """ clusters = np.array(sorted(list(clusterer.condensed_tree_._select_clusters()))).astype(np.intp) all_points = clusterer.prediction_data_.raw_data # When no clusters found, return array of 0's if clusters.size == 0: return np.zeros(all_points.shape[0]) distance_vecs = all_points_dist_membership_vector( all_points, clusterer.prediction_data_.exemplars, clusterer.prediction_data_.dist_metric) outlier_vecs = all_points_outlier_membership_vector( clusters, clusterer.condensed_tree_._raw_tree, clusterer.prediction_data_.leaf_max_lambdas, clusterer.prediction_data_.cluster_tree) in_cluster_probs = all_points_prob_in_some_cluster( clusters, clusterer.condensed_tree_._raw_tree, clusterer.prediction_data_.leaf_max_lambdas, clusterer.prediction_data_.cluster_tree) result = distance_vecs * outlier_vecs row_sums = result.sum(axis=1) result = result / row_sums[:, np.newaxis] result *= in_cluster_probs[:, np.newaxis] return result
hdbscan/prediction.py
21,034
Extra data that allows for faster prediction if cached. Parameters ---------- data : array (n_samples, n_features) The original data set that was clustered condensed_tree : CondensedTree The condensed tree object created by a clustering min_samples : int The min_samples value used in clustering tree_type : string, optional Which type of space tree to use for core distance computation. One of: * ``kdtree`` * ``balltree`` metric : string, optional The metric used to determine distance for the clustering. This is the metric that will be used for the space tree to determine core distances etc. **kwargs : Any further arguments to the metric. Attributes ---------- raw_data : array (n_samples, n_features) The original data set that was clustered tree : KDTree or BallTree A space partitioning tree that can be queried for nearest neighbors. core_distances : array (n_samples,) The core distances for every point in the original data set. cluster_map : dict A dictionary mapping cluster numbers in the condensed tree to labels in the final selected clustering. cluster_tree : structured array A version of the condensed tree that only contains clusters, not individual points. max_lambdas : dict A dictionary mapping cluster numbers in the condensed tree to the maximum lambda value seen in that cluster. Create a new condensed tree with an additional point added, allowing for computations as if this point had been part of the original tree. Note that this makes as little change to the tree as possible, with no re-optimizing/re-condensing so that the selected clusters remain effectively unchanged. Parameters ---------- tree : structured array The raw format condensed tree to update. neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points min_samples : int The min_samples value used to generate core distances. Returns ------- new_tree : structured array The original tree with an extra row providing the parent cluster and lambda information for a new point given index -1. Return the cluster label (of the original clustering) and membership probability of a new data point. Parameters ---------- tree : CondensedTree The condensed tree associated with the clustering. cluster_tree : structured_array The raw form of the condensed tree with only cluster information (no data on individual points). This is significantly more compact. neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points cluster_map : dict A dictionary mapping cluster numbers in the condensed tree to labels in the final selected clustering. max_lambdas : dict A dictionary mapping cluster numbers in the condensed tree to the maximum lambda value seen in that cluster. min_samples : int The min_samples value used to generate core distances. Find the nearest mutual reachability neighbor of a point, and compute the associated lambda value for the point, given the mutual reachability distance to a nearest neighbor. Parameters ---------- neighbor_indices : array (2 * min_samples, ) An array of raw distance based nearest neighbor indices. neighbor_distances : array (2 * min_samples, ) An array of raw distances to the nearest neighbors. core_distances : array (n_samples, ) An array of core distances for all points min_samples : int The min_samples value used to generate core distances. Returns ------- neighbor : int The index into the full raw data set of the nearest mutual reachability distance neighbor of the point. lambda_ : float The lambda value at which this point joins/merges with `neighbor`. Predict soft cluster membership vectors for all points in the original dataset the clusterer was trained on. This function is more efficient by making use of the fact that all points are already in the condensed tree, and processing in bulk. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. This method does not work if the clusterer was trained with ``metric='precomputed'``. Returns ------- membership_vectors : array (n_samples, n_clusters) The probability that point ``i`` of the original dataset is a member of cluster ``j`` is in ``membership_vectors[i, j]``. See Also -------- :py:func:`hdbscan.predict.predict` :py:func:`hdbscan.predict.all_points_membership_vectors` Predict the cluster label of new points. The returned labels will be those of the original clustering found by ``clusterer``, and therefore are not (necessarily) the cluster labels that would be found by clustering the original data combined with ``points_to_predict``, hence the 'approximate' label. If you simply wish to assign new points to an existing clustering in the 'best' way possible, this is the function to use. If you want to predict how ``points_to_predict`` would cluster with the original data under HDBSCAN the most efficient existing approach is to simply recluster with the new point(s) added to the original dataset. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. points_to_predict : array, or array-like (n_samples, n_features) The new data points to predict cluster labels for. They should have the same dimensionality as the original dataset over which clusterer was fit. Returns ------- labels : array (n_samples,) The predicted labels of the ``points_to_predict`` probabilities : array (n_samples,) The soft cluster scores for each of the ``points_to_predict`` See Also -------- :py:func:`hdbscan.predict.membership_vector` :py:func:`hdbscan.predict.all_points_membership_vectors` Predict soft cluster membership. The result produces a vector for each point in ``points_to_predict`` that gives a probability that the given point is a member of a cluster for each of the selected clusters of the ``clusterer``. Parameters ---------- clusterer : HDBSCAN A clustering object that has been fit to the data and either had ``prediction_data=True`` set, or called the ``generate_prediction_data`` method after the fact. points_to_predict : array, or array-like (n_samples, n_features) The new data points to predict cluster labels for. They should have the same dimensionality as the original dataset over which clusterer was fit. Returns ------- membership_vectors : array (n_samples, n_clusters) The probability that point ``i`` is a member of cluster ``j`` is in ``membership_vectors[i, j]``. See Also -------- :py:func:`hdbscan.predict.predict` :py:func:`hdbscan.predict.all_points_membership_vectors` Support various prediction methods for predicting cluster membership of new or unseen points. There are several ways to interpret how to do this correctly, so we provide several methods for the different use cases that may arise. raw_condensed_tree = condensed_tree.to_numpy() New point departs with the old Find appropriate cluster based on lambda of new point Find appropriate cluster based on lambda of new point We need to find where in the tree the new point would go for the purposes of outlier membership approximation When no clusters found, return array of 0's
7,951
en
0.819102
""" Views for PubSite app. """ from django.conf import settings from django.contrib.auth.views import ( PasswordResetView, PasswordResetDoneView, PasswordResetConfirmView, PasswordResetCompleteView, ) from django.shortcuts import render import requests import logging logger = logging.getLogger(__name__) def _get_context(page_name): return { "pages": settings.PUBLIC_PAGES, "current_page_name": page_name, } # Regular index # def index(request): # """ # View for the static index page # """ # return render(request, 'public/home.html', _get_context('Home')) def index(request): """ View for the static index page """ return render(request, "public/home.html", _get_context("Home")) def about(request): """ View for the static chapter history page. """ return render(request, "public/about.html", _get_context("About")) def activities(request): """ View for the static chapter service page. """ return render( request, "public/activities.html", _get_context("Service & Activities"), ) def rush(request): """ View for the static chapter service page. """ return render( request, "public/rush.html", _get_context("Rush"), ) def campaign(request): """ View for the campaign service page. """ # Overrride requests Session authentication handling class NoRebuildAuthSession(requests.Session): def rebuild_auth(self, prepared_request, response): """ No code here means requests will always preserve the Authorization header when redirected. Be careful not to leak your credentials to untrusted hosts! """ url = "https://api.givebutter.com/v1/transactions/" headers = {"Authorization": f"Bearer {settings.GIVEBUTTER_API_KEY}"} response = None # Create custom requests session session = NoRebuildAuthSession() # Make GET request to server, timeout in seconds try: r = session.get(url, headers=headers, timeout=0.75) if r.status_code == 200: response = r.json() else: logger.error(f"ERROR in request: {r.status_code}") except requests.exceptions.Timeout: logger.warning("Connection to GiveButter API Timed out") except requests.ConnectionError: logger.warning("Connection to GiveButter API could not be resolved") except requests.exceptions.RequestException: logger.error( "An unknown issue occurred while trying to retrieve GiveButter Donor List" ) # Grab context object to use later ctx = _get_context("Campaign") # Check for successful response, if so - filter, sort, and format data if response and "data" in response: response = response["data"] # Pull data from GET response object logger.debug(f"GiveButter API Response: {response}") # Filter by only successful transactions, then sort by amount descending successful_txs = [tx for tx in response if tx["status"] == "succeeded"] sorted_txs = sorted(successful_txs, key=lambda tx: tx["amount"], reverse=True) # Clean data to a list of dictionaries & remove unnecessary data transactions = [ { "name": tx["giving_space"]["name"], "amount": tx["giving_space"]["amount"], "message": tx["giving_space"]["message"], } for tx in sorted_txs[:20] ] # Attach transaction dictionary & length to context object ctx["transactions"] = transactions ctx["num_txs"] = len(successful_txs) return render( request, "public/campaign.html", ctx, ) def permission_denied(request): """ View for 403 (Permission Denied) error. """ return render( request, "common/403.html", _get_context("Permission Denied"), ) def handler404(request, exception): """ """ return render(request, "common/404.html", _get_context("Page Not Found")) class ResetPassword(PasswordResetView): template_name = "password_reset/password_reset_form.html" class ResetPasswordDone(PasswordResetDoneView): template_name = "password_reset/password_reset_done.html" class ResetPasswordConfirm(PasswordResetConfirmView): template_name = "password_reset/password_reset_confirm.html" class ResetPasswordComplete(PasswordResetCompleteView): template_name = "password_reset/password_reset_complete.html"
sigmapiweb/apps/PubSite/views.py
4,610
View for the static chapter history page. View for the static chapter service page. View for the campaign service page. View for the static index page View for 403 (Permission Denied) error. No code here means requests will always preserve the Authorization header when redirected. Be careful not to leak your credentials to untrusted hosts! View for the static chapter service page. Views for PubSite app. Regular index def index(request): """ View for the static index page """ return render(request, 'public/home.html', _get_context('Home')) Overrride requests Session authentication handling Create custom requests session Make GET request to server, timeout in seconds Grab context object to use later Check for successful response, if so - filter, sort, and format data Pull data from GET response object Filter by only successful transactions, then sort by amount descending Clean data to a list of dictionaries & remove unnecessary data Attach transaction dictionary & length to context object
1,019
en
0.732355
import requests import datetime class BearerAuth(requests.auth.AuthBase): def __init__(self, token): self.token = token def __call__(self, r): r.headers["authorization"] = "Bearer " + self.token return r class MintMobile: def __init__(self, phone_number, password): self.phone_number = phone_number self.password = password self.token="" self.id="" self.family_members=[] self.info={} def login(self): #print("Logging Into " + self.phone_number) r=requests.post('https://w3b-api.ultramobile.com/v1/mint/login?', json = {"msisdn":self.phone_number,"password":self.password}) if r.status_code == 200: response=r.json() self.id=response['id'] self.token=response['token'] self.info[self.id]={"phone_number":self.phone_number} self.master_account_details() return True else: return False def master_account_details(self): r=requests.get('https://w3b-api.ultramobile.com/v1/mint/account/'+str(self.id)+'?', auth=BearerAuth(str(self.token))) response=r.json() self.info[self.id]['line_name']=response['firstName'] self.info[self.id]['endOfCycle']=self.epoch_days_remaining(response['plan']['endOfCycle']) self.info[self.id]['months']=response['plan']['months'] self.info[self.id]['exp']=self.epoch_days_remaining(response['plan']['exp']) def data_remaining(self): r=requests.get('https://w3b-api.ultramobile.com/v1/mint/account/'+str(self.id)+'/data?', auth=BearerAuth(str(self.token))) response=r.json() response['remaining4G_GB']=self.conv_MB_to_GB(response['remaining4G']) self.info[self.id]['remaining4G']=response['remaining4G_GB'] return self.info def conv_MB_to_GB(self,input_megabyte): gigabyte = 1.0/1024 convert_gb = gigabyte * input_megabyte convert_gb=round(convert_gb, 2) return convert_gb def epoch_days_remaining(self,epoch): dt1 = datetime.datetime.fromtimestamp(epoch) dt2 = datetime.datetime.now() delta = dt1 - dt2 return delta.days def get_family_members(self): r=requests.get('https://w3b-api.ultramobile.com/v1/mint/account/'+str(self.id)+'/multi-line?', auth=BearerAuth(str(self.token))) response=r.json() for activeMembers in response['activeMembers']: self.family_members.append(activeMembers['id']) self.info[activeMembers['id']]={} #self.info[activeMembers['id']]={"phone_number":activeMembers['msisdn'],"line_name":activeMembers['nickName']} self.info[activeMembers['id']]["phone_number"]=activeMembers['msisdn'] self.info[activeMembers['id']]["line_name"]=activeMembers['nickName'] self.info[activeMembers['id']]["endOfCycle"]=self.epoch_days_remaining(activeMembers['currentPlan']["rechargeDate"]) self.info[activeMembers['id']]["months"]=activeMembers['currentPlan']["duration"] self.info[activeMembers['id']]["exp"]=self.epoch_days_remaining(activeMembers['nextPlan']["renewalDate"]) self.family_data_remaining() def family_data_remaining(self): for member in self.family_members: r=requests.get('https://w3b-api.ultramobile.com/v1/mint/account/'+self.id+'/multi-line/'+member+'/usage?', auth=BearerAuth(str(self.token))) response=r.json() response['remaining4G_GB']=self.conv_MB_to_GB(response['data']['remaining4G']) self.info[member]['remaining4G']=response['remaining4G_GB'] def get_all_data_remaining(self): self.login() self.data_remaining() self.get_family_members() return self.info def lines(self): self.login() self.get_family_members() return self.info.keys()
custom_components/mintmobile/api.py
3,928
print("Logging Into " + self.phone_number)self.info[activeMembers['id']]={"phone_number":activeMembers['msisdn'],"line_name":activeMembers['nickName']}
151
en
0.683048
import re class Normalizer: """Normalizer return the text replaced with 'repl'. If 'repl' is None, normalization is not applied to the pattern corresponding to 'repl'. Args: url_repl (str): replace all urls in text with this tag_repl (str): replace all tags in text with this emoji_repl (str): replace all emojis in text with this email_repl (str): replace all emails in text with this tel_repl (str): replace all tels in text with this """ def __init__(self, url_repl='[URL]', tag_repl='[TAG]', emoji_repl='[EMOJI]', email_repl='[EMAIL]', tel_repl='[TEL]'): # repls self.url_repl = url_repl self.tag_repl = tag_repl self.emoji_repl = emoji_repl self.email_repl = email_repl self.tel_repl = tel_repl self._normalize = [] self._init_normalize() def normalize(self, text: str) -> str: """Normalize text. Args: text (str): text to be normalized """ for normalize_fn, repl in self._normalize: text = normalize_fn(text, repl) return text def _init_normalize(self) -> None: """Initialize normalize function. If 'repl' is None, normalization is not applied to the pattern corresponding to 'repl'. """ if self.url_repl is not None: self._normalize.append((self._url_normalize, self.url_repl)) if self.tag_repl is not None: self._normalize.append((self._tag_normalize, self.tag_repl)) if self.emoji_repl is not None: self._normalize.append((self._emoji_normalize, self.emoji_repl)) if self.email_repl is not None: self._normalize.append((self._email_normalize, self.email_repl)) if self.tel_repl is not None: self._normalize.append((self._tel_normalize, self.tel_repl)) def _url_normalize(self, text: str, repl: str, regex=re.compile(r'(https?|ftp|www)\S+')) -> str: """Return the string obtained by replacing all urls in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all urls in text with 'repl' """ text = regex.sub(repl, text) return text def _tag_normalize(self, text: str, repl: str, regex=re.compile(r'<[^>]*>')) -> str: """Return the string obtained by replacing all HTML tags in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all HTML tags in text with 'repl' """ text = regex.sub(repl, text) return text def _emoji_normalize(self, text: str, repl: str, regex=re.compile(r'\U0001f469\u200d\u2764\ufe0f\u200d\U0001f48b\u200d\U0001f468|\U0001f468\u200d\u2764\ufe0f\u200d\U0001f48b\u200d\U0001f468|\U0001f469\u200d\u2764\ufe0f\u200d\U0001f48b\u200d\U0001f469|\U0001f9d1\U0001f3fb\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3fc\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3fc\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fc|\U0001f9d1\U0001f3fd\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3fd\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fc|\U0001f9d1\U0001f3fd\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fd|\U0001f9d1\U0001f3fe\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3fe\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fc|\U0001f9d1\U0001f3fe\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fd|\U0001f9d1\U0001f3fe\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fe|\U0001f9d1\U0001f3ff\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3ff\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fc|\U0001f9d1\U0001f3ff\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fd|\U0001f9d1\U0001f3ff\u200d\U0001f91d\u200d\U0001f9d1\U0001f3fe|\U0001f9d1\U0001f3ff\u200d\U0001f91d\u200d\U0001f9d1\U0001f3ff|\U0001f469\U0001f3fc\u200d\U0001f91d\u200d\U0001f469\U0001f3fb|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f469\U0001f3fb|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f469\U0001f3fc|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f469\U0001f3fb|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f469\U0001f3fc|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f469\U0001f3fd|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f469\U0001f3fb|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f469\U0001f3fc|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f469\U0001f3fd|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f469\U0001f3fe|\U0001f469\U0001f3fb\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f469\U0001f3fb\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f469\U0001f3fb\u200d\U0001f91d\u200d\U0001f468\U0001f3fe|\U0001f469\U0001f3fb\u200d\U0001f91d\u200d\U0001f468\U0001f3ff|\U0001f469\U0001f3fc\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f469\U0001f3fc\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f469\U0001f3fc\u200d\U0001f91d\u200d\U0001f468\U0001f3fe|\U0001f469\U0001f3fc\u200d\U0001f91d\u200d\U0001f468\U0001f3ff|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3fe|\U0001f469\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3ff|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f469\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3ff|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f469\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fe|\U0001f468\U0001f3fc\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f468\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f468\U0001f3fd\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f468\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f468\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f468\U0001f3fe\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f468\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fb|\U0001f468\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fc|\U0001f468\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fd|\U0001f468\U0001f3ff\u200d\U0001f91d\u200d\U0001f468\U0001f3fe|\U0001f469\u200d\u2764\u200d\U0001f48b\u200d\U0001f468|\U0001f468\u200d\u2764\u200d\U0001f48b\u200d\U0001f468|\U0001f469\u200d\u2764\u200d\U0001f48b\u200d\U0001f469|\U0001f468\u200d\U0001f469\u200d\U0001f467\u200d\U0001f466|\U0001f468\u200d\U0001f469\u200d\U0001f466\u200d\U0001f466|\U0001f468\u200d\U0001f469\u200d\U0001f467\u200d\U0001f467|\U0001f468\u200d\U0001f468\u200d\U0001f467\u200d\U0001f466|\U0001f468\u200d\U0001f468\u200d\U0001f466\u200d\U0001f466|\U0001f468\u200d\U0001f468\u200d\U0001f467\u200d\U0001f467|\U0001f469\u200d\U0001f469\u200d\U0001f467\u200d\U0001f466|\U0001f469\u200d\U0001f469\u200d\U0001f466\u200d\U0001f466|\U0001f469\u200d\U0001f469\u200d\U0001f467\u200d\U0001f467|\U0001f3f4\U000e0067\U000e0062\U000e0065\U000e006e\U000e0067\U000e007f|\U0001f3f4\U000e0067\U000e0062\U000e0073\U000e0063\U000e0074\U000e007f|\U0001f3f4\U000e0067\U000e0062\U000e0077\U000e006c\U000e0073\U000e007f|\U0001f469\u200d\u2764\ufe0f\u200d\U0001f468|\U0001f468\u200d\u2764\ufe0f\u200d\U0001f468|\U0001f469\u200d\u2764\ufe0f\u200d\U0001f469|\U0001f441\ufe0f\u200d\U0001f5e8\ufe0f|\U0001f471\U0001f3fb\u200d\u2642\ufe0f|\U0001f471\U0001f3fc\u200d\u2642\ufe0f|\U0001f471\U0001f3fd\u200d\u2642\ufe0f|\U0001f471\U0001f3fe\u200d\u2642\ufe0f|\U0001f471\U0001f3ff\u200d\u2642\ufe0f|\U0001f471\U0001f3fb\u200d\u2640\ufe0f|\U0001f471\U0001f3fc\u200d\u2640\ufe0f|\U0001f471\U0001f3fd\u200d\u2640\ufe0f|\U0001f471\U0001f3fe\u200d\u2640\ufe0f|\U0001f471\U0001f3ff\u200d\u2640\ufe0f|\U0001f64d\U0001f3fb\u200d\u2642\ufe0f|\U0001f64d\U0001f3fc\u200d\u2642\ufe0f|\U0001f64d\U0001f3fd\u200d\u2642\ufe0f|\U0001f64d\U0001f3fe\u200d\u2642\ufe0f|\U0001f64d\U0001f3ff\u200d\u2642\ufe0f|\U0001f64d\U0001f3fb\u200d\u2640\ufe0f|\U0001f64d\U0001f3fc\u200d\u2640\ufe0f|\U0001f64d\U0001f3fd\u200d\u2640\ufe0f|\U0001f64d\U0001f3fe\u200d\u2640\ufe0f|\U0001f64d\U0001f3ff\u200d\u2640\ufe0f|\U0001f64e\U0001f3fb\u200d\u2642\ufe0f|\U0001f64e\U0001f3fc\u200d\u2642\ufe0f|\U0001f64e\U0001f3fd\u200d\u2642\ufe0f|\U0001f64e\U0001f3fe\u200d\u2642\ufe0f|\U0001f64e\U0001f3ff\u200d\u2642\ufe0f|\U0001f64e\U0001f3fb\u200d\u2640\ufe0f|\U0001f64e\U0001f3fc\u200d\u2640\ufe0f|\U0001f64e\U0001f3fd\u200d\u2640\ufe0f|\U0001f64e\U0001f3fe\u200d\u2640\ufe0f|\U0001f64e\U0001f3ff\u200d\u2640\ufe0f|\U0001f645\U0001f3fb\u200d\u2642\ufe0f|\U0001f645\U0001f3fc\u200d\u2642\ufe0f|\U0001f645\U0001f3fd\u200d\u2642\ufe0f|\U0001f645\U0001f3fe\u200d\u2642\ufe0f|\U0001f645\U0001f3ff\u200d\u2642\ufe0f|\U0001f645\U0001f3fb\u200d\u2640\ufe0f|\U0001f645\U0001f3fc\u200d\u2640\ufe0f|\U0001f645\U0001f3fd\u200d\u2640\ufe0f|\U0001f645\U0001f3fe\u200d\u2640\ufe0f|\U0001f645\U0001f3ff\u200d\u2640\ufe0f|\U0001f646\U0001f3fb\u200d\u2642\ufe0f|\U0001f646\U0001f3fc\u200d\u2642\ufe0f|\U0001f646\U0001f3fd\u200d\u2642\ufe0f|\U0001f646\U0001f3fe\u200d\u2642\ufe0f|\U0001f646\U0001f3ff\u200d\u2642\ufe0f|\U0001f646\U0001f3fb\u200d\u2640\ufe0f|\U0001f646\U0001f3fc\u200d\u2640\ufe0f|\U0001f646\U0001f3fd\u200d\u2640\ufe0f|\U0001f646\U0001f3fe\u200d\u2640\ufe0f|\U0001f646\U0001f3ff\u200d\u2640\ufe0f|\U0001f481\U0001f3fb\u200d\u2642\ufe0f|\U0001f481\U0001f3fc\u200d\u2642\ufe0f|\U0001f481\U0001f3fd\u200d\u2642\ufe0f|\U0001f481\U0001f3fe\u200d\u2642\ufe0f|\U0001f481\U0001f3ff\u200d\u2642\ufe0f|\U0001f481\U0001f3fb\u200d\u2640\ufe0f|\U0001f481\U0001f3fc\u200d\u2640\ufe0f|\U0001f481\U0001f3fd\u200d\u2640\ufe0f|\U0001f481\U0001f3fe\u200d\u2640\ufe0f|\U0001f481\U0001f3ff\u200d\u2640\ufe0f|\U0001f64b\U0001f3fb\u200d\u2642\ufe0f|\U0001f64b\U0001f3fc\u200d\u2642\ufe0f|\U0001f64b\U0001f3fd\u200d\u2642\ufe0f|\U0001f64b\U0001f3fe\u200d\u2642\ufe0f|\U0001f64b\U0001f3ff\u200d\u2642\ufe0f|\U0001f64b\U0001f3fb\u200d\u2640\ufe0f|\U0001f64b\U0001f3fc\u200d\u2640\ufe0f|\U0001f64b\U0001f3fd\u200d\u2640\ufe0f|\U0001f64b\U0001f3fe\u200d\u2640\ufe0f|\U0001f64b\U0001f3ff\u200d\u2640\ufe0f|\U0001f9cf\U0001f3fb\u200d\u2642\ufe0f|\U0001f9cf\U0001f3fc\u200d\u2642\ufe0f|\U0001f9cf\U0001f3fd\u200d\u2642\ufe0f|\U0001f9cf\U0001f3fe\u200d\u2642\ufe0f|\U0001f9cf\U0001f3ff\u200d\u2642\ufe0f|\U0001f9cf\U0001f3fb\u200d\u2640\ufe0f|\U0001f9cf\U0001f3fc\u200d\u2640\ufe0f|\U0001f9cf\U0001f3fd\u200d\u2640\ufe0f|\U0001f9cf\U0001f3fe\u200d\u2640\ufe0f|\U0001f9cf\U0001f3ff\u200d\u2640\ufe0f|\U0001f647\U0001f3fb\u200d\u2642\ufe0f|\U0001f647\U0001f3fc\u200d\u2642\ufe0f|\U0001f647\U0001f3fd\u200d\u2642\ufe0f|\U0001f647\U0001f3fe\u200d\u2642\ufe0f|\U0001f647\U0001f3ff\u200d\u2642\ufe0f|\U0001f647\U0001f3fb\u200d\u2640\ufe0f|\U0001f647\U0001f3fc\u200d\u2640\ufe0f|\U0001f647\U0001f3fd\u200d\u2640\ufe0f|\U0001f647\U0001f3fe\u200d\u2640\ufe0f|\U0001f647\U0001f3ff\u200d\u2640\ufe0f|\U0001f926\U0001f3fb\u200d\u2642\ufe0f|\U0001f926\U0001f3fc\u200d\u2642\ufe0f|\U0001f926\U0001f3fd\u200d\u2642\ufe0f|\U0001f926\U0001f3fe\u200d\u2642\ufe0f|\U0001f926\U0001f3ff\u200d\u2642\ufe0f|\U0001f926\U0001f3fb\u200d\u2640\ufe0f|\U0001f926\U0001f3fc\u200d\u2640\ufe0f|\U0001f926\U0001f3fd\u200d\u2640\ufe0f|\U0001f926\U0001f3fe\u200d\u2640\ufe0f|\U0001f926\U0001f3ff\u200d\u2640\ufe0f|\U0001f937\U0001f3fb\u200d\u2642\ufe0f|\U0001f937\U0001f3fc\u200d\u2642\ufe0f|\U0001f937\U0001f3fd\u200d\u2642\ufe0f|\U0001f937\U0001f3fe\u200d\u2642\ufe0f|\U0001f937\U0001f3ff\u200d\u2642\ufe0f|\U0001f937\U0001f3fb\u200d\u2640\ufe0f|\U0001f937\U0001f3fc\u200d\u2640\ufe0f|\U0001f937\U0001f3fd\u200d\u2640\ufe0f|\U0001f937\U0001f3fe\u200d\u2640\ufe0f|\U0001f937\U0001f3ff\u200d\u2640\ufe0f|\U0001f468\U0001f3fb\u200d\u2695\ufe0f|\U0001f468\U0001f3fc\u200d\u2695\ufe0f|\U0001f468\U0001f3fd\u200d\u2695\ufe0f|\U0001f468\U0001f3fe\u200d\u2695\ufe0f|\U0001f468\U0001f3ff\u200d\u2695\ufe0f|\U0001f469\U0001f3fb\u200d\u2695\ufe0f|\U0001f469\U0001f3fc\u200d\u2695\ufe0f|\U0001f469\U0001f3fd\u200d\u2695\ufe0f|\U0001f469\U0001f3fe\u200d\u2695\ufe0f|\U0001f469\U0001f3ff\u200d\u2695\ufe0f|\U0001f468\U0001f3fb\u200d\u2696\ufe0f|\U0001f468\U0001f3fc\u200d\u2696\ufe0f|\U0001f468\U0001f3fd\u200d\u2696\ufe0f|\U0001f468\U0001f3fe\u200d\u2696\ufe0f|\U0001f468\U0001f3ff\u200d\u2696\ufe0f|\U0001f469\U0001f3fb\u200d\u2696\ufe0f|\U0001f469\U0001f3fc\u200d\u2696\ufe0f|\U0001f469\U0001f3fd\u200d\u2696\ufe0f|\U0001f469\U0001f3fe\u200d\u2696\ufe0f|\U0001f469\U0001f3ff\u200d\u2696\ufe0f|\U0001f468\U0001f3fb\u200d\u2708\ufe0f|\U0001f468\U0001f3fc\u200d\u2708\ufe0f|\U0001f468\U0001f3fd\u200d\u2708\ufe0f|\U0001f468\U0001f3fe\u200d\u2708\ufe0f|\U0001f468\U0001f3ff\u200d\u2708\ufe0f|\U0001f469\U0001f3fb\u200d\u2708\ufe0f|\U0001f469\U0001f3fc\u200d\u2708\ufe0f|\U0001f469\U0001f3fd\u200d\u2708\ufe0f|\U0001f469\U0001f3fe\u200d\u2708\ufe0f|\U0001f469\U0001f3ff\u200d\u2708\ufe0f|\U0001f46e\U0001f3fb\u200d\u2642\ufe0f|\U0001f46e\U0001f3fc\u200d\u2642\ufe0f|\U0001f46e\U0001f3fd\u200d\u2642\ufe0f|\U0001f46e\U0001f3fe\u200d\u2642\ufe0f|\U0001f46e\U0001f3ff\u200d\u2642\ufe0f|\U0001f46e\U0001f3fb\u200d\u2640\ufe0f|\U0001f46e\U0001f3fc\u200d\u2640\ufe0f|\U0001f46e\U0001f3fd\u200d\u2640\ufe0f|\U0001f46e\U0001f3fe\u200d\u2640\ufe0f|\U0001f46e\U0001f3ff\u200d\u2640\ufe0f|\U0001f575\ufe0f\u200d\u2642\ufe0f|\U0001f575\U0001f3fb\u200d\u2642\ufe0f|\U0001f575\U0001f3fc\u200d\u2642\ufe0f|\U0001f575\U0001f3fd\u200d\u2642\ufe0f|\U0001f575\U0001f3fe\u200d\u2642\ufe0f|\U0001f575\U0001f3ff\u200d\u2642\ufe0f|\U0001f575\ufe0f\u200d\u2640\ufe0f|\U0001f575\U0001f3fb\u200d\u2640\ufe0f|\U0001f575\U0001f3fc\u200d\u2640\ufe0f|\U0001f575\U0001f3fd\u200d\u2640\ufe0f|\U0001f575\U0001f3fe\u200d\u2640\ufe0f|\U0001f575\U0001f3ff\u200d\u2640\ufe0f|\U0001f482\U0001f3fb\u200d\u2642\ufe0f|\U0001f482\U0001f3fc\u200d\u2642\ufe0f|\U0001f482\U0001f3fd\u200d\u2642\ufe0f|\U0001f482\U0001f3fe\u200d\u2642\ufe0f|\U0001f482\U0001f3ff\u200d\u2642\ufe0f|\U0001f482\U0001f3fb\u200d\u2640\ufe0f|\U0001f482\U0001f3fc\u200d\u2640\ufe0f|\U0001f482\U0001f3fd\u200d\u2640\ufe0f|\U0001f482\U0001f3fe\u200d\u2640\ufe0f|\U0001f482\U0001f3ff\u200d\u2640\ufe0f|\U0001f477\U0001f3fb\u200d\u2642\ufe0f|\U0001f477\U0001f3fc\u200d\u2642\ufe0f|\U0001f477\U0001f3fd\u200d\u2642\ufe0f|\U0001f477\U0001f3fe\u200d\u2642\ufe0f|\U0001f477\U0001f3ff\u200d\u2642\ufe0f|\U0001f477\U0001f3fb\u200d\u2640\ufe0f|\U0001f477\U0001f3fc\u200d\u2640\ufe0f|\U0001f477\U0001f3fd\u200d\u2640\ufe0f|\U0001f477\U0001f3fe\u200d\u2640\ufe0f|\U0001f477\U0001f3ff\u200d\u2640\ufe0f|\U0001f473\U0001f3fb\u200d\u2642\ufe0f|\U0001f473\U0001f3fc\u200d\u2642\ufe0f|\U0001f473\U0001f3fd\u200d\u2642\ufe0f|\U0001f473\U0001f3fe\u200d\u2642\ufe0f|\U0001f473\U0001f3ff\u200d\u2642\ufe0f|\U0001f473\U0001f3fb\u200d\u2640\ufe0f|\U0001f473\U0001f3fc\u200d\u2640\ufe0f|\U0001f473\U0001f3fd\u200d\u2640\ufe0f|\U0001f473\U0001f3fe\u200d\u2640\ufe0f|\U0001f473\U0001f3ff\u200d\u2640\ufe0f|\U0001f9b8\U0001f3fb\u200d\u2642\ufe0f|\U0001f9b8\U0001f3fc\u200d\u2642\ufe0f|\U0001f9b8\U0001f3fd\u200d\u2642\ufe0f|\U0001f9b8\U0001f3fe\u200d\u2642\ufe0f|\U0001f9b8\U0001f3ff\u200d\u2642\ufe0f|\U0001f9b8\U0001f3fb\u200d\u2640\ufe0f|\U0001f9b8\U0001f3fc\u200d\u2640\ufe0f|\U0001f9b8\U0001f3fd\u200d\u2640\ufe0f|\U0001f9b8\U0001f3fe\u200d\u2640\ufe0f|\U0001f9b8\U0001f3ff\u200d\u2640\ufe0f|\U0001f9b9\U0001f3fb\u200d\u2642\ufe0f|\U0001f9b9\U0001f3fc\u200d\u2642\ufe0f|\U0001f9b9\U0001f3fd\u200d\u2642\ufe0f|\U0001f9b9\U0001f3fe\u200d\u2642\ufe0f|\U0001f9b9\U0001f3ff\u200d\u2642\ufe0f|\U0001f9b9\U0001f3fb\u200d\u2640\ufe0f|\U0001f9b9\U0001f3fc\u200d\u2640\ufe0f|\U0001f9b9\U0001f3fd\u200d\u2640\ufe0f|\U0001f9b9\U0001f3fe\u200d\u2640\ufe0f|\U0001f9b9\U0001f3ff\u200d\u2640\ufe0f|\U0001f9d9\U0001f3fb\u200d\u2642\ufe0f|\U0001f9d9\U0001f3fc\u200d\u2642\ufe0f|\U0001f9d9\U0001f3fd\u200d\u2642\ufe0f|\U0001f9d9\U0001f3fe\u200d\u2642\ufe0f|\U0001f9d9\U0001f3ff\u200d\u2642\ufe0f|\U0001f9d9\U0001f3fb\u200d\u2640\ufe0f|\U0001f9d9\U0001f3fc\u200d\u2640\ufe0f|\U0001f9d9\U0001f3fd\u200d\u2640\ufe0f|\U0001f9d9\U0001f3fe\u200d\u2640\ufe0f|\U0001f9d9\U0001f3ff\u200d\u2640\ufe0f|\U0001f9da\U0001f3fb\u200d\u2642\ufe0f|\U0001f9da\U0001f3fc\u200d\u2642\ufe0f|\U0001f9da\U0001f3fd\u200d\u2642\ufe0f|\U0001f9da\U0001f3fe\u200d\u2642\ufe0f|\U0001f9da\U0001f3ff\u200d\u2642\ufe0f|\U0001f9da\U0001f3fb\u200d\u2640\ufe0f|\U0001f9da\U0001f3fc\u200d\u2640\ufe0f|\U0001f9da\U0001f3fd\u200d\u2640\ufe0f|\U0001f9da\U0001f3fe\u200d\u2640\ufe0f|\U0001f9da\U0001f3ff\u200d\u2640\ufe0f|\U0001f9db\U0001f3fb\u200d\u2642\ufe0f|\U0001f9db\U0001f3fc\u200d\u2642\ufe0f|\U0001f9db\U0001f3fd\u200d\u2642\ufe0f|\U0001f9db\U0001f3fe\u200d\u2642\ufe0f|\U0001f9db\U0001f3ff\u200d\u2642\ufe0f|\U0001f9db\U0001f3fb\u200d\u2640\ufe0f|\U0001f9db\U0001f3fc\u200d\u2640\ufe0f|\U0001f9db\U0001f3fd\u200d\u2640\ufe0f|\U0001f9db\U0001f3fe\u200d\u2640\ufe0f|\U0001f9db\U0001f3ff\u200d\u2640\ufe0f|\U0001f9dc\U0001f3fb\u200d\u2642\ufe0f|\U0001f9dc\U0001f3fc\u200d\u2642\ufe0f|\U0001f9dc\U0001f3fd\u200d\u2642\ufe0f|\U0001f9dc\U0001f3fe\u200d\u2642\ufe0f|\U0001f9dc\U0001f3ff\u200d\u2642\ufe0f|\U0001f9dc\U0001f3fb\u200d\u2640\ufe0f|\U0001f9dc\U0001f3fc\u200d\u2640\ufe0f|\U0001f9dc\U0001f3fd\u200d\u2640\ufe0f|\U0001f9dc\U0001f3fe\u200d\u2640\ufe0f|\U0001f9dc\U0001f3ff\u200d\u2640\ufe0f|\U0001f9dd\U0001f3fb\u200d\u2642\ufe0f|\U0001f9dd\U0001f3fc\u200d\u2642\ufe0f|\U0001f9dd\U0001f3fd\u200d\u2642\ufe0f|\U0001f9dd\U0001f3fe\u200d\u2642\ufe0f|\U0001f9dd\U0001f3ff\u200d\u2642\ufe0f|\U0001f9dd\U0001f3fb\u200d\u2640\ufe0f|\U0001f9dd\U0001f3fc\u200d\u2640\ufe0f|\U0001f9dd\U0001f3fd\u200d\u2640\ufe0f|\U0001f9dd\U0001f3fe\u200d\u2640\ufe0f|\U0001f9dd\U0001f3ff\u200d\u2640\ufe0f|\U0001f486\U0001f3fb\u200d\u2642\ufe0f|\U0001f486\U0001f3fc\u200d\u2642\ufe0f|\U0001f486\U0001f3fd\u200d\u2642\ufe0f|\U0001f486\U0001f3fe\u200d\u2642\ufe0f|\U0001f486\U0001f3ff\u200d\u2642\ufe0f|\U0001f486\U0001f3fb\u200d\u2640\ufe0f|\U0001f486\U0001f3fc\u200d\u2640\ufe0f|\U0001f486\U0001f3fd\u200d\u2640\ufe0f|\U0001f486\U0001f3fe\u200d\u2640\ufe0f|\U0001f486\U0001f3ff\u200d\u2640\ufe0f|\U0001f487\U0001f3fb\u200d\u2642\ufe0f|\U0001f487\U0001f3fc\u200d\u2642\ufe0f|\U0001f487\U0001f3fd\u200d\u2642\ufe0f|\U0001f487\U0001f3fe\u200d\u2642\ufe0f|\U0001f487\U0001f3ff\u200d\u2642\ufe0f|\U0001f487\U0001f3fb\u200d\u2640\ufe0f|\U0001f487\U0001f3fc\u200d\u2640\ufe0f|\U0001f487\U0001f3fd\u200d\u2640\ufe0f|\U0001f487\U0001f3fe\u200d\u2640\ufe0f|\U0001f487\U0001f3ff\u200d\u2640\ufe0f|\U0001f6b6\U0001f3fb\u200d\u2642\ufe0f|\U0001f6b6\U0001f3fc\u200d\u2642\ufe0f|\U0001f6b6\U0001f3fd\u200d\u2642\ufe0f|\U0001f6b6\U0001f3fe\u200d\u2642\ufe0f|\U0001f6b6\U0001f3ff\u200d\u2642\ufe0f|\U0001f6b6\U0001f3fb\u200d\u2640\ufe0f|\U0001f6b6\U0001f3fc\u200d\u2640\ufe0f|\U0001f6b6\U0001f3fd\u200d\u2640\ufe0f|\U0001f6b6\U0001f3fe\u200d\u2640\ufe0f|\U0001f6b6\U0001f3ff\u200d\u2640\ufe0f|\U0001f9cd\U0001f3fb\u200d\u2642\ufe0f|\U0001f9cd\U0001f3fc\u200d\u2642\ufe0f|\U0001f9cd\U0001f3fd\u200d\u2642\ufe0f|\U0001f9cd\U0001f3fe\u200d\u2642\ufe0f|\U0001f9cd\U0001f3ff\u200d\u2642\ufe0f|\U0001f9cd\U0001f3fb\u200d\u2640\ufe0f|\U0001f9cd\U0001f3fc\u200d\u2640\ufe0f|\U0001f9cd\U0001f3fd\u200d\u2640\ufe0f|\U0001f9cd\U0001f3fe\u200d\u2640\ufe0f|\U0001f9cd\U0001f3ff\u200d\u2640\ufe0f|\U0001f9ce\U0001f3fb\u200d\u2642\ufe0f|\U0001f9ce\U0001f3fc\u200d\u2642\ufe0f|\U0001f9ce\U0001f3fd\u200d\u2642\ufe0f|\U0001f9ce\U0001f3fe\u200d\u2642\ufe0f|\U0001f9ce\U0001f3ff\u200d\u2642\ufe0f|\U0001f9ce\U0001f3fb\u200d\u2640\ufe0f|\U0001f9ce\U0001f3fc\u200d\u2640\ufe0f|\U0001f9ce\U0001f3fd\u200d\u2640\ufe0f|\U0001f9ce\U0001f3fe\u200d\u2640\ufe0f|\U0001f9ce\U0001f3ff\u200d\u2640\ufe0f|\U0001f3c3\U0001f3fb\u200d\u2642\ufe0f|\U0001f3c3\U0001f3fc\u200d\u2642\ufe0f|\U0001f3c3\U0001f3fd\u200d\u2642\ufe0f|\U0001f3c3\U0001f3fe\u200d\u2642\ufe0f|\U0001f3c3\U0001f3ff\u200d\u2642\ufe0f|\U0001f3c3\U0001f3fb\u200d\u2640\ufe0f|\U0001f3c3\U0001f3fc\u200d\u2640\ufe0f|\U0001f3c3\U0001f3fd\u200d\u2640\ufe0f|\U0001f3c3\U0001f3fe\u200d\u2640\ufe0f|\U0001f3c3\U0001f3ff\u200d\u2640\ufe0f|\U0001f9d6\U0001f3fb\u200d\u2642\ufe0f|\U0001f9d6\U0001f3fc\u200d\u2642\ufe0f|\U0001f9d6\U0001f3fd\u200d\u2642\ufe0f|\U0001f9d6\U0001f3fe\u200d\u2642\ufe0f|\U0001f9d6\U0001f3ff\u200d\u2642\ufe0f|\U0001f9d6\U0001f3fb\u200d\u2640\ufe0f|\U0001f9d6\U0001f3fc\u200d\u2640\ufe0f|\U0001f9d6\U0001f3fd\u200d\u2640\ufe0f|\U0001f9d6\U0001f3fe\u200d\u2640\ufe0f|\U0001f9d6\U0001f3ff\u200d\u2640\ufe0f|\U0001f9d7\U0001f3fb\u200d\u2642\ufe0f|\U0001f9d7\U0001f3fc\u200d\u2642\ufe0f|\U0001f9d7\U0001f3fd\u200d\u2642\ufe0f|\U0001f9d7\U0001f3fe\u200d\u2642\ufe0f|\U0001f9d7\U0001f3ff\u200d\u2642\ufe0f|\U0001f9d7\U0001f3fb\u200d\u2640\ufe0f|\U0001f9d7\U0001f3fc\u200d\u2640\ufe0f|\U0001f9d7\U0001f3fd\u200d\u2640\ufe0f|\U0001f9d7\U0001f3fe\u200d\u2640\ufe0f|\U0001f9d7\U0001f3ff\u200d\u2640\ufe0f|\U0001f3cc\ufe0f\u200d\u2642\ufe0f|\U0001f3cc\U0001f3fb\u200d\u2642\ufe0f|\U0001f3cc\U0001f3fc\u200d\u2642\ufe0f|\U0001f3cc\U0001f3fd\u200d\u2642\ufe0f|\U0001f3cc\U0001f3fe\u200d\u2642\ufe0f|\U0001f3cc\U0001f3ff\u200d\u2642\ufe0f|\U0001f3cc\ufe0f\u200d\u2640\ufe0f|\U0001f3cc\U0001f3fb\u200d\u2640\ufe0f|\U0001f3cc\U0001f3fc\u200d\u2640\ufe0f|\U0001f3cc\U0001f3fd\u200d\u2640\ufe0f|\U0001f3cc\U0001f3fe\u200d\u2640\ufe0f|\U0001f3cc\U0001f3ff\u200d\u2640\ufe0f|\U0001f3c4\U0001f3fb\u200d\u2642\ufe0f|\U0001f3c4\U0001f3fc\u200d\u2642\ufe0f|\U0001f3c4\U0001f3fd\u200d\u2642\ufe0f|\U0001f3c4\U0001f3fe\u200d\u2642\ufe0f|\U0001f3c4\U0001f3ff\u200d\u2642\ufe0f|\U0001f3c4\U0001f3fb\u200d\u2640\ufe0f|\U0001f3c4\U0001f3fc\u200d\u2640\ufe0f|\U0001f3c4\U0001f3fd\u200d\u2640\ufe0f|\U0001f3c4\U0001f3fe\u200d\u2640\ufe0f|\U0001f3c4\U0001f3ff\u200d\u2640\ufe0f|\U0001f6a3\U0001f3fb\u200d\u2642\ufe0f|\U0001f6a3\U0001f3fc\u200d\u2642\ufe0f|\U0001f6a3\U0001f3fd\u200d\u2642\ufe0f|\U0001f6a3\U0001f3fe\u200d\u2642\ufe0f|\U0001f6a3\U0001f3ff\u200d\u2642\ufe0f|\U0001f6a3\U0001f3fb\u200d\u2640\ufe0f|\U0001f6a3\U0001f3fc\u200d\u2640\ufe0f|\U0001f6a3\U0001f3fd\u200d\u2640\ufe0f|\U0001f6a3\U0001f3fe\u200d\u2640\ufe0f|\U0001f6a3\U0001f3ff\u200d\u2640\ufe0f|\U0001f3ca\U0001f3fb\u200d\u2642\ufe0f|\U0001f3ca\U0001f3fc\u200d\u2642\ufe0f|\U0001f3ca\U0001f3fd\u200d\u2642\ufe0f|\U0001f3ca\U0001f3fe\u200d\u2642\ufe0f|\U0001f3ca\U0001f3ff\u200d\u2642\ufe0f|\U0001f3ca\U0001f3fb\u200d\u2640\ufe0f|\U0001f3ca\U0001f3fc\u200d\u2640\ufe0f|\U0001f3ca\U0001f3fd\u200d\u2640\ufe0f|\U0001f3ca\U0001f3fe\u200d\u2640\ufe0f|\U0001f3ca\U0001f3ff\u200d\u2640\ufe0f|\u26f9\ufe0f\u200d\u2642\ufe0f|\u26f9\U0001f3fb\u200d\u2642\ufe0f|\u26f9\U0001f3fc\u200d\u2642\ufe0f|\u26f9\U0001f3fd\u200d\u2642\ufe0f|\u26f9\U0001f3fe\u200d\u2642\ufe0f|\u26f9\U0001f3ff\u200d\u2642\ufe0f|\u26f9\ufe0f\u200d\u2640\ufe0f|\u26f9\U0001f3fb\u200d\u2640\ufe0f|\u26f9\U0001f3fc\u200d\u2640\ufe0f|\u26f9\U0001f3fd\u200d\u2640\ufe0f|\u26f9\U0001f3fe\u200d\u2640\ufe0f|\u26f9\U0001f3ff\u200d\u2640\ufe0f|\U0001f3cb\ufe0f\u200d\u2642\ufe0f|\U0001f3cb\U0001f3fb\u200d\u2642\ufe0f|\U0001f3cb\U0001f3fc\u200d\u2642\ufe0f|\U0001f3cb\U0001f3fd\u200d\u2642\ufe0f|\U0001f3cb\U0001f3fe\u200d\u2642\ufe0f|\U0001f3cb\U0001f3ff\u200d\u2642\ufe0f|\U0001f3cb\ufe0f\u200d\u2640\ufe0f|\U0001f3cb\U0001f3fb\u200d\u2640\ufe0f|\U0001f3cb\U0001f3fc\u200d\u2640\ufe0f|\U0001f3cb\U0001f3fd\u200d\u2640\ufe0f|\U0001f3cb\U0001f3fe\u200d\u2640\ufe0f|\U0001f3cb\U0001f3ff\u200d\u2640\ufe0f|\U0001f6b4\U0001f3fb\u200d\u2642\ufe0f|\U0001f6b4\U0001f3fc\u200d\u2642\ufe0f|\U0001f6b4\U0001f3fd\u200d\u2642\ufe0f|\U0001f6b4\U0001f3fe\u200d\u2642\ufe0f|\U0001f6b4\U0001f3ff\u200d\u2642\ufe0f|\U0001f6b4\U0001f3fb\u200d\u2640\ufe0f|\U0001f6b4\U0001f3fc\u200d\u2640\ufe0f|\U0001f6b4\U0001f3fd\u200d\u2640\ufe0f|\U0001f6b4\U0001f3fe\u200d\u2640\ufe0f|\U0001f6b4\U0001f3ff\u200d\u2640\ufe0f|\U0001f6b5\U0001f3fb\u200d\u2642\ufe0f|\U0001f6b5\U0001f3fc\u200d\u2642\ufe0f|\U0001f6b5\U0001f3fd\u200d\u2642\ufe0f|\U0001f6b5\U0001f3fe\u200d\u2642\ufe0f|\U0001f6b5\U0001f3ff\u200d\u2642\ufe0f|\U0001f6b5\U0001f3fb\u200d\u2640\ufe0f|\U0001f6b5\U0001f3fc\u200d\u2640\ufe0f|\U0001f6b5\U0001f3fd\u200d\u2640\ufe0f|\U0001f6b5\U0001f3fe\u200d\u2640\ufe0f|\U0001f6b5\U0001f3ff\u200d\u2640\ufe0f|\U0001f938\U0001f3fb\u200d\u2642\ufe0f|\U0001f938\U0001f3fc\u200d\u2642\ufe0f|\U0001f938\U0001f3fd\u200d\u2642\ufe0f|\U0001f938\U0001f3fe\u200d\u2642\ufe0f|\U0001f938\U0001f3ff\u200d\u2642\ufe0f|\U0001f938\U0001f3fb\u200d\u2640\ufe0f|\U0001f938\U0001f3fc\u200d\u2640\ufe0f|\U0001f938\U0001f3fd\u200d\u2640\ufe0f|\U0001f938\U0001f3fe\u200d\u2640\ufe0f|\U0001f938\U0001f3ff\u200d\u2640\ufe0f|\U0001f93d\U0001f3fb\u200d\u2642\ufe0f|\U0001f93d\U0001f3fc\u200d\u2642\ufe0f|\U0001f93d\U0001f3fd\u200d\u2642\ufe0f|\U0001f93d\U0001f3fe\u200d\u2642\ufe0f|\U0001f93d\U0001f3ff\u200d\u2642\ufe0f|\U0001f93d\U0001f3fb\u200d\u2640\ufe0f|\U0001f93d\U0001f3fc\u200d\u2640\ufe0f|\U0001f93d\U0001f3fd\u200d\u2640\ufe0f|\U0001f93d\U0001f3fe\u200d\u2640\ufe0f|\U0001f93d\U0001f3ff\u200d\u2640\ufe0f|\U0001f93e\U0001f3fb\u200d\u2642\ufe0f|\U0001f93e\U0001f3fc\u200d\u2642\ufe0f|\U0001f93e\U0001f3fd\u200d\u2642\ufe0f|\U0001f93e\U0001f3fe\u200d\u2642\ufe0f|\U0001f93e\U0001f3ff\u200d\u2642\ufe0f|\U0001f93e\U0001f3fb\u200d\u2640\ufe0f|\U0001f93e\U0001f3fc\u200d\u2640\ufe0f|\U0001f93e\U0001f3fd\u200d\u2640\ufe0f|\U0001f93e\U0001f3fe\u200d\u2640\ufe0f|\U0001f93e\U0001f3ff\u200d\u2640\ufe0f|\U0001f939\U0001f3fb\u200d\u2642\ufe0f|\U0001f939\U0001f3fc\u200d\u2642\ufe0f|\U0001f939\U0001f3fd\u200d\u2642\ufe0f|\U0001f939\U0001f3fe\u200d\u2642\ufe0f|\U0001f939\U0001f3ff\u200d\u2642\ufe0f|\U0001f939\U0001f3fb\u200d\u2640\ufe0f|\U0001f939\U0001f3fc\u200d\u2640\ufe0f|\U0001f939\U0001f3fd\u200d\u2640\ufe0f|\U0001f939\U0001f3fe\u200d\u2640\ufe0f|\U0001f939\U0001f3ff\u200d\u2640\ufe0f|\U0001f9d8\U0001f3fb\u200d\u2642\ufe0f|\U0001f9d8\U0001f3fc\u200d\u2642\ufe0f|\U0001f9d8\U0001f3fd\u200d\u2642\ufe0f|\U0001f9d8\U0001f3fe\u200d\u2642\ufe0f|\U0001f9d8\U0001f3ff\u200d\u2642\ufe0f|\U0001f9d8\U0001f3fb\u200d\u2640\ufe0f|\U0001f9d8\U0001f3fc\u200d\u2640\ufe0f|\U0001f9d8\U0001f3fd\u200d\u2640\ufe0f|\U0001f9d8\U0001f3fe\u200d\u2640\ufe0f|\U0001f9d8\U0001f3ff\u200d\u2640\ufe0f|\U0001f9d1\u200d\U0001f91d\u200d\U0001f9d1|\U0001f469\u200d\u2764\u200d\U0001f468|\U0001f468\u200d\u2764\u200d\U0001f468|\U0001f469\u200d\u2764\u200d\U0001f469|\U0001f468\u200d\U0001f469\u200d\U0001f466|\U0001f468\u200d\U0001f469\u200d\U0001f467|\U0001f468\u200d\U0001f468\u200d\U0001f466|\U0001f468\u200d\U0001f468\u200d\U0001f467|\U0001f469\u200d\U0001f469\u200d\U0001f466|\U0001f469\u200d\U0001f469\u200d\U0001f467|\U0001f468\u200d\U0001f466\u200d\U0001f466|\U0001f468\u200d\U0001f467\u200d\U0001f466|\U0001f468\u200d\U0001f467\u200d\U0001f467|\U0001f469\u200d\U0001f466\u200d\U0001f466|\U0001f469\u200d\U0001f467\u200d\U0001f466|\U0001f469\u200d\U0001f467\u200d\U0001f467|\U0001f441\u200d\U0001f5e8\ufe0f|\U0001f441\ufe0f\u200d\U0001f5e8|\U0001f471\u200d\u2642\ufe0f|\U0001f471\U0001f3fb\u200d\u2642|\U0001f471\U0001f3fc\u200d\u2642|\U0001f471\U0001f3fd\u200d\u2642|\U0001f471\U0001f3fe\u200d\u2642|\U0001f471\U0001f3ff\u200d\u2642|\U0001f468\U0001f3fb\u200d\U0001f9b0|\U0001f468\U0001f3fc\u200d\U0001f9b0|\U0001f468\U0001f3fd\u200d\U0001f9b0|\U0001f468\U0001f3fe\u200d\U0001f9b0|\U0001f468\U0001f3ff\u200d\U0001f9b0|\U0001f468\U0001f3fb\u200d\U0001f9b1|\U0001f468\U0001f3fc\u200d\U0001f9b1|\U0001f468\U0001f3fd\u200d\U0001f9b1|\U0001f468\U0001f3fe\u200d\U0001f9b1|\U0001f468\U0001f3ff\u200d\U0001f9b1|\U0001f468\U0001f3fb\u200d\U0001f9b3|\U0001f468\U0001f3fc\u200d\U0001f9b3|\U0001f468\U0001f3fd\u200d\U0001f9b3|\U0001f468\U0001f3fe\u200d\U0001f9b3|\U0001f468\U0001f3ff\u200d\U0001f9b3|\U0001f468\U0001f3fb\u200d\U0001f9b2|\U0001f468\U0001f3fc\u200d\U0001f9b2|\U0001f468\U0001f3fd\u200d\U0001f9b2|\U0001f468\U0001f3fe\u200d\U0001f9b2|\U0001f468\U0001f3ff\u200d\U0001f9b2|\U0001f471\u200d\u2640\ufe0f|\U0001f471\U0001f3fb\u200d\u2640|\U0001f471\U0001f3fc\u200d\u2640|\U0001f471\U0001f3fd\u200d\u2640|\U0001f471\U0001f3fe\u200d\u2640|\U0001f471\U0001f3ff\u200d\u2640|\U0001f469\U0001f3fb\u200d\U0001f9b0|\U0001f469\U0001f3fc\u200d\U0001f9b0|\U0001f469\U0001f3fd\u200d\U0001f9b0|\U0001f469\U0001f3fe\u200d\U0001f9b0|\U0001f469\U0001f3ff\u200d\U0001f9b0|\U0001f469\U0001f3fb\u200d\U0001f9b1|\U0001f469\U0001f3fc\u200d\U0001f9b1|\U0001f469\U0001f3fd\u200d\U0001f9b1|\U0001f469\U0001f3fe\u200d\U0001f9b1|\U0001f469\U0001f3ff\u200d\U0001f9b1|\U0001f469\U0001f3fb\u200d\U0001f9b3|\U0001f469\U0001f3fc\u200d\U0001f9b3|\U0001f469\U0001f3fd\u200d\U0001f9b3|\U0001f469\U0001f3fe\u200d\U0001f9b3|\U0001f469\U0001f3ff\u200d\U0001f9b3|\U0001f469\U0001f3fb\u200d\U0001f9b2|\U0001f469\U0001f3fc\u200d\U0001f9b2|\U0001f469\U0001f3fd\u200d\U0001f9b2|\U0001f469\U0001f3fe\u200d\U0001f9b2|\U0001f469\U0001f3ff\u200d\U0001f9b2|\U0001f64d\u200d\u2642\ufe0f|\U0001f64d\U0001f3fb\u200d\u2642|\U0001f64d\U0001f3fc\u200d\u2642|\U0001f64d\U0001f3fd\u200d\u2642|\U0001f64d\U0001f3fe\u200d\u2642|\U0001f64d\U0001f3ff\u200d\u2642|\U0001f64d\u200d\u2640\ufe0f|\U0001f64d\U0001f3fb\u200d\u2640|\U0001f64d\U0001f3fc\u200d\u2640|\U0001f64d\U0001f3fd\u200d\u2640|\U0001f64d\U0001f3fe\u200d\u2640|\U0001f64d\U0001f3ff\u200d\u2640|\U0001f64e\u200d\u2642\ufe0f|\U0001f64e\U0001f3fb\u200d\u2642|\U0001f64e\U0001f3fc\u200d\u2642|\U0001f64e\U0001f3fd\u200d\u2642|\U0001f64e\U0001f3fe\u200d\u2642|\U0001f64e\U0001f3ff\u200d\u2642|\U0001f64e\u200d\u2640\ufe0f|\U0001f64e\U0001f3fb\u200d\u2640|\U0001f64e\U0001f3fc\u200d\u2640|\U0001f64e\U0001f3fd\u200d\u2640|\U0001f64e\U0001f3fe\u200d\u2640|\U0001f64e\U0001f3ff\u200d\u2640|\U0001f645\u200d\u2642\ufe0f|\U0001f645\U0001f3fb\u200d\u2642|\U0001f645\U0001f3fc\u200d\u2642|\U0001f645\U0001f3fd\u200d\u2642|\U0001f645\U0001f3fe\u200d\u2642|\U0001f645\U0001f3ff\u200d\u2642|\U0001f645\u200d\u2640\ufe0f|\U0001f645\U0001f3fb\u200d\u2640|\U0001f645\U0001f3fc\u200d\u2640|\U0001f645\U0001f3fd\u200d\u2640|\U0001f645\U0001f3fe\u200d\u2640|\U0001f645\U0001f3ff\u200d\u2640|\U0001f646\u200d\u2642\ufe0f|\U0001f646\U0001f3fb\u200d\u2642|\U0001f646\U0001f3fc\u200d\u2642|\U0001f646\U0001f3fd\u200d\u2642|\U0001f646\U0001f3fe\u200d\u2642|\U0001f646\U0001f3ff\u200d\u2642|\U0001f646\u200d\u2640\ufe0f|\U0001f646\U0001f3fb\u200d\u2640|\U0001f646\U0001f3fc\u200d\u2640|\U0001f646\U0001f3fd\u200d\u2640|\U0001f646\U0001f3fe\u200d\u2640|\U0001f646\U0001f3ff\u200d\u2640|\U0001f481\u200d\u2642\ufe0f|\U0001f481\U0001f3fb\u200d\u2642|\U0001f481\U0001f3fc\u200d\u2642|\U0001f481\U0001f3fd\u200d\u2642|\U0001f481\U0001f3fe\u200d\u2642|\U0001f481\U0001f3ff\u200d\u2642|\U0001f481\u200d\u2640\ufe0f|\U0001f481\U0001f3fb\u200d\u2640|\U0001f481\U0001f3fc\u200d\u2640|\U0001f481\U0001f3fd\u200d\u2640|\U0001f481\U0001f3fe\u200d\u2640|\U0001f481\U0001f3ff\u200d\u2640|\U0001f64b\u200d\u2642\ufe0f|\U0001f64b\U0001f3fb\u200d\u2642|\U0001f64b\U0001f3fc\u200d\u2642|\U0001f64b\U0001f3fd\u200d\u2642|\U0001f64b\U0001f3fe\u200d\u2642|\U0001f64b\U0001f3ff\u200d\u2642|\U0001f64b\u200d\u2640\ufe0f|\U0001f64b\U0001f3fb\u200d\u2640|\U0001f64b\U0001f3fc\u200d\u2640|\U0001f64b\U0001f3fd\u200d\u2640|\U0001f64b\U0001f3fe\u200d\u2640|\U0001f64b\U0001f3ff\u200d\u2640|\U0001f9cf\u200d\u2642\ufe0f|\U0001f9cf\U0001f3fb\u200d\u2642|\U0001f9cf\U0001f3fc\u200d\u2642|\U0001f9cf\U0001f3fd\u200d\u2642|\U0001f9cf\U0001f3fe\u200d\u2642|\U0001f9cf\U0001f3ff\u200d\u2642|\U0001f9cf\u200d\u2640\ufe0f|\U0001f9cf\U0001f3fb\u200d\u2640|\U0001f9cf\U0001f3fc\u200d\u2640|\U0001f9cf\U0001f3fd\u200d\u2640|\U0001f9cf\U0001f3fe\u200d\u2640|\U0001f9cf\U0001f3ff\u200d\u2640|\U0001f647\u200d\u2642\ufe0f|\U0001f647\U0001f3fb\u200d\u2642|\U0001f647\U0001f3fc\u200d\u2642|\U0001f647\U0001f3fd\u200d\u2642|\U0001f647\U0001f3fe\u200d\u2642|\U0001f647\U0001f3ff\u200d\u2642|\U0001f647\u200d\u2640\ufe0f|\U0001f647\U0001f3fb\u200d\u2640|\U0001f647\U0001f3fc\u200d\u2640|\U0001f647\U0001f3fd\u200d\u2640|\U0001f647\U0001f3fe\u200d\u2640|\U0001f647\U0001f3ff\u200d\u2640|\U0001f926\u200d\u2642\ufe0f|\U0001f926\U0001f3fb\u200d\u2642|\U0001f926\U0001f3fc\u200d\u2642|\U0001f926\U0001f3fd\u200d\u2642|\U0001f926\U0001f3fe\u200d\u2642|\U0001f926\U0001f3ff\u200d\u2642|\U0001f926\u200d\u2640\ufe0f|\U0001f926\U0001f3fb\u200d\u2640|\U0001f926\U0001f3fc\u200d\u2640|\U0001f926\U0001f3fd\u200d\u2640|\U0001f926\U0001f3fe\u200d\u2640|\U0001f926\U0001f3ff\u200d\u2640|\U0001f937\u200d\u2642\ufe0f|\U0001f937\U0001f3fb\u200d\u2642|\U0001f937\U0001f3fc\u200d\u2642|\U0001f937\U0001f3fd\u200d\u2642|\U0001f937\U0001f3fe\u200d\u2642|\U0001f937\U0001f3ff\u200d\u2642|\U0001f937\u200d\u2640\ufe0f|\U0001f937\U0001f3fb\u200d\u2640|\U0001f937\U0001f3fc\u200d\u2640|\U0001f937\U0001f3fd\u200d\u2640|\U0001f937\U0001f3fe\u200d\u2640|\U0001f937\U0001f3ff\u200d\u2640|\U0001f468\u200d\u2695\ufe0f|\U0001f468\U0001f3fb\u200d\u2695|\U0001f468\U0001f3fc\u200d\u2695|\U0001f468\U0001f3fd\u200d\u2695|\U0001f468\U0001f3fe\u200d\u2695|\U0001f468\U0001f3ff\u200d\u2695|\U0001f469\u200d\u2695\ufe0f|\U0001f469\U0001f3fb\u200d\u2695|\U0001f469\U0001f3fc\u200d\u2695|\U0001f469\U0001f3fd\u200d\u2695|\U0001f469\U0001f3fe\u200d\u2695|\U0001f469\U0001f3ff\u200d\u2695|\U0001f468\U0001f3fb\u200d\U0001f393|\U0001f468\U0001f3fc\u200d\U0001f393|\U0001f468\U0001f3fd\u200d\U0001f393|\U0001f468\U0001f3fe\u200d\U0001f393|\U0001f468\U0001f3ff\u200d\U0001f393|\U0001f469\U0001f3fb\u200d\U0001f393|\U0001f469\U0001f3fc\u200d\U0001f393|\U0001f469\U0001f3fd\u200d\U0001f393|\U0001f469\U0001f3fe\u200d\U0001f393|\U0001f469\U0001f3ff\u200d\U0001f393|\U0001f468\U0001f3fb\u200d\U0001f3eb|\U0001f468\U0001f3fc\u200d\U0001f3eb|\U0001f468\U0001f3fd\u200d\U0001f3eb|\U0001f468\U0001f3fe\u200d\U0001f3eb|\U0001f468\U0001f3ff\u200d\U0001f3eb|\U0001f469\U0001f3fb\u200d\U0001f3eb|\U0001f469\U0001f3fc\u200d\U0001f3eb|\U0001f469\U0001f3fd\u200d\U0001f3eb|\U0001f469\U0001f3fe\u200d\U0001f3eb|\U0001f469\U0001f3ff\u200d\U0001f3eb|\U0001f468\u200d\u2696\ufe0f|\U0001f468\U0001f3fb\u200d\u2696|\U0001f468\U0001f3fc\u200d\u2696|\U0001f468\U0001f3fd\u200d\u2696|\U0001f468\U0001f3fe\u200d\u2696|\U0001f468\U0001f3ff\u200d\u2696|\U0001f469\u200d\u2696\ufe0f|\U0001f469\U0001f3fb\u200d\u2696|\U0001f469\U0001f3fc\u200d\u2696|\U0001f469\U0001f3fd\u200d\u2696|\U0001f469\U0001f3fe\u200d\u2696|\U0001f469\U0001f3ff\u200d\u2696|\U0001f468\U0001f3fb\u200d\U0001f33e|\U0001f468\U0001f3fc\u200d\U0001f33e|\U0001f468\U0001f3fd\u200d\U0001f33e|\U0001f468\U0001f3fe\u200d\U0001f33e|\U0001f468\U0001f3ff\u200d\U0001f33e|\U0001f469\U0001f3fb\u200d\U0001f33e|\U0001f469\U0001f3fc\u200d\U0001f33e|\U0001f469\U0001f3fd\u200d\U0001f33e|\U0001f469\U0001f3fe\u200d\U0001f33e|\U0001f469\U0001f3ff\u200d\U0001f33e|\U0001f468\U0001f3fb\u200d\U0001f373|\U0001f468\U0001f3fc\u200d\U0001f373|\U0001f468\U0001f3fd\u200d\U0001f373|\U0001f468\U0001f3fe\u200d\U0001f373|\U0001f468\U0001f3ff\u200d\U0001f373|\U0001f469\U0001f3fb\u200d\U0001f373|\U0001f469\U0001f3fc\u200d\U0001f373|\U0001f469\U0001f3fd\u200d\U0001f373|\U0001f469\U0001f3fe\u200d\U0001f373|\U0001f469\U0001f3ff\u200d\U0001f373|\U0001f468\U0001f3fb\u200d\U0001f527|\U0001f468\U0001f3fc\u200d\U0001f527|\U0001f468\U0001f3fd\u200d\U0001f527|\U0001f468\U0001f3fe\u200d\U0001f527|\U0001f468\U0001f3ff\u200d\U0001f527|\U0001f469\U0001f3fb\u200d\U0001f527|\U0001f469\U0001f3fc\u200d\U0001f527|\U0001f469\U0001f3fd\u200d\U0001f527|\U0001f469\U0001f3fe\u200d\U0001f527|\U0001f469\U0001f3ff\u200d\U0001f527|\U0001f468\U0001f3fb\u200d\U0001f3ed|\U0001f468\U0001f3fc\u200d\U0001f3ed|\U0001f468\U0001f3fd\u200d\U0001f3ed|\U0001f468\U0001f3fe\u200d\U0001f3ed|\U0001f468\U0001f3ff\u200d\U0001f3ed|\U0001f469\U0001f3fb\u200d\U0001f3ed|\U0001f469\U0001f3fc\u200d\U0001f3ed|\U0001f469\U0001f3fd\u200d\U0001f3ed|\U0001f469\U0001f3fe\u200d\U0001f3ed|\U0001f469\U0001f3ff\u200d\U0001f3ed|\U0001f468\U0001f3fb\u200d\U0001f4bc|\U0001f468\U0001f3fc\u200d\U0001f4bc|\U0001f468\U0001f3fd\u200d\U0001f4bc|\U0001f468\U0001f3fe\u200d\U0001f4bc|\U0001f468\U0001f3ff\u200d\U0001f4bc|\U0001f469\U0001f3fb\u200d\U0001f4bc|\U0001f469\U0001f3fc\u200d\U0001f4bc|\U0001f469\U0001f3fd\u200d\U0001f4bc|\U0001f469\U0001f3fe\u200d\U0001f4bc|\U0001f469\U0001f3ff\u200d\U0001f4bc|\U0001f468\U0001f3fb\u200d\U0001f52c|\U0001f468\U0001f3fc\u200d\U0001f52c|\U0001f468\U0001f3fd\u200d\U0001f52c|\U0001f468\U0001f3fe\u200d\U0001f52c|\U0001f468\U0001f3ff\u200d\U0001f52c|\U0001f469\U0001f3fb\u200d\U0001f52c|\U0001f469\U0001f3fc\u200d\U0001f52c|\U0001f469\U0001f3fd\u200d\U0001f52c|\U0001f469\U0001f3fe\u200d\U0001f52c|\U0001f469\U0001f3ff\u200d\U0001f52c|\U0001f468\U0001f3fb\u200d\U0001f4bb|\U0001f468\U0001f3fc\u200d\U0001f4bb|\U0001f468\U0001f3fd\u200d\U0001f4bb|\U0001f468\U0001f3fe\u200d\U0001f4bb|\U0001f468\U0001f3ff\u200d\U0001f4bb|\U0001f469\U0001f3fb\u200d\U0001f4bb|\U0001f469\U0001f3fc\u200d\U0001f4bb|\U0001f469\U0001f3fd\u200d\U0001f4bb|\U0001f469\U0001f3fe\u200d\U0001f4bb|\U0001f469\U0001f3ff\u200d\U0001f4bb|\U0001f468\U0001f3fb\u200d\U0001f3a4|\U0001f468\U0001f3fc\u200d\U0001f3a4|\U0001f468\U0001f3fd\u200d\U0001f3a4|\U0001f468\U0001f3fe\u200d\U0001f3a4|\U0001f468\U0001f3ff\u200d\U0001f3a4|\U0001f469\U0001f3fb\u200d\U0001f3a4|\U0001f469\U0001f3fc\u200d\U0001f3a4|\U0001f469\U0001f3fd\u200d\U0001f3a4|\U0001f469\U0001f3fe\u200d\U0001f3a4|\U0001f469\U0001f3ff\u200d\U0001f3a4|\U0001f468\U0001f3fb\u200d\U0001f3a8|\U0001f468\U0001f3fc\u200d\U0001f3a8|\U0001f468\U0001f3fd\u200d\U0001f3a8|\U0001f468\U0001f3fe\u200d\U0001f3a8|\U0001f468\U0001f3ff\u200d\U0001f3a8|\U0001f469\U0001f3fb\u200d\U0001f3a8|\U0001f469\U0001f3fc\u200d\U0001f3a8|\U0001f469\U0001f3fd\u200d\U0001f3a8|\U0001f469\U0001f3fe\u200d\U0001f3a8|\U0001f469\U0001f3ff\u200d\U0001f3a8|\U0001f468\u200d\u2708\ufe0f|\U0001f468\U0001f3fb\u200d\u2708|\U0001f468\U0001f3fc\u200d\u2708|\U0001f468\U0001f3fd\u200d\u2708|\U0001f468\U0001f3fe\u200d\u2708|\U0001f468\U0001f3ff\u200d\u2708|\U0001f469\u200d\u2708\ufe0f|\U0001f469\U0001f3fb\u200d\u2708|\U0001f469\U0001f3fc\u200d\u2708|\U0001f469\U0001f3fd\u200d\u2708|\U0001f469\U0001f3fe\u200d\u2708|\U0001f469\U0001f3ff\u200d\u2708|\U0001f468\U0001f3fb\u200d\U0001f680|\U0001f468\U0001f3fc\u200d\U0001f680|\U0001f468\U0001f3fd\u200d\U0001f680|\U0001f468\U0001f3fe\u200d\U0001f680|\U0001f468\U0001f3ff\u200d\U0001f680|\U0001f469\U0001f3fb\u200d\U0001f680|\U0001f469\U0001f3fc\u200d\U0001f680|\U0001f469\U0001f3fd\u200d\U0001f680|\U0001f469\U0001f3fe\u200d\U0001f680|\U0001f469\U0001f3ff\u200d\U0001f680|\U0001f468\U0001f3fb\u200d\U0001f692|\U0001f468\U0001f3fc\u200d\U0001f692|\U0001f468\U0001f3fd\u200d\U0001f692|\U0001f468\U0001f3fe\u200d\U0001f692|\U0001f468\U0001f3ff\u200d\U0001f692|\U0001f469\U0001f3fb\u200d\U0001f692|\U0001f469\U0001f3fc\u200d\U0001f692|\U0001f469\U0001f3fd\u200d\U0001f692|\U0001f469\U0001f3fe\u200d\U0001f692|\U0001f469\U0001f3ff\u200d\U0001f692|\U0001f46e\u200d\u2642\ufe0f|\U0001f46e\U0001f3fb\u200d\u2642|\U0001f46e\U0001f3fc\u200d\u2642|\U0001f46e\U0001f3fd\u200d\u2642|\U0001f46e\U0001f3fe\u200d\u2642|\U0001f46e\U0001f3ff\u200d\u2642|\U0001f46e\u200d\u2640\ufe0f|\U0001f46e\U0001f3fb\u200d\u2640|\U0001f46e\U0001f3fc\u200d\u2640|\U0001f46e\U0001f3fd\u200d\u2640|\U0001f46e\U0001f3fe\u200d\u2640|\U0001f46e\U0001f3ff\u200d\u2640|\U0001f575\u200d\u2642\ufe0f|\U0001f575\ufe0f\u200d\u2642|\U0001f575\U0001f3fb\u200d\u2642|\U0001f575\U0001f3fc\u200d\u2642|\U0001f575\U0001f3fd\u200d\u2642|\U0001f575\U0001f3fe\u200d\u2642|\U0001f575\U0001f3ff\u200d\u2642|\U0001f575\u200d\u2640\ufe0f|\U0001f575\ufe0f\u200d\u2640|\U0001f575\U0001f3fb\u200d\u2640|\U0001f575\U0001f3fc\u200d\u2640|\U0001f575\U0001f3fd\u200d\u2640|\U0001f575\U0001f3fe\u200d\u2640|\U0001f575\U0001f3ff\u200d\u2640|\U0001f482\u200d\u2642\ufe0f|\U0001f482\U0001f3fb\u200d\u2642|\U0001f482\U0001f3fc\u200d\u2642|\U0001f482\U0001f3fd\u200d\u2642|\U0001f482\U0001f3fe\u200d\u2642|\U0001f482\U0001f3ff\u200d\u2642|\U0001f482\u200d\u2640\ufe0f|\U0001f482\U0001f3fb\u200d\u2640|\U0001f482\U0001f3fc\u200d\u2640|\U0001f482\U0001f3fd\u200d\u2640|\U0001f482\U0001f3fe\u200d\u2640|\U0001f482\U0001f3ff\u200d\u2640|\U0001f477\u200d\u2642\ufe0f|\U0001f477\U0001f3fb\u200d\u2642|\U0001f477\U0001f3fc\u200d\u2642|\U0001f477\U0001f3fd\u200d\u2642|\U0001f477\U0001f3fe\u200d\u2642|\U0001f477\U0001f3ff\u200d\u2642|\U0001f477\u200d\u2640\ufe0f|\U0001f477\U0001f3fb\u200d\u2640|\U0001f477\U0001f3fc\u200d\u2640|\U0001f477\U0001f3fd\u200d\u2640|\U0001f477\U0001f3fe\u200d\u2640|\U0001f477\U0001f3ff\u200d\u2640|\U0001f473\u200d\u2642\ufe0f|\U0001f473\U0001f3fb\u200d\u2642|\U0001f473\U0001f3fc\u200d\u2642|\U0001f473\U0001f3fd\u200d\u2642|\U0001f473\U0001f3fe\u200d\u2642|\U0001f473\U0001f3ff\u200d\u2642|\U0001f473\u200d\u2640\ufe0f|\U0001f473\U0001f3fb\u200d\u2640|\U0001f473\U0001f3fc\u200d\u2640|\U0001f473\U0001f3fd\u200d\u2640|\U0001f473\U0001f3fe\u200d\u2640|\U0001f473\U0001f3ff\u200d\u2640|\U0001f9b8\u200d\u2642\ufe0f|\U0001f9b8\U0001f3fb\u200d\u2642|\U0001f9b8\U0001f3fc\u200d\u2642|\U0001f9b8\U0001f3fd\u200d\u2642|\U0001f9b8\U0001f3fe\u200d\u2642|\U0001f9b8\U0001f3ff\u200d\u2642|\U0001f9b8\u200d\u2640\ufe0f|\U0001f9b8\U0001f3fb\u200d\u2640|\U0001f9b8\U0001f3fc\u200d\u2640|\U0001f9b8\U0001f3fd\u200d\u2640|\U0001f9b8\U0001f3fe\u200d\u2640|\U0001f9b8\U0001f3ff\u200d\u2640|\U0001f9b9\u200d\u2642\ufe0f|\U0001f9b9\U0001f3fb\u200d\u2642|\U0001f9b9\U0001f3fc\u200d\u2642|\U0001f9b9\U0001f3fd\u200d\u2642|\U0001f9b9\U0001f3fe\u200d\u2642|\U0001f9b9\U0001f3ff\u200d\u2642|\U0001f9b9\u200d\u2640\ufe0f|\U0001f9b9\U0001f3fb\u200d\u2640|\U0001f9b9\U0001f3fc\u200d\u2640|\U0001f9b9\U0001f3fd\u200d\u2640|\U0001f9b9\U0001f3fe\u200d\u2640|\U0001f9b9\U0001f3ff\u200d\u2640|\U0001f9d9\u200d\u2642\ufe0f|\U0001f9d9\U0001f3fb\u200d\u2642|\U0001f9d9\U0001f3fc\u200d\u2642|\U0001f9d9\U0001f3fd\u200d\u2642|\U0001f9d9\U0001f3fe\u200d\u2642|\U0001f9d9\U0001f3ff\u200d\u2642|\U0001f9d9\u200d\u2640\ufe0f|\U0001f9d9\U0001f3fb\u200d\u2640|\U0001f9d9\U0001f3fc\u200d\u2640|\U0001f9d9\U0001f3fd\u200d\u2640|\U0001f9d9\U0001f3fe\u200d\u2640|\U0001f9d9\U0001f3ff\u200d\u2640|\U0001f9da\u200d\u2642\ufe0f|\U0001f9da\U0001f3fb\u200d\u2642|\U0001f9da\U0001f3fc\u200d\u2642|\U0001f9da\U0001f3fd\u200d\u2642|\U0001f9da\U0001f3fe\u200d\u2642|\U0001f9da\U0001f3ff\u200d\u2642|\U0001f9da\u200d\u2640\ufe0f|\U0001f9da\U0001f3fb\u200d\u2640|\U0001f9da\U0001f3fc\u200d\u2640|\U0001f9da\U0001f3fd\u200d\u2640|\U0001f9da\U0001f3fe\u200d\u2640|\U0001f9da\U0001f3ff\u200d\u2640|\U0001f9db\u200d\u2642\ufe0f|\U0001f9db\U0001f3fb\u200d\u2642|\U0001f9db\U0001f3fc\u200d\u2642|\U0001f9db\U0001f3fd\u200d\u2642|\U0001f9db\U0001f3fe\u200d\u2642|\U0001f9db\U0001f3ff\u200d\u2642|\U0001f9db\u200d\u2640\ufe0f|\U0001f9db\U0001f3fb\u200d\u2640|\U0001f9db\U0001f3fc\u200d\u2640|\U0001f9db\U0001f3fd\u200d\u2640|\U0001f9db\U0001f3fe\u200d\u2640|\U0001f9db\U0001f3ff\u200d\u2640|\U0001f9dc\u200d\u2642\ufe0f|\U0001f9dc\U0001f3fb\u200d\u2642|\U0001f9dc\U0001f3fc\u200d\u2642|\U0001f9dc\U0001f3fd\u200d\u2642|\U0001f9dc\U0001f3fe\u200d\u2642|\U0001f9dc\U0001f3ff\u200d\u2642|\U0001f9dc\u200d\u2640\ufe0f|\U0001f9dc\U0001f3fb\u200d\u2640|\U0001f9dc\U0001f3fc\u200d\u2640|\U0001f9dc\U0001f3fd\u200d\u2640|\U0001f9dc\U0001f3fe\u200d\u2640|\U0001f9dc\U0001f3ff\u200d\u2640|\U0001f9dd\u200d\u2642\ufe0f|\U0001f9dd\U0001f3fb\u200d\u2642|\U0001f9dd\U0001f3fc\u200d\u2642|\U0001f9dd\U0001f3fd\u200d\u2642|\U0001f9dd\U0001f3fe\u200d\u2642|\U0001f9dd\U0001f3ff\u200d\u2642|\U0001f9dd\u200d\u2640\ufe0f|\U0001f9dd\U0001f3fb\u200d\u2640|\U0001f9dd\U0001f3fc\u200d\u2640|\U0001f9dd\U0001f3fd\u200d\u2640|\U0001f9dd\U0001f3fe\u200d\u2640|\U0001f9dd\U0001f3ff\u200d\u2640|\U0001f9de\u200d\u2642\ufe0f|\U0001f9de\u200d\u2640\ufe0f|\U0001f9df\u200d\u2642\ufe0f|\U0001f9df\u200d\u2640\ufe0f|\U0001f486\u200d\u2642\ufe0f|\U0001f486\U0001f3fb\u200d\u2642|\U0001f486\U0001f3fc\u200d\u2642|\U0001f486\U0001f3fd\u200d\u2642|\U0001f486\U0001f3fe\u200d\u2642|\U0001f486\U0001f3ff\u200d\u2642|\U0001f486\u200d\u2640\ufe0f|\U0001f486\U0001f3fb\u200d\u2640|\U0001f486\U0001f3fc\u200d\u2640|\U0001f486\U0001f3fd\u200d\u2640|\U0001f486\U0001f3fe\u200d\u2640|\U0001f486\U0001f3ff\u200d\u2640|\U0001f487\u200d\u2642\ufe0f|\U0001f487\U0001f3fb\u200d\u2642|\U0001f487\U0001f3fc\u200d\u2642|\U0001f487\U0001f3fd\u200d\u2642|\U0001f487\U0001f3fe\u200d\u2642|\U0001f487\U0001f3ff\u200d\u2642|\U0001f487\u200d\u2640\ufe0f|\U0001f487\U0001f3fb\u200d\u2640|\U0001f487\U0001f3fc\u200d\u2640|\U0001f487\U0001f3fd\u200d\u2640|\U0001f487\U0001f3fe\u200d\u2640|\U0001f487\U0001f3ff\u200d\u2640|\U0001f6b6\u200d\u2642\ufe0f|\U0001f6b6\U0001f3fb\u200d\u2642|\U0001f6b6\U0001f3fc\u200d\u2642|\U0001f6b6\U0001f3fd\u200d\u2642|\U0001f6b6\U0001f3fe\u200d\u2642|\U0001f6b6\U0001f3ff\u200d\u2642|\U0001f6b6\u200d\u2640\ufe0f|\U0001f6b6\U0001f3fb\u200d\u2640|\U0001f6b6\U0001f3fc\u200d\u2640|\U0001f6b6\U0001f3fd\u200d\u2640|\U0001f6b6\U0001f3fe\u200d\u2640|\U0001f6b6\U0001f3ff\u200d\u2640|\U0001f9cd\u200d\u2642\ufe0f|\U0001f9cd\U0001f3fb\u200d\u2642|\U0001f9cd\U0001f3fc\u200d\u2642|\U0001f9cd\U0001f3fd\u200d\u2642|\U0001f9cd\U0001f3fe\u200d\u2642|\U0001f9cd\U0001f3ff\u200d\u2642|\U0001f9cd\u200d\u2640\ufe0f|\U0001f9cd\U0001f3fb\u200d\u2640|\U0001f9cd\U0001f3fc\u200d\u2640|\U0001f9cd\U0001f3fd\u200d\u2640|\U0001f9cd\U0001f3fe\u200d\u2640|\U0001f9cd\U0001f3ff\u200d\u2640|\U0001f9ce\u200d\u2642\ufe0f|\U0001f9ce\U0001f3fb\u200d\u2642|\U0001f9ce\U0001f3fc\u200d\u2642|\U0001f9ce\U0001f3fd\u200d\u2642|\U0001f9ce\U0001f3fe\u200d\u2642|\U0001f9ce\U0001f3ff\u200d\u2642|\U0001f9ce\u200d\u2640\ufe0f|\U0001f9ce\U0001f3fb\u200d\u2640|\U0001f9ce\U0001f3fc\u200d\u2640|\U0001f9ce\U0001f3fd\u200d\u2640|\U0001f9ce\U0001f3fe\u200d\u2640|\U0001f9ce\U0001f3ff\u200d\u2640|\U0001f468\U0001f3fb\u200d\U0001f9af|\U0001f468\U0001f3fc\u200d\U0001f9af|\U0001f468\U0001f3fd\u200d\U0001f9af|\U0001f468\U0001f3fe\u200d\U0001f9af|\U0001f468\U0001f3ff\u200d\U0001f9af|\U0001f469\U0001f3fb\u200d\U0001f9af|\U0001f469\U0001f3fc\u200d\U0001f9af|\U0001f469\U0001f3fd\u200d\U0001f9af|\U0001f469\U0001f3fe\u200d\U0001f9af|\U0001f469\U0001f3ff\u200d\U0001f9af|\U0001f468\U0001f3fb\u200d\U0001f9bc|\U0001f468\U0001f3fc\u200d\U0001f9bc|\U0001f468\U0001f3fd\u200d\U0001f9bc|\U0001f468\U0001f3fe\u200d\U0001f9bc|\U0001f468\U0001f3ff\u200d\U0001f9bc|\U0001f469\U0001f3fb\u200d\U0001f9bc|\U0001f469\U0001f3fc\u200d\U0001f9bc|\U0001f469\U0001f3fd\u200d\U0001f9bc|\U0001f469\U0001f3fe\u200d\U0001f9bc|\U0001f469\U0001f3ff\u200d\U0001f9bc|\U0001f468\U0001f3fb\u200d\U0001f9bd|\U0001f468\U0001f3fc\u200d\U0001f9bd|\U0001f468\U0001f3fd\u200d\U0001f9bd|\U0001f468\U0001f3fe\u200d\U0001f9bd|\U0001f468\U0001f3ff\u200d\U0001f9bd|\U0001f469\U0001f3fb\u200d\U0001f9bd|\U0001f469\U0001f3fc\u200d\U0001f9bd|\U0001f469\U0001f3fd\u200d\U0001f9bd|\U0001f469\U0001f3fe\u200d\U0001f9bd|\U0001f469\U0001f3ff\u200d\U0001f9bd|\U0001f3c3\u200d\u2642\ufe0f|\U0001f3c3\U0001f3fb\u200d\u2642|\U0001f3c3\U0001f3fc\u200d\u2642|\U0001f3c3\U0001f3fd\u200d\u2642|\U0001f3c3\U0001f3fe\u200d\u2642|\U0001f3c3\U0001f3ff\u200d\u2642|\U0001f3c3\u200d\u2640\ufe0f|\U0001f3c3\U0001f3fb\u200d\u2640|\U0001f3c3\U0001f3fc\u200d\u2640|\U0001f3c3\U0001f3fd\u200d\u2640|\U0001f3c3\U0001f3fe\u200d\u2640|\U0001f3c3\U0001f3ff\u200d\u2640|\U0001f46f\u200d\u2642\ufe0f|\U0001f46f\u200d\u2640\ufe0f|\U0001f9d6\u200d\u2642\ufe0f|\U0001f9d6\U0001f3fb\u200d\u2642|\U0001f9d6\U0001f3fc\u200d\u2642|\U0001f9d6\U0001f3fd\u200d\u2642|\U0001f9d6\U0001f3fe\u200d\u2642|\U0001f9d6\U0001f3ff\u200d\u2642|\U0001f9d6\u200d\u2640\ufe0f|\U0001f9d6\U0001f3fb\u200d\u2640|\U0001f9d6\U0001f3fc\u200d\u2640|\U0001f9d6\U0001f3fd\u200d\u2640|\U0001f9d6\U0001f3fe\u200d\u2640|\U0001f9d6\U0001f3ff\u200d\u2640|\U0001f9d7\u200d\u2642\ufe0f|\U0001f9d7\U0001f3fb\u200d\u2642|\U0001f9d7\U0001f3fc\u200d\u2642|\U0001f9d7\U0001f3fd\u200d\u2642|\U0001f9d7\U0001f3fe\u200d\u2642|\U0001f9d7\U0001f3ff\u200d\u2642|\U0001f9d7\u200d\u2640\ufe0f|\U0001f9d7\U0001f3fb\u200d\u2640|\U0001f9d7\U0001f3fc\u200d\u2640|\U0001f9d7\U0001f3fd\u200d\u2640|\U0001f9d7\U0001f3fe\u200d\u2640|\U0001f9d7\U0001f3ff\u200d\u2640|\U0001f3cc\u200d\u2642\ufe0f|\U0001f3cc\ufe0f\u200d\u2642|\U0001f3cc\U0001f3fb\u200d\u2642|\U0001f3cc\U0001f3fc\u200d\u2642|\U0001f3cc\U0001f3fd\u200d\u2642|\U0001f3cc\U0001f3fe\u200d\u2642|\U0001f3cc\U0001f3ff\u200d\u2642|\U0001f3cc\u200d\u2640\ufe0f|\U0001f3cc\ufe0f\u200d\u2640|\U0001f3cc\U0001f3fb\u200d\u2640|\U0001f3cc\U0001f3fc\u200d\u2640|\U0001f3cc\U0001f3fd\u200d\u2640|\U0001f3cc\U0001f3fe\u200d\u2640|\U0001f3cc\U0001f3ff\u200d\u2640|\U0001f3c4\u200d\u2642\ufe0f|\U0001f3c4\U0001f3fb\u200d\u2642|\U0001f3c4\U0001f3fc\u200d\u2642|\U0001f3c4\U0001f3fd\u200d\u2642|\U0001f3c4\U0001f3fe\u200d\u2642|\U0001f3c4\U0001f3ff\u200d\u2642|\U0001f3c4\u200d\u2640\ufe0f|\U0001f3c4\U0001f3fb\u200d\u2640|\U0001f3c4\U0001f3fc\u200d\u2640|\U0001f3c4\U0001f3fd\u200d\u2640|\U0001f3c4\U0001f3fe\u200d\u2640|\U0001f3c4\U0001f3ff\u200d\u2640|\U0001f6a3\u200d\u2642\ufe0f|\U0001f6a3\U0001f3fb\u200d\u2642|\U0001f6a3\U0001f3fc\u200d\u2642|\U0001f6a3\U0001f3fd\u200d\u2642|\U0001f6a3\U0001f3fe\u200d\u2642|\U0001f6a3\U0001f3ff\u200d\u2642|\U0001f6a3\u200d\u2640\ufe0f|\U0001f6a3\U0001f3fb\u200d\u2640|\U0001f6a3\U0001f3fc\u200d\u2640|\U0001f6a3\U0001f3fd\u200d\u2640|\U0001f6a3\U0001f3fe\u200d\u2640|\U0001f6a3\U0001f3ff\u200d\u2640|\U0001f3ca\u200d\u2642\ufe0f|\U0001f3ca\U0001f3fb\u200d\u2642|\U0001f3ca\U0001f3fc\u200d\u2642|\U0001f3ca\U0001f3fd\u200d\u2642|\U0001f3ca\U0001f3fe\u200d\u2642|\U0001f3ca\U0001f3ff\u200d\u2642|\U0001f3ca\u200d\u2640\ufe0f|\U0001f3ca\U0001f3fb\u200d\u2640|\U0001f3ca\U0001f3fc\u200d\u2640|\U0001f3ca\U0001f3fd\u200d\u2640|\U0001f3ca\U0001f3fe\u200d\u2640|\U0001f3ca\U0001f3ff\u200d\u2640|\u26f9\u200d\u2642\ufe0f|\u26f9\ufe0f\u200d\u2642|\u26f9\U0001f3fb\u200d\u2642|\u26f9\U0001f3fc\u200d\u2642|\u26f9\U0001f3fd\u200d\u2642|\u26f9\U0001f3fe\u200d\u2642|\u26f9\U0001f3ff\u200d\u2642|\u26f9\u200d\u2640\ufe0f|\u26f9\ufe0f\u200d\u2640|\u26f9\U0001f3fb\u200d\u2640|\u26f9\U0001f3fc\u200d\u2640|\u26f9\U0001f3fd\u200d\u2640|\u26f9\U0001f3fe\u200d\u2640|\u26f9\U0001f3ff\u200d\u2640|\U0001f3cb\u200d\u2642\ufe0f|\U0001f3cb\ufe0f\u200d\u2642|\U0001f3cb\U0001f3fb\u200d\u2642|\U0001f3cb\U0001f3fc\u200d\u2642|\U0001f3cb\U0001f3fd\u200d\u2642|\U0001f3cb\U0001f3fe\u200d\u2642|\U0001f3cb\U0001f3ff\u200d\u2642|\U0001f3cb\u200d\u2640\ufe0f|\U0001f3cb\ufe0f\u200d\u2640|\U0001f3cb\U0001f3fb\u200d\u2640|\U0001f3cb\U0001f3fc\u200d\u2640|\U0001f3cb\U0001f3fd\u200d\u2640|\U0001f3cb\U0001f3fe\u200d\u2640|\U0001f3cb\U0001f3ff\u200d\u2640|\U0001f6b4\u200d\u2642\ufe0f|\U0001f6b4\U0001f3fb\u200d\u2642|\U0001f6b4\U0001f3fc\u200d\u2642|\U0001f6b4\U0001f3fd\u200d\u2642|\U0001f6b4\U0001f3fe\u200d\u2642|\U0001f6b4\U0001f3ff\u200d\u2642|\U0001f6b4\u200d\u2640\ufe0f|\U0001f6b4\U0001f3fb\u200d\u2640|\U0001f6b4\U0001f3fc\u200d\u2640|\U0001f6b4\U0001f3fd\u200d\u2640|\U0001f6b4\U0001f3fe\u200d\u2640|\U0001f6b4\U0001f3ff\u200d\u2640|\U0001f6b5\u200d\u2642\ufe0f|\U0001f6b5\U0001f3fb\u200d\u2642|\U0001f6b5\U0001f3fc\u200d\u2642|\U0001f6b5\U0001f3fd\u200d\u2642|\U0001f6b5\U0001f3fe\u200d\u2642|\U0001f6b5\U0001f3ff\u200d\u2642|\U0001f6b5\u200d\u2640\ufe0f|\U0001f6b5\U0001f3fb\u200d\u2640|\U0001f6b5\U0001f3fc\u200d\u2640|\U0001f6b5\U0001f3fd\u200d\u2640|\U0001f6b5\U0001f3fe\u200d\u2640|\U0001f6b5\U0001f3ff\u200d\u2640|\U0001f938\u200d\u2642\ufe0f|\U0001f938\U0001f3fb\u200d\u2642|\U0001f938\U0001f3fc\u200d\u2642|\U0001f938\U0001f3fd\u200d\u2642|\U0001f938\U0001f3fe\u200d\u2642|\U0001f938\U0001f3ff\u200d\u2642|\U0001f938\u200d\u2640\ufe0f|\U0001f938\U0001f3fb\u200d\u2640|\U0001f938\U0001f3fc\u200d\u2640|\U0001f938\U0001f3fd\u200d\u2640|\U0001f938\U0001f3fe\u200d\u2640|\U0001f938\U0001f3ff\u200d\u2640|\U0001f93c\u200d\u2642\ufe0f|\U0001f93c\u200d\u2640\ufe0f|\U0001f93d\u200d\u2642\ufe0f|\U0001f93d\U0001f3fb\u200d\u2642|\U0001f93d\U0001f3fc\u200d\u2642|\U0001f93d\U0001f3fd\u200d\u2642|\U0001f93d\U0001f3fe\u200d\u2642|\U0001f93d\U0001f3ff\u200d\u2642|\U0001f93d\u200d\u2640\ufe0f|\U0001f93d\U0001f3fb\u200d\u2640|\U0001f93d\U0001f3fc\u200d\u2640|\U0001f93d\U0001f3fd\u200d\u2640|\U0001f93d\U0001f3fe\u200d\u2640|\U0001f93d\U0001f3ff\u200d\u2640|\U0001f93e\u200d\u2642\ufe0f|\U0001f93e\U0001f3fb\u200d\u2642|\U0001f93e\U0001f3fc\u200d\u2642|\U0001f93e\U0001f3fd\u200d\u2642|\U0001f93e\U0001f3fe\u200d\u2642|\U0001f93e\U0001f3ff\u200d\u2642|\U0001f93e\u200d\u2640\ufe0f|\U0001f93e\U0001f3fb\u200d\u2640|\U0001f93e\U0001f3fc\u200d\u2640|\U0001f93e\U0001f3fd\u200d\u2640|\U0001f93e\U0001f3fe\u200d\u2640|\U0001f93e\U0001f3ff\u200d\u2640|\U0001f939\u200d\u2642\ufe0f|\U0001f939\U0001f3fb\u200d\u2642|\U0001f939\U0001f3fc\u200d\u2642|\U0001f939\U0001f3fd\u200d\u2642|\U0001f939\U0001f3fe\u200d\u2642|\U0001f939\U0001f3ff\u200d\u2642|\U0001f939\u200d\u2640\ufe0f|\U0001f939\U0001f3fb\u200d\u2640|\U0001f939\U0001f3fc\u200d\u2640|\U0001f939\U0001f3fd\u200d\u2640|\U0001f939\U0001f3fe\u200d\u2640|\U0001f939\U0001f3ff\u200d\u2640|\U0001f9d8\u200d\u2642\ufe0f|\U0001f9d8\U0001f3fb\u200d\u2642|\U0001f9d8\U0001f3fc\u200d\u2642|\U0001f9d8\U0001f3fd\u200d\u2642|\U0001f9d8\U0001f3fe\u200d\u2642|\U0001f9d8\U0001f3ff\u200d\u2642|\U0001f9d8\u200d\u2640\ufe0f|\U0001f9d8\U0001f3fb\u200d\u2640|\U0001f9d8\U0001f3fc\u200d\u2640|\U0001f9d8\U0001f3fd\u200d\u2640|\U0001f9d8\U0001f3fe\u200d\u2640|\U0001f9d8\U0001f3ff\u200d\u2640|\U0001f3f3\ufe0f\u200d\U0001f308|\U0001f3f4\u200d\u2620\ufe0f|\U0001f441\u200d\U0001f5e8|\U0001f471\u200d\u2642|\U0001f468\u200d\U0001f9b0|\U0001f468\u200d\U0001f9b1|\U0001f468\u200d\U0001f9b3|\U0001f468\u200d\U0001f9b2|\U0001f471\u200d\u2640|\U0001f469\u200d\U0001f9b0|\U0001f469\u200d\U0001f9b1|\U0001f469\u200d\U0001f9b3|\U0001f469\u200d\U0001f9b2|\U0001f64d\u200d\u2642|\U0001f64d\u200d\u2640|\U0001f64e\u200d\u2642|\U0001f64e\u200d\u2640|\U0001f645\u200d\u2642|\U0001f645\u200d\u2640|\U0001f646\u200d\u2642|\U0001f646\u200d\u2640|\U0001f481\u200d\u2642|\U0001f481\u200d\u2640|\U0001f64b\u200d\u2642|\U0001f64b\u200d\u2640|\U0001f9cf\u200d\u2642|\U0001f9cf\u200d\u2640|\U0001f647\u200d\u2642|\U0001f647\u200d\u2640|\U0001f926\u200d\u2642|\U0001f926\u200d\u2640|\U0001f937\u200d\u2642|\U0001f937\u200d\u2640|\U0001f468\u200d\u2695|\U0001f469\u200d\u2695|\U0001f468\u200d\U0001f393|\U0001f469\u200d\U0001f393|\U0001f468\u200d\U0001f3eb|\U0001f469\u200d\U0001f3eb|\U0001f468\u200d\u2696|\U0001f469\u200d\u2696|\U0001f468\u200d\U0001f33e|\U0001f469\u200d\U0001f33e|\U0001f468\u200d\U0001f373|\U0001f469\u200d\U0001f373|\U0001f468\u200d\U0001f527|\U0001f469\u200d\U0001f527|\U0001f468\u200d\U0001f3ed|\U0001f469\u200d\U0001f3ed|\U0001f468\u200d\U0001f4bc|\U0001f469\u200d\U0001f4bc|\U0001f468\u200d\U0001f52c|\U0001f469\u200d\U0001f52c|\U0001f468\u200d\U0001f4bb|\U0001f469\u200d\U0001f4bb|\U0001f468\u200d\U0001f3a4|\U0001f469\u200d\U0001f3a4|\U0001f468\u200d\U0001f3a8|\U0001f469\u200d\U0001f3a8|\U0001f468\u200d\u2708|\U0001f469\u200d\u2708|\U0001f468\u200d\U0001f680|\U0001f469\u200d\U0001f680|\U0001f468\u200d\U0001f692|\U0001f469\u200d\U0001f692|\U0001f46e\u200d\u2642|\U0001f46e\u200d\u2640|\U0001f575\u200d\u2642|\U0001f575\u200d\u2640|\U0001f482\u200d\u2642|\U0001f482\u200d\u2640|\U0001f477\u200d\u2642|\U0001f477\u200d\u2640|\U0001f473\u200d\u2642|\U0001f473\u200d\u2640|\U0001f9b8\u200d\u2642|\U0001f9b8\u200d\u2640|\U0001f9b9\u200d\u2642|\U0001f9b9\u200d\u2640|\U0001f9d9\u200d\u2642|\U0001f9d9\u200d\u2640|\U0001f9da\u200d\u2642|\U0001f9da\u200d\u2640|\U0001f9db\u200d\u2642|\U0001f9db\u200d\u2640|\U0001f9dc\u200d\u2642|\U0001f9dc\u200d\u2640|\U0001f9dd\u200d\u2642|\U0001f9dd\u200d\u2640|\U0001f9de\u200d\u2642|\U0001f9de\u200d\u2640|\U0001f9df\u200d\u2642|\U0001f9df\u200d\u2640|\U0001f486\u200d\u2642|\U0001f486\u200d\u2640|\U0001f487\u200d\u2642|\U0001f487\u200d\u2640|\U0001f6b6\u200d\u2642|\U0001f6b6\u200d\u2640|\U0001f9cd\u200d\u2642|\U0001f9cd\u200d\u2640|\U0001f9ce\u200d\u2642|\U0001f9ce\u200d\u2640|\U0001f468\u200d\U0001f9af|\U0001f469\u200d\U0001f9af|\U0001f468\u200d\U0001f9bc|\U0001f469\u200d\U0001f9bc|\U0001f468\u200d\U0001f9bd|\U0001f469\u200d\U0001f9bd|\U0001f3c3\u200d\u2642|\U0001f3c3\u200d\u2640|\U0001f46f\u200d\u2642|\U0001f46f\u200d\u2640|\U0001f9d6\u200d\u2642|\U0001f9d6\u200d\u2640|\U0001f9d7\u200d\u2642|\U0001f9d7\u200d\u2640|\U0001f3cc\u200d\u2642|\U0001f3cc\u200d\u2640|\U0001f3c4\u200d\u2642|\U0001f3c4\u200d\u2640|\U0001f6a3\u200d\u2642|\U0001f6a3\u200d\u2640|\U0001f3ca\u200d\u2642|\U0001f3ca\u200d\u2640|\u26f9\u200d\u2642|\u26f9\u200d\u2640|\U0001f3cb\u200d\u2642|\U0001f3cb\u200d\u2640|\U0001f6b4\u200d\u2642|\U0001f6b4\u200d\u2640|\U0001f6b5\u200d\u2642|\U0001f6b5\u200d\u2640|\U0001f938\u200d\u2642|\U0001f938\u200d\u2640|\U0001f93c\u200d\u2642|\U0001f93c\u200d\u2640|\U0001f93d\u200d\u2642|\U0001f93d\u200d\u2640|\U0001f93e\u200d\u2642|\U0001f93e\u200d\u2640|\U0001f939\u200d\u2642|\U0001f939\u200d\u2640|\U0001f9d8\u200d\u2642|\U0001f9d8\u200d\u2640|\U0001f468\u200d\U0001f466|\U0001f468\u200d\U0001f467|\U0001f469\u200d\U0001f466|\U0001f469\u200d\U0001f467|\U0001f415\u200d\U0001f9ba|\\#\ufe0f\u20e3|\\*\ufe0f\u20e3|0\ufe0f\u20e3|1\ufe0f\u20e3|2\ufe0f\u20e3|3\ufe0f\u20e3|4\ufe0f\u20e3|5\ufe0f\u20e3|6\ufe0f\u20e3|7\ufe0f\u20e3|8\ufe0f\u20e3|9\ufe0f\u20e3|\U0001f3f3\u200d\U0001f308|\U0001f3f4\u200d\u2620|\u263a\ufe0f|\u2639\ufe0f|\u2620\ufe0f|\u2763\ufe0f|\u2764\ufe0f|\U0001f573\ufe0f|\U0001f5e8\ufe0f|\U0001f5ef\ufe0f|\U0001f44b\U0001f3fb|\U0001f44b\U0001f3fc|\U0001f44b\U0001f3fd|\U0001f44b\U0001f3fe|\U0001f44b\U0001f3ff|\U0001f91a\U0001f3fb|\U0001f91a\U0001f3fc|\U0001f91a\U0001f3fd|\U0001f91a\U0001f3fe|\U0001f91a\U0001f3ff|\U0001f590\ufe0f|\U0001f590\U0001f3fb|\U0001f590\U0001f3fc|\U0001f590\U0001f3fd|\U0001f590\U0001f3fe|\U0001f590\U0001f3ff|\u270b\U0001f3fb|\u270b\U0001f3fc|\u270b\U0001f3fd|\u270b\U0001f3fe|\u270b\U0001f3ff|\U0001f596\U0001f3fb|\U0001f596\U0001f3fc|\U0001f596\U0001f3fd|\U0001f596\U0001f3fe|\U0001f596\U0001f3ff|\U0001f44c\U0001f3fb|\U0001f44c\U0001f3fc|\U0001f44c\U0001f3fd|\U0001f44c\U0001f3fe|\U0001f44c\U0001f3ff|\U0001f90f\U0001f3fb|\U0001f90f\U0001f3fc|\U0001f90f\U0001f3fd|\U0001f90f\U0001f3fe|\U0001f90f\U0001f3ff|\u270c\ufe0f|\u270c\U0001f3fb|\u270c\U0001f3fc|\u270c\U0001f3fd|\u270c\U0001f3fe|\u270c\U0001f3ff|\U0001f91e\U0001f3fb|\U0001f91e\U0001f3fc|\U0001f91e\U0001f3fd|\U0001f91e\U0001f3fe|\U0001f91e\U0001f3ff|\U0001f91f\U0001f3fb|\U0001f91f\U0001f3fc|\U0001f91f\U0001f3fd|\U0001f91f\U0001f3fe|\U0001f91f\U0001f3ff|\U0001f918\U0001f3fb|\U0001f918\U0001f3fc|\U0001f918\U0001f3fd|\U0001f918\U0001f3fe|\U0001f918\U0001f3ff|\U0001f919\U0001f3fb|\U0001f919\U0001f3fc|\U0001f919\U0001f3fd|\U0001f919\U0001f3fe|\U0001f919\U0001f3ff|\U0001f448\U0001f3fb|\U0001f448\U0001f3fc|\U0001f448\U0001f3fd|\U0001f448\U0001f3fe|\U0001f448\U0001f3ff|\U0001f449\U0001f3fb|\U0001f449\U0001f3fc|\U0001f449\U0001f3fd|\U0001f449\U0001f3fe|\U0001f449\U0001f3ff|\U0001f446\U0001f3fb|\U0001f446\U0001f3fc|\U0001f446\U0001f3fd|\U0001f446\U0001f3fe|\U0001f446\U0001f3ff|\U0001f595\U0001f3fb|\U0001f595\U0001f3fc|\U0001f595\U0001f3fd|\U0001f595\U0001f3fe|\U0001f595\U0001f3ff|\U0001f447\U0001f3fb|\U0001f447\U0001f3fc|\U0001f447\U0001f3fd|\U0001f447\U0001f3fe|\U0001f447\U0001f3ff|\u261d\ufe0f|\u261d\U0001f3fb|\u261d\U0001f3fc|\u261d\U0001f3fd|\u261d\U0001f3fe|\u261d\U0001f3ff|\U0001f44d\U0001f3fb|\U0001f44d\U0001f3fc|\U0001f44d\U0001f3fd|\U0001f44d\U0001f3fe|\U0001f44d\U0001f3ff|\U0001f44e\U0001f3fb|\U0001f44e\U0001f3fc|\U0001f44e\U0001f3fd|\U0001f44e\U0001f3fe|\U0001f44e\U0001f3ff|\u270a\U0001f3fb|\u270a\U0001f3fc|\u270a\U0001f3fd|\u270a\U0001f3fe|\u270a\U0001f3ff|\U0001f44a\U0001f3fb|\U0001f44a\U0001f3fc|\U0001f44a\U0001f3fd|\U0001f44a\U0001f3fe|\U0001f44a\U0001f3ff|\U0001f91b\U0001f3fb|\U0001f91b\U0001f3fc|\U0001f91b\U0001f3fd|\U0001f91b\U0001f3fe|\U0001f91b\U0001f3ff|\U0001f91c\U0001f3fb|\U0001f91c\U0001f3fc|\U0001f91c\U0001f3fd|\U0001f91c\U0001f3fe|\U0001f91c\U0001f3ff|\U0001f44f\U0001f3fb|\U0001f44f\U0001f3fc|\U0001f44f\U0001f3fd|\U0001f44f\U0001f3fe|\U0001f44f\U0001f3ff|\U0001f64c\U0001f3fb|\U0001f64c\U0001f3fc|\U0001f64c\U0001f3fd|\U0001f64c\U0001f3fe|\U0001f64c\U0001f3ff|\U0001f450\U0001f3fb|\U0001f450\U0001f3fc|\U0001f450\U0001f3fd|\U0001f450\U0001f3fe|\U0001f450\U0001f3ff|\U0001f932\U0001f3fb|\U0001f932\U0001f3fc|\U0001f932\U0001f3fd|\U0001f932\U0001f3fe|\U0001f932\U0001f3ff|\U0001f64f\U0001f3fb|\U0001f64f\U0001f3fc|\U0001f64f\U0001f3fd|\U0001f64f\U0001f3fe|\U0001f64f\U0001f3ff|\u270d\ufe0f|\u270d\U0001f3fb|\u270d\U0001f3fc|\u270d\U0001f3fd|\u270d\U0001f3fe|\u270d\U0001f3ff|\U0001f485\U0001f3fb|\U0001f485\U0001f3fc|\U0001f485\U0001f3fd|\U0001f485\U0001f3fe|\U0001f485\U0001f3ff|\U0001f933\U0001f3fb|\U0001f933\U0001f3fc|\U0001f933\U0001f3fd|\U0001f933\U0001f3fe|\U0001f933\U0001f3ff|\U0001f4aa\U0001f3fb|\U0001f4aa\U0001f3fc|\U0001f4aa\U0001f3fd|\U0001f4aa\U0001f3fe|\U0001f4aa\U0001f3ff|\U0001f9b5\U0001f3fb|\U0001f9b5\U0001f3fc|\U0001f9b5\U0001f3fd|\U0001f9b5\U0001f3fe|\U0001f9b5\U0001f3ff|\U0001f9b6\U0001f3fb|\U0001f9b6\U0001f3fc|\U0001f9b6\U0001f3fd|\U0001f9b6\U0001f3fe|\U0001f9b6\U0001f3ff|\U0001f442\U0001f3fb|\U0001f442\U0001f3fc|\U0001f442\U0001f3fd|\U0001f442\U0001f3fe|\U0001f442\U0001f3ff|\U0001f9bb\U0001f3fb|\U0001f9bb\U0001f3fc|\U0001f9bb\U0001f3fd|\U0001f9bb\U0001f3fe|\U0001f9bb\U0001f3ff|\U0001f443\U0001f3fb|\U0001f443\U0001f3fc|\U0001f443\U0001f3fd|\U0001f443\U0001f3fe|\U0001f443\U0001f3ff|\U0001f441\ufe0f|\U0001f476\U0001f3fb|\U0001f476\U0001f3fc|\U0001f476\U0001f3fd|\U0001f476\U0001f3fe|\U0001f476\U0001f3ff|\U0001f9d2\U0001f3fb|\U0001f9d2\U0001f3fc|\U0001f9d2\U0001f3fd|\U0001f9d2\U0001f3fe|\U0001f9d2\U0001f3ff|\U0001f466\U0001f3fb|\U0001f466\U0001f3fc|\U0001f466\U0001f3fd|\U0001f466\U0001f3fe|\U0001f466\U0001f3ff|\U0001f467\U0001f3fb|\U0001f467\U0001f3fc|\U0001f467\U0001f3fd|\U0001f467\U0001f3fe|\U0001f467\U0001f3ff|\U0001f9d1\U0001f3fb|\U0001f9d1\U0001f3fc|\U0001f9d1\U0001f3fd|\U0001f9d1\U0001f3fe|\U0001f9d1\U0001f3ff|\U0001f471\U0001f3fb|\U0001f471\U0001f3fc|\U0001f471\U0001f3fd|\U0001f471\U0001f3fe|\U0001f471\U0001f3ff|\U0001f468\U0001f3fb|\U0001f468\U0001f3fc|\U0001f468\U0001f3fd|\U0001f468\U0001f3fe|\U0001f468\U0001f3ff|\U0001f9d4\U0001f3fb|\U0001f9d4\U0001f3fc|\U0001f9d4\U0001f3fd|\U0001f9d4\U0001f3fe|\U0001f9d4\U0001f3ff|\U0001f469\U0001f3fb|\U0001f469\U0001f3fc|\U0001f469\U0001f3fd|\U0001f469\U0001f3fe|\U0001f469\U0001f3ff|\U0001f9d3\U0001f3fb|\U0001f9d3\U0001f3fc|\U0001f9d3\U0001f3fd|\U0001f9d3\U0001f3fe|\U0001f9d3\U0001f3ff|\U0001f474\U0001f3fb|\U0001f474\U0001f3fc|\U0001f474\U0001f3fd|\U0001f474\U0001f3fe|\U0001f474\U0001f3ff|\U0001f475\U0001f3fb|\U0001f475\U0001f3fc|\U0001f475\U0001f3fd|\U0001f475\U0001f3fe|\U0001f475\U0001f3ff|\U0001f64d\U0001f3fb|\U0001f64d\U0001f3fc|\U0001f64d\U0001f3fd|\U0001f64d\U0001f3fe|\U0001f64d\U0001f3ff|\U0001f64e\U0001f3fb|\U0001f64e\U0001f3fc|\U0001f64e\U0001f3fd|\U0001f64e\U0001f3fe|\U0001f64e\U0001f3ff|\U0001f645\U0001f3fb|\U0001f645\U0001f3fc|\U0001f645\U0001f3fd|\U0001f645\U0001f3fe|\U0001f645\U0001f3ff|\U0001f646\U0001f3fb|\U0001f646\U0001f3fc|\U0001f646\U0001f3fd|\U0001f646\U0001f3fe|\U0001f646\U0001f3ff|\U0001f481\U0001f3fb|\U0001f481\U0001f3fc|\U0001f481\U0001f3fd|\U0001f481\U0001f3fe|\U0001f481\U0001f3ff|\U0001f64b\U0001f3fb|\U0001f64b\U0001f3fc|\U0001f64b\U0001f3fd|\U0001f64b\U0001f3fe|\U0001f64b\U0001f3ff|\U0001f9cf\U0001f3fb|\U0001f9cf\U0001f3fc|\U0001f9cf\U0001f3fd|\U0001f9cf\U0001f3fe|\U0001f9cf\U0001f3ff|\U0001f647\U0001f3fb|\U0001f647\U0001f3fc|\U0001f647\U0001f3fd|\U0001f647\U0001f3fe|\U0001f647\U0001f3ff|\U0001f926\U0001f3fb|\U0001f926\U0001f3fc|\U0001f926\U0001f3fd|\U0001f926\U0001f3fe|\U0001f926\U0001f3ff|\U0001f937\U0001f3fb|\U0001f937\U0001f3fc|\U0001f937\U0001f3fd|\U0001f937\U0001f3fe|\U0001f937\U0001f3ff|\U0001f46e\U0001f3fb|\U0001f46e\U0001f3fc|\U0001f46e\U0001f3fd|\U0001f46e\U0001f3fe|\U0001f46e\U0001f3ff|\U0001f575\ufe0f|\U0001f575\U0001f3fb|\U0001f575\U0001f3fc|\U0001f575\U0001f3fd|\U0001f575\U0001f3fe|\U0001f575\U0001f3ff|\U0001f482\U0001f3fb|\U0001f482\U0001f3fc|\U0001f482\U0001f3fd|\U0001f482\U0001f3fe|\U0001f482\U0001f3ff|\U0001f477\U0001f3fb|\U0001f477\U0001f3fc|\U0001f477\U0001f3fd|\U0001f477\U0001f3fe|\U0001f477\U0001f3ff|\U0001f934\U0001f3fb|\U0001f934\U0001f3fc|\U0001f934\U0001f3fd|\U0001f934\U0001f3fe|\U0001f934\U0001f3ff|\U0001f478\U0001f3fb|\U0001f478\U0001f3fc|\U0001f478\U0001f3fd|\U0001f478\U0001f3fe|\U0001f478\U0001f3ff|\U0001f473\U0001f3fb|\U0001f473\U0001f3fc|\U0001f473\U0001f3fd|\U0001f473\U0001f3fe|\U0001f473\U0001f3ff|\U0001f472\U0001f3fb|\U0001f472\U0001f3fc|\U0001f472\U0001f3fd|\U0001f472\U0001f3fe|\U0001f472\U0001f3ff|\U0001f9d5\U0001f3fb|\U0001f9d5\U0001f3fc|\U0001f9d5\U0001f3fd|\U0001f9d5\U0001f3fe|\U0001f9d5\U0001f3ff|\U0001f935\U0001f3fb|\U0001f935\U0001f3fc|\U0001f935\U0001f3fd|\U0001f935\U0001f3fe|\U0001f935\U0001f3ff|\U0001f470\U0001f3fb|\U0001f470\U0001f3fc|\U0001f470\U0001f3fd|\U0001f470\U0001f3fe|\U0001f470\U0001f3ff|\U0001f930\U0001f3fb|\U0001f930\U0001f3fc|\U0001f930\U0001f3fd|\U0001f930\U0001f3fe|\U0001f930\U0001f3ff|\U0001f931\U0001f3fb|\U0001f931\U0001f3fc|\U0001f931\U0001f3fd|\U0001f931\U0001f3fe|\U0001f931\U0001f3ff|\U0001f47c\U0001f3fb|\U0001f47c\U0001f3fc|\U0001f47c\U0001f3fd|\U0001f47c\U0001f3fe|\U0001f47c\U0001f3ff|\U0001f385\U0001f3fb|\U0001f385\U0001f3fc|\U0001f385\U0001f3fd|\U0001f385\U0001f3fe|\U0001f385\U0001f3ff|\U0001f936\U0001f3fb|\U0001f936\U0001f3fc|\U0001f936\U0001f3fd|\U0001f936\U0001f3fe|\U0001f936\U0001f3ff|\U0001f9b8\U0001f3fb|\U0001f9b8\U0001f3fc|\U0001f9b8\U0001f3fd|\U0001f9b8\U0001f3fe|\U0001f9b8\U0001f3ff|\U0001f9b9\U0001f3fb|\U0001f9b9\U0001f3fc|\U0001f9b9\U0001f3fd|\U0001f9b9\U0001f3fe|\U0001f9b9\U0001f3ff|\U0001f9d9\U0001f3fb|\U0001f9d9\U0001f3fc|\U0001f9d9\U0001f3fd|\U0001f9d9\U0001f3fe|\U0001f9d9\U0001f3ff|\U0001f9da\U0001f3fb|\U0001f9da\U0001f3fc|\U0001f9da\U0001f3fd|\U0001f9da\U0001f3fe|\U0001f9da\U0001f3ff|\U0001f9db\U0001f3fb|\U0001f9db\U0001f3fc|\U0001f9db\U0001f3fd|\U0001f9db\U0001f3fe|\U0001f9db\U0001f3ff|\U0001f9dc\U0001f3fb|\U0001f9dc\U0001f3fc|\U0001f9dc\U0001f3fd|\U0001f9dc\U0001f3fe|\U0001f9dc\U0001f3ff|\U0001f9dd\U0001f3fb|\U0001f9dd\U0001f3fc|\U0001f9dd\U0001f3fd|\U0001f9dd\U0001f3fe|\U0001f9dd\U0001f3ff|\U0001f486\U0001f3fb|\U0001f486\U0001f3fc|\U0001f486\U0001f3fd|\U0001f486\U0001f3fe|\U0001f486\U0001f3ff|\U0001f487\U0001f3fb|\U0001f487\U0001f3fc|\U0001f487\U0001f3fd|\U0001f487\U0001f3fe|\U0001f487\U0001f3ff|\U0001f6b6\U0001f3fb|\U0001f6b6\U0001f3fc|\U0001f6b6\U0001f3fd|\U0001f6b6\U0001f3fe|\U0001f6b6\U0001f3ff|\U0001f9cd\U0001f3fb|\U0001f9cd\U0001f3fc|\U0001f9cd\U0001f3fd|\U0001f9cd\U0001f3fe|\U0001f9cd\U0001f3ff|\U0001f9ce\U0001f3fb|\U0001f9ce\U0001f3fc|\U0001f9ce\U0001f3fd|\U0001f9ce\U0001f3fe|\U0001f9ce\U0001f3ff|\U0001f3c3\U0001f3fb|\U0001f3c3\U0001f3fc|\U0001f3c3\U0001f3fd|\U0001f3c3\U0001f3fe|\U0001f3c3\U0001f3ff|\U0001f483\U0001f3fb|\U0001f483\U0001f3fc|\U0001f483\U0001f3fd|\U0001f483\U0001f3fe|\U0001f483\U0001f3ff|\U0001f57a\U0001f3fb|\U0001f57a\U0001f3fc|\U0001f57a\U0001f3fd|\U0001f57a\U0001f3fe|\U0001f57a\U0001f3ff|\U0001f574\ufe0f|\U0001f574\U0001f3fb|\U0001f574\U0001f3fc|\U0001f574\U0001f3fd|\U0001f574\U0001f3fe|\U0001f574\U0001f3ff|\U0001f9d6\U0001f3fb|\U0001f9d6\U0001f3fc|\U0001f9d6\U0001f3fd|\U0001f9d6\U0001f3fe|\U0001f9d6\U0001f3ff|\U0001f9d7\U0001f3fb|\U0001f9d7\U0001f3fc|\U0001f9d7\U0001f3fd|\U0001f9d7\U0001f3fe|\U0001f9d7\U0001f3ff|\U0001f3c7\U0001f3fb|\U0001f3c7\U0001f3fc|\U0001f3c7\U0001f3fd|\U0001f3c7\U0001f3fe|\U0001f3c7\U0001f3ff|\u26f7\ufe0f|\U0001f3c2\U0001f3fb|\U0001f3c2\U0001f3fc|\U0001f3c2\U0001f3fd|\U0001f3c2\U0001f3fe|\U0001f3c2\U0001f3ff|\U0001f3cc\ufe0f|\U0001f3cc\U0001f3fb|\U0001f3cc\U0001f3fc|\U0001f3cc\U0001f3fd|\U0001f3cc\U0001f3fe|\U0001f3cc\U0001f3ff|\U0001f3c4\U0001f3fb|\U0001f3c4\U0001f3fc|\U0001f3c4\U0001f3fd|\U0001f3c4\U0001f3fe|\U0001f3c4\U0001f3ff|\U0001f6a3\U0001f3fb|\U0001f6a3\U0001f3fc|\U0001f6a3\U0001f3fd|\U0001f6a3\U0001f3fe|\U0001f6a3\U0001f3ff|\U0001f3ca\U0001f3fb|\U0001f3ca\U0001f3fc|\U0001f3ca\U0001f3fd|\U0001f3ca\U0001f3fe|\U0001f3ca\U0001f3ff|\u26f9\ufe0f|\u26f9\U0001f3fb|\u26f9\U0001f3fc|\u26f9\U0001f3fd|\u26f9\U0001f3fe|\u26f9\U0001f3ff|\U0001f3cb\ufe0f|\U0001f3cb\U0001f3fb|\U0001f3cb\U0001f3fc|\U0001f3cb\U0001f3fd|\U0001f3cb\U0001f3fe|\U0001f3cb\U0001f3ff|\U0001f6b4\U0001f3fb|\U0001f6b4\U0001f3fc|\U0001f6b4\U0001f3fd|\U0001f6b4\U0001f3fe|\U0001f6b4\U0001f3ff|\U0001f6b5\U0001f3fb|\U0001f6b5\U0001f3fc|\U0001f6b5\U0001f3fd|\U0001f6b5\U0001f3fe|\U0001f6b5\U0001f3ff|\U0001f938\U0001f3fb|\U0001f938\U0001f3fc|\U0001f938\U0001f3fd|\U0001f938\U0001f3fe|\U0001f938\U0001f3ff|\U0001f93d\U0001f3fb|\U0001f93d\U0001f3fc|\U0001f93d\U0001f3fd|\U0001f93d\U0001f3fe|\U0001f93d\U0001f3ff|\U0001f93e\U0001f3fb|\U0001f93e\U0001f3fc|\U0001f93e\U0001f3fd|\U0001f93e\U0001f3fe|\U0001f93e\U0001f3ff|\U0001f939\U0001f3fb|\U0001f939\U0001f3fc|\U0001f939\U0001f3fd|\U0001f939\U0001f3fe|\U0001f939\U0001f3ff|\U0001f9d8\U0001f3fb|\U0001f9d8\U0001f3fc|\U0001f9d8\U0001f3fd|\U0001f9d8\U0001f3fe|\U0001f9d8\U0001f3ff|\U0001f6c0\U0001f3fb|\U0001f6c0\U0001f3fc|\U0001f6c0\U0001f3fd|\U0001f6c0\U0001f3fe|\U0001f6c0\U0001f3ff|\U0001f6cc\U0001f3fb|\U0001f6cc\U0001f3fc|\U0001f6cc\U0001f3fd|\U0001f6cc\U0001f3fe|\U0001f6cc\U0001f3ff|\U0001f46d\U0001f3fb|\U0001f46d\U0001f3fc|\U0001f46d\U0001f3fd|\U0001f46d\U0001f3fe|\U0001f46d\U0001f3ff|\U0001f46b\U0001f3fb|\U0001f46b\U0001f3fc|\U0001f46b\U0001f3fd|\U0001f46b\U0001f3fe|\U0001f46b\U0001f3ff|\U0001f46c\U0001f3fb|\U0001f46c\U0001f3fc|\U0001f46c\U0001f3fd|\U0001f46c\U0001f3fe|\U0001f46c\U0001f3ff|\U0001f5e3\ufe0f|\U0001f43f\ufe0f|\U0001f54a\ufe0f|\U0001f577\ufe0f|\U0001f578\ufe0f|\U0001f3f5\ufe0f|\u2618\ufe0f|\U0001f336\ufe0f|\U0001f37d\ufe0f|\U0001f5fa\ufe0f|\U0001f3d4\ufe0f|\u26f0\ufe0f|\U0001f3d5\ufe0f|\U0001f3d6\ufe0f|\U0001f3dc\ufe0f|\U0001f3dd\ufe0f|\U0001f3de\ufe0f|\U0001f3df\ufe0f|\U0001f3db\ufe0f|\U0001f3d7\ufe0f|\U0001f3d8\ufe0f|\U0001f3da\ufe0f|\u26e9\ufe0f|\U0001f3d9\ufe0f|\u2668\ufe0f|\U0001f3ce\ufe0f|\U0001f3cd\ufe0f|\U0001f6e3\ufe0f|\U0001f6e4\ufe0f|\U0001f6e2\ufe0f|\U0001f6f3\ufe0f|\u26f4\ufe0f|\U0001f6e5\ufe0f|\u2708\ufe0f|\U0001f6e9\ufe0f|\U0001f6f0\ufe0f|\U0001f6ce\ufe0f|\u23f1\ufe0f|\u23f2\ufe0f|\U0001f570\ufe0f|\U0001f321\ufe0f|\u2600\ufe0f|\u2601\ufe0f|\u26c8\ufe0f|\U0001f324\ufe0f|\U0001f325\ufe0f|\U0001f326\ufe0f|\U0001f327\ufe0f|\U0001f328\ufe0f|\U0001f329\ufe0f|\U0001f32a\ufe0f|\U0001f32b\ufe0f|\U0001f32c\ufe0f|\u2602\ufe0f|\u26f1\ufe0f|\u2744\ufe0f|\u2603\ufe0f|\u2604\ufe0f|\U0001f397\ufe0f|\U0001f39f\ufe0f|\U0001f396\ufe0f|\u26f8\ufe0f|\U0001f579\ufe0f|\u2660\ufe0f|\u2665\ufe0f|\u2666\ufe0f|\u2663\ufe0f|\u265f\ufe0f|\U0001f5bc\ufe0f|\U0001f576\ufe0f|\U0001f6cd\ufe0f|\u26d1\ufe0f|\U0001f399\ufe0f|\U0001f39a\ufe0f|\U0001f39b\ufe0f|\u260e\ufe0f|\U0001f5a5\ufe0f|\U0001f5a8\ufe0f|\u2328\ufe0f|\U0001f5b1\ufe0f|\U0001f5b2\ufe0f|\U0001f39e\ufe0f|\U0001f4fd\ufe0f|\U0001f56f\ufe0f|\U0001f5de\ufe0f|\U0001f3f7\ufe0f|\u2709\ufe0f|\U0001f5f3\ufe0f|\u270f\ufe0f|\u2712\ufe0f|\U0001f58b\ufe0f|\U0001f58a\ufe0f|\U0001f58c\ufe0f|\U0001f58d\ufe0f|\U0001f5c2\ufe0f|\U0001f5d2\ufe0f|\U0001f5d3\ufe0f|\U0001f587\ufe0f|\u2702\ufe0f|\U0001f5c3\ufe0f|\U0001f5c4\ufe0f|\U0001f5d1\ufe0f|\U0001f5dd\ufe0f|\u26cf\ufe0f|\u2692\ufe0f|\U0001f6e0\ufe0f|\U0001f5e1\ufe0f|\u2694\ufe0f|\U0001f6e1\ufe0f|\u2699\ufe0f|\U0001f5dc\ufe0f|\u2696\ufe0f|\u26d3\ufe0f|\u2697\ufe0f|\U0001f6cf\ufe0f|\U0001f6cb\ufe0f|\u26b0\ufe0f|\u26b1\ufe0f|\u26a0\ufe0f|\u2622\ufe0f|\u2623\ufe0f|\u2b06\ufe0f|\u2197\ufe0f|\u27a1\ufe0f|\u2198\ufe0f|\u2b07\ufe0f|\u2199\ufe0f|\u2b05\ufe0f|\u2196\ufe0f|\u2195\ufe0f|\u2194\ufe0f|\u21a9\ufe0f|\u21aa\ufe0f|\u2934\ufe0f|\u2935\ufe0f|\u269b\ufe0f|\U0001f549\ufe0f|\u2721\ufe0f|\u2638\ufe0f|\u262f\ufe0f|\u271d\ufe0f|\u2626\ufe0f|\u262a\ufe0f|\u262e\ufe0f|\u25b6\ufe0f|\u23ed\ufe0f|\u23ef\ufe0f|\u25c0\ufe0f|\u23ee\ufe0f|\u23f8\ufe0f|\u23f9\ufe0f|\u23fa\ufe0f|\u23cf\ufe0f|\u2640\ufe0f|\u2642\ufe0f|\u2695\ufe0f|\u267e\ufe0f|\u267b\ufe0f|\u269c\ufe0f|\u2611\ufe0f|\u2714\ufe0f|\u2716\ufe0f|\u303d\ufe0f|\u2733\ufe0f|\u2734\ufe0f|\u2747\ufe0f|\u203c\ufe0f|\u2049\ufe0f|\u3030\ufe0f|\xa9\ufe0f|\xae\ufe0f|\u2122\ufe0f|\\#\u20e3|\\*\u20e3|0\u20e3|1\u20e3|2\u20e3|3\u20e3|4\u20e3|5\u20e3|6\u20e3|7\u20e3|8\u20e3|9\u20e3|\U0001f170\ufe0f|\U0001f171\ufe0f|\u2139\ufe0f|\u24c2\ufe0f|\U0001f17e\ufe0f|\U0001f17f\ufe0f|\U0001f202\ufe0f|\U0001f237\ufe0f|\u3297\ufe0f|\u3299\ufe0f|\u25fc\ufe0f|\u25fb\ufe0f|\u25aa\ufe0f|\u25ab\ufe0f|\U0001f3f3\ufe0f|\U0001f1e6\U0001f1e8|\U0001f1e6\U0001f1e9|\U0001f1e6\U0001f1ea|\U0001f1e6\U0001f1eb|\U0001f1e6\U0001f1ec|\U0001f1e6\U0001f1ee|\U0001f1e6\U0001f1f1|\U0001f1e6\U0001f1f2|\U0001f1e6\U0001f1f4|\U0001f1e6\U0001f1f6|\U0001f1e6\U0001f1f7|\U0001f1e6\U0001f1f8|\U0001f1e6\U0001f1f9|\U0001f1e6\U0001f1fa|\U0001f1e6\U0001f1fc|\U0001f1e6\U0001f1fd|\U0001f1e6\U0001f1ff|\U0001f1e7\U0001f1e6|\U0001f1e7\U0001f1e7|\U0001f1e7\U0001f1e9|\U0001f1e7\U0001f1ea|\U0001f1e7\U0001f1eb|\U0001f1e7\U0001f1ec|\U0001f1e7\U0001f1ed|\U0001f1e7\U0001f1ee|\U0001f1e7\U0001f1ef|\U0001f1e7\U0001f1f1|\U0001f1e7\U0001f1f2|\U0001f1e7\U0001f1f3|\U0001f1e7\U0001f1f4|\U0001f1e7\U0001f1f6|\U0001f1e7\U0001f1f7|\U0001f1e7\U0001f1f8|\U0001f1e7\U0001f1f9|\U0001f1e7\U0001f1fb|\U0001f1e7\U0001f1fc|\U0001f1e7\U0001f1fe|\U0001f1e7\U0001f1ff|\U0001f1e8\U0001f1e6|\U0001f1e8\U0001f1e8|\U0001f1e8\U0001f1e9|\U0001f1e8\U0001f1eb|\U0001f1e8\U0001f1ec|\U0001f1e8\U0001f1ed|\U0001f1e8\U0001f1ee|\U0001f1e8\U0001f1f0|\U0001f1e8\U0001f1f1|\U0001f1e8\U0001f1f2|\U0001f1e8\U0001f1f3|\U0001f1e8\U0001f1f4|\U0001f1e8\U0001f1f5|\U0001f1e8\U0001f1f7|\U0001f1e8\U0001f1fa|\U0001f1e8\U0001f1fb|\U0001f1e8\U0001f1fc|\U0001f1e8\U0001f1fd|\U0001f1e8\U0001f1fe|\U0001f1e8\U0001f1ff|\U0001f1e9\U0001f1ea|\U0001f1e9\U0001f1ec|\U0001f1e9\U0001f1ef|\U0001f1e9\U0001f1f0|\U0001f1e9\U0001f1f2|\U0001f1e9\U0001f1f4|\U0001f1e9\U0001f1ff|\U0001f1ea\U0001f1e6|\U0001f1ea\U0001f1e8|\U0001f1ea\U0001f1ea|\U0001f1ea\U0001f1ec|\U0001f1ea\U0001f1ed|\U0001f1ea\U0001f1f7|\U0001f1ea\U0001f1f8|\U0001f1ea\U0001f1f9|\U0001f1ea\U0001f1fa|\U0001f1eb\U0001f1ee|\U0001f1eb\U0001f1ef|\U0001f1eb\U0001f1f0|\U0001f1eb\U0001f1f2|\U0001f1eb\U0001f1f4|\U0001f1eb\U0001f1f7|\U0001f1ec\U0001f1e6|\U0001f1ec\U0001f1e7|\U0001f1ec\U0001f1e9|\U0001f1ec\U0001f1ea|\U0001f1ec\U0001f1eb|\U0001f1ec\U0001f1ec|\U0001f1ec\U0001f1ed|\U0001f1ec\U0001f1ee|\U0001f1ec\U0001f1f1|\U0001f1ec\U0001f1f2|\U0001f1ec\U0001f1f3|\U0001f1ec\U0001f1f5|\U0001f1ec\U0001f1f6|\U0001f1ec\U0001f1f7|\U0001f1ec\U0001f1f8|\U0001f1ec\U0001f1f9|\U0001f1ec\U0001f1fa|\U0001f1ec\U0001f1fc|\U0001f1ec\U0001f1fe|\U0001f1ed\U0001f1f0|\U0001f1ed\U0001f1f2|\U0001f1ed\U0001f1f3|\U0001f1ed\U0001f1f7|\U0001f1ed\U0001f1f9|\U0001f1ed\U0001f1fa|\U0001f1ee\U0001f1e8|\U0001f1ee\U0001f1e9|\U0001f1ee\U0001f1ea|\U0001f1ee\U0001f1f1|\U0001f1ee\U0001f1f2|\U0001f1ee\U0001f1f3|\U0001f1ee\U0001f1f4|\U0001f1ee\U0001f1f6|\U0001f1ee\U0001f1f7|\U0001f1ee\U0001f1f8|\U0001f1ee\U0001f1f9|\U0001f1ef\U0001f1ea|\U0001f1ef\U0001f1f2|\U0001f1ef\U0001f1f4|\U0001f1ef\U0001f1f5|\U0001f1f0\U0001f1ea|\U0001f1f0\U0001f1ec|\U0001f1f0\U0001f1ed|\U0001f1f0\U0001f1ee|\U0001f1f0\U0001f1f2|\U0001f1f0\U0001f1f3|\U0001f1f0\U0001f1f5|\U0001f1f0\U0001f1f7|\U0001f1f0\U0001f1fc|\U0001f1f0\U0001f1fe|\U0001f1f0\U0001f1ff|\U0001f1f1\U0001f1e6|\U0001f1f1\U0001f1e7|\U0001f1f1\U0001f1e8|\U0001f1f1\U0001f1ee|\U0001f1f1\U0001f1f0|\U0001f1f1\U0001f1f7|\U0001f1f1\U0001f1f8|\U0001f1f1\U0001f1f9|\U0001f1f1\U0001f1fa|\U0001f1f1\U0001f1fb|\U0001f1f1\U0001f1fe|\U0001f1f2\U0001f1e6|\U0001f1f2\U0001f1e8|\U0001f1f2\U0001f1e9|\U0001f1f2\U0001f1ea|\U0001f1f2\U0001f1eb|\U0001f1f2\U0001f1ec|\U0001f1f2\U0001f1ed|\U0001f1f2\U0001f1f0|\U0001f1f2\U0001f1f1|\U0001f1f2\U0001f1f2|\U0001f1f2\U0001f1f3|\U0001f1f2\U0001f1f4|\U0001f1f2\U0001f1f5|\U0001f1f2\U0001f1f6|\U0001f1f2\U0001f1f7|\U0001f1f2\U0001f1f8|\U0001f1f2\U0001f1f9|\U0001f1f2\U0001f1fa|\U0001f1f2\U0001f1fb|\U0001f1f2\U0001f1fc|\U0001f1f2\U0001f1fd|\U0001f1f2\U0001f1fe|\U0001f1f2\U0001f1ff|\U0001f1f3\U0001f1e6|\U0001f1f3\U0001f1e8|\U0001f1f3\U0001f1ea|\U0001f1f3\U0001f1eb|\U0001f1f3\U0001f1ec|\U0001f1f3\U0001f1ee|\U0001f1f3\U0001f1f1|\U0001f1f3\U0001f1f4|\U0001f1f3\U0001f1f5|\U0001f1f3\U0001f1f7|\U0001f1f3\U0001f1fa|\U0001f1f3\U0001f1ff|\U0001f1f4\U0001f1f2|\U0001f1f5\U0001f1e6|\U0001f1f5\U0001f1ea|\U0001f1f5\U0001f1eb|\U0001f1f5\U0001f1ec|\U0001f1f5\U0001f1ed|\U0001f1f5\U0001f1f0|\U0001f1f5\U0001f1f1|\U0001f1f5\U0001f1f2|\U0001f1f5\U0001f1f3|\U0001f1f5\U0001f1f7|\U0001f1f5\U0001f1f8|\U0001f1f5\U0001f1f9|\U0001f1f5\U0001f1fc|\U0001f1f5\U0001f1fe|\U0001f1f6\U0001f1e6|\U0001f1f7\U0001f1ea|\U0001f1f7\U0001f1f4|\U0001f1f7\U0001f1f8|\U0001f1f7\U0001f1fa|\U0001f1f7\U0001f1fc|\U0001f1f8\U0001f1e6|\U0001f1f8\U0001f1e7|\U0001f1f8\U0001f1e8|\U0001f1f8\U0001f1e9|\U0001f1f8\U0001f1ea|\U0001f1f8\U0001f1ec|\U0001f1f8\U0001f1ed|\U0001f1f8\U0001f1ee|\U0001f1f8\U0001f1ef|\U0001f1f8\U0001f1f0|\U0001f1f8\U0001f1f1|\U0001f1f8\U0001f1f2|\U0001f1f8\U0001f1f3|\U0001f1f8\U0001f1f4|\U0001f1f8\U0001f1f7|\U0001f1f8\U0001f1f8|\U0001f1f8\U0001f1f9|\U0001f1f8\U0001f1fb|\U0001f1f8\U0001f1fd|\U0001f1f8\U0001f1fe|\U0001f1f8\U0001f1ff|\U0001f1f9\U0001f1e6|\U0001f1f9\U0001f1e8|\U0001f1f9\U0001f1e9|\U0001f1f9\U0001f1eb|\U0001f1f9\U0001f1ec|\U0001f1f9\U0001f1ed|\U0001f1f9\U0001f1ef|\U0001f1f9\U0001f1f0|\U0001f1f9\U0001f1f1|\U0001f1f9\U0001f1f2|\U0001f1f9\U0001f1f3|\U0001f1f9\U0001f1f4|\U0001f1f9\U0001f1f7|\U0001f1f9\U0001f1f9|\U0001f1f9\U0001f1fb|\U0001f1f9\U0001f1fc|\U0001f1f9\U0001f1ff|\U0001f1fa\U0001f1e6|\U0001f1fa\U0001f1ec|\U0001f1fa\U0001f1f2|\U0001f1fa\U0001f1f3|\U0001f1fa\U0001f1f8|\U0001f1fa\U0001f1fe|\U0001f1fa\U0001f1ff|\U0001f1fb\U0001f1e6|\U0001f1fb\U0001f1e8|\U0001f1fb\U0001f1ea|\U0001f1fb\U0001f1ec|\U0001f1fb\U0001f1ee|\U0001f1fb\U0001f1f3|\U0001f1fb\U0001f1fa|\U0001f1fc\U0001f1eb|\U0001f1fc\U0001f1f8|\U0001f1fd\U0001f1f0|\U0001f1fe\U0001f1ea|\U0001f1fe\U0001f1f9|\U0001f1ff\U0001f1e6|\U0001f1ff\U0001f1f2|\U0001f1ff\U0001f1fc|\U0001f600|\U0001f603|\U0001f604|\U0001f601|\U0001f606|\U0001f605|\U0001f923|\U0001f602|\U0001f642|\U0001f643|\U0001f609|\U0001f60a|\U0001f607|\U0001f970|\U0001f60d|\U0001f929|\U0001f618|\U0001f617|\u263a|\U0001f61a|\U0001f619|\U0001f60b|\U0001f61b|\U0001f61c|\U0001f92a|\U0001f61d|\U0001f911|\U0001f917|\U0001f92d|\U0001f92b|\U0001f914|\U0001f910|\U0001f928|\U0001f610|\U0001f611|\U0001f636|\U0001f60f|\U0001f612|\U0001f644|\U0001f62c|\U0001f925|\U0001f60c|\U0001f614|\U0001f62a|\U0001f924|\U0001f634|\U0001f637|\U0001f912|\U0001f915|\U0001f922|\U0001f92e|\U0001f927|\U0001f975|\U0001f976|\U0001f974|\U0001f635|\U0001f92f|\U0001f920|\U0001f973|\U0001f60e|\U0001f913|\U0001f9d0|\U0001f615|\U0001f61f|\U0001f641|\u2639|\U0001f62e|\U0001f62f|\U0001f632|\U0001f633|\U0001f97a|\U0001f626|\U0001f627|\U0001f628|\U0001f630|\U0001f625|\U0001f622|\U0001f62d|\U0001f631|\U0001f616|\U0001f623|\U0001f61e|\U0001f613|\U0001f629|\U0001f62b|\U0001f971|\U0001f624|\U0001f621|\U0001f620|\U0001f92c|\U0001f608|\U0001f47f|\U0001f480|\u2620|\U0001f4a9|\U0001f921|\U0001f479|\U0001f47a|\U0001f47b|\U0001f47d|\U0001f47e|\U0001f916|\U0001f63a|\U0001f638|\U0001f639|\U0001f63b|\U0001f63c|\U0001f63d|\U0001f640|\U0001f63f|\U0001f63e|\U0001f648|\U0001f649|\U0001f64a|\U0001f48b|\U0001f48c|\U0001f498|\U0001f49d|\U0001f496|\U0001f497|\U0001f493|\U0001f49e|\U0001f495|\U0001f49f|\u2763|\U0001f494|\u2764|\U0001f9e1|\U0001f49b|\U0001f49a|\U0001f499|\U0001f49c|\U0001f90e|\U0001f5a4|\U0001f90d|\U0001f4af|\U0001f4a2|\U0001f4a5|\U0001f4ab|\U0001f4a6|\U0001f4a8|\U0001f573|\U0001f4a3|\U0001f4ac|\U0001f5e8|\U0001f5ef|\U0001f4ad|\U0001f4a4|\U0001f44b|\U0001f91a|\U0001f590|\u270b|\U0001f596|\U0001f44c|\U0001f90f|\u270c|\U0001f91e|\U0001f91f|\U0001f918|\U0001f919|\U0001f448|\U0001f449|\U0001f446|\U0001f595|\U0001f447|\u261d|\U0001f44d|\U0001f44e|\u270a|\U0001f44a|\U0001f91b|\U0001f91c|\U0001f44f|\U0001f64c|\U0001f450|\U0001f932|\U0001f91d|\U0001f64f|\u270d|\U0001f485|\U0001f933|\U0001f4aa|\U0001f9be|\U0001f9bf|\U0001f9b5|\U0001f9b6|\U0001f442|\U0001f9bb|\U0001f443|\U0001f9e0|\U0001f9b7|\U0001f9b4|\U0001f440|\U0001f441|\U0001f445|\U0001f444|\U0001f476|\U0001f9d2|\U0001f466|\U0001f467|\U0001f9d1|\U0001f471|\U0001f468|\U0001f9d4|\U0001f469|\U0001f9d3|\U0001f474|\U0001f475|\U0001f64d|\U0001f64e|\U0001f645|\U0001f646|\U0001f481|\U0001f64b|\U0001f9cf|\U0001f647|\U0001f926|\U0001f937|\U0001f46e|\U0001f575|\U0001f482|\U0001f477|\U0001f934|\U0001f478|\U0001f473|\U0001f472|\U0001f9d5|\U0001f935|\U0001f470|\U0001f930|\U0001f931|\U0001f47c|\U0001f385|\U0001f936|\U0001f9b8|\U0001f9b9|\U0001f9d9|\U0001f9da|\U0001f9db|\U0001f9dc|\U0001f9dd|\U0001f9de|\U0001f9df|\U0001f486|\U0001f487|\U0001f6b6|\U0001f9cd|\U0001f9ce|\U0001f3c3|\U0001f483|\U0001f57a|\U0001f574|\U0001f46f|\U0001f9d6|\U0001f9d7|\U0001f93a|\U0001f3c7|\u26f7|\U0001f3c2|\U0001f3cc|\U0001f3c4|\U0001f6a3|\U0001f3ca|\u26f9|\U0001f3cb|\U0001f6b4|\U0001f6b5|\U0001f938|\U0001f93c|\U0001f93d|\U0001f93e|\U0001f939|\U0001f9d8|\U0001f6c0|\U0001f6cc|\U0001f46d|\U0001f46b|\U0001f46c|\U0001f48f|\U0001f491|\U0001f46a|\U0001f5e3|\U0001f464|\U0001f465|\U0001f463|\U0001f3fb|\U0001f3fc|\U0001f3fd|\U0001f3fe|\U0001f3ff|\U0001f9b0|\U0001f9b1|\U0001f9b3|\U0001f9b2|\U0001f435|\U0001f412|\U0001f98d|\U0001f9a7|\U0001f436|\U0001f415|\U0001f9ae|\U0001f429|\U0001f43a|\U0001f98a|\U0001f99d|\U0001f431|\U0001f408|\U0001f981|\U0001f42f|\U0001f405|\U0001f406|\U0001f434|\U0001f40e|\U0001f984|\U0001f993|\U0001f98c|\U0001f42e|\U0001f402|\U0001f403|\U0001f404|\U0001f437|\U0001f416|\U0001f417|\U0001f43d|\U0001f40f|\U0001f411|\U0001f410|\U0001f42a|\U0001f42b|\U0001f999|\U0001f992|\U0001f418|\U0001f98f|\U0001f99b|\U0001f42d|\U0001f401|\U0001f400|\U0001f439|\U0001f430|\U0001f407|\U0001f43f|\U0001f994|\U0001f987|\U0001f43b|\U0001f428|\U0001f43c|\U0001f9a5|\U0001f9a6|\U0001f9a8|\U0001f998|\U0001f9a1|\U0001f43e|\U0001f983|\U0001f414|\U0001f413|\U0001f423|\U0001f424|\U0001f425|\U0001f426|\U0001f427|\U0001f54a|\U0001f985|\U0001f986|\U0001f9a2|\U0001f989|\U0001f9a9|\U0001f99a|\U0001f99c|\U0001f438|\U0001f40a|\U0001f422|\U0001f98e|\U0001f40d|\U0001f432|\U0001f409|\U0001f995|\U0001f996|\U0001f433|\U0001f40b|\U0001f42c|\U0001f41f|\U0001f420|\U0001f421|\U0001f988|\U0001f419|\U0001f41a|\U0001f40c|\U0001f98b|\U0001f41b|\U0001f41c|\U0001f41d|\U0001f41e|\U0001f997|\U0001f577|\U0001f578|\U0001f982|\U0001f99f|\U0001f9a0|\U0001f490|\U0001f338|\U0001f4ae|\U0001f3f5|\U0001f339|\U0001f940|\U0001f33a|\U0001f33b|\U0001f33c|\U0001f337|\U0001f331|\U0001f332|\U0001f333|\U0001f334|\U0001f335|\U0001f33e|\U0001f33f|\u2618|\U0001f340|\U0001f341|\U0001f342|\U0001f343|\U0001f347|\U0001f348|\U0001f349|\U0001f34a|\U0001f34b|\U0001f34c|\U0001f34d|\U0001f96d|\U0001f34e|\U0001f34f|\U0001f350|\U0001f351|\U0001f352|\U0001f353|\U0001f95d|\U0001f345|\U0001f965|\U0001f951|\U0001f346|\U0001f954|\U0001f955|\U0001f33d|\U0001f336|\U0001f952|\U0001f96c|\U0001f966|\U0001f9c4|\U0001f9c5|\U0001f344|\U0001f95c|\U0001f330|\U0001f35e|\U0001f950|\U0001f956|\U0001f968|\U0001f96f|\U0001f95e|\U0001f9c7|\U0001f9c0|\U0001f356|\U0001f357|\U0001f969|\U0001f953|\U0001f354|\U0001f35f|\U0001f355|\U0001f32d|\U0001f96a|\U0001f32e|\U0001f32f|\U0001f959|\U0001f9c6|\U0001f95a|\U0001f373|\U0001f958|\U0001f372|\U0001f963|\U0001f957|\U0001f37f|\U0001f9c8|\U0001f9c2|\U0001f96b|\U0001f371|\U0001f358|\U0001f359|\U0001f35a|\U0001f35b|\U0001f35c|\U0001f35d|\U0001f360|\U0001f362|\U0001f363|\U0001f364|\U0001f365|\U0001f96e|\U0001f361|\U0001f95f|\U0001f960|\U0001f961|\U0001f980|\U0001f99e|\U0001f990|\U0001f991|\U0001f9aa|\U0001f366|\U0001f367|\U0001f368|\U0001f369|\U0001f36a|\U0001f382|\U0001f370|\U0001f9c1|\U0001f967|\U0001f36b|\U0001f36c|\U0001f36d|\U0001f36e|\U0001f36f|\U0001f37c|\U0001f95b|\u2615|\U0001f375|\U0001f376|\U0001f37e|\U0001f377|\U0001f378|\U0001f379|\U0001f37a|\U0001f37b|\U0001f942|\U0001f943|\U0001f964|\U0001f9c3|\U0001f9c9|\U0001f9ca|\U0001f962|\U0001f37d|\U0001f374|\U0001f944|\U0001f52a|\U0001f3fa|\U0001f30d|\U0001f30e|\U0001f30f|\U0001f310|\U0001f5fa|\U0001f5fe|\U0001f9ed|\U0001f3d4|\u26f0|\U0001f30b|\U0001f5fb|\U0001f3d5|\U0001f3d6|\U0001f3dc|\U0001f3dd|\U0001f3de|\U0001f3df|\U0001f3db|\U0001f3d7|\U0001f9f1|\U0001f3d8|\U0001f3da|\U0001f3e0|\U0001f3e1|\U0001f3e2|\U0001f3e3|\U0001f3e4|\U0001f3e5|\U0001f3e6|\U0001f3e8|\U0001f3e9|\U0001f3ea|\U0001f3eb|\U0001f3ec|\U0001f3ed|\U0001f3ef|\U0001f3f0|\U0001f492|\U0001f5fc|\U0001f5fd|\u26ea|\U0001f54c|\U0001f6d5|\U0001f54d|\u26e9|\U0001f54b|\u26f2|\u26fa|\U0001f301|\U0001f303|\U0001f3d9|\U0001f304|\U0001f305|\U0001f306|\U0001f307|\U0001f309|\u2668|\U0001f3a0|\U0001f3a1|\U0001f3a2|\U0001f488|\U0001f3aa|\U0001f682|\U0001f683|\U0001f684|\U0001f685|\U0001f686|\U0001f687|\U0001f688|\U0001f689|\U0001f68a|\U0001f69d|\U0001f69e|\U0001f68b|\U0001f68c|\U0001f68d|\U0001f68e|\U0001f690|\U0001f691|\U0001f692|\U0001f693|\U0001f694|\U0001f695|\U0001f696|\U0001f697|\U0001f698|\U0001f699|\U0001f69a|\U0001f69b|\U0001f69c|\U0001f3ce|\U0001f3cd|\U0001f6f5|\U0001f9bd|\U0001f9bc|\U0001f6fa|\U0001f6b2|\U0001f6f4|\U0001f6f9|\U0001f68f|\U0001f6e3|\U0001f6e4|\U0001f6e2|\u26fd|\U0001f6a8|\U0001f6a5|\U0001f6a6|\U0001f6d1|\U0001f6a7|\u2693|\u26f5|\U0001f6f6|\U0001f6a4|\U0001f6f3|\u26f4|\U0001f6e5|\U0001f6a2|\u2708|\U0001f6e9|\U0001f6eb|\U0001f6ec|\U0001fa82|\U0001f4ba|\U0001f681|\U0001f69f|\U0001f6a0|\U0001f6a1|\U0001f6f0|\U0001f680|\U0001f6f8|\U0001f6ce|\U0001f9f3|\u231b|\u23f3|\u231a|\u23f0|\u23f1|\u23f2|\U0001f570|\U0001f55b|\U0001f567|\U0001f550|\U0001f55c|\U0001f551|\U0001f55d|\U0001f552|\U0001f55e|\U0001f553|\U0001f55f|\U0001f554|\U0001f560|\U0001f555|\U0001f561|\U0001f556|\U0001f562|\U0001f557|\U0001f563|\U0001f558|\U0001f564|\U0001f559|\U0001f565|\U0001f55a|\U0001f566|\U0001f311|\U0001f312|\U0001f313|\U0001f314|\U0001f315|\U0001f316|\U0001f317|\U0001f318|\U0001f319|\U0001f31a|\U0001f31b|\U0001f31c|\U0001f321|\u2600|\U0001f31d|\U0001f31e|\U0001fa90|\u2b50|\U0001f31f|\U0001f320|\U0001f30c|\u2601|\u26c5|\u26c8|\U0001f324|\U0001f325|\U0001f326|\U0001f327|\U0001f328|\U0001f329|\U0001f32a|\U0001f32b|\U0001f32c|\U0001f300|\U0001f308|\U0001f302|\u2602|\u2614|\u26f1|\u26a1|\u2744|\u2603|\u26c4|\u2604|\U0001f525|\U0001f4a7|\U0001f30a|\U0001f383|\U0001f384|\U0001f386|\U0001f387|\U0001f9e8|\u2728|\U0001f388|\U0001f389|\U0001f38a|\U0001f38b|\U0001f38d|\U0001f38e|\U0001f38f|\U0001f390|\U0001f391|\U0001f9e7|\U0001f380|\U0001f381|\U0001f397|\U0001f39f|\U0001f3ab|\U0001f396|\U0001f3c6|\U0001f3c5|\U0001f947|\U0001f948|\U0001f949|\u26bd|\u26be|\U0001f94e|\U0001f3c0|\U0001f3d0|\U0001f3c8|\U0001f3c9|\U0001f3be|\U0001f94f|\U0001f3b3|\U0001f3cf|\U0001f3d1|\U0001f3d2|\U0001f94d|\U0001f3d3|\U0001f3f8|\U0001f94a|\U0001f94b|\U0001f945|\u26f3|\u26f8|\U0001f3a3|\U0001f93f|\U0001f3bd|\U0001f3bf|\U0001f6f7|\U0001f94c|\U0001f3af|\U0001fa80|\U0001fa81|\U0001f3b1|\U0001f52e|\U0001f9ff|\U0001f3ae|\U0001f579|\U0001f3b0|\U0001f3b2|\U0001f9e9|\U0001f9f8|\u2660|\u2665|\u2666|\u2663|\u265f|\U0001f0cf|\U0001f004|\U0001f3b4|\U0001f3ad|\U0001f5bc|\U0001f3a8|\U0001f9f5|\U0001f9f6|\U0001f453|\U0001f576|\U0001f97d|\U0001f97c|\U0001f9ba|\U0001f454|\U0001f455|\U0001f456|\U0001f9e3|\U0001f9e4|\U0001f9e5|\U0001f9e6|\U0001f457|\U0001f458|\U0001f97b|\U0001fa71|\U0001fa72|\U0001fa73|\U0001f459|\U0001f45a|\U0001f45b|\U0001f45c|\U0001f45d|\U0001f6cd|\U0001f392|\U0001f45e|\U0001f45f|\U0001f97e|\U0001f97f|\U0001f460|\U0001f461|\U0001fa70|\U0001f462|\U0001f451|\U0001f452|\U0001f3a9|\U0001f393|\U0001f9e2|\u26d1|\U0001f4ff|\U0001f484|\U0001f48d|\U0001f48e|\U0001f507|\U0001f508|\U0001f509|\U0001f50a|\U0001f4e2|\U0001f4e3|\U0001f4ef|\U0001f514|\U0001f515|\U0001f3bc|\U0001f3b5|\U0001f3b6|\U0001f399|\U0001f39a|\U0001f39b|\U0001f3a4|\U0001f3a7|\U0001f4fb|\U0001f3b7|\U0001f3b8|\U0001f3b9|\U0001f3ba|\U0001f3bb|\U0001fa95|\U0001f941|\U0001f4f1|\U0001f4f2|\u260e|\U0001f4de|\U0001f4df|\U0001f4e0|\U0001f50b|\U0001f50c|\U0001f4bb|\U0001f5a5|\U0001f5a8|\u2328|\U0001f5b1|\U0001f5b2|\U0001f4bd|\U0001f4be|\U0001f4bf|\U0001f4c0|\U0001f9ee|\U0001f3a5|\U0001f39e|\U0001f4fd|\U0001f3ac|\U0001f4fa|\U0001f4f7|\U0001f4f8|\U0001f4f9|\U0001f4fc|\U0001f50d|\U0001f50e|\U0001f56f|\U0001f4a1|\U0001f526|\U0001f3ee|\U0001fa94|\U0001f4d4|\U0001f4d5|\U0001f4d6|\U0001f4d7|\U0001f4d8|\U0001f4d9|\U0001f4da|\U0001f4d3|\U0001f4d2|\U0001f4c3|\U0001f4dc|\U0001f4c4|\U0001f4f0|\U0001f5de|\U0001f4d1|\U0001f516|\U0001f3f7|\U0001f4b0|\U0001f4b4|\U0001f4b5|\U0001f4b6|\U0001f4b7|\U0001f4b8|\U0001f4b3|\U0001f9fe|\U0001f4b9|\U0001f4b1|\U0001f4b2|\u2709|\U0001f4e7|\U0001f4e8|\U0001f4e9|\U0001f4e4|\U0001f4e5|\U0001f4e6|\U0001f4eb|\U0001f4ea|\U0001f4ec|\U0001f4ed|\U0001f4ee|\U0001f5f3|\u270f|\u2712|\U0001f58b|\U0001f58a|\U0001f58c|\U0001f58d|\U0001f4dd|\U0001f4bc|\U0001f4c1|\U0001f4c2|\U0001f5c2|\U0001f4c5|\U0001f4c6|\U0001f5d2|\U0001f5d3|\U0001f4c7|\U0001f4c8|\U0001f4c9|\U0001f4ca|\U0001f4cb|\U0001f4cc|\U0001f4cd|\U0001f4ce|\U0001f587|\U0001f4cf|\U0001f4d0|\u2702|\U0001f5c3|\U0001f5c4|\U0001f5d1|\U0001f512|\U0001f513|\U0001f50f|\U0001f510|\U0001f511|\U0001f5dd|\U0001f528|\U0001fa93|\u26cf|\u2692|\U0001f6e0|\U0001f5e1|\u2694|\U0001f52b|\U0001f3f9|\U0001f6e1|\U0001f527|\U0001f529|\u2699|\U0001f5dc|\u2696|\U0001f9af|\U0001f517|\u26d3|\U0001f9f0|\U0001f9f2|\u2697|\U0001f9ea|\U0001f9eb|\U0001f9ec|\U0001f52c|\U0001f52d|\U0001f4e1|\U0001f489|\U0001fa78|\U0001f48a|\U0001fa79|\U0001fa7a|\U0001f6aa|\U0001f6cf|\U0001f6cb|\U0001fa91|\U0001f6bd|\U0001f6bf|\U0001f6c1|\U0001fa92|\U0001f9f4|\U0001f9f7|\U0001f9f9|\U0001f9fa|\U0001f9fb|\U0001f9fc|\U0001f9fd|\U0001f9ef|\U0001f6d2|\U0001f6ac|\u26b0|\u26b1|\U0001f5ff|\U0001f3e7|\U0001f6ae|\U0001f6b0|\u267f|\U0001f6b9|\U0001f6ba|\U0001f6bb|\U0001f6bc|\U0001f6be|\U0001f6c2|\U0001f6c3|\U0001f6c4|\U0001f6c5|\u26a0|\U0001f6b8|\u26d4|\U0001f6ab|\U0001f6b3|\U0001f6ad|\U0001f6af|\U0001f6b1|\U0001f6b7|\U0001f4f5|\U0001f51e|\u2622|\u2623|\u2b06|\u2197|\u27a1|\u2198|\u2b07|\u2199|\u2b05|\u2196|\u2195|\u2194|\u21a9|\u21aa|\u2934|\u2935|\U0001f503|\U0001f504|\U0001f519|\U0001f51a|\U0001f51b|\U0001f51c|\U0001f51d|\U0001f6d0|\u269b|\U0001f549|\u2721|\u2638|\u262f|\u271d|\u2626|\u262a|\u262e|\U0001f54e|\U0001f52f|\u2648|\u2649|\u264a|\u264b|\u264c|\u264d|\u264e|\u264f|\u2650|\u2651|\u2652|\u2653|\u26ce|\U0001f500|\U0001f501|\U0001f502|\u25b6|\u23e9|\u23ed|\u23ef|\u25c0|\u23ea|\u23ee|\U0001f53c|\u23eb|\U0001f53d|\u23ec|\u23f8|\u23f9|\u23fa|\u23cf|\U0001f3a6|\U0001f505|\U0001f506|\U0001f4f6|\U0001f4f3|\U0001f4f4|\u2640|\u2642|\u2695|\u267e|\u267b|\u269c|\U0001f531|\U0001f4db|\U0001f530|\u2b55|\u2705|\u2611|\u2714|\u2716|\u274c|\u274e|\u2795|\u2796|\u2797|\u27b0|\u27bf|\u303d|\u2733|\u2734|\u2747|\u203c|\u2049|\u2753|\u2754|\u2755|\u2757|\u3030|\xa9|\xae|\u2122|\U0001f51f|\U0001f520|\U0001f521|\U0001f522|\U0001f523|\U0001f524|\U0001f170|\U0001f18e|\U0001f171|\U0001f191|\U0001f192|\U0001f193|\u2139|\U0001f194|\u24c2|\U0001f195|\U0001f196|\U0001f17e|\U0001f197|\U0001f17f|\U0001f198|\U0001f199|\U0001f19a|\U0001f201|\U0001f202|\U0001f237|\U0001f236|\U0001f22f|\U0001f250|\U0001f239|\U0001f21a|\U0001f232|\U0001f251|\U0001f238|\U0001f234|\U0001f233|\u3297|\u3299|\U0001f23a|\U0001f235|\U0001f534|\U0001f7e0|\U0001f7e1|\U0001f7e2|\U0001f535|\U0001f7e3|\U0001f7e4|\u26ab|\u26aa|\U0001f7e5|\U0001f7e7|\U0001f7e8|\U0001f7e9|\U0001f7e6|\U0001f7ea|\U0001f7eb|\u2b1b|\u2b1c|\u25fc|\u25fb|\u25fe|\u25fd|\u25aa|\u25ab|\U0001f536|\U0001f537|\U0001f538|\U0001f539|\U0001f53a|\U0001f53b|\U0001f4a0|\U0001f518|\U0001f533|\U0001f532|\U0001f3c1|\U0001f6a9|\U0001f38c|\U0001f3f4|\U0001f3f3')) -> str: """Return the string obtained by replacing all emojis in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all emojis in text with 'repl' Reference: akkez/emoji.py: Python emoji regexp / python emoji detection https://gist.github.com/akkez/99ceeae2f13c9d8d9be7df0279e2c438 """ text = regex.sub(repl, text) return text def _email_normalize(self, text: str, repl: str, regex=re.compile(r'[a-zA-Z0-9.!#$%&\'*+/=?^_`{|}~-]+@[a-zA-Z0-9-]+\.[a-zA-Z0-9.]+')) -> str: """Return the string obtained by replacing all email addresses in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all email addresses in text with 'repl' """ text = regex.sub(repl, text) return text def _tel_normalize(self, text: str, repl: str, regex=re.compile(r'[()+\d.\-]*[ ]?\d{2,4}[-. ]+\d{3,4}[-. ]+\d{3,4}')) -> str: """Return the string obtained by replacing all phone numbers in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all phone numbers in text with 'repl' """ text = regex.sub(repl, text) return text
prenlp/data/normalizer.py
91,819
Normalizer return the text replaced with 'repl'. If 'repl' is None, normalization is not applied to the pattern corresponding to 'repl'. Args: url_repl (str): replace all urls in text with this tag_repl (str): replace all tags in text with this emoji_repl (str): replace all emojis in text with this email_repl (str): replace all emails in text with this tel_repl (str): replace all tels in text with this Return the string obtained by replacing all email addresses in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all email addresses in text with 'repl' Return the string obtained by replacing all emojis in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all emojis in text with 'repl' Reference: akkez/emoji.py: Python emoji regexp / python emoji detection https://gist.github.com/akkez/99ceeae2f13c9d8d9be7df0279e2c438 Initialize normalize function. If 'repl' is None, normalization is not applied to the pattern corresponding to 'repl'. Return the string obtained by replacing all HTML tags in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all HTML tags in text with 'repl' Return the string obtained by replacing all phone numbers in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all phone numbers in text with 'repl' Return the string obtained by replacing all urls in 'text' by the replacement 'repl'. Args: text (str): text to be replaced repl (str): replace all urls in text with 'repl' Normalize text. Args: text (str): text to be normalized repls
1,732
en
0.620159
"""BERT Training Script.""" import functools from typing import Any, Callable, Dict, Tuple, Optional, Type from absl import logging from clu import metric_writers from clu import periodic_actions from flax import jax_utils import flax.linen as nn import jax from jax.experimental import optimizers as jax_optimizers import jax.numpy as jnp import jax.profiler import ml_collections import numpy as np from scenic.dataset_lib import dataset_utils from scenic.projects.baselines.bert import bert_base_model from scenic.projects.baselines.bert import train_utils as bert_train_utils from scenic.train_lib import lr_schedules from scenic.train_lib import optimizers from scenic.train_lib import pretrain_utils from scenic.train_lib import train_utils def train_step( *, flax_model: nn.Module, train_state: train_utils.TrainState, batch: bert_base_model.Batch, learning_rate_fn: Callable[[int], float], loss_fn: bert_base_model.LossFn, metrics_fn: bert_base_model.MetricFn, config: ml_collections.ConfigDict, debug: Optional[bool] = False ) -> Tuple[train_utils.TrainState, Dict[str, Tuple[float, int]], float]: """Runs a single step of training. Given the state of the training and a batch of data, computes the loss and updates the parameters of the model. Note that in this code, the buffers of the first (train_state) and second (batch) arguments are donated to the computation. Args: flax_model: A Flax model. train_state: The state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data. The buffer of this argument can be donated to the computation. learning_rate_fn: Learning rate scheduler which given the global_step generates the learning rate. loss_fn: A loss function that given logits, a batch, and parameters of the model calculates the loss. metrics_fn: A metrics function that given logits and batch of data, calculates the metrics as well as the loss. config: Configurations of the experiment. debug: Whether the debug mode is enabled during training. `debug=True` enables model specific logging/storing some values using jax.host_callback. Returns: Updated state of training, computed metrics, and learning rate for logging. """ new_rng, rng = jax.random.split(train_state.rng) # Bind the rng to the host/device we are on. dropout_rng = train_utils.bind_rng_to_host_device( rng, axis_name='batch', bind_to='device') def training_loss_fn(params): variables = {'params': params, **train_state.model_state} output, new_model_state = flax_model.apply( variables, batch, mutable=['batch_stats'], train=True, rngs={'dropout': dropout_rng}, debug=debug) loss = loss_fn(output, batch, variables['params']) return loss, (new_model_state, output) compute_gradient_fn = jax.value_and_grad(training_loss_fn, has_aux=True) step = train_state.global_step lr = learning_rate_fn(step) (train_cost, (new_model_state, output)), grad = compute_gradient_fn(train_state.optimizer.target) del train_cost # We clip gradients before pmean in BERT. if config.get('max_grad_norm', None) is not None: grad = jax_optimizers.clip_grads(grad, config.max_grad_norm) # Re-use same axis_name as in the call to `pmap(...train_step...)` below. grad = jax.lax.pmean(grad, axis_name='batch') new_optimizer = train_state.optimizer.apply_gradient(grad, learning_rate=lr) # Explicit weight decay, if necessary. if config.get('explicit_weight_decay', None) is not None: new_optimizer = new_optimizer.replace( target=optimizers.tree_map_with_names( functools.partial( optimizers.decay_weight_fn, lr=lr, decay=config.explicit_weight_decay), new_optimizer.target, match_name_fn=lambda name: 'kernel' in name)) metrics = metrics_fn(output, batch) new_train_state = train_state.replace( # pytype: disable=attribute-error global_step=step + 1, optimizer=new_optimizer, model_state=new_model_state, rng=new_rng) return new_train_state, metrics, lr def eval_step( *, flax_model: nn.Module, train_state: train_utils.TrainState, batch: bert_base_model.Batch, metrics_fn: bert_base_model.MetricFn, all_gather: bool = False, debug: Optional[bool] = False ) -> Tuple[Dict[str, Tuple[float, int]], Optional[jnp.ndarray], Optional[jnp.ndarray]]: """Runs a single step of training. Note that in this code, the buffer of the second argument (batch) is donated to the computation. Assumed API of metrics_fn is: ```metrics = metrics_fn(logits, batch) where batch is yielded by the batch iterator, and metrics is a dictionary mapping metric name to a vector of per example measurements. eval_step will aggregate (by summing) all per example measurements and divide by the aggregated normalizers. For each given metric we compute: 1/N sum_{b in batch_iter} metric(b), where N is the sum of normalizer over all batches. Args: flax_model: A Flax model. train_state: TrainState, the state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data. a metrics function, that given logits and batch of data, calculates the metrics as well as the loss. metrics_fn: A metrics function, that given logits and batch of data, calculates the metrics as well as the loss. all_gather: If True, the function gather batch and output of model in from all hosts, using `jax.lax.all_gather` and return it, e.g., for computing global metrics on CPU. debug: Whether the debug mode is enabled during evaluation. `debug=True` enables model specific logging/storing some values using jax.host_callback. Returns: Calculated metrics and optionally output, and batch after all_gather. """ variables = { 'params': train_state.optimizer.target, **train_state.model_state } output = flax_model.apply( variables, batch, train=False, mutable=False, debug=debug) metrics = metrics_fn(output, batch) if all_gather: output = jax.lax.all_gather(output, 'batch') batch = jax.lax.all_gather(batch, 'batch') return metrics, output, batch else: return metrics, None, None def representation_fn( *, flax_model: nn.Module, train_state: train_utils.TrainState, batch: bert_base_model.Batch, representation_layer: str, gather_to_host: bool = True ) -> Tuple[jnp.ndarray, jnp.ndarray, jnp.ndarray]: """Feeds the inputs to the model and returns their representations. Args: flax_model: A Flax model. train_state: TrainState, the state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data from the dataset. representation_layer: The name of the layer to use as the representation. gather_to_host: Whether to gather results from all devices to the host, rather than leaving them distributed. Returns: Representation learned by the model for the given inputs and the labels and masks. If `gather_to_host` is True, these are collected from all hosts. """ variables = { 'params': train_state.optimizer.target, **train_state.model_state } representation_layer_parts = representation_layer.split('/') filter_rep = lambda mdl, _: mdl.name == representation_layer_parts[-1] _, model_state = flax_model.apply( variables, batch, train=False, capture_intermediates=filter_rep, mutable=['intermediates'], transfer_mode=True, debug=False) if 'intermediates' not in model_state: raise ValueError(f'Layer with name "{representation_layer}"' ' does not exist in your model.') representation = model_state['intermediates'] for rep_layer in representation_layer_parts: if rep_layer: representation = representation[rep_layer] representation = representation['__call__'][0] if gather_to_host: representation = jax.lax.all_gather(representation, 'batch') batch = jax.lax.all_gather(batch, 'batch') return representation, batch['label'], batch['batch_mask'] def train( *, rng: jnp.ndarray, config: ml_collections.ConfigDict, model_cls: Type[bert_base_model.BERTBaseModel], dataset: dataset_utils.Dataset, workdir: str, writer: metric_writers.MetricWriter, ) -> Tuple[train_utils.TrainState, Dict[str, Any], Dict[str, Any]]: """Main training loop lives in this function. Given the model class and dataset, it prepares the items needed to run the training, including the TrainState. Args: rng: Jax rng key. config: Configurations of the experiment. model_cls: Model class; A model has a flax_module, a loss_fn, and a metrics_fn associated with it. dataset: The dataset that has train_iter, eval_iter, meta_data, and optionally, test_iter. workdir: Directory for checkpointing. writer: CLU metrics writer instance. Returns: train_state that has the state of training (including current global_step, model_state, rng, and the optimizer), train_summary and eval_summary which are dict of metrics. These outputs are used for regression testing. """ lead_host = jax.process_index() == 0 # Build the loss_fn, metrics, and flax_model. model = model_cls(config, dataset.meta_data) # Initialize model. rng, init_rng = jax.random.split(rng) (params, model_state, num_trainable_params, gflops) = bert_train_utils.initialize_bert_model( model_def=model.flax_model, input_spec=dataset.meta_data['input_spec'], config=config, rngs=init_rng) # Create optimizer. # We jit this, such that the arrays that are created are created on the same # device as the input is, in this case the CPU. Else they'd be on device[0]. optimizer = jax.jit( optimizers.get_optimizer(config).create, backend='cpu')( params) rng, train_rng = jax.random.split(rng) train_state = train_utils.TrainState( global_step=0, optimizer=optimizer, model_state=model_state, rng=train_rng, accum_train_time=0) start_step = train_state.global_step if config.checkpoint: train_state, start_step = train_utils.restore_checkpoint( workdir, train_state) if (start_step == 0 # Which means "no" checkpoint is restored! and config.get('init_from') is not None): restored_model_cfg = config.init_from.get('model_config') init_checkpoint_path = config.init_from.get('checkpoint_path') restored_train_state = pretrain_utils.restore_pretrained_checkpoint( init_checkpoint_path, train_state, assert_exist=True) # Load params from the init_model. train_state = model.init_from_train_state( # pytype: disable=attribute-error train_state, restored_train_state, restored_model_cfg) del restored_train_state # Replicate the optimzier, state, and rng. train_state = jax_utils.replicate(train_state) del params # Do not keep a copy of the initial params. # Calculate the total number of training steps. total_steps, steps_per_epoch = train_utils.get_num_training_steps( config, dataset.meta_data) # Get learning rate scheduler. learning_rate_fn = lr_schedules.get_learning_rate_fn(config) train_step_pmapped = jax.pmap( functools.partial( train_step, flax_model=model.flax_model, learning_rate_fn=learning_rate_fn, loss_fn=model.loss_function, metrics_fn=model.get_metrics_fn('train'), config=config, debug=config.debug_train), axis_name='batch', # We can donate both buffers of train_state and train_batch. donate_argnums=(0, 1), ) eval_step_pmapped = jax.pmap( functools.partial( eval_step, flax_model=model.flax_model, metrics_fn=model.get_metrics_fn('validation'), all_gather=config.get('global_metrics', False), debug=config.debug_eval), axis_name='batch', # We can donate the eval_batch's buffer. donate_argnums=(1,), ) if 'fewshot' in config: representation_fn_pmaped = jax.pmap( functools.partial( representation_fn, flax_model=model.flax_model, representation_layer=config.fewshot.representation_layer), # We can donate the batch's buffer. donate_argnums=(1,), axis_name='batch') fewshotter = bert_train_utils.BERTFewShotEvaluator(representation_fn_pmaped, config.fewshot) log_eval_steps = config.get('log_eval_steps') or steps_per_epoch if not log_eval_steps: raise ValueError("'log_eval_steps' should be specified in the config.") checkpoint_steps = config.get('checkpoint_steps') or log_eval_steps log_summary_steps = config.get('log_summary_steps') or log_eval_steps # Ceil rounding such that we include the last incomplete batch. total_eval_steps = int( np.ceil(dataset.meta_data['num_eval_examples'] / config.batch_size)) steps_per_eval = config.get('steps_per_eval') or total_eval_steps # If `global_metrics` are set in the config and we are the the lead host compute_global_metrics = False if config.get('global_metrics', False) and lead_host: compute_global_metrics = True if compute_global_metrics: global_metrics_evaluator = bert_train_utils.BERTGlobalEvaluator( config.global_metrics) train_metrics, extra_training_logs = [], [] train_summary, eval_summary = None, None chrono = train_utils.Chrono( first_step=start_step, total_steps=total_steps, steps_per_epoch=steps_per_epoch, global_bs=config.batch_size, accum_train_time=int(jax_utils.unreplicate(train_state.accum_train_time)), example_type='example') logging.info('Starting training loop at step %d.', start_step + 1) report_progress = periodic_actions.ReportProgress( num_train_steps=total_steps, writer=writer) hooks = [report_progress] if config.get('xprof', True) and lead_host: hooks.append(periodic_actions.Profile(num_profile_steps=5, logdir=workdir)) if start_step == 0: step0_log = {'num_trainable_params': num_trainable_params} if gflops: step0_log['gflops'] = gflops writer.write_scalars(1, step0_log) for step in range(start_step + 1, total_steps + 1): with jax.profiler.StepTraceContext('train', step_num=step): train_batch = next(dataset.train_iter) train_state, t_metrics, lr = train_step_pmapped( train_state=train_state, batch=train_batch) # This will accumulate metrics in TPU memory up to the point that we log # them. This is no problem for small metrics but may be a problem for # large (e.g. segmentation) metrics. An alternative is to set # `log_summary_steps` to a small number, or to use # `train_utils.unreplicate_and_get` here instead of right before writing # summaries, but that means in each step, we have data transfer between # tpu and host, which might slow down the training. train_metrics.append(t_metrics) # Additional training logs: learning rate: extra_training_logs.append({'learning_rate': lr}) for h in hooks: h(step) chrono.pause() # Below are once-in-a-while ops -> pause. ###################### LOG TRAIN SUMMARY ######################## if (step % log_summary_steps == 1) or (step == total_steps): if lead_host: chrono.tick(step, writer=writer) # train_metrics is list of a dictionaries of metrics, where the shape of # the metrics[key] is [n_local_devices]. However, because metric functions # have a psum, we have already summed across the whole sharded batch, and # what's returned is n_local_devices copies of the same summed metric. # So we do unreplicate and fetch them to host using `unreplicate_and_get`. train_summary = train_utils.log_train_summary( step=step, train_metrics=jax.tree_map(train_utils.unreplicate_and_get, train_metrics), extra_training_logs=jax.tree_map(train_utils.unreplicate_and_get, extra_training_logs), writer=writer) # Reset metric accumulation for next evaluation cycle. train_metrics, extra_training_logs = [], [] ################### EVALUATION ####################### if (step % log_eval_steps == 1) or (step == total_steps): with report_progress.timed('eval'): eval_metrics = [] # Sync model state across replicas. train_state = train_utils.sync_model_state_across_replicas( train_state) for _ in range(steps_per_eval): eval_batch = next(dataset.valid_iter) e_metrics, e_output, e_batch = eval_step_pmapped( train_state=train_state, batch=eval_batch) eval_metrics.append(train_utils.unreplicate_and_get(e_metrics)) if compute_global_metrics: # Unreplicate outputs of eval_step_pmapped that are coming from # `lax.all_gather`, fetch to the host and add to the Evaluator: e_batch_mask = train_utils.unreplicate_and_get( e_batch['batch_mask']).astype(bool) # Classification: 'label', regression: 'target' t_key = 'label' if 'label' in e_batch else 'targets' global_metrics_evaluator.add_batch_of_examples( target=train_utils.unreplicate_and_get( e_batch[t_key])[e_batch_mask], output=train_utils.unreplicate_and_get(e_output) [e_batch_mask]) del e_batch, e_output, e_batch_mask eval_global_metrics_summary = None if compute_global_metrics: if (len(global_metrics_evaluator) != dataset.meta_data['num_eval_examples']): # Make sure no example is lost (specially in multi-host setup). raise ValueError(f'Number of eval examples should be ' f'{dataset.meta_data["num_eval_examples"]}, ' f'but it is {len(global_metrics_evaluator)}.') eval_global_metrics_summary = ( global_metrics_evaluator.compute_metrics( clear_annotations=True)) eval_summary = train_utils.log_eval_summary( step=step, eval_metrics=eval_metrics, extra_eval_summary=eval_global_metrics_summary, writer=writer) writer.flush() del eval_metrics, eval_global_metrics_summary ##################### CHECKPOINTING ################### if ((step % checkpoint_steps == 0 and step > 0) or (step == total_steps)) and config.checkpoint: with report_progress.timed('checkpoint'): # Sync model state across replicas. train_state = train_utils.sync_model_state_across_replicas(train_state) if lead_host: train_state.replace( # pytype: disable=attribute-error accum_train_time=chrono.accum_train_time) train_utils.save_checkpoint(workdir, train_state) ##################### FEWSHOT EVALUATION ############################ if 'fewshot' in config: # Compute few-shot on-the-fly evaluation. if (step % config.fewshot.log_eval_steps == 1) or (step == total_steps): with report_progress.timed('fewshot'): results = fewshotter.run_all(train_state, config.fewshot.datasets) fewshotter.log_fewshot_summary( writer=writer, step=step, results=results) del results writer.write_scalars(step, {'zz/epoch': step / steps_per_epoch}) writer.flush() chrono.resume() # un-pause now # Wait until computations are done before exiting. jax.random.normal(jax.random.PRNGKey(0), ()).block_until_ready() # Return the train and eval summary after last step for regresesion testing. return train_state, train_summary, eval_summary
scenic/projects/baselines/bert/trainer.py
20,582
Runs a single step of training. Note that in this code, the buffer of the second argument (batch) is donated to the computation. Assumed API of metrics_fn is: ```metrics = metrics_fn(logits, batch) where batch is yielded by the batch iterator, and metrics is a dictionary mapping metric name to a vector of per example measurements. eval_step will aggregate (by summing) all per example measurements and divide by the aggregated normalizers. For each given metric we compute: 1/N sum_{b in batch_iter} metric(b), where N is the sum of normalizer over all batches. Args: flax_model: A Flax model. train_state: TrainState, the state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data. a metrics function, that given logits and batch of data, calculates the metrics as well as the loss. metrics_fn: A metrics function, that given logits and batch of data, calculates the metrics as well as the loss. all_gather: If True, the function gather batch and output of model in from all hosts, using `jax.lax.all_gather` and return it, e.g., for computing global metrics on CPU. debug: Whether the debug mode is enabled during evaluation. `debug=True` enables model specific logging/storing some values using jax.host_callback. Returns: Calculated metrics and optionally output, and batch after all_gather. Feeds the inputs to the model and returns their representations. Args: flax_model: A Flax model. train_state: TrainState, the state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data from the dataset. representation_layer: The name of the layer to use as the representation. gather_to_host: Whether to gather results from all devices to the host, rather than leaving them distributed. Returns: Representation learned by the model for the given inputs and the labels and masks. If `gather_to_host` is True, these are collected from all hosts. Main training loop lives in this function. Given the model class and dataset, it prepares the items needed to run the training, including the TrainState. Args: rng: Jax rng key. config: Configurations of the experiment. model_cls: Model class; A model has a flax_module, a loss_fn, and a metrics_fn associated with it. dataset: The dataset that has train_iter, eval_iter, meta_data, and optionally, test_iter. workdir: Directory for checkpointing. writer: CLU metrics writer instance. Returns: train_state that has the state of training (including current global_step, model_state, rng, and the optimizer), train_summary and eval_summary which are dict of metrics. These outputs are used for regression testing. Runs a single step of training. Given the state of the training and a batch of data, computes the loss and updates the parameters of the model. Note that in this code, the buffers of the first (train_state) and second (batch) arguments are donated to the computation. Args: flax_model: A Flax model. train_state: The state of training including the current global_step, model_state, rng, and optimizer. The buffer of this argument can be donated to the computation. batch: A single batch of data. The buffer of this argument can be donated to the computation. learning_rate_fn: Learning rate scheduler which given the global_step generates the learning rate. loss_fn: A loss function that given logits, a batch, and parameters of the model calculates the loss. metrics_fn: A metrics function that given logits and batch of data, calculates the metrics as well as the loss. config: Configurations of the experiment. debug: Whether the debug mode is enabled during training. `debug=True` enables model specific logging/storing some values using jax.host_callback. Returns: Updated state of training, computed metrics, and learning rate for logging. BERT Training Script. Bind the rng to the host/device we are on. We clip gradients before pmean in BERT. Re-use same axis_name as in the call to `pmap(...train_step...)` below. Explicit weight decay, if necessary. pytype: disable=attribute-error Build the loss_fn, metrics, and flax_model. Initialize model. Create optimizer. We jit this, such that the arrays that are created are created on the same device as the input is, in this case the CPU. Else they'd be on device[0]. Which means "no" checkpoint is restored! Load params from the init_model. pytype: disable=attribute-error Replicate the optimzier, state, and rng. Do not keep a copy of the initial params. Calculate the total number of training steps. Get learning rate scheduler. We can donate both buffers of train_state and train_batch. We can donate the eval_batch's buffer. We can donate the batch's buffer. Ceil rounding such that we include the last incomplete batch. If `global_metrics` are set in the config and we are the the lead host This will accumulate metrics in TPU memory up to the point that we log them. This is no problem for small metrics but may be a problem for large (e.g. segmentation) metrics. An alternative is to set `log_summary_steps` to a small number, or to use `train_utils.unreplicate_and_get` here instead of right before writing summaries, but that means in each step, we have data transfer between tpu and host, which might slow down the training. Additional training logs: learning rate: Below are once-in-a-while ops -> pause. LOG TRAIN SUMMARY train_metrics is list of a dictionaries of metrics, where the shape of the metrics[key] is [n_local_devices]. However, because metric functions have a psum, we have already summed across the whole sharded batch, and what's returned is n_local_devices copies of the same summed metric. So we do unreplicate and fetch them to host using `unreplicate_and_get`. Reset metric accumulation for next evaluation cycle. EVALUATION Sync model state across replicas. Unreplicate outputs of eval_step_pmapped that are coming from `lax.all_gather`, fetch to the host and add to the Evaluator: Classification: 'label', regression: 'target' Make sure no example is lost (specially in multi-host setup). CHECKPOINTING Sync model state across replicas. pytype: disable=attribute-error FEWSHOT EVALUATION Compute few-shot on-the-fly evaluation. un-pause now Wait until computations are done before exiting. Return the train and eval summary after last step for regresesion testing.
6,623
en
0.887796
__source__ = 'https://leetcode.com/problems/delete-node-in-a-linked-list/description/' # https://github.com/kamyu104/LeetCode/blob/master/Python/delete-node-in-a-linked-list.py # Time: O(1) # Space: O(1) # # Description: Leetcode # 237. Delete Node in a Linked List # # Write a function to delete a node (except the tail) in a singly linked list, # given only access to that node. # # Supposed the linked list is 1 -> 2 -> 3 -> 4 and you are given the third node # with value 3, the linked list should become 1 -> 2 -> 4 after calling your function. # # Companies # Adobe Apple Microsoft # Related Topics # Linked List # Similar Questions # Remove Linked List Elements # # Definition for singly-linked list. class ListNode: def __init__(self, x): self.val = x self.next = None import unittest class Solution: # @param {ListNode} node # @return {void} Do not return anything, modify node in-place instead. def deleteNode(self, node): if node and node.next: node_to_delete = node.next node.val = node_to_delete.val node.next = node_to_delete.next del node_to_delete class TestMethods(unittest.TestCase): def test_Local(self): self.assertEqual(1, 1) if __name__ == '__main__': unittest.main() Java = ''' # Thought: https://leetcode.com/problems/delete-node-in-a-linked-list/solution/ Thought: We can't really delete the node, but we can kinda achieve the same effect by instead removing the next node after copying its data into the node that we were asked to delete. /** * Definition for singly-linked list. * public class ListNode { * int val; * ListNode next; * ListNode(int x) { val = x; } * } */ # 0ms 100% class Solution { public void deleteNode(ListNode node) { node.val = node.next.val; node.next = node.next.next; } } # 0ms 100% class Solution { public void deleteNode(ListNode node) { if (node == null || node.next == null) { return; } while (node.next.next != null) { node.val = node.next.val; node = node.next; } node.val = node.next.val; node.next = null; } } '''
cs15211/DeleteNodrinaLinkedList.py
2,221
https://github.com/kamyu104/LeetCode/blob/master/Python/delete-node-in-a-linked-list.py Time: O(1) Space: O(1) Description: Leetcode 237. Delete Node in a Linked List Write a function to delete a node (except the tail) in a singly linked list, given only access to that node. Supposed the linked list is 1 -> 2 -> 3 -> 4 and you are given the third node with value 3, the linked list should become 1 -> 2 -> 4 after calling your function. Companies Adobe Apple Microsoft Related Topics Linked List Similar Questions Remove Linked List Elements Definition for singly-linked list. @param {ListNode} node @return {void} Do not return anything, modify node in-place instead.
672
en
0.806736
from unittest import TestCase from btcmagic import transaction, convert import os import json class TestTransaction(TestCase): def setUp(self): self.tx_bin = convert.hex_to_bytes( '0100000001637aaf20d708fcff67bb688af6e41d1807e6883f736c50eacb6042bf6e6c829c010000008c493046022100da1e59d78bb88ca7c3e13a4a6f4e259d5dd8cb177d5f79199bf024b1f57121d50221008d1d9838606a62ed4bd011a6ce8a2042ae2dc38fd05381b50aa388a1c8bd9150014104d3b615c609e48ae81389f6617b50473bf4c93f63c9853cd038aa4f00a989ebd62ae8253555e24c88b939817da18cd4e7263fda6a0e815097589bb90a5a6b3ff1ffffffff03b9000000000000001976a9149fe14d50c95abd6ecddc5d61255cfe5aebeba7e988ac57300f00000000001976a914c0492db5f283a22274ef378cdffbe5ecbe29862b88ac00000000000000000a6a0810e2cdc1af05180100000000') self.tx_obj = { 'ins': [ { 'sequence': 4294967295, 'script': b'I0F\x02!\x00\xda\x1eY\xd7\x8b\xb8\x8c\xa7\xc3\xe1:JoN%\x9d]\xd8\xcb\x17}_y\x19\x9b\xf0$\xb1\xf5q!\xd5\x02!\x00\x8d\x1d\x988`jb\xedK\xd0\x11\xa6\xce\x8a B\xae-\xc3\x8f\xd0S\x81\xb5\n\xa3\x88\xa1\xc8\xbd\x91P\x01A\x04\xd3\xb6\x15\xc6\t\xe4\x8a\xe8\x13\x89\xf6a{PG;\xf4\xc9?c\xc9\x85<\xd08\xaaO\x00\xa9\x89\xeb\xd6*\xe8%5U\xe2L\x88\xb99\x81}\xa1\x8c\xd4\xe7&?\xdaj\x0e\x81P\x97X\x9b\xb9\nZk?\xf1', 'outpoint': {'index': 1, 'hash': b'\x9c\x82ln\xbfB`\xcb\xeaPls?\x88\xe6\x07\x18\x1d\xe4\xf6\x8ah\xbbg\xff\xfc\x08\xd7 \xafzc'} } ], 'locktime': 0, 'version': 1, 'outs': [ { 'value': 185, 'script': b'v\xa9\x14\x9f\xe1MP\xc9Z\xbdn\xcd\xdc]a%\\\xfeZ\xeb\xeb\xa7\xe9\x88\xac' }, { 'value': 995415, 'script': b'v\xa9\x14\xc0I-\xb5\xf2\x83\xa2"t\xef7\x8c\xdf\xfb\xe5\xec\xbe)\x86+\x88\xac' }, { 'value': 0, 'script': b'j\x08\x10\xe2\xcd\xc1\xaf\x05\x18\x01' } ] } def test_deserialization(self): tx_obj = transaction.deserialize(self.tx_bin) self.assertEqual(tx_obj, self.tx_obj) def test_serialization(self): tx_bin = transaction.serialize(self.tx_obj) self.assertEqual(tx_bin, self.tx_bin) class TestSighash(TestCase): def setUp(self): loc = os.path.realpath(os.path.join(os.getcwd(), os.path.dirname(__file__))) with open(os.path.join(loc, 'sighash.json')) as f: self.data = json.load(f) def test_sighash(self): first = True for vector in self.data: # Ignore first header row in the JSON. if first: first = False continue tx = transaction.deserialize(convert.hex_to_bytes(vector[0])) script = convert.hex_to_bytes(vector[1]) index = int(vector[2]) hashtype = int(vector[3]) & 0xffffffff # This must be unsigned int sighash = convert.hex_to_bytes(vector[4])[::-1] # It's reversed for some reason? my_sighash = transaction.sighash(tx, index, script, hashtype) self.assertEqual( sighash, my_sighash, 'hashtype = {:x}'.format(hashtype) )
btcmagic/test_transaction.py
3,355
Ignore first header row in the JSON. This must be unsigned int It's reversed for some reason?
93
en
0.925768
from os import environ from pathlib import Path # Build paths inside the project like this: BASE_DIR / 'subdir'. BASE_DIR = Path(__file__).resolve().parent.parent.parent DEBUG = True CRON_ENABLED = False if 'SIMONE_DB_NAME' in environ: DATABASES = { 'default': { 'ENGINE': 'mysql.connector.django', 'NAME': environ['SIMONE_DB_NAME'], 'USER': environ['SIMONE_DB_USER'], 'PASSWORD': environ['SIMONE_DB_PASSWORD'], 'HOST': environ['SIMONE_DB_HOST'], 'PORT': environ.get('SIMONE_DB_PORT', '3306'), 'CONN_MAX_AGE': 300, } } else: DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': BASE_DIR / 'db' / 'db.sqlite3', } } LOGGING = { 'version': 1, 'disable_existing_loggers': False, 'formatters': { 'simple': { 'format': '%(asctime)s %(levelname)-5s %(name)s %(message)s', 'datefmt': '%Y-%m-%dT%H:%M:%SZ', } }, 'handlers': { 'console': { 'class': 'logging.StreamHandler', 'level': 'DEBUG', 'formatter': 'simple', }, 'file': { 'class': 'logging.handlers.WatchedFileHandler', 'level': 'DEBUG', 'formatter': 'simple', 'filename': 'django.log', }, }, 'root': {'level': 'DEBUG', 'handlers': ('console', 'file')}, 'loggers': { 'django.db.backends': { # comment out to see db queries 'level': 'INFO' }, 'slack_bolt': { # super noisy 'level': 'INFO' }, }, }
simone/settings/dev.py
1,698
Build paths inside the project like this: BASE_DIR / 'subdir'. comment out to see db queries super noisy
104
en
0.491747
"""Nyamuk event.""" import socket import nyamuk_const as NC #mqtt event EV_CONNACK = NC.CMD_CONNACK EV_PUBLISH = NC.CMD_PUBLISH EV_SUBACK = NC.CMD_SUBACK #non mqtt event EV_NET_ERR = 1000 class BaseEvent: """Event Base Class.""" def __init__(self, tipe): self.type = tipe class EventConnack(BaseEvent): """CONNACK received.""" def __init__(self, ret_code, session_present = 0): BaseEvent.__init__(self, NC.CMD_CONNACK) self.ret_code = ret_code # v3.1.1 only self.session_present = session_present class EventPublish(BaseEvent): """PUBLISH received.""" def __init__(self, msg): BaseEvent.__init__(self, NC.CMD_PUBLISH) self.msg = msg class EventSuback(BaseEvent): """SUBACK received.""" def __init__(self, mid, granted_qos): BaseEvent.__init__(self, NC.CMD_SUBACK) self.mid = mid self.granted_qos = granted_qos class EventUnsuback(BaseEvent): """UNSUBACK received.""" def __init__(self, mid): BaseEvent.__init__(self, NC.CMD_UNSUBACK) self.mid = mid class EventPuback(BaseEvent): """PUBACK received.""" def __init__(self, mid): BaseEvent.__init__(self, NC.CMD_PUBACK) self.mid = mid class EventPubrec(BaseEvent): """PUBREC received.""" def __init__(self, mid): BaseEvent.__init__(self, NC.CMD_PUBREC) self.mid = mid class EventPubrel(BaseEvent): """PUBREL received.""" def __init__(self, mid): BaseEvent.__init__(self, NC.CMD_PUBREL) self.mid = mid class EventPubcomp(BaseEvent): """PUBCOMP received.""" def __init__(self, mid): BaseEvent.__init__(self, NC.CMD_PUBCOMP) self.mid = mid class EventNeterr(BaseEvent): """Network error event.""" def __init__(self, errnum, msg): BaseEvent.__init__(self, EV_NET_ERR) self.errnum = errnum self.msg = msg class EventPingResp(BaseEvent): """PINGRESP received.""" def __init__(self): BaseEvent.__init__(self, NC.CMD_PINGRESP)
nyamuk/event.py
2,060
Event Base Class. CONNACK received. Network error event. PINGRESP received. PUBACK received. PUBCOMP received. PUBLISH received. PUBREC received. PUBREL received. SUBACK received. UNSUBACK received. Nyamuk event. mqtt eventnon mqtt event v3.1.1 only
250
en
0.984124
# Generated by Django 2.2.7 on 2019-11-20 17:08 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('quiz', '0002_question_image'), ] operations = [ migrations.RemoveField( model_name='question', name='answer', ), migrations.AddField( model_name='choice', name='is_correct', field=models.BooleanField(default=False, help_text='Mark right if this is the right choice'), ), ]
quiz/migrations/0003_auto_20191120_2238.py
541
Generated by Django 2.2.7 on 2019-11-20 17:08
45
en
0.614117
#!/usr/bin/python # coding: utf-8 """A simple webserver.""" # python 2.7 compatibility from __future__ import print_function, unicode_literals # based on tornado import tornado.ioloop import tornado.web import tornado.websocket import sys import json def make_app(): """Create and return the main Tornado web application. It will listen on the port assigned via `app.listen(port)`, and will run on Tornado's main ioloop, which can be started with `tornado.ioloop.IOLoop.current().start()`. """ return tornado.web.Application([ (r"/connect", ClientSocket), (r"/(.*)", tornado.web.StaticFileHandler, { "path": "client", "default_filename": "index.html" }), ], debug=True) class ClientSocket(tornado.websocket.WebSocketHandler): """ClientSocket represents an active websocket connection to a client. """ def open(self): """Called when a websocket connection is initiated.""" # print some info about the opened connection print("WebSocket opened", "from user at {}".format(self.request.remote_ip)) def on_message(self, message): """Called when a websocket client sends a message.""" # print the message to the console print("client sent: {!r}".format(message)) # try to parse the message try: parsed_message = json.loads(message) except ValueError: print("Failed to parse message: {!r}".format(message)) return # if there's a "message" in the message, echo it if "message" in parsed_message: response = { "client" : str(self.request.remote_ip), "message" : parsed_message["message"] } # respond to the message m = json.dumps(response) self.write_message(m) else: print("message unhandled.") def on_close(self): """Called when a client connection is closed for any reason.""" # print some info about the closed connection print("WebSocket closed", "by user at {}".format(self.request.remote_ip)) print("close code: {}".format(self.close_code)) print("close reason: {!r}".format(self.close_reason)) if __name__ == "__main__": # print some basic info about the system print("Running Tornado Web Server {}".format(tornado.version)) print("Using Python {}".format(sys.version)) # start the webapp on port 8888 app = make_app() app.listen(8888) print("webapp started on port 8888") tornado.ioloop.IOLoop.current().start()
6/server.py
2,720
ClientSocket represents an active websocket connection to a client. Create and return the main Tornado web application. It will listen on the port assigned via `app.listen(port)`, and will run on Tornado's main ioloop, which can be started with `tornado.ioloop.IOLoop.current().start()`. Called when a client connection is closed for any reason. Called when a websocket client sends a message. Called when a websocket connection is initiated. A simple webserver. !/usr/bin/python coding: utf-8 python 2.7 compatibility based on tornado print some info about the opened connection print the message to the console try to parse the message if there's a "message" in the message, echo it respond to the message print some info about the closed connection print some basic info about the system start the webapp on port 8888
827
en
0.826251
"""Dataset, producer, and config metadata.""" import logging import warnings import sqlalchemy as sa from .._globals import REGISTRY as registry from .. import _tools from .. import backend as _backend __all__ = ['Dataset', 'Producer', 'Config'] log = logging.getLogger(__name__) @registry.mapped class Dataset: """Git commit loaded into the database.""" __tablename__ = '__dataset__' id = sa.Column(sa.Integer, sa.CheckConstraint('id = 1'), primary_key=True) title = sa.Column(sa.Text, sa.CheckConstraint("title != ''"), nullable=False) git_commit = sa.Column(sa.String(40), sa.CheckConstraint('length(git_commit) = 40'), nullable=False, unique=True) git_describe = sa.Column(sa.Text, sa.CheckConstraint("git_describe != ''"), nullable=False, unique=True) clean = sa.Column(sa.Boolean(create_constraint=True), nullable=False) version = sa.Column(sa.Text, sa.CheckConstraint("version != ''")) exclude_raw = sa.Column(sa.Boolean(create_constraint=True), nullable=False) @classmethod def get_dataset(cls, *, bind, strict, fallback=None): table = cls.__tablename__ log.debug('read %r from %r', table, bind) try: result, = _backend.iterrows(sa.select(cls), mappings=True, bind=bind) except sa.exc.OperationalError as e: if 'no such table' in e.orig.args[0]: pass else: log.exception('error selecting %r', table) if strict: # pragma: no cover raise RuntimeError('failed to select %r from %r', table, bind) from e return fallback except ValueError as e: log.exception('error selecting %r', table) if 'not enough values to unpack' in e.args[0] and not strict: return fallback else: # pragma: no cover raise RuntimeError('failed to select %r from %r', table, bind) from e except Exception as e: # pragma: no cover log.exception('error selecting %r', table) raise RuntimeError('failed to select %r from %r', table, bind) from e else: return result @classmethod def log_dataset(cls, params, *, ignore_dirty: bool = False, also_print: bool = False, print_file=None): name = cls.__tablename__ log.info('git describe %(git_describe)r clean: %(clean)r', params) log.debug('%s.title: %r', name, params['title']) log.info('%s.git_commit: %r', name, params['git_commit']) if 'version' in params: log.info('%s.version: %r', name, params['version']) log.debug('%s.exclude_raw: %r', name, params['exclude_raw']) if also_print or print_file is not None: print('git describe {git_describe!r}' ' clean: {clean!r}'.format_map(params), file=print_file) print(f"{name}.title: {params['title']!r}'", file=print_file) print(f"{name}.git_commit: {params['git_commit']!r}", file=print_file) if 'version' in params: print(f"{name}.version: {params['version']!r}", file=print_file) print(f"{name}.exclude_raw: {params['exclude_raw']!r}", file=print_file) if not params['clean'] and not ignore_dirty: warnings.warn(f'{name} not clean,' ' pass ignore_dirty=True to disable') # pragma: no cover @registry.mapped class Producer: """Name and version of the package that created a __dataset__.""" __tablename__ = '__producer__' id = sa.Column(sa.Integer, sa.CheckConstraint('id = 1'), primary_key=True) name = sa.Column(sa.Text, sa.CheckConstraint("name != ''"), unique=True, nullable=False) version = sa.Column(sa.Text, sa.CheckConstraint("version != ''"), nullable=False) @classmethod def get_producer(cls, *, bind): result, = _backend.iterrows(sa.select(cls), mappings=True, bind=bind) return result @classmethod def log_producer(cls, params, *, also_print=False, print_file=None): name = cls.__tablename__ log.info('%s.name: %s', name, params['name']) log.info('%s.version: %s', name, params['version']) if also_print or print_file is not None: print(f"{name}.name: {params['name']}", file=print_file) print(f"{name}.version: {params['version']}", file=print_file) @registry.mapped class Config: """Configuration setting from ``glottolog/config/*.ini``.""" __tablename__ = '_config' filename = sa.Column(sa.String, sa.CheckConstraint("filename != ''"), primary_key=True) section = sa.Column(sa.String, sa.CheckConstraint("section != ''"), primary_key=True) option = sa.Column(sa.String, sa.CheckConstraint("option != ''"), primary_key=True) value = sa.Column(sa.Text, sa.CheckConstraint("value != ''"), nullable=False) line = sa.Column(sa.Integer, sa.CheckConstraint('line > 0'), nullable=False) __table_args__ = (sa.UniqueConstraint(filename, line), {'info': {'without_rowid': True}}) @classmethod def load(cls, filename: str, *, bind, _groupby_section=_tools.groupby_itemgetter(0)): select_values = (sa.select(Config.section, Config.option, Config.value) .filter_by(filename=filename) .order_by('section', 'option')) result = _backend.iterrows(select_values, bind=bind) return {section: {option: value for _, option, value in grp} for section, grp in _groupby_section(result)}
treedb/backend/models.py
5,941
Configuration setting from ``glottolog/config/*.ini``. Git commit loaded into the database. Name and version of the package that created a __dataset__. Dataset, producer, and config metadata. pragma: no cover pragma: no cover pragma: no cover pragma: no cover
261
en
0.634529
import logging import multiprocessing import os import signal import sys import time from typing import Any from datastore.reader.app import register_services from gunicorn.app.base import BaseApplication from .shared.env import is_dev_mode from .shared.interfaces.logging import LoggingModule from .shared.interfaces.wsgi import WSGIApplication register_services() # ATTENTION: We use the Python builtin logging module. To change this use # something like "import custom_logging as logging". DEFAULT_ADDRESSES = { "ActionView": "0.0.0.0:9002", "PresenterView": "0.0.0.0:9003", } class OpenSlidesBackendGunicornApplication(BaseApplication): # pragma: no cover """ Standalone application class for Gunicorn. It prepares Gunicorn for using OpenSlidesBackendWSGIApplication via OpenSlidesBackendWSGIContainer either with action component or with presenter component. """ def __init__(self, view_name: str, *args: Any, **kwargs: Any) -> None: # Setup global loglevel. if is_dev_mode(): logging.basicConfig(level=logging.DEBUG) logger = logging.getLogger(__name__) self.view_name = view_name if self.view_name not in ("ActionView", "PresenterView"): raise ValueError( f"View name has to be ActionView or PresenterView, not {self.view_name}." ) logger.debug(f"Create gunicorn application for {self.view_name}.") super().__init__(*args, **kwargs) def load_config(self) -> None: dev_mode = is_dev_mode() options = { "bind": DEFAULT_ADDRESSES[self.view_name], "worker_tmp_dir": "/dev/shm", # See https://pythonspeed.com/articles/gunicorn-in-docker/ "timeout": int(os.environ.get("OPENSLIDES_BACKEND_WORKER_TIMEOUT", "30")), "loglevel": "debug" if dev_mode else "info", "reload": dev_mode, "reload_engine": "auto", # This is the default however. } for key, value in options.items(): self.cfg.set(key, value) def load(self) -> WSGIApplication: # We import this here so Gunicorn can use its reload feature properly. from .wsgi import create_wsgi_application # TODO: Fix this typing problem. logging_module: LoggingModule = logging # type: ignore return create_wsgi_application(logging_module, self.view_name) def start_action_server() -> None: # pragma: no cover OpenSlidesBackendGunicornApplication(view_name="ActionView").run() def start_presenter_server() -> None: # pragma: no cover OpenSlidesBackendGunicornApplication(view_name="PresenterView").run() def start_them_all() -> None: # pragma: no cover print( f"Start all components in child processes. Parent process id is {os.getpid()}." ) processes = { "action": multiprocessing.Process(target=start_action_server), "presenter": multiprocessing.Process(target=start_presenter_server), } for process in processes.values(): process.start() def sigterm_handler(signalnum: int, current_stack_frame: Any) -> None: strsignal = signal.strsignal # type: ignore print( f"Parent process {os.getpid()} received {strsignal(signalnum)} " "signal. Terminate all child processes first." ) for child in multiprocessing.active_children(): child.terminate() child.join() print(f"Parent process {os.getpid()} terminated successfully.") sys.exit(0) signal.signal(signal.SIGTERM, sigterm_handler) signal.signal(signal.SIGINT, sigterm_handler) while True: for name, process in processes.items(): if not process.is_alive(): process.join() print( f"Component {name} terminated. Terminate all other components now." ) for other_name, other_process in processes.items(): if name != other_name: other_process.terminate() other_process.join() print("Parent process terminated.") sys.exit(1) time.sleep(0.1) def main() -> None: # pragma: no cover component = os.environ.get("OPENSLIDES_BACKEND_COMPONENT", "all") if component == "action": start_action_server() elif component == "presenter": start_presenter_server() elif component == "all": start_them_all() else: print( f"Error: OPENSLIDES_BACKEND_COMPONENT must not be {component}.", file=sys.stderr, ) sys.stderr.flush() sys.exit(1) sys.exit(0)
openslides_backend/main.py
4,739
Standalone application class for Gunicorn. It prepares Gunicorn for using OpenSlidesBackendWSGIApplication via OpenSlidesBackendWSGIContainer either with action component or with presenter component. ATTENTION: We use the Python builtin logging module. To change this use something like "import custom_logging as logging". pragma: no cover Setup global loglevel. See https://pythonspeed.com/articles/gunicorn-in-docker/ This is the default however. We import this here so Gunicorn can use its reload feature properly. TODO: Fix this typing problem. type: ignore pragma: no cover pragma: no cover pragma: no cover type: ignore pragma: no cover
644
en
0.739255
import logging from datetime import datetime from dateutil import parser as DatetimeParser def dicom_name(names: list) -> str: s = "^".join(names).upper() return s def dicom_date(dt: datetime) -> str: s = dt.strftime("%Y%m%d") return s def dicom_time(dt: datetime) -> str: s = dt.strftime("%H%M%S") return s def dicom_datetime(dt: datetime) -> (str, str): d = dicom_date(dt) t = dicom_time(dt) return d, t def parse_dicom_datetime(dts: str, tms: str = None) -> datetime: if tms: dts = dts + tms # GE Scanner dt format try: ts = datetime.strptime( dts, "%Y%m%d%H%M%S") return ts except ValueError: # Wrong format pass # Siemens scanners use fractional seconds try: ts = datetime.strptime( dts, "%Y%m%d%H%M%S.%f") return ts except ValueError: # Wrong format pass # Unknown format, fall back on guessing try: # Parser does _not_ like fractional seconds dts = dts.split(".")[0] ts = DatetimeParser.parse(dts) return ts except ValueError: # Wrong format pass logger = logging.getLogger("DcmStrings") logger.error(f"Failed to parse date time string: {dts}") def date_str_to_dicom(dstr): dt = DatetimeParser.parse(dstr) dcm_dt = dicom_date(dt) return dcm_dt
package/diana/utils/dicom/strings.py
1,384
GE Scanner dt format Wrong format Siemens scanners use fractional seconds Wrong format Unknown format, fall back on guessing Parser does _not_ like fractional seconds Wrong format
179
en
0.57792
# -*- encoding: utf-8 -*- # # Copyright © 2013 Red Hat, Inc # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # # Copyright (c) 2013-2016 Wind River Systems, Inc. # from cgtsclient.common import base from cgtsclient import exc CREATION_ATTRIBUTES = ['servicename', 'state'] class SmNodes(base.Resource): def __repr__(self): return "<SmNodes %s>" % self._info class SmNodesManager(base.Manager): resource_class = SmNodes @staticmethod def _path(id=None): return '/v1/servicenodes/%s' % id if id else '/v1/servicenodes' def list(self): return self._list(self._path(), "nodes") def get(self, nodes_id): try: return self._list(self._path(nodes_id))[0] except IndexError: return None def create(self, **kwargs): new = {} for (key, value) in kwargs.items(): if key in CREATION_ATTRIBUTES: new[key] = value else: raise exc.InvalidAttribute() return self._create(self._path(), new) def delete(self, nodes_id): return self._delete(self._path(nodes_id)) def update(self, nodes_id, patch): return self._update(self._path(nodes_id), patch)
cgtsclient/v1/sm_service_nodes.py
1,763
-*- encoding: utf-8 -*- Copyright © 2013 Red Hat, Inc Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Copyright (c) 2013-2016 Wind River Systems, Inc.
652
en
0.855674
import codecs import csv import datetime import logging from django.contrib.auth import get_user_model from django.contrib.auth.models import Group from todo.models import Task, TaskList log = logging.getLogger(__name__) class CSVImporter: """Core upsert functionality for CSV import, for re-use by `import_csv` management command, web UI and tests. Supplies a detailed log of what was and was not imported at the end. See README for usage notes. """ def __init__(self): self.errors = [] self.upserts = [] self.summaries = [] self.line_count = 0 self.upsert_count = 0 def upsert(self, fileobj, as_string_obj=False): """Expects a file *object*, not a file path. This is important because this has to work for both the management command and the web uploader; the web uploader will pass in in-memory file with no path! Header row is: Title, Group, Task List, Created Date, Due Date, Completed, Created By, Assigned To, Note, Priority """ if as_string_obj: # fileobj comes from mgmt command csv_reader = csv.DictReader(fileobj) else: # fileobj comes from browser upload (in-memory) csv_reader = csv.DictReader(codecs.iterdecode(fileobj, "utf-8")) # DI check: Do we have expected header row? header = csv_reader.fieldnames expected = [ "Title", "Group", "Task List", "Created By", "Created Date", "Due Date", "Completed", "Assigned To", "Note", "Priority", ] if header != expected: self.errors.append( f"Inbound data does not have expected columns.\nShould be: {expected}" ) return for row in csv_reader: self.line_count += 1 newrow = self.validate_row(row) if newrow: # newrow at this point is fully validated, and all FK relations exist, # e.g. `newrow.get("Assigned To")`, is a Django User instance. assignee = newrow.get("Assigned To") if newrow.get("Assigned To") else None created_at = ( newrow.get("Created Date") if newrow.get("Created Date") else datetime.datetime.today() ) due_date = newrow.get("Due Date") if newrow.get("Due Date") else None priority = newrow.get("Priority") if newrow.get("Priority") else None obj, created = Task.objects.update_or_create( created_by=newrow.get("Created By"), task_list=newrow.get("Task List"), title=newrow.get("Title"), defaults={ "assigned_to": assignee, "completed": newrow.get("Completed"), "created_at": created_at, "due_date": due_date, "note": newrow.get("Note"), "priority": priority, }, ) self.upsert_count += 1 msg = ( f'Upserted task {obj.id}: "{obj.title}"' f' in list "{obj.task_list}" (group "{obj.task_list.group}")' ) self.upserts.append(msg) self.summaries.append(f"Processed {self.line_count} CSV rows") self.summaries.append(f"Upserted {self.upsert_count} rows") self.summaries.append(f"Skipped {self.line_count - self.upsert_count} rows") return {"summaries": self.summaries, "upserts": self.upserts, "errors": self.errors} def validate_row(self, row): """Perform data integrity checks and set default values. Returns a valid object for insertion, or False. Errors are stored for later display. Intentionally not broken up into separate validator functions because there are interdpendencies, such as checking for existing `creator` in one place and then using that creator for group membership check in others.""" row_errors = [] # ####################### # Task creator must exist if not row.get("Created By"): msg = f"Missing required task creator." row_errors.append(msg) creator = get_user_model().objects.filter(username=row.get("Created By")).first() if not creator: msg = f"Invalid task creator {row.get('Created By')}" row_errors.append(msg) # ####################### # If specified, Assignee must exist assignee = None # Perfectly valid if row.get("Assigned To"): assigned = get_user_model().objects.filter(username=row.get("Assigned To")) if assigned.exists(): assignee = assigned.first() else: msg = f"Missing or invalid task assignee {row.get('Assigned To')}" row_errors.append(msg) # ####################### # Group must exist try: target_group = Group.objects.get(name=row.get("Group")) except Group.DoesNotExist: msg = f"Could not find group {row.get('Group')}." row_errors.append(msg) target_group = None # ####################### # Task creator must be in the target group if creator and target_group not in creator.groups.all(): msg = f"{creator} is not in group {target_group}" row_errors.append(msg) # ####################### # Assignee must be in the target group if assignee and target_group not in assignee.groups.all(): msg = f"{assignee} is not in group {target_group}" row_errors.append(msg) # ####################### # Task list must exist in the target group try: tasklist = TaskList.objects.get(name=row.get("Task List"), group=target_group) row["Task List"] = tasklist except TaskList.DoesNotExist: msg = f"Task list {row.get('Task List')} in group {target_group} does not exist" row_errors.append(msg) # ####################### # Validate Dates datefields = ["Due Date", "Created Date"] for datefield in datefields: datestring = row.get(datefield) if datestring: valid_date = self.validate_date(datestring) if valid_date: row[datefield] = valid_date else: msg = f"Could not convert {datefield} {datestring} to valid date instance" row_errors.append(msg) # ####################### # Group membership checks have passed row["Created By"] = creator row["Group"] = target_group if assignee: row["Assigned To"] = assignee # Set Completed row["Completed"] = row["Completed"] == "Yes" # ####################### if row_errors: self.errors.append({self.line_count: row_errors}) return False # No errors: return row def validate_date(self, datestring): """Inbound date string from CSV translates to a valid python date.""" try: date_obj = datetime.datetime.strptime(datestring, "%Y-%m-%d") return date_obj except ValueError: return False
todo/operations/csv_importer.py
7,623
Core upsert functionality for CSV import, for re-use by `import_csv` management command, web UI and tests. Supplies a detailed log of what was and was not imported at the end. See README for usage notes. Expects a file *object*, not a file path. This is important because this has to work for both the management command and the web uploader; the web uploader will pass in in-memory file with no path! Header row is: Title, Group, Task List, Created Date, Due Date, Completed, Created By, Assigned To, Note, Priority Inbound date string from CSV translates to a valid python date. Perform data integrity checks and set default values. Returns a valid object for insertion, or False. Errors are stored for later display. Intentionally not broken up into separate validator functions because there are interdpendencies, such as checking for existing `creator` in one place and then using that creator for group membership check in others. fileobj comes from mgmt command fileobj comes from browser upload (in-memory) DI check: Do we have expected header row? newrow at this point is fully validated, and all FK relations exist, e.g. `newrow.get("Assigned To")`, is a Django User instance. Task creator must exist If specified, Assignee must exist Perfectly valid Group must exist Task creator must be in the target group Assignee must be in the target group Task list must exist in the target group Validate Dates Group membership checks have passed Set Completed No errors:
1,484
en
0.89412
import fiftyone as fo import fiftyone.zoo as foz # Load Dataset dataset = foz.load_zoo_dataset("coco-2017", split="validation") # Randomly select 20 samples on which to generate predictions view = dataset.take(20) # Load zoo model model = foz.load_zoo_model("keypoint-rcnn-resnet50-fpn-coco-torch") # Run Inference view.apply_model(model, label_field="predictions") # Launch the FiftyOne App to visualize your dataset session = fo.launch_app(dataset) session.view = view
scripts/fiftyone_sample.py
481
Load Dataset Randomly select 20 samples on which to generate predictions Load zoo model Run Inference Launch the FiftyOne App to visualize your dataset
151
en
0.736952
#!/usr/bin/env python3 import argparse from botocore.exceptions import ClientError import os from pacu.core.lib import downloads_dir module_info = { # Name of the module (should be the same as the filename) "name": "lightsail__generate_temp_access", # Name and any other notes about the author "author": "Alexander Morgenstern alexander.morgenstern@rhinosecuritylabs.com", # Category of the module. Make sure the name matches an existing category. "category": "EXPLOIT", # One liner description of the module functionality. This shows up when a user searches for modules. "one_liner": "Creates temporary SSH keys for available instances in AWS Lightsail.", # Full description about what the module does and how it works "description": "This module creates temporary SSH keys that can be used to connect to Lightsail instances, and downloads them into the session's download directory.", # A list of AWS services that the module utilizes during its execution "services": ["Lightsail"], # For prerequisite modules, try and see if any existing modules return the data that is required for your module before writing that code yourself, that way, session data can stay separated and modular. "prerequisite_modules": ["lightsail__enum"], # External resources that the module depends on. Valid options are either a GitHub URL (must end in .git) or single file URL. "external_dependencies": [], # Module arguments to autocomplete when the user hits tab "arguments_to_autocomplete": ["--instances", "--regions"], } parser = argparse.ArgumentParser(add_help=False, description=module_info["description"]) parser.add_argument( "--instances", required=False, help="One or more Lightsail instance names, their regions, and their access protocol in the format instanceid@region@protocol. Windows instances will use the RDP protocol, and others use SSH. Defaults to all instances.", ) parser.add_argument( "--regions", required=False, default=None, help="One or more (comma separated) AWS regions in the format us-east-1. Defaults to all session regions.", ) def write_keys_to_file(created_keys, session): for region in created_keys: ssh_key_dir = os.path.join(downloads_dir(), module_info["name"], region) if not os.path.exists(ssh_key_dir): os.makedirs(ssh_key_dir) for credential in created_keys[region]: if credential["protocol"] == "rdp": windows_file_dir = os.path.join(ssh_key_dir, credential["instanceName"]) try: with open(windows_file_dir, "w") as windows_file: # Create header for file. windows_file.write("instanceName,ipAddress,username,password\n") windows_file.write(credential["instanceName"] + ",") windows_file.write(credential["ipAddress"] + ",") windows_file.write(credential["username"] + ",") windows_file.write(credential["password"] + "\n") except IOError: print( "Error writing credential file for {}.".format( credential["instanceName"] ) ) continue else: private_key_file_dir = os.path.join( ssh_key_dir, credential["instanceName"] ) cert_key_file_dir = os.path.join( ssh_key_dir, credential["instanceName"] + "-cert.pub" ) try: with open(private_key_file_dir, "w") as private_key_file: private_key_file.write(credential["privateKey"]) with open(cert_key_file_dir, "w") as cert_key_file: cert_key_file.write(credential["certKey"]) except IOError: print( "Error writing credential file for {}.".format( credential["instanceName"] ) ) continue def main(args, pacu_main): session = pacu_main.get_active_session() print = pacu_main.print get_regions = pacu_main.get_regions fetch_data = pacu_main.fetch_data args = parser.parse_args(args) regions = args.regions.split(",") if args.regions else get_regions("lightsail") instances = [] if ( args.instances is not None ): # need to update this to include the regions of these IDs for instance in args.instances.split(","): instance_name = instance.split("@")[0] region = instance.split("@")[1] protocol = instance.split("@")[2] if region not in regions: print(" {} is not a valid region".format(region)) continue else: instances.append( { "name": instance_name, "protocol": protocol, "region": region, } ) else: print("Targeting all Lightsail instances...") if ( fetch_data( ["Lightsail"], module_info["prerequisite_modules"][0], "--instances" ) is False ): print("Pre-req module not run successfully. Exiting...") return for instance in session.Lightsail["instances"]: if instance["region"] in regions: protocol = "rdp" if "Windows" in instance["blueprintName"] else "ssh" instances.append( { "name": instance["name"], "protocol": protocol, "region": instance["region"], } ) temp_keys = {} for instance in instances: temp_keys[instance["region"]] = [] for instance in instances: client = pacu_main.get_boto3_client("lightsail", instance["region"]) print(" Instance {}".format(instance["name"])) try: response = client.get_instance_access_details( instanceName=instance["name"], protocol=instance["protocol"] ) temp_keys[instance["region"]].append(response["accessDetails"]) print( " Successfully created temporary access for {}".format( instance["name"] ) ) except ClientError as error: code = error.response["Error"]["Code"] if code == "AccessDeniedException": print(" Unauthorized to generate temporary access.") return elif code == "OperationFailureException": print(" FAILED: Unable to interact with non-running instance.") continue else: print(error) break write_keys_to_file(temp_keys, session) windows_count = 0 ssh_count = 0 for region in temp_keys: for credential in temp_keys[region]: if credential["protocol"] == "rdp": windows_count += 1 else: ssh_count += 1 if windows_count or ssh_count: written_file_path = os.path.join(downloads_dir(), module_info["name"]) else: written_file_path = None summary_data = { "windows": windows_count, "linux": ssh_count, "written_file_path": written_file_path, } return summary_data def summary(data, pacu_main): out = " Created temporary access for {} Windows instances.\n".format( data["windows"] ) out += " Created temporary access for {} Linux instances.\n".format(data["linux"]) if data["written_file_path"] is not None: out += "\n Credential files written to:\n {}{}".format( data["written_file_path"], os.path.sep ) return out
pacu/modules/lightsail__generate_temp_access/main.py
8,140
!/usr/bin/env python3 Name of the module (should be the same as the filename) Name and any other notes about the author Category of the module. Make sure the name matches an existing category. One liner description of the module functionality. This shows up when a user searches for modules. Full description about what the module does and how it works A list of AWS services that the module utilizes during its execution For prerequisite modules, try and see if any existing modules return the data that is required for your module before writing that code yourself, that way, session data can stay separated and modular. External resources that the module depends on. Valid options are either a GitHub URL (must end in .git) or single file URL. Module arguments to autocomplete when the user hits tab Create header for file. need to update this to include the regions of these IDs
882
en
0.803181
# -*- coding: utf-8 -*- import sys import argparse from cgate.reader import readfile, readschema, get_dtype from cgate.validation import validate def main(): parser = argparse.ArgumentParser() parser.add_argument('target', help='Table name or File path') parser.add_argument('--schema', '-s', help='Cerberus schema file') parser.add_argument('--null', '-n', help='Null character', default='NULL,\\N') args = parser.parse_args() schema = readschema(args.schema) try: header = schema['header'] except: header = None na_values = args.null.split(',') dtype, date_cols = get_dtype(schema['schema']) dfs = readfile(args.target, header=header, dtype=dtype, parse_dates=date_cols, na_values=na_values) fail_count = validate(dfs, schema['schema']) if fail_count != 0: print('Failed {0} error...'.format(fail_count), file=sys.stderr) return 1 print('Success!', file=sys.stderr) return 0
cgate/cgate.py
973
-*- coding: utf-8 -*-
21
en
0.767281
# Generated by Django 2.2.6 on 2020-09-03 03:23 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('comment', '0001_initial'), ] operations = [ migrations.AlterField( model_name='comment', name='target', field=models.CharField(max_length=100, verbose_name='评论目标'), ), ]
comment/migrations/0002_auto_20200903_0323.py
406
Generated by Django 2.2.6 on 2020-09-03 03:23
45
en
0.737223
from django.contrib import admin # Register your models here. from .models import Join class JoinAdmin(admin.ModelAdmin): list_display = ['email', 'friend', 'timestamp', 'updated'] class Meta: model = Join admin.site.register(Join, JoinAdmin)
trader/joins/admin.py
263
Register your models here.
26
en
0.957485
import logging import os import boto3 from lib.cleanup_resource_handler import CleanupResourceHandler from lib.queue_handler import QueueHandler logging.getLogger().setLevel(logging.INFO) def queue_handler(event, context): """ Handler for the event queue lambda trigger """ ec2_client = boto3.client('ec2') dynamodb_resource = boto3.resource('dynamodb') route53_client = boto3.client('route53') handler = QueueHandler(ec2_client=ec2_client, dynamodb_resource=dynamodb_resource, route53_client=route53_client, environ=os.environ) return handler.handle(event, context) def cleanup_resource_handler(event, context): """ Event handler for the custom resource. """ route53_client = boto3.client('route53') handler = CleanupResourceHandler(route53_client=route53_client) handler.handle_event(event, context)
packages/@aws-cdk-containers/ecs-service-extensions/lib/extensions/assign-public-ip/lambda/index.py
896
Event handler for the custom resource. Handler for the event queue lambda trigger
81
en
0.802655
# inclass/mongo_queries.py import pymongo import os from dotenv import load_dotenv import sqlite3 load_dotenv() DB_USER = os.getenv("MONGO_USER", default="OOPS") DB_PASSWORD = os.getenv("MONGO_PASSWORD", default="OOPS") CLUSTER_NAME = os.getenv("MONGO_CLUSTER_NAME", default="OOPS") connection_uri = f"mongodb+srv://{DB_USER}:{DB_PASSWORD}@{CLUSTER_NAME}.mongodb.net/test?retryWrites=true&w=majority&ssl=true&ssl_cert_reqs=CERT_NONE" print("----------------") print("URI:", connection_uri) client = pymongo.MongoClient(connection_uri) print("----------------") print("CLIENT:", type(client), client) # print(dir(client)) # print("DB NAMES:", client.list_database_names()) #> ['admin', 'local'] db = client.ds14_db # "ds14_db" or whatever you want to call it # print("----------------") # print("DB:", type(db), db) # collection = db.ds14_pokemon_collection # "ds14_collection" or whatever you want to call it # print("----------------") # print("COLLECTION:", type(collection), collection) # print("----------------") # # print("COLLECTIONS:") # # print(db.list_collection_names()) # print("--------------------------------------") ################## ASSIGNMENT III ############################# # INSERT RPG DATA INTO MONGODB INSTANCE # Create RPG database db = client.rpg_data_db # Establish sqlite3 connection to access rpg data sl_conn = sqlite3.connect("data/rpg_db_original.sqlite3") sl_curs = sl_conn.cursor() ################# CHARACTERS ########################### # ## Create new collection for RPG data # col_characters = db.character_collection # ## Establish SQL syntax for query # rpg_characters = 'SELECT * FROM charactercreator_character' # # Function to loop through characters and return list of dictionaries # def all_chars(): # query = rpg_characters # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "character_id": row[0], # "name": row[1], # "level": row[2], # "exp": row[3], # "hp": row[4], # "strength": row[5], # "intelligence": row[6], # "dexterity": row[7], # "wisdom": row[8] # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # # print(character_dict_list) # col_characters.insert_many(character_dict_list) # print("DOCS(Num Characters):", col_characters.count_documents({})) # # SELECT count(distinct id) from characters ################# MAGES ########################### # col_mage = db.mage_collection # mages = 'SELECT * FROM charactercreator_mage' # def all_chars(): # query = mages # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "character_ptr_id": row[0], # "has_pet": row[1], # "mana": row[2], # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_mage.insert_many(character_dict_list) # print("DOCS:", col_mage.count_documents({})) ################# THIEVES ########################### # col_thief = db.thief_collection # thieves = 'SELECT * FROM charactercreator_thief' # def all_chars(): # query = thieves # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "character_ptr_id": row[0], # "is_sneaking": row[1], # "energy": row[2], # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_thief.insert_many(character_dict_list) # print("DOCS:", col_thief.count_documents({})) ################# CLERICS ########################### # col_cleric = db.cleric_collection # clerics = 'SELECT * FROM charactercreator_cleric' # def all_chars(): # query = clerics # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "character_ptr_id": row[0], # "using_shield": row[1], # "mana": row[2], # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_cleric.insert_many(character_dict_list) # print("DOCS:", col_cleric.count_documents({})) ################# FIGHTERS ########################### # col_fighter = db.fighter_collection # fighters = 'SELECT * FROM charactercreator_fighter' # def all_chars(): # query = fighters # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "character_ptr_id": row[0], # "using_shield": row[1], # "rage": row[2], # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_fighter.insert_many(character_dict_list) # print("DOCS:", col_fighter.count_documents({})) ################# NECROMANCERS ########################### # col_mancer = db.mancer_collection # mancers = 'SELECT * FROM charactercreator_necromancer' # def all_chars(): # query = mancers # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "mage_ptr_id": row[0], # "talisman_charged": row[1], # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_mancer.insert_many(character_dict_list) # print("DOCS:", col_mancer.count_documents({})) ################# ITEMS ########################### # col_items = db.items_collection # items = 'SELECT * FROM armory_item' # def all_chars(): # query = items # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "item_id": row[0], # "name": row[1], # "value": row[2], # "weight": row[3] # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_items.insert_many(character_dict_list) # print("DOCS:", col_items.count_documents({})) ################# WEAPONS ########################### # col_weapons = db.weapons_collection # weapons = 'SELECT * FROM armory_weapon' # def all_chars(): # query = weapons # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "item_ptr_id": row[0], # "power": row[1] # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_weapons.insert_many(character_dict_list) # print("DOCS:", col_weapons.count_documents({})) ################# INVENTORY ########################### # col_inventory = db.inventory_collection # records = 'SELECT * FROM charactercreator_character_inventory' # def all_chars(): # query = records # chars = sl_curs.execute(query) # char_data = [] # for row in chars: # character = { # "id": row[0], # "character_id": row[1], # "item_id": row[2] # } # char_data.append(character) # result = char_data # return result # character_dict_list = all_chars() # col_inventory.insert_many(character_dict_list) # print("DOCS:", col_inventory.count_documents({})) # print("COLLECTIONS:") # print(db.list_collection_names()) #################### IN-CLASS POKEMON INSERTS ############################# # collection.insert_one({ # "name": "Pikachu", # "level": 30, # "exp": 76000000000, # "hp": 400, # "fav_icecream_flavors":["vanila_bean", "choc"], # "stats":{"a":1,"b":2,"c":[1,2,3]} # }) # print("DOCS:", collection.count_documents({})) # SELECT count(distinct id) from pokemon # print(collection.count_documents({"name": "Pikachu"})) # SELECT # count(distinct id) from pokemon WHERE name = "Pikachu" # mewtwo = { # "name": "Mewtwo", # "level": 100, # "exp": 76000000000, # "hp": 450, # "strength": 550, # "intelligence": 450, # "dexterity": 300, # "wisdom": 575 # } # blastoise = { # "name": "Blastoise", # "lvl": 70, # OOPS we made a mistake with the structure of this dict # } # charmander = { # "nameeeeeee": "Charmander", # "level": 70, # "random_stat": {"a":2} # } # skarmory = { # "name": "Skarmory", # "level": 22, # "exp": 42000, # "hp": 85, # "strength": 750, # "intelligence": 8, # "dexterity": 57 # } # cubone = { # "name": "Cubone", # "level": 20, # "exp": 35000, # "hp": 80, # "strength": 600, # "intelligence": 60, # "dexterity": 200, # "wisdom": 200 # } # scyther = { # "name": "Scyther", # "level": 99, # "exp": 7000, # "hp": 40, # "strength": 50, # "intelligence": 40, # "dexterity": 30, # "wisdom": 57 # } # slowpoke = { # "name": "Slowpoke", # "level": 1, # "exp": 100, # "hp": 80, # "strength": 100, # "intelligence": 10, # "dexterity": 50, # "wisdom": 200 # } # pokemon_team = [mewtwo, blastoise, skarmory, cubone, scyther, slowpoke, charmander] # collection.insert_many(pokemon_team) # print("DOCS:", collection.count_documents({})) # SELECT count(distinct id) from pokemon # #collection.insert_one({"_id": "OURVAL", "name":"TEST"}) # # can overwrite the _id but not insert duplicate _id values # #breakpoint() # pikas = list(collection.find({"name": "Pikachu"})) # SELECT * FROM pokemon WHERE name = "Pikachu" # # print(len(pikas), "PIKAS") # # print(pikas[0]["_id"]) #> ObjectId('5ebc31c79c171e43bb5ed469') # # print(pikas[0]["name"]) # # strong = list(collection.find({"level": {"$gte": 60}} $or {"lvl": {"$gte": 60}})) # # strong = list(collection.find({"level": {"$gte": 60}, "$or" "lvl": {"$gte": 60}})) # strong = list(collection.find({"$or": [{"level": {"$gte": 60}}, {"lvl": {"$gte": 60}}]})) # # TODO: also try to account for our mistakes "lvl" vs "level" # breakpoint() # print(strong)
assignment3/a3_mongo_queries_abw.py
10,194
inclass/mongo_queries.py print(dir(client)) print("DB NAMES:", client.list_database_names()) > ['admin', 'local'] "ds14_db" or whatever you want to call it print("----------------") print("DB:", type(db), db) collection = db.ds14_pokemon_collection "ds14_collection" or whatever you want to call it print("----------------") print("COLLECTION:", type(collection), collection) print("----------------") print("COLLECTIONS:") print(db.list_collection_names()) print("--------------------------------------") ASSIGNMENT III INSERT RPG DATA INTO MONGODB INSTANCE Create RPG database Establish sqlite3 connection to access rpg data CHARACTERS Create new collection for RPG data col_characters = db.character_collection Establish SQL syntax for query rpg_characters = 'SELECT * FROM charactercreator_character' Function to loop through characters and return list of dictionaries def all_chars(): query = rpg_characters chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "character_id": row[0], "name": row[1], "level": row[2], "exp": row[3], "hp": row[4], "strength": row[5], "intelligence": row[6], "dexterity": row[7], "wisdom": row[8] } char_data.append(character) result = char_data return result character_dict_list = all_chars() print(character_dict_list) col_characters.insert_many(character_dict_list) print("DOCS(Num Characters):", col_characters.count_documents({})) SELECT count(distinct id) from characters MAGES col_mage = db.mage_collection mages = 'SELECT * FROM charactercreator_mage' def all_chars(): query = mages chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "character_ptr_id": row[0], "has_pet": row[1], "mana": row[2], } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_mage.insert_many(character_dict_list) print("DOCS:", col_mage.count_documents({})) THIEVES col_thief = db.thief_collection thieves = 'SELECT * FROM charactercreator_thief' def all_chars(): query = thieves chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "character_ptr_id": row[0], "is_sneaking": row[1], "energy": row[2], } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_thief.insert_many(character_dict_list) print("DOCS:", col_thief.count_documents({})) CLERICS col_cleric = db.cleric_collection clerics = 'SELECT * FROM charactercreator_cleric' def all_chars(): query = clerics chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "character_ptr_id": row[0], "using_shield": row[1], "mana": row[2], } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_cleric.insert_many(character_dict_list) print("DOCS:", col_cleric.count_documents({})) FIGHTERS col_fighter = db.fighter_collection fighters = 'SELECT * FROM charactercreator_fighter' def all_chars(): query = fighters chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "character_ptr_id": row[0], "using_shield": row[1], "rage": row[2], } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_fighter.insert_many(character_dict_list) print("DOCS:", col_fighter.count_documents({})) NECROMANCERS col_mancer = db.mancer_collection mancers = 'SELECT * FROM charactercreator_necromancer' def all_chars(): query = mancers chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "mage_ptr_id": row[0], "talisman_charged": row[1], } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_mancer.insert_many(character_dict_list) print("DOCS:", col_mancer.count_documents({})) ITEMS col_items = db.items_collection items = 'SELECT * FROM armory_item' def all_chars(): query = items chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "item_id": row[0], "name": row[1], "value": row[2], "weight": row[3] } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_items.insert_many(character_dict_list) print("DOCS:", col_items.count_documents({})) WEAPONS col_weapons = db.weapons_collection weapons = 'SELECT * FROM armory_weapon' def all_chars(): query = weapons chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "item_ptr_id": row[0], "power": row[1] } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_weapons.insert_many(character_dict_list) print("DOCS:", col_weapons.count_documents({})) INVENTORY col_inventory = db.inventory_collection records = 'SELECT * FROM charactercreator_character_inventory' def all_chars(): query = records chars = sl_curs.execute(query) char_data = [] for row in chars: character = { "id": row[0], "character_id": row[1], "item_id": row[2] } char_data.append(character) result = char_data return result character_dict_list = all_chars() col_inventory.insert_many(character_dict_list) print("DOCS:", col_inventory.count_documents({})) print("COLLECTIONS:") print(db.list_collection_names()) IN-CLASS POKEMON INSERTS collection.insert_one({ "name": "Pikachu", "level": 30, "exp": 76000000000, "hp": 400, "fav_icecream_flavors":["vanila_bean", "choc"], "stats":{"a":1,"b":2,"c":[1,2,3]} }) print("DOCS:", collection.count_documents({})) SELECT count(distinct id) from pokemon print(collection.count_documents({"name": "Pikachu"})) SELECT count(distinct id) from pokemon WHERE name = "Pikachu" mewtwo = { "name": "Mewtwo", "level": 100, "exp": 76000000000, "hp": 450, "strength": 550, "intelligence": 450, "dexterity": 300, "wisdom": 575 } blastoise = { "name": "Blastoise", "lvl": 70, OOPS we made a mistake with the structure of this dict } charmander = { "nameeeeeee": "Charmander", "level": 70, "random_stat": {"a":2} } skarmory = { "name": "Skarmory", "level": 22, "exp": 42000, "hp": 85, "strength": 750, "intelligence": 8, "dexterity": 57 } cubone = { "name": "Cubone", "level": 20, "exp": 35000, "hp": 80, "strength": 600, "intelligence": 60, "dexterity": 200, "wisdom": 200 } scyther = { "name": "Scyther", "level": 99, "exp": 7000, "hp": 40, "strength": 50, "intelligence": 40, "dexterity": 30, "wisdom": 57 } slowpoke = { "name": "Slowpoke", "level": 1, "exp": 100, "hp": 80, "strength": 100, "intelligence": 10, "dexterity": 50, "wisdom": 200 } pokemon_team = [mewtwo, blastoise, skarmory, cubone, scyther, slowpoke, charmander] collection.insert_many(pokemon_team) print("DOCS:", collection.count_documents({})) SELECT count(distinct id) from pokemon collection.insert_one({"_id": "OURVAL", "name":"TEST"}) can overwrite the _id but not insert duplicate _id values breakpoint() pikas = list(collection.find({"name": "Pikachu"})) SELECT * FROM pokemon WHERE name = "Pikachu" print(len(pikas), "PIKAS") print(pikas[0]["_id"]) > ObjectId('5ebc31c79c171e43bb5ed469') print(pikas[0]["name"]) strong = list(collection.find({"level": {"$gte": 60}} $or {"lvl": {"$gte": 60}})) strong = list(collection.find({"level": {"$gte": 60}, "$or" "lvl": {"$gte": 60}})) strong = list(collection.find({"$or": [{"level": {"$gte": 60}}, {"lvl": {"$gte": 60}}]})) TODO: also try to account for our mistakes "lvl" vs "level" breakpoint() print(strong)
8,333
en
0.531613
########################################################### # # Copyright (c) 2005, Southpaw Technology # All Rights Reserved # # PROPRIETARY INFORMATION. This software is proprietary to # Southpaw Technology, and is not to be reproduced, transmitted, # or disclosed in any way without written permission. # # # __all__ = ['TabWdg', 'TabSaveStateCmd'] from pyasm.common import TacticException, Xml, Common, Environment, Container from pyasm.web import DivWdg, SpanWdg, WebState, WebContainer, WidgetSettings, HtmlElement from pyasm.search import Search from pyasm.widget import WidgetConfigView, WidgetConfig, IconWdg from tactic.ui.common import BaseRefreshWdg import types, sys, re, os import six basestring = six.string_types class TabWdg(BaseRefreshWdg): ARGS_KEYS = { 'show_add': { 'description': 'show the + button', 'values': 'true|false', 'category': 'Display' }, 'show_context_menu': { 'description': 'show the context menu', 'values': 'true|false', 'category': 'Display' }, 'show_remove': { 'description': 'show the close button', 'values': 'true|false', 'category': 'Display' }, 'save_state': { 'description': 'key which is used to save state [ie: "save_state|main_tab" is the default]', 'category': 'Display' }, } def __init__(self, **kwargs): super(TabWdg, self).__init__(**kwargs) self.unique_id = self.generate_unique_id(is_random=True) self.header = HtmlElement.ul() self.header_id = self.header.set_unique_id() self.use_default_style = kwargs.get("use_default_style") if self.use_default_style not in [False, 'false']: self.use_default_style = True else: self.use_default_style = False def get_onload_js(self): return r''' if (spt.tab) { return; } spt.Environment.get().add_library("spt_tab"); spt.tab = {}; spt.tab.top = null; spt.tab.set_main_body_tab = function() { spt.tab.top = document.id(document.body).getElement(".spt_tab_top"); return spt.tab.top; } // this is to be deprecated spt.tab.set_main_body_top = function() { spt.tab.top = document.id('main_body').getElement(".spt_tab_top"); } spt.tab.set_tab_top = function( tab_top ) { // if this is not really a tab top, then find a child if (! tab_top.hasClass("spt_tab_top") ) { tab_top = tab_top.getElement(".spt_tab_top"); } spt.tab.top = tab_top; return spt.tab.top; } spt.tab.set_tab_top_from_child = function( el ) { // if this is not really a tab top, then find a parent if (! el.hasClass("spt_tab_top") ) { el = el.getParent(".spt_tab_top"); } spt.tab.top = el; return spt.tab.top; } spt.tab.get_headers = function() { var top = spt.tab.top; var tab_id = top.getAttribute("id"); var header_top = top.getElement(".spt_tab_header_top"); var all_headers = header_top.getElements(".spt_tab_header"); //return all_headers; var headers = []; for (var i = 0; i < all_headers.length; i++ ) { var header_tab_id = all_headers[i].getAttribute("spt_tab_id"); if (header_tab_id != tab_id) { continue; } headers.push(all_headers[i]); } return headers; } spt.tab.get_header = function(name) { var top = spt.tab.top; var header_top = top.getElement(".spt_tab_header_top"); var headers = header_top.getElements(".spt_tab_header"); for (var i = 0; i < headers.length; i++) { if (name == headers[i].getAttribute("spt_element_name") ) { return headers[i]; } } return null; } spt.tab.resize_headers = function() { var top = spt.tab.top; var header_top = top.getElement(".spt_tab_header_top"); var offset = 120; var size = top.getSize(); top.spt_last_width = size.x; var els = header_top.getElements(".spt_tab_header"); var count = els.length; var width = parseInt((size.x-offset) / (count)); if (width > 120) { width = 120; } if (width < 30) { width = 30; } for (var i = 0; i < els.length; i++) { els[i].setStyle("width", width + 'px'); //var title_el = els[i].getElement(".spt_tab_header_label"); els[i].setStyle("width", width + 'px'); } } spt.tab.get_content = function(name) { var top = spt.tab.top; var tab_id = top.getAttribute("spt_tab_id"); var content_top = top.getElement(".spt_tab_content_top"); var all_contents = content_top.getElements(".spt_tab_content"); // FIXME: this breaks when opening new tabs for some reason //return all_contents; for (var i = 0; i < all_contents.length; i++ ) { var content_tab_id = all_contents[i].getAttribute("spt_tab_id"); var content_name = all_contents[i].getAttribute("spt_element_name"); if (content_name == name) { return all_contents[i]; } } return null; } spt.tab.get_contents = function() { var top = spt.tab.top; var tab_id = top.getAttribute("spt_tab_id"); var content_top = top.getElement(".spt_tab_content_top"); var all_contents = content_top.getElements(".spt_tab_content"); // FIXME: this breaks when opening new tabs for some reason //return all_contents; var contents = []; for (var i = 0; i < all_contents.length; i++ ) { var content_tab_id = all_contents[i].getAttribute("spt_tab_id"); if (content_tab_id == null) { alert(all_contents[i].getAttribute("spt_element_name")); } if (content_tab_id != tab_id) { continue; } contents.push(all_contents[i]); } return contents; } spt.tab.set_attribute = function(element_name, name, value) { var header = spt.tab.get_header(element_name); var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs; if (kwargs_str != '') { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"') kwargs = JSON.parse(kwargs_str); } else { kwargs = {}; } kwargs[name] = value; header.setAttribute("spt_"+name, value); kwargs_str = JSON.stringify(kwargs); kwargs_str = kwargs_str.replace(/"/g,"\&amp;quot\;"); header.setAttribute("spt_kwargs", kwargs_str); } spt.tab.add_new = function(element_name, title, class_name, kwargs, values, hash) { if (typeof(title) == 'undefined') { title = '(Untitled)'; } if (typeof(element_name) == 'undefined') { //alert("No element name provided"); //return; element_name = "__default__"; } if (typeof(class_name) == 'undefined') { class_name = ''; } if (typeof(kwargs) == 'undefined') { kwargs = {}; } if (typeof(values) == 'undefined') { values = {}; } var top = spt.tab.top; if (!top) { spt.tab.set_main_body_tab(); top = spt.tab.top; } if (!hash && hash != false && kwargs.hash) { hash = kwargs.hash; } if (hash == "__link__") { hash = "link/" + element_name; } var orig_element_name = element_name; var mode = top.getAttribute("spt_tab_mode"); if (mode == "hidden") { element_name = "__default__"; } var unique = kwargs.unique; if (unique == true || unique == "true") { var header = spt.tab.get_header(element_name); if (header) { var num = Math.floor((Math.random()*10000)+1); element_name = element_name + num; } } var top_id = top.getAttribute("spt_tab_id"); // disable sub tabs for now full_element_name = element_name; subelement_name = ""; /* if (element_name.indexOf("/") != -1) { var full_element_name = element_name; var parts = element_name.split("/"); element_name = parts[0]; var subelement_name = parts[1]; } else { var full_element_name = element_name; var subelement_name = ""; } */ var subelement_title; var full_title; if (title.indexOf("/") != -1) { full_title = title; var parts = title.split("/"); title = parts[0]; subelement_title = parts[1]; } else { full_title = title; subelement_title = title; } //var headers = header_top.getElements(".spt_tab_header"); var headers = spt.tab.get_headers(); if (headers.length > 20) { spt.alert("You have too many tabs open. Please close before opening any others."); return; } var header; var found = false; for (var k=0; k < headers.length; k++){ var existing_header = headers[k]; if (existing_header.getAttribute('spt_element_name')==element_name){ header = existing_header; found = true; break; } } // add a new tab if (!found) { var template_top = top.getElement(".spt_tab_template_top"); var header_template = template_top.getElement(".spt_tab_header"); // clone the header template var header = spt.behavior.clone(header_template); var header_id = Math.floor(Math.random()*10000000+1); header.setAttribute("id", header_id); if (kwargs.hidden == "true") { header.setStyle("display", "none"); } // add a subheader template for each header var subheader_template = template_top.getElement(".spt_tab_subheader"); if (subheader_template) { var subheader = spt.behavior.clone(subheader_template); var subheader_id = Math.floor(Math.random()*10000000+1); header.setAttribute("spt_subheader_id", subheader_id); subheader.setAttribute("id", subheader_id); subheader.setStyle("display", "none"); subheader.setAttribute("spt_header_id", header_id); subheader_top = top.getElement(".spt_tab_subheader_top") subheader.inject(subheader_top); } var last_header = headers[headers.length -1]; // set the new label var label = header.getElement(".spt_tab_header_label"); var display_title = title; label.setAttribute("title", title); label.innerHTML = display_title; header.setAttribute("spt_class_name", class_name); var kwargs_str = JSON.stringify(kwargs); kwargs_str = kwargs_str.replace(/\"/,"\&amp;quot\;"); header.setAttribute("spt_kwargs", kwargs_str); header.setAttribute("spt_element_name", element_name); header.setAttribute("spt_title", title); header.setAttribute("spt_tab_id", top_id); header.removeClass("spt_content_loaded"); header.inject(last_header, "after"); spt.tab.resize_headers(); var selected_header = spt.tab.get_selected_header(); if (selected_header) { var opener = selected_header.getAttribute("spt_element_name"); header.setAttribute("spt_tab_opener", opener); } // copy the content from template var content_top = top.getElement(".spt_tab_content_top"); var content_template = template_top.getElement(".spt_tab_content"); var content_box = spt.behavior.clone(content_template); content_box.setAttribute("spt_element_name", element_name); content_box.setAttribute("spt_title", title); content_box.setAttribute("spt_tab_id", top_id); var content_boxes = spt.tab.get_contents(); var last_content = content_boxes[content_boxes.length -1]; content_box.inject(last_content, "after"); if (kwargs.count) { var count_div = header.getElement(".spt_tab_header_count"); var expression = kwargs.count; var search_key = kwargs.search_key; var server = TacticServerStub.get(); var count = server.eval(expression, {search_keys: search_key}); count_div.innerText = count; var update_data = { expression: expression, expr_key: search_key }; spt.update.add(count_div, update_data); } } // if a subtab is needed, create that if (subelement_name) { // find out if the subheader exists var subheader_id = header.getAttribute("spt_subheader_id"); var subheader_top = document.id(subheader_id); var subheaders = subheader_top.getElements(".spt_tab_subheader_item"); var subheader_exists = false; var subheader = null; for (var i = 0; i < subheaders.length; i++) { var box_name = subheaders[i].getAttribute("spt_element_name"); if (full_element_name == box_name) { subheader_exists = true; subheader = subheaders[i]; break; } } if (subheader_exists == false) { // create a new one var subheader = document.id(document.createElement("div")); subheader.innerHTML = "<div style='padding: 5px 5px'><div class='spt_tab_header_label'>"+subelement_name+"</div></div>"; subheader_top.appendChild(subheader); subheader.addClass("spt_tab_subheader_item"); // set the new label var label = subheader.getElement(".spt_tab_header_label"); var display_title = subelement_title; if (display_title.length > 20) { display_title = subelement_title.substr(0,18) + "..."; } title = subelement_name; label.setAttribute("title", subelement_title); label.innerHTML = display_title; subheader.setAttribute("spt_class_name", class_name); var kwargs_str = JSON.stringify(kwargs); kwargs_str = kwargs_str.replace(/\"/,"\&amp;quot\;"); subheader.setAttribute("spt_kwargs", kwargs_str); subheader.setAttribute("spt_element_name", full_element_name); subheader.setAttribute("spt_title", full_title); subheader.setAttribute("spt_tab_id", top_id); subheader.removeClass("spt_content_loaded"); // copy the content from template var template_top = top.getElement(".spt_tab_template_top"); var content_top = top.getElement(".spt_tab_content_top"); var content_template = template_top.getElement(".spt_tab_content"); var content_box = spt.behavior.clone(content_template); content_box.setAttribute("spt_element_name", full_element_name); content_box.setAttribute("spt_title", full_title); content_box.setAttribute("spt_tab_id", top_id); var content_boxes = spt.tab.get_contents(); var last_content = content_boxes[content_boxes.length -1]; content_box.inject(last_content, "after"); } } if (! class_name) { spt.tab.select(element_name); } else if (subelement_name) { var force = true; spt.tab.load_class(subheader, class_name, kwargs, values, force); } else { var force = true; spt.tab.load_class(header, class_name, kwargs, values, force); } // FIXME: this should only move on the main table //var top_pos = spt.tab.getY(header_top); //scroll(0,top_pos-20); // register the hash if (hash) { var state = { element_name: orig_element_name, title: title, class_name: class_name, kwargs: kwargs, hash: hash, mode: 'tab', } spt.hash.set_hash(state, title, hash); } if (top.hasClass("spt_tab_save_state") ) { spt.tab.save_state(); } return header; } // TEST spt.tab.getY = function(oElement) { var iReturnValue = 0; while( oElement != null ) { iReturnValue += oElement.offsetTop; oElement = oElement.offsetParent; } return iReturnValue; } spt.tab.load_selected = function(element_name, title, class_name, kwargs, values) { var top = spt.tab.top; var header = spt.tab.get_selected_header(); // if none are selected, use the last one if (header == null) { var headers = spt.tab.get_headers(); header = headers[headers.length - 1]; } var old_element_name = header.getAttribute("spt_element_name"); header.setAttribute("spt_element_name", element_name); header.setAttribute("spt_title", title); header.setAttribute("spt_class_name", class_name); if (typeof(kwargs) == 'undefined') { kwargs = {}; } var kwargs_str = JSON.stringify(kwargs) header.setAttribute("spt_kwargs", kwargs_str); var label = header.getElement(".spt_tab_header_label"); var display_title = title; if (display_title.length > 20) { display_title = title.substr(0,18) + "..."; } label.innerHTML = display_title; var content_top = top.getElement(".spt_tab_content_top"); var content_boxes = content_top.getElements(".spt_tab_content"); for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; var box_name = content_box.getAttribute("spt_element_name") if (box_name == old_element_name) { content_box.setAttribute("spt_element_name", element_name) break; } } var force = true; spt.tab.load_class(header, class_name, kwargs, values, force); } // add a DOM node to the named content spt.tab.load_node = function(element_name, node) { var top = spt.tab.top; var content_top = top.getElement(".spt_tab_content_top"); var content_boxes = spt.tab.get_contents(); for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; var box_name = content_box.getAttribute("spt_element_name") if (box_name == element_name) { if(content_box.hasChildNodes()) { while(content_box.childNodes.length >= 1 ) { content_box.removeChild(content_box.firstChild); } } content_box.appendChild(node); break; } } } // add raw HTML to the named content spt.tab.load_html = function(element_name, html) { var top = spt.tab.top; var content_top = top.getElement(".spt_tab_content_top"); var content_boxes = spt.tab.get_contents(); for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; var box_name = content_box.getAttribute("spt_element_name") if (box_name == element_name) { spt.behavior.replace_inner_html(content_box, html); } } } spt.tab.select = function(element_name) { var header = spt.tab.get_header(element_name); var top = spt.tab.top; var header_top = top.getElement(".spt_tab_header_top"); var headers = spt.tab.get_headers(); for (var i=0; i < headers.length; i++) { headers[i].setStyle("opacity", "0.4"); headers[i].setStyle("font-weight", "normal"); headers[i].removeClass("spt_is_selected"); headers[i].removeClass("spt_tab_selected"); headers[i].addClass("spt_tab_unselected"); } if (header) { header.setStyle("opacity", "1.0"); header.addClass("spt_is_selected"); header.addClass("spt_tab_selected"); header.removeClass("spt_tab_unselected"); } var content_top = top.getElement(".spt_tab_content_top"); var content_boxes = spt.tab.get_contents(); for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; content_box.setStyle("display", "none"); } for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; var box_name = content_box.getAttribute("spt_element_name") if (box_name == element_name) { content_box.setStyle("display", ""); if (!content_box.hasClass("spt_content_loaded")) { spt.tab.load_class(header); } break; } } var kwargs_str = header ? header.getAttribute("spt_kwargs") : ''; if (!kwargs_str) { kwargs = {}; } else { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } bvr.options = { element_name: element_name, alias: kwargs.help_alias } spt.named_events.fire_event("tab|select", bvr); // usually a tab contains a table and layout. it's better to set to that. var tab_content = top.getElement('.spt_tab_content[spt_element_name=' + element_name + ']'); if (tab_content) { var table = tab_content.getElement('.spt_table_table'); if (table) { var layout = table.getParent(".spt_layout"); spt.table.set_layout(layout); } var last_element_name = spt.tab.get_selected_element_name(); if (last_element_name) { top.setAttribute("spt_last_element_name", last_element_name); } return true; } else { return false; } } spt.tab.load_class = function(header, class_name, kwargs, values, force) { var title = header.getAttribute("spt_title"); var tab_element_name = header.getAttribute("spt_element_name"); if (typeof(force) == 'undefined') { force = false; } if (typeof(class_name) == 'undefined') { var class_name = header.getAttribute("SPT_WIDGET_KEY"); if (! class_name) { class_name = header.getAttribute("spt_class_name"); } } if (typeof(kwargs) == 'undefined') { kwargs = {}; var kwargs_str = header.getAttribute("spt_kwargs"); if (kwargs_str) { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } } var top = spt.tab.top; var header_top = top.getElement(".spt_tab_header_top"); var top_id = top.getAttribute("id"); //spt.api.app_busy_show("Loading " + title, ''); setTimeout( function() { var header_top = header.getParent(".spt_tab_header_top"); var headers = spt.tab.get_headers(); for (var i=0; i < headers.length; i++) { headers[i].setStyle("opacity", "0.4"); headers[i].setStyle("font-weight", "normal"); headers[i].removeClass("spt_is_selected"); headers[i].removeClass("spt_tab_selected"); headers[i].addClass("spt_tab_unselected"); } // select the header if (header.hasClass("spt_tab_subheader_item")) { var subheader_top = header.getParent(".spt_tab_subheader"); header_id = subheader_top.getAttribute("spt_header_id"); select_header = document.id(header_id); } else { select_header = header; } // select the header select_header.setStyle("opacity", "1.0"); select_header.addClass("spt_is_selected"); select_header.addClass("spt_tab_selected"); select_header.removeClass("spt_tab_unselected"); var content_top = top.getElement(".spt_tab_content_top"); var content_boxes = spt.tab.get_contents(); // make all of the content boxes disappear for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; content_box.setStyle("display", "none"); } for (var i=0; i < content_boxes.length; i++) { var content_box = content_boxes[i]; var box_name = content_box.getAttribute("spt_element_name") if (box_name == tab_element_name) { content_box.setStyle("display", ""); // if no class name is defined, then break out if (typeof(class_name) == 'undefined' || class_name == '') { break; } if (force || ! content_box.hasClass("spt_content_loaded")) { var resize_offset = content_box.getAttribute("spt_window_resize_offset"); spt.panel.load(content_box, class_name, kwargs, values); // update info on header header.setAttribute("spt_class_name", class_name); var kwargs_str = JSON.stringify(kwargs); kwargs_str = kwargs_str.replace(/\"/,"\&amp;quot\;"); header.setAttribute("spt_kwargs", kwargs_str); header.setAttribute("spt_element_name", tab_element_name); header.setAttribute("spt_title", title); content_box.addClass("spt_content_loaded"); // have to set this again because load removes it content_box.setAttribute("spt_element_name", tab_element_name); content_box.setAttribute("spt_tab_id", top_id); content_box.setAttribute("spt_title", title); if (resize_offset) { content_box.setAttribute("spt_window_resize_offset", resize_offset); } } break; } } var bvr = {}; var parts = tab_element_name.split("/"); var element_name = parts[parts.length-1]; var alias = kwargs.help_alias; bvr.options = { element_name: element_name, alias: alias } spt.named_events.fire_event("tab|select", bvr); spt.api.app_busy_hide(); }, 10 ); } spt.tab.reload_selected = function() { var header = spt.tab.get_selected_header(); var class_name = header.getAttribute("spt_class_name"); var kwargs = header.getAttribute("spt_kwargs"); var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs; if (kwargs_str != '') { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } else { kwargs = {}; } var values = null; var force = true; spt.tab.load_class(header, class_name, kwargs, values, force); } spt.tab.get_selected_header = function() { var top = spt.tab.top; var header_top = top.getElement(".spt_tab_header_top"); var headers = header_top.getElements(".spt_tab_header"); for (var i = 0; i < headers.length; i++) { var header = headers[i]; if ( header.hasClass("spt_is_selected") ) { return header; } } return null; } spt.tab.get_selected_element_name = function() { var header = spt.tab.get_selected_header(); if (header) { var element_name = header.getAttribute("spt_element_name"); return element_name; } return ""; } spt.tab.get_last_selected_element_name = function() { var top = spt.tab.top; return top.getAttribute("spt_last_element_name"); } spt.tab.save_state = function() { var top = spt.tab.top; var save_state = top.getAttribute("spt_tab_save_state"); var header_top = top.getElement(".spt_tab_header_top"); var headers = header_top.getElements(".spt_tab_header"); var class_names = []; var attrs_list = []; var kwargs_list = []; for (var i = 0; i < headers.length; i++) { var header = headers[i]; var element_name = header.getAttribute("spt_element_name") var title = header.getAttribute("spt_title") var attrs = { name: element_name, title: title }; var class_name = header.getAttribute("spt_class_name"); class_names.push(class_name); var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs; if (kwargs_str) { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } else { kwargs = {}; } kwargs_list.push(kwargs); if (kwargs.count) attrs.count = kwargs.count; attrs_list.push(attrs) } var server = TacticServerStub.get(); var command = 'tactic.ui.container.TabSaveStateCmd'; var kwargs = { class_names: class_names, attrs_list: attrs_list, kwargs_list: kwargs_list, save_state: save_state }; server.execute_cmd(command, kwargs, {}, { on_complete: function(ret_val) {console.log(ret_val)}, on_error: function(err) {console.log(err)} }); } spt.tab.header_pos = null; spt.tab.mouse_pos = null; spt.tab.dragging = false; spt.tab.header_drag_setup = function( evt, bvr, mouse_411) { spt.tab.top = bvr.src_el.getParent(".spt_tab_top"); spt.tab.header_pos = bvr.src_el.getPosition(spt.tab.top); spt.tab.mouse_pos = {x: mouse_411.curr_x, y: mouse_411.curr_y}; var header = bvr.src_el; var element_name = header.getAttribute("spt_element_name"); spt.tab.select(element_name); } spt.tab.header_drag_motion = function( evt, bvr, mouse_411) { //var header = bvr.drag_el; var header = bvr.src_el; var dx = mouse_411.curr_x - spt.tab.mouse_pos.x; var dy = mouse_411.curr_y - spt.tab.mouse_pos.y; if (Math.abs(dx) < 20) { spt.tab.dragging = false; return; } spt.tab.dragging = true; header.setStyle("position", "absolute"); header.setStyle("z-index", "100"); header.setStyle("opacity", "1.0"); header.setStyle("left", spt.tab.header_pos.x + dx - 10 ); //header.setStyle("top", spt.tab.header_pos.y + dy ); } spt.tab.header_drag_action = function( evt, bvr, mouse_411) { var header = bvr.src_el; var drag_pos = header.getPosition(); if (spt.tab.dragging == false) return; var headers = spt.tab.get_headers(); for ( var i = headers.length-1; i >= 0; i-- ) { if (headers[i] == header) { continue; } if (headers[i].getStyle("display") == "none") { continue; } var pos = headers[i].getPosition(); var size = headers[i].getSize(); // the y ensures 2nd row tabs don't jump to first row on click if (drag_pos.x > pos.x + size.x/2 && drag_pos.y >= pos.y) { header.inject(headers[i], "after"); break; } if (drag_pos.x > pos.x && drag_pos.y >= pos.y ) { header.inject(headers[i], "before"); break; } } bvr.drag_el.setStyle("position", "relative"); bvr.drag_el.setStyle("z-index", ""); bvr.drag_el.setStyle("top", ""); bvr.drag_el.setStyle("left", ""); bvr.drag_el.setStyle("background", bvr.gradient); var top = spt.tab.top; if (top.hasClass("spt_tab_save_state") ) { spt.tab.save_state(); } } spt.tab.close = function(src_el) { // src_el should be a child of spt_tab_content or spt_tab_header if (!src_el) { spt.error('src_el passed in to spt.tab.close() does not exist.'); return; } spt.tab.top = src_el.getParent(".spt_tab_top"); var top = spt.tab.top; var headers = spt.tab.get_headers(); if (headers.length == 1) { return; } var content = src_el.getParent(".spt_tab_content"); var element_name; // check if it's a header child if (src_el.hasClass("spt_tab_header")) { var header = src_el; } else { var header = src_el.getParent(".spt_tab_header"); } var subheader = src_el.getParent(".spt_tab_subheader"); if (header) { element_name = header.getAttribute("spt_element_name"); content = spt.tab.get_content(element_name); } else if (subheader) { element_name = header.getAttribute("spt_element_name"); content = spt.tab.get_content(element_name); } else if (content) { element_name = content.getAttribute("spt_element_name"); header = spt.tab.get_selected_header(element_name); } if (!header || !content) { spt.error('Tab close cannot find the header or content. Abort'); return; } /* If there are changed elements in the current tab, changedParameters * is a list with index 0 containing changed element, and index 1 containing * change type class. Otherwise, changedParameters is false. */ function ok(changedParameters) { //Remove unsaved changes flags if (changedParameters) { var changed_element = changedParameters[0]; var changed_type = changedParameters[1]; changed_element.removeClass(changed_type); } var opener = header.getAttribute("spt_tab_opener"); var element_name = header.getAttribute("spt_element_name"); if (header) { var subheader = document.id(header.getAttribute("spt_subheader_id")); if (subheader) { var items = subheader.getElements(".spt_tab_subheader_item"); for (var i = 0; i < items.length; i++) { var subheader_element_name = items[i].getAttribute("spt_element_name"); var subheader_content = spt.tab.get_content(subheader_element_name); spt.behavior.destroy_element(subheader_content); } spt.behavior.destroy_element(subheader); } } //header.destroy(); //content.destroy(); spt.behavior.destroy_element(header); spt.behavior.destroy_element(content); //spt.tab.resize_headers(); var last_element_name = spt.tab.get_last_selected_element_name(); last_element_name = null; // make the opener active if (opener) { spt.tab.select(opener); } else if (last_element_name) { spt.tab.select(last_element_name); } else { // select last one from the remaining headers = spt.tab.get_headers(); var last = headers[headers.length - 1].getAttribute("spt_element_name"); spt.tab.select(last); } if (top.hasClass("spt_tab_save_state") ) { spt.tab.save_state(); } } var changed_el = content.getElement(".spt_has_changes"); var changed_row = content.getElement(".spt_row_changed"); if (changed_el) { spt.confirm("There are unsaved changes in the current tab. Continue without saving?", ok, null, {ok_args : [changed_el, "spt_has_changed"]}); } else if (changed_row) { spt.confirm("There are unsaved changes in the current tab. Continue without saving?", ok, null, {ok_args: [changed_row, "spt_row_changed"]}); } else { ok(false); } } spt.tab.view_definition = function(bvr) { var activator = spt.smenu.get_activator(bvr); var header = activator; var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs; if (kwargs_str != '') { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } else { kwargs = {}; } var class_name = header.getAttribute("spt_class_name_decoded"); /* TEST: show widget editor var class_name2 = 'tactic.ui.tools.WidgetEditorWdg'; var kwargs2 = { 'editor_id': bvr.editor_id, 'display_handler': class_name, 'display_options': kwargs, } spt.panel.load_popup("Widget Editor", class_name2, kwargs2); */ var br = '\n'; var xml = ''; var placeholder = "element"; xml += '<' + placeholder + '>' + br; xml += ' <display class="'+class_name+'">' + br; for (var name in kwargs) { if (name == 'class_name') { continue; } xml += ' <'+name+'>'+kwargs[name]+'</'+name+'>' + br; } xml += ' </display>' + br; xml += '</' + placeholder + '>'; var html = spt.convert_to_html_display(xml); spt.alert(html, {type:'html'}); } ''' def get_config_xml(self): return ''' <config> <tab> <element name="untitled" title="(Untitled)"/> </tab> </config> ''' def add_styles(self): if self.use_default_style: palette = self.top.get_palette() border = palette.color("border") color = palette.color("color") background = palette.color("background") data = { 'border': border, 'color': color, 'background': background, 'header_id': self.header_id, } from pyasm.web import HtmlElement style = HtmlElement.style() self.top.add(style) style.add(''' #%(header_id)s .spt_tab_header { border-style: solid; position: relative; border-color: %(border)s; border-width: 1px 1px 0px 1px; padding: 7px 5px; color: %(color)s; background: %(background)s; height: 30px; box-sizing: border-box; } #%(header_id)s .spt_tab_selected { opacity: 1.0; #border-bottom: none; } #%(header_id)s .spt_tab_unselected { opacity: 0.4 ; #border-bottom: solid 1px %(border)s; } #%(header_id)s .spt_tab_hover { } ''' % data) def get_bootstrap_styles(self): styles = HtmlElement.style() color = styles.get_color("color") styles.add(''' .nav-pills .nav-link, .nav-tabs .nav-link { padding: .5em .8575em; font-size: 12px; height: 31px; } .spt_tab_header_top { height: 31px; } .spt_tab_header_top { height: 31px; } .spt_tab_header { display: flex; align-items: center; justify-content: space-between; user-select: none; -webkit-touch-callout: none; /* iOS Safari */ -webkit-user-select: none; /* Safari */ -khtml-user-select: none; /* Konqueror HTML */ -moz-user-select: none; /* Old versions of Firefox */ -ms-user-select: none; /* Internet Explorer/Edge */ } .spt_tab_header_top .spt_tab_selected { height: 31px; background: var(--spt_palette_background); color: #000; } .spt_tab_content_top .spt_tab_header { border-bottom: solid .214rem transparent; } .spt_popup_content .spt_tab_header { border-bottom: solid .214rem transparent; background: inherit; color: inherit; } .spt_tab_content_top .spt_tab_header_top .spt_tab_selected { border-bottom: solid .214rem var(--spt_palette_md_secondary); background: inherit; color: inherit; } .spt_popup_content .spt_tab_header_top .spt_tab_selected { border-bottom: solid .214rem var(--spt_palette_md_secondary); background: inherit; color: inherit; } .spt_tab_header_top .spt_tab_selected .nav-link { color: #fff; border-bottom: 0px solid transparent; } .spt_tab_header_top .spt_tab_selected .nav-link:hover { color: #fff; border-bottom: 0px solid transparent; } .nav-tabs .nav-link { border-bottom: .214rem solid transparent; } .nav-tabs .nav-link:hover { border-bottom: .214rem solid transparent; } .spt_tab_remove { display: none; position: absolute; right: 3px; } .spt_tab_selected .spt_tab_remove { display: block; } .spt_tab_selected .spt_tab_remove { display: block; } .spt_tab_header:hover .spt_tab_remove { display: block; } .spt_tab_header_label_container { display: flex; align-items: center; } .spt_tab_header_label { color: %s; text-overflow: ellipsis; overflow-x: hidden; white-space: nowrap; } .spt_tab_selected .spt_tab_header_label { color: %s !important; } .spt_tab_header_count { margin-left: 10px; } ''' % (color, color)) return styles def get_styles(self): styles = HtmlElement.style(''' /* NEW */ .spt_tab_header_top { white-space: nowrap; height: auto; float: left; position: relative; z-index: 1; margin-bottom: -1px; width: 100%; overflow: hidden; text-align: left; box-sizing: border-box; } .spt_tab_header { vertical-align: top; margin-right: 1px; box-sizing: border-box; overflow: hidden; display: inline-block; } .spt_tab_header.rounded-top-corners { border-top-left-radius: 5px; border-top-right-radius: 5px; } .spt_tab_header.drag-header { position: relative; } .spt_tab_header_label { text-align: left; overflow: hidden; text-overflow: ellipsis; white-space: nowrap; z-index: 1; } .spt_add_tab { margin-left: -2px; display: inline-block; } .spt_add_tab_inner { padding: 0px 2px 0px 2px; border-top-right-radius: 12px; opacity: 0.5; background: linear-gradient(180deg, #f2f2f2, #FFFFFF); float: left; margin-top: 2px; padding-top: 4px; height: 21px; width: 22px; margin-left: 4px; border-style: solid; border-width: 1px 1px 0px 1px; border-color: #BBB; text-align: center; } .spt_tab_remove { position: absolute; right: 2px; top: 8px; z-index: 2; width: 16px; height: 16px; padding-left: 2px; border-radius: 10px; box-sizing: border-box; color: #000; background: #FFF; border: solid 1px transparent; } .spt_icon_active { margin: auto; position: absolute; top: 0; bottom: 0; max-height: 100%; opacity: 0.3; } .spt_tab_content_top { border-bottom-left-radius: 5px; color: #000; border-bottom-right-radius: 5px; margin-top: -1px; min-height: 500px; width: 100%; z-index: 1; background: var(--spt_palette_background); border-top-right-radius: 5px; border: 1px solid #BBB; } ''') return styles def get_tab_id(self): return self.unique_id def get_header_id(self): return self.header_id def get_display(self): top = self.top top.add_class("spt_tab_top") if self._use_bootstrap(): top.add(self.get_bootstrap_styles()) else: top.add(self.get_styles()) self.search_type = None self.view = self.kwargs.get("view") config_xml = self.kwargs.get("config_xml") config = self.kwargs.get("config") # save state overrides saved_config_xml = None self.save_state = self.kwargs.get("save_state") if self.save_state in [True, 'true']: self.save_state = "save_state|main_tab" if self.save_state: saved_config_xml = WidgetSettings.get_value_by_key(self.save_state) if saved_config_xml: if not config_xml: config_xml = saved_config_xml else: saved_xml_data = Xml() saved_xml_data.read_string(saved_config_xml) saved_xml_root = saved_xml_data.get_root_node() saved_xml_tab = Xml.get_first_child(saved_xml_root) nodes = Xml.get_children(saved_xml_tab) xml_data = Xml() xml_data.read_string(config_xml) xml_root = xml_data.get_root_node() xml_tab = Xml.get_first_child(xml_root) curr_nodes = Xml.get_children(xml_tab) curr_node_names = [Xml.get_attribute(node, "name") for node in curr_nodes] for node in nodes: node_name = Xml.get_attribute(node, "name") if node_name not in curr_node_names: Xml.append_child(xml_tab, node) config_xml = xml_data.get_xml() saved_config_xml = None top.add_class("spt_tab_save_state") top.add_attr("spt_tab_save_state", self.save_state) self.mode = self.kwargs.get('mode') if not self.mode: self.mode = "default" if self.view and self.view != 'tab' and not config_xml: config = None # if it is not defined in the database, look at a config file includes = self.kwargs.get("include") if includes: includes = includes.split("|") for include in includes: tmp_path = __file__ dir_name = os.path.dirname(tmp_path) file_path="%s/../config/%s" % (dir_name, include) config = WidgetConfig.get(file_path=file_path, view=self.view) if config and config.has_view(self.view): pass else: config = None if not config: search = Search("config/widget_config") search.add_filter("category", "TabWdg") search.add_filter("view", self.view) config_sobj = search.get_sobject() if not config_sobj: config_xml = "<config><%s></%s></config>" % (self.view, self.view) else: config_xml = config_sobj.get_value("config") config = WidgetConfig.get(view=self.view, xml=config_xml) else: if saved_config_xml: # this is for custom config_xml with a matching custom view if not self.view: self.view = 'tab' config = WidgetConfig.get(view=self.view, xml=saved_config_xml) elif config: pass elif config_xml: # this is for custom config_xml with a matching custom view if not self.view: self.view = 'tab' config = WidgetConfig.get(view=self.view, xml=config_xml) elif self.widgets: config_xml = ''' <config> <tab></tab> </config> ''' else: config_xml = ''' <config> <tab> <element name="untitled" title="(Untitled)"/> </tab> </config> ''' self.view = 'tab' config = WidgetConfig.get(view=self.view, xml=config_xml) element_names = None if self.save_state and config: element_names = config.get_element_names() if not element_names: element_names = self.kwargs.get("element_names") if element_names and isinstance(element_names, basestring): element_names = element_names.split(",") if not element_names and config: element_names = config.get_element_names() if not element_names: element_names = [] #top.add_style("padding: 10px") top.set_id(self.unique_id) top.set_attr("spt_tab_id", self.unique_id) top.set_attr("spt_tab_mode", self.mode) gradient = top.get_gradient("background", -5, 5) inner = DivWdg() top.add(inner) inner.add_style("position: relative") inner.add_style("width: auto") if not Container.get_dict("JSLibraries", "spt_tab"): inner.add_behavior( { 'type': 'load', 'gradient': gradient, 'cbjs_action': self.get_onload_js() } ) header_div = self.header header_id = self.get_header_id() inner.add(header_div) header_div.add_class("spt_tab_header_top") header_div.add_class("nav nav-tabs") subheader_div = DivWdg() subheader_div.add_class("spt_tab_subheader_top") subheader_div.add_class("SPT_TEMPLATE") inner.add(subheader_div) self.add_subheader_behaviors(subheader_div) if not self._use_bootstrap(): self.add_styles() # if a search_key has been passed in, add it to the state. state = self.kwargs.get("state") if not state: state = self.kwargs search_key = self.kwargs.get("search_key") if search_key: state['search_key'] = search_key selected = self.kwargs.get("selected") if not selected: if element_names: selected = element_names[0] else: selected = '' offset = self.kwargs.get("tab_offset") if offset: header_div.add_style("padding-left: %s" % offset) if self.mode == "hidden": # TODO: The display is problematic because bmd is using !important header_div.add_style("display: none") header_div.add_style("visibility: hidden") header_div.add_style("height: 0px") header_defs = {} title_dict = {} self.add_context_menu( header_div ) min_width = self.kwargs.get("min_width") if min_width: try: min_width = int(min_width) min_width = str(min_width) + "px" except ValueError: pass header_div.add_style("min-width", min_width) resize_headers = self.kwargs.get("resize_headers") if resize_headers: offset = 120 header_div.add_behavior( { 'type': 'load', 'offset': offset, 'cbjs_action': ''' var top = bvr.src_el; top.spt_last_width = 0; var offset = bvr.offset; setInterval( function() { if (!top.isVisible() ) { return; } var size = top.getSize(); if (size.x == top.spt_last_width) { return; } top.spt_last_width = size.x; var els = bvr.src_el.getElements(".spt_tab_header"); var count = els.length; var width = parseInt((size.x-offset) / (count)); if (width > 120) { width = 120; } if (width < 30) { width = 30; } for (var i = 0; i < els.length; i++) { els[i].setStyle("width", width + "px"); } }, 250); ''' } ) show_remove = self.kwargs.get("show_remove") # resize headers on leave resize_headers = self.kwargs.get("resize_headers") resize_headers = True if resize_headers: header_div.add_behavior( { 'type': 'mouseleave', 'cbjs_action': ''' spt.tab.resize_headers(); ''' } ) if show_remove == "hover": header_div.add_relay_behavior( { 'type': 'mouseenter', 'bvr_match_class': 'spt_tab_header', 'cbjs_action': ''' var el = bvr.src_el.getElement(".spt_tab_remove"); el.setStyle("display", ""); ''' } ) header_div.add_relay_behavior( { 'type': 'mouseleave', 'bvr_match_class': 'spt_tab_header', 'cbjs_action': ''' var el = bvr.src_el.getElement(".spt_tab_remove"); el.setStyle("display", "none"); ''' } ) header_div.add_relay_behavior( { 'type': 'mouseenter', 'bvr_match_class': 'spt_tab_remove', 'cbjs_action': ''' var active = bvr.src_el.getElement(".spt_icon_active"); active.setStyle("opacity", 1.0); ''' } ) header_div.add_relay_behavior( { 'type': 'mouseleave', 'bvr_match_class': 'spt_tab_remove', 'cbjs_action': ''' var active = bvr.src_el.getElement(".spt_icon_active"); active.setStyle("opacity", 0.3); ''' } ) loaded_dict = {} for element_name in element_names: attrs = config.get_element_attributes(element_name) title = attrs.get("title") if not title: title = Common.get_display_title(element_name) title = _(title) if attrs.get("display") == "false": continue load_now = attrs.get('load') is_loaded = load_now =='true' loaded_dict[element_name] = is_loaded display_class = config.get_display_handler(element_name) display_options = config.get_display_options(element_name) header_defs[element_name] = { 'display_class': display_class, 'display_options': display_options } # FIXME: this is already defined in get_display_options # process the display options for name, value in display_options.items(): # so it allows JSON string to pass thru without eval as expression if re.search("^{[@$]", value) and re.search("}$", value): value = Search.eval(value, state=state) display_options[name] = value # DEPRECATED: this should not really be used. It is likely # better to use expressions set above to explicitly set # the values if display_options.get("use_state") in [True, 'true']: # add the state items to the display options for state_name, state_value in state.items(): display_options[state_name] = state_value if element_name == selected: is_selected = True else: is_selected = False header = self.get_tab_header(element_name, title, display_class, display_options, is_selected=is_selected, is_loaded=is_loaded, is_template=False, attrs=attrs) header_div.add(header) # add widgets that have been manually added for i, widget in enumerate(self.widgets): name = widget.get_name() if not name: num = Common.randint(0, 10000) name = "noname%s" % num widget.set_name(name) title = "(Untitled)" else: title = Common.get_display_title(name) if not title: title = "(Untitled)" title_dict[name] = title if name == selected: is_selected = True else: is_selected = False class_name = Common.get_full_class_name(widget) if isinstance(widget, BaseRefreshWdg): kwargs = widget.get_kwargs() else: kwargs = {} header = self.get_tab_header(name, title, class_name, kwargs, is_selected=is_selected, is_loaded=True, is_template=False) header_div.add(header) show_add = self.kwargs.get("show_add") not in [False, "false"] if show_add: header_div.add( self.get_add_wdg() ) content_top = DivWdg() content_top.add_class("spt_tab_content_top") content_top.add_attr("spt_tab_id", self.get_tab_id()) inner.add(content_top) resize_offset = self.kwargs.get("resize_offset") resize_attr = self.kwargs.get("resize_attr") or "height" if resize_offset != None: content_top.add_class("spt_window_resize") content_top.add_attr("spt_window_resize_offset", resize_offset) content_top.add_attr("spt_window_resize_attr", resize_attr) #content_top.add_style("overflow: auto") content_top.add_style("overflow: auto") else: height = self.kwargs.get("height") if height: try: height = int(height) height = str(height) + "px" except ValueError: pass content_top.add_style("height: %s" % height) content_top.add_style("overflow-y: auto") content_top.add_style("min-height: %s" % height) width = self.kwargs.get("width") if not width: width = self.kwargs.get("min_width") if width: try: width = int(width) width = str(width) + "px" except ValueError: pass content_top.add_style("min-width: %s" % width) content_top.add_class("tab_content_top") color_mode = self.kwargs.get("color_mode") if color_mode == "transparent": pass else: content_top.add_color("color", "color") # content_top.add_color("background", "background") # put in a content box for each element for element_name in element_names: content_div = DivWdg() content_top.add(content_div) content_div.add_class("spt_tab_content") content_div.add_attr("spt_tab_id", self.unique_id) content_div.add_attr("spt_element_name", element_name) resize_offset = self.kwargs.get("resize_offset") if resize_offset != None: content_div.add_class("spt_window_resize") content_div.add_attr("spt_window_resize_offset", resize_offset) content_div.add_style("overflow: auto") content_div.add_style("width: 100%") content_div.add_style("text-align: left") is_loaded = loaded_dict.get(element_name) if element_name == selected or is_loaded: header_def = header_defs[element_name] display_class = header_def.get("display_class") if not display_class: widget = DivWdg() widget.add_color("background", "background") widget.add_style("height: 300px") widget.add_style("padding-top: 50px") inner = DivWdg() widget.add(inner) inner.add_style("margin-left: auto") inner.add_style("margin-right: auto") inner.add_style("width: 500px") inner.add_style("height: 100px") inner.add_border() inner.add_style("text-align: center") inner.add_style("padding-top: 50px") inner.add_color("color", "color3") inner.add_color("background", "background3") inner.add( IconWdg("WARNING", IconWdg.WARNING) ) inner.add(" <b>Nothing to display</b>") else: display_options = header_def.get("display_options") widget = Common.create_from_class_path(display_class, kwargs=display_options) content_div.add(widget) content_div.add_class("spt_content_loaded") if is_loaded and element_name != selected: # hide preloaded tabs or non-selected tabs content_div.add_style("display: none") else: content_div.add("&nbsp;") content_div.add_style("display: none") for widget in self.widgets: name = widget.get_name() content_div = DivWdg() content_div.add_class("spt_tab_content") content_div.add_attr("spt_tab_id", self.unique_id) content_div.add_class("spt_content_loaded") content_div.add_attr("spt_element_name", name) resize_offset = self.kwargs.get("resize_offset") if resize_offset != None: content_div.add_class("spt_window_resize") content_div.add_attr("spt_window_resize_offset", resize_offset) content_div.add_style("overflow: auto") title = title_dict.get(name) content_div.add_attr("spt_title", title) if name != selected: content_div.add_style("display: none") content_div.add(widget) content_div.add_style("width: 100%") # content_div.add_style("height: 100%") content_div.add_style("text-align: left") content_top.add(content_div) # Add in a template template_div = DivWdg() template_div.add_class("spt_tab_template_top") template_div.add_style("display: none") template_div.add_class("SPT_TEMPLATE") name = "" title = "" is_selected = False header = self.get_tab_header(name, title, None, None, is_selected=is_selected, is_template=True) template_div.add(header) # subheader test subheader = self.get_tab_subheader(name, title, None, None, is_selected=is_selected, is_template=True, config=config) template_div.add(subheader) subheader.add_style("z-index: 3") #header.add_behavior( { header.add_relay_behavior( { 'type': 'click', 'bvr_match_class': 'spt_tab_header', 'cbjs_action': ''' var header_top = bvr.src_el.getParent(".spt_tab_header_top"); var top = bvr.src_el.getParent(".spt_tab_top"); var subheader_id = bvr.src_el.getAttribute("spt_subheader_id") var subheaders = top.getElements(".spt_tab_subheader"); for ( var i = 0; i < subheaders.length; i++) { subheaders[i].setStyle("display", "none"); } var el = document.id(subheader_id); var items = el.getElements(".spt_tab_subheader_item"); if (items.length == 0) { return; } var size = bvr.src_el.getSize(); var pos = bvr.src_el.getPosition(header_top); if (el) { el.setStyle("display", ""); spt.body.add_focus_element(el); el.position({x: pos.x, y: pos.y+size.y-1}, el); } ''' } ) top.add(template_div) content_div = DivWdg() content_div.add_class("spt_tab_content") content_div.add_attr("spt_element_name", "NEW") content_div.add_attr("spt_tab_id", self.unique_id) content_div.add("") content_div.add_style("width: 100%") # content_div.add_style("height: 100%") content_div.add_style("text-align: left") template_div.add(content_div) return top def get_add_wdg(self): style = HtmlElement.style(''' .spt_add_tab { display: flex; align-items: center; padding-left: 10px; } ''') div = DivWdg() div.add(style) div.add_class("spt_add_tab") icon_div = DivWdg() icon_div.add_class("spt_add_tab_inner") from tactic.ui.widget import IconButtonWdg, ButtonNewWdg icon = ButtonNewWdg(title="New Tab", icon="FA_PLUS", width=20) #icon = IconWdg("New Tab", IconWdg.PLUS) #icon.add_style("top: -1px") #icon.add_style("left: 0px") #icon.add_style("position: absolute") icon.add_style("display: block") icon_div.add_class("hand") icon_div.add(icon) add_bvr = self.kwargs.get("add_bvr") if not add_bvr: add_bvr = """ spt.tab.top = bvr.src_el.getParent(".spt_tab_top"); spt.tab.add_new(); """ icon_div.add_behavior( { 'type': 'click_up', 'cbjs_action': add_bvr } ) div.add(icon_div) self.extra_menu = self.kwargs.get("extra_menu") if self.extra_menu: icon_div = DivWdg() icon_div.set_round_corners(3, corners=['TR']) from tactic.ui.widget import IconButtonWdg icon = IconWdg("More Options", IconWdg.ARROWHEAD_DARK_DOWN) icon.add_style("margin-left: -2px") icon_div.add(icon) from smart_menu_wdg import SmartMenu smenu_set = SmartMenu.add_smart_menu_set( icon_div, { 'BUTTON_MENU': self.extra_menu } ) SmartMenu.assign_as_local_activator( icon_div, "BUTTON_MENU", True ) icon_div.add_style("padding-top: 4px") icon_div.add_style("margin-top: 10px") icon_div.add_style("float: left") icon_div.add_style("height: 16px") icon_div.add_style("width: 10px") icon_div.add_style("margin-left: -1px") icon_div.add_gradient("background", "background", -5, 5) icon_div.add_border() icon_div.add_style("text-align: center") icon_div.add_style("opacity: 0.5") div.add(icon_div) return div def get_edit_wdg(self): div = DivWdg() div.add_style("margin-left: -2px") icon_div = DivWdg() icon_div.add_style("padding: 0px 2px 0px 2px") icon_div.set_round_corners(3, corners=['TR','TL']) from tactic.ui.widget import IconButtonWdg icon = IconButtonWdg(title="New Tab", icon=IconWdg.EDIT) icon = IconWdg("Edit Tab Definition", IconWdg.EDIT) icon.add_style("margin-top: -1px") icon.add_style("margin-left: 1px") icon_div.add_class("hand") icon_div.add(icon) icon.add_behavior( { 'type': 'click_up', 'cbjs_action': ''' var class_name = 'tactic.ui.tools.tab_edit_wdg.TabEditWdg'; var kwargs = {} spt.panel.load_popup("Tab Edit", class_name, kwargs) ''' } ) icon_div.add_style("padding-top: 4px") icon_div.add_style("float: left") icon_div.add_style("height: 20px") icon_div.add_style("width: 18px") icon_div.add_style("margin-left: 2px") icon_div.add_gradient("background", "background", -5, 5) icon_div.add_border() icon_div.add_style("text-align: center") div.add(icon_div) return div def add_context_menu(self, header_div): from .menu_wdg import Menu, MenuItem menu = Menu(width=180) #menu.set_allow_icons(False) #menu.set_setup_cbfn( 'spt.tab.smenu_ctx.setup_cbk' ) menu_item = MenuItem(type='title', label='Actions') menu.add(menu_item) menu_item = MenuItem(type='action', label='Reload Tab') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var header = activator; var class_name = header.getAttribute("spt_class_name"); var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs; if (kwargs_str != '') { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } else { kwargs = {}; } var values = null; var force = true; spt.tab.load_class(header, class_name, kwargs, values, force); ''' } ) menu.add(menu_item) menu_item = MenuItem(type='separator') menu.add(menu_item) menu_item = MenuItem(type='action', label='Rename Tab') menu_item.add_behavior( { 'cbjs_action': ''' var class_name = 'tactic.ui.container.TabRenameWdg'; var kwargs = {}; var activator = spt.smenu.get_activator(bvr); var label = activator.getElement(".spt_tab_header_label"); name = label.innerHTML; title = "Raname Tab ["+name+"]"; var popup = spt.panel.load_popup(title, class_name, kwargs); popup.activator = activator; ''' } ) menu.add(menu_item) """ menu_item = MenuItem(type='action', label='New Tab') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; spt.tab.add_new(); ''' } ) menu.add(menu_item) """ menu_item = MenuItem(type='action', label='Tear Off') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); // add new if this is the last oni var headers = spt.tab.get_headers(); if (headers.length == 1) { spt.tab.add_new(); } spt.tab.top = top; var header = activator; var element_name = header.getAttribute("spt_element_name"); spt.behavior.destroy_element(header); var contents = spt.tab.get_contents(); for (var i=0; i<contents.length; i++) { var content = contents[i]; if (content.getAttribute("spt_element_name") == element_name) { spt.panel.load_popup_with_html( element_name, content.innerHTML ); spt.behavior.destroy_element(content); } } ''' } ) menu.add(menu_item) menu_item = MenuItem(type='action', label='Copy To Main Tab') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var html; var header = activator; var element_name = header.getAttribute("spt_element_name"); var title = header.getAttribute("spt_title"); var class_name = header.getAttribute("spt_class_name"); var kwargs_str = header.getAttribute("spt_kwargs"); var kwargs = {}; if (kwargs_str) { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } var contents = spt.tab.get_contents(); for (var i=0; i<contents.length; i++) { var content = contents[i]; if (content.getAttribute("spt_element_name") == element_name) { html = content.innerHTML; break; } } spt.tab.set_main_body_tab(); spt.tab.add_new(element_name, title, class_name, kwargs); ''' } ) menu.add(menu_item) if self.kwargs.get("show_remove") not in ['false', False]: menu_item = MenuItem(type='separator') menu.add(menu_item) menu_item = MenuItem(type='action', label='Close Tab') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var header = activator; var element_name = header.getAttribute("spt_element_name"); spt.behavior.destroy_element(header); var contents = top.getElements(".spt_tab_content"); for (var i=0; i<contents.length; i++) { var content = contents[i]; if (content.getAttribute("element_name") == element_name) { spt.behavior.destroy_element(content); } } ''' } ) menu.add(menu_item) menu_item = MenuItem(type='action', label='Close All Except This Tab') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var headers = spt.tab.get_headers(); for (var i=0; i < headers.length; i++) { var element_name = headers[i].getAttribute("spt_element_name"); if (activator.getAttribute('spt_element_name') != element_name) { spt.tab.close(headers[i]); } } var element_name = activator.getAttribute("spt_element_name"); spt.tab.select(element_name); ''' } ) menu.add(menu_item) security = Environment.get_security() if security.check_access("builtin", "view_site_admin", "allow"): menu_item = MenuItem(type='separator') menu.add(menu_item) menu_item = MenuItem(type='action', label='View Definition') menu_item.add_behavior( { 'cbjs_action': '''spt.tab.view_definition(bvr);''' } ) menu.add(menu_item) menu_item = MenuItem(type='action', label='Add to Side Bar') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var header = activator; var element_name = header.getAttribute("spt_element_name"); var title = header.getAttribute("spt_title"); var kwargs = header.getAttribute("spt_kwargs"); kwargs = kwargs.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs); var view = element_name; var element_name = element_name.replace(/ /g, "_"); element_name = element_name.replace(/\//g, "_"); var kwargs = { class_name: 'LinkWdg', display_options: kwargs, element_attrs: { title: title } } try { var server = TacticServerStub.get(); var info = server.add_config_element("SideBarWdg", "definition", element_name, kwargs); var info = server.add_config_element("SideBarWdg", "project_view", element_name, kwargs); spt.panel.refresh("side_bar"); } catch(e) { alert(e); throw(e); } ''' } ) menu.add(menu_item) has_my_views = True if has_my_views: menu_item = MenuItem(type='action', label='Add to My Views') menu_item.add_behavior( { 'cbjs_action': ''' var activator = spt.smenu.get_activator(bvr); var top = activator.getParent(".spt_tab_top"); spt.tab.top = top; var header = activator; var element_name = header.getAttribute("spt_element_name"); var title = header.getAttribute("spt_title"); var kwargs = header.getAttribute("spt_kwargs"); kwargs = kwargs.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs); var login = 'admin'; var class_name = kwargs.class_name; if (!class_name) { class_name = "tactic.ui.panel.CustomLayoutWdg"; } var view = element_name; var element_name = element_name.replace(/ /g, "_"); element_name = element_name.replace(/\//g, "_"); element_name = login + "." + element_name; var kwargs = { class_name: class_name, display_options: kwargs, element_attrs: { title: title }, login: login, unique: false, } var view = "self_view_" + login; try { var server = TacticServerStub.get(); var info = server.add_config_element("SideBarWdg", "definition", element_name, kwargs); var info = server.add_config_element("SideBarWdg", view, element_name, kwargs); spt.panel.refresh("side_bar"); } catch(e) { alert(e); throw(e); } ''' } ) menu.add(menu_item) if self.kwargs.get("show_context_menu") not in ['false', False]: menus = [menu.get_data()] menus_in = { 'DG_HEADER_CTX': menus, } from .smart_menu_wdg import SmartMenu SmartMenu.attach_smart_context_menu( header_div, menus_in, False ) def get_tab_header(self, element_name, title, class_name=None, kwargs=None, is_selected=False, is_loaded=False, is_template=False, attrs={}): web = WebContainer.get_web() #header = DivWdg() header = HtmlElement.li() header.add_class("spt_tab_header") header.add_class("nav-item") header.add_attr("spt_tab_id", self.unique_id) header.add_class("hand") header.add_style("width: 120px") header.add_behavior({ 'type': 'load', 'cbjs_action': '''$(bvr.src_el).bmdRipples();''' }) if self.use_default_style: header.add_class("rounded-top-corners") if is_selected: header.add_class("spt_tab_selected") header.add_class("spt_is_selected") else: header.add_class("spt_tab_unselected") count = attrs.get("count") header.add_attr("spt_element_name", element_name) header.add_attr("spt_title", title) if not is_template: if kwargs: if count: kwargs['count'] = count kwargs_str = Common.convert_to_json(kwargs) header.add_attr("spt_kwargs", kwargs_str) # Disable widget key for now #widget_key = header.generate_widget_key(class_name, inputs=kwargs) #header.add_attr("spt_class_name", widget_key) header.add_attr("spt_class_name", class_name) # For viewing definition with widget_key header.add_attr("spt_class_name_decoded", class_name) else: header.add_attr("spt_kwargs", '') widget_key = None header.add_behavior( { 'type': 'click_up', 'cbjs_action': ''' var header = bvr.src_el; spt.tab.top = header.getParent(".spt_tab_top"); var this_name = header.getAttribute('spt_element_name'); spt.tab.select(this_name); ''' } ) from .smart_menu_wdg import SmartMenu SmartMenu.assign_as_local_activator( header, 'DG_HEADER_CTX' ) title_container = DivWdg(css="spt_tab_header_label_container") title_container.add_class("nav-link") show_remove = self.kwargs.get("show_remove") if show_remove not in [False, "false"]: title_container.add_style("width: 95%") else: title_container.add_style("width: 100%") header.add(title_container) title_div = DivWdg() title_div.add_class("spt_tab_header_label") display_title = title title_div.add(display_title) title_container.add(title_div) title_div.add_attr("title", "%s" % (title)) count_wdg = SpanWdg() count_wdg.add_class("badge badge-secondary spt_tab_header_count") title_container.add(count_wdg) icon = None if icon: icon = IconWdg(name="whatever", icon=icon) title_div.add(icon) if count: count_color = attrs.get("count_color") state = self.kwargs.get("state") or {} search_key = state.get("search_key") if not search_key: search_key = self.kwargs.get("search_key") if search_key: sobject = Search.get_by_search_key(search_key) else: sobject = None if sobject: value = Search.eval(count, sobject) count_wdg.add(value) if count_color: count_wdg.add_style("background", count_color) count_wdg.add_update( { 'expression': count, 'expr_key': search_key, } ) remove_wdg = DivWdg() remove_wdg.add_class("spt_tab_remove") remove_wdg.add_behavior( { 'type': 'click', 'cbjs_action': ''' spt.tab.close(bvr.src_el); ''' } ) show_remove = self.kwargs.get("show_remove") show_remove_tabs = self.kwargs.get("show_remove_tabs") if show_remove_tabs: show_remove_tabs_list = show_remove_tabs.split("|") if element_name in show_remove_tabs_list: show_remove = False if is_template or show_remove not in [False, 'false']: header.add(remove_wdg) if show_remove == "hover": remove_wdg.add_style("display: none") remove_icon_path = self.kwargs.get("remove_icon_path") if (remove_icon_path): icon = HtmlElement.img(remove_icon_path) icon.add_styles("padding: 2px; width: 11px") remove_wdg.add_style("right: 6px;") else: from tactic.ui.widget import ButtonNewWdg icon = ButtonNewWdg(icon="FA_TIMES", size=12, width=20) icon.add_class("spt_icon_active") remove_wdg.add(icon) divider_div = DivWdg() divider_div.add_style("width: 1px") divider_div.add_style("background: #fff") divider_div.add_style("height: 20px") header.add(divider_div) # add a drag behavior allow_drag = self.kwargs.get("allow_drag") if allow_drag not in [False, 'false']: header.add_class("drag-header") header.add_behavior( { 'type': 'drag', "drag_el": '@', "cb_set_prefix": 'spt.tab.header_drag' } ) return header def get_tab_subheader(self, element_name, title, class_name=None, kwargs=None, is_selected=False, is_loaded=False, is_template=False, config=None): subheader_div = DivWdg() subheader_div.add_class("spt_tab_subheader") subheader_div.add_style("width: 200px") subheader_div.add_style("height: auto") subheader_div.add_border() subheader_div.add_style("position: absolute") subheader_div.add_style("left: 5px") subheader_div.add_color("background", "background") subheader_div.add_style("top: 28px") subheader_div.add_style("padding: 10px 5px") #element_names = ['my_tasks','all_orders','all_deliverables'] element_names = [] for element_name in element_names: attrs = config.get_element_attributes(element_name) title = attrs.get("title") if not title: title = Common.get_display_title(element_name) subheader = DivWdg() subheader.add_style("position: relative") subheader.add_attr("spt_element_name", element_name) subheader.add_class("spt_tab_subheader_item") icon = IconWdg("Remove Tab", "FA_TIMES", opacity=0.3) subheader.add(icon) icon.add_class("spt_icon_inactive") icon.add_styles("position: absolute; right: 0; top: 3px;") subheader_div.add( subheader ) subheader.add_style("padding: 5px") subheader.add(title) display_class = config.get_display_handler(element_name) display_options = config.get_display_options(element_name) """ subheader.add_behavior( { 'type': 'click', 'title': title, 'display_class': display_class, 'display_options': display_options, 'cbjs_action': ''' spt.panel.load_popup(bvr.title, bvr.display_class, bvr.display_options); ''' } ) subheader.add_behavior( { 'type': 'mouseenter', 'cbjs_action': ''' bvr.src_el.setStyle("background", "#DDD"); ''' } ) subheader.add_behavior( { 'type': 'mouseleave', 'cbjs_action': ''' bvr.src_el.setStyle("background", ""); ''' } ) """ return subheader_div def add_subheader_behaviors(self, subheader_top): subheader_top.set_unique_id() subheader_top.add_smart_style("spt_tab_subheader_item", "pointer", "cursor") subheader_top.add_relay_behavior( { 'type': 'click', 'bvr_match_class': 'spt_tab_subheader_item', 'cbjs_action': ''' var element_name = bvr.src_el.getAttribute("spt_element_name"); var title = bvr.src_el.getAttribute("spt_title"); var display_class = bvr.src_el.getAttribute("spt_class_name"); var kwargs_str = bvr.src_el.getAttribute("spt_kwargs"); if (!kwargs_str) { kwargs = {} } else { kwargs_str = kwargs_str.replace(/\&amp;quot\;/g, '"'); kwargs = JSON.parse(kwargs_str); } spt.tab.load_selected(element_name, title, display_class, kwargs); ''' } ) subheader_top.add_relay_behavior( { 'type': 'mouseenter', 'bvr_match_class': 'spt_tab_subheader_item', 'cbjs_action': ''' bvr.src_el.setStyle("background", "#DDD"); ''' } ) subheader_top.add_relay_behavior( { 'type': 'mouseleave', 'bvr_match_class': 'spt_tab_subheader_item', 'cbjs_action': ''' bvr.src_el.setStyle("background", ""); ''' } ) subheader_top.add_relay_behavior( { 'type': 'mouseleave', 'bvr_match_class': 'spt_tab_subheader', 'cbjs_action': ''' bvr.src_el.setStyle("display", "none"); ''' } ) __all__.append("TabRenameWdg") class TabRenameWdg(BaseRefreshWdg): def get_display(self): top = self.top top.add_style("margin: 20px") top.add_style("min-width: 250px") top.add_class("spt_tab_rename_top") top.add("<div>New Name:</div>") from tactic.ui.input import TextInputWdg from tactic.ui.widget import ActionButtonWdg text = TextInputWdg(name="new_name") text.add_class("spt_tab_new_name") top.add(text) text.add_behavior( { 'type': 'load', 'cbjs_action': 'bvr.src_el.focus()' } ) top.add("<br/>") button = ActionButtonWdg(title="Rename", color="primary") top.add(button) button.add_style("float: right") button.add_behavior( { 'type': 'click', 'cbjs_action': ''' var popup = bvr.src_el.getParent(".spt_popup"); var activator = popup.activator var rename_top = bvr.src_el.getParent(".spt_tab_rename_top"); var input = rename_top.getElement(".spt_tab_new_name"); new_name = input.value spt.popup.close(popup); var label = activator.getElement(".spt_tab_header_label"); label.innerHTML = new_name; label.setAttribute("title", new_name); activator.setAttribute("spt_title", new_name); var top = spt.tab.top; if (!top) { spt.tab.set_main_body_tab(); top = spt.tab.top; } if (top.hasClass("spt_tab_save_state") ) { spt.tab.save_state(); } ''' } ) top.add("<br clear='all'/>") return top from pyasm.command import Command class TabSaveStateCmd(Command): def __init__(self, **kwargs): super(TabSaveStateCmd, self).__init__(**kwargs) self.update = True def execute(self): class_names = self.kwargs.get("class_names") attrs_list = self.kwargs.get("attrs_list") kwargs_list = self.kwargs.get("kwargs_list") save_state = self.kwargs.get("save_state") xml = Xml() xml.create_doc("config") root = xml.get_root_node() view = xml.create_element("tab") xml.append_child(root, view) for class_name, attrs, kwargs in zip(class_names, attrs_list, kwargs_list): element = xml.create_element("element") xml.append_child(view, element) for key, value in attrs.items(): xml.set_attribute(element, key, value) display = xml.create_element("display") xml.append_child(element, display) if class_name.startswith("$"): from pyasm.common import jsonloads key = class_name key = key.lstrip("$") tmp_dir = Environment.get_tmp_dir(include_ticket=True) path = "%s/%s_key_%s.txt" % (tmp_dir, "widget", key) if not os.path.exists(path): print("ERROR: %s path [%s] not found" % ("widget", path)) raise Exception("widget key not valid") f = open(path, 'r') data = f.read() f.close() data = jsonloads(data) class_name = data.get("method") xml.set_attribute(display, "class", class_name) for key, value in kwargs.items(): if (key == 'count'): xml.set_attribute(element, key, value) else: attr = xml.create_text_element(key, value) xml.append_child(display, attr) xml_string = xml.to_string() WidgetSettings.set_value_by_key(save_state, xml_string)
src/tactic/ui/container/tab_wdg.py
93,668
Copyright (c) 2005, Southpaw Technology All Rights Reserved PROPRIETARY INFORMATION. This software is proprietary to Southpaw Technology, and is not to be reproduced, transmitted, or disclosed in any way without written permission. save state overrides if it is not defined in the database, look at a config file this is for custom config_xml with a matching custom view this is for custom config_xml with a matching custom viewtop.add_style("padding: 10px") if a search_key has been passed in, add it to the state. TODO: The display is problematic because bmd is using !important resize headers on leave FIXME: this is already defined in get_display_options process the display options so it allows JSON string to pass thru without eval as expression DEPRECATED: this should not really be used. It is likely better to use expressions set above to explicitly set the values add the state items to the display options add widgets that have been manually addedcontent_top.add_style("overflow: auto") content_top.add_color("background", "background") put in a content box for each element hide preloaded tabs or non-selected tabs content_div.add_style("height: 100%") Add in a template subheader testheader.add_behavior( { content_div.add_style("height: 100%")icon = IconWdg("New Tab", IconWdg.PLUS)icon.add_style("top: -1px")icon.add_style("left: 0px")icon.add_style("position: absolute")menu.set_allow_icons(False)menu.set_setup_cbfn( 'spt.tab.smenu_ctx.setup_cbk' )header = DivWdg() Disable widget key for nowwidget_key = header.generate_widget_key(class_name, inputs=kwargs)header.add_attr("spt_class_name", widget_key) For viewing definition with widget_key add a drag behaviorelement_names = ['my_tasks','all_orders','all_deliverables']
1,761
en
0.605619
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities __all__ = [ 'WorkloadCrrAccessTokenResponse', ] @pulumi.output_type class WorkloadCrrAccessTokenResponse(dict): def __init__(__self__, *, object_type: str, access_token_string: Optional[str] = None, b_ms_active_region: Optional[str] = None, backup_management_type: Optional[str] = None, container_id: Optional[str] = None, container_name: Optional[str] = None, container_type: Optional[str] = None, coordinator_service_stamp_id: Optional[str] = None, coordinator_service_stamp_uri: Optional[str] = None, datasource_container_name: Optional[str] = None, datasource_id: Optional[str] = None, datasource_name: Optional[str] = None, datasource_type: Optional[str] = None, policy_id: Optional[str] = None, policy_name: Optional[str] = None, protectable_object_container_host_os_name: Optional[str] = None, protectable_object_friendly_name: Optional[str] = None, protectable_object_parent_logical_container_name: Optional[str] = None, protectable_object_protection_state: Optional[str] = None, protectable_object_unique_name: Optional[str] = None, protectable_object_workload_type: Optional[str] = None, protection_container_id: Optional[float] = None, protection_service_stamp_id: Optional[str] = None, protection_service_stamp_uri: Optional[str] = None, recovery_point_id: Optional[str] = None, recovery_point_time: Optional[str] = None, resource_group_name: Optional[str] = None, resource_id: Optional[str] = None, resource_name: Optional[str] = None, rp_is_managed_virtual_machine: Optional[bool] = None, rp_original_sa_option: Optional[bool] = None, rp_tier_information: Optional[Mapping[str, str]] = None, rp_vm_size_description: Optional[str] = None, subscription_id: Optional[str] = None, token_extended_information: Optional[str] = None): """ :param str object_type: Type of the specific object - used for deserializing Expected value is 'WorkloadCrrAccessToken'. :param str access_token_string: Access token used for authentication :param str b_ms_active_region: Active region name of BMS Stamp :param str backup_management_type: Backup Management Type :param str container_id: Container Id :param str container_name: Container Unique name :param str container_type: Container Type :param str coordinator_service_stamp_id: CoordinatorServiceStampId to be used by BCM in restore call :param str coordinator_service_stamp_uri: CoordinatorServiceStampUri to be used by BCM in restore call :param str datasource_container_name: Datasource Container Unique Name :param str datasource_id: Datasource Id :param str datasource_name: Datasource Friendly Name :param str datasource_type: Datasource Type :param str policy_id: Policy Id :param str policy_name: Policy Name :param float protection_container_id: Protected item container id :param str protection_service_stamp_id: ProtectionServiceStampId to be used by BCM in restore call :param str protection_service_stamp_uri: ProtectionServiceStampUri to be used by BCM in restore call :param str recovery_point_id: Recovery Point Id :param str recovery_point_time: Recovery Point Time :param str resource_group_name: Resource Group name of the source vault :param str resource_id: Resource Id of the source vault :param str resource_name: Resource Name of the source vault :param bool rp_is_managed_virtual_machine: Recovery point information: Managed virtual machine :param bool rp_original_sa_option: Recovery point information: Original SA option :param Mapping[str, str] rp_tier_information: Recovery point Tier Information :param str rp_vm_size_description: Recovery point information: VM size description :param str subscription_id: Subscription Id of the source vault :param str token_extended_information: Extended Information about the token like FileSpec etc. """ pulumi.set(__self__, "object_type", 'WorkloadCrrAccessToken') if access_token_string is not None: pulumi.set(__self__, "access_token_string", access_token_string) if b_ms_active_region is not None: pulumi.set(__self__, "b_ms_active_region", b_ms_active_region) if backup_management_type is not None: pulumi.set(__self__, "backup_management_type", backup_management_type) if container_id is not None: pulumi.set(__self__, "container_id", container_id) if container_name is not None: pulumi.set(__self__, "container_name", container_name) if container_type is not None: pulumi.set(__self__, "container_type", container_type) if coordinator_service_stamp_id is not None: pulumi.set(__self__, "coordinator_service_stamp_id", coordinator_service_stamp_id) if coordinator_service_stamp_uri is not None: pulumi.set(__self__, "coordinator_service_stamp_uri", coordinator_service_stamp_uri) if datasource_container_name is not None: pulumi.set(__self__, "datasource_container_name", datasource_container_name) if datasource_id is not None: pulumi.set(__self__, "datasource_id", datasource_id) if datasource_name is not None: pulumi.set(__self__, "datasource_name", datasource_name) if datasource_type is not None: pulumi.set(__self__, "datasource_type", datasource_type) if policy_id is not None: pulumi.set(__self__, "policy_id", policy_id) if policy_name is not None: pulumi.set(__self__, "policy_name", policy_name) if protectable_object_container_host_os_name is not None: pulumi.set(__self__, "protectable_object_container_host_os_name", protectable_object_container_host_os_name) if protectable_object_friendly_name is not None: pulumi.set(__self__, "protectable_object_friendly_name", protectable_object_friendly_name) if protectable_object_parent_logical_container_name is not None: pulumi.set(__self__, "protectable_object_parent_logical_container_name", protectable_object_parent_logical_container_name) if protectable_object_protection_state is not None: pulumi.set(__self__, "protectable_object_protection_state", protectable_object_protection_state) if protectable_object_unique_name is not None: pulumi.set(__self__, "protectable_object_unique_name", protectable_object_unique_name) if protectable_object_workload_type is not None: pulumi.set(__self__, "protectable_object_workload_type", protectable_object_workload_type) if protection_container_id is not None: pulumi.set(__self__, "protection_container_id", protection_container_id) if protection_service_stamp_id is not None: pulumi.set(__self__, "protection_service_stamp_id", protection_service_stamp_id) if protection_service_stamp_uri is not None: pulumi.set(__self__, "protection_service_stamp_uri", protection_service_stamp_uri) if recovery_point_id is not None: pulumi.set(__self__, "recovery_point_id", recovery_point_id) if recovery_point_time is not None: pulumi.set(__self__, "recovery_point_time", recovery_point_time) if resource_group_name is not None: pulumi.set(__self__, "resource_group_name", resource_group_name) if resource_id is not None: pulumi.set(__self__, "resource_id", resource_id) if resource_name is not None: pulumi.set(__self__, "resource_name", resource_name) if rp_is_managed_virtual_machine is not None: pulumi.set(__self__, "rp_is_managed_virtual_machine", rp_is_managed_virtual_machine) if rp_original_sa_option is not None: pulumi.set(__self__, "rp_original_sa_option", rp_original_sa_option) if rp_tier_information is not None: pulumi.set(__self__, "rp_tier_information", rp_tier_information) if rp_vm_size_description is not None: pulumi.set(__self__, "rp_vm_size_description", rp_vm_size_description) if subscription_id is not None: pulumi.set(__self__, "subscription_id", subscription_id) if token_extended_information is not None: pulumi.set(__self__, "token_extended_information", token_extended_information) @property @pulumi.getter(name="objectType") def object_type(self) -> str: """ Type of the specific object - used for deserializing Expected value is 'WorkloadCrrAccessToken'. """ return pulumi.get(self, "object_type") @property @pulumi.getter(name="accessTokenString") def access_token_string(self) -> Optional[str]: """ Access token used for authentication """ return pulumi.get(self, "access_token_string") @property @pulumi.getter(name="bMSActiveRegion") def b_ms_active_region(self) -> Optional[str]: """ Active region name of BMS Stamp """ return pulumi.get(self, "b_ms_active_region") @property @pulumi.getter(name="backupManagementType") def backup_management_type(self) -> Optional[str]: """ Backup Management Type """ return pulumi.get(self, "backup_management_type") @property @pulumi.getter(name="containerId") def container_id(self) -> Optional[str]: """ Container Id """ return pulumi.get(self, "container_id") @property @pulumi.getter(name="containerName") def container_name(self) -> Optional[str]: """ Container Unique name """ return pulumi.get(self, "container_name") @property @pulumi.getter(name="containerType") def container_type(self) -> Optional[str]: """ Container Type """ return pulumi.get(self, "container_type") @property @pulumi.getter(name="coordinatorServiceStampId") def coordinator_service_stamp_id(self) -> Optional[str]: """ CoordinatorServiceStampId to be used by BCM in restore call """ return pulumi.get(self, "coordinator_service_stamp_id") @property @pulumi.getter(name="coordinatorServiceStampUri") def coordinator_service_stamp_uri(self) -> Optional[str]: """ CoordinatorServiceStampUri to be used by BCM in restore call """ return pulumi.get(self, "coordinator_service_stamp_uri") @property @pulumi.getter(name="datasourceContainerName") def datasource_container_name(self) -> Optional[str]: """ Datasource Container Unique Name """ return pulumi.get(self, "datasource_container_name") @property @pulumi.getter(name="datasourceId") def datasource_id(self) -> Optional[str]: """ Datasource Id """ return pulumi.get(self, "datasource_id") @property @pulumi.getter(name="datasourceName") def datasource_name(self) -> Optional[str]: """ Datasource Friendly Name """ return pulumi.get(self, "datasource_name") @property @pulumi.getter(name="datasourceType") def datasource_type(self) -> Optional[str]: """ Datasource Type """ return pulumi.get(self, "datasource_type") @property @pulumi.getter(name="policyId") def policy_id(self) -> Optional[str]: """ Policy Id """ return pulumi.get(self, "policy_id") @property @pulumi.getter(name="policyName") def policy_name(self) -> Optional[str]: """ Policy Name """ return pulumi.get(self, "policy_name") @property @pulumi.getter(name="protectableObjectContainerHostOsName") def protectable_object_container_host_os_name(self) -> Optional[str]: return pulumi.get(self, "protectable_object_container_host_os_name") @property @pulumi.getter(name="protectableObjectFriendlyName") def protectable_object_friendly_name(self) -> Optional[str]: return pulumi.get(self, "protectable_object_friendly_name") @property @pulumi.getter(name="protectableObjectParentLogicalContainerName") def protectable_object_parent_logical_container_name(self) -> Optional[str]: return pulumi.get(self, "protectable_object_parent_logical_container_name") @property @pulumi.getter(name="protectableObjectProtectionState") def protectable_object_protection_state(self) -> Optional[str]: return pulumi.get(self, "protectable_object_protection_state") @property @pulumi.getter(name="protectableObjectUniqueName") def protectable_object_unique_name(self) -> Optional[str]: return pulumi.get(self, "protectable_object_unique_name") @property @pulumi.getter(name="protectableObjectWorkloadType") def protectable_object_workload_type(self) -> Optional[str]: return pulumi.get(self, "protectable_object_workload_type") @property @pulumi.getter(name="protectionContainerId") def protection_container_id(self) -> Optional[float]: """ Protected item container id """ return pulumi.get(self, "protection_container_id") @property @pulumi.getter(name="protectionServiceStampId") def protection_service_stamp_id(self) -> Optional[str]: """ ProtectionServiceStampId to be used by BCM in restore call """ return pulumi.get(self, "protection_service_stamp_id") @property @pulumi.getter(name="protectionServiceStampUri") def protection_service_stamp_uri(self) -> Optional[str]: """ ProtectionServiceStampUri to be used by BCM in restore call """ return pulumi.get(self, "protection_service_stamp_uri") @property @pulumi.getter(name="recoveryPointId") def recovery_point_id(self) -> Optional[str]: """ Recovery Point Id """ return pulumi.get(self, "recovery_point_id") @property @pulumi.getter(name="recoveryPointTime") def recovery_point_time(self) -> Optional[str]: """ Recovery Point Time """ return pulumi.get(self, "recovery_point_time") @property @pulumi.getter(name="resourceGroupName") def resource_group_name(self) -> Optional[str]: """ Resource Group name of the source vault """ return pulumi.get(self, "resource_group_name") @property @pulumi.getter(name="resourceId") def resource_id(self) -> Optional[str]: """ Resource Id of the source vault """ return pulumi.get(self, "resource_id") @property @pulumi.getter(name="resourceName") def resource_name(self) -> Optional[str]: """ Resource Name of the source vault """ return pulumi.get(self, "resource_name") @property @pulumi.getter(name="rpIsManagedVirtualMachine") def rp_is_managed_virtual_machine(self) -> Optional[bool]: """ Recovery point information: Managed virtual machine """ return pulumi.get(self, "rp_is_managed_virtual_machine") @property @pulumi.getter(name="rpOriginalSAOption") def rp_original_sa_option(self) -> Optional[bool]: """ Recovery point information: Original SA option """ return pulumi.get(self, "rp_original_sa_option") @property @pulumi.getter(name="rpTierInformation") def rp_tier_information(self) -> Optional[Mapping[str, str]]: """ Recovery point Tier Information """ return pulumi.get(self, "rp_tier_information") @property @pulumi.getter(name="rpVMSizeDescription") def rp_vm_size_description(self) -> Optional[str]: """ Recovery point information: VM size description """ return pulumi.get(self, "rp_vm_size_description") @property @pulumi.getter(name="subscriptionId") def subscription_id(self) -> Optional[str]: """ Subscription Id of the source vault """ return pulumi.get(self, "subscription_id") @property @pulumi.getter(name="tokenExtendedInformation") def token_extended_information(self) -> Optional[str]: """ Extended Information about the token like FileSpec etc. """ return pulumi.get(self, "token_extended_information")
sdk/python/pulumi_azure_native/recoveryservices/v20181220/outputs.py
17,512
:param str object_type: Type of the specific object - used for deserializing Expected value is 'WorkloadCrrAccessToken'. :param str access_token_string: Access token used for authentication :param str b_ms_active_region: Active region name of BMS Stamp :param str backup_management_type: Backup Management Type :param str container_id: Container Id :param str container_name: Container Unique name :param str container_type: Container Type :param str coordinator_service_stamp_id: CoordinatorServiceStampId to be used by BCM in restore call :param str coordinator_service_stamp_uri: CoordinatorServiceStampUri to be used by BCM in restore call :param str datasource_container_name: Datasource Container Unique Name :param str datasource_id: Datasource Id :param str datasource_name: Datasource Friendly Name :param str datasource_type: Datasource Type :param str policy_id: Policy Id :param str policy_name: Policy Name :param float protection_container_id: Protected item container id :param str protection_service_stamp_id: ProtectionServiceStampId to be used by BCM in restore call :param str protection_service_stamp_uri: ProtectionServiceStampUri to be used by BCM in restore call :param str recovery_point_id: Recovery Point Id :param str recovery_point_time: Recovery Point Time :param str resource_group_name: Resource Group name of the source vault :param str resource_id: Resource Id of the source vault :param str resource_name: Resource Name of the source vault :param bool rp_is_managed_virtual_machine: Recovery point information: Managed virtual machine :param bool rp_original_sa_option: Recovery point information: Original SA option :param Mapping[str, str] rp_tier_information: Recovery point Tier Information :param str rp_vm_size_description: Recovery point information: VM size description :param str subscription_id: Subscription Id of the source vault :param str token_extended_information: Extended Information about the token like FileSpec etc. Access token used for authentication Active region name of BMS Stamp Backup Management Type Container Id Container Unique name Container Type CoordinatorServiceStampId to be used by BCM in restore call CoordinatorServiceStampUri to be used by BCM in restore call Datasource Container Unique Name Datasource Id Datasource Friendly Name Datasource Type Type of the specific object - used for deserializing Expected value is 'WorkloadCrrAccessToken'. Policy Id Policy Name Protected item container id ProtectionServiceStampId to be used by BCM in restore call ProtectionServiceStampUri to be used by BCM in restore call Recovery Point Id Recovery Point Time Resource Group name of the source vault Resource Id of the source vault Resource Name of the source vault Recovery point information: Managed virtual machine Recovery point information: Original SA option Recovery point Tier Information Recovery point information: VM size description Subscription Id of the source vault Extended Information about the token like FileSpec etc. coding=utf-8 *** WARNING: this file was generated by the Pulumi SDK Generator. *** *** Do not edit by hand unless you're certain you know what you are doing! ***
3,173
en
0.630396
import os import numpy as np from netCDF4 import Dataset def load_region(region_id, local=False, return_regions=False): if local: _vr = Dataset( os.path.join(os.path.dirname(os.path.abspath(__file__)), r"data/terrain_parameters/VarslingsOmr_2017.nc"), "r") # flip up-down because Meps data is upside down #_regions = np.flipud(_vr.variables["LokalOmr_2018"][:]) _regions = _vr.variables["LokalOmr_2018"][:] else: _vr = Dataset( os.path.join(os.path.dirname(os.path.abspath(__file__)), r"data/terrain_parameters/VarslingsOmr_2019.nc"), "r") # flip up-down because Meps data is upside down #_regions = np.flipud(_vr.variables["skredomr19_km"][:]) _regions = _vr.variables["skredomr19_km"][:] print("Missing value: {mv}".format(mv=_vr.variables["skredomr19_km"].missing_value)) _region_bounds = np.where(_regions == region_id) # just to get the bounding box # get the lower left and upper right corner of a rectangle around the region y_min, y_max, x_min, x_max = min(_region_bounds[0].flatten()), max(_region_bounds[0].flatten()), \ min(_region_bounds[1].flatten()), max(_region_bounds[1].flatten()) #reg_mask = np.ma.masked_where(_regions[y_min:y_max, x_min:x_max] == region_id, _regions[y_min:y_max, x_min:x_max]).mask #reg_mask = np.where(_regions[y_min:y_max, x_min:x_max] == region_id, _regions[y_min:y_max, x_min:x_max], np.nan) reg_mask = np.where(_regions[y_min:y_max, x_min:x_max] == region_id, 1., np.nan) #reg_mask = np.ma.masked_where(_reg_mask == region_id).mask _vr.close() if return_regions: return _regions, reg_mask, y_min, y_max, x_min, x_max else: return reg_mask, y_min, y_max, x_min, x_max def clip_region(nc_variable, region_mask, t_index, y_min, y_max, x_min, x_max): s = len(nc_variable.shape) if s == 2: #return np.flipud(region_mask * nc_variable[y_min:y_max, x_min:x_max]) return (region_mask * nc_variable[y_min:y_max, x_min:x_max]) elif s == 3: #return np.flipud(region_mask * nc_variable[t_index, y_min:y_max, x_min:x_max]) return (region_mask * nc_variable[t_index, y_min:y_max, x_min:x_max]) elif s == 4: #return np.flipud(region_mask * nc_variable[t_index, 0, y_min:y_max, x_min:x_max]) return (region_mask * nc_variable[t_index, 0, y_min:y_max, x_min:x_max]) else: print('Input array needs to have 2- to 4-dimensions: {0} were given.'.format(s)) if __name__ == "__main__": import matplotlib.pyplot as plt regions, region_mask, y_min, y_max, x_min, x_max = load_region(3013, return_regions=True) print(region_mask, type(region_mask), np.unique(region_mask)) clp = clip_region(regions, region_mask, 0, y_min, y_max, x_min, x_max) plt.imshow(clp) plt.show() k = 'm'
aps/load_region.py
2,930
flip up-down because Meps data is upside down_regions = np.flipud(_vr.variables["LokalOmr_2018"][:]) flip up-down because Meps data is upside down_regions = np.flipud(_vr.variables["skredomr19_km"][:]) just to get the bounding box get the lower left and upper right corner of a rectangle around the regionreg_mask = np.ma.masked_where(_regions[y_min:y_max, x_min:x_max] == region_id, _regions[y_min:y_max, x_min:x_max]).maskreg_mask = np.where(_regions[y_min:y_max, x_min:x_max] == region_id, _regions[y_min:y_max, x_min:x_max], np.nan)reg_mask = np.ma.masked_where(_reg_mask == region_id).maskreturn np.flipud(region_mask * nc_variable[y_min:y_max, x_min:x_max])return np.flipud(region_mask * nc_variable[t_index, y_min:y_max, x_min:x_max])return np.flipud(region_mask * nc_variable[t_index, 0, y_min:y_max, x_min:x_max])
822
en
0.090889
# Generated by Django 2.1.11 on 2020-06-24 06:55 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('products', '0008_auto_20190919_1521'), ] operations = [ migrations.AddField( model_name='product', name='is_deleted', field=models.BooleanField(default=False), ), ]
products/migrations/0009_product_is_deleted.py
394
Generated by Django 2.1.11 on 2020-06-24 06:55
46
en
0.69865
import time import torch from hpc_rll.origin.td import iqn_nstep_td_error, iqn_nstep_td_data from hpc_rll.rl_utils.td import IQNNStepTDError from testbase import mean_relative_error, times assert torch.cuda.is_available() use_cuda = True tau = 33 tauPrime = 34 T = 10 B = 64 N = 8 gamma = 0.95 kappa = 0.9 def iqn_val(): ori_q = torch.randn(tau, B, N) ori_next_n_q = torch.randn(tauPrime, B, N) ori_action = torch.randint(0, N, size=(B, )) ori_next_n_action = torch.randint(0, N, size=(B, )) ori_reward = torch.randn(T, B) ori_done = torch.randn(B) ori_r_q = torch.randn(tau, B) ori_weight = torch.randn(B) ori_value_gamma = torch.randn(B) hpc_q = ori_q.clone().detach() hpc_next_n_q = ori_next_n_q.clone().detach() hpc_action = ori_action.clone().detach() hpc_next_n_action = ori_next_n_action.clone().detach() hpc_reward = ori_reward.clone().detach() hpc_done = ori_done.clone().detach() hpc_r_q = ori_r_q.clone().detach() hpc_weight = ori_weight.clone().detach() hpc_value_gamma = ori_value_gamma.clone().detach() hpc_iqn = IQNNStepTDError(tau, tauPrime, T, B, N) if use_cuda: ori_q = ori_q.cuda() ori_next_n_q = ori_next_n_q.cuda() ori_action = ori_action.cuda() ori_next_n_action = ori_next_n_action.cuda() ori_reward = ori_reward.cuda() ori_done = ori_done.cuda() ori_r_q = ori_r_q.cuda() ori_weight = ori_weight.cuda() ori_value_gamma = ori_value_gamma.cuda() hpc_q = hpc_q.cuda() hpc_next_n_q = hpc_next_n_q.cuda() hpc_action = hpc_action.cuda() hpc_next_n_action = hpc_next_n_action.cuda() hpc_reward = hpc_reward.cuda() hpc_done = hpc_done.cuda() hpc_r_q = hpc_r_q.cuda() hpc_weight = hpc_weight.cuda() hpc_value_gamma = hpc_value_gamma.cuda() hpc_iqn = hpc_iqn.cuda() ori_q.requires_grad_(True) ori_loss, ori_ = iqn_nstep_td_error(iqn_nstep_td_data(ori_q, ori_next_n_q, ori_action, ori_next_n_action, ori_reward, ori_done, ori_r_q, ori_weight), gamma, T, kappa, ori_value_gamma) ori_loss = ori_loss.mean() ori_loss.backward() if use_cuda: torch.cuda.synchronize() torch.cuda.cudart().cudaProfilerStart() hpc_q.requires_grad_(True) hpc_loss, hpc_ = hpc_iqn(hpc_q, hpc_next_n_q, hpc_action, hpc_next_n_action, hpc_reward, hpc_done, hpc_r_q, gamma, kappa, hpc_weight, hpc_value_gamma) hpc_loss = hpc_loss.mean() hpc_loss.backward() if use_cuda: torch.cuda.synchronize() torch.cuda.cudart().cudaProfilerStop() mre = mean_relative_error(torch.flatten(ori_loss).cpu().detach().numpy(), torch.flatten(hpc_loss).cpu().detach().numpy()) print("iqn fp mean_relative_error: " + str(mre)) mre = mean_relative_error(torch.flatten(ori_q.grad).cpu().detach().numpy(), torch.flatten(hpc_q.grad).cpu().detach().numpy()) print("iqn bp mean_relative_error: " + str(mre)) def iqn_perf(): ori_q = torch.randn(tau, B, N) ori_next_n_q = torch.randn(tauPrime, B, N) ori_action = torch.randint(0, N, size=(B, )) ori_next_n_action = torch.randint(0, N, size=(B, )) ori_reward = torch.randn(T, B) ori_done = torch.randn(B) ori_r_q = torch.randn(tau, B) ori_weight = torch.randn(B) ori_value_gamma = torch.randn(B) hpc_q = ori_q.clone().detach() hpc_next_n_q = ori_next_n_q.clone().detach() hpc_action = ori_action.clone().detach() hpc_next_n_action = ori_next_n_action.clone().detach() hpc_reward = ori_reward.clone().detach() hpc_done = ori_done.clone().detach() hpc_r_q = ori_r_q.clone().detach() hpc_weight = ori_weight.clone().detach() hpc_value_gamma = ori_value_gamma.clone().detach() hpc_iqn = IQNNStepTDError(tau, tauPrime, T, B, N) if use_cuda: ori_q = ori_q.cuda() ori_next_n_q = ori_next_n_q.cuda() ori_action = ori_action.cuda() ori_next_n_action = ori_next_n_action.cuda() ori_reward = ori_reward.cuda() ori_done = ori_done.cuda() ori_r_q = ori_r_q.cuda() ori_weight = ori_weight.cuda() ori_value_gamma = ori_value_gamma.cuda() hpc_q = hpc_q.cuda() hpc_next_n_q = hpc_next_n_q.cuda() hpc_action = hpc_action.cuda() hpc_next_n_action = hpc_next_n_action.cuda() hpc_reward = hpc_reward.cuda() hpc_done = hpc_done.cuda() hpc_r_q = hpc_r_q.cuda() hpc_weight = hpc_weight.cuda() hpc_iqn = hpc_iqn.cuda() hpc_value_gamma = hpc_value_gamma.cuda() ori_q.requires_grad_(True) for i in range(times): t = time.time() ori_loss, ori_ = iqn_nstep_td_error(iqn_nstep_td_data(ori_q, ori_next_n_q, ori_action, ori_next_n_action, ori_reward, ori_done, ori_r_q, ori_weight), gamma, T, kappa, ori_value_gamma) ori_loss = ori_loss.mean() ori_loss.backward() if use_cuda: torch.cuda.synchronize() print('epoch: {}, original iqn cost time: {}'.format(i, time.time() - t)) #torch.cuda.cudart().cudaProfilerStart() hpc_q.requires_grad_(True) for i in range(times): t = time.time() hpc_loss, hpc_ = hpc_iqn(hpc_q, hpc_next_n_q, hpc_action, hpc_next_n_action, hpc_reward, hpc_done, hpc_r_q, gamma, kappa, hpc_weight, hpc_value_gamma) hpc_loss = hpc_loss.mean() hpc_loss.backward() if use_cuda: torch.cuda.synchronize() print('epoch: {}, hpc iqn cost time: {}'.format(i, time.time() - t)) #torch.cuda.cudart().cudaProfilerStop() mre = mean_relative_error(torch.flatten(ori_loss).cpu().detach().numpy(), torch.flatten(hpc_loss).cpu().detach().numpy()) print("iqn fp mean_relative_error: " + str(mre)) mre = mean_relative_error(torch.flatten(ori_q.grad).cpu().detach().numpy(), torch.flatten(hpc_q.grad).cpu().detach().numpy()) print("iqn bp mean_relative_error: " + str(mre)) if __name__ == '__main__': print("target problem: tau = {}, tauPrime = {}, T = {}, B = {}, N = {}, gamma = {}, kappa = {}".format(tau, tauPrime, T, B, N, gamma, kappa)) print("================run iqn validation test================") iqn_val() print("================run iqn performance test================") iqn_perf()
tests/test_iqn_nstep_td_error.py
6,308
torch.cuda.cudart().cudaProfilerStart()torch.cuda.cudart().cudaProfilerStop()
77
zh
0.110549
#!/usr/bin/env python3 # Copyright (c) 2015-2016 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. """Utilities for manipulating blocks and transactions.""" from test_framework.mininode import * from test_framework.script import CScript, OP_TRUE, OP_CHECKSIG # Create a block (with regtest difficulty) def create_block(hashprev, coinbase, nTime=None): block = CBlock() if nTime is None: import time block.nTime = int(time.time()+600) else: block.nTime = nTime block.hashPrevBlock = hashprev block.nBits = 0x1e0ffff0 # Will break after a difficulty adjustment... block.vtx.append(coinbase) block.hashMerkleRoot = block.calc_merkle_root() block.calc_sha256() return block def serialize_script_num(value): r = bytearray(0) if value == 0: return r neg = value < 0 absvalue = -value if neg else value while (absvalue): r.append(int(absvalue & 0xff)) absvalue >>= 8 if r[-1] & 0x80: r.append(0x80 if neg else 0) elif neg: r[-1] |= 0x80 return r def cbase_scriptsig(height): return ser_string(serialize_script_num(height)) def cbase_value(height): #return ((50 * COIN) >> int(height/150)) return (250 * COIN) # Create a coinbase transaction, assuming no miner fees. # If pubkey is passed in, the coinbase output will be a P2PK output; # otherwise an anyone-can-spend output. def create_coinbase(height, pubkey = None): coinbase = CTransaction() coinbase.vin = [CTxIn(NullOutPoint, cbase_scriptsig(height), 0xffffffff)] coinbaseoutput = CTxOut() coinbaseoutput.nValue = cbase_value(height) if (pubkey != None): coinbaseoutput.scriptPubKey = CScript([pubkey, OP_CHECKSIG]) else: coinbaseoutput.scriptPubKey = CScript([OP_TRUE]) coinbase.vout = [coinbaseoutput] coinbase.calc_sha256() return coinbase # Create a transaction. # If the scriptPubKey is not specified, make it anyone-can-spend. def create_transaction(prevtx, n, sig, value, scriptPubKey=CScript()): tx = CTransaction() assert(n < len(prevtx.vout)) tx.vin.append(CTxIn(COutPoint(prevtx.sha256, n), sig, 0xffffffff)) tx.vout.append(CTxOut(value, scriptPubKey)) tx.calc_sha256() return tx def create_transaction_from_outpoint(outPoint, sig, value, scriptPubKey=CScript()): tx = CTransaction() tx.vin.append(CTxIn(outPoint, sig, 0xffffffff)) tx.vout.append(CTxOut(value, scriptPubKey)) tx.calc_sha256() return tx def get_legacy_sigopcount_block(block, fAccurate=True): count = 0 for tx in block.vtx: count += get_legacy_sigopcount_tx(tx, fAccurate) return count def get_legacy_sigopcount_tx(tx, fAccurate=True): count = 0 for i in tx.vout: count += i.scriptPubKey.GetSigOpCount(fAccurate) for j in tx.vin: # scriptSig might be of type bytes, so convert to CScript for the moment count += CScript(j.scriptSig).GetSigOpCount(fAccurate) return count ### SupernodeCoin specific blocktools ### def create_coinbase_pos(height): coinbase = CTransaction() coinbase.vin = [CTxIn(NullOutPoint, cbase_scriptsig(height), 0xffffffff)] coinbase.vout = [CTxOut(0, b"")] coinbase.calc_sha256() return coinbase def is_zerocoin(uniqueness): ulen = len(uniqueness) if ulen == 32: return True if ulen == 36: return False raise Exception("Wrong uniqueness len: %d" % ulen)
test/functional/test_framework/blocktools.py
3,564
Utilities for manipulating blocks and transactions. !/usr/bin/env python3 Copyright (c) 2015-2016 The Bitcoin Core developers Distributed under the MIT software license, see the accompanying file COPYING or http://www.opensource.org/licenses/mit-license.php. Create a block (with regtest difficulty) Will break after a difficulty adjustment...return ((50 * COIN) >> int(height/150)) Create a coinbase transaction, assuming no miner fees. If pubkey is passed in, the coinbase output will be a P2PK output; otherwise an anyone-can-spend output. Create a transaction. If the scriptPubKey is not specified, make it anyone-can-spend. scriptSig might be of type bytes, so convert to CScript for the moment SupernodeCoin specific blocktools
734
en
0.655877
# -*- coding: utf-8 -*- import datetime from unittest.mock import Mock import pytest from h.activity import bucketing from tests.common import factories UTCNOW = datetime.datetime(year=1970, month=2, day=21, hour=19, minute=30) FIVE_MINS_AGO = UTCNOW - datetime.timedelta(minutes=5) YESTERDAY = UTCNOW - datetime.timedelta(days=1) THIRD_MARCH_1968 = datetime.datetime(year=1968, month=3, day=3) FIFTH_NOVEMBER_1969 = datetime.datetime(year=1969, month=11, day=5) class timeframe_with: # noqa: N801 def __init__(self, label, document_buckets): self.label = label self.document_buckets = document_buckets def __eq__(self, timeframe): return ( self.label == timeframe.label and self.document_buckets == timeframe.document_buckets ) def __repr__(self): return '{class_} "{label}" with {n} document buckets'.format( class_=self.__class__, label=self.label, n=len(self.document_buckets) ) @pytest.mark.usefixtures("factories") class TestDocumentBucket: def test_init_sets_the_document_title(self, db_session, document): title_meta = factories.DocumentMeta( type="title", value=["The Document Title"], document=document ) document.title = "The Document Title" db_session.add(title_meta) db_session.flush() bucket = bucketing.DocumentBucket(document) assert bucket.title == "The Document Title" def test_init_uses_the_document_web_uri(self, db_session, document): document.web_uri = "http://example.com" bucket = bucketing.DocumentBucket(document) assert bucket.uri == "http://example.com" def test_init_sets_None_uri_when_no_http_or_https_can_be_found( self, db_session, document ): document.web_uri = None bucket = bucketing.DocumentBucket(document) assert bucket.uri is None def test_init_sets_the_domain_from_the_extracted_uri(self, db_session, document): document.web_uri = "https://www.example.com/foobar.html" bucket = bucketing.DocumentBucket(document) assert bucket.domain == "www.example.com" def test_init_sets_domain_to_local_file_when_no_uri_is_set( self, db_session, document ): docuri_pdf = factories.DocumentURI( uri="urn:x-pdf:fingerprint", document=document ) db_session.add(docuri_pdf) db_session.flush() bucket = bucketing.DocumentBucket(document) assert bucket.domain == "Local file" def test_annotations_count_returns_count_of_annotations(self, db_session, document): bucket = bucketing.DocumentBucket(document) for _ in range(7): annotation = factories.Annotation() bucket.append(annotation) assert bucket.annotations_count == 7 def test_append_appends_the_annotation(self, document): bucket = bucketing.DocumentBucket(document) annotations = [] for _ in range(7): annotation = factories.Annotation() annotations.append(annotation) bucket.append(annotation) assert bucket.annotations == annotations def test_append_adds_unique_annotation_tag_to_bucket(self, document): ann_1 = factories.Annotation(tags=["foo", "bar"]) ann_2 = factories.Annotation(tags=["foo", "baz"]) bucket = bucketing.DocumentBucket(document) bucket.append(ann_1) bucket.append(ann_2) assert bucket.tags == {"foo", "bar", "baz"} def test_append_adds_unique_annotation_user_to_bucket(self, document): ann_1 = factories.Annotation(userid="luke") ann_2 = factories.Annotation(userid="alice") ann_3 = factories.Annotation(userid="luke") bucket = bucketing.DocumentBucket(document) bucket.append(ann_1) bucket.append(ann_2) bucket.append(ann_3) assert bucket.users == {"luke", "alice"} def test_eq(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) for _ in range(5): annotation = factories.Annotation() bucket_1.append(annotation) bucket_2.append(annotation) assert bucket_1 == bucket_2 def test_eq_annotations_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.annotations = [1, 2, 3] bucket_2.annotations = [2, 3, 4] assert not bucket_1 == bucket_2 def test_eq_tags_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.tags.update(["foo", "bar"]) bucket_2.tags.update(["foo", "baz"]) assert not bucket_1 == bucket_2 def test_eq_users_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.users.update(["alice", "luke"]) bucket_2.users.update(["luke", "paula"]) assert not bucket_1 == bucket_2 def test_eq_uri_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.uri = "http://example.com" bucket_2.uri = "http://example.org" assert not bucket_1 == bucket_2 def test_eq_domain_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.domain = "example.com" bucket_2.domain = "example.org" assert not bucket_1 == bucket_2 def test_eq_title_mismatch(self, document): bucket_1 = bucketing.DocumentBucket(document) bucket_2 = bucketing.DocumentBucket(document) bucket_1.title = "First Title" bucket_2.title = "Second Title" assert not bucket_1 == bucket_2 def test_incontext_link_returns_link_to_first_annotation(self, document, patch): incontext_link = patch("h.links.incontext_link") bucket = bucketing.DocumentBucket(document) ann = factories.Annotation() bucket.append(ann) request = Mock() assert bucket.incontext_link(request) == incontext_link.return_value def test_incontext_link_returns_none_if_bucket_empty(self, document, patch): patch("h.links.incontext_link") bucket = bucketing.DocumentBucket(document) request = Mock() assert bucket.incontext_link(request) is None @pytest.fixture def document(self, db_session): document = factories.Document() db_session.add(document) db_session.flush() return document @pytest.mark.usefixtures("factories", "utcnow") class TestBucket: def test_no_annotations(self): assert bucketing.bucket([]) == [] @pytest.mark.parametrize( "annotation_datetime,timeframe_label", [(FIVE_MINS_AGO, "Last 7 days"), (THIRD_MARCH_1968, "Mar 1968")], ) def test_one_annotation(self, annotation_datetime, timeframe_label): annotation = factories.Annotation( document=factories.Document(), updated=annotation_datetime ) timeframes = bucketing.bucket([annotation]) assert timeframes == [ timeframe_with( timeframe_label, { annotation.document: bucketing.DocumentBucket( annotation.document, [annotation] ) }, ) ] @pytest.mark.parametrize( "annotation_datetime,timeframe_label", [(FIVE_MINS_AGO, "Last 7 days"), (THIRD_MARCH_1968, "Mar 1968")], ) def test_multiple_annotations_of_one_document_in_one_timeframe( self, annotation_datetime, timeframe_label ): results = [ factories.Annotation( target_uri="https://example.com", updated=annotation_datetime ) for _ in range(3) ] timeframes = bucketing.bucket(results) document = results[0].document assert timeframes == [ timeframe_with( timeframe_label, {document: bucketing.DocumentBucket(document, results)} ) ] @pytest.mark.parametrize( "annotation_datetime,timeframe_label", [(YESTERDAY, "Last 7 days"), (THIRD_MARCH_1968, "Mar 1968")], ) def test_annotations_of_multiple_documents_in_one_timeframe( self, annotation_datetime, timeframe_label ): annotation_1 = factories.Annotation( target_uri="http://example1.com", updated=annotation_datetime ) annotation_2 = factories.Annotation( target_uri="http://example2.com", updated=annotation_datetime ) annotation_3 = factories.Annotation( target_uri="http://example3.com", updated=annotation_datetime ) timeframes = bucketing.bucket([annotation_1, annotation_2, annotation_3]) assert timeframes == [ timeframe_with( timeframe_label, { annotation_1.document: bucketing.DocumentBucket( annotation_1.document, [annotation_1] ), annotation_2.document: bucketing.DocumentBucket( annotation_2.document, [annotation_2] ), annotation_3.document: bucketing.DocumentBucket( annotation_3.document, [annotation_3] ), }, ) ] def test_annotations_of_the_same_document_in_different_timeframes(self): results = [ factories.Annotation(), factories.Annotation(updated=FIFTH_NOVEMBER_1969), factories.Annotation(updated=THIRD_MARCH_1968), ] document = factories.Document() for annotation in results: annotation.document = document timeframes = bucketing.bucket(results) expected_bucket_1 = bucketing.DocumentBucket(document, [results[0]]) expected_bucket_2 = bucketing.DocumentBucket(document, [results[1]]) expected_bucket_3 = bucketing.DocumentBucket(document, [results[2]]) assert timeframes == [ timeframe_with("Last 7 days", {document: expected_bucket_1}), timeframe_with("Nov 1969", {document: expected_bucket_2}), timeframe_with("Mar 1968", {document: expected_bucket_3}), ] def test_recent_and_older_annotations_together(self): results = [ factories.Annotation(target_uri="http://example1.com"), factories.Annotation(target_uri="http://example2.com"), factories.Annotation(target_uri="http://example3.com"), factories.Annotation( target_uri="http://example4.com", updated=THIRD_MARCH_1968 ), factories.Annotation( target_uri="http://example5.com", updated=THIRD_MARCH_1968 ), factories.Annotation( target_uri="http://example6.com", updated=THIRD_MARCH_1968 ), ] timeframes = bucketing.bucket(results) expected_bucket_1 = bucketing.DocumentBucket(results[0].document, [results[0]]) expected_bucket_2 = bucketing.DocumentBucket(results[1].document, [results[1]]) expected_bucket_3 = bucketing.DocumentBucket(results[2].document, [results[2]]) expected_bucket_4 = bucketing.DocumentBucket(results[3].document, [results[3]]) expected_bucket_5 = bucketing.DocumentBucket(results[4].document, [results[4]]) expected_bucket_6 = bucketing.DocumentBucket(results[5].document, [results[5]]) assert timeframes == [ timeframe_with( "Last 7 days", { results[0].document: expected_bucket_1, results[1].document: expected_bucket_2, results[2].document: expected_bucket_3, }, ), timeframe_with( "Mar 1968", { results[3].document: expected_bucket_4, results[4].document: expected_bucket_5, results[5].document: expected_bucket_6, }, ), ] def test_annotations_from_different_days_in_same_month(self): """ Test bucketing multiple annotations from different days of same month. Annotations from different days of the same month should go into one bucket. """ one_month_ago = UTCNOW - datetime.timedelta(days=30) annotations = [ factories.Annotation( target_uri="http://example.com", updated=one_month_ago ), factories.Annotation( target_uri="http://example.com", updated=one_month_ago - datetime.timedelta(days=1), ), factories.Annotation( target_uri="http://example.com", updated=one_month_ago - datetime.timedelta(days=2), ), ] timeframes = bucketing.bucket(annotations) expected_bucket = bucketing.DocumentBucket(annotations[0].document) expected_bucket.update(annotations) assert timeframes == [ timeframe_with("Jan 1970", {annotations[0].document: expected_bucket}) ] @pytest.fixture def utcnow(self, patch): utcnow = patch("h.activity.bucketing.utcnow") utcnow.return_value = UTCNOW return utcnow
tests/h/activity/bucketing_test.py
13,834
Test bucketing multiple annotations from different days of same month. Annotations from different days of the same month should go into one bucket. -*- coding: utf-8 -*- noqa: N801
183
en
0.858458
# models.py from flask_login import UserMixin from . import db class User(UserMixin, db.Model): id = db.Column(db.Integer, primary_key=True) # primary keys are required by SQLAlchemy email = db.Column(db.String(100), unique=True) password = db.Column(db.String(100)) name = db.Column(db.String(1000)) home = db.Column(db.String(1000))
webapp/models.py
356
models.py primary keys are required by SQLAlchemy
49
en
0.893456
# -*- coding: utf-8 -*- from . import fields from . import integrators from . import points from . import system from . import utils from . import visualizer from .system import NBodySystem from .visualizer import Visualizer, run
fieldbillard/__init__.py
232
-*- coding: utf-8 -*-
21
en
0.767281
import json import logging import requests from kube_hunter.core.events import handler from kube_hunter.core.events.types import Event, OpenPortEvent, Service from kube_hunter.core.types import Discovery class EtcdAccessEvent(Service, Event): """Etcd is a DB that stores cluster's data, it contains configuration and current state information, and might contain secrets""" def __init__(self): Service.__init__(self, name="Etcd") @handler.subscribe(OpenPortEvent, predicate= lambda p: p.port == 2379) class EtcdRemoteAccess(Discovery): """Etcd service check for the existence of etcd service """ def __init__(self, event): self.event = event def execute(self): self.publish_event(EtcdAccessEvent())
kube_hunter/modules/discovery/etcd.py
760
Etcd is a DB that stores cluster's data, it contains configuration and current state information, and might contain secrets Etcd service check for the existence of etcd service
176
en
0.843072
# -*- coding: utf-8 -*- # This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. # # Copyright (c) 2021-present Kaleidos Ventures SL from __future__ import unicode_literals from django.db import connection, migrations, models def set_finished_date_for_tasks(apps, schema_editor): # Updates the finished date from tasks according to the history_entries associated # It takes the last history change updateing the status of a task and if it's a closed # one it updates the finished_date attribute sql=""" WITH status_update AS( WITH status_update AS( WITH history_entries AS ( SELECT diff #>>'{status, 1}' new_status_id, regexp_split_to_array(key, ':') as split_key, created_at as date FROM history_historyentry WHERE diff #>>'{status, 1}' != '' ) SELECT split_key[2] as object_id, new_status_id::int, MAX(date) as status_change_datetime FROM history_entries WHERE split_key[1] = 'tasks.task' GROUP BY object_id, new_status_id, date ) SELECT status_update.* FROM status_update INNER JOIN projects_taskstatus ON projects_taskstatus.id = new_status_id AND projects_taskstatus.is_closed = True ) UPDATE tasks_task SET finished_date = status_update.status_change_datetime FROM status_update WHERE tasks_task.id = status_update.object_id::int """ cursor = connection.cursor() cursor.execute(sql) class Migration(migrations.Migration): dependencies = [ ('tasks', '0008_remove_task_watchers'), ] operations = [ migrations.RunPython(set_finished_date_for_tasks), ]
taiga/projects/tasks/migrations/0009_auto_20151104_1131.py
1,711
-*- coding: utf-8 -*- This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. Copyright (c) 2021-present Kaleidos Ventures SL Updates the finished date from tasks according to the history_entries associated It takes the last history change updateing the status of a task and if it's a closed one it updates the finished_date attribute
471
en
0.871513
from gpiozero import CPUTemperature from tabulate import tabulate from math import floor import numpy as np import termplotlib as tpl import time import shutil def roundNum(num, digits): return floor(num * 10 ** digits) / (10 ** digits) def CtoF(temp): fahrenheit = (temp + 1.8) + 32 rounded = roundNum(fahrenheit, 3) return str(rounded) cpu = CPUTemperature() colors = { 'HEADER': '\033[95m', 'OKBLUE': '\033[94m', 'OKCYAN': '\033[96m', 'OKGREEN': '\033[92m', 'WARNING': '\033[93m', 'FAIL': '\033[91m', 'ENDC': '\033[0m', 'BOLD': '\033[1m', 'UNDERLINE': '\033[4m', } times = [0] temps = [cpu.temperature] while True: tickRate = 2 #takes data every {tickRate} seconds minutes = 5 numPoints = int(60 / tickRate * minutes) width, height = shutil.get_terminal_size() if len(temps) > numPoints: temps = temps[-numPoints:] times = times[-numPoints:] temps.append(cpu.temperature) times.append(times[-1] + tickRate) averageTemp = roundNum(np.average(temps), 3) cpuTempColor = '' if cpu.temperature < 50: cpuTempColor = colors['OKBLUE'] elif cpu.temperature < 65: cpuTempColor = colors['OKCYAN'] elif cpu.temperature < 80: cpuTempColor = colors['OKGREEN'] else: cpuTempColor = colors['FAIL'] + colors['BOLD'] table = [[ f"{cpuTempColor}{str(cpu.temperature)}\N{DEGREE SIGN}C / {CtoF(cpu.temperature)}\N{DEGREE SIGN}F\n", f"{colors['OKGREEN']}{averageTemp} / {CtoF(averageTemp)}\N{DEGREE SIGN}F\n", f"{colors['OKGREEN']}{np.amax(temps)} / {CtoF(np.amax(temps))}\N{DEGREE SIGN}F\n", f"{colors['OKGREEN']}{np.amin(temps)} / {CtoF(np.amin(temps))}\N{DEGREE SIGN}F" ]] headers = [ f"{colors['OKGREEN']}CPU TEMPERATURE", f"{colors['OKGREEN']}Average Temperature (last {minutes} minutes)", f"{colors['FAIL']}Peak Temperature (last {minutes} minutes)", f"{colors['OKCYAN']}Lowest Temperature (last {minutes} minutes){colors['OKGREEN']}", #OKGREEN at end is to make sure table lines are green, not cyan ] print('\n') fig = tpl.figure() plotConfig = { 'width': width-2, 'height': height-5, 'label': 'CPU Temperature', 'xlabel': 'Time (s)', 'xlim': [times[0], times[-1:]], 'ylim': [np.amin(temps)-2, np.amax(temps)+2], 'title': f"CPU Temperature over last {minutes} minutes", } fig.plot(times, temps, **plotConfig) fig.show() # width=width-2, height=height-5, label='CPU Temperature', xlabel='Time (s)', , ylim=[np.amin(temps)-2, np.amax(temps)+2], title='CPU Temperature over last 5 minutes' print('\n') print(tabulate(table, headers=headers)) time.sleep(tickRate)
monitor_temp.py
2,789
takes data every {tickRate} secondsOKGREEN at end is to make sure table lines are green, not cyan width=width-2, height=height-5, label='CPU Temperature', xlabel='Time (s)', , ylim=[np.amin(temps)-2, np.amax(temps)+2], title='CPU Temperature over last 5 minutes'
262
en
0.473205
# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """Nexmark launcher. The Nexmark suite is a series of queries (streaming pipelines) performed on a simulation of auction events. The launcher orchestrates the generation and parsing of streaming events and the running of queries. Model - Person: Author of an auction or a bid. - Auction: Item under auction. - Bid: A bid for an item under auction. Events - Create Person - Create Auction - Create Bid Queries - Query0: Pass through (send and receive auction events). Usage - DirectRunner python nexmark_launcher.py \ --query/q <query number> \ --project <project id> \ --loglevel=DEBUG (optional) \ --wait_until_finish_duration <time_in_ms> \ --streaming - DataflowRunner python nexmark_launcher.py \ --query/q <query number> \ --project <project id> \ --region <GCE region> \ --loglevel=DEBUG (optional) \ --wait_until_finish_duration <time_in_ms> \ --streaming \ --sdk_location <apache_beam tar.gz> \ --staging_location=gs://... \ --temp_location=gs:// """ # pytype: skip-file from __future__ import absolute_import from __future__ import print_function import argparse import logging import sys import uuid from google.cloud import pubsub import apache_beam as beam from apache_beam.options.pipeline_options import GoogleCloudOptions from apache_beam.options.pipeline_options import PipelineOptions from apache_beam.options.pipeline_options import SetupOptions from apache_beam.options.pipeline_options import StandardOptions from apache_beam.options.pipeline_options import TestOptions from apache_beam.testing.benchmarks.nexmark.nexmark_util import Command from apache_beam.testing.benchmarks.nexmark.queries import query0 from apache_beam.testing.benchmarks.nexmark.queries import query1 from apache_beam.testing.benchmarks.nexmark.queries import query2 class NexmarkLauncher(object): def __init__(self): self.parse_args() self.uuid = str(uuid.uuid4()) self.topic_name = self.args.topic_name + self.uuid self.subscription_name = self.args.subscription_name + self.uuid publish_client = pubsub.Client(project=self.project) topic = publish_client.topic(self.topic_name) if topic.exists(): logging.info('deleting topic %s', self.topic_name) topic.delete() logging.info('creating topic %s', self.topic_name) topic.create() sub = topic.subscription(self.subscription_name) if sub.exists(): logging.info('deleting sub %s', self.topic_name) sub.delete() logging.info('creating sub %s', self.topic_name) sub.create() def parse_args(self): parser = argparse.ArgumentParser() parser.add_argument( '--query', '-q', type=int, action='append', required=True, choices=[0, 1, 2], help='Query to run') parser.add_argument( '--subscription_name', type=str, help='Pub/Sub subscription to read from') parser.add_argument( '--topic_name', type=str, help='Pub/Sub topic to read from') parser.add_argument( '--loglevel', choices=['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL'], default='INFO', help='Set logging level to debug') parser.add_argument( '--input', type=str, required=True, help='Path to the data file containing nexmark events.') self.args, self.pipeline_args = parser.parse_known_args() logging.basicConfig( level=getattr(logging, self.args.loglevel, None), format='(%(threadName)-10s) %(message)s') self.pipeline_options = PipelineOptions(self.pipeline_args) logging.debug('args, pipeline_args: %s, %s', self.args, self.pipeline_args) # Usage with Dataflow requires a project to be supplied. self.project = self.pipeline_options.view_as(GoogleCloudOptions).project if self.project is None: parser.print_usage() print(sys.argv[0] + ': error: argument --project is required') sys.exit(1) # Pub/Sub is currently available for use only in streaming pipelines. self.streaming = self.pipeline_options.view_as(StandardOptions).streaming if self.streaming is None: parser.print_usage() print(sys.argv[0] + ': error: argument --streaming is required') sys.exit(1) # wait_until_finish ensures that the streaming job is canceled. self.wait_until_finish_duration = ( self.pipeline_options.view_as(TestOptions).wait_until_finish_duration) if self.wait_until_finish_duration is None: parser.print_usage() print(sys.argv[0] + ': error: argument --wait_until_finish_duration is required') # pylint: disable=line-too-long sys.exit(1) # We use the save_main_session option because one or more DoFn's in this # workflow rely on global context (e.g., a module imported at module level). self.pipeline_options.view_as(SetupOptions).save_main_session = True def generate_events(self): publish_client = pubsub.Client(project=self.project) topic = publish_client.topic(self.topic_name) sub = topic.subscription(self.subscription_name) logging.info('Generating auction events to topic %s', topic.name) if self.args.input.startswith('gs://'): from apache_beam.io.gcp.gcsfilesystem import GCSFileSystem fs = GCSFileSystem(self.pipeline_options) with fs.open(self.args.input) as infile: for line in infile: topic.publish(line) else: with open(self.args.input) as infile: for line in infile: topic.publish(line) logging.info('Finished event generation.') # Read from PubSub into a PCollection. if self.args.subscription_name: raw_events = self.pipeline | 'ReadPubSub' >> beam.io.ReadFromPubSub( subscription=sub.full_name) else: raw_events = self.pipeline | 'ReadPubSub' >> beam.io.ReadFromPubSub( topic=topic.full_name) return raw_events def run_query(self, query, query_args, query_errors): try: self.parse_args() self.pipeline = beam.Pipeline(options=self.pipeline_options) raw_events = self.generate_events() query.load(raw_events, query_args) result = self.pipeline.run() job_duration = ( self.pipeline_options.view_as(TestOptions).wait_until_finish_duration) if self.pipeline_options.view_as(StandardOptions).runner == 'DataflowRunner': # pylint: disable=line-too-long result.wait_until_finish(duration=job_duration) result.cancel() else: result.wait_until_finish() except Exception as exc: query_errors.append(str(exc)) raise def cleanup(self): publish_client = pubsub.Client(project=self.project) topic = publish_client.topic(self.topic_name) if topic.exists(): topic.delete() sub = topic.subscription(self.subscription_name) if sub.exists(): sub.delete() def run(self): queries = { 0: query0, 1: query1, 2: query2, # TODO(mariagh): Add more queries. } # TODO(mariagh): Move to a config file. query_args = {2: {'auction_id': 'a1003'}} query_errors = [] for i in self.args.query: self.parse_args() logging.info('Running query %d', i) # The DirectRunner is the default runner, and it needs # special handling to cancel streaming jobs. launch_from_direct_runner = self.pipeline_options.view_as( StandardOptions).runner in [None, 'DirectRunner'] query_duration = self.pipeline_options.view_as(TestOptions).wait_until_finish_duration # pylint: disable=line-too-long if launch_from_direct_runner: command = Command( self.run_query, args=[queries[i], query_args.get(i), query_errors]) command.run(timeout=query_duration // 1000) else: try: self.run_query(queries[i], query_args.get(i), query_errors=None) except Exception as exc: query_errors.append(exc) if query_errors: logging.error('Query failed with %s', ', '.join(query_errors)) else: logging.info('Queries run: %s', self.args.query) if __name__ == '__main__': launcher = NexmarkLauncher() launcher.run() launcher.cleanup()
sdks/python/apache_beam/testing/benchmarks/nexmark/nexmark_launcher.py
9,122
Nexmark launcher. The Nexmark suite is a series of queries (streaming pipelines) performed on a simulation of auction events. The launcher orchestrates the generation and parsing of streaming events and the running of queries. Model - Person: Author of an auction or a bid. - Auction: Item under auction. - Bid: A bid for an item under auction. Events - Create Person - Create Auction - Create Bid Queries - Query0: Pass through (send and receive auction events). Usage - DirectRunner python nexmark_launcher.py --query/q <query number> --project <project id> --loglevel=DEBUG (optional) --wait_until_finish_duration <time_in_ms> --streaming - DataflowRunner python nexmark_launcher.py --query/q <query number> --project <project id> --region <GCE region> --loglevel=DEBUG (optional) --wait_until_finish_duration <time_in_ms> --streaming --sdk_location <apache_beam tar.gz> --staging_location=gs://... --temp_location=gs:// Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. pytype: skip-file Usage with Dataflow requires a project to be supplied. Pub/Sub is currently available for use only in streaming pipelines. wait_until_finish ensures that the streaming job is canceled. pylint: disable=line-too-long We use the save_main_session option because one or more DoFn's in this workflow rely on global context (e.g., a module imported at module level). Read from PubSub into a PCollection. pylint: disable=line-too-long TODO(mariagh): Add more queries. TODO(mariagh): Move to a config file. The DirectRunner is the default runner, and it needs special handling to cancel streaming jobs. pylint: disable=line-too-long
2,493
en
0.748332
# # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # from rlstructures.logger import Logger, TFLogger from rlstructures import DictTensor, TemporalDictTensor from rlstructures import logging from rlstructures.tools import weight_init from rlstructures.batchers import EpisodeBatcher, Batcher import torch.nn as nn import copy import torch import time import numpy as np import torch.nn.functional as F from tutorial.tutorial_reinforce.agent import * class A2C: def __init__(self, config, create_env, create_train_env, create_agent): self.config = config # Creation of the Logger (that saves in tensorboard and CSV) self.logger = TFLogger(log_dir=self.config["logdir"], hps=self.config) self._create_env = create_env self._create_train_env = create_train_env self._create_agent = create_agent # Creation of one env instance to get the dimensionnality of observations and number of actions env = self._create_env( self.config["n_envs"], seed=0, env_name=self.config["env_name"] ) self.n_actions = env.action_space.n self.obs_dim = env.reset()[0]["frame"].size()[1] del env def run(self): # Instantiate the learning model abd the baseline model self.learning_model = AgentModel(self.obs_dim, self.n_actions, 32) self.critic_model = BaselineModel(self.obs_dim, 32) # We create a batcher dedicated to evaluation model = copy.deepcopy(self.learning_model) self.evaluation_batcher = EpisodeBatcher( n_timesteps=self.config["max_episode_steps"], n_slots=self.config["n_evaluation_episodes"], create_agent=self._create_agent, create_env=self._create_env, env_args={ "n_envs": self.config["n_envs"], "max_episode_steps": self.config["max_episode_steps"], "env_name": self.config["env_name"], }, agent_args={"n_actions": self.n_actions, "model": model}, n_threads=self.config["n_evaluation_threads"], seeds=[ self.config["env_seed"] + k * 10 for k in range(self.config["n_evaluation_threads"]) ], ) # Creation of the batcher for sampling complete pieces of trajectories (i.e Batcher) # The batcher will sample n_threads*n_envs trajectories at each call # To have a fast batcher, we have to configure it with n_timesteps=self.config["max_episode_steps"] model = copy.deepcopy(self.learning_model) self.train_batcher = Batcher( n_timesteps=self.config["a2c_timesteps"], n_slots=self.config["n_envs"] * self.config["n_threads"], create_agent=self._create_agent, create_env=self._create_train_env, env_args={ "n_envs": self.config["n_envs"], "max_episode_steps": self.config["max_episode_steps"], "env_name": self.config["env_name"], }, agent_args={"n_actions": self.n_actions, "model": model}, n_threads=self.config["n_threads"], seeds=[ self.config["env_seed"] + k * 10 for k in range(self.config["n_threads"]) ], ) # Creation of the optimizer optimizer = torch.optim.Adam( nn.Sequential(self.learning_model, self.critic_model).parameters(), lr=self.config["lr"], ) # Training Loop: _start_time = time.time() self.iteration = 0 # #We launch the evaluation batcher (in deterministic mode) n_episodes = self.config["n_evaluation_episodes"] agent_info = DictTensor( {"stochastic": torch.tensor([False]).repeat(n_episodes)} ) self.evaluation_batcher.execute(n_episodes=n_episodes, agent_info=agent_info) self.evaluation_iteration = self.iteration # Initialize the training batcher such that agents will start to acqire pieces of episodes self.train_batcher.update(self.learning_model.state_dict()) n_episodes = self.config["n_envs"] * self.config["n_threads"] agent_info = DictTensor({"stochastic": torch.tensor([True]).repeat(n_episodes)}) self.train_batcher.reset(agent_info=agent_info) while time.time() - _start_time < self.config["time_limit"]: # Call the batcher to get a sample of trajectories # 2) We get the pieces of episodes. Since the env is an infinite env, we will always receive a new piece of episode self.train_batcher.execute() trajectories = self.train_batcher.get(blocking=True) # 3) Now, we compute the loss dt = self.get_loss(trajectories) [self.logger.add_scalar(k, dt[k].item(), self.iteration) for k in dt.keys()] # Computation of final loss ld = self.config["critic_coef"] * dt["critic_loss"] lr = self.config["a2c_coef"] * dt["a2c_loss"] le = self.config["entropy_coef"] * dt["entropy_loss"] floss = ld - le - lr floss = floss / n_episodes * trajectories.n_elems() optimizer.zero_grad() floss.backward() optimizer.step() # Update the train batcher with the updated model self.train_batcher.update(self.learning_model.state_dict()) self.iteration += 1 # We check the evaluation batcher evaluation_trajectories = self.evaluation_batcher.get(blocking=False) if not evaluation_trajectories is None: # trajectories are available # Compute the cumulated reward cumulated_reward = ( ( evaluation_trajectories["_reward"] * evaluation_trajectories.mask() ) .sum(1) .mean() ) self.logger.add_scalar( "evaluation_reward", cumulated_reward.item(), self.evaluation_iteration, ) print( "At iteration %d, reward is %f" % (self.evaluation_iteration, cumulated_reward.item()) ) # We reexecute the evaluation batcher (with same value of agent_info and same number of episodes) self.evaluation_batcher.update(self.learning_model.state_dict()) self.evaluation_iteration = self.iteration self.evaluation_batcher.reexecute() self.train_batcher.close() self.evaluation_batcher.get() # To wait for the last trajectories self.evaluation_batcher.close() self.logger.update_csv() # To save as a CSV file in logdir self.logger.close() def get_loss(self, trajectories): # First, we want to compute the cumulated reward per trajectory # The reward is a t+1 in each iteration (since it is btained after the aaction), so we use the '_reward' field in the trajectory # The 'reward' field corresopnds to the reward at time t reward = trajectories["_reward"] # We get the mask that tells which transition is in a trajectory (1) or not (0) mask = trajectories.mask() # We remove the reward values that are not in the trajectories reward = reward * mask max_length = trajectories.lengths.max().item() # Now, we want to compute the action probabilities over the trajectories such that we will be able to do 'backward' action_probabilities = [] for t in range(max_length): proba = self.learning_model(trajectories["frame"][:, t]) action_probabilities.append( proba.unsqueeze(1) ) # We append the probability, and introduces the temporal dimension (2nde dimension) action_probabilities = torch.cat( action_probabilities, dim=1 ) # Now, we have a B x T x n_actions tensor # We compute the critic value for t=0 to T (i.e including the very last observation) critic = [] for t in range(max_length): b = self.critic_model(trajectories["frame"][:, t]) critic.append(b.unsqueeze(1)) critic = torch.cat(critic + [b.unsqueeze(1)], dim=1).squeeze( -1 ) # Now, we have a B x (T+1) tensor # We also need to compute the critic value at for the last observation of the trajectories (to compute the TD) # It may be the last element of the trajectories (if episode is not finished), or on the last frame of the episode idx = torch.arange(trajectories.n_elems()) last_critic = self.critic_model( trajectories["_frame"][idx, trajectories.lengths - 1] ).squeeze(-1) critic[idx, trajectories.lengths] = last_critic # We compute the temporal difference target = ( reward + self.config["discount_factor"] * (1 - trajectories["_done"].float()) * critic[:, 1:].detach() ) td = critic[:, :-1] - target critic_loss = td ** 2 # We sum the loss for each episode (considering the mask) critic_loss = (critic_loss * mask).sum(1) / mask.sum(1) # We average the loss over all the trajectories avg_critic_loss = critic_loss.mean() # We do the same on the reinforce loss action_distribution = torch.distributions.Categorical(action_probabilities) log_proba = action_distribution.log_prob(trajectories["action"]) a2c_loss = -log_proba * td.detach() a2c_loss = (a2c_loss * mask).sum(1) / mask.sum(1) avg_a2c_loss = a2c_loss.mean() # We compute the entropy loss entropy = action_distribution.entropy() entropy = (entropy * mask).sum(1) / mask.sum(1) avg_entropy = entropy.mean() return DictTensor( { "critic_loss": avg_critic_loss, "a2c_loss": avg_a2c_loss, "entropy_loss": avg_entropy, } )
tutorial/deprecated/tutorial_a2c_with_infinite_env/a2c.py
10,403
Copyright (c) Facebook, Inc. and its affiliates. This source code is licensed under the MIT license found in the LICENSE file in the root directory of this source tree. Creation of the Logger (that saves in tensorboard and CSV) Creation of one env instance to get the dimensionnality of observations and number of actions Instantiate the learning model abd the baseline model We create a batcher dedicated to evaluation Creation of the batcher for sampling complete pieces of trajectories (i.e Batcher) The batcher will sample n_threads*n_envs trajectories at each call To have a fast batcher, we have to configure it with n_timesteps=self.config["max_episode_steps"] Creation of the optimizer Training Loop: We launch the evaluation batcher (in deterministic mode) Initialize the training batcher such that agents will start to acqire pieces of episodes Call the batcher to get a sample of trajectories 2) We get the pieces of episodes. Since the env is an infinite env, we will always receive a new piece of episode 3) Now, we compute the loss Computation of final loss Update the train batcher with the updated model We check the evaluation batcher trajectories are available Compute the cumulated reward We reexecute the evaluation batcher (with same value of agent_info and same number of episodes) To wait for the last trajectories To save as a CSV file in logdir First, we want to compute the cumulated reward per trajectory The reward is a t+1 in each iteration (since it is btained after the aaction), so we use the '_reward' field in the trajectory The 'reward' field corresopnds to the reward at time t We get the mask that tells which transition is in a trajectory (1) or not (0) We remove the reward values that are not in the trajectories Now, we want to compute the action probabilities over the trajectories such that we will be able to do 'backward' We append the probability, and introduces the temporal dimension (2nde dimension) Now, we have a B x T x n_actions tensor We compute the critic value for t=0 to T (i.e including the very last observation) Now, we have a B x (T+1) tensor We also need to compute the critic value at for the last observation of the trajectories (to compute the TD) It may be the last element of the trajectories (if episode is not finished), or on the last frame of the episode We compute the temporal difference We sum the loss for each episode (considering the mask) We average the loss over all the trajectories We do the same on the reinforce loss We compute the entropy loss
2,527
en
0.880288
"""Utility functions for parcinging Freesurfer output files.""" from os.path import join import nibabel as nb import numpy as np def _vectorize_fs_surf(file_path): """ Read surface information from a file and turn it into a vector. Parameters ---------- file_path : str The path to a file with surface data. Returns ------- vectorized_data : numpy.ndarray Extracted data. """ img = nb.load(file_path) in_data = img.get_fdata().squeeze() return in_data def get_area(subject_dir, n_points): """ Read area information for the given subject and turn it into a vector. Data for left and right hemispheres are concatenated. Parameters ---------- subject_dir : str The directory to files with surface data. n_points : int Defines how many points to take from cortex surface. Returns ------- : numpy.ndarray Extracted data. """ AREA_FILES = ('lh.area.mgh', 'rh.area.mgh') lh_data = _vectorize_fs_surf(join(subject_dir, AREA_FILES[0])) rh_data = _vectorize_fs_surf(join(subject_dir, AREA_FILES[1])) n_points = n_points // 2 return np.concatenate((lh_data[:n_points], rh_data[:n_points]), 0) def get_thickness(subject_dir, n_points): """ Read thickness information for the given subject and turn it into a vector. Data for left and right hemispheres are concatenated. Parameters ---------- subject_dir : str The directory to files with surface data. n_points : int Defines how many points to take from cortex surface. Returns ------- : numpy.ndarray Extracted data. """ THICKNESS_FILES = ('rh.thickness.mgh', 'lh.thickness.mgh') lh_data = _vectorize_fs_surf(join(subject_dir, THICKNESS_FILES[0])) rh_data = _vectorize_fs_surf(join(subject_dir, THICKNESS_FILES[1])) n_points = n_points // 2 return np.concatenate((lh_data[:n_points], rh_data[:n_points]), 0)
camcan/utils/file_parsing.py
2,013
Read surface information from a file and turn it into a vector. Parameters ---------- file_path : str The path to a file with surface data. Returns ------- vectorized_data : numpy.ndarray Extracted data. Read area information for the given subject and turn it into a vector. Data for left and right hemispheres are concatenated. Parameters ---------- subject_dir : str The directory to files with surface data. n_points : int Defines how many points to take from cortex surface. Returns ------- : numpy.ndarray Extracted data. Read thickness information for the given subject and turn it into a vector. Data for left and right hemispheres are concatenated. Parameters ---------- subject_dir : str The directory to files with surface data. n_points : int Defines how many points to take from cortex surface. Returns ------- : numpy.ndarray Extracted data. Utility functions for parcinging Freesurfer output files.
952
en
0.677209
import builtins import os import sys from array import array from collections import Counter, defaultdict, deque from dataclasses import dataclass, fields, is_dataclass from itertools import islice from typing import ( TYPE_CHECKING, Any, Callable, Dict, Iterable, List, Optional, Set, Union, Tuple, ) from rich.highlighter import ReprHighlighter from . import get_console from ._loop import loop_last from ._pick import pick_bool from .abc import RichRenderable from .cells import cell_len from .highlighter import ReprHighlighter from .jupyter import JupyterMixin, JupyterRenderable from .measure import Measurement from .text import Text if TYPE_CHECKING: from .console import ( Console, ConsoleOptions, HighlighterType, JustifyMethod, OverflowMethod, RenderResult, ) def install( console: "Console" = None, overflow: "OverflowMethod" = "ignore", crop: bool = False, indent_guides: bool = False, max_length: int = None, max_string: int = None, expand_all: bool = False, ) -> None: """Install automatic pretty printing in the Python REPL. Args: console (Console, optional): Console instance or ``None`` to use global console. Defaults to None. overflow (Optional[OverflowMethod], optional): Overflow method. Defaults to "ignore". crop (Optional[bool], optional): Enable cropping of long lines. Defaults to False. indent_guides (bool, optional): Enable indentation guides. Defaults to False. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable. Defaults to None. expand_all (bool, optional): Expand all containers. Defaults to False """ from rich import get_console from .console import ConsoleRenderable # needed here to prevent circular import console = console or get_console() assert console is not None def display_hook(value: Any) -> None: """Replacement sys.displayhook which prettifies objects with Rich.""" if value is not None: assert console is not None builtins._ = None # type: ignore console.print( value if isinstance(value, RichRenderable) else Pretty( value, overflow=overflow, indent_guides=indent_guides, max_length=max_length, max_string=max_string, expand_all=expand_all, ), crop=crop, ) builtins._ = value # type: ignore def ipy_display_hook(value: Any) -> None: # pragma: no cover assert console is not None # always skip rich generated jupyter renderables or None values if isinstance(value, JupyterRenderable) or value is None: return # on jupyter rich display, if using one of the special representations dont use rich if console.is_jupyter and any(attr.startswith("_repr_") for attr in dir(value)): return if hasattr(value, "_repr_mimebundle_"): return # certain renderables should start on a new line if isinstance(value, ConsoleRenderable): console.line() console.print( value if isinstance(value, RichRenderable) else Pretty( value, overflow=overflow, indent_guides=indent_guides, max_length=max_length, max_string=max_string, expand_all=expand_all, margin=12, ), crop=crop, ) try: # pragma: no cover ip = get_ipython() # type: ignore from IPython.core.formatters import BaseFormatter # replace plain text formatter with rich formatter rich_formatter = BaseFormatter() rich_formatter.for_type(object, func=ipy_display_hook) ip.display_formatter.formatters["text/plain"] = rich_formatter except Exception: sys.displayhook = display_hook class Pretty(JupyterMixin): """A rich renderable that pretty prints an object. Args: _object (Any): An object to pretty print. highlighter (HighlighterType, optional): Highlighter object to apply to result, or None for ReprHighlighter. Defaults to None. indent_size (int, optional): Number of spaces in indent. Defaults to 4. justify (JustifyMethod, optional): Justify method, or None for default. Defaults to None. overflow (OverflowMethod, optional): Overflow method, or None for default. Defaults to None. no_wrap (Optional[bool], optional): Disable word wrapping. Defaults to False. indent_guides (bool, optional): Enable indentation guides. Defaults to False. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable. Defaults to None. expand_all (bool, optional): Expand all containers. Defaults to False. margin (int, optional): Subtrace a margin from width to force containers to expand earlier. Defaults to 0. insert_line (bool, optional): Insert a new line if the output has multiple new lines. Defaults to False. """ def __init__( self, _object: Any, highlighter: "HighlighterType" = None, *, indent_size: int = 4, justify: "JustifyMethod" = None, overflow: Optional["OverflowMethod"] = None, no_wrap: Optional[bool] = False, indent_guides: bool = False, max_length: int = None, max_string: int = None, expand_all: bool = False, margin: int = 0, insert_line: bool = False, ) -> None: self._object = _object self.highlighter = highlighter or ReprHighlighter() self.indent_size = indent_size self.justify = justify self.overflow = overflow self.no_wrap = no_wrap self.indent_guides = indent_guides self.max_length = max_length self.max_string = max_string self.expand_all = expand_all self.margin = margin self.insert_line = insert_line def __rich_console__( self, console: "Console", options: "ConsoleOptions" ) -> "RenderResult": pretty_str = pretty_repr( self._object, max_width=options.max_width - self.margin, indent_size=self.indent_size, max_length=self.max_length, max_string=self.max_string, expand_all=self.expand_all, ) pretty_text = Text( pretty_str, justify=self.justify or options.justify, overflow=self.overflow or options.overflow, no_wrap=pick_bool(self.no_wrap, options.no_wrap), style="pretty", ) pretty_text = ( self.highlighter(pretty_text) if pretty_text else Text( f"{type(self._object)}.__repr__ returned empty string", style="dim italic", ) ) if self.indent_guides and not options.ascii_only: pretty_text = pretty_text.with_indent_guides( self.indent_size, style="repr.indent" ) if self.insert_line and "\n" in pretty_text: yield "" yield pretty_text def __rich_measure__( self, console: "Console", options: "ConsoleOptions" ) -> "Measurement": pretty_str = pretty_repr( self._object, max_width=options.max_width, indent_size=self.indent_size, max_length=self.max_length, max_string=self.max_string, ) text_width = ( max(cell_len(line) for line in pretty_str.splitlines()) if pretty_str else 0 ) return Measurement(text_width, text_width) def _get_braces_for_defaultdict(_object: defaultdict) -> Tuple[str, str, str]: return ( f"defaultdict({_object.default_factory!r}, {{", "})", f"defaultdict({_object.default_factory!r}, {{}})", ) def _get_braces_for_array(_object: array) -> Tuple[str, str, str]: return (f"array({_object.typecode!r}, [", "])", "array({_object.typecode!r})") _BRACES: Dict[type, Callable[[Any], Tuple[str, str, str]]] = { os._Environ: lambda _object: ("environ({", "})", "environ({})"), array: _get_braces_for_array, defaultdict: _get_braces_for_defaultdict, Counter: lambda _object: ("Counter({", "})", "Counter()"), deque: lambda _object: ("deque([", "])", "deque()"), dict: lambda _object: ("{", "}", "{}"), frozenset: lambda _object: ("frozenset({", "})", "frozenset()"), list: lambda _object: ("[", "]", "[]"), set: lambda _object: ("{", "}", "set()"), tuple: lambda _object: ("(", ")", "()"), } _CONTAINERS = tuple(_BRACES.keys()) _MAPPING_CONTAINERS = (dict, os._Environ) def is_expandable(obj: Any) -> bool: """Check if an object may be expanded by pretty print.""" return ( isinstance(obj, _CONTAINERS) or (is_dataclass(obj) and not isinstance(obj, type)) or hasattr(obj, "__rich_repr__") ) @dataclass class Node: """A node in a repr tree. May be atomic or a container.""" key_repr: str = "" value_repr: str = "" open_brace: str = "" close_brace: str = "" empty: str = "" last: bool = False is_tuple: bool = False children: Optional[List["Node"]] = None key_separator = ": " @property def separator(self) -> str: """Get separator between items.""" return "" if self.last else "," def iter_tokens(self) -> Iterable[str]: """Generate tokens for this node.""" if self.key_repr: yield self.key_repr yield self.key_separator if self.value_repr: yield self.value_repr elif self.children is not None: if self.children: yield self.open_brace if self.is_tuple and len(self.children) == 1: yield from self.children[0].iter_tokens() yield "," else: for child in self.children: yield from child.iter_tokens() if not child.last: yield ", " yield self.close_brace else: yield self.empty def check_length(self, start_length: int, max_length: int) -> bool: """Check the length fits within a limit. Args: start_length (int): Starting length of the line (indent, prefix, suffix). max_length (int): Maximum length. Returns: bool: True if the node can be rendered within max length, otherwise False. """ total_length = start_length for token in self.iter_tokens(): total_length += cell_len(token) if total_length > max_length: return False return True def __str__(self) -> str: repr_text = "".join(self.iter_tokens()) return repr_text def render( self, max_width: int = 80, indent_size: int = 4, expand_all: bool = False ) -> str: """Render the node to a pretty repr. Args: max_width (int, optional): Maximum width of the repr. Defaults to 80. indent_size (int, optional): Size of indents. Defaults to 4. expand_all (bool, optional): Expand all levels. Defaults to False. Returns: str: A repr string of the original object. """ lines = [_Line(node=self, is_root=True)] line_no = 0 while line_no < len(lines): line = lines[line_no] if line.expandable and not line.expanded: if expand_all or not line.check_length(max_width): lines[line_no : line_no + 1] = line.expand(indent_size) line_no += 1 repr_str = "\n".join(str(line) for line in lines) return repr_str @dataclass class _Line: """A line in repr output.""" is_root: bool = False node: Optional[Node] = None text: str = "" suffix: str = "" whitespace: str = "" expanded: bool = False @property def expandable(self) -> bool: """Check if the line may be expanded.""" return bool(self.node is not None and self.node.children) def check_length(self, max_length: int) -> bool: """Check this line fits within a given number of cells.""" start_length = ( len(self.whitespace) + cell_len(self.text) + cell_len(self.suffix) ) assert self.node is not None return self.node.check_length(start_length, max_length) def expand(self, indent_size: int) -> Iterable["_Line"]: """Expand this line by adding children on their own line.""" node = self.node assert node is not None whitespace = self.whitespace assert node.children if node.key_repr: yield _Line( text=f"{node.key_repr}{node.key_separator}{node.open_brace}", whitespace=whitespace, ) else: yield _Line(text=node.open_brace, whitespace=whitespace) child_whitespace = self.whitespace + " " * indent_size tuple_of_one = node.is_tuple and len(node.children) == 1 for child in node.children: separator = "," if tuple_of_one else child.separator line = _Line( node=child, whitespace=child_whitespace, suffix=separator, ) yield line yield _Line( text=node.close_brace, whitespace=whitespace, suffix="," if (tuple_of_one and not self.is_root) else node.separator, ) def __str__(self) -> str: return f"{self.whitespace}{self.text}{self.node or ''}{self.suffix}" def traverse(_object: Any, max_length: int = None, max_string: int = None) -> Node: """Traverse object and generate a tree. Args: _object (Any): Object to be traversed. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable truncating. Defaults to None. Returns: Node: The root of a tree structure which can be used to render a pretty repr. """ def to_repr(obj: Any) -> str: """Get repr string for an object, but catch errors.""" if ( max_string is not None and isinstance(obj, (bytes, str)) and len(obj) > max_string ): truncated = len(obj) - max_string obj_repr = f"{obj[:max_string]!r}+{truncated}" else: try: obj_repr = repr(obj) except Exception as error: obj_repr = f"<repr-error '{error}'>" return obj_repr visited_ids: Set[int] = set() push_visited = visited_ids.add pop_visited = visited_ids.remove def _traverse(obj: Any, root: bool = False) -> Node: """Walk the object depth first.""" obj_type = type(obj) py_version = (sys.version_info.major, sys.version_info.minor) children: List[Node] def iter_rich_args(rich_args) -> Iterable[Union[Any, Tuple[str, Any]]]: for arg in rich_args: if isinstance(arg, tuple): if len(arg) == 3: key, child, default = arg if default == child: continue yield key, child elif len(arg) == 2: key, child = arg yield key, child elif len(arg) == 1: yield arg[0] else: yield arg if hasattr(obj, "__rich_repr__"): args = list(iter_rich_args(obj.__rich_repr__())) if args: children = [] append = children.append node = Node( open_brace=f"{obj.__class__.__name__}(", close_brace=")", children=children, last=root, ) for last, arg in loop_last(args): if isinstance(arg, tuple): key, child = arg child_node = _traverse(child) child_node.last = last child_node.key_repr = key child_node.last = last child_node.key_separator = "=" append(child_node) else: child_node = _traverse(arg) child_node.last = last append(child_node) else: node = Node( value_repr=f"{obj.__class__.__name__}()", children=[], last=root ) elif ( is_dataclass(obj) and not isinstance(obj, type) and ( "__create_fn__" in obj.__repr__.__qualname__ or py_version == (3, 6) ) # Check if __repr__ wasn't overriden ): obj_id = id(obj) if obj_id in visited_ids: # Recursion detected return Node(value_repr="...") push_visited(obj_id) children = [] append = children.append node = Node( open_brace=f"{obj.__class__.__name__}(", close_brace=")", children=children, last=root, ) for last, field in loop_last(fields(obj)): if field.repr: child_node = _traverse(getattr(obj, field.name)) child_node.key_repr = field.name child_node.last = last child_node.key_separator = "=" append(child_node) pop_visited(obj_id) elif obj_type in _CONTAINERS: obj_id = id(obj) if obj_id in visited_ids: # Recursion detected return Node(value_repr="...") push_visited(obj_id) open_brace, close_brace, empty = _BRACES[obj_type](obj) if obj: children = [] node = Node( open_brace=open_brace, close_brace=close_brace, children=children, last=root, ) append = children.append num_items = len(obj) last_item_index = num_items - 1 if isinstance(obj, _MAPPING_CONTAINERS): iter_items = iter(obj.items()) if max_length is not None: iter_items = islice(iter_items, max_length) for index, (key, child) in enumerate(iter_items): child_node = _traverse(child) child_node.key_repr = to_repr(key) child_node.last = index == last_item_index append(child_node) else: iter_values = iter(obj) if max_length is not None: iter_values = islice(iter_values, max_length) for index, child in enumerate(iter_values): child_node = _traverse(child) child_node.last = index == last_item_index append(child_node) if max_length is not None and num_items > max_length: append(Node(value_repr=f"... +{num_items-max_length}", last=True)) else: node = Node(empty=empty, children=[], last=root) pop_visited(obj_id) else: node = Node(value_repr=to_repr(obj), last=root) node.is_tuple = isinstance(obj, tuple) return node node = _traverse(_object, root=True) return node def pretty_repr( _object: Any, *, max_width: int = 80, indent_size: int = 4, max_length: int = None, max_string: int = None, expand_all: bool = False, ) -> str: """Prettify repr string by expanding on to new lines to fit within a given width. Args: _object (Any): Object to repr. max_width (int, optional): Desired maximum width of repr string. Defaults to 80. indent_size (int, optional): Number of spaces to indent. Defaults to 4. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable truncating. Defaults to None. expand_all (bool, optional): Expand all containers regardless of available width. Defaults to False. Returns: str: A possibly multi-line representation of the object. """ if isinstance(_object, Node): node = _object else: node = traverse(_object, max_length=max_length, max_string=max_string) repr_str = node.render( max_width=max_width, indent_size=indent_size, expand_all=expand_all ) return repr_str def pprint( _object: Any, *, console: "Console" = None, indent_guides: bool = True, max_length: int = None, max_string: int = None, expand_all: bool = False, ): """A convenience function for pretty printing. Args: _object (Any): Object to pretty print. console (Console, optional): Console instance, or None to use default. Defaults to None. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of strings before truncating, or None to disable. Defaults to None. indent_guides (bool, optional): Enable indentation guides. Defaults to True. expand_all (bool, optional): Expand all containers. Defaults to False. """ _console = get_console() if console is None else console _console.print( Pretty( _object, max_length=max_length, max_string=max_string, indent_guides=indent_guides, expand_all=expand_all, overflow="ignore", ), soft_wrap=True, ) if __name__ == "__main__": # pragma: no cover class BrokenRepr: def __repr__(self): 1 / 0 d = defaultdict(int) d["foo"] = 5 data = { "foo": [ 1, "Hello World!", 100.123, 323.232, 432324.0, {5, 6, 7, (1, 2, 3, 4), 8}, ], "bar": frozenset({1, 2, 3}), "defaultdict": defaultdict( list, {"crumble": ["apple", "rhubarb", "butter", "sugar", "flour"]} ), "counter": Counter( [ "apple", "orange", "pear", "kumquat", "kumquat", "durian" * 100, ] ), "atomic": (False, True, None), "Broken": BrokenRepr(), } data["foo"].append(data) # type: ignore from rich import print print(Pretty(data, indent_guides=True, max_string=20))
rich/pretty.py
24,145
A node in a repr tree. May be atomic or a container. A rich renderable that pretty prints an object. Args: _object (Any): An object to pretty print. highlighter (HighlighterType, optional): Highlighter object to apply to result, or None for ReprHighlighter. Defaults to None. indent_size (int, optional): Number of spaces in indent. Defaults to 4. justify (JustifyMethod, optional): Justify method, or None for default. Defaults to None. overflow (OverflowMethod, optional): Overflow method, or None for default. Defaults to None. no_wrap (Optional[bool], optional): Disable word wrapping. Defaults to False. indent_guides (bool, optional): Enable indentation guides. Defaults to False. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable. Defaults to None. expand_all (bool, optional): Expand all containers. Defaults to False. margin (int, optional): Subtrace a margin from width to force containers to expand earlier. Defaults to 0. insert_line (bool, optional): Insert a new line if the output has multiple new lines. Defaults to False. A line in repr output. Walk the object depth first. Check the length fits within a limit. Args: start_length (int): Starting length of the line (indent, prefix, suffix). max_length (int): Maximum length. Returns: bool: True if the node can be rendered within max length, otherwise False. Check this line fits within a given number of cells. Replacement sys.displayhook which prettifies objects with Rich. Expand this line by adding children on their own line. Check if the line may be expanded. Install automatic pretty printing in the Python REPL. Args: console (Console, optional): Console instance or ``None`` to use global console. Defaults to None. overflow (Optional[OverflowMethod], optional): Overflow method. Defaults to "ignore". crop (Optional[bool], optional): Enable cropping of long lines. Defaults to False. indent_guides (bool, optional): Enable indentation guides. Defaults to False. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable. Defaults to None. expand_all (bool, optional): Expand all containers. Defaults to False Check if an object may be expanded by pretty print. Generate tokens for this node. A convenience function for pretty printing. Args: _object (Any): Object to pretty print. console (Console, optional): Console instance, or None to use default. Defaults to None. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of strings before truncating, or None to disable. Defaults to None. indent_guides (bool, optional): Enable indentation guides. Defaults to True. expand_all (bool, optional): Expand all containers. Defaults to False. Prettify repr string by expanding on to new lines to fit within a given width. Args: _object (Any): Object to repr. max_width (int, optional): Desired maximum width of repr string. Defaults to 80. indent_size (int, optional): Number of spaces to indent. Defaults to 4. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable truncating. Defaults to None. expand_all (bool, optional): Expand all containers regardless of available width. Defaults to False. Returns: str: A possibly multi-line representation of the object. Render the node to a pretty repr. Args: max_width (int, optional): Maximum width of the repr. Defaults to 80. indent_size (int, optional): Size of indents. Defaults to 4. expand_all (bool, optional): Expand all levels. Defaults to False. Returns: str: A repr string of the original object. Get separator between items. Get repr string for an object, but catch errors. Traverse object and generate a tree. Args: _object (Any): Object to be traversed. max_length (int, optional): Maximum length of containers before abbreviating, or None for no abbreviation. Defaults to None. max_string (int, optional): Maximum length of string before truncating, or None to disable truncating. Defaults to None. Returns: Node: The root of a tree structure which can be used to render a pretty repr. needed here to prevent circular import type: ignore type: ignore pragma: no cover always skip rich generated jupyter renderables or None values on jupyter rich display, if using one of the special representations dont use rich certain renderables should start on a new line pragma: no cover type: ignore replace plain text formatter with rich formatter Check if __repr__ wasn't overriden Recursion detected Recursion detected pragma: no cover type: ignore
5,217
en
0.629515
import json import os import sys import time import torch from training.training import Trainer from data.conversion import GridDataConverter, PointCloudDataConverter, ERA5Converter from data.dataloaders import mnist, celebahq from data.dataloaders_era5 import era5 from data.dataloaders3d import shapenet_voxels, shapenet_point_clouds from models.discriminator import PointConvDiscriminator from models.function_distribution import HyperNetwork, FunctionDistribution from models.function_representation import FunctionRepresentation, FourierFeatures def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad) device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') # Get config file from command line arguments if len(sys.argv) != 2: raise(RuntimeError("Wrong arguments, use python main.py <config_path>")) config_path = sys.argv[1] # Open config file with open(config_path) as f: config = json.load(f) if config["path_to_data"] == "": raise(RuntimeError("Path to data not specified. Modify path_to_data attribute in config to point to data.")) # Create a folder to store experiment results timestamp = time.strftime("%Y-%m-%d_%H-%M") directory = "{}_{}".format(timestamp, config["id"]) if not os.path.exists(directory): os.makedirs(directory) # Save config file in experiment directory with open(directory + '/config.json', 'w') as f: json.dump(config, f) # Setup dataloader is_voxel = False is_point_cloud = False is_era5 = False if config["dataset"] == 'mnist': dataloader = mnist(path_to_data=config["path_to_data"], batch_size=config["training"]["batch_size"], size=config["resolution"], train=True) input_dim = 2 output_dim = 1 data_shape = (1, config["resolution"], config["resolution"]) elif config["dataset"] == 'celebahq': dataloader = celebahq(path_to_data=config["path_to_data"], batch_size=config["training"]["batch_size"], size=config["resolution"]) input_dim = 2 output_dim = 3 data_shape = (3, config["resolution"], config["resolution"]) elif config["dataset"] == 'shapenet_voxels': dataloader = shapenet_voxels(path_to_data=config["path_to_data"], batch_size=config["training"]["batch_size"], size=config["resolution"]) input_dim = 3 output_dim = 1 data_shape = (1, config["resolution"], config["resolution"], config["resolution"]) is_voxel = True elif config["dataset"] == 'shapenet_point_clouds': dataloader = shapenet_point_clouds(path_to_data=config["path_to_data"], batch_size=config["training"]["batch_size"]) input_dim = 3 output_dim = 1 data_shape = (1, config["resolution"], config["resolution"], config["resolution"]) is_point_cloud = True elif config["dataset"] == 'era5': dataloader = era5(path_to_data=config["path_to_data"], batch_size=config["training"]["batch_size"]) input_dim = 3 output_dim = 1 data_shape = (46, 90) is_era5 = True # Setup data converter if is_point_cloud: data_converter = PointCloudDataConverter(device, data_shape, normalize_features=True) elif is_era5: data_converter = ERA5Converter(device, data_shape, normalize_features=True) else: data_converter = GridDataConverter(device, data_shape, normalize_features=True) # Setup encoding for function distribution num_frequencies = config["generator"]["encoding"]["num_frequencies"] std_dev = config["generator"]["encoding"]["std_dev"] if num_frequencies: frequency_matrix = torch.normal(mean=torch.zeros(num_frequencies, input_dim), std=std_dev).to(device) encoding = FourierFeatures(frequency_matrix) else: encoding = torch.nn.Identity() # Setup generator models final_non_linearity = torch.nn.Tanh() non_linearity = torch.nn.LeakyReLU(0.1) function_representation = FunctionRepresentation(input_dim, output_dim, config["generator"]["layer_sizes"], encoding, non_linearity, final_non_linearity).to(device) hypernetwork = HyperNetwork(function_representation, config["generator"]["latent_dim"], config["generator"]["hypernet_layer_sizes"], non_linearity).to(device) function_distribution = FunctionDistribution(hypernetwork).to(device) # Setup discriminator discriminator = PointConvDiscriminator(input_dim, output_dim, config["discriminator"]["layer_configs"], linear_layer_sizes=config["discriminator"]["linear_layer_sizes"], norm_order=config["discriminator"]["norm_order"], add_sigmoid=True, add_batchnorm=config["discriminator"]["add_batchnorm"], add_weightnet_batchnorm=config["discriminator"]["add_weightnet_batchnorm"], deterministic=config["discriminator"]["deterministic"], same_coordinates=config["discriminator"]["same_coordinates"]).to(device) print("\nFunction distribution") print(hypernetwork) print("Number of parameters: {}".format(count_parameters(hypernetwork))) print("\nDiscriminator") print(discriminator) print("Number of parameters: {}".format(count_parameters(discriminator))) # Setup trainer trainer = Trainer(device, function_distribution, discriminator, data_converter, lr=config["training"]["lr"], lr_disc=config["training"]["lr_disc"], r1_weight=config["training"]["r1_weight"], max_num_points=config["training"]["max_num_points"], print_freq=config["training"]["print_freq"], save_dir=directory, model_save_freq=config["training"]["model_save_freq"], is_voxel=is_voxel, is_point_cloud=is_point_cloud, is_era5=is_era5) trainer.train(dataloader, config["training"]["epochs"])
main.py
6,283
Get config file from command line arguments Open config file Create a folder to store experiment results Save config file in experiment directory Setup dataloader Setup data converter Setup encoding for function distribution Setup generator models Setup discriminator Setup trainer
281
en
0.627002
#! /usr/bin/env python # Convert OpenSSH known_hosts and known_hosts2 files to "new format" PuTTY # host keys. # usage: # kh2reg.py [ --win ] known_hosts1 2 3 4 ... > hosts.reg # Creates a Windows .REG file (double-click to install). # kh2reg.py --unix known_hosts1 2 3 4 ... > sshhostkeys # Creates data suitable for storing in ~/.putty/sshhostkeys (Unix). # Line endings are someone else's problem as is traditional. # Originally developed for Python 1.5.2, but probably won't run on that # any more. import fileinput import base64 import struct import string import re import sys import getopt def winmungestr(s): "Duplicate of PuTTY's mungestr() in winstore.c:1.10 for Registry keys" candot = 0 r = "" for c in s: if c in ' \*?%~' or ord(c)<ord(' ') or (c == '.' and not candot): r = r + ("%%%02X" % ord(c)) else: r = r + c candot = 1 return r def strtolong(s): "Convert arbitrary-length big-endian binary data to a Python long" bytes = struct.unpack(">%luB" % len(s), s) return reduce ((lambda a, b: (long(a) << 8) + long(b)), bytes) def longtohex(n): """Convert long int to lower-case hex. Ick, Python (at least in 1.5.2) doesn't appear to have a way to turn a long int into an unadorned hex string -- % gets upset if the number is too big, and raw hex() uses uppercase (sometimes), and adds unwanted "0x...L" around it.""" plain=string.lower(re.match(r"0x([0-9A-Fa-f]*)l?$", hex(n), re.I).group(1)) return "0x" + plain output_type = 'windows' try: optlist, args = getopt.getopt(sys.argv[1:], '', [ 'win', 'unix' ]) if filter(lambda x: x[0] == '--unix', optlist): output_type = 'unix' except getopt.error, e: sys.stderr.write(str(e) + "\n") sys.exit(1) if output_type == 'windows': # Output REG file header. sys.stdout.write("""REGEDIT4 [HKEY_CURRENT_USER\Software\SimonTatham\PuTTY\SshHostKeys] """) class BlankInputLine(Exception): pass class UnknownKeyType(Exception): def __init__(self, keytype): self.keytype = keytype # Now process all known_hosts input. for line in fileinput.input(args): try: # Remove leading/trailing whitespace (should zap CR and LF) line = string.strip (line) # Skip blanks and comments if line == '' or line[0] == '#': raise BlankInputLine # Split line on spaces. fields = string.split (line, ' ') # Common fields hostpat = fields[0] magicnumbers = [] # placeholder keytype = "" # placeholder # Grotty heuristic to distinguish known_hosts from known_hosts2: # is second field entirely decimal digits? if re.match (r"\d*$", fields[1]): # Treat as SSH-1-type host key. # Format: hostpat bits10 exp10 mod10 comment... # (PuTTY doesn't store the number of bits.) magicnumbers = map (long, fields[2:4]) keytype = "rsa" else: # Treat as SSH-2-type host key. # Format: hostpat keytype keyblob64 comment... sshkeytype, blob = fields[1], base64.decodestring (fields[2]) # 'blob' consists of a number of # uint32 N (big-endian) # uint8[N] field_data subfields = [] while blob: sizefmt = ">L" (size,) = struct.unpack (sizefmt, blob[0:4]) size = int(size) # req'd for slicage (data,) = struct.unpack (">%lus" % size, blob[4:size+4]) subfields.append(data) blob = blob [struct.calcsize(sizefmt) + size : ] # The first field is keytype again, and the rest we can treat as # an opaque list of bignums (same numbers and order as stored # by PuTTY). (currently embedded keytype is ignored entirely) magicnumbers = map (strtolong, subfields[1:]) # Translate key type into something PuTTY can use. if sshkeytype == "ssh-rsa": keytype = "rsa2" elif sshkeytype == "ssh-dss": keytype = "dss" else: raise UnknownKeyType(sshkeytype) # Now print out one line per host pattern, discarding wildcards. for host in string.split (hostpat, ','): if re.search (r"[*?!]", host): sys.stderr.write("Skipping wildcard host pattern '%s'\n" % host) continue elif re.match (r"\|", host): sys.stderr.write("Skipping hashed hostname '%s'\n" % host) continue else: m = re.match (r"\[([^]]*)\]:(\d*)$", host) if m: (host, port) = m.group(1,2) port = int(port) else: port = 22 # Slightly bizarre output key format: 'type@port:hostname' # XXX: does PuTTY do anything useful with literal IP[v4]s? key = keytype + ("@%d:%s" % (port, host)) value = string.join (map (longtohex, magicnumbers), ',') if output_type == 'unix': # Unix format. sys.stdout.write('%s %s\n' % (key, value)) else: # Windows format. # XXX: worry about double quotes? sys.stdout.write("\"%s\"=\"%s\"\n" % (winmungestr(key), value)) except UnknownKeyType, k: sys.stderr.write("Unknown SSH key type '%s', skipping\n" % k.keytype) except BlankInputLine: pass
contrib/kh2reg.py
5,752
! /usr/bin/env python Convert OpenSSH known_hosts and known_hosts2 files to "new format" PuTTY host keys. usage: kh2reg.py [ --win ] known_hosts1 2 3 4 ... > hosts.reg Creates a Windows .REG file (double-click to install). kh2reg.py --unix known_hosts1 2 3 4 ... > sshhostkeys Creates data suitable for storing in ~/.putty/sshhostkeys (Unix). Line endings are someone else's problem as is traditional. Originally developed for Python 1.5.2, but probably won't run on that any more. Output REG file header. Now process all known_hosts input. Remove leading/trailing whitespace (should zap CR and LF) Skip blanks and comments Split line on spaces. Common fields placeholder placeholder Grotty heuristic to distinguish known_hosts from known_hosts2: is second field entirely decimal digits? Treat as SSH-1-type host key. Format: hostpat bits10 exp10 mod10 comment... (PuTTY doesn't store the number of bits.) Treat as SSH-2-type host key. Format: hostpat keytype keyblob64 comment... 'blob' consists of a number of uint32 N (big-endian) uint8[N] field_data req'd for slicage The first field is keytype again, and the rest we can treat as an opaque list of bignums (same numbers and order as stored by PuTTY). (currently embedded keytype is ignored entirely) Translate key type into something PuTTY can use. Now print out one line per host pattern, discarding wildcards. Slightly bizarre output key format: 'type@port:hostname' XXX: does PuTTY do anything useful with literal IP[v4]s? Unix format. Windows format. XXX: worry about double quotes?
1,574
en
0.833724
""" Copyright 2020 The Magma Authors. This source code is licensed under the BSD-style license found in the LICENSE file in the root directory of this source tree. Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import logging import subprocess from typing import Optional # noqa from lte.protos.policydb_pb2 import FlowMatch from .tc_ops_cmd import TcOpsCmd, argSplit, run_cmd from .tc_ops_pyroute2 import TcOpsPyRoute2 from .types import QosInfo from .utils import IdManager LOG = logging.getLogger('pipelined.qos.qos_tc_impl') # LOG.setLevel(logging.DEBUG) # TODO - replace this implementation with pyroute2 tc ROOT_QID = 65534 DEFAULT_RATE = '80Kbit' DEFAULT_INTF_SPEED = '1000' class TrafficClass: """ Creates/Deletes queues in linux. Using Qdiscs for flow based rate limiting(traffic shaping) of user traffic. """ tc_ops = None @staticmethod def delete_class(intf: str, qid: int, skip_filter=False) -> int: qid_hex = hex(qid) if not skip_filter: TrafficClass.tc_ops.del_filter(intf, qid_hex, qid_hex) return TrafficClass.tc_ops.del_htb(intf, qid_hex) @staticmethod def create_class( intf: str, qid: int, max_bw: int, rate=None, parent_qid=None, skip_filter=False, ) -> int: if not rate: rate = DEFAULT_RATE if not parent_qid: parent_qid = ROOT_QID if parent_qid == qid: # parent qid should only be self for root case, everything else # should be the child of root class LOG.error('parent and self qid equal, setting parent_qid to root') parent_qid = ROOT_QID qid_hex = hex(qid) parent_qid_hex = '1:' + hex(parent_qid) err = TrafficClass.tc_ops.create_htb(intf, qid_hex, max_bw, rate, parent_qid_hex) if err < 0 or skip_filter: return err # add filter return TrafficClass.tc_ops.create_filter(intf, qid_hex, qid_hex) @staticmethod def init_qdisc( intf: str, show_error=False, enable_pyroute2=False, default_gbr=DEFAULT_RATE, ) -> int: # TODO: Convert this class into an object. if TrafficClass.tc_ops is None: if enable_pyroute2: TrafficClass.tc_ops = TcOpsPyRoute2() else: TrafficClass.tc_ops = TcOpsCmd() cmd_list = [] speed = DEFAULT_INTF_SPEED qid_hex = hex(ROOT_QID) fn = "/sys/class/net/{intf}/speed".format(intf=intf) try: with open(fn, encoding="utf-8") as f: speed = f.read().strip() except OSError: LOG.error('unable to read speed from %s defaulting to %s', fn, speed) # qdisc does not support replace, so check it before creating the HTB qdisc. qdisc_type = TrafficClass._get_qdisc_type(intf) if qdisc_type != "htb": qdisc_cmd = "tc qdisc add dev {intf} root handle 1: htb".format(intf=intf) cmd_list.append(qdisc_cmd) LOG.info("Created root qdisc") parent_q_cmd = "tc class replace dev {intf} parent 1: classid 1:{root_qid} htb " parent_q_cmd += "rate {speed}Mbit ceil {speed}Mbit" parent_q_cmd = parent_q_cmd.format(intf=intf, root_qid=qid_hex, speed=speed) cmd_list.append(parent_q_cmd) tc_cmd = "tc class replace dev {intf} parent 1:{root_qid} classid 1:1 htb " tc_cmd += "rate {rate} ceil {speed}Mbit" tc_cmd = tc_cmd.format( intf=intf, root_qid=qid_hex, rate=default_gbr, speed=speed, ) cmd_list.append(tc_cmd) return run_cmd(cmd_list, show_error) @staticmethod def read_all_classes(intf: str): qid_list = [] # example output of this command # b'class htb 1:1 parent 1:fffe prio 0 rate 12Kbit ceil 1Gbit burst \ # 1599b cburst 1375b \nclass htb 1:fffe root rate 1Gbit ceil 1Gbit \ # burst 1375b cburst 1375b \n' # we need to parse this output and extract class ids from here tc_cmd = "tc class show dev {}".format(intf) args = argSplit(tc_cmd) try: output = subprocess.check_output(args) for ln in output.decode('utf-8').split("\n"): ln = ln.strip() if not ln: continue tok = ln.split() if len(tok) < 5: continue if tok[1] != "htb": continue if tok[3] == 'root': continue qid_str = tok[2].split(':')[1] qid = int(qid_str, 16) pqid_str = tok[4].split(':')[1] pqid = int(pqid_str, 16) qid_list.append((qid, pqid)) LOG.debug("TC-dump: %s qid %d pqid %d", ln, qid, pqid) except subprocess.CalledProcessError as e: LOG.error('failed extracting classids from tc %s', e) return qid_list @staticmethod def dump_class_state(intf: str, qid: int): qid_hex = hex(qid) tc_cmd = "tc -s -d class show dev {} classid 1:{}".format( intf, qid_hex, ) args = argSplit(tc_cmd) try: output = subprocess.check_output(args) print(output.decode()) except subprocess.CalledProcessError: print("Exception dumping Qos State for %s", intf) @staticmethod def dump_root_class_stats(intf: str): tc_cmd = "tc -s -s -d q ls dev {}".format(intf) args = argSplit(tc_cmd) try: output = subprocess.check_output(args) print(output.decode()) except subprocess.CalledProcessError: print("Exception dumping Qos State for %s", intf) @staticmethod def get_class_rate(intf: str, qid: int) -> Optional[str]: qid_hex = hex(qid) tc_cmd = "tc class show dev {} classid 1:{}".format(intf, qid_hex) args = argSplit(tc_cmd) try: # output: class htb 1:3 parent 1:2 prio 2 rate 250Kbit ceil 500Kbit burst 1600b cburst 1600b raw_output = subprocess.check_output(args) output = raw_output.decode('utf-8') # return all config from 'rate' onwards config = output.split("rate") try: return config[1] except IndexError: LOG.error("could not find rate: %s", output) except subprocess.CalledProcessError: LOG.error("Exception dumping Qos State for %s", tc_cmd) @staticmethod def _get_qdisc_type(intf: str) -> Optional[str]: tc_cmd = "tc qdisc show dev {}".format(intf) args = argSplit(tc_cmd) try: # output: qdisc htb 1: root refcnt 2 r2q 10 default 0 direct_packets_stat 314 direct_qlen 1000 raw_output = subprocess.check_output(args) output = raw_output.decode('utf-8') config = output.split() try: return config[1] except IndexError: LOG.error("could not qdisc type: %s", output) except subprocess.CalledProcessError: LOG.error("Exception dumping Qos State for %s", tc_cmd) class TCManager(object): """ Creates/Deletes queues in linux. Using Qdiscs for flow based rate limiting(traffic shaping) of user traffic. Queues are created on an egress interface and flows in OVS are programmed with qid to filter traffic to the queue. Traffic matching a specific flow is filtered to a queue and is rate limited based on configured value. Traffic to flows with no QoS configuration are sent to a default queue and are not rate limited. """ def __init__( self, datapath, config, ) -> None: self._datapath = datapath self._uplink = config['nat_iface'] self._downlink = config['enodeb_iface'] self._max_rate = config["qos"]["max_rate"] self._gbr_rate = config["qos"].get("gbr_rate", DEFAULT_RATE) self._enable_pyroute2 = config["qos"].get('enable_pyroute2', False) self._start_idx, self._max_idx = ( config['qos']['linux_tc']['min_idx'], config['qos']['linux_tc']['max_idx'], ) self._id_manager = IdManager(self._start_idx, self._max_idx) self._initialized = True LOG.info( "Init LinuxTC module uplink:%s downlink:%s", config['nat_iface'], config['enodeb_iface'], ) def destroy(self): if not TrafficClass.tc_ops: LOG.info("TC not initialized, skip destroying existing qos classes") return LOG.info("destroying existing leaf qos classes") # ensure ordering during deletion of classes, children should be deleted # prior to the parent class ids p_qids = set() for intf in [self._uplink, self._downlink]: qid_list = TrafficClass.read_all_classes(intf) for qid_tuple in qid_list: (qid, pqid) = qid_tuple if self._start_idx <= qid < (self._max_idx - 1): LOG.info("Attempting to delete class idx %d", qid) TrafficClass.delete_class(intf, qid) if self._start_idx <= pqid < (self._max_idx - 1): p_qids.add((intf, pqid)) LOG.info("destroying existing parent classes") for p_qid_tuple in p_qids: (intf, pqid) = p_qid_tuple LOG.info("Attempting to delete parent class idx %d", pqid) TrafficClass.delete_class(intf, pqid, skip_filter=True) LOG.info("destroying All qos classes: done") def setup(self): # initialize new qdisc TrafficClass.init_qdisc( self._uplink, enable_pyroute2=self._enable_pyroute2, default_gbr=self._gbr_rate, ) TrafficClass.init_qdisc( self._downlink, enable_pyroute2=self._enable_pyroute2, default_gbr=self._gbr_rate, ) def get_action_instruction(self, qid: int): # return an action and an instruction corresponding to this qid if qid < self._start_idx or qid > (self._max_idx - 1): LOG.error("invalid qid %d, no action/inst returned", qid) return None, None parser = self._datapath.ofproto_parser return parser.OFPActionSetField(pkt_mark=qid), None, qid def create_class_async( self, d: FlowMatch.Direction, qos_info: QosInfo, qid, parent, skip_filter, cleanup_rule, ): intf = self._uplink if d == FlowMatch.UPLINK else self._downlink gbr = qos_info.gbr if gbr is None: gbr = self._gbr_rate err = TrafficClass.create_class( intf, qid, qos_info.mbr, rate=gbr, parent_qid=parent, skip_filter=skip_filter, ) # typecast to int to avoid MagicMock related error in unit test err_no = int(err) if err_no < 0: if cleanup_rule: cleanup_rule() LOG.error("qos create error: qid %d err %d", qid, err_no) return LOG.debug("create done: if: %s qid %d err %s", intf, qid, err_no) def add_qos( self, d: FlowMatch.Direction, qos_info: QosInfo, cleanup_rule=None, parent=None, skip_filter=False, ) -> int: LOG.debug("add QoS: %s", qos_info) qid = self._id_manager.allocate_idx() self.create_class_async( d, qos_info, qid, parent, skip_filter, cleanup_rule, ) LOG.debug("assigned qid: %d", qid) return qid def remove_qos( self, qid: int, d: FlowMatch.Direction, recovery_mode=False, skip_filter=False, ): if not self._initialized and not recovery_mode: return if qid < self._start_idx or qid > (self._max_idx - 1): LOG.error("invalid qid %d, removal failed", qid) return LOG.debug("deleting qos_handle %s, skip_filter %s", qid, skip_filter) intf = self._uplink if d == FlowMatch.UPLINK else self._downlink err = TrafficClass.delete_class(intf, qid, skip_filter) if err == 0: self._id_manager.release_idx(qid) else: LOG.error('error deleting class %d, not releasing idx', qid) return def read_all_state(self): LOG.debug("read_all_state") st = {} apn_qid_list = set() ul_qid_list = TrafficClass.read_all_classes(self._uplink) dl_qid_list = TrafficClass.read_all_classes(self._downlink) for (d, qid_list) in ( (FlowMatch.UPLINK, ul_qid_list), (FlowMatch.DOWNLINK, dl_qid_list), ): for qid_tuple in qid_list: qid, pqid = qid_tuple if qid < self._start_idx or qid > (self._max_idx - 1): LOG.debug("qid %d out of range: (%d - %d)", qid, self._start_idx, self._max_idx) continue apn_qid = pqid if pqid != self._max_idx else 0 st[qid] = { 'direction': d, 'ambr_qid': apn_qid, } if apn_qid != 0: apn_qid_list.add(apn_qid) self._id_manager.restore_state(st) return st, apn_qid_list def same_qos_config( self, d: FlowMatch.Direction, qid1: int, qid2: int, ) -> bool: intf = self._uplink if d == FlowMatch.UPLINK else self._downlink config1 = TrafficClass.get_class_rate(intf, qid1) config2 = TrafficClass.get_class_rate(intf, qid2) return config1 == config2
lte/gateway/python/magma/pipelined/qos/qos_tc_impl.py
14,093
Creates/Deletes queues in linux. Using Qdiscs for flow based rate limiting(traffic shaping) of user traffic. Queues are created on an egress interface and flows in OVS are programmed with qid to filter traffic to the queue. Traffic matching a specific flow is filtered to a queue and is rate limited based on configured value. Traffic to flows with no QoS configuration are sent to a default queue and are not rate limited. Creates/Deletes queues in linux. Using Qdiscs for flow based rate limiting(traffic shaping) of user traffic. Copyright 2020 The Magma Authors. This source code is licensed under the BSD-style license found in the LICENSE file in the root directory of this source tree. Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. noqa LOG.setLevel(logging.DEBUG) TODO - replace this implementation with pyroute2 tc parent qid should only be self for root case, everything else should be the child of root class add filter TODO: Convert this class into an object. qdisc does not support replace, so check it before creating the HTB qdisc. example output of this command b'class htb 1:1 parent 1:fffe prio 0 rate 12Kbit ceil 1Gbit burst \ 1599b cburst 1375b \nclass htb 1:fffe root rate 1Gbit ceil 1Gbit \ burst 1375b cburst 1375b \n' we need to parse this output and extract class ids from here output: class htb 1:3 parent 1:2 prio 2 rate 250Kbit ceil 500Kbit burst 1600b cburst 1600b return all config from 'rate' onwards output: qdisc htb 1: root refcnt 2 r2q 10 default 0 direct_packets_stat 314 direct_qlen 1000 ensure ordering during deletion of classes, children should be deleted prior to the parent class ids initialize new qdisc return an action and an instruction corresponding to this qid typecast to int to avoid MagicMock related error in unit test
2,034
en
0.803903
# coding=utf-8 # This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. import shutil import mock import pytest from callee import Contains from .conftest import git_out, search_diff, search_rev from mozphab import environment, exceptions, mozphab arc_call_conduit = mock.Mock() call_conduit = mock.Mock() def by_line_mock(*args, **_kwargs): # join args to catch unicode errors " ".join(*args) return ["Revision URI: http://example.test/D123"] check_call_by_line = mock.Mock() check_call_by_line.side_effect = by_line_mock initial_sha = None def test_submit_create_arc(in_process, git_repo_path, init_sha): call_conduit.side_effect = ( dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), [{"userName": "alice", "phid": "PHID-USER-1"}], ) testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") git_out("commit", "--message", "A r?alice") testfile = git_repo_path / "untracked" testfile.write_text("a") mozphab.main( ["submit", "--arc", "--yes", "--bug", "1", init_sha], is_development=True ) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - A r?alice Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() def test_submit_create(in_process, git_repo_path, init_sha): call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # user search [dict(userName="alice", phid="PHID-USER-1")], # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) (git_repo_path / "X").write_text(u"ą\nb\nc\n", encoding="utf-8") (git_repo_path / "Y").write_text("no line ending") git_out("add", ".") (git_repo_path / "msg").write_text(u"Ą r?alice", encoding="utf-8") git_out("commit", "--file", "msg") (git_repo_path / "untracked").write_text("a\n") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - Ą r?alice Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() assert mock.call("conduit.ping", {}) in call_conduit.call_args_list assert ( mock.call("user.query", dict(usernames=["alice"])) in call_conduit.call_args_list ) assert ( mock.call( "diffusion.repository.search", dict(limit=1, constraints=dict(callsigns=["TEST"])), ) in call_conduit.call_args_list ) assert ( mock.call( "differential.creatediff", { "sourceControlPath": "/", "sourceControlSystem": "git", "lintStatus": "none", "sourcePath": mock.ANY, "unitStatus": "none", "sourceMachine": "http://example.test", "sourceControlBaseRevision": mock.ANY, "repositoryPHID": "PHID-REPO-1", "branch": "HEAD", "changes": [ { "commitHash": mock.ANY, "awayPaths": [], "newProperties": {"unix:filemode": "100644"}, "oldPath": None, "hunks": [ { "oldOffset": 0, "oldLength": 0, "newOffset": 1, "newLength": 3, "addLines": 3, "delLines": 0, "corpus": "+ą\n+b\n+c\n", "isMissingOldNewline": False, "isMissingNewNewline": False, } ], "oldProperties": {}, "currentPath": "X", "fileType": 1, "type": 1, "metadata": {}, }, { "commitHash": mock.ANY, "awayPaths": [], "newProperties": {"unix:filemode": "100644"}, "oldPath": None, "hunks": [ { "oldOffset": 0, "oldLength": 0, "newOffset": 1, "newLength": 1, "addLines": 1, "delLines": 0, "corpus": ( "+no line ending\n\\ No newline at end of file\n" ), "isMissingOldNewline": False, "isMissingNewNewline": True, } ], "oldProperties": {}, "currentPath": "Y", "fileType": 1, "type": 1, "metadata": {}, }, ], "creationMethod": "moz-phab-git", }, ) in call_conduit.call_args_list ) assert ( mock.call( "differential.setdiffproperty", {"diff_id": "1", "name": "local:commits", "data": ~Contains('"rev":')}, ) in call_conduit.call_args_list ) assert ( call_conduit.call_args_list.count( mock.call("differential.setdiffproperty", mock.ANY) ) == 2 ) def test_submit_create_added_not_commited(in_process, git_repo_path, init_sha): call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # user search [dict(userName="alice", phid="PHID-USER-1")], # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) (git_repo_path / "X").write_text("ą\r\nb\nc\n", encoding="utf-8") (git_repo_path / "Y").write_text("no line ending") git_out("add", ".") (git_repo_path / "msg").write_text("Ą r?alice", encoding="utf-8") git_out("commit", "--file", "msg") (git_repo_path / "untracked").write_text("a\n") git_out("add", "untracked") with pytest.raises(exceptions.Error) as excinfo: mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) assert "Uncommitted changes present." in str(excinfo.value) def test_submit_create_no_bug(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # user search [dict(userName="alice", phid="PHID-USER-1")], # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) testfile = git_repo_path / "X" testfile.write_text("a\n") git_out("add", ".") msgfile = git_repo_path / "msg" msgfile.write_text("A r?alice") git_out("commit", "--file", "msg") mozphab.main(["submit", "--yes", "--no-bug", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ A r?alice Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() def test_submit_create_binary_arc(in_process, git_repo_path, init_sha, data_file): call_conduit.side_effect = ( dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), [{"userName": "alice", "phid": "PHID-USER-1"}], ) shutil.copyfile(str(data_file), str(git_repo_path / "img.png")) git_out("add", ".") git_out("commit", "--message", "IMG") mozphab.main( ["submit", "--arc", "--yes", "--bug", "1", init_sha], is_development=True ) expected = """ Bug 1 - IMG Differential Revision: http://example.test/D123 """ log = git_out("log", "--format=%s%n%n%b", "-1") assert log.strip() == expected.strip() def test_submit_create_binary(in_process, git_repo_path, init_sha, data_file): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # file.allocate dict(dict(filePHID=None, upload=True)), # file.upload dict(), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) shutil.copyfile(str(data_file), str(git_repo_path / "img.png")) git_out("add", ".") git_out("commit", "-m", "IMG") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - IMG Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() assert ( mock.call( "file.allocate", {"name": "img.png", "contentHash": mock.ANY, "contentLength": 182}, ) in call_conduit.call_args_list ) assert ( mock.call("file.upload", {"data_base64": mock.ANY, "name": "img.png"}) in call_conduit.call_args_list ) def test_submit_create_binary_existing(in_process, git_repo_path, init_sha, data_file): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # file.allocate dict(dict(filePHID="PHID-FILE-1", upload=False)), # no file.upload call # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) shutil.copyfile(str(data_file), str(git_repo_path / "img.png")) git_out("add", ".") git_out("commit", "-m", "IMG") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - IMG Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() assert ( mock.call( "file.allocate", {"name": "img.png", "contentHash": mock.ANY, "contentLength": 182}, ) in call_conduit.call_args_list ) assert mock.call("file.upload", mock.ANY) not in call_conduit.call_args_list def test_submit_create_binary_chunked(in_process, git_repo_path, init_sha, data_file): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # file.allocate dict(dict(filePHID="PHID-FILE-1", upload=True)), # file.querychunks [ dict(byteStart="0", byteEnd="4194304", complete=False), dict(byteStart="4194304", byteEnd="8388608", complete=False), dict(byteStart="8388608", byteEnd="8425160", complete=False), ], # file.uploadchunk dict(), # file.uploadchunk dict(), # file.uploadchunk dict(), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) shutil.copyfile(str(data_file), str(git_repo_path / "img.png")) git_out("add", ".") git_out("commit", "-m", "IMG") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - IMG Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() assert ( mock.call( "file.allocate", {"name": "img.png", "contentHash": mock.ANY, "contentLength": 182}, ) in call_conduit.call_args_list ) assert ( mock.call("file.querychunks", {"filePHID": "PHID-FILE-1"}) in call_conduit.call_args_list ) assert ( mock.call( "file.uploadchunk", { "filePHID": "PHID-FILE-1", "byteStart": 0, "data": mock.ANY, "dataEncoding": "base64", }, ) in call_conduit.call_args_list ) assert ( mock.call( "file.uploadchunk", { "filePHID": "PHID-FILE-1", "byteStart": 4194304, "data": mock.ANY, "dataEncoding": "base64", }, ) in call_conduit.call_args_list ) assert ( mock.call( "file.uploadchunk", { "filePHID": "PHID-FILE-1", "byteStart": 8388608, "data": mock.ANY, "dataEncoding": "base64", }, ) in call_conduit.call_args_list ) def test_submit_update(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # diffusion.revision.search dict(data=[search_rev(rev=123)]), # diffusion.diff.search dict(data=[search_diff()]), # whoami dict(phid="PHID-USER-1"), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) testfile = git_repo_path / "X" testfile.write_text("ą", encoding="utf-8") git_out("add", ".") msgfile = git_repo_path / "msg" msgfile.write_text( """\ Bug 1 - Ą Differential Revision: http://example.test/D123 """, encoding="utf-8", ) git_out("commit", "--file", "msg") mozphab.main( ["submit", "--yes"] + ["--bug", "1"] + ["--message", "update message ćwikła"] + [init_sha], is_development=True, ) assert call_conduit.call_count == 9 log = git_out("log", "--format=%s%n%n%b", "-1") expected = """\ Bug 1 - Ą Differential Revision: http://example.test/D123 """ assert log == expected def test_submit_update_no_change(in_process, git_repo_path, init_sha, git_sha): testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") msgfile = git_repo_path / "msg" msgfile.write_text( """\ Bug 1 - A Differential Revision: http://example.test/D123 """ ) git_out("commit", "--file", "msg") sha = git_sha() call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # diffusion.revision.search dict(data=[search_rev(rev=123)]), # diffusion.diff.search dict(data=[search_diff(node=sha)]), # whoami dict(phid="PHID-USER-1"), ) mozphab.main( ["submit", "--yes"] + [init_sha], is_development=True, ) assert call_conduit.call_count == 5 def test_submit_remove_cr(in_process, git_repo_path, init_sha): if environment.IS_WINDOWS: pytest.skip("Removing CR will not work on Windows.") call_conduit.side_effect = ( # CREATE # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # user.search [dict(userName="alice", phid="PHID-USER-1")], # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), # UPDATE # no need to ping (checked) # no need to check reviewer # no need to search for repository repository data is saved in .hg # differential.creatediff dict(dict(phid="PHID-DIFF-2", diffid="2")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="124")), # differential.setdiffproperty dict(), ) test_a = git_repo_path / "X" test_a.write_text("a\r\nb\n") git_out("add", "X") git_out("commit", "-am", "A r?alice") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) call_conduit.reset_mock() # removing CR, leaving LF test_a.write_text("a\nb\n") git_out("commit", "-am", "B r?alice") mozphab.main(["submit", "--yes", "--bug", "1", "HEAD~"], is_development=True) assert ( mock.call( "differential.creatediff", { "changes": [ { "metadata": {}, "oldPath": "X", "currentPath": "X", "awayPaths": [], "oldProperties": {}, "newProperties": {}, "commitHash": mock.ANY, "type": 2, "fileType": 1, "hunks": [ { "oldOffset": 1, "oldLength": 2, "newOffset": 1, "newLength": 2, "addLines": 1, "delLines": 1, "isMissingOldNewline": False, "isMissingNewNewline": False, "corpus": "-a\r\n+a\n b\n", } ], } ], "sourceMachine": "http://example.test", "sourceControlSystem": "git", "sourceControlPath": "/", "sourceControlBaseRevision": mock.ANY, "creationMethod": "moz-phab-git", "lintStatus": "none", "unitStatus": "none", "repositoryPHID": "PHID-REPO-1", "sourcePath": mock.ANY, "branch": "HEAD", }, ) in call_conduit.call_args_list ) assert ( mock.call( "differential.setdiffproperty", { "diff_id": "2", "name": "local:commits", "data": Contains('"summary": "Bug 1 - B r?alice"') & Contains( '"message": "' "Bug 1 - B r?alice\\n\\n" "Summary:\\n\\n\\n\\n\\n" "Test Plan:\\n\\n" "Reviewers: alice\\n\\n" "Subscribers:\\n\\n" 'Bug #: 1"' ), }, ) in call_conduit.call_args_list ) def test_submit_single_last(in_process, git_repo_path, init_sha): call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) (git_repo_path / "X").write_text("a\n") git_out("add", "X") git_out("commit", "-am", "A") (git_repo_path / "X").write_text("b\n") git_out("commit", "-am", "B") mozphab.main(["submit", "--yes", "--bug", "1", "--single"], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-2") expected = """\ Bug 1 - B Differential Revision: http://example.test/D123 A """ assert log == expected def test_submit_single_first(in_process, git_repo_path, init_sha, git_sha): call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) (git_repo_path / "X").write_text("a\n") git_out("add", "X") git_out("commit", "-am", "A") sha = git_sha() (git_repo_path / "X").write_text("b\n") git_out("commit", "-am", "B") mozphab.main( ["submit", "--yes", "--bug", "1", "--single", sha], is_development=True ) log = git_out("log", "--format=%s%n%n%b", "-2") expected = """\ B Bug 1 - A Differential Revision: http://example.test/D123 """ assert log == expected def test_submit_update_no_message(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), dict(data=[search_rev(rev=123)]), dict(data=[search_diff()]), dict(phid="PHID-USER-1"), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) (git_repo_path / "X").write_text(u"ą", encoding="utf-8") git_out("add", ".") (git_repo_path / "msg").write_text( u"""\ Bug 1 - Ą Differential Revision: http://example.test/D123 """, encoding="utf-8", ) git_out("commit", "--file", "msg") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """\ Bug 1 - Ą Differential Revision: http://example.test/D123 """ assert log == expected def test_submit_different_author_arc(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), [{"userName": "alice", "phid": "PHID-USER-1"}], ) testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") git_out( "commit", "--date", "Tue, 22 Jan 2019 13:42:48 +0000", "--author", "foo <foo@bar.com>", "--message", "A r?alice", ) testfile.write_text("b") git_out( "commit", "--date", "Tue, 22 Jan 2019 13:43:48 +0000", "--author", "bar <bar@foo.com>", "--all", "--message", "B r?alice", ) mozphab.main( ["submit", "--arc", "--yes", "--bug", "1", init_sha], is_development=True ) log = git_out("log", "--format=%aD+++%an+++%ae", "-2") expected = """\ Tue, 22 Jan 2019 13:43:48 +0000+++bar+++bar@foo.com Tue, 22 Jan 2019 13:42:48 +0000+++foo+++foo@bar.com """ assert log == expected def test_submit_utf8_author_arc(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), [{"userName": "alice", "phid": "PHID-USER-1"}], ) testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") git_out( "commit", "--date", "Tue, 22 Jan 2019 13:42:48 +0000", "--author", "ćwikła <ćwikła@bar.com>", "--message", "A r?alice", ) mozphab.main( ["submit", "--arc", "--yes", "--bug", "1", init_sha], is_development=True ) log = git_out("log", "--format=%aD+++%an+++%ae", "-1") expected = "Tue, 22 Jan 2019 13:42:48 +0000+++ćwikła+++ćwikła@bar.com\n" assert log == expected def test_submit_update_arc(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( {}, # ping dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), dict(data=[search_rev(rev=123)]), dict(data=[search_diff()]), dict(phid="PHID-USER-1"), ) testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") # Write out our commit message as if the program had already run and appended # a Differential Revision keyword to the commit body for tracking. git_out( "commit", "--message", """\ Bug 1 - A Differential Revision: http://example.test/D123 """, ) mozphab.main( ["submit", "--arc", "--yes"] + ["--bug", "1"] + ["--message", "update message ćwikła"] + [init_sha], is_development=True, ) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """\ Bug 1 - A Differential Revision: http://example.test/D123 """ assert log == expected def test_submit_update_bug_id_arc(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), dict(data=[search_rev(rev=123)]), dict(data=[search_diff()]), # get reviewers for updated revision dict(phid="PHID-USER-1"), ) arc_call_conduit.reset_mock() arc_call_conduit.side_effect = ( {}, {"data": {}}, ) testfile = git_repo_path / "X" testfile.write_text("a") git_out("add", ".") # Write out our commit message as if the program had already run and appended # a Differential Revision keyword to the commit body for tracking. git_out( "commit", "--message", """\ Bug 1 - A Differential Revision: http://example.test/D123 """, ) mozphab.main( ["submit", "--arc", "--yes", "--bug", "2", init_sha], is_development=True ) arc_call_conduit.assert_called_with( "differential.revision.edit", { "objectIdentifier": "D123", "transactions": [{"type": "bugzilla.bug-id", "value": "2"}], }, mock.ANY, ) def test_submit_update_revision_not_found(in_process, git_repo_path, init_sha): call_conduit.reset_mock() call_conduit.side_effect = ( # ping dict(), dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # response for searching D123 and D124 dict(data=[search_rev(rev=123)]), dict(data=[search_diff()]), # moz-phab asks again for D124 dict(data=[]), # whoami dict(phid="PHID-USER-1"), # moz-phab asks again for D124 dict(data=[]), # moz-phab asks again for D124 dict(data=[]), ) testfile = git_repo_path / "X" testfile.write_text(u"ą", encoding="utf-8") git_out("add", ".") msgfile = git_repo_path / "msg" msgfile.write_text( u"""\ Bug 1 - Ą Differential Revision: http://example.test/D123 """, encoding="utf-8", ) git_out("commit", "--file", "msg") testfile.write_text(u"missing repo") msgfile.write_text( u"""\ Bug 1 - missing revision Differential Revision: http://example.test/D124 """ ) git_out("commit", "--all", "--file", "./msg") with pytest.raises(exceptions.Error) as excinfo: mozphab.main( ["submit", "--yes"] + ["--bug", "1"] + ["--message", "update message ćwikła"] + [init_sha], is_development=True, ) assert "query result for revision D124" in str(excinfo.value) def test_empty_file(in_process, git_repo_path, init_sha): # Add an empty file call_conduit.side_effect = ( # ping dict(), # diffusion.repository.search dict(data=[dict(phid="PHID-REPO-1", fields=dict(vcs="git"))]), # differential.creatediff dict(dict(phid="PHID-DIFF-1", diffid="1")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="123")), # differential.setdiffproperty dict(), ) testfile = git_repo_path / "X" testfile.touch() git_out("add", ".") git_out("commit", "--message", "A") mozphab.main(["submit", "--yes", "--bug", "1", init_sha], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - A Differential Revision: http://example.test/D123 """ assert log.strip() == expected.strip() # Rempve an empty file call_conduit.reset_mock() call_conduit.side_effect = ( # differential.creatediff dict(dict(phid="PHID-DIFF-2", diffid="2")), # differential.setdiffproperty dict(), # differential.revision.edit dict(object=dict(id="124")), # differential.setdiffproperty dict(), ) testfile.unlink() git_out("commit", "-a", "--message", "B") mozphab.main(["submit", "--yes", "--bug", "1", "HEAD~"], is_development=True) log = git_out("log", "--format=%s%n%n%b", "-1") expected = """ Bug 1 - B Differential Revision: http://example.test/D124 """ assert log.strip() == expected.strip()
tests/test_integration_git.py
31,249
coding=utf-8 This Source Code Form is subject to the terms of the Mozilla Public License, v. 2.0. If a copy of the MPL was not distributed with this file, You can obtain one at http://mozilla.org/MPL/2.0/. join args to catch unicode errors ping diffusion.repository.search user search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search user search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search user search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search file.allocate file.upload differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search file.allocate no file.upload call differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search file.allocate file.querychunks file.uploadchunk file.uploadchunk file.uploadchunk differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search diffusion.revision.search diffusion.diff.search whoami differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search diffusion.revision.search diffusion.diff.search whoami CREATE ping diffusion.repository.search user.search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty UPDATE no need to ping (checked) no need to check reviewer no need to search for repository repository data is saved in .hg differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty removing CR, leaving LF ping diffusion.repository.search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping diffusion.repository.search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty ping Write out our commit message as if the program had already run and appended a Differential Revision keyword to the commit body for tracking. get reviewers for updated revision Write out our commit message as if the program had already run and appended a Differential Revision keyword to the commit body for tracking. ping response for searching D123 and D124 moz-phab asks again for D124 whoami moz-phab asks again for D124 moz-phab asks again for D124 Add an empty file ping diffusion.repository.search differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty Rempve an empty file differential.creatediff differential.setdiffproperty differential.revision.edit differential.setdiffproperty
3,146
en
0.721162
#!/usr/bin/env python # -*- coding: utf-8 -*- """ :mod:`accelerometer` ================== Updated by lkasso <hello@mbientlab.com> Created by hbldh <henrik.blidh@nedomkull.com> Created on 2016-04-10 """ from __future__ import division from __future__ import print_function from __future__ import absolute_import import time from pymetawear.discover import select_device from pymetawear.client import MetaWearClient address = select_device() c = MetaWearClient(str(address), 'pygatt', debug=True) print("New client created: {0}".format(c)) def acc_callback(data): """Handle a (epoch, (x,y,z)) accelerometer tuple.""" print("Epoch time: [{0}] - X: {1}, Y: {2}, Z: {3}".format(data[0], *data[1])) print("Get possible accelerometer settings...") settings = c.accelerometer.get_possible_settings() print(settings) time.sleep(1.0) print("Write accelerometer settings...") c.accelerometer.set_settings(data_rate=3.125, data_range=4.0) time.sleep(1.0) print("Check accelerometer settings...") settings = c.accelerometer.get_current_settings() print(settings) print("Subscribing to accelerometer signal notifications...") c.accelerometer.high_frequency_stream = False c.accelerometer.notifications(acc_callback) time.sleep(10.0) print("Unsubscribe to notification...") c.accelerometer.notifications(None) time.sleep(5.0) c.disconnect()
examples/accelerometer.py
1,353
Handle a (epoch, (x,y,z)) accelerometer tuple. :mod:`accelerometer` ================== Updated by lkasso <hello@mbientlab.com> Created by hbldh <henrik.blidh@nedomkull.com> Created on 2016-04-10 !/usr/bin/env python -*- coding: utf-8 -*-
239
en
0.520104
# Natural Language Toolkit: SVM-based classifier # # Copyright (C) 2001-2022 NLTK Project # Author: Leon Derczynski <leon@dcs.shef.ac.uk> # # URL: <https://www.nltk.org/> # For license information, see LICENSE.TXT """ nltk.classify.svm was deprecated. For classification based on support vector machines SVMs use nltk.classify.scikitlearn (or `scikit-learn <https://scikit-learn.org>`_ directly). """ class SvmClassifier: def __init__(self, *args, **kwargs): raise NotImplementedError(__doc__)
nltk/classify/svm.py
508
nltk.classify.svm was deprecated. For classification based on support vector machines SVMs use nltk.classify.scikitlearn (or `scikit-learn <https://scikit-learn.org>`_ directly). Natural Language Toolkit: SVM-based classifier Copyright (C) 2001-2022 NLTK Project Author: Leon Derczynski <leon@dcs.shef.ac.uk> URL: <https://www.nltk.org/> For license information, see LICENSE.TXT
380
en
0.739263
# -*- coding: utf-8 -*- from setuptools import setup, find_packages with open('README.md') as f: readme = f.read() with open('LICENSE') as f: license = f.read() with open('VERSION.py') as f: exec(f.read()) setup( name='upsere analysis', version=__version__, description='10X Genomics CLI', long_description=readme, author='Eddie Belter', author_email='ebetler@gmail.com', license=license, url='https://github.com/ebelter/upserve-analysis.git', install_requires=[ 'click==7.0', 'pyyaml==5.1', 'Jinja2>=2.10.1', ], entry_points=''' [console_scripts] upserve=upserve.cli:cli ''', setup_requires=["pytest-runner"], tests_require=["pytest"], packages=find_packages(exclude=('tests', 'docs')), include_package_data=True, )
setup.py
838
-*- coding: utf-8 -*-
21
en
0.767281
#!/usr/bin/python ################################################################################ # 20de4144-5cc5-11e4-af55-00155d01fe08 # # Justin Dierking # justindierking@hardbitsolutions.com # phnomcobra@gmail.com # # 10/24/2014 Original Construction ################################################################################ class Finding: def __init__(self): self.output = [] self.is_compliant = False self.uuid = "20de4144-5cc5-11e4-af55-00155d01fe08" def check(self, cli): # Initialize Compliance self.is_compliant = False # Get Registry DWORD dword = cli.get_reg_dword(r'HKLM:\Software\Policies\Microsoft\EMET\SysSettings', 'ASLR') # Output Lines self.output = [r'HKLM:\Software\Policies\Microsoft\EMET\SysSettings', ('ASLR=' + str(dword))] if dword == 3: self.is_compliant = True return self.is_compliant def fix(self, cli): cli.powershell(r"New-Item -path 'HKLM:\Software\Policies\Microsoft'") cli.powershell(r"New-Item -path 'HKLM:\Software\Policies\Microsoft\EMET'") cli.powershell(r"New-Item -path 'HKLM:\Software\Policies\Microsoft\EMET\SysSettings'") cli.powershell(r"Set-ItemProperty -path 'HKLM:\Software\Policies\Microsoft\EMET\SysSettings' -name 'ASLR' -value 3 -Type DWord")
pcat2py/class/20de4144-5cc5-11e4-af55-00155d01fe08.py
1,362
!/usr/bin/python 20de4144-5cc5-11e4-af55-00155d01fe08 Justin Dierking justindierking@hardbitsolutions.com phnomcobra@gmail.com 10/24/2014 Original Construction Initialize Compliance Get Registry DWORD Output Lines
213
en
0.3191
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import print_function import io import os import sys from setuptools import find_packages, setup def read_req(req_file): with open(os.path.join('requirements', req_file)) as req: return [line.strip() for line in req.readlines() if line.strip() and not line.strip().startswith('#')] with io.open('README.rst', encoding='utf-8') as readme: description = readme.read() requirements = read_req('base.txt') requirements_validation = read_req('validation.txt') def find_versions_from_readme(prefix): for line in description.splitlines(): line = line.strip() if line.startswith(prefix): versions = [v.strip() for v in line[len(prefix):].split(',')] if versions: return versions raise RuntimeError("failed to find supported versions list for '{}'".format(prefix)) python_versions = find_versions_from_readme("- **Python**: ") django_versions = find_versions_from_readme("- **Django**: ") python_requires = ">=" + python_versions[0] python_classifiers = [ 'Programming Language :: Python', 'Programming Language :: Python :: 3', ] + ['Programming Language :: Python :: {}'.format(v) for v in python_versions] django_classifiers = [ 'Framework :: Django', ] + ['Framework :: Django :: {}'.format(v) for v in django_versions] def drf_yasg_setup(**kwargs): setup( name='drf-yasg', packages=find_packages('src'), package_dir={'': 'src'}, include_package_data=True, install_requires=requirements, extras_require={ 'validation': requirements_validation, }, license='BSD License', description='Automated generation of real Swagger/OpenAPI 2.0 schemas from Django Rest Framework code.', long_description=description, long_description_content_type='text/x-rst', url='https://github.com/axnsan12/drf-yasg', author='Cristi V.', author_email='cristi@cvjd.me', keywords='drf django django-rest-framework schema swagger openapi codegen swagger-codegen ' 'documentation drf-yasg django-rest-swagger drf-openapi', python_requires=python_requires, classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: BSD License', 'Development Status :: 5 - Production/Stable', 'Operating System :: OS Independent', 'Environment :: Web Environment', 'Topic :: Documentation', 'Topic :: Software Development :: Code Generators', ] + python_classifiers + django_classifiers, **kwargs ) try: # noinspection PyUnresolvedReferences import setuptools_scm # noqa: F401 drf_yasg_setup(use_scm_version=True) except (ImportError, LookupError) as e: if os.getenv('CI', 'false') == 'true' or os.getenv('TRAVIS', 'false') == 'true': # don't silently fail on travis - we don't want to accidentally push a dummy version to PyPI raise err_msg = str(e) if 'setuptools-scm' in err_msg or 'setuptools_scm' in err_msg: import time import traceback timestamp_ms = int(time.time() * 1000) timestamp_str = hex(timestamp_ms)[2:].zfill(16) dummy_version = '1!0.0.0.dev0+noscm.' + timestamp_str drf_yasg_setup(version=dummy_version) traceback.print_exc(file=sys.stderr) print("failed to detect version, package was built with dummy version " + dummy_version, file=sys.stderr) else: raise
setup.py
3,611
!/usr/bin/env python -*- coding: utf-8 -*- noinspection PyUnresolvedReferences noqa: F401 don't silently fail on travis - we don't want to accidentally push a dummy version to PyPI
180
en
0.839002
from tourapi.list import TourAPI from tourapi.config import ServiceKey, MobileOS, MobileApp, Languages from mysql_config import MysqlHost, MysqlUser, MysqlPass, MysqlDB import pymysql import json def upload_category_codes(codes, language="Kor", level=0, cat1="", cat2="", cat3=""): global conn, curs query = """ INSERT INTO category_code (code, cat1, cat2, cat3, level, name_{0}) VALUES (%s, %s, %s, %s, %s, %s) ON DUPLICATE KEY UPDATE name_{0}=%s """.format(language.lower()) for code in codes: curs.execute(query, (code["code"], cat1, cat2, cat3, level, code["name"], code["name"])) # print(code["name"], code["code"]) conn.commit() return conn = pymysql.connect(host = MysqlHost, user = MysqlUser, password = MysqlPass, db = MysqlDB) curs = conn.cursor() for lan in Languages: language = lan["code"] api = TourAPI(ServiceKey, language) # 대분류 카테고리 cat1_codes = api.list_category_code() upload_category_codes(cat1_codes, language, 1) for cat1 in cat1_codes: cat2_codes = api.list_category_code(cat1["code"]) upload_category_codes(cat2_codes, language, 2, cat1["code"]) print(cat2_codes) for cat2 in cat2_codes: cat3_codes = api.list_category_code(cat1["code"], cat2["code"]) upload_category_codes(cat3_codes, language, 3, cat1["code"], cat2["code"]) conn.commit() conn.close()
02_category_code.py
1,373
print(code["name"], code["code"]) 대분류 카테고리
42
ko
0.967001
import os import shutil import subprocess import re import string import pathlib import timeit import jmhbenchmark class JHaskellBenchmark(jmhbenchmark.JMHBenchmark): def __init__(self, name, source_path, compiler_args=None): if compiler_args is None: compiler_args = [] source_path = pathlib.Path(source_path) super().__init__(name, source_path.stem.lower(), source_path.stem.capitalize()) self._source_path = source_path self._compiler_args = compiler_args.copy() def __enter__(self): ret = super().__enter__() self._output_jar = (self._temp_dir / self._name).with_suffix(".jar") return ret def get_run_args(self): return ["-jar", f"{self._name}.jar"] def _compile(self): self._run_jhaskell_compiler() def _post_compile(self): self._results["size"] = jmhbenchmark.get_jar_entry_size( self._output_jar, [ f"{self._package_name}/{s}.class" for s in [self._class_name, "Data", "Function", "BoxedData", "HeapObject"] ], ) return super()._post_compile() def _get_classpath(self): return [f"{self._name}.jar"] def _run_jhaskell_compiler(self, extra_args=None): if extra_args is None: extra_args = [] original_dir = pathlib.Path.cwd() # Build the source program args = ( [ "compiler-exe", "--build-dir", f"{self._temp_dir / 'out'}", "--output-jar", str(self._output_jar), "--output-class", self._class_name, "--runtime-file-dir", str(original_dir.parent / "runtime"), ] + self._compiler_args + extra_args + [f"programs/{self._package_name}.hs"] ) try: return subprocess.check_output(args) except subprocess.CalledProcessError as e: print(e.stdout.decode()) raise # For JHaskell, time for each stage of the compiler def _benchmark_compilation(self, iterations=50): number = 1 # Record the output of each invocation outputs = [] def bench_func(): outputs.append(self._run_jhaskell_compiler(["--time-stages"]).decode()) overall_times = timeit.repeat(stmt=bench_func, setup=self._pre_compile, number=number, repeat=iterations) time_data = [] data_extractor = re.compile(r"(.+): (.+)ms") for output, overall_time in zip(outputs, overall_times): cumulative_time = 0 this_run_data = [] for line in output.splitlines(): match = data_extractor.fullmatch(line) if match is None: raise RuntimeError("Invalid line from compiler: " + line) this_time = float(match.group(2)) this_run_data.append((match.group(1), this_time)) cumulative_time += this_time #this_run_data.append(("Other", overall_time * 1000 - cumulative_time)) time_data.append(this_run_data) self._results["times"] = time_data
benchmarks/jhaskellbenchmark.py
3,270
Build the source program For JHaskell, time for each stage of the compiler Record the output of each invocationthis_run_data.append(("Other", overall_time * 1000 - cumulative_time))
181
en
0.609848
import copy import logging from collections import defaultdict import dask.dataframe as dd import numpy as np import pandas as pd from pandas.api.types import is_dtype_equal, is_numeric_dtype import featuretools.variable_types.variable as vtypes from featuretools.entityset import deserialize, serialize from featuretools.entityset.entity import Entity from featuretools.entityset.relationship import Relationship, RelationshipPath from featuretools.utils.gen_utils import import_or_raise pd.options.mode.chained_assignment = None # default='warn' logger = logging.getLogger('featuretools.entityset') class EntitySet(object): """ Stores all actual data for a entityset Attributes: id entity_dict relationships time_type Properties: metadata """ def __init__(self, id=None, entities=None, relationships=None): """Creates EntitySet Args: id (str) : Unique identifier to associate with this instance entities (dict[str -> tuple(pd.DataFrame, str, str, dict[str -> Variable])]): dictionary of entities. Entries take the format {entity id -> (dataframe, id column, (time_index), (variable_types), (make_index))}. Note that time_index, variable_types and make_index are optional. relationships (list[(str, str, str, str)]): List of relationships between entities. List items are a tuple with the format (parent entity id, parent variable, child entity id, child variable). Example: .. code-block:: python entities = { "cards" : (card_df, "id"), "transactions" : (transactions_df, "id", "transaction_time") } relationships = [("cards", "id", "transactions", "card_id")] ft.EntitySet("my-entity-set", entities, relationships) """ self.id = id self.entity_dict = {} self.relationships = [] self.time_type = None entities = entities or {} relationships = relationships or [] for entity in entities: df = entities[entity][0] index_column = entities[entity][1] time_index = None variable_types = None make_index = None if len(entities[entity]) > 2: time_index = entities[entity][2] if len(entities[entity]) > 3: variable_types = entities[entity][3] if len(entities[entity]) > 4: make_index = entities[entity][4] self.entity_from_dataframe(entity_id=entity, dataframe=df, index=index_column, time_index=time_index, variable_types=variable_types, make_index=make_index) for relationship in relationships: parent_variable = self[relationship[0]][relationship[1]] child_variable = self[relationship[2]][relationship[3]] self.add_relationship(Relationship(parent_variable, child_variable)) self.reset_data_description() def __sizeof__(self): return sum([entity.__sizeof__() for entity in self.entities]) def __dask_tokenize__(self): return (EntitySet, serialize.entityset_to_description(self.metadata)) def __eq__(self, other, deep=False): if len(self.entity_dict) != len(other.entity_dict): return False for eid, e in self.entity_dict.items(): if eid not in other.entity_dict: return False if not e.__eq__(other[eid], deep=deep): return False for r in other.relationships: if r not in other.relationships: return False return True def __ne__(self, other, deep=False): return not self.__eq__(other, deep=deep) def __getitem__(self, entity_id): """Get entity instance from entityset Args: entity_id (str): Id of entity. Returns: :class:`.Entity` : Instance of entity. None if entity doesn't exist. """ if entity_id in self.entity_dict: return self.entity_dict[entity_id] name = self.id or "entity set" raise KeyError('Entity %s does not exist in %s' % (entity_id, name)) @property def entities(self): return list(self.entity_dict.values()) @property def metadata(self): '''Returns the metadata for this EntitySet. The metadata will be recomputed if it does not exist.''' if self._data_description is None: description = serialize.entityset_to_description(self) self._data_description = deserialize.description_to_entityset(description) return self._data_description def reset_data_description(self): self._data_description = None def to_pickle(self, path, compression=None, profile_name=None): '''Write entityset in the pickle format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str): location on disk to write to (will be created as a directory) compression (str) : Name of the compression to use. Possible values are: {'gzip', 'bz2', 'zip', 'xz', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. ''' serialize.write_data_description(self, path, format='pickle', compression=compression, profile_name=profile_name) return self def to_parquet(self, path, engine='auto', compression=None, profile_name=None): '''Write entityset to disk in the parquet format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str): location on disk to write to (will be created as a directory) engine (str) : Name of the engine to use. Possible values are: {'auto', 'pyarrow', 'fastparquet'}. compression (str) : Name of the compression to use. Possible values are: {'snappy', 'gzip', 'brotli', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. ''' serialize.write_data_description(self, path, format='parquet', engine=engine, compression=compression, profile_name=profile_name) return self def to_csv(self, path, sep=',', encoding='utf-8', engine='python', compression=None, profile_name=None): '''Write entityset to disk in the csv format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str) : Location on disk to write to (will be created as a directory) sep (str) : String of length 1. Field delimiter for the output file. encoding (str) : A string representing the encoding to use in the output file, defaults to 'utf-8'. engine (str) : Name of the engine to use. Possible values are: {'c', 'python'}. compression (str) : Name of the compression to use. Possible values are: {'gzip', 'bz2', 'zip', 'xz', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. ''' serialize.write_data_description(self, path, format='csv', index=False, sep=sep, encoding=encoding, engine=engine, compression=compression, profile_name=profile_name) return self def to_dictionary(self): return serialize.entityset_to_description(self) ########################################################################### # Public getter/setter methods ######################################### ########################################################################### def __repr__(self): repr_out = u"Entityset: {}\n".format(self.id) repr_out += u" Entities:" for e in self.entities: if e.df.shape: repr_out += u"\n {} [Rows: {}, Columns: {}]".format( e.id, e.df.shape[0], e.df.shape[1]) else: repr_out += u"\n {} [Rows: None, Columns: None]".format( e.id) repr_out += "\n Relationships:" if len(self.relationships) == 0: repr_out += u"\n No relationships" for r in self.relationships: repr_out += u"\n %s.%s -> %s.%s" % \ (r._child_entity_id, r._child_variable_id, r._parent_entity_id, r._parent_variable_id) return repr_out def add_relationships(self, relationships): """Add multiple new relationships to a entityset Args: relationships (list[Relationship]) : List of new relationships. """ return [self.add_relationship(r) for r in relationships][-1] def add_relationship(self, relationship): """Add a new relationship between entities in the entityset Args: relationship (Relationship) : Instance of new relationship to be added. """ if relationship in self.relationships: logger.warning( "Not adding duplicate relationship: %s", relationship) return self # _operations? # this is a new pair of entities child_e = relationship.child_entity child_v = relationship.child_variable.id parent_e = relationship.parent_entity parent_v = relationship.parent_variable.id if not isinstance(child_e[child_v], vtypes.Id): child_e.convert_variable_type(variable_id=child_v, new_type=vtypes.Id, convert_data=False) if not isinstance(parent_e[parent_v], vtypes.Index): parent_e.convert_variable_type(variable_id=parent_v, new_type=vtypes.Index, convert_data=False) # Empty dataframes (as a result of accessing Entity.metadata) # default to object dtypes for discrete variables, but # indexes/ids default to ints. In this case, we convert # the empty column's type to int if isinstance(child_e.df, pd.DataFrame) and \ (child_e.df.empty and child_e.df[child_v].dtype == object and is_numeric_dtype(parent_e.df[parent_v])): child_e.df[child_v] = pd.Series(name=child_v, dtype=np.int64) parent_dtype = parent_e.df[parent_v].dtype child_dtype = child_e.df[child_v].dtype msg = u"Unable to add relationship because {} in {} is Pandas dtype {}"\ u" and {} in {} is Pandas dtype {}." if not is_dtype_equal(parent_dtype, child_dtype): raise ValueError(msg.format(parent_v, parent_e.id, parent_dtype, child_v, child_e.id, child_dtype)) self.relationships.append(relationship) self.reset_data_description() return self ########################################################################### # Relationship access/helper methods ################################### ########################################################################### def find_forward_paths(self, start_entity_id, goal_entity_id): """ Generator which yields all forward paths between a start and goal entity. Does not include paths which contain cycles. Args: start_entity_id (str) : id of entity to start the search from goal_entity_id (str) : if of entity to find forward path to See Also: :func:`BaseEntitySet.find_backward_paths` """ for sub_entity_id, path in self._forward_entity_paths(start_entity_id): if sub_entity_id == goal_entity_id: yield path def find_backward_paths(self, start_entity_id, goal_entity_id): """ Generator which yields all backward paths between a start and goal entity. Does not include paths which contain cycles. Args: start_entity_id (str) : Id of entity to start the search from. goal_entity_id (str) : Id of entity to find backward path to. See Also: :func:`BaseEntitySet.find_forward_paths` """ for path in self.find_forward_paths(goal_entity_id, start_entity_id): # Reverse path yield path[::-1] def _forward_entity_paths(self, start_entity_id, seen_entities=None): """ Generator which yields the ids of all entities connected through forward relationships, and the path taken to each. An entity will be yielded multiple times if there are multiple paths to it. Implemented using depth first search. """ if seen_entities is None: seen_entities = set() if start_entity_id in seen_entities: return seen_entities.add(start_entity_id) yield start_entity_id, [] for relationship in self.get_forward_relationships(start_entity_id): next_entity = relationship.parent_entity.id # Copy seen entities for each next node to allow multiple paths (but # not cycles). descendants = self._forward_entity_paths(next_entity, seen_entities.copy()) for sub_entity_id, sub_path in descendants: yield sub_entity_id, [relationship] + sub_path def get_forward_entities(self, entity_id, deep=False): """ Get entities that are in a forward relationship with entity Args: entity_id (str): Id entity of entity to search from. deep (bool): if True, recursively find forward entities. Yields a tuple of (descendent_id, path from entity_id to descendant). """ for relationship in self.get_forward_relationships(entity_id): parent_eid = relationship.parent_entity.id direct_path = RelationshipPath([(True, relationship)]) yield parent_eid, direct_path if deep: sub_entities = self.get_forward_entities(parent_eid, deep=True) for sub_eid, path in sub_entities: yield sub_eid, direct_path + path def get_backward_entities(self, entity_id, deep=False): """ Get entities that are in a backward relationship with entity Args: entity_id (str): Id entity of entity to search from. deep (bool): if True, recursively find backward entities. Yields a tuple of (descendent_id, path from entity_id to descendant). """ for relationship in self.get_backward_relationships(entity_id): child_eid = relationship.child_entity.id direct_path = RelationshipPath([(False, relationship)]) yield child_eid, direct_path if deep: sub_entities = self.get_backward_entities(child_eid, deep=True) for sub_eid, path in sub_entities: yield sub_eid, direct_path + path def get_forward_relationships(self, entity_id): """Get relationships where entity "entity_id" is the child Args: entity_id (str): Id of entity to get relationships for. Returns: list[:class:`.Relationship`]: List of forward relationships. """ return [r for r in self.relationships if r.child_entity.id == entity_id] def get_backward_relationships(self, entity_id): """ get relationships where entity "entity_id" is the parent. Args: entity_id (str): Id of entity to get relationships for. Returns: list[:class:`.Relationship`]: list of backward relationships """ return [r for r in self.relationships if r.parent_entity.id == entity_id] def has_unique_forward_path(self, start_entity_id, end_entity_id): """ Is the forward path from start to end unique? This will raise if there is no such path. """ paths = self.find_forward_paths(start_entity_id, end_entity_id) next(paths) second_path = next(paths, None) return not second_path ########################################################################### # Entity creation methods ############################################## ########################################################################### def entity_from_dataframe(self, entity_id, dataframe, index=None, variable_types=None, make_index=False, time_index=None, secondary_time_index=None, already_sorted=False): """ Load the data for a specified entity from a Pandas DataFrame. Args: entity_id (str) : Unique id to associate with this entity. dataframe (pandas.DataFrame) : Dataframe containing the data. index (str, optional): Name of the variable used to index the entity. If None, take the first column. variable_types (dict[str -> Variable/str], optional): Keys are of variable ids and values are variable types or type_strings. Used to to initialize an entity's store. make_index (bool, optional) : If True, assume index does not exist as a column in dataframe, and create a new column of that name using integers. Otherwise, assume index exists. time_index (str, optional): Name of the variable containing time data. Type must be in :class:`variables.DateTime` or be able to be cast to datetime (e.g. str, float, or numeric.) secondary_time_index (dict[str -> Variable]): Name of variable containing time data to use a second time index for the entity. already_sorted (bool, optional) : If True, assumes that input dataframe is already sorted by time. Defaults to False. Notes: Will infer variable types from Pandas dtype Example: .. ipython:: python import featuretools as ft import pandas as pd transactions_df = pd.DataFrame({"id": [1, 2, 3, 4, 5, 6], "session_id": [1, 2, 1, 3, 4, 5], "amount": [100.40, 20.63, 33.32, 13.12, 67.22, 1.00], "transaction_time": pd.date_range(start="10:00", periods=6, freq="10s"), "fraud": [True, False, True, False, True, True]}) es = ft.EntitySet("example") es.entity_from_dataframe(entity_id="transactions", index="id", time_index="transaction_time", dataframe=transactions_df) es["transactions"] es["transactions"].df """ variable_types = variable_types or {} if time_index is not None and time_index == index: raise ValueError("time_index and index cannot be the same value, %s" % (time_index)) if time_index is None: for variable, variable_type in variable_types.items(): if variable_type == vtypes.DatetimeTimeIndex: raise ValueError("DatetimeTimeIndex variable %s must be set using time_index parameter" % (variable)) if len(self.entities) > 0: if not isinstance(dataframe, type(self.entities[0].df)): raise ValueError("All entity dataframes must be of the same type. " "Cannot add entity of type {} to an entityset with existing entities " "of type {}".format(type(dataframe), type(self.entities[0].df))) entity = Entity( entity_id, dataframe, self, variable_types=variable_types, index=index, time_index=time_index, secondary_time_index=secondary_time_index, already_sorted=already_sorted, make_index=make_index) self.entity_dict[entity.id] = entity self.reset_data_description() return self def normalize_entity(self, base_entity_id, new_entity_id, index, additional_variables=None, copy_variables=None, make_time_index=None, make_secondary_time_index=None, new_entity_time_index=None, new_entity_secondary_time_index=None): """Create a new entity and relationship from unique values of an existing variable. Args: base_entity_id (str) : Entity id from which to split. new_entity_id (str): Id of the new entity. index (str): Variable in old entity that will become index of new entity. Relationship will be created across this variable. additional_variables (list[str]): List of variable ids to remove from base_entity and move to new entity. copy_variables (list[str]): List of variable ids to copy from old entity and move to new entity. make_time_index (bool or str, optional): Create time index for new entity based on time index in base_entity, optionally specifying which variable in base_entity to use for time_index. If specified as True without a specific variable, uses the primary time index. Defaults to True if base entity has a time index. make_secondary_time_index (dict[str -> list[str]], optional): Create a secondary time index from key. Values of dictionary are the variables to associate with the secondary time index. Only one secondary time index is allowed. If None, only associate the time index. new_entity_time_index (str, optional): Rename new entity time index. new_entity_secondary_time_index (str, optional): Rename new entity secondary time index. """ base_entity = self.entity_dict[base_entity_id] additional_variables = additional_variables or [] copy_variables = copy_variables or [] # Check base entity to make sure time index is valid if base_entity.time_index is not None: t_index = base_entity[base_entity.time_index] if not isinstance(t_index, (vtypes.NumericTimeIndex, vtypes.DatetimeTimeIndex)): base_error = "Time index '{0}' is not a NumericTimeIndex or DatetimeTimeIndex, but type {1}. Use set_time_index on entity '{2}' to set the time_index." raise TypeError(base_error.format(base_entity.time_index, type(t_index), str(base_entity.id))) if not isinstance(additional_variables, list): raise TypeError("'additional_variables' must be a list, but received type {}" .format(type(additional_variables))) if len(additional_variables) != len(set(additional_variables)): raise ValueError("'additional_variables' contains duplicate variables. All variables must be unique.") if not isinstance(copy_variables, list): raise TypeError("'copy_variables' must be a list, but received type {}" .format(type(copy_variables))) if len(copy_variables) != len(set(copy_variables)): raise ValueError("'copy_variables' contains duplicate variables. All variables must be unique.") for v in additional_variables + copy_variables: if v == index: raise ValueError("Not copying {} as both index and variable".format(v)) for v in additional_variables: if v == base_entity.time_index: raise ValueError("Not moving {} as it is the base time index variable. Perhaps, move the variable to the copy_variables.".format(v)) if isinstance(make_time_index, str): if make_time_index not in base_entity.df.columns: raise ValueError("'make_time_index' must be a variable in the base entity") elif make_time_index not in additional_variables + copy_variables: raise ValueError("'make_time_index' must be specified in 'additional_variables' or 'copy_variables'") if index == base_entity.index: raise ValueError("'index' must be different from the index column of the base entity") transfer_types = {} transfer_types[index] = type(base_entity[index]) for v in additional_variables + copy_variables: if type(base_entity[v]) == vtypes.DatetimeTimeIndex: transfer_types[v] = vtypes.Datetime elif type(base_entity[v]) == vtypes.NumericTimeIndex: transfer_types[v] = vtypes.Numeric else: transfer_types[v] = type(base_entity[v]) # create and add new entity new_entity_df = self[base_entity_id].df.copy() if make_time_index is None and base_entity.time_index is not None: make_time_index = True if isinstance(make_time_index, str): # Set the new time index to make_time_index. base_time_index = make_time_index new_entity_time_index = make_time_index already_sorted = (new_entity_time_index == base_entity.time_index) elif make_time_index: # Create a new time index based on the base entity time index. base_time_index = base_entity.time_index if new_entity_time_index is None: new_entity_time_index = "first_%s_time" % (base_entity.id) already_sorted = True assert base_entity.time_index is not None, \ "Base entity doesn't have time_index defined" if base_time_index not in [v for v in additional_variables]: copy_variables.append(base_time_index) transfer_types[new_entity_time_index] = type(base_entity[base_entity.time_index]) else: new_entity_time_index = None already_sorted = False if new_entity_time_index is not None and new_entity_time_index == index: raise ValueError("time_index and index cannot be the same value, %s" % (new_entity_time_index)) selected_variables = [index] +\ [v for v in additional_variables] +\ [v for v in copy_variables] new_entity_df2 = new_entity_df. \ drop_duplicates(index, keep='first')[selected_variables] if make_time_index: new_entity_df2 = new_entity_df2.rename(columns={base_time_index: new_entity_time_index}) if make_secondary_time_index: assert len(make_secondary_time_index) == 1, "Can only provide 1 secondary time index" secondary_time_index = list(make_secondary_time_index.keys())[0] secondary_variables = [index, secondary_time_index] + list(make_secondary_time_index.values())[0] secondary_df = new_entity_df. \ drop_duplicates(index, keep='last')[secondary_variables] if new_entity_secondary_time_index: secondary_df = secondary_df.rename(columns={secondary_time_index: new_entity_secondary_time_index}) secondary_time_index = new_entity_secondary_time_index else: new_entity_secondary_time_index = secondary_time_index secondary_df = secondary_df.set_index(index) new_entity_df = new_entity_df2.join(secondary_df, on=index) else: new_entity_df = new_entity_df2 base_entity_index = index transfer_types[index] = vtypes.Categorical if make_secondary_time_index: old_ti_name = list(make_secondary_time_index.keys())[0] ti_cols = list(make_secondary_time_index.values())[0] ti_cols = [c if c != old_ti_name else secondary_time_index for c in ti_cols] make_secondary_time_index = {secondary_time_index: ti_cols} self.entity_from_dataframe( new_entity_id, new_entity_df, index, already_sorted=already_sorted, time_index=new_entity_time_index, secondary_time_index=make_secondary_time_index, variable_types=transfer_types) self.entity_dict[base_entity_id].delete_variables(additional_variables) new_entity = self.entity_dict[new_entity_id] base_entity.convert_variable_type(base_entity_index, vtypes.Id, convert_data=False) self.add_relationship(Relationship(new_entity[index], base_entity[base_entity_index])) self.reset_data_description() return self ########################################################################### # Data wrangling methods ############################################### ########################################################################### def concat(self, other, inplace=False): '''Combine entityset with another to create a new entityset with the combined data of both entitysets. ''' assert_string = "Entitysets must have the same entities, relationships"\ ", and variable_ids" assert (self.__eq__(other) and self.relationships == other.relationships), assert_string for entity in self.entities: assert entity.id in other.entity_dict, assert_string assert (len(self[entity.id].variables) == len(other[entity.id].variables)), assert_string other_variable_ids = [o_variable.id for o_variable in other[entity.id].variables] assert (all([variable.id in other_variable_ids for variable in self[entity.id].variables])), assert_string if inplace: combined_es = self else: combined_es = copy.deepcopy(self) has_last_time_index = [] for entity in self.entities: self_df = entity.df other_df = other[entity.id].df combined_df = pd.concat([self_df, other_df]) if entity.created_index == entity.index: columns = [col for col in combined_df.columns if col != entity.index or col != entity.time_index] else: columns = [entity.index] combined_df.drop_duplicates(columns, inplace=True) if entity.time_index: combined_df.sort_values([entity.time_index, entity.index], inplace=True) else: combined_df.sort_index(inplace=True) if (entity.last_time_index is not None or other[entity.id].last_time_index is not None): has_last_time_index.append(entity.id) combined_es[entity.id].update_data(df=combined_df, recalculate_last_time_indexes=False) combined_es.add_last_time_indexes(updated_entities=has_last_time_index) self.reset_data_description() return combined_es ########################################################################### # Indexing methods ############################################### ########################################################################### def add_last_time_indexes(self, updated_entities=None): """ Calculates the last time index values for each entity (the last time an instance or children of that instance were observed). Used when calculating features using training windows Args: updated_entities (list[str]): List of entity ids to update last_time_index for (will update all parents of those entities as well) """ # Generate graph of entities to find leaf entities children = defaultdict(list) # parent --> child mapping child_vars = defaultdict(dict) for r in self.relationships: children[r.parent_entity.id].append(r.child_entity) child_vars[r.parent_entity.id][r.child_entity.id] = r.child_variable updated_entities = updated_entities or [] if updated_entities: # find parents of updated_entities parent_queue = updated_entities[:] parents = set() while len(parent_queue): e = parent_queue.pop(0) if e in parents: continue parents.add(e) for parent_id, _ in self.get_forward_entities(e): parent_queue.append(parent_id) queue = [self[p] for p in parents] to_explore = parents else: to_explore = set([e.id for e in self.entities[:]]) queue = self.entities[:] explored = set() for e in queue: e.last_time_index = None # We will explore children of entities on the queue, # which may not be in the to_explore set. Therefore, # we check whether all elements of to_explore are in # explored, rather than just comparing length while not to_explore.issubset(explored): entity = queue.pop(0) if entity.last_time_index is None: if entity.time_index is not None: lti = entity.df[entity.time_index].copy() if isinstance(entity.df, dd.DataFrame): # The current Dask implementation doesn't set the index of the dataframe # to the entity's index, so we have to do it manually here lti.index = entity.df[entity.index].copy() else: lti = entity.df[entity.index].copy() if isinstance(entity.df, dd.DataFrame): lti.index = entity.df[entity.index].copy() lti = lti.apply(lambda x: None) else: lti[:] = None entity.last_time_index = lti if entity.id in children: child_entities = children[entity.id] # if all children not explored, skip for now if not set([e.id for e in child_entities]).issubset(explored): # Now there is a possibility that a child entity # was not explicitly provided in updated_entities, # and never made it onto the queue. If updated_entities # is None then we just load all entities onto the queue # so we didn't need this logic for e in child_entities: if e.id not in explored and e.id not in [q.id for q in queue]: queue.append(e) queue.append(entity) continue # updated last time from all children for child_e in child_entities: if child_e.last_time_index is None: continue link_var = child_vars[entity.id][child_e.id].id if isinstance(child_e.last_time_index, dd.Series): to_join = child_e.df[link_var] to_join.index = child_e.df[child_e.index] lti_df = child_e.last_time_index.to_frame(name='last_time').join( to_join.to_frame(name=entity.index) ) new_index = lti_df.index.copy() new_index.name = None lti_df.index = new_index lti_df = lti_df.groupby(lti_df[entity.index]).agg('max') lti_df = entity.last_time_index.to_frame(name='last_time_old').join(lti_df) else: lti_df = pd.DataFrame({'last_time': child_e.last_time_index, entity.index: child_e.df[link_var]}) # sort by time and keep only the most recent lti_df.sort_values(['last_time', entity.index], kind="mergesort", inplace=True) lti_df.drop_duplicates(entity.index, keep='last', inplace=True) lti_df.set_index(entity.index, inplace=True) lti_df = lti_df.reindex(entity.last_time_index.index) lti_df['last_time_old'] = entity.last_time_index if not isinstance(lti_df, dd.DataFrame) and lti_df.empty: # Pandas errors out if it tries to do fillna and then max on an empty dataframe lti_df = pd.Series() else: lti_df['last_time'] = lti_df['last_time'].astype('datetime64[ns]') lti_df['last_time_old'] = lti_df['last_time_old'].astype('datetime64[ns]') lti_df = lti_df.fillna(pd.to_datetime('1800-01-01 00:00')).max(axis=1) lti_df = lti_df.replace(pd.to_datetime('1800-01-01 00:00'), pd.NaT) # lti_df = lti_df.apply(lambda x: x.dropna().max(), axis=1) entity.last_time_index = lti_df entity.last_time_index.name = 'last_time' explored.add(entity.id) self.reset_data_description() ########################################################################### # Other ############################################### ########################################################################### def add_interesting_values(self, max_values=5, verbose=False): """Find interesting values for categorical variables, to be used to generate "where" clauses Args: max_values (int) : Maximum number of values per variable to add. verbose (bool) : If True, print summary of interesting values found. Returns: None """ for entity in self.entities: entity.add_interesting_values(max_values=max_values, verbose=verbose) self.reset_data_description() def plot(self, to_file=None): """ Create a UML diagram-ish graph of the EntitySet. Args: to_file (str, optional) : Path to where the plot should be saved. If set to None (as by default), the plot will not be saved. Returns: graphviz.Digraph : Graph object that can directly be displayed in Jupyter notebooks. """ GRAPHVIZ_ERR_MSG = ('Please install graphviz to plot entity sets.' + ' (See https://docs.featuretools.com/en/stable/getting_started/install.html#installing-graphviz for' + ' details)') graphviz = import_or_raise("graphviz", GRAPHVIZ_ERR_MSG) # Try rendering a dummy graph to see if a working backend is installed try: graphviz.Digraph().pipe() except graphviz.backend.ExecutableNotFound: raise RuntimeError( "To plot entity sets, a graphviz backend is required.\n" + "Install the backend using one of the following commands:\n" + " Mac OS: brew install graphviz\n" + " Linux (Ubuntu): sudo apt-get install graphviz\n" + " Windows: conda install python-graphviz\n" + " For more details visit: https://docs.featuretools.com/en/stable/getting_started/install.html" ) if to_file: # Explicitly cast to str in case a Path object was passed in to_file = str(to_file) split_path = to_file.split('.') if len(split_path) < 2: raise ValueError("Please use a file extension like '.pdf'" + " so that the format can be inferred") format = split_path[-1] valid_formats = graphviz.backend.FORMATS if format not in valid_formats: raise ValueError("Unknown format. Make sure your format is" + " amongst the following: %s" % valid_formats) else: format = None # Initialize a new directed graph graph = graphviz.Digraph(self.id, format=format, graph_attr={'splines': 'ortho'}) # Draw entities for entity in self.entities: variables_string = '\l'.join([var.id + ' : ' + var.type_string # noqa: W605 for var in entity.variables]) nrows = entity.shape[0] label = '{%s (%d row%s)|%s\l}' % (entity.id, nrows, 's' * (nrows > 1), variables_string) # noqa: W605 graph.node(entity.id, shape='record', label=label) # Draw relationships for rel in self.relationships: # Display the key only once if is the same for both related entities if rel._parent_variable_id == rel._child_variable_id: label = rel._parent_variable_id else: label = '%s -> %s' % (rel._parent_variable_id, rel._child_variable_id) graph.edge(rel._child_entity_id, rel._parent_entity_id, xlabel=label) if to_file: # Graphviz always appends the format to the file name, so we need to # remove it manually to avoid file names like 'file_name.pdf.pdf' offset = len(format) + 1 # Add 1 for the dot output_path = to_file[:-offset] graph.render(output_path, cleanup=True) return graph
featuretools/entityset/entityset.py
43,227
Stores all actual data for a entityset Attributes: id entity_dict relationships time_type Properties: metadata Get entity instance from entityset Args: entity_id (str): Id of entity. Returns: :class:`.Entity` : Instance of entity. None if entity doesn't exist. Creates EntitySet Args: id (str) : Unique identifier to associate with this instance entities (dict[str -> tuple(pd.DataFrame, str, str, dict[str -> Variable])]): dictionary of entities. Entries take the format {entity id -> (dataframe, id column, (time_index), (variable_types), (make_index))}. Note that time_index, variable_types and make_index are optional. relationships (list[(str, str, str, str)]): List of relationships between entities. List items are a tuple with the format (parent entity id, parent variable, child entity id, child variable). Example: .. code-block:: python entities = { "cards" : (card_df, "id"), "transactions" : (transactions_df, "id", "transaction_time") } relationships = [("cards", "id", "transactions", "card_id")] ft.EntitySet("my-entity-set", entities, relationships) Generator which yields the ids of all entities connected through forward relationships, and the path taken to each. An entity will be yielded multiple times if there are multiple paths to it. Implemented using depth first search. Find interesting values for categorical variables, to be used to generate "where" clauses Args: max_values (int) : Maximum number of values per variable to add. verbose (bool) : If True, print summary of interesting values found. Returns: None Calculates the last time index values for each entity (the last time an instance or children of that instance were observed). Used when calculating features using training windows Args: updated_entities (list[str]): List of entity ids to update last_time_index for (will update all parents of those entities as well) Add a new relationship between entities in the entityset Args: relationship (Relationship) : Instance of new relationship to be added. Add multiple new relationships to a entityset Args: relationships (list[Relationship]) : List of new relationships. Combine entityset with another to create a new entityset with the combined data of both entitysets. Load the data for a specified entity from a Pandas DataFrame. Args: entity_id (str) : Unique id to associate with this entity. dataframe (pandas.DataFrame) : Dataframe containing the data. index (str, optional): Name of the variable used to index the entity. If None, take the first column. variable_types (dict[str -> Variable/str], optional): Keys are of variable ids and values are variable types or type_strings. Used to to initialize an entity's store. make_index (bool, optional) : If True, assume index does not exist as a column in dataframe, and create a new column of that name using integers. Otherwise, assume index exists. time_index (str, optional): Name of the variable containing time data. Type must be in :class:`variables.DateTime` or be able to be cast to datetime (e.g. str, float, or numeric.) secondary_time_index (dict[str -> Variable]): Name of variable containing time data to use a second time index for the entity. already_sorted (bool, optional) : If True, assumes that input dataframe is already sorted by time. Defaults to False. Notes: Will infer variable types from Pandas dtype Example: .. ipython:: python import featuretools as ft import pandas as pd transactions_df = pd.DataFrame({"id": [1, 2, 3, 4, 5, 6], "session_id": [1, 2, 1, 3, 4, 5], "amount": [100.40, 20.63, 33.32, 13.12, 67.22, 1.00], "transaction_time": pd.date_range(start="10:00", periods=6, freq="10s"), "fraud": [True, False, True, False, True, True]}) es = ft.EntitySet("example") es.entity_from_dataframe(entity_id="transactions", index="id", time_index="transaction_time", dataframe=transactions_df) es["transactions"] es["transactions"].df Generator which yields all backward paths between a start and goal entity. Does not include paths which contain cycles. Args: start_entity_id (str) : Id of entity to start the search from. goal_entity_id (str) : Id of entity to find backward path to. See Also: :func:`BaseEntitySet.find_forward_paths` Generator which yields all forward paths between a start and goal entity. Does not include paths which contain cycles. Args: start_entity_id (str) : id of entity to start the search from goal_entity_id (str) : if of entity to find forward path to See Also: :func:`BaseEntitySet.find_backward_paths` Get entities that are in a backward relationship with entity Args: entity_id (str): Id entity of entity to search from. deep (bool): if True, recursively find backward entities. Yields a tuple of (descendent_id, path from entity_id to descendant). get relationships where entity "entity_id" is the parent. Args: entity_id (str): Id of entity to get relationships for. Returns: list[:class:`.Relationship`]: list of backward relationships Get entities that are in a forward relationship with entity Args: entity_id (str): Id entity of entity to search from. deep (bool): if True, recursively find forward entities. Yields a tuple of (descendent_id, path from entity_id to descendant). Get relationships where entity "entity_id" is the child Args: entity_id (str): Id of entity to get relationships for. Returns: list[:class:`.Relationship`]: List of forward relationships. Is the forward path from start to end unique? This will raise if there is no such path. Returns the metadata for this EntitySet. The metadata will be recomputed if it does not exist. Create a new entity and relationship from unique values of an existing variable. Args: base_entity_id (str) : Entity id from which to split. new_entity_id (str): Id of the new entity. index (str): Variable in old entity that will become index of new entity. Relationship will be created across this variable. additional_variables (list[str]): List of variable ids to remove from base_entity and move to new entity. copy_variables (list[str]): List of variable ids to copy from old entity and move to new entity. make_time_index (bool or str, optional): Create time index for new entity based on time index in base_entity, optionally specifying which variable in base_entity to use for time_index. If specified as True without a specific variable, uses the primary time index. Defaults to True if base entity has a time index. make_secondary_time_index (dict[str -> list[str]], optional): Create a secondary time index from key. Values of dictionary are the variables to associate with the secondary time index. Only one secondary time index is allowed. If None, only associate the time index. new_entity_time_index (str, optional): Rename new entity time index. new_entity_secondary_time_index (str, optional): Rename new entity secondary time index. Create a UML diagram-ish graph of the EntitySet. Args: to_file (str, optional) : Path to where the plot should be saved. If set to None (as by default), the plot will not be saved. Returns: graphviz.Digraph : Graph object that can directly be displayed in Jupyter notebooks. Write entityset to disk in the csv format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str) : Location on disk to write to (will be created as a directory) sep (str) : String of length 1. Field delimiter for the output file. encoding (str) : A string representing the encoding to use in the output file, defaults to 'utf-8'. engine (str) : Name of the engine to use. Possible values are: {'c', 'python'}. compression (str) : Name of the compression to use. Possible values are: {'gzip', 'bz2', 'zip', 'xz', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. Write entityset to disk in the parquet format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str): location on disk to write to (will be created as a directory) engine (str) : Name of the engine to use. Possible values are: {'auto', 'pyarrow', 'fastparquet'}. compression (str) : Name of the compression to use. Possible values are: {'snappy', 'gzip', 'brotli', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. Write entityset in the pickle format, location specified by `path`. Path could be a local path or a S3 path. If writing to S3 a tar archive of files will be written. Args: path (str): location on disk to write to (will be created as a directory) compression (str) : Name of the compression to use. Possible values are: {'gzip', 'bz2', 'zip', 'xz', None}. profile_name (str) : Name of AWS profile to use, False to use an anonymous profile, or None. default='warn' Public getter/setter methods _operations? this is a new pair of entities Empty dataframes (as a result of accessing Entity.metadata) default to object dtypes for discrete variables, but indexes/ids default to ints. In this case, we convert the empty column's type to int Relationship access/helper methods Reverse path Copy seen entities for each next node to allow multiple paths (but not cycles). Entity creation methods Check base entity to make sure time index is valid create and add new entity Set the new time index to make_time_index. Create a new time index based on the base entity time index. Data wrangling methods Indexing methods Generate graph of entities to find leaf entities parent --> child mapping find parents of updated_entities We will explore children of entities on the queue, which may not be in the to_explore set. Therefore, we check whether all elements of to_explore are in explored, rather than just comparing length The current Dask implementation doesn't set the index of the dataframe to the entity's index, so we have to do it manually here if all children not explored, skip for now Now there is a possibility that a child entity was not explicitly provided in updated_entities, and never made it onto the queue. If updated_entities is None then we just load all entities onto the queue so we didn't need this logic updated last time from all children sort by time and keep only the most recent Pandas errors out if it tries to do fillna and then max on an empty dataframe lti_df = lti_df.apply(lambda x: x.dropna().max(), axis=1) Other Try rendering a dummy graph to see if a working backend is installed Explicitly cast to str in case a Path object was passed in Initialize a new directed graph Draw entities noqa: W605 noqa: W605 Draw relationships Display the key only once if is the same for both related entities Graphviz always appends the format to the file name, so we need to remove it manually to avoid file names like 'file_name.pdf.pdf' Add 1 for the dot
11,731
en
0.748461
from __future__ import unicode_literals from future.builtins import int, zip from functools import reduce from operator import ior, iand from string import punctuation from django.core.exceptions import ImproperlyConfigured from django.db.models import Manager, Q, CharField, TextField from django.db.models.loading import get_models from django.db.models.manager import ManagerDescriptor from django.db.models.query import QuerySet from django.contrib.sites.managers import CurrentSiteManager as DjangoCSM from django.utils.timezone import now from django.utils.translation import ugettext_lazy as _ from mezzanine.conf import settings from mezzanine.utils.models import get_model from mezzanine.utils.sites import current_site_id from mezzanine.utils.urls import home_slug class PublishedManager(Manager): """ Provides filter for restricting items returned by status and publish date when the given user is not a staff member. """ def published(self, for_user=None): """ For non-staff users, return items with a published status and whose publish and expiry dates fall before and after the current date when specified. """ from mezzanine.core.models import CONTENT_STATUS_PUBLISHED if for_user is not None and for_user.is_staff: return self.all() return self.filter( Q(publish_date__lte=now()) | Q(publish_date__isnull=True), Q(expiry_date__gte=now()) | Q(expiry_date__isnull=True), Q(status=CONTENT_STATUS_PUBLISHED)) def get_by_natural_key(self, slug): return self.get(slug=slug) def search_fields_to_dict(fields): """ In ``SearchableQuerySet`` and ``SearchableManager``, search fields can either be a sequence, or a dict of fields mapped to weights. This function converts sequences to a dict mapped to even weights, so that we're consistently dealing with a dict of fields mapped to weights, eg: ("title", "content") -> {"title": 1, "content": 1} """ if not fields: return {} try: int(list(dict(fields).values())[0]) except (TypeError, ValueError): fields = dict(zip(fields, [1] * len(fields))) return fields class SearchableQuerySet(QuerySet): """ QuerySet providing main search functionality for ``SearchableManager``. """ def __init__(self, *args, **kwargs): self._search_ordered = False self._search_terms = set() self._search_fields = kwargs.pop("search_fields", {}) super(SearchableQuerySet, self).__init__(*args, **kwargs) def search(self, query, search_fields=None): """ Build a queryset matching words in the given search query, treating quoted terms as exact phrases and taking into account + and - symbols as modifiers controlling which terms to require and exclude. """ # ### DETERMINE FIELDS TO SEARCH ### # Use search_fields arg if given, otherwise use search_fields # initially configured by the manager class. if search_fields: self._search_fields = search_fields_to_dict(search_fields) if not self._search_fields: return self.none() # ### BUILD LIST OF TERMS TO SEARCH FOR ### # Remove extra spaces, put modifiers inside quoted terms. terms = " ".join(query.split()).replace("+ ", "+") \ .replace('+"', '"+') \ .replace("- ", "-") \ .replace('-"', '"-') \ .split('"') # Strip punctuation other than modifiers from terms and create # terms list, first from quoted terms and then remaining words. terms = [("" if t[0:1] not in "+-" else t[0:1]) + t.strip(punctuation) for t in terms[1::2] + "".join(terms[::2]).split()] # Remove stop words from terms that aren't quoted or use # modifiers, since words with these are an explicit part of # the search query. If doing so ends up with an empty term # list, then keep the stop words. terms_no_stopwords = [t for t in terms if t.lower() not in settings.STOP_WORDS] get_positive_terms = lambda terms: [t.lower().strip(punctuation) for t in terms if t[0:1] != "-"] positive_terms = get_positive_terms(terms_no_stopwords) if positive_terms: terms = terms_no_stopwords else: positive_terms = get_positive_terms(terms) # Append positive terms (those without the negative modifier) # to the internal list for sorting when results are iterated. if not positive_terms: return self.none() else: self._search_terms.update(positive_terms) # ### BUILD QUERYSET FILTER ### # Create the queryset combining each set of terms. excluded = [reduce(iand, [~Q(**{"%s__icontains" % f: t[1:]}) for f in self._search_fields.keys()]) for t in terms if t[0:1] == "-"] required = [reduce(ior, [Q(**{"%s__icontains" % f: t[1:]}) for f in self._search_fields.keys()]) for t in terms if t[0:1] == "+"] optional = [reduce(ior, [Q(**{"%s__icontains" % f: t}) for f in self._search_fields.keys()]) for t in terms if t[0:1] not in "+-"] queryset = self if excluded: queryset = queryset.filter(reduce(iand, excluded)) if required: queryset = queryset.filter(reduce(iand, required)) # Optional terms aren't relevant to the filter if there are # terms that are explicitly required. elif optional: queryset = queryset.filter(reduce(ior, optional)) return queryset.distinct() def _clone(self, *args, **kwargs): """ Ensure attributes are copied to subsequent queries. """ for attr in ("_search_terms", "_search_fields", "_search_ordered"): kwargs[attr] = getattr(self, attr) return super(SearchableQuerySet, self)._clone(*args, **kwargs) def order_by(self, *field_names): """ Mark the filter as being ordered if search has occurred. """ if not self._search_ordered: self._search_ordered = len(self._search_terms) > 0 return super(SearchableQuerySet, self).order_by(*field_names) def iterator(self): """ If search has occurred and no ordering has occurred, decorate each result with the number of search terms so that it can be sorted by the number of occurrence of terms. In the case of search fields that span model relationships, we cannot accurately match occurrences without some very complicated traversal code, which we won't attempt. So in this case, namely when there are no matches for a result (count=0), and search fields contain relationships (double underscores), we assume one match for one of the fields, and use the average weight of all search fields with relationships. """ results = super(SearchableQuerySet, self).iterator() if self._search_terms and not self._search_ordered: results = list(results) for i, result in enumerate(results): count = 0 related_weights = [] for (field, weight) in self._search_fields.items(): if "__" in field: related_weights.append(weight) for term in self._search_terms: field_value = getattr(result, field, None) if field_value: count += field_value.lower().count(term) * weight if not count and related_weights: count = int(sum(related_weights) / len(related_weights)) results[i].result_count = count return iter(results) return results class SearchableManager(Manager): """ Manager providing a chainable queryset. Adapted from http://www.djangosnippets.org/snippets/562/ search method supports spanning across models that subclass the model being used to search. """ def __init__(self, *args, **kwargs): self._search_fields = kwargs.pop("search_fields", {}) super(SearchableManager, self).__init__(*args, **kwargs) def get_search_fields(self): """ Returns the search field names mapped to weights as a dict. Used in ``get_queryset`` below to tell ``SearchableQuerySet`` which search fields to use. Also used by ``DisplayableAdmin`` to populate Django admin's ``search_fields`` attribute. Search fields can be populated via ``SearchableManager.__init__``, which then get stored in ``SearchableManager._search_fields``, which serves as an approach for defining an explicit set of fields to be used. Alternatively and more commonly, ``search_fields`` can be defined on models themselves. In this case, we look at the model and all its base classes, and build up the search fields from all of those, so the search fields are implicitly built up from the inheritence chain. Finally if no search fields have been defined at all, we fall back to any fields that are ``CharField`` or ``TextField`` instances. """ search_fields = self._search_fields.copy() if not search_fields: for cls in reversed(self.model.__mro__): super_fields = getattr(cls, "search_fields", {}) search_fields.update(search_fields_to_dict(super_fields)) if not search_fields: search_fields = [] for f in self.model._meta.fields: if isinstance(f, (CharField, TextField)): search_fields.append(f.name) search_fields = search_fields_to_dict(search_fields) return search_fields def get_queryset(self): search_fields = self.get_search_fields() return SearchableQuerySet(self.model, search_fields=search_fields) def contribute_to_class(self, model, name): """ Django 1.5 explicitly prevents managers being accessed from abstract classes, which is behaviour the search API has relied on for years. Here we reinstate it. """ super(SearchableManager, self).contribute_to_class(model, name) setattr(model, name, ManagerDescriptor(self)) def search(self, *args, **kwargs): """ Proxy to queryset's search method for the manager's model and any models that subclass from this manager's model if the model is abstract. """ if not settings.SEARCH_MODEL_CHOICES: # No choices defined - build a list of leaf models (those # without subclasses) that inherit from Displayable. models = [m for m in get_models() if issubclass(m, self.model)] parents = reduce(ior, [m._meta.get_parent_list() for m in models]) models = [m for m in models if m not in parents] elif getattr(self.model._meta, "abstract", False): # When we're combining model subclasses for an abstract # model (eg Displayable), we only want to use models that # are represented by the ``SEARCH_MODEL_CHOICES`` setting. # Now this setting won't contain an exact list of models # we should use, since it can define superclass models such # as ``Page``, so we check the parent class list of each # model when determining whether a model falls within the # ``SEARCH_MODEL_CHOICES`` setting. search_choices = set() models = set() parents = set() errors = [] for name in settings.SEARCH_MODEL_CHOICES: try: model = get_model(*name.split(".", 1)) except LookupError: errors.append(name) else: search_choices.add(model) if errors: raise ImproperlyConfigured("Could not load the model(s) " "%s defined in the 'SEARCH_MODEL_CHOICES' setting." % ", ".join(errors)) for model in get_models(): # Model is actually a subclasses of what we're # searching (eg Displayabale) is_subclass = issubclass(model, self.model) # Model satisfies the search choices list - either # there are no search choices, model is directly in # search choices, or its parent is. this_parents = set(model._meta.get_parent_list()) in_choices = not search_choices or model in search_choices in_choices = in_choices or this_parents & search_choices if is_subclass and (in_choices or not search_choices): # Add to models we'll seach. Also maintain a parent # set, used below for further refinement of models # list to search. models.add(model) parents.update(this_parents) # Strip out any models that are superclasses of models, # specifically the Page model which will generally be the # superclass for all custom content types, since if we # query the Page model as well, we will get duplicate # results. models -= parents else: models = [self.model] all_results = [] user = kwargs.pop("for_user", None) for model in models: try: queryset = model.objects.published(for_user=user) except AttributeError: queryset = model.objects.get_queryset() all_results.extend(queryset.search(*args, **kwargs)) return sorted(all_results, key=lambda r: r.result_count, reverse=True) class CurrentSiteManager(DjangoCSM): """ Extends Django's site manager to first look up site by ID stored in the request, the session, then domain for the current request (accessible via threadlocals in ``mezzanine.core.request``), the environment variable ``MEZZANINE_SITE_ID`` (which can be used by management commands with the ``--site`` arg, finally falling back to ``settings.SITE_ID`` if none of those match a site. """ def __init__(self, field_name=None, *args, **kwargs): super(DjangoCSM, self).__init__(*args, **kwargs) self.__field_name = field_name self.__is_validated = False def get_queryset(self): if not self.__is_validated: try: # Django <= 1.6 self._validate_field_name() except AttributeError: # Django >= 1.7: will populate "self.__field_name". self._get_field_name() lookup = {self.__field_name + "__id__exact": current_site_id()} return super(DjangoCSM, self).get_queryset().filter(**lookup) class DisplayableManager(CurrentSiteManager, PublishedManager, SearchableManager): """ Manually combines ``CurrentSiteManager``, ``PublishedManager`` and ``SearchableManager`` for the ``Displayable`` model. """ def url_map(self, for_user=None, **kwargs): """ Returns a dictionary of urls mapped to Displayable subclass instances, including a fake homepage instance if none exists. Used in ``mezzanine.core.sitemaps``. """ home = self.model(title=_("Home")) setattr(home, "get_absolute_url", home_slug) items = {home.get_absolute_url(): home} for model in get_models(): if issubclass(model, self.model): for item in (model.objects.published(for_user=for_user) .filter(**kwargs) .exclude(slug__startswith="http://") .exclude(slug__startswith="https://")): items[item.get_absolute_url()] = item return items
mezzanine/core/managers.py
16,435
Extends Django's site manager to first look up site by ID stored in the request, the session, then domain for the current request (accessible via threadlocals in ``mezzanine.core.request``), the environment variable ``MEZZANINE_SITE_ID`` (which can be used by management commands with the ``--site`` arg, finally falling back to ``settings.SITE_ID`` if none of those match a site. Manually combines ``CurrentSiteManager``, ``PublishedManager`` and ``SearchableManager`` for the ``Displayable`` model. Provides filter for restricting items returned by status and publish date when the given user is not a staff member. Manager providing a chainable queryset. Adapted from http://www.djangosnippets.org/snippets/562/ search method supports spanning across models that subclass the model being used to search. QuerySet providing main search functionality for ``SearchableManager``. Ensure attributes are copied to subsequent queries. Django 1.5 explicitly prevents managers being accessed from abstract classes, which is behaviour the search API has relied on for years. Here we reinstate it. Returns the search field names mapped to weights as a dict. Used in ``get_queryset`` below to tell ``SearchableQuerySet`` which search fields to use. Also used by ``DisplayableAdmin`` to populate Django admin's ``search_fields`` attribute. Search fields can be populated via ``SearchableManager.__init__``, which then get stored in ``SearchableManager._search_fields``, which serves as an approach for defining an explicit set of fields to be used. Alternatively and more commonly, ``search_fields`` can be defined on models themselves. In this case, we look at the model and all its base classes, and build up the search fields from all of those, so the search fields are implicitly built up from the inheritence chain. Finally if no search fields have been defined at all, we fall back to any fields that are ``CharField`` or ``TextField`` instances. If search has occurred and no ordering has occurred, decorate each result with the number of search terms so that it can be sorted by the number of occurrence of terms. In the case of search fields that span model relationships, we cannot accurately match occurrences without some very complicated traversal code, which we won't attempt. So in this case, namely when there are no matches for a result (count=0), and search fields contain relationships (double underscores), we assume one match for one of the fields, and use the average weight of all search fields with relationships. Mark the filter as being ordered if search has occurred. For non-staff users, return items with a published status and whose publish and expiry dates fall before and after the current date when specified. Build a queryset matching words in the given search query, treating quoted terms as exact phrases and taking into account + and - symbols as modifiers controlling which terms to require and exclude. Proxy to queryset's search method for the manager's model and any models that subclass from this manager's model if the model is abstract. In ``SearchableQuerySet`` and ``SearchableManager``, search fields can either be a sequence, or a dict of fields mapped to weights. This function converts sequences to a dict mapped to even weights, so that we're consistently dealing with a dict of fields mapped to weights, eg: ("title", "content") -> {"title": 1, "content": 1} Returns a dictionary of urls mapped to Displayable subclass instances, including a fake homepage instance if none exists. Used in ``mezzanine.core.sitemaps``. DETERMINE FIELDS TO SEARCH Use search_fields arg if given, otherwise use search_fields initially configured by the manager class. BUILD LIST OF TERMS TO SEARCH FOR Remove extra spaces, put modifiers inside quoted terms. Strip punctuation other than modifiers from terms and create terms list, first from quoted terms and then remaining words. Remove stop words from terms that aren't quoted or use modifiers, since words with these are an explicit part of the search query. If doing so ends up with an empty term list, then keep the stop words. Append positive terms (those without the negative modifier) to the internal list for sorting when results are iterated. BUILD QUERYSET FILTER Create the queryset combining each set of terms. Optional terms aren't relevant to the filter if there are terms that are explicitly required. No choices defined - build a list of leaf models (those without subclasses) that inherit from Displayable. When we're combining model subclasses for an abstract model (eg Displayable), we only want to use models that are represented by the ``SEARCH_MODEL_CHOICES`` setting. Now this setting won't contain an exact list of models we should use, since it can define superclass models such as ``Page``, so we check the parent class list of each model when determining whether a model falls within the ``SEARCH_MODEL_CHOICES`` setting. Model is actually a subclasses of what we're searching (eg Displayabale) Model satisfies the search choices list - either there are no search choices, model is directly in search choices, or its parent is. Add to models we'll seach. Also maintain a parent set, used below for further refinement of models list to search. Strip out any models that are superclasses of models, specifically the Page model which will generally be the superclass for all custom content types, since if we query the Page model as well, we will get duplicate results. Django <= 1.6 Django >= 1.7: will populate "self.__field_name".
5,542
en
0.893695
import pdfplumber import re import pandas as pd from datetime import datetime import sys # AUTHOR: Simon Rosen # ----------------------------------- # DEPENDENCIES # This module requires 'pdfplumber' # # Install: pip install pdfplumber # ----------------------------------- def extract_data(file_path): pdfp_obj = pdfplumber.open(file_path) # Helper functions # text - string you are finding substring in def get_string_between_2_strings(text, string1, string2): # print("text: {}\n string1: {}, string2:{}".format("text", string1, string2)) try: regex_str = string1 + '(.+?)' + string2 # print('regex_str: {}'.format(regex_str)) # all_found = [x.group() for x in re.finditer(regex_str, text)] all_found = re.search(regex_str, text, re.DOTALL).group(1) # print(all_found) except AttributeError: # no text found between two substrings # print('Not found') all_found = [] # apply your error handling return all_found # GP data contained in paragraph under following heading # GAUTENG CONFIRMED COVID-19 CASES DISTRICT BREAKDOWN # GP cases, recoveries, deaths, contacts traced, people de-isolated & hospitalisations def get_gp_breakdown_data(): district_pg =0 first_page_txt = pdfp_obj.pages[0].extract_text() # GAUTENG CONFIRMED COVID-19 CASES DISTRICT BREAKDOWN heading_txt_1 = "GAUTENG CONFIRMED COVID-19 CASES DISTRICT BREAKDOWN" heading_txt_2 = "BREAKDOWN PER DISTRICT" breakdown_txt = get_string_between_2_strings(first_page_txt, heading_txt_1, heading_txt_2) if len(breakdown_txt)==0: breakdown_txt = get_string_between_2_strings(pdfp_obj.pages[1].extract_text(), heading_txt_1, heading_txt_2) district_pg=1 if len(breakdown_txt)==0: breakdown_txt = get_string_between_2_strings(pdfp_obj.pages[1].extract_text(), "^", heading_txt_2) district_pg=1 str_list = list(filter(lambda x: False if x == ' ' else True, breakdown_txt.splitlines())) str_body = "".join(str_list) sentences = str_body.split('.') def find_date(text): return re.search(r'(\d{2}|\d{1}) [a-zA-Z]* \d{4}', text).group(0) def get_nums(text, exclude_texts=['COVID-19']): for exclude_text in exclude_texts: text = text.replace(exclude_text, '') num_tuples = re.findall(r'(\d{3}|\d{2}|\d{1})( \d{3}|\d{2}|\d{1})*', text) num_list = [int(x[0] + x[1].replace(' ', '')) for x in num_tuples] return num_list date_txt = get_string_between_2_strings(pdfp_obj.pages[0].extract_text(), heading_txt_1, "$") sentences = "".join(date_txt).split(".") _gp_covid_stats = {"date": find_date(date_txt)} # First Sentence tmp_dict = dict(zip(['cases', 'recoveries', 'deaths'], get_nums(sentences[0])[2:])) _gp_covid_stats.update(tmp_dict) # Second Sentence tmp_dict = dict(zip(['traced', 'de_isolated'], get_nums(sentences[1])[:2])) _gp_covid_stats.update(tmp_dict) # Third Sentence tmp_dict = dict(zip(['hospitalised'], get_nums(sentences[2]))) _gp_covid_stats.update(tmp_dict) return district_pg, _gp_covid_stats district_pg, gp_covid_stats = get_gp_breakdown_data() # DISTRICT BREAKDOWN def get_district_data(): district_table_list = pdfp_obj.pages[district_pg].extract_tables()[0] print(type(district_table_list)) dl = [] for i, row in enumerate(district_table_list): print(i,row) dl.append(list(filter(lambda x: x != None and len(x) !=0, row))) dl[-2]=dl[-2]+[0,0,0] print(dl) all_list = [[x[i] for x in dl] for i in range(0, len(dl[0]))] print(all_list,"*******") gp_breakdown_dict = {curr_list[0]: curr_list[1:] for curr_list in all_list} gp_breakdown_df = pd.DataFrame.from_dict(gp_breakdown_dict) print(gp_breakdown_df) gp_breakdown_df.fillna(0, inplace=True) gp_breakdown_df.set_index("DISTRICT", inplace=True) gp_breakdown_df.rename(inplace=True, columns={gp_breakdown_df.columns[0]: "CASES", gp_breakdown_df.columns[1]: "NEW CASES"}) for i in range(0, 4): gp_breakdown_df.iloc[:, i] = gp_breakdown_df.iloc[:, i].apply(lambda x: x if type(x)==int else x.replace(' ', '')) return gp_breakdown_df gp_district_df = get_district_data() # --------------- # SUB-DISTRICTS # --------------- def get_extracted_raw_list(page_no): currPage = pdfp_obj.pages[page_no] bounding_box = (300, 0, currPage.width, currPage.height) cropped_page = currPage.crop(bounding_box) # table_settings = {"vertical_strategy": "text"} table_settings = {"snap_tolerance": 10, "join_tolerance": 15} extracted_raw_list = cropped_page.extract_tables(table_settings)[0] return extracted_raw_list def get_sub_districts_data(raw_list): sub_districts_list = [] curr_sub_district = [] prev_sub_district = [] for i in range(1, len(raw_list)): curr_list = raw_list[i] if curr_sub_district == [] or not (curr_list[0] == None or curr_list[0] == ''): # print(prev_sub_district) if prev_sub_district != []: sub_districts_list.append(curr_sub_district) curr_sub_district = curr_list prev_sub_district = curr_sub_district # print(curr_sub_district) if (curr_sub_district[1] == '' and curr_list[1] != '' and curr_list[1] != None): curr_sub_district[1] = curr_list[1] if (curr_sub_district[2] == '' and curr_list[2] != '' and curr_list[2] != None): curr_sub_district[2] = curr_list[2] if (i == len(raw_list) - 1): sub_districts_list.append(curr_sub_district) # Check if first item of list is valid e.g. total and/or recoveries has values prev_sub_district = sub_districts_list[0] if (prev_sub_district[1] == '' or prev_sub_district[1] == None) and (prev_sub_district[2] == '' or \ prev_sub_district[2] == None): sub_districts_list.pop(0) return sub_districts_list def get_table_list(page_no): currPage = pdfp_obj.pages[page_no] bounding_box = (300, 0, currPage.width, currPage.height) cropped_page = currPage.crop(bounding_box) # table_settings = {"vertical_strategy": "text"} table_settings = {"snap_tolerance": 10, "join_tolerance": 15} extracted_raw_list = cropped_page.extract_tables(table_settings)[0] return extracted_raw_list def get_all_sub_districts(page_start, page_end): all_sub_districts = [] for i in range(page_start, page_end + 1): all_sub_districts.extend(get_sub_districts_data(get_table_list(i))) def remove_spaces(str_no): if type(str_no)==str: return str_no.replace(" ", "") else: return str_no all_sub_districts = [[x[0], remove_spaces(x[1]), remove_spaces(x[2])] for x in all_sub_districts] return all_sub_districts all_sub_dists = get_all_sub_districts(district_pg+1, district_pg+4) pdfp_obj.close() def get_district_map(): # Johannesburg jhb_dict = dict(zip(['A', 'B', 'C', 'D', 'E', 'F', 'G', 'Unallocated'], [[x[1], x[2]] for x in all_sub_dists[0:8]])) # Tshwane tsh_keys = list(range(1, 8)) tsh_keys.append('Unallocated') tsh_dict = dict(zip(tsh_keys, [[x[1], x[2]] for x in all_sub_dists[8:16]])) # Ekurhuleni eku_keys = "e1 e2 n1 n2 s1 s2 Unallocated".split(" ") eku_dict = dict(zip(eku_keys, [[x[1], x[2]] for x in all_sub_dists[16:23]])) # Sedibeng sed_keys = "Lesedi Emfuleni Midvaal Unallocated".split(" ") sed_dict = dict(zip(sed_keys, [[x[1], x[2]] for x in all_sub_dists[23:27]])) # West Rand wr_keys = "Mogale Rand_West Merafong Unallocated".split(" ") wr_dict = dict(zip(wr_keys, [[x[1], x[2]] for x in all_sub_dists[27:31]])) # All Districts district_map = { 'Johannesburg': jhb_dict, 'Tshwane': tsh_dict, 'Ekurhuleni': eku_dict, 'Sedibeng': sed_dict, 'West Rand': wr_dict } return district_map district_map = get_district_map() # DATE curr_date = datetime.strptime(gp_covid_stats['date'], '%d %B %Y') date_formatted = datetime.strftime(curr_date, '%d-%m-%Y') date_yyyymmdd = datetime.strftime(curr_date, '%Y%m%d') # print(gp_covid_stats['date'], date_formatted, date_yyyymmdd) ############################## # OUT LIST # # DETERMINES ORDER OF OUTPUT # ############################## # List later gets converted to formatted string jhb_districts = [x for x in 'ABCDEFG']+['Unallocated'] tsh_districts = [x for x in range(1,8)]+['Unallocated'] wr_districts=['Mogale',"Rand_West","Merafong","Unallocated"] out_list = [ # Date date_yyyymmdd, date_formatted, # Gauteng Data gp_covid_stats['cases'], 'Check', 'Check', gp_covid_stats['recoveries'], gp_covid_stats['deaths'], 'Check','Check', gp_covid_stats['hospitalised'], # DISTRICT TOTALS DATA # ---------------------- # Johannesburg gp_district_df.loc['Johannesburg']['CASES'], gp_district_df.loc['Ekurhuleni']['CASES'], gp_district_df.loc['Tshwane']['CASES'], gp_district_df.loc['Sedibeng']['CASES'], gp_district_df.loc['West Rand']['CASES'], gp_district_df.loc['Unallocated']['CASES'], ' Check', gp_district_df.loc['Johannesburg']['DEATHS'], gp_district_df.loc['Ekurhuleni']['DEATHS'], gp_district_df.loc['Tshwane']['DEATHS'], gp_district_df.loc['Sedibeng']['DEATHS'], gp_district_df.loc['West Rand']['DEATHS'], gp_district_df.loc['Johannesburg']['RECOVERIES'], gp_district_df.loc['Ekurhuleni']['RECOVERIES'], gp_district_df.loc['Tshwane']['RECOVERIES'], gp_district_df.loc['Sedibeng']['RECOVERIES'], gp_district_df.loc['West Rand']['RECOVERIES'], ' Check', ' Check'] + \ [district_map['Johannesburg'][x][0] for x in jhb_districts]+\ ['Check']+\ [district_map['Johannesburg'][x][1] for x in jhb_districts]+\ ['Check']+\ [district_map['Tshwane'][x][0] for x in tsh_districts]+\ ['Check']+\ [district_map['Tshwane'][x][1] for x in tsh_districts]+\ ['Check']+\ [district_map['Ekurhuleni'][x][0] for x in ['e1','e2','n1','n2','s1','s2','Unallocated']]+\ ['Check']+\ [district_map['Ekurhuleni'][x][1] for x in ['e1','e2','n1','n2','s1','s2','Unallocated']]+\ ['Check']+\ [district_map['Sedibeng'][x][0] for x in ['Lesedi','Emfuleni','Midvaal','Unallocated']]+\ ['Check']+\ [district_map['Sedibeng'][x][1] for x in ['Lesedi','Emfuleni','Midvaal','Unallocated']]+\ ['Check']+\ [district_map['West Rand'][x][0] for x in wr_districts]+\ [district_map['West Rand'][x][1] for x in wr_districts]+\ ['Check'] def list_to_formatted(in_list, delimiter='\t'): return delimiter.join(map(str, in_list)) out_str = list_to_formatted(out_list) # return district_map return out_str if __name__ == "__main__": print(extract_data(sys.argv[1]))
scripts/gp_pdf_extractor.py
12,011
AUTHOR: Simon Rosen ----------------------------------- DEPENDENCIES This module requires 'pdfplumber' Install: pip install pdfplumber ----------------------------------- Helper functions text - string you are finding substring in print("text: {}\n string1: {}, string2:{}".format("text", string1, string2)) print('regex_str: {}'.format(regex_str)) all_found = [x.group() for x in re.finditer(regex_str, text)] print(all_found) no text found between two substrings print('Not found') apply your error handling GP data contained in paragraph under following heading GAUTENG CONFIRMED COVID-19 CASES DISTRICT BREAKDOWN GP cases, recoveries, deaths, contacts traced, people de-isolated & hospitalisations GAUTENG CONFIRMED COVID-19 CASES DISTRICT BREAKDOWN First Sentence Second Sentence Third Sentence DISTRICT BREAKDOWN --------------- SUB-DISTRICTS --------------- table_settings = {"vertical_strategy": "text"} print(prev_sub_district) print(curr_sub_district) Check if first item of list is valid e.g. total and/or recoveries has values table_settings = {"vertical_strategy": "text"} Johannesburg Tshwane Ekurhuleni Sedibeng West Rand All Districts DATE print(gp_covid_stats['date'], date_formatted, date_yyyymmdd) OUT LIST DETERMINES ORDER OF OUTPUT List later gets converted to formatted string Date Gauteng Data DISTRICT TOTALS DATA ---------------------- Johannesburg return district_map
1,460
en
0.541883
# # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # pylint: disable=not-callable import re import unittest from unittest import mock import pytest from google.cloud.bigquery import DEFAULT_RETRY, DatasetReference, Table, TableReference from google.cloud.bigquery.dataset import AccessEntry, Dataset, DatasetListItem from google.cloud.exceptions import NotFound from parameterized import parameterized from airflow import AirflowException from airflow.providers.google.cloud.hooks.bigquery import ( BigQueryCursor, BigQueryHook, _api_resource_configs_duplication_check, _cleanse_time_partitioning, _split_tablename, _validate_src_fmt_configs, _validate_value, ) PROJECT_ID = "bq-project" CREDENTIALS = "bq-credentials" DATASET_ID = "bq_dataset" TABLE_ID = "bq_table" PARTITION_ID = "20200101" VIEW_ID = 'bq_view' JOB_ID = "1234" LOCATION = 'europe-north1' TABLE_REFERENCE_REPR = { 'tableId': TABLE_ID, 'datasetId': DATASET_ID, 'projectId': PROJECT_ID, } TABLE_REFERENCE = TableReference.from_api_repr(TABLE_REFERENCE_REPR) class _BigQueryBaseTestClass(unittest.TestCase): def setUp(self) -> None: class MockedBigQueryHook(BigQueryHook): def _get_credentials_and_project_id(self): return CREDENTIALS, PROJECT_ID self.hook = MockedBigQueryHook() class TestBigQueryHookMethods(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryConnection") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook._authorize") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.build") def test_bigquery_client_creation(self, mock_build, mock_authorize, mock_bigquery_connection): result = self.hook.get_conn() mock_build.assert_called_once_with( 'bigquery', 'v2', http=mock_authorize.return_value, cache_discovery=False ) mock_bigquery_connection.assert_called_once_with( service=mock_build.return_value, project_id=PROJECT_ID, hook=self.hook, use_legacy_sql=self.hook.use_legacy_sql, location=self.hook.location, num_retries=self.hook.num_retries, ) assert mock_bigquery_connection.return_value == result @mock.patch("airflow.providers.google.common.hooks.base_google.GoogleBaseHook.__init__") def test_bigquery_bigquery_conn_id_deprecation_warning( self, mock_base_hook_init, ): bigquery_conn_id = "bigquery conn id" warning_message = ( "The bigquery_conn_id parameter has been deprecated. " "You should pass the gcp_conn_id parameter." ) with pytest.warns(DeprecationWarning) as warnings: BigQueryHook(bigquery_conn_id=bigquery_conn_id) mock_base_hook_init.assert_called_once_with( delegate_to=None, gcp_conn_id='bigquery conn id', impersonation_chain=None, ) assert warning_message == str(warnings[0].message) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_location_propagates_properly(self, run_with_config, _): # TODO: this creates side effect assert self.hook.location is None self.hook.run_query(sql='select 1', location='US') assert run_with_config.call_count == 1 assert self.hook.location == 'US' def test_bigquery_insert_rows_not_implemented(self): with pytest.raises(NotImplementedError): self.hook.insert_rows(table="table", rows=[1, 2]) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_bigquery_table_exists_true(self, mock_client): result = self.hook.table_exists(project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID) mock_client.return_value.get_table.assert_called_once_with(TABLE_REFERENCE) mock_client.assert_called_once_with(project_id=PROJECT_ID) assert result is True @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_bigquery_table_exists_false(self, mock_client): mock_client.return_value.get_table.side_effect = NotFound("Dataset not found") result = self.hook.table_exists(project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID) mock_client.return_value.get_table.assert_called_once_with(TABLE_REFERENCE) mock_client.assert_called_once_with(project_id=PROJECT_ID) assert result is False @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_bigquery_table_partition_exists_true(self, mock_client): mock_client.return_value.list_partitions.return_value = [PARTITION_ID] result = self.hook.table_partition_exists( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, partition_id=PARTITION_ID ) mock_client.return_value.list_partitions.assert_called_once_with(TABLE_REFERENCE) mock_client.assert_called_once_with(project_id=PROJECT_ID) assert result is True @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_bigquery_table_partition_exists_false_no_table(self, mock_client): mock_client.return_value.get_table.side_effect = NotFound("Dataset not found") result = self.hook.table_partition_exists( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, partition_id=PARTITION_ID ) mock_client.return_value.list_partitions.assert_called_once_with(TABLE_REFERENCE) mock_client.assert_called_once_with(project_id=PROJECT_ID) assert result is False @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_bigquery_table_partition_exists_false_no_partition(self, mock_client): mock_client.return_value.list_partitions.return_value = [] result = self.hook.table_partition_exists( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, partition_id=PARTITION_ID ) mock_client.return_value.list_partitions.assert_called_once_with(TABLE_REFERENCE) mock_client.assert_called_once_with(project_id=PROJECT_ID) assert result is False @mock.patch('airflow.providers.google.cloud.hooks.bigquery.read_gbq') def test_get_pandas_df(self, mock_read_gbq): self.hook.get_pandas_df('select 1') mock_read_gbq.assert_called_once_with( 'select 1', credentials=CREDENTIALS, dialect='legacy', project_id=PROJECT_ID, verbose=False ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_invalid_schema_update_options(self, mock_get_service): with pytest.raises( Exception, match=( r"\['THIS IS NOT VALID'\] contains invalid schema update options." r"Please only use one or more of the following options: " r"\['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION'\]" ), ): self.hook.run_load( "test.test", "test_schema.json", ["test_data.json"], schema_update_options=["THIS IS NOT VALID"], ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_invalid_schema_update_and_write_disposition(self, mock_get_service): with pytest.raises( Exception, match="schema_update_options is only allowed if" " write_disposition is 'WRITE_APPEND' or 'WRITE_TRUNCATE'.", ): self.hook.run_load( "test.test", "test_schema.json", ["test_data.json"], schema_update_options=['ALLOW_FIELD_ADDITION'], write_disposition='WRITE_EMPTY', ) @mock.patch( "airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.poll_job_complete", side_effect=[False, True], ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_cancel_queries(self, mock_client, mock_poll_job_complete): running_job_id = 3 self.hook.running_job_id = running_job_id self.hook.cancel_query() mock_poll_job_complete.has_calls(mock.call(running_job_id), mock.call(running_job_id)) mock_client.assert_called_once_with(project_id=PROJECT_ID, location=None) mock_client.return_value.cancel_job.assert_called_once_with(job_id=running_job_id) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_sql_dialect_default( self, mock_insert, _, ): self.hook.run_query('query') _, kwargs = mock_insert.call_args assert kwargs['configuration']['query']['useLegacySql'] is True @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_sql_dialect(self, mock_insert, _): self.hook.run_query('query', use_legacy_sql=False) _, kwargs = mock_insert.call_args assert kwargs['configuration']['query']['useLegacySql'] is False @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_sql_dialect_legacy_with_query_params(self, mock_insert, _): params = [ { 'name': "param_name", 'parameterType': {'type': "STRING"}, 'parameterValue': {'value': "param_value"}, } ] self.hook.run_query('query', use_legacy_sql=False, query_params=params) _, kwargs = mock_insert.call_args assert kwargs['configuration']['query']['useLegacySql'] is False @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_run_query_sql_dialect_legacy_with_query_params_fails(self, _): params = [ { 'name': "param_name", 'parameterType': {'type': "STRING"}, 'parameterValue': {'value': "param_value"}, } ] with pytest.raises(ValueError, match="Query parameters are not allowed when using legacy SQL"): self.hook.run_query('query', use_legacy_sql=True, query_params=params) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_run_query_without_sql_fails(self, _): with pytest.raises( TypeError, match=r"`BigQueryBaseCursor.run_query` missing 1 required positional argument: `sql`" ): self.hook.run_query(sql=None) @parameterized.expand( [ (['ALLOW_FIELD_ADDITION'], 'WRITE_APPEND'), (['ALLOW_FIELD_RELAXATION'], 'WRITE_APPEND'), (['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION'], 'WRITE_APPEND'), (['ALLOW_FIELD_ADDITION'], 'WRITE_TRUNCATE'), (['ALLOW_FIELD_RELAXATION'], 'WRITE_TRUNCATE'), (['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION'], 'WRITE_TRUNCATE'), ] ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_schema_update_options( self, schema_update_options, write_disposition, mock_insert, mock_get_service, ): self.hook.run_query( sql='query', destination_dataset_table='my_dataset.my_table', schema_update_options=schema_update_options, write_disposition=write_disposition, ) _, kwargs = mock_insert.call_args assert kwargs['configuration']['query']['schemaUpdateOptions'] == schema_update_options assert kwargs['configuration']['query']['writeDisposition'] == write_disposition @parameterized.expand( [ ( ['INCORRECT_OPTION'], None, r"\['INCORRECT_OPTION'\] contains invalid schema update options\. " r"Please only use one or more of the following options: " r"\['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION'\]", ), ( ['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION', 'INCORRECT_OPTION'], None, r"\['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION', 'INCORRECT_OPTION'\] contains invalid " r"schema update options\. Please only use one or more of the following options: " r"\['ALLOW_FIELD_ADDITION', 'ALLOW_FIELD_RELAXATION'\]", ), ( ['ALLOW_FIELD_ADDITION'], None, r"schema_update_options is only allowed if write_disposition is " r"'WRITE_APPEND' or 'WRITE_TRUNCATE'", ), ] ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_run_query_schema_update_options_incorrect( self, schema_update_options, write_disposition, expected_regex, mock_get_service, ): with pytest.raises(ValueError, match=expected_regex): self.hook.run_query( sql='query', destination_dataset_table='my_dataset.my_table', schema_update_options=schema_update_options, write_disposition=write_disposition, ) @parameterized.expand([(True,), (False,)]) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_api_resource_configs( self, bool_val, mock_insert, _, ): self.hook.run_query('query', api_resource_configs={'query': {'useQueryCache': bool_val}}) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['query']['useQueryCache'] is bool_val assert kwargs["configuration"]['query']['useLegacySql'] is True @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_api_resource_configs_duplication_warning(self, mock_get_service): with pytest.raises( ValueError, match=( r"Values of useLegacySql param are duplicated\. api_resource_configs " r"contained useLegacySql param in `query` config and useLegacySql was " r"also provided with arg to run_query\(\) method\. Please remove duplicates\." ), ): self.hook.run_query( 'query', use_legacy_sql=True, api_resource_configs={'query': {'useLegacySql': False}} ) def test_validate_value(self): with pytest.raises( TypeError, match="case_1 argument must have a type <class 'dict'> not <class 'str'>" ): _validate_value("case_1", "a", dict) assert _validate_value("case_2", 0, int) is None def test_duplication_check(self): with pytest.raises( ValueError, match=r"Values of key_one param are duplicated. api_resource_configs contained key_one param in" r" `query` config and key_one was also provided with arg to run_query\(\) method. " r"Please remove duplicates.", ): key_one = True _api_resource_configs_duplication_check("key_one", key_one, {"key_one": False}) assert _api_resource_configs_duplication_check("key_one", key_one, {"key_one": True}) is None def test_validate_src_fmt_configs(self): source_format = "test_format" valid_configs = ["test_config_known", "compatibility_val"] backward_compatibility_configs = {"compatibility_val": "val"} with pytest.raises( ValueError, match="test_config_unknown is not a valid src_fmt_configs for type test_format." ): # This config should raise a value error. src_fmt_configs = {"test_config_unknown": "val"} _validate_src_fmt_configs( source_format, src_fmt_configs, valid_configs, backward_compatibility_configs ) src_fmt_configs = {"test_config_known": "val"} src_fmt_configs = _validate_src_fmt_configs( source_format, src_fmt_configs, valid_configs, backward_compatibility_configs ) assert ( "test_config_known" in src_fmt_configs ), "src_fmt_configs should contain al known src_fmt_configs" assert ( "compatibility_val" in src_fmt_configs ), "_validate_src_fmt_configs should add backward_compatibility config" @parameterized.expand([("AVRO",), ("PARQUET",), ("NEWLINE_DELIMITED_JSON",), ("DATASTORE_BACKUP",)]) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_with_non_csv_as_src_fmt(self, fmt, _): try: self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', source_uris=[], source_format=fmt, autodetect=True, ) except ValueError: self.fail("run_load() raised ValueError unexpectedly!") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_extract(self, mock_insert): source_project_dataset_table = f"{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}" destination_cloud_storage_uris = ["gs://bucket/file.csv"] expected_configuration = { "extract": { "sourceTable": { "projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": TABLE_ID, }, "compression": "NONE", "destinationUris": destination_cloud_storage_uris, "destinationFormat": "CSV", "fieldDelimiter": ",", "printHeader": True, } } self.hook.run_extract( source_project_dataset_table=source_project_dataset_table, destination_cloud_storage_uris=destination_cloud_storage_uris, ) mock_insert.assert_called_once_with(configuration=expected_configuration, project_id=PROJECT_ID) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.SchemaField") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_list_rows(self, mock_client, mock_schema, mock_table): self.hook.list_rows( dataset_id=DATASET_ID, table_id=TABLE_ID, max_results=10, selected_fields=["field_1", "field_2"], page_token="page123", start_index=5, location=LOCATION, ) mock_table.from_api_repr.assert_called_once_with({"tableReference": TABLE_REFERENCE_REPR}) mock_schema.has_calls([mock.call(x, "") for x in ["field_1", "field_2"]]) mock_client.return_value.list_rows.assert_called_once_with( table=mock_table.from_api_repr.return_value, max_results=10, selected_fields=mock.ANY, page_token='page123', start_index=5, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_list_rows_with_empty_selected_fields(self, mock_client, mock_table): self.hook.list_rows( dataset_id=DATASET_ID, table_id=TABLE_ID, max_results=10, page_token="page123", selected_fields=[], start_index=5, location=LOCATION, ) mock_table.from_api_repr.assert_called_once_with({"tableReference": TABLE_REFERENCE_REPR}) mock_client.return_value.list_rows.assert_called_once_with( table=mock_table.from_api_repr.return_value, max_results=10, page_token='page123', selected_fields=None, start_index=5, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_run_table_delete(self, mock_client, mock_table): source_project_dataset_table = f"{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}" self.hook.run_table_delete(source_project_dataset_table, ignore_if_missing=False) mock_table.from_string.assert_called_once_with(source_project_dataset_table) mock_client.return_value.delete_table.assert_called_once_with( table=mock_table.from_string.return_value, not_found_ok=False ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.create_empty_table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_dataset_tables") def test_table_upsert_create_new_table(self, mock_get, mock_create): table_resource = {"tableReference": {"tableId": TABLE_ID}} mock_get.return_value = [] self.hook.run_table_upsert(dataset_id=DATASET_ID, table_resource=table_resource) mock_get.assert_called_once_with(project_id=PROJECT_ID, dataset_id=DATASET_ID) mock_create.assert_called_once_with(table_resource=table_resource, project_id=PROJECT_ID) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.update_table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_dataset_tables") def test_table_upsert_already_exists(self, mock_get, mock_update): table_resource = {"tableReference": {"tableId": TABLE_ID}} mock_get.return_value = [{"tableId": TABLE_ID}] self.hook.run_table_upsert(dataset_id=DATASET_ID, table_resource=table_resource) mock_get.assert_called_once_with(project_id=PROJECT_ID, dataset_id=DATASET_ID) mock_update.assert_called_once_with(table_resource=table_resource) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.update_dataset") def test_run_grant_dataset_view_access_granting(self, mock_update, mock_get): view_table = f"{TABLE_ID}_view" view_dataset = f"{DATASET_ID}_view" view_access = AccessEntry( role=None, entity_type="view", entity_id={'projectId': PROJECT_ID, 'datasetId': view_dataset, 'tableId': view_table}, ) dataset = Dataset(DatasetReference.from_string(DATASET_ID, PROJECT_ID)) dataset.access_entries = [] mock_get.return_value = dataset self.hook.run_grant_dataset_view_access( source_dataset=DATASET_ID, view_dataset=view_dataset, view_table=view_table ) mock_get.assert_called_once_with(project_id=PROJECT_ID, dataset_id=DATASET_ID) assert view_access in dataset.access_entries mock_update.assert_called_once_with( fields=["access"], dataset_resource=dataset.to_api_repr(), project_id=PROJECT_ID, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.update_dataset") def test_run_grant_dataset_view_access_already_granted(self, mock_update, mock_get): view_table = f"{TABLE_ID}_view" view_dataset = f"{DATASET_ID}_view" view_access = AccessEntry( role=None, entity_type="view", entity_id={'projectId': PROJECT_ID, 'datasetId': view_dataset, 'tableId': view_table}, ) dataset = Dataset(DatasetReference.from_string(DATASET_ID, PROJECT_ID)) dataset.access_entries = [view_access] mock_get.return_value = dataset self.hook.run_grant_dataset_view_access( source_dataset=DATASET_ID, view_dataset=view_dataset, view_table=view_table ) mock_get.assert_called_once_with(project_id=PROJECT_ID, dataset_id=DATASET_ID) assert len(mock_update.calls) == 0 @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_get_dataset_tables_list(self, mock_client): table_list = [ {"projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": "a-1"}, {"projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": "b-1"}, {"projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": "a-2"}, {"projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": "b-2"}, ] table_list_response = [Table.from_api_repr({"tableReference": t}) for t in table_list] mock_client.return_value.list_tables.return_value = table_list_response dataset_reference = DatasetReference(PROJECT_ID, DATASET_ID) result = self.hook.get_dataset_tables_list(dataset_id=DATASET_ID, project_id=PROJECT_ID) mock_client.return_value.list_tables.assert_called_once_with( dataset=dataset_reference, max_results=None ) assert table_list == result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_poll_job_complete(self, mock_client): self.hook.poll_job_complete(job_id=JOB_ID, location=LOCATION, project_id=PROJECT_ID) mock_client.assert_called_once_with(location=LOCATION, project_id=PROJECT_ID) mock_client.return_value.get_job.assert_called_once_with(job_id=JOB_ID) mock_client.return_value.get_job.return_value.done.assert_called_once_with(retry=DEFAULT_RETRY) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.poll_job_complete") @mock.patch("logging.Logger.info") def test_cancel_query_jobs_to_cancel( self, mock_logger_info, poll_job_complete, ): poll_job_complete.return_value = True self.hook.running_job_id = JOB_ID self.hook.cancel_query() poll_job_complete.assert_called_once_with(job_id=JOB_ID) mock_logger_info.has_call(mock.call("No running BigQuery jobs to cancel.")) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.poll_job_complete") @mock.patch("time.sleep") @mock.patch("logging.Logger.info") def test_cancel_query_cancel_timeout( self, mock_logger_info, mock_sleep, poll_job_complete, mock_client, ): poll_job_complete.side_effect = [False] * 13 self.hook.running_job_id = JOB_ID self.hook.cancel_query() mock_client.return_value.cancel_job.assert_called_once_with(job_id=JOB_ID) assert poll_job_complete.call_count == 13 assert mock_sleep.call_count == 11 mock_logger_info.has_call( mock.call( f"Stopping polling due to timeout. Job with id {JOB_ID} " "has not completed cancel and may or may not finish." ) ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.poll_job_complete") @mock.patch("time.sleep") @mock.patch("logging.Logger.info") def test_cancel_query_cancel_completed( self, mock_logger_info, mock_sleep, poll_job_complete, mock_client, ): poll_job_complete.side_effect = [False] * 12 + [True] self.hook.running_job_id = JOB_ID self.hook.cancel_query() mock_client.return_value.cancel_job.assert_called_once_with(job_id=JOB_ID) assert poll_job_complete.call_count == 13 assert mock_sleep.call_count == 11 mock_logger_info.has_call(mock.call(f"Job successfully canceled: {PROJECT_ID}, {PROJECT_ID}")) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_get_schema(self, mock_client): table = { "tableReference": TABLE_REFERENCE_REPR, "schema": { "fields": [ {'name': 'id', 'type': 'STRING', 'mode': 'REQUIRED'}, {'name': 'name', 'type': 'STRING', 'mode': 'NULLABLE'}, ] }, } mock_client.return_value.get_table.return_value = Table.from_api_repr(table) result = self.hook.get_schema(dataset_id=DATASET_ID, table_id=TABLE_ID) mock_client.return_value.get_table.assert_called_once_with(TABLE_REFERENCE) assert "fields" in result assert len(result["fields"]) == 2 @mock.patch('airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_schema') @mock.patch('airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.update_table') def test_update_table_schema_with_policy_tags(self, mock_update, mock_get_schema): mock_get_schema.return_value = { "fields": [ {'name': 'emp_name', 'type': 'STRING', 'mode': 'REQUIRED'}, { 'name': 'salary', 'type': 'INTEGER', 'mode': 'REQUIRED', 'policyTags': {'names': ['sensitive']}, }, {'name': 'not_changed', 'type': 'INTEGER', 'mode': 'REQUIRED'}, { 'name': 'subrecord', 'type': 'RECORD', 'mode': 'REQUIRED', 'fields': [ { 'name': 'field_1', 'type': 'STRING', 'mode': 'REQUIRED', 'policyTags': {'names': ['sensitive']}, }, ], }, ] } schema_fields_updates = [ {'name': 'emp_name', 'description': 'Name of employee', 'policyTags': {'names': ['sensitive']}}, { 'name': 'salary', 'description': 'Monthly salary in USD', 'policyTags': {}, }, { 'name': 'subrecord', 'description': 'Some Desc', 'fields': [ {'name': 'field_1', 'description': 'Some nested desc'}, ], }, ] expected_result_schema = { 'fields': [ { 'name': 'emp_name', 'type': 'STRING', 'mode': 'REQUIRED', 'description': 'Name of employee', 'policyTags': {'names': ['sensitive']}, }, { 'name': 'salary', 'type': 'INTEGER', 'mode': 'REQUIRED', 'description': 'Monthly salary in USD', 'policyTags': {}, }, {'name': 'not_changed', 'type': 'INTEGER', 'mode': 'REQUIRED'}, { 'name': 'subrecord', 'type': 'RECORD', 'mode': 'REQUIRED', 'description': 'Some Desc', 'fields': [ { 'name': 'field_1', 'type': 'STRING', 'mode': 'REQUIRED', 'description': 'Some nested desc', 'policyTags': {'names': ['sensitive']}, } ], }, ] } self.hook.update_table_schema( schema_fields_updates=schema_fields_updates, include_policy_tags=True, dataset_id=DATASET_ID, table_id=TABLE_ID, ) mock_update.assert_called_once_with( dataset_id=DATASET_ID, table_id=TABLE_ID, project_id=PROJECT_ID, table_resource={'schema': expected_result_schema}, fields=['schema'], ) @mock.patch('airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_schema') @mock.patch('airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.update_table') def test_update_table_schema_without_policy_tags(self, mock_update, mock_get_schema): mock_get_schema.return_value = { "fields": [ {'name': 'emp_name', 'type': 'STRING', 'mode': 'REQUIRED'}, {'name': 'salary', 'type': 'INTEGER', 'mode': 'REQUIRED'}, {'name': 'not_changed', 'type': 'INTEGER', 'mode': 'REQUIRED'}, { 'name': 'subrecord', 'type': 'RECORD', 'mode': 'REQUIRED', 'fields': [ {'name': 'field_1', 'type': 'STRING', 'mode': 'REQUIRED'}, ], }, ] } schema_fields_updates = [ {'name': 'emp_name', 'description': 'Name of employee'}, { 'name': 'salary', 'description': 'Monthly salary in USD', 'policyTags': {'names': ['sensitive']}, }, { 'name': 'subrecord', 'description': 'Some Desc', 'fields': [ {'name': 'field_1', 'description': 'Some nested desc'}, ], }, ] expected_result_schema = { 'fields': [ {'name': 'emp_name', 'type': 'STRING', 'mode': 'REQUIRED', 'description': 'Name of employee'}, { 'name': 'salary', 'type': 'INTEGER', 'mode': 'REQUIRED', 'description': 'Monthly salary in USD', }, {'name': 'not_changed', 'type': 'INTEGER', 'mode': 'REQUIRED'}, { 'name': 'subrecord', 'type': 'RECORD', 'mode': 'REQUIRED', 'description': 'Some Desc', 'fields': [ { 'name': 'field_1', 'type': 'STRING', 'mode': 'REQUIRED', 'description': 'Some nested desc', } ], }, ] } self.hook.update_table_schema( schema_fields_updates=schema_fields_updates, include_policy_tags=False, dataset_id=DATASET_ID, table_id=TABLE_ID, ) mock_update.assert_called_once_with( dataset_id=DATASET_ID, table_id=TABLE_ID, project_id=PROJECT_ID, table_resource={'schema': expected_result_schema}, fields=['schema'], ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_invalid_source_format(self, mock_get_service): with pytest.raises( Exception, match=r"JSON is not a valid source format. Please use one of the following types: \['CSV', " r"'NEWLINE_DELIMITED_JSON', 'AVRO', 'GOOGLE_SHEETS', 'DATASTORE_BACKUP', 'PARQUET'\]", ): self.hook.run_load("test.test", "test_schema.json", ["test_data.json"], source_format="json") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_insert_all_succeed(self, mock_client): rows = [{"json": {"a_key": "a_value_0"}}] self.hook.insert_all( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, rows=rows, ignore_unknown_values=True, skip_invalid_rows=True, ) mock_client.return_value.get_table.assert_called_once_with(TABLE_REFERENCE) mock_client.return_value.insert_rows.assert_called_once_with( table=mock_client.return_value.get_table.return_value, rows=rows, ignore_unknown_values=True, skip_invalid_rows=True, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_insert_all_fail(self, mock_client): rows = [{"json": {"a_key": "a_value_0"}}] mock_client.return_value.insert_rows.return_value = ["some", "errors"] with pytest.raises(AirflowException, match="insert error"): self.hook.insert_all( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, rows=rows, fail_on_error=True ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_with_arg(self, mock_insert): self.hook.run_query( sql='select 1', destination_dataset_table='my_dataset.my_table', labels={'label1': 'test1', 'label2': 'test2'}, ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['labels'] == {'label1': 'test1', 'label2': 'test2'} @mock.patch("airflow.providers.google.cloud.hooks.bigquery.QueryJob") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_insert_job(self, mock_client, mock_query_job): job_conf = { "query": { "query": "SELECT * FROM test", "useLegacySql": "False", } } mock_query_job._JOB_TYPE = "query" self.hook.insert_job( configuration=job_conf, job_id=JOB_ID, project_id=PROJECT_ID, location=LOCATION, ) mock_client.assert_called_once_with( project_id=PROJECT_ID, location=LOCATION, ) mock_query_job.from_api_repr.assert_called_once_with( { 'configuration': job_conf, 'jobReference': {'jobId': JOB_ID, 'projectId': PROJECT_ID, 'location': LOCATION}, }, mock_client.return_value, ) mock_query_job.from_api_repr.return_value.result.assert_called_once_with() class TestBigQueryTableSplitter(unittest.TestCase): def test_internal_need_default_project(self): with pytest.raises(Exception, match="INTERNAL: No default project is specified"): _split_tablename("dataset.table", None) @parameterized.expand( [ ("project", "dataset", "table", "dataset.table"), ("alternative", "dataset", "table", "alternative:dataset.table"), ("alternative", "dataset", "table", "alternative.dataset.table"), ("alt1:alt", "dataset", "table", "alt1:alt.dataset.table"), ("alt1:alt", "dataset", "table", "alt1:alt:dataset.table"), ] ) def test_split_tablename(self, project_expected, dataset_expected, table_expected, table_input): default_project_id = "project" project, dataset, table = _split_tablename(table_input, default_project_id) assert project_expected == project assert dataset_expected == dataset assert table_expected == table @parameterized.expand( [ ("alt1:alt2:alt3:dataset.table", None, "Use either : or . to specify project got {}"), ( "alt1.alt.dataset.table", None, r"Expect format of \(<project\.\|<project\:\)<dataset>\.<table>, got {}", ), ( "alt1:alt2:alt.dataset.table", "var_x", "Format exception for var_x: Use either : or . to specify project got {}", ), ( "alt1:alt2:alt:dataset.table", "var_x", "Format exception for var_x: Use either : or . to specify project got {}", ), ( "alt1.alt.dataset.table", "var_x", r"Format exception for var_x: Expect format of " r"\(<project\.\|<project:\)<dataset>.<table>, got {}", ), ] ) def test_invalid_syntax(self, table_input, var_name, exception_message): default_project_id = "project" with pytest.raises(Exception, match=exception_message.format(table_input)): _split_tablename(table_input, default_project_id, var_name) class TestTableOperations(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_view(self, mock_bq_client, mock_table): view = { 'query': 'SELECT * FROM `test-project-id.test_dataset_id.test_table_prefix*`', "useLegacySql": False, } self.hook.create_empty_table( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, view=view, retry=DEFAULT_RETRY ) body = {'tableReference': TABLE_REFERENCE_REPR, 'view': view} mock_table.from_api_repr.assert_called_once_with(body) mock_bq_client.return_value.create_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, exists_ok=True, retry=DEFAULT_RETRY, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_patch_table(self, mock_client, mock_table): description_patched = 'Test description.' expiration_time_patched = 2524608000000 friendly_name_patched = 'Test friendly name.' labels_patched = {'label1': 'test1', 'label2': 'test2'} schema_patched = [ {'name': 'id', 'type': 'STRING', 'mode': 'REQUIRED'}, {'name': 'name', 'type': 'STRING', 'mode': 'NULLABLE'}, {'name': 'balance', 'type': 'FLOAT', 'mode': 'NULLABLE'}, {'name': 'new_field', 'type': 'STRING', 'mode': 'NULLABLE'}, ] time_partitioning_patched = {'expirationMs': 10000000} require_partition_filter_patched = True view_patched = { 'query': "SELECT * FROM `test-project-id.test_dataset_id.test_table_prefix*` LIMIT 500", 'useLegacySql': False, } self.hook.patch_table( dataset_id=DATASET_ID, table_id=TABLE_ID, project_id=PROJECT_ID, description=description_patched, expiration_time=expiration_time_patched, friendly_name=friendly_name_patched, labels=labels_patched, schema=schema_patched, time_partitioning=time_partitioning_patched, require_partition_filter=require_partition_filter_patched, view=view_patched, ) body = { "description": description_patched, "expirationTime": expiration_time_patched, "friendlyName": friendly_name_patched, "labels": labels_patched, "schema": {"fields": schema_patched}, "timePartitioning": time_partitioning_patched, "view": view_patched, "requirePartitionFilter": require_partition_filter_patched, } fields = list(body.keys()) body["tableReference"] = TABLE_REFERENCE_REPR mock_table.from_api_repr.assert_called_once_with(body) mock_client.return_value.update_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, fields=fields ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_table_succeed(self, mock_bq_client, mock_table): self.hook.create_empty_table(project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID) body = { 'tableReference': { 'tableId': TABLE_ID, 'projectId': PROJECT_ID, 'datasetId': DATASET_ID, } } mock_table.from_api_repr.assert_called_once_with(body) mock_bq_client.return_value.create_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, exists_ok=True, retry=DEFAULT_RETRY ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_table_with_extras_succeed(self, mock_bq_client, mock_table): schema_fields = [ {'name': 'id', 'type': 'STRING', 'mode': 'REQUIRED'}, {'name': 'name', 'type': 'STRING', 'mode': 'NULLABLE'}, {'name': 'created', 'type': 'DATE', 'mode': 'REQUIRED'}, ] time_partitioning = {"field": "created", "type": "DAY"} cluster_fields = ['name'] self.hook.create_empty_table( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, schema_fields=schema_fields, time_partitioning=time_partitioning, cluster_fields=cluster_fields, ) body = { 'tableReference': { 'tableId': TABLE_ID, 'projectId': PROJECT_ID, 'datasetId': DATASET_ID, }, 'schema': {'fields': schema_fields}, 'timePartitioning': time_partitioning, 'clustering': {'fields': cluster_fields}, } mock_table.from_api_repr.assert_called_once_with(body) mock_bq_client.return_value.create_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, exists_ok=True, retry=DEFAULT_RETRY ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_get_tables_list(self, mock_client): table_list = [ { "kind": "bigquery#table", "id": "your-project:your_dataset.table1", "tableReference": { "projectId": "your-project", "datasetId": "your_dataset", "tableId": "table1", }, "type": "TABLE", "creationTime": "1565781859261", }, { "kind": "bigquery#table", "id": "your-project:your_dataset.table2", "tableReference": { "projectId": "your-project", "datasetId": "your_dataset", "tableId": "table2", }, "type": "TABLE", "creationTime": "1565782713480", }, ] table_list_response = [Table.from_api_repr(t) for t in table_list] mock_client.return_value.list_tables.return_value = table_list_response dataset_reference = DatasetReference(PROJECT_ID, DATASET_ID) result = self.hook.get_dataset_tables(dataset_id=DATASET_ID, project_id=PROJECT_ID) mock_client.return_value.list_tables.assert_called_once_with( dataset=dataset_reference, max_results=None, retry=DEFAULT_RETRY, ) for res, exp in zip(result, table_list): assert res["tableId"] == exp["tableReference"]["tableId"] @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_materialized_view(self, mock_bq_client, mock_table): query = """ SELECT product, SUM(amount) FROM `test-project-id.test_dataset_id.test_table_prefix*` GROUP BY product """ materialized_view = { 'query': query, 'enableRefresh': True, 'refreshIntervalMs': 2000000, } self.hook.create_empty_table( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, materialized_view=materialized_view, retry=DEFAULT_RETRY, ) body = {'tableReference': TABLE_REFERENCE_REPR, 'materializedView': materialized_view} mock_table.from_api_repr.assert_called_once_with(body) mock_bq_client.return_value.create_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, exists_ok=True, retry=DEFAULT_RETRY, ) class TestBigQueryCursor(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_execute_with_parameters(self, mock_insert, _): bq_cursor = self.hook.get_cursor() bq_cursor.execute("SELECT %(foo)s", {"foo": "bar"}) conf = { 'query': { 'query': "SELECT 'bar'", 'priority': 'INTERACTIVE', 'useLegacySql': True, 'schemaUpdateOptions': [], } } mock_insert.assert_called_once_with(configuration=conf, project_id=PROJECT_ID) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_execute_many(self, mock_insert, _): bq_cursor = self.hook.get_cursor() bq_cursor.executemany("SELECT %(foo)s", [{"foo": "bar"}, {"foo": "baz"}]) assert mock_insert.call_count == 2 assert mock_insert.has_calls( mock.call( configuration={ 'query': { 'query': "SELECT 'bar'", 'priority': 'INTERACTIVE', 'useLegacySql': True, 'schemaUpdateOptions': [], } }, project_id=PROJECT_ID, ), mock.call( configuration={ 'query': { 'query': "SELECT 'baz'", 'priority': 'INTERACTIVE', 'useLegacySql': True, 'schemaUpdateOptions': [], } }, project_id=PROJECT_ID, ), ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_description(self, mock_get_service): bq_cursor = self.hook.get_cursor() with pytest.raises(NotImplementedError): bq_cursor.description @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_close(self, mock_get_service): bq_cursor = self.hook.get_cursor() result = bq_cursor.close() # pylint: disable=assignment-from-no-return assert result is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_rowcount(self, mock_get_service): bq_cursor = self.hook.get_cursor() result = bq_cursor.rowcount assert -1 == result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryCursor.next") def test_fetchone(self, mock_next, mock_get_service): bq_cursor = self.hook.get_cursor() result = bq_cursor.fetchone() mock_next.call_count == 1 assert mock_next.return_value == result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch( "airflow.providers.google.cloud.hooks.bigquery.BigQueryCursor.fetchone", side_effect=[1, 2, 3, None] ) def test_fetchall(self, mock_fetchone, mock_get_service): bq_cursor = self.hook.get_cursor() result = bq_cursor.fetchall() assert [1, 2, 3] == result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryCursor.fetchone") def test_fetchmany(self, mock_fetchone, mock_get_service): side_effect_values = [1, 2, 3, None] bq_cursor = self.hook.get_cursor() mock_fetchone.side_effect = side_effect_values result = bq_cursor.fetchmany() assert [1] == result mock_fetchone.side_effect = side_effect_values result = bq_cursor.fetchmany(2) assert [1, 2] == result mock_fetchone.side_effect = side_effect_values result = bq_cursor.fetchmany(5) assert [1, 2, 3] == result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_next_no_jobid(self, mock_get_service): bq_cursor = self.hook.get_cursor() bq_cursor.job_id = None result = bq_cursor.next() assert result is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_next_buffer(self, mock_get_service): bq_cursor = self.hook.get_cursor() bq_cursor.job_id = JOB_ID bq_cursor.buffer = [1, 2] result = bq_cursor.next() assert 1 == result result = bq_cursor.next() assert 2 == result bq_cursor.all_pages_loaded = True result = bq_cursor.next() assert result is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_next(self, mock_get_service): mock_get_query_results = mock_get_service.return_value.jobs.return_value.getQueryResults mock_execute = mock_get_query_results.return_value.execute mock_execute.return_value = { "rows": [ {"f": [{"v": "one"}, {"v": 1}]}, {"f": [{"v": "two"}, {"v": 2}]}, ], "pageToken": None, "schema": { "fields": [ {"name": "field_1", "type": "STRING"}, {"name": "field_2", "type": "INTEGER"}, ] }, } bq_cursor = self.hook.get_cursor() bq_cursor.job_id = JOB_ID bq_cursor.location = LOCATION result = bq_cursor.next() assert ['one', 1] == result result = bq_cursor.next() assert ['two', 2] == result mock_get_query_results.assert_called_once_with( jobId=JOB_ID, location=LOCATION, pageToken=None, projectId='bq-project' ) mock_execute.assert_called_once_with(num_retries=bq_cursor.num_retries) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryCursor.flush_results") def test_next_no_rows(self, mock_flush_results, mock_get_service): mock_get_query_results = mock_get_service.return_value.jobs.return_value.getQueryResults mock_execute = mock_get_query_results.return_value.execute mock_execute.return_value = {} bq_cursor = self.hook.get_cursor() bq_cursor.job_id = JOB_ID result = bq_cursor.next() assert result is None mock_get_query_results.assert_called_once_with( jobId=JOB_ID, location=None, pageToken=None, projectId='bq-project' ) mock_execute.assert_called_once_with(num_retries=bq_cursor.num_retries) assert mock_flush_results.call_count == 1 @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryCursor.flush_results") def test_flush_cursor_in_execute(self, _, mock_insert, mock_get_service): bq_cursor = self.hook.get_cursor() bq_cursor.execute("SELECT %(foo)s", {"foo": "bar"}) assert mock_insert.call_count == 1 @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_flush_cursor(self, mock_get_service): bq_cursor = self.hook.get_cursor() bq_cursor.page_token = '456dcea9-fcbf-4f02-b570-83f5297c685e' bq_cursor.job_id = 'c0a79ae4-0e72-4593-a0d0-7dbbf726f193' bq_cursor.all_pages_loaded = True bq_cursor.buffer = [('a', 100, 200), ('b', 200, 300)] bq_cursor.flush_results() assert bq_cursor.page_token is None assert bq_cursor.job_id is None assert not bq_cursor.all_pages_loaded assert bq_cursor.buffer == [] @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_arraysize(self, mock_get_service): bq_cursor = self.hook.get_cursor() assert bq_cursor.buffersize is None assert bq_cursor.arraysize == 1 bq_cursor.set_arraysize(10) assert bq_cursor.buffersize == 10 assert bq_cursor.arraysize == 10 class TestDatasetsOperations(_BigQueryBaseTestClass): def test_create_empty_dataset_no_dataset_id_err(self): with pytest.raises(ValueError, match=r"Please specify `datasetId`"): self.hook.create_empty_dataset(dataset_id=None, project_id=None) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_dataset_with_params(self, mock_client, mock_dataset): self.hook.create_empty_dataset(project_id=PROJECT_ID, dataset_id=DATASET_ID, location=LOCATION) expected_body = { "location": LOCATION, "datasetReference": {"datasetId": DATASET_ID, "projectId": PROJECT_ID}, } api_repr = mock_dataset.from_api_repr api_repr.assert_called_once_with(expected_body) mock_client.return_value.create_dataset.assert_called_once_with( dataset=api_repr.return_value, exists_ok=True ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_dataset_with_object(self, mock_client, mock_dataset): dataset = { "location": "LOCATION", "datasetReference": {"datasetId": "DATASET_ID", "projectId": "PROJECT_ID"}, } self.hook.create_empty_dataset(dataset_reference=dataset) api_repr = mock_dataset.from_api_repr api_repr.assert_called_once_with(dataset) mock_client.return_value.create_dataset.assert_called_once_with( dataset=api_repr.return_value, exists_ok=True ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_dataset_use_values_from_object(self, mock_client, mock_dataset): dataset = { "location": "LOCATION", "datasetReference": {"datasetId": "DATASET_ID", "projectId": "PROJECT_ID"}, } self.hook.create_empty_dataset( dataset_reference=dataset, location="Unknown location", dataset_id="Fashionable Dataset", project_id="Amazing Project", ) api_repr = mock_dataset.from_api_repr api_repr.assert_called_once_with(dataset) mock_client.return_value.create_dataset.assert_called_once_with( dataset=api_repr.return_value, exists_ok=True ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_get_dataset(self, mock_client): _expected_result = { "kind": "bigquery#dataset", "location": "US", "id": "your-project:dataset_2_test", "datasetReference": {"projectId": "your-project", "datasetId": "dataset_2_test"}, } expected_result = Dataset.from_api_repr(_expected_result) mock_client.return_value.get_dataset.return_value = expected_result result = self.hook.get_dataset(dataset_id=DATASET_ID, project_id=PROJECT_ID) mock_client.return_value.get_dataset.assert_called_once_with( dataset_ref=DatasetReference(PROJECT_ID, DATASET_ID) ) assert result == expected_result @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_get_datasets_list(self, mock_client): datasets = [ { "kind": "bigquery#dataset", "location": "US", "id": "your-project:dataset_2_test", "datasetReference": {"projectId": "your-project", "datasetId": "dataset_2_test"}, }, { "kind": "bigquery#dataset", "location": "US", "id": "your-project:dataset_1_test", "datasetReference": {"projectId": "your-project", "datasetId": "dataset_1_test"}, }, ] return_value = [DatasetListItem(d) for d in datasets] mock_client.return_value.list_datasets.return_value = return_value result = self.hook.get_datasets_list(project_id=PROJECT_ID) mock_client.return_value.list_datasets.assert_called_once_with( project=PROJECT_ID, include_all=False, filter=None, max_results=None, page_token=None, retry=DEFAULT_RETRY, ) for exp, res in zip(datasets, result): assert res.full_dataset_id == exp["id"] @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_delete_dataset(self, mock_client): delete_contents = True self.hook.delete_dataset( project_id=PROJECT_ID, dataset_id=DATASET_ID, delete_contents=delete_contents ) mock_client.return_value.delete_dataset.assert_called_once_with( dataset=DatasetReference(PROJECT_ID, DATASET_ID), delete_contents=delete_contents, retry=DEFAULT_RETRY, not_found_ok=True, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") def test_patch_dataset(self, mock_get_service): dataset_resource = {"access": [{"role": "WRITER", "groupByEmail": "cloud-logs@google.com"}]} method = mock_get_service.return_value.datasets.return_value.patch self.hook.patch_dataset( dataset_id=DATASET_ID, project_id=PROJECT_ID, dataset_resource=dataset_resource ) method.assert_called_once_with(projectId=PROJECT_ID, datasetId=DATASET_ID, body=dataset_resource) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Dataset") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_update_dataset(self, mock_client, mock_dataset): dataset_resource = { "kind": "bigquery#dataset", "location": "US", "id": "your-project:dataset_2_test", "datasetReference": {"projectId": "your-project", "datasetId": "dataset_2_test"}, } method = mock_client.return_value.update_dataset dataset = Dataset.from_api_repr(dataset_resource) mock_dataset.from_api_repr.return_value = dataset method.return_value = dataset result = self.hook.update_dataset( dataset_id=DATASET_ID, project_id=PROJECT_ID, dataset_resource=dataset_resource, fields=["location"], ) mock_dataset.from_api_repr.assert_called_once_with(dataset_resource) method.assert_called_once_with( dataset=dataset, fields=["location"], retry=DEFAULT_RETRY, ) assert result == dataset class TestTimePartitioningInRunJob(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_default(self, mock_insert): self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load'].get('timePartitioning') is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_with_auto_detect(self, mock_insert): destination_project_dataset_table = "autodetect.table" self.hook.run_load(destination_project_dataset_table, [], [], autodetect=True) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load']['autodetect'] is True @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_with_arg(self, mock_insert): self.hook.run_load( destination_project_dataset_table=f"{DATASET_ID}.{TABLE_ID}", schema_fields=[], source_uris=[], time_partitioning={'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000}, ) configuration = { 'load': { 'autodetect': False, 'createDisposition': 'CREATE_IF_NEEDED', 'destinationTable': {'projectId': PROJECT_ID, 'datasetId': DATASET_ID, 'tableId': TABLE_ID}, 'sourceFormat': 'CSV', 'sourceUris': [], 'writeDisposition': 'WRITE_EMPTY', 'ignoreUnknownValues': False, 'timePartitioning': {'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000}, 'skipLeadingRows': 0, 'fieldDelimiter': ',', 'quote': None, 'allowQuotedNewlines': False, 'encoding': 'UTF-8', } } mock_insert.assert_called_once_with(configuration=configuration, project_id=PROJECT_ID) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_with_arg(self, mock_insert): self.hook.run_query( sql='select 1', destination_dataset_table=f"{DATASET_ID}.{TABLE_ID}", time_partitioning={'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000}, ) configuration = { 'query': { 'query': 'select 1', 'priority': 'INTERACTIVE', 'useLegacySql': True, 'timePartitioning': {'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000}, 'schemaUpdateOptions': [], 'destinationTable': {'projectId': PROJECT_ID, 'datasetId': DATASET_ID, 'tableId': TABLE_ID}, 'allowLargeResults': False, 'flattenResults': None, 'writeDisposition': 'WRITE_EMPTY', 'createDisposition': 'CREATE_IF_NEEDED', } } mock_insert.assert_called_once_with(configuration=configuration, project_id=PROJECT_ID) def test_dollar_makes_partition(self): tp_out = _cleanse_time_partitioning('test.teast$20170101', {}) expect = {'type': 'DAY'} assert tp_out == expect def test_extra_time_partitioning_options(self): tp_out = _cleanse_time_partitioning( 'test.teast', {'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000} ) expect = {'type': 'DAY', 'field': 'test_field', 'expirationMs': 1000} assert tp_out == expect class TestClusteringInRunJob(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_default(self, mock_insert): self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load'].get('clustering') is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_with_arg(self, mock_insert): self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], cluster_fields=['field1', 'field2'], time_partitioning={'type': 'DAY'}, ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load']['clustering'] == {'fields': ['field1', 'field2']} @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_default(self, mock_insert): self.hook.run_query(sql='select 1') _, kwargs = mock_insert.call_args assert kwargs["configuration"]['query'].get('clustering') is None @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_with_arg(self, mock_insert): self.hook.run_query( sql='select 1', destination_dataset_table='my_dataset.my_table', cluster_fields=['field1', 'field2'], time_partitioning={'type': 'DAY'}, ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['query']['clustering'] == {'fields': ['field1', 'field2']} class TestBigQueryHookLegacySql(_BigQueryBaseTestClass): """Ensure `use_legacy_sql` param in `BigQueryHook` propagates properly.""" @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_hook_uses_legacy_sql_by_default(self, mock_insert, _): self.hook.get_first('query') _, kwargs = mock_insert.call_args assert kwargs["configuration"]['query']['useLegacySql'] is True @mock.patch( 'airflow.providers.google.common.hooks.base_google.GoogleBaseHook._get_credentials_and_project_id', return_value=(CREDENTIALS, PROJECT_ID), ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_service") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_legacy_sql_override_propagates_properly( self, mock_insert, mock_get_service, mock_get_creds_and_proj_id ): bq_hook = BigQueryHook(use_legacy_sql=False) bq_hook.get_first('query') _, kwargs = mock_insert.call_args assert kwargs["configuration"]['query']['useLegacySql'] is False class TestBigQueryHookRunWithConfiguration(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.LoadJob") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.get_client") def test_run_with_configuration_location(self, mock_client, mock_job): running_job_id = 'job_vjdi28vskdui2onru23' location = 'asia-east1' mock_job._JOB_TYPE = "load" conf = {"load": {}} self.hook.running_job_id = running_job_id self.hook.location = location self.hook.run_with_configuration(conf) mock_client.assert_called_once_with(project_id=PROJECT_ID, location=location) mock_job.from_api_repr.assert_called_once_with( { "configuration": conf, "jobReference": {"jobId": mock.ANY, "projectId": PROJECT_ID, "location": location}, }, mock_client.return_value, ) class TestBigQueryWithKMS(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_create_empty_table_with_kms(self, mock_bq_client, mock_table): schema_fields = [{"name": "id", "type": "STRING", "mode": "REQUIRED"}] encryption_configuration = {"kms_key_name": "projects/p/locations/l/keyRings/k/cryptoKeys/c"} self.hook.create_empty_table( project_id=PROJECT_ID, dataset_id=DATASET_ID, table_id=TABLE_ID, schema_fields=schema_fields, encryption_configuration=encryption_configuration, ) body = { "tableReference": {"tableId": TABLE_ID, 'projectId': PROJECT_ID, 'datasetId': DATASET_ID}, "schema": {"fields": schema_fields}, "encryptionConfiguration": encryption_configuration, } mock_table.from_api_repr.assert_called_once_with(body) mock_bq_client.return_value.create_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, exists_ok=True, retry=DEFAULT_RETRY, ) # pylint: disable=too-many-locals @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.create_empty_table") def test_create_external_table_with_kms(self, mock_create): external_project_dataset_table = f"{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}" source_uris = ['test_data.csv'] source_format = 'CSV' autodetect = False compression = 'NONE' ignore_unknown_values = False max_bad_records = 10 skip_leading_rows = 1 field_delimiter = ',' quote_character = None allow_quoted_newlines = False allow_jagged_rows = False encoding = "UTF-8" labels = {'label1': 'test1', 'label2': 'test2'} schema_fields = [{'mode': 'REQUIRED', 'name': 'id', 'type': 'STRING', 'description': None}] encryption_configuration = {"kms_key_name": "projects/p/locations/l/keyRings/k/cryptoKeys/c"} self.hook.create_external_table( external_project_dataset_table=external_project_dataset_table, source_uris=source_uris, source_format=source_format, autodetect=autodetect, compression=compression, ignore_unknown_values=ignore_unknown_values, max_bad_records=max_bad_records, skip_leading_rows=skip_leading_rows, field_delimiter=field_delimiter, quote_character=quote_character, allow_jagged_rows=allow_jagged_rows, encoding=encoding, allow_quoted_newlines=allow_quoted_newlines, labels=labels, schema_fields=schema_fields, encryption_configuration=encryption_configuration, ) body = { 'externalDataConfiguration': { 'autodetect': autodetect, 'sourceFormat': source_format, 'sourceUris': source_uris, 'compression': compression, 'ignoreUnknownValues': ignore_unknown_values, 'schema': {'fields': schema_fields}, 'maxBadRecords': max_bad_records, 'csvOptions': { 'skipLeadingRows': skip_leading_rows, 'fieldDelimiter': field_delimiter, 'quote': quote_character, 'allowQuotedNewlines': allow_quoted_newlines, 'allowJaggedRows': allow_jagged_rows, 'encoding': encoding, }, }, 'tableReference': { 'projectId': PROJECT_ID, 'datasetId': DATASET_ID, 'tableId': TABLE_ID, }, 'labels': labels, "encryptionConfiguration": encryption_configuration, } mock_create.assert_called_once_with( table_resource=body, project_id=PROJECT_ID, location=None, exists_ok=True, ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Table") @mock.patch("airflow.providers.google.cloud.hooks.bigquery.Client") def test_update_table(self, mock_client, mock_table): description_patched = 'Test description.' expiration_time_patched = 2524608000000 friendly_name_patched = 'Test friendly name.' labels_patched = {'label1': 'test1', 'label2': 'test2'} schema_patched = [ {'name': 'id', 'type': 'STRING', 'mode': 'REQUIRED'}, {'name': 'name', 'type': 'STRING', 'mode': 'NULLABLE'}, {'name': 'balance', 'type': 'FLOAT', 'mode': 'NULLABLE'}, {'name': 'new_field', 'type': 'STRING', 'mode': 'NULLABLE'}, ] time_partitioning_patched = {'expirationMs': 10000000} require_partition_filter_patched = True view_patched = { 'query': "SELECT * FROM `test-project-id.test_dataset_id.test_table_prefix*` LIMIT 500", 'useLegacySql': False, } body = { "tableReference": { "projectId": PROJECT_ID, "datasetId": DATASET_ID, "tableId": TABLE_ID, }, "description": description_patched, "expirationTime": expiration_time_patched, "friendlyName": friendly_name_patched, "labels": labels_patched, "schema": {"fields": schema_patched}, "timePartitioning": time_partitioning_patched, "view": view_patched, "requirePartitionFilter": require_partition_filter_patched, } fields = list(body.keys()) self.hook.update_table( table_resource=body, fields=fields, dataset_id=DATASET_ID, table_id=TABLE_ID, project_id=PROJECT_ID, ) mock_table.from_api_repr.assert_called_once_with(body) mock_client.return_value.update_table.assert_called_once_with( table=mock_table.from_api_repr.return_value, fields=fields ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_query_with_kms(self, mock_insert): encryption_configuration = {"kms_key_name": "projects/p/locations/l/keyRings/k/cryptoKeys/c"} self.hook.run_query(sql='query', encryption_configuration=encryption_configuration) _, kwargs = mock_insert.call_args assert ( kwargs["configuration"]['query']['destinationEncryptionConfiguration'] is encryption_configuration ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_copy_with_kms(self, mock_insert): encryption_configuration = {"kms_key_name": "projects/p/locations/l/keyRings/k/cryptoKeys/c"} self.hook.run_copy( source_project_dataset_tables='p.d.st', destination_project_dataset_table='p.d.dt', encryption_configuration=encryption_configuration, ) _, kwargs = mock_insert.call_args assert ( kwargs["configuration"]['copy']['destinationEncryptionConfiguration'] is encryption_configuration ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_with_kms(self, mock_insert): encryption_configuration = {"kms_key_name": "projects/p/locations/l/keyRings/k/cryptoKeys/c"} self.hook.run_load( destination_project_dataset_table='p.d.dt', source_uris=['abc.csv'], autodetect=True, encryption_configuration=encryption_configuration, ) _, kwargs = mock_insert.call_args assert ( kwargs["configuration"]['load']['destinationEncryptionConfiguration'] is encryption_configuration ) class TestBigQueryBaseCursorMethodsDeprecationWarning(unittest.TestCase): @parameterized.expand( [ ("create_empty_table",), ("create_empty_dataset",), ("get_dataset_tables",), ("delete_dataset",), ("create_external_table",), ("patch_table",), ("insert_all",), ("update_dataset",), ("patch_dataset",), ("get_dataset_tables_list",), ("get_datasets_list",), ("get_dataset",), ("run_grant_dataset_view_access",), ("run_table_upsert",), ("run_table_delete",), ("get_tabledata",), ("get_schema",), ("poll_job_complete",), ("cancel_query",), ("run_with_configuration",), ("run_load",), ("run_copy",), ("run_extract",), ("run_query",), ] ) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook") def test_deprecation_warning(self, func_name, mock_bq_hook): args, kwargs = [1], {"param1": "val1"} new_path = re.escape(f"`airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.{func_name}`") message_pattern = fr"This method is deprecated\.\s+Please use {new_path}" message_regex = re.compile(message_pattern, re.MULTILINE) mocked_func = getattr(mock_bq_hook, func_name) bq_cursor = BigQueryCursor(mock.MagicMock(), PROJECT_ID, mock_bq_hook) func = getattr(bq_cursor, func_name) with pytest.warns(DeprecationWarning, match=message_regex): _ = func(*args, **kwargs) mocked_func.assert_called_once_with(*args, **kwargs) assert re.search(f".*{new_path}.*", func.__doc__) class TestBigQueryWithLabelsAndDescription(_BigQueryBaseTestClass): @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_labels(self, mock_insert): labels = {'label1': 'test1', 'label2': 'test2'} self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], labels=labels, ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load']['destinationTableProperties']['labels'] is labels @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.insert_job") def test_run_load_description(self, mock_insert): description = "Test Description" self.hook.run_load( destination_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], description=description, ) _, kwargs = mock_insert.call_args assert kwargs["configuration"]['load']['destinationTableProperties']['description'] is description @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.create_empty_table") def test_create_external_table_labels(self, mock_create): labels = {'label1': 'test1', 'label2': 'test2'} self.hook.create_external_table( external_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], labels=labels, ) _, kwargs = mock_create.call_args self.assertDictEqual(kwargs['table_resource']['labels'], labels) @mock.patch("airflow.providers.google.cloud.hooks.bigquery.BigQueryHook.create_empty_table") def test_create_external_table_description(self, mock_create): description = "Test Description" self.hook.create_external_table( external_project_dataset_table='my_dataset.my_table', schema_fields=[], source_uris=[], description=description, ) _, kwargs = mock_create.call_args assert kwargs['table_resource']['description'] is description
tests/providers/google/cloud/hooks/test_bigquery.py
85,313
Ensure `use_legacy_sql` param in `BigQueryHook` propagates properly. Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. pylint: disable=not-callable TODO: this creates side effect This config should raise a value error. pylint: disable=assignment-from-no-return pylint: disable=too-many-locals
997
en
0.839319
# AUTOGENERATED BY NBDEV! DO NOT EDIT! __all__ = ["index", "modules", "custom_doc_links", "git_url"] index = {"Config": "00_learner.ipynb", "energy_score": "00_learner.ipynb", "EnsemblePredict": "00_learner.ipynb", "EnsembleLearner": "00_learner.ipynb", "ARCHITECTURES": "01_models.ipynb", "ENCODERS": "01_models.ipynb", "get_pretrained_options": "01_models.ipynb", "create_smp_model": "01_models.ipynb", "save_smp_model": "01_models.ipynb", "load_smp_model": "01_models.ipynb", "show": "02_data.ipynb", "preprocess_mask": "02_data.ipynb", "DeformationField": "02_data.ipynb", "BaseDataset": "02_data.ipynb", "RandomTileDataset": "02_data.ipynb", "TileDataset": "02_data.ipynb", "Dice": "03_metrics.ipynb", "Iou": "03_metrics.ipynb", "Recorder.plot_metrics": "03_metrics.ipynb", "LOSSES": "05_losses.ipynb", "FastaiLoss": "05_losses.ipynb", "WeightedLoss": "05_losses.ipynb", "JointLoss": "05_losses.ipynb", "get_loss": "05_losses.ipynb", "unzip": "06_utils.ipynb", "install_package": "06_utils.ipynb", "import_package": "06_utils.ipynb", "compose_albumentations": "06_utils.ipynb", "ensemble_results": "06_utils.ipynb", "plot_results": "06_utils.ipynb", "iou": "06_utils.ipynb", "label_mask": "06_utils.ipynb", "get_candidates": "06_utils.ipynb", "iou_mapping": "06_utils.ipynb", "calculate_roi_measures": "06_utils.ipynb", "export_roi_set": "06_utils.ipynb", "calc_iterations": "06_utils.ipynb", "get_label_fn": "06_utils.ipynb", "save_mask": "06_utils.ipynb", "save_unc": "06_utils.ipynb", "rot90": "07_tta.ipynb", "hflip": "07_tta.ipynb", "vflip": "07_tta.ipynb", "BaseTransform": "07_tta.ipynb", "Chain": "07_tta.ipynb", "Transformer": "07_tta.ipynb", "Compose": "07_tta.ipynb", "Merger": "07_tta.ipynb", "HorizontalFlip": "07_tta.ipynb", "VerticalFlip": "07_tta.ipynb", "Rotate90": "07_tta.ipynb", "GRID_COLS": "08_gui.ipynb", "set_css_in_cell_output": "08_gui.ipynb", "tooltip_css": "08_gui.ipynb", "ZipUpload": "08_gui.ipynb", "ItemsPerPage": "08_gui.ipynb", "BaseParamWidget": "08_gui.ipynb", "BaseUI": "08_gui.ipynb", "PathSelector": "08_gui.ipynb", "PathDownloads": "08_gui.ipynb", "PathConfig": "08_gui.ipynb", "GTDataSB": "08_gui.ipynb", "GTEstSB": "08_gui.ipynb", "GTEstUI": "08_gui.ipynb", "TrainDataSB": "08_gui.ipynb", "TrainModelSB": "08_gui.ipynb", "TrainValidSB": "08_gui.ipynb", "LRWidget": "08_gui.ipynb", "BasePopUpParamWidget": "08_gui.ipynb", "ParamWidget": "08_gui.ipynb", "MWWidget": "08_gui.ipynb", "TrainUI": "08_gui.ipynb", "PredInputSB": "08_gui.ipynb", "PredSB": "08_gui.ipynb", "PredUI": "08_gui.ipynb", "GUI": "08_gui.ipynb", "import_sitk": "09_gt.ipynb", "staple": "09_gt.ipynb", "m_voting": "09_gt.ipynb", "msk_show": "09_gt.ipynb", "GTEstimator": "09_gt.ipynb"} modules = ["learner.py", "models.py", "data.py", "metrics.py", "losses.py", "utils.py", "tta.py", "gui.py", "gt.py"] doc_url = "https://matjesg.github.io/deepflash2/" git_url = "https://github.com/matjesg/deepflash2/tree/master/" def custom_doc_links(name): return None
deepflash2/_nbdev.py
3,783
AUTOGENERATED BY NBDEV! DO NOT EDIT!
36
en
0.8261
from django.db import models from django.contrib.auth import get_user_model User = get_user_model() class Group(models.Model): title = models.CharField(max_length=200) slug = models.SlugField(unique=True) description = models.TextField() class Post(models.Model): text = models.TextField() pub_date = models.DateTimeField("date published", auto_now_add=True) author = models.ForeignKey(User, on_delete=models.CASCADE, related_name="post_author") group = models.ForeignKey(Group, on_delete=models.CASCADE, blank=True, null=True) image = models.ImageField(upload_to='posts/', blank=True, null=True) class Comment(models.Model): post = models.ForeignKey(Post, on_delete=models.CASCADE, related_name='comments') author = models.ForeignKey(User, on_delete=models.CASCADE, related_name='comments') text = models.TextField() created = models.DateTimeField('Дата и время публикации', auto_now_add=True, db_index=True) def __str__(self): return self.text class Follow(models.Model): user = models.ForeignKey(User, on_delete=models.CASCADE, related_name='follower') #тот который подписывается author = models.ForeignKey(User, on_delete=models.CASCADE, related_name='following') #тот на которого подписываются def __str__(self): return self.text
posts/models.py
1,397
тот который подписываетсятот на которого подписываются
54
ru
0.999288
import sys import numpy as np from numpy.lib import recfunctions as recFunc from ..frequency_domain.survey import Survey from ...data import Data as BaseData from ...utils import mkvc from .sources import Planewave_xy_1Dprimary, Planewave_xy_1DhomotD from .receivers import Point3DImpedance, Point3DTipper from .utils.plot_utils import DataNSEMPlotMethods ######### # Survey ######### # class Survey(BaseSurvey): # """ # Survey class for NSEM. # **Requried** # :param list srcList: List of sources associated with the survey # **Optional** # """ # srcPair = BaseNSEMSrc # def __init__(self, srcList, **kwargs): # # Sort these by frequency # self.source_list = srcList # BaseSurvey.__init__(self, **kwargs) # _freqDict = {} # for src in srcList: # if src.freq not in _freqDict: # _freqDict[src.freq] = [] # _freqDict[src.freq] += [src] # self._freqDict = _freqDict # self._freqs = sorted([f for f in self._freqDict]) # @property # def freqs(self): # """Frequencies""" # return self._freqs # @property # def nFreq(self): # """Number of frequencies""" # return len(self._freqDict) # def getSrcByFreq(self, freq): # """Returns the sources associated with a specific frequency.""" # assert freq in self._freqDict, "The requested frequency is not in this survey." # return self._freqDict[freq] # def eval(self, f): # """ # Evalute and return Data given calculated fields # :param SimPEG.electromagnetics.frequency_domain.fields.FieldsFDEM f: A NSEM fileds object to evaluate data from # :retype: SimPEG.EM.NSEM.Data # :return: NSEM Data object # """ # data = Data(self) # for src in self.source_list: # sys.stdout.flush() # for rx in src.receiver_list: # data[src, rx] = rx.eval(src, self.mesh, f) # return data # def evalDeriv(self, f): # raise Exception('Use Sources to project fields deriv.') ######### # Data ######### class Data(BaseData, DataNSEMPlotMethods): """ Data class for NSEMdata. Stores the data vector indexed by the survey. """ def __init__(self, survey, dobs=None, relative_error=None, noise_floor=None): BaseData.__init__(self, survey, dobs, relative_error, noise_floor) def toRecArray(self, returnType="RealImag"): """ Returns a numpy.recarray for a SimpegNSEM impedance data object. :param returnType: Switches between returning a rec array where the impedance is split to real and imaginary ('RealImag') or is a complex ('Complex') :type returnType: str, optional :rtype: numpy.recarray :return: Record array with data, with indexed columns """ # Define the record fields dtRI = [ ("freq", float), ("x", float), ("y", float), ("z", float), ("zxxr", float), ("zxxi", float), ("zxyr", float), ("zxyi", float), ("zyxr", float), ("zyxi", float), ("zyyr", float), ("zyyi", float), ("tzxr", float), ("tzxi", float), ("tzyr", float), ("tzyi", float), ] dtCP = [ ("freq", float), ("x", float), ("y", float), ("z", float), ("zxx", complex), ("zxy", complex), ("zyx", complex), ("zyy", complex), ("tzx", complex), ("tzy", complex), ] for src in self.survey.source_list: # Temp array for all the receivers of the source. # Note: needs to be written more generally, # using diffterent rxTypes and not all the data at the locations # Assume the same locs for all RX locs = src.receiver_list[0].locations if locs.shape[1] == 1: locs = np.hstack((np.array([[0.0, 0.0]]), locs)) elif locs.shape[1] == 2: locs = np.hstack((np.array([[0.0]]), locs)) tArrRec = np.concatenate( ( src.freq * np.ones((locs.shape[0], 1)), locs, np.nan * np.ones((locs.shape[0], 12)), ), axis=1, ).view(dtRI) # Get the type and the value for the DataNSEM object as a list typeList = [ [rx.orientation, rx.component, self[src, rx]] for rx in src.receiver_list ] # Insert the values to the temp array for nr, (k, c, val) in enumerate(typeList): zt_type = "t" if "z" in k else "z" key = zt_type + k + c[0] tArrRec[key] = mkvc(val, 2) # Masked array try: outTemp = recFunc.stack_arrays((outTemp, tArrRec)) except NameError: outTemp = tArrRec.copy() if "RealImag" in returnType: outArr = outTemp.copy() elif "Complex" in returnType: # Add the real and imaginary to a complex number outArr = np.empty(outTemp.shape, dtype=dtCP) for comp in ["freq", "x", "y", "z"]: outArr[comp] = outTemp[comp].copy() for comp in ["zxx", "zxy", "zyx", "zyy", "tzx", "tzy"]: outArr[comp] = ( outTemp[comp + "r"].copy() + 1j * outTemp[comp + "i"].copy() ) else: raise NotImplementedError( "{:s} is not implemented, as to be RealImag or Complex." ) # Return return outArr @classmethod def fromRecArray(cls, recArray, srcType="primary"): """ Class method that reads in a numpy record array to NSEMdata object. :param recArray: Record array with the data. Has to have ('freq','x','y','z') columns and some ('zxx','zxy','zyx','zyy','tzx','tzy') :type recArray: numpy.recarray :param srcType: The type of SimPEG.EM.NSEM.SrcNSEM to be used :type srcType: str, optional """ if srcType == "primary": src = Planewave_xy_1Dprimary elif srcType == "total": src = Planewave_xy_1DhomotD else: raise NotImplementedError("{:s} is not a valid source type for NSEMdata") # Find all the frequencies in recArray uniFreq = np.unique(recArray["freq"].copy()) srcList = [] dataList = [] for freq in uniFreq: # Initiate rxList rxList = [] # Find that data for freq dFreq = recArray[recArray["freq"] == freq].copy() # Find the impedance rxTypes in the recArray. rxTypes = [ comp for comp in recArray.dtype.names if (len(comp) == 4 or len(comp) == 3) and "z" in comp ] for rxType in rxTypes: # Find index of not nan values in rxType notNaNind = ~np.isnan(dFreq[rxType].copy()) if np.any(notNaNind): # Make sure that there is any data to add. locs = _rec_to_ndarr(dFreq[["x", "y", "z"]][notNaNind].copy()) if dFreq[rxType].dtype.name in "complex128": if "t" in rxType: rxList.append(Point3DTipper(locs, rxType[1:3], "real")) dataList.append(dFreq[rxType][notNaNind].real.copy()) rxList.append(Point3DTipper(locs, rxType[1:3], "imag")) dataList.append(dFreq[rxType][notNaNind].imag.copy()) elif "z" in rxType: rxList.append(Point3DImpedance(locs, rxType[1:3], "real")) dataList.append(dFreq[rxType][notNaNind].real.copy()) rxList.append(Point3DImpedance(locs, rxType[1:3], "imag")) dataList.append(dFreq[rxType][notNaNind].imag.copy()) else: component = "real" if "r" in rxType else "imag" if "z" in rxType: rxList.append( Point3DImpedance(locs, rxType[1:3], component) ) dataList.append(dFreq[rxType][notNaNind].copy()) if "t" in rxType: rxList.append(Point3DTipper(locs, rxType[1:3], component)) dataList.append(dFreq[rxType][notNaNind].copy()) srcList.append(src(rxList, freq)) # Make a survey survey = Survey(srcList) dataVec = np.hstack(dataList) return cls(survey, dataVec) def _rec_to_ndarr(rec_arr, data_type=float): """ Function to transform a numpy record array to a nd array. dupe of SimPEG.electromagnetics.natural_source.utils.rec_to_ndarr to avoid circular import """ # fix for numpy >= 1.16.0 # https://numpy.org/devdocs/release/1.16.0-notes.html#multi-field-views-return-a-view-instead-of-a-copy return np.array(recFunc.structured_to_unstructured(recFunc.repack_fields(rec_arr[list(rec_arr.dtype.names)])), dtype=data_type)
SimPEG/electromagnetics/natural_source/survey.py
9,624
Data class for NSEMdata. Stores the data vector indexed by the survey. Function to transform a numpy record array to a nd array. dupe of SimPEG.electromagnetics.natural_source.utils.rec_to_ndarr to avoid circular import Class method that reads in a numpy record array to NSEMdata object. :param recArray: Record array with the data. Has to have ('freq','x','y','z') columns and some ('zxx','zxy','zyx','zyy','tzx','tzy') :type recArray: numpy.recarray :param srcType: The type of SimPEG.EM.NSEM.SrcNSEM to be used :type srcType: str, optional Returns a numpy.recarray for a SimpegNSEM impedance data object. :param returnType: Switches between returning a rec array where the impedance is split to real and imaginary ('RealImag') or is a complex ('Complex') :type returnType: str, optional :rtype: numpy.recarray :return: Record array with data, with indexed columns Survey class Survey(BaseSurvey): """ Survey class for NSEM. **Requried** :param list srcList: List of sources associated with the survey **Optional** """ srcPair = BaseNSEMSrc def __init__(self, srcList, **kwargs): Sort these by frequency self.source_list = srcList BaseSurvey.__init__(self, **kwargs) _freqDict = {} for src in srcList: if src.freq not in _freqDict: _freqDict[src.freq] = [] _freqDict[src.freq] += [src] self._freqDict = _freqDict self._freqs = sorted([f for f in self._freqDict]) @property def freqs(self): """Frequencies""" return self._freqs @property def nFreq(self): """Number of frequencies""" return len(self._freqDict) def getSrcByFreq(self, freq): """Returns the sources associated with a specific frequency.""" assert freq in self._freqDict, "The requested frequency is not in this survey." return self._freqDict[freq] def eval(self, f): """ Evalute and return Data given calculated fields :param SimPEG.electromagnetics.frequency_domain.fields.FieldsFDEM f: A NSEM fileds object to evaluate data from :retype: SimPEG.EM.NSEM.Data :return: NSEM Data object """ data = Data(self) for src in self.source_list: sys.stdout.flush() for rx in src.receiver_list: data[src, rx] = rx.eval(src, self.mesh, f) return data def evalDeriv(self, f): raise Exception('Use Sources to project fields deriv.') Data Define the record fields Temp array for all the receivers of the source. Note: needs to be written more generally, using diffterent rxTypes and not all the data at the locations Assume the same locs for all RX Get the type and the value for the DataNSEM object as a list Insert the values to the temp array Masked array Add the real and imaginary to a complex number Return Find all the frequencies in recArray Initiate rxList Find that data for freq Find the impedance rxTypes in the recArray. Find index of not nan values in rxType Make sure that there is any data to add. Make a survey fix for numpy >= 1.16.0 https://numpy.org/devdocs/release/1.16.0-notes.htmlmulti-field-views-return-a-view-instead-of-a-copy
3,241
en
0.588412
# (C) Copyright 1996-2016 ECMWF. # # This software is licensed under the terms of the Apache Licence Version 2.0 # which can be obtained at http://www.apache.org/licenses/LICENSE-2.0. # In applying this licence, ECMWF does not waive the privileges and immunities # granted to it by virtue of its status as an intergovernmental organisation nor # does it submit to any jurisdiction. # importing Magics module from Magics.macro import * ref = 'era' # Setting of the output file name output = output(output_formats=['png'], output_name_first_page_number='off', output_name=ref) # Setting the coordinates of the geographical area projection = mmap( subpage_x_length=24., subpage_upper_right_longitude=50.00, subpage_upper_right_latitude=65.00, subpage_lower_left_latitude=25.00, subpage_lower_left_longitude=-20.0, subpage_map_projection='cylindrical', ) # Coastlines setting coast = mcoast(map_grid='on', map_grid_colour='grey', map_grid_thickness=2, map_coastline_colour='RGB(0.4,0.4,0.4)', map_coastline_thickness=3) obs = mobs( obsjson_input_filename = "small.json", obs_template_file_name = "obs.template", obs_identification = "on", obs_size = 0.5, obs_distance_apart = 0. ) title = mtext(text_lines=["Observation plotting ..." ], text_justification='left', text_font_size=0.8, text_colour='charcoal') # To the plot plot( output, projection, obs, coast, title, )
regression/era/era.py
1,527
(C) Copyright 1996-2016 ECMWF. This software is licensed under the terms of the Apache Licence Version 2.0 which can be obtained at http://www.apache.org/licenses/LICENSE-2.0. In applying this licence, ECMWF does not waive the privileges and immunities granted to it by virtue of its status as an intergovernmental organisation nor does it submit to any jurisdiction. importing Magics module Setting of the output file name Setting the coordinates of the geographical area Coastlines setting To the plot
506
en
0.886521
#!/usr/bin/env python3 """A utility script for automating the beets release process. """ import click import os import re import subprocess from contextlib import contextmanager import datetime BASE = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) CHANGELOG = os.path.join(BASE, 'docs', 'changelog.rst') @contextmanager def chdir(d): """A context manager that temporary changes the working directory. """ olddir = os.getcwd() os.chdir(d) yield os.chdir(olddir) @click.group() def release(): pass # Locations (filenames and patterns) of the version number. VERSION_LOCS = [ ( os.path.join(BASE, 'beets', '__init__.py'), [ ( r'__version__\s*=\s*u[\'"]([0-9\.]+)[\'"]', "__version__ = '{version}'", ) ] ), ( os.path.join(BASE, 'docs', 'conf.py'), [ ( r'version\s*=\s*[\'"]([0-9\.]+)[\'"]', "version = '{minor}'", ), ( r'release\s*=\s*[\'"]([0-9\.]+)[\'"]', "release = '{version}'", ), ] ), ( os.path.join(BASE, 'setup.py'), [ ( r'\s*version\s*=\s*[\'"]([0-9\.]+)[\'"]', " version='{version}',", ) ] ), ] GITHUB_USER = 'beetbox' GITHUB_REPO = 'beets' def bump_version(version): """Update the version number in setup.py, docs config, changelog, and root module. """ version_parts = [int(p) for p in version.split('.')] assert len(version_parts) == 3, "invalid version number" minor = '{}.{}'.format(*version_parts) major = '{}'.format(*version_parts) # Replace the version each place where it lives. for filename, locations in VERSION_LOCS: # Read and transform the file. out_lines = [] with open(filename) as f: found = False for line in f: for pattern, template in locations: match = re.match(pattern, line) if match: # Check that this version is actually newer. old_version = match.group(1) old_parts = [int(p) for p in old_version.split('.')] assert version_parts > old_parts, \ "version must be newer than {}".format( old_version ) # Insert the new version. out_lines.append(template.format( version=version, major=major, minor=minor, ) + '\n') found = True break else: # Normal line. out_lines.append(line) if not found: print(f"No pattern found in {filename}") # Write the file back. with open(filename, 'w') as f: f.write(''.join(out_lines)) # Generate bits to insert into changelog. header_line = f'{version} (in development)' header = '\n\n' + header_line + '\n' + '-' * len(header_line) + '\n\n' header += 'Changelog goes here!\n' # Insert into the right place. with open(CHANGELOG) as f: contents = f.read() location = contents.find('\n\n') # First blank line. contents = contents[:location] + header + contents[location:] # Write back. with open(CHANGELOG, 'w') as f: f.write(contents) @release.command() @click.argument('version') def bump(version): """Bump the version number. """ bump_version(version) def get_latest_changelog(): """Extract the first section of the changelog. """ started = False lines = [] with open(CHANGELOG) as f: for line in f: if re.match(r'^--+$', line.strip()): # Section boundary. Start or end. if started: # Remove last line, which is the header of the next # section. del lines[-1] break else: started = True elif started: lines.append(line) return ''.join(lines).strip() def rst2md(text): """Use Pandoc to convert text from ReST to Markdown. """ pandoc = subprocess.Popen( ['pandoc', '--from=rst', '--to=markdown', '--wrap=none'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE ) stdout, _ = pandoc.communicate(text.encode('utf-8')) md = stdout.decode('utf-8').strip() # Fix up odd spacing in lists. return re.sub(r'^- ', '- ', md, flags=re.M) def changelog_as_markdown(): """Get the latest changelog entry as hacked up Markdown. """ rst = get_latest_changelog() # Replace plugin links with plugin names. rst = re.sub(r':doc:`/plugins/(\w+)`', r'``\1``', rst) # References with text. rst = re.sub(r':ref:`([^<]+)(<[^>]+>)`', r'\1', rst) # Other backslashes with verbatim ranges. rst = re.sub(r'(\s)`([^`]+)`([^_])', r'\1``\2``\3', rst) # Command links with command names. rst = re.sub(r':ref:`(\w+)-cmd`', r'``\1``', rst) # Bug numbers. rst = re.sub(r':bug:`(\d+)`', r'#\1', rst) # Users. rst = re.sub(r':user:`(\w+)`', r'@\1', rst) # Convert with Pandoc. md = rst2md(rst) # Restore escaped issue numbers. md = re.sub(r'\\#(\d+)\b', r'#\1', md) return md @release.command() def changelog(): """Get the most recent version's changelog as Markdown. """ print(changelog_as_markdown()) def get_version(index=0): """Read the current version from the changelog. """ with open(CHANGELOG) as f: cur_index = 0 for line in f: match = re.search(r'^\d+\.\d+\.\d+', line) if match: if cur_index == index: return match.group(0) else: cur_index += 1 @release.command() def version(): """Display the current version. """ print(get_version()) @release.command() def datestamp(): """Enter today's date as the release date in the changelog. """ dt = datetime.datetime.now() stamp = '({} {}, {})'.format(dt.strftime('%B'), dt.day, dt.year) marker = '(in development)' lines = [] underline_length = None with open(CHANGELOG) as f: for line in f: if marker in line: # The header line. line = line.replace(marker, stamp) lines.append(line) underline_length = len(line.strip()) elif underline_length: # This is the line after the header. Rewrite the dashes. lines.append('-' * underline_length + '\n') underline_length = None else: lines.append(line) with open(CHANGELOG, 'w') as f: for line in lines: f.write(line) @release.command() def prep(): """Run all steps to prepare a release. - Tag the commit. - Build the sdist package. - Generate the Markdown changelog to ``changelog.md``. - Bump the version number to the next version. """ cur_version = get_version() # Tag. subprocess.check_call(['git', 'tag', f'v{cur_version}']) # Build. with chdir(BASE): subprocess.check_call(['python', 'setup.py', 'sdist']) # Generate Markdown changelog. cl = changelog_as_markdown() with open(os.path.join(BASE, 'changelog.md'), 'w') as f: f.write(cl) # Version number bump. # FIXME It should be possible to specify this as an argument. version_parts = [int(n) for n in cur_version.split('.')] version_parts[-1] += 1 next_version = '.'.join(map(str, version_parts)) bump_version(next_version) @release.command() def publish(): """Unleash a release unto the world. - Push the tag to GitHub. - Upload to PyPI. """ version = get_version(1) # Push to GitHub. with chdir(BASE): subprocess.check_call(['git', 'push']) subprocess.check_call(['git', 'push', '--tags']) # Upload to PyPI. path = os.path.join(BASE, 'dist', f'beets-{version}.tar.gz') subprocess.check_call(['twine', 'upload', path]) @release.command() def ghrelease(): """Create a GitHub release using the `github-release` command-line tool. Reads the changelog to upload from `changelog.md`. Uploads the tarball from the `dist` directory. """ version = get_version(1) tag = 'v' + version # Load the changelog. with open(os.path.join(BASE, 'changelog.md')) as f: cl_md = f.read() # Create the release. subprocess.check_call([ 'github-release', 'release', '-u', GITHUB_USER, '-r', GITHUB_REPO, '--tag', tag, '--name', f'{GITHUB_REPO} {version}', '--description', cl_md, ]) # Attach the release tarball. tarball = os.path.join(BASE, 'dist', f'beets-{version}.tar.gz') subprocess.check_call([ 'github-release', 'upload', '-u', GITHUB_USER, '-r', GITHUB_REPO, '--tag', tag, '--name', os.path.basename(tarball), '--file', tarball, ]) if __name__ == '__main__': release()
extra/release.py
9,525
Bump the version number. Update the version number in setup.py, docs config, changelog, and root module. Get the most recent version's changelog as Markdown. Get the latest changelog entry as hacked up Markdown. A context manager that temporary changes the working directory. Enter today's date as the release date in the changelog. Extract the first section of the changelog. Read the current version from the changelog. Create a GitHub release using the `github-release` command-line tool. Reads the changelog to upload from `changelog.md`. Uploads the tarball from the `dist` directory. Run all steps to prepare a release. - Tag the commit. - Build the sdist package. - Generate the Markdown changelog to ``changelog.md``. - Bump the version number to the next version. Unleash a release unto the world. - Push the tag to GitHub. - Upload to PyPI. Use Pandoc to convert text from ReST to Markdown. Display the current version. A utility script for automating the beets release process. !/usr/bin/env python3 Locations (filenames and patterns) of the version number. Replace the version each place where it lives. Read and transform the file. Check that this version is actually newer. Insert the new version. Normal line. Write the file back. Generate bits to insert into changelog. Insert into the right place. First blank line. Write back. Section boundary. Start or end. Remove last line, which is the header of the next section. Fix up odd spacing in lists. Replace plugin links with plugin names. References with text. Other backslashes with verbatim ranges. Command links with command names. Bug numbers. Users. Convert with Pandoc. Restore escaped issue numbers. The header line. This is the line after the header. Rewrite the dashes. Tag. Build. Generate Markdown changelog. Version number bump. FIXME It should be possible to specify this as an argument. Push to GitHub. Upload to PyPI. Load the changelog. Create the release. Attach the release tarball.
2,015
en
0.795307
import re from Error import Error4,Error6,Error9 from DataBase import BaseDatos,BdRow from .precompilada import precompilada from typing import Pattern def getEtiqueta(linea:str)->str: """Obtiene el nombre de la captura Args: linea (str): Linea donde se va a buscar la etiqueta Returns: str: Regresa el nombre de la etiqueta """ # Buscamos el mnemonico pattern='\s+([a-z]{1,5})\s+([a-z]{1,24})' busqueda=re.search(pattern,linea,re.IGNORECASE) # Obtenemos el mnemonico------------------------------- etiqueta =busqueda.group(2) return etiqueta def calcularEtiqueta(sustraendo:str,minuendo:str)-> str: """Resta la diferencia entre dos PC en hexadecimal sustraendo - minuendo - Si - Sustraendo - minuendo - En caso de error regresa 'e10' operando muy grande Args: sustraendo (str): Ejemplo '0x7' minuendo (str): Ejemplo '0x1' Returns: str: Ejemplo '0x06' """ print(sustraendo) print(minuendo) sustraendo=int(sustraendo,16) minuendo=int(minuendo,16) resultado:int= sustraendo-minuendo print(resultado) if resultado <-127 or 128<resultado: return 'e10' #E10 el salto relativo es muy lejano # Si es negativa elif resultado<0: return convertirA2Hex(resultado) # si es positiva else: return hex(resultado) def bindigits(n:int, bits:int)->str: """Convierte a binario un numero de complemento A2 en caso de negativo, normal en caso de ser positivo Args: n (int): E.g 7 bits (int): eg 3 Returns: str: E.g '001' """ s = bin(n & int("1"*bits, 2))[2:] return ("{0:0>%s}" % (bits)).format(s) def convertirA2Hex(numero:int)-> str: """Convierte un numero decimal a hexadecimal - Si el número es decimal lo convierte a complemento A2 Args: numero (int): Número decimal que se quiere convertir Eg. 07 Returns: str: Eg. 0x07 """ # cuantos bits ocupa el número hexadecimal cuantosBits=(len(hex(numero))-2) *4 # el -2 es 0x, el 4 es porque 1 hex equivale a 4 bits #numero convertido a binario binario=bindigits(numero,cuantosBits) return hex(int(binario, 2)) def precompilarPasada1(numLinea:int,modo:str,linea:str,pc: str)->precompilada: # variables globales # Buscamos el mnemonico pattern='\s+([a-z]{1,5})\s+([a-z]{1,24})' busqueda=re.search(pattern,linea,re.IGNORECASE) # Obtenemos el mnemonico------------------------------- mnemonico =busqueda.group(1) etiqueta=busqueda.group(2) # Consulta a la base de datos------------------------------- consultaBd:BdRow = BaseDatos.bdSearch(mnemonico,6) # obtenemos el Pc Actual=pc + bytesOcupados pcActual=hex(int(pc,16) +2) # El más 2 es porque todas las relativos usan 2 bytes # Datos directos-------------------------------------- lineaPrecompilada=precompilada(numLinea,modo,pcActual,consultaBd.opcode,etiqueta,consultaBd.byte) # Datos detivados----------------------------------- lineaPrecompilada.bytesOcupados=consultaBd.byte return lineaPrecompilada def precompilarPasada2(lineaPrecompilada:precompilada,pcEtiqueta:str)->precompilada: # obtenemos el Pc Actual=pc + bytesOcupados pcActual=hex(int(lineaPrecompilada.pcActual,16) ) # El más 2 es porque todas las relativos usan 2 bytes lineaPrecompilada1:precompilada # Calculamos el operando operandoPrecompilado=calcularEtiqueta(pcEtiqueta,pcActual) # Verificamos si el salto relaitvo no es tan grande if operandoPrecompilado=='e10': # en caso de error salto muy lejando lineaPrecompilada1=precompilada(0,'','','','',0) lineaPrecompilada1.error='e10' else: operandoPrecompilado=operandoPrecompilado[2:] # hacer una copia lineaPrecompilada1=precompilada(lineaPrecompilada.numLinea,lineaPrecompilada.modo,hex(int(lineaPrecompilada.pcActual,16)-2),lineaPrecompilada.opcode,operandoPrecompilado,lineaPrecompilada.byte) print(operandoPrecompilado) return lineaPrecompilada1 #return lineaPrecompilada1
Precompilar/relativo.py
4,178
Convierte a binario un numero de complemento A2 en caso de negativo, normal en caso de ser positivo Args: n (int): E.g 7 bits (int): eg 3 Returns: str: E.g '001' Resta la diferencia entre dos PC en hexadecimal sustraendo - minuendo - Si - Sustraendo - minuendo - En caso de error regresa 'e10' operando muy grande Args: sustraendo (str): Ejemplo '0x7' minuendo (str): Ejemplo '0x1' Returns: str: Ejemplo '0x06' Convierte un numero decimal a hexadecimal - Si el número es decimal lo convierte a complemento A2 Args: numero (int): Número decimal que se quiere convertir Eg. 07 Returns: str: Eg. 0x07 Obtiene el nombre de la captura Args: linea (str): Linea donde se va a buscar la etiqueta Returns: str: Regresa el nombre de la etiqueta Buscamos el mnemonico Obtenemos el mnemonico-------------------------------E10 el salto relativo es muy lejano Si es negativa si es positiva cuantos bits ocupa el número hexadecimal el -2 es 0x, el 4 es porque 1 hex equivale a 4 bitsnumero convertido a binario variables globales Buscamos el mnemonico Obtenemos el mnemonico------------------------------- Consulta a la base de datos------------------------------- obtenemos el Pc Actual=pc + bytesOcupados El más 2 es porque todas las relativos usan 2 bytes Datos directos-------------------------------------- Datos detivados----------------------------------- obtenemos el Pc Actual=pc + bytesOcupados El más 2 es porque todas las relativos usan 2 bytes Calculamos el operando Verificamos si el salto relaitvo no es tan grande en caso de error salto muy lejando hacer una copiareturn lineaPrecompilada1
1,647
es
0.832802
# Copyright 2016 Hewlett Packard Enterprise Development LP # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import sys from _restobject import RestObject def ex17_mount_virtual_media_iso(restobj, iso_url, boot_on_next_server_reset): sys.stdout.write("\nEXAMPLE 17: Mount iLO Virtual Media DVD ISO from URL\n") instances = restobj.search_for_type("Manager.") for instance in instances: rsp = restobj.rest_get(instance["href"]) rsp = restobj.rest_get(rsp.dict["links"]["VirtualMedia"]["href"]) for vmlink in rsp.dict["links"]["Member"]: response = restobj.rest_get(vmlink["href"]) if response.status == 200 and "DVD" in response.dict["MediaTypes"]: body = {"Image": iso_url} if (iso_url is not None and \ boot_on_next_server_reset is not None): body["Oem"] = {"Hp": {"BootOnNextServerReset": \ boot_on_next_server_reset}} response = restobj.rest_patch(vmlink["href"], body) restobj.error_handler(response) elif response.status != 200: restobj.error_handler(response) if __name__ == "__main__": # When running on the server locally use the following commented values # iLO_https_url = "blobstore://." # iLO_account = "None" # iLO_password = "None" # When running remotely connect using the iLO secured (https://) address, # iLO account name, and password to send https requests # iLO_https_url acceptable examples: # "https://10.0.0.100" # "https://f250asha.americas.hpqcorp.net" iLO_https_url = "https://10.0.0.100" iLO_account = "admin" iLO_password = "password" #Create a REST object REST_OBJ = RestObject(iLO_https_url, iLO_account, iLO_password) ex17_mount_virtual_media_iso(REST_OBJ, "http://10.0.0.100/test.iso", True)
examples/Rest/ex17_mount_virtual_media_iso.py
2,563
Copyright 2016 Hewlett Packard Enterprise Development LP Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. When running on the server locally use the following commented values iLO_https_url = "blobstore://." iLO_account = "None" iLO_password = "None" When running remotely connect using the iLO secured (https://) address, iLO account name, and password to send https requests iLO_https_url acceptable examples: "https://10.0.0.100" "https://f250asha.americas.hpqcorp.net"Create a REST object
967
en
0.833323
from gpiozero import Servo from gpiozero import LED from time import sleep from WeatherDataCW import WeatherData class WeatherDashboard: servo_pin = 17 led_pin = 14 servoCorrection=0.5 maxPW=(2.0+servoCorrection)/1000 minPW=(1.0-servoCorrection)/1000 def __init__(self, servo_position=0, led_status=0): self.servo = Servo(self.servo_pin, min_pulse_width=self.minPW, max_pulse_width=self.maxPW) self.led = LED(self.led_pin) self.move_servo(servo_position) self.set_led_status(led_status) def move_servo(self, servo_position=0): self.servo.value = self.convert_percentage_to_integer(servo_position) def turnOffServo(self): sleep(2) self.servo.close() def set_led_status(self, led_status=0): if(led_status==0): self.led.off() elif (led_status==1): self.led.on() else: self.led.blink() def convert_percentage_to_integer(self, percentage_amount): #adjust for servos that turn counter clockwise by default adjusted_percentage_amount = 100 - percentage_amount return (adjusted_percentage_amount*0.02)-1 if __name__=="__main__": weather_data = WeatherData('Yekaterinburg') print("%s %sC %s wind speed %s km/h" %(weather_data.getCity(), weather_data.getTemperature(), weather_data.getWeatherConditions(), weather_data.getWindSpeed())) print(weather_data.getServoValue()) print(weather_data.getLEDValue()) weather_dashboard = WeatherDashboard( weather_data.getServoValue(), weather_data.getLEDValue()) weather_dashboard.turnOffServo()
WeatherDashboardCW.py
1,779
adjust for servos that turn counter clockwise by default
56
en
0.846388
# Generated by Django 2.0.3 on 2018-05-28 23:42 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('api', '0003_task_order'), ] operations = [ migrations.AlterField( model_name='task', name='order', field=models.IntegerField(blank=True), ), ]
backend/api/migrations/0004_auto_20180528_2342.py
371
Generated by Django 2.0.3 on 2018-05-28 23:42
45
en
0.619848
import datetime from django.conf import settings from django.core.cache import cache from django.db import models from django.db.models import Sum import commonware.log import waffle import amo import mkt.constants.comm as comm from amo.utils import cache_ns_key from mkt.comm.utils import create_comm_note from mkt.site.mail import send_mail_jinja from mkt.site.models import ManagerBase, ModelBase, skip_cache from mkt.tags.models import Tag from mkt.translations.fields import save_signal, TranslatedField from mkt.users.models import UserProfile from mkt.webapps.indexers import WebappIndexer from mkt.webapps.models import Webapp user_log = commonware.log.getLogger('z.users') QUEUE_TARAKO = 'tarako' class CannedResponse(ModelBase): name = TranslatedField() response = TranslatedField(short=False) sort_group = models.CharField(max_length=255) class Meta: db_table = 'cannedresponses' def __unicode__(self): return unicode(self.name) models.signals.pre_save.connect(save_signal, sender=CannedResponse, dispatch_uid='cannedresponses_translations') class EditorSubscription(ModelBase): user = models.ForeignKey(UserProfile) addon = models.ForeignKey(Webapp) class Meta: db_table = 'editor_subscriptions' class ReviewerScore(ModelBase): user = models.ForeignKey(UserProfile, related_name='_reviewer_scores') addon = models.ForeignKey(Webapp, blank=True, null=True, related_name='+') score = models.SmallIntegerField() # For automated point rewards. note_key = models.SmallIntegerField(choices=amo.REVIEWED_CHOICES.items(), default=0) # For manual point rewards with a note. note = models.CharField(max_length=255, blank=True) class Meta: db_table = 'reviewer_scores' ordering = ('-created',) @classmethod def get_key(cls, key=None, invalidate=False): namespace = 'riscore' if not key: # Assuming we're invalidating the namespace. cache_ns_key(namespace, invalidate) return else: # Using cache_ns_key so each cache val is invalidated together. ns_key = cache_ns_key(namespace, invalidate) return '%s:%s' % (ns_key, key) @classmethod def get_event(cls, addon, status, **kwargs): """Return the review event type constant. This is determined by the app type and the queue the addon is currently in (which is determined from the status). Note: We're not using addon.status because this is called after the status has been updated by the reviewer action. """ if addon.is_packaged: if status in amo.WEBAPPS_APPROVED_STATUSES: return amo.REVIEWED_WEBAPP_UPDATE else: # If it's not PUBLIC, assume it's a new submission. return amo.REVIEWED_WEBAPP_PACKAGED else: # It's a hosted app. in_rereview = kwargs.pop('in_rereview', False) if status in amo.WEBAPPS_APPROVED_STATUSES and in_rereview: return amo.REVIEWED_WEBAPP_REREVIEW else: return amo.REVIEWED_WEBAPP_HOSTED @classmethod def award_points(cls, user, addon, status, **kwargs): """Awards points to user based on an event and the queue. `event` is one of the `REVIEWED_` keys in constants. `status` is one of the `STATUS_` keys in constants. """ event = cls.get_event(addon, status, **kwargs) score = amo.REVIEWED_SCORES.get(event) if score: cls.objects.create(user=user, addon=addon, score=score, note_key=event) cls.get_key(invalidate=True) user_log.info( (u'Awarding %s points to user %s for "%s" for addon %s' % (score, user, amo.REVIEWED_CHOICES[event], addon.id)) .encode('utf-8')) return score @classmethod def award_moderation_points(cls, user, addon, review_id): """Awards points to user based on moderated review.""" event = amo.REVIEWED_APP_REVIEW score = amo.REVIEWED_SCORES.get(event) cls.objects.create(user=user, addon=addon, score=score, note_key=event) cls.get_key(invalidate=True) user_log.info( u'Awarding %s points to user %s for "%s" for review %s' % ( score, user, amo.REVIEWED_CHOICES[event], review_id)) @classmethod def get_total(cls, user): """Returns total points by user.""" key = cls.get_key('get_total:%s' % user.id) val = cache.get(key) if val is not None: return val val = (ReviewerScore.objects.no_cache().filter(user=user) .aggregate(total=Sum('score')) .values())[0] if val is None: val = 0 cache.set(key, val, None) return val @classmethod def get_recent(cls, user, limit=5): """Returns most recent ReviewerScore records.""" key = cls.get_key('get_recent:%s' % user.id) val = cache.get(key) if val is not None: return val val = ReviewerScore.objects.no_cache().filter(user=user) val = list(val[:limit]) cache.set(key, val, None) return val @classmethod def get_performance(cls, user): """Returns sum of reviewer points.""" key = cls.get_key('get_performance:%s' % user.id) val = cache.get(key) if val is not None: return val sql = """ SELECT `reviewer_scores`.*, SUM(`reviewer_scores`.`score`) AS `total` FROM `reviewer_scores` LEFT JOIN `addons` ON (`reviewer_scores`.`addon_id`=`addons`.`id`) WHERE `reviewer_scores`.`user_id` = %s ORDER BY `total` DESC """ with skip_cache(): val = list(ReviewerScore.objects.raw(sql, [user.id])) cache.set(key, val, None) return val @classmethod def get_performance_since(cls, user, since): """ Returns sum of reviewer points since the given datetime. """ key = cls.get_key('get_performance:%s:%s' % (user.id, since.isoformat())) val = cache.get(key) if val is not None: return val sql = """ SELECT `reviewer_scores`.*, SUM(`reviewer_scores`.`score`) AS `total` FROM `reviewer_scores` LEFT JOIN `addons` ON (`reviewer_scores`.`addon_id`=`addons`.`id`) WHERE `reviewer_scores`.`user_id` = %s AND `reviewer_scores`.`created` >= %s ORDER BY `total` DESC """ with skip_cache(): val = list(ReviewerScore.objects.raw(sql, [user.id, since])) cache.set(key, val, 3600) return val @classmethod def _leaderboard_query(cls, since=None, types=None): """ Returns common SQL to leaderboard calls. """ query = (cls.objects .values_list('user__id', 'user__display_name') .annotate(total=Sum('score')) .exclude(user__groups__name__in=('No Reviewer Incentives', 'Staff', 'Admins')) .order_by('-total')) if since is not None: query = query.filter(created__gte=since) if types is not None: query = query.filter(note_key__in=types) return query @classmethod def get_leaderboards(cls, user, days=7, types=None): """Returns leaderboards with ranking for the past given days. This will return a dict of 3 items:: {'leader_top': [...], 'leader_near: [...], 'user_rank': (int)} If the user is not in the leaderboard, or if the user is in the top 5, 'leader_near' will be an empty list and 'leader_top' will contain 5 elements instead of the normal 3. """ key = cls.get_key('get_leaderboards:%s' % user.id) val = cache.get(key) if val is not None: return val week_ago = datetime.date.today() - datetime.timedelta(days=days) leader_top = [] leader_near = [] query = cls._leaderboard_query(since=week_ago, types=types) scores = [] user_rank = 0 in_leaderboard = False for rank, row in enumerate(query, 1): user_id, name, total = row scores.append({ 'user_id': user_id, 'name': name, 'rank': rank, 'total': int(total), }) if user_id == user.id: user_rank = rank in_leaderboard = True if not in_leaderboard: leader_top = scores[:5] else: if user_rank <= 5: # User is in top 5, show top 5. leader_top = scores[:5] else: leader_top = scores[:3] leader_near = [scores[user_rank - 2], scores[user_rank - 1]] try: leader_near.append(scores[user_rank]) except IndexError: pass # User is last on the leaderboard. val = { 'leader_top': leader_top, 'leader_near': leader_near, 'user_rank': user_rank, } cache.set(key, val, None) return val @classmethod def all_users_by_score(cls): """ Returns reviewers ordered by highest total points first. """ query = cls._leaderboard_query() scores = [] for row in query: user_id, name, total = row user_level = len(amo.REVIEWED_LEVELS) - 1 for i, level in enumerate(amo.REVIEWED_LEVELS): if total < level['points']: user_level = i - 1 break # Only show level if it changes. if user_level < 0: level = '' else: level = amo.REVIEWED_LEVELS[user_level]['name'] scores.append({ 'user_id': user_id, 'name': name, 'total': int(total), 'level': level, }) prev = None for score in reversed(scores): if score['level'] == prev: score['level'] = '' else: prev = score['level'] return scores class EscalationQueue(ModelBase): addon = models.ForeignKey(Webapp) class Meta: db_table = 'escalation_queue' class RereviewQueue(ModelBase): addon = models.ForeignKey(Webapp) class Meta: db_table = 'rereview_queue' @classmethod def flag(cls, addon, event, message=None): cls.objects.get_or_create(addon=addon) if message: amo.log(event, addon, addon.current_version, details={'comments': message}) else: amo.log(event, addon, addon.current_version) # TODO: if we ever get rid of ActivityLog for reviewer notes, replace # all flag calls to use the comm constant and not have to use # ACTION_MAP. create_comm_note(addon, addon.current_version, None, message, note_type=comm.ACTION_MAP(event)) def send_tarako_mail(review): if not waffle.switch_is_active('comm-dashboard'): send_mail_jinja( 'Low-memory devices review {passed}'.format( passed='passed' if review.passed else 'failed'), 'reviewers/emails/tarako_review_complete.txt', {'review': review}, recipient_list=[a.email for a in review.app.authors.all()], from_email=settings.MKT_REVIEWERS_EMAIL) def tarako_passed(review): """Add the tarako tag to the app.""" tag = Tag(tag_text='tarako') tag.save_tag(review.app) WebappIndexer.index_ids([review.app.pk]) send_tarako_mail(review) def tarako_failed(review): """Remove the tarako tag from the app.""" tag = Tag(tag_text='tarako') tag.remove_tag(review.app) WebappIndexer.index_ids([review.app.pk]) send_tarako_mail(review) class AdditionalReviewManager(ManagerBase): def unreviewed(self, queue, and_approved=False): query = { 'passed': None, 'queue': queue, } if and_approved: query['app__status__in'] = amo.WEBAPPS_APPROVED_STATUSES return self.get_queryset().no_cache().filter(**query) def latest_for_queue(self, queue): try: return self.get_queryset().filter(queue=queue).latest() except AdditionalReview.DoesNotExist: return None class AdditionalReview(ModelBase): app = models.ForeignKey(Webapp) queue = models.CharField(max_length=30) passed = models.NullBooleanField() review_completed = models.DateTimeField(null=True) comment = models.CharField(null=True, blank=True, max_length=255) reviewer = models.ForeignKey('users.UserProfile', null=True, blank=True) objects = AdditionalReviewManager() class Meta: db_table = 'additional_review' get_latest_by = 'created' @property def pending(self): return self.passed is None @property def failed(self): return self.passed is False def __init__(self, *args, **kwargs): super(AdditionalReview, self).__init__(*args, **kwargs) from mkt.reviewers.utils import log_reviewer_action self.log_reviewer_action = log_reviewer_action def execute_post_review_task(self): """ Call the correct post-review function for the queue. """ # TODO: Pull this function from somewhere based on self.queue. if self.passed is None: raise ValueError('cannot execute post-review task when unreviewed') elif self.passed: tarako_passed(self) action = amo.LOG.PASS_ADDITIONAL_REVIEW else: tarako_failed(self) action = amo.LOG.FAIL_ADDITIONAL_REVIEW self.log_reviewer_action( self.app, self.reviewer, self.comment or '', action, queue=self.queue) def cleanup_queues(sender, instance, **kwargs): RereviewQueue.objects.filter(addon=instance).delete() EscalationQueue.objects.filter(addon=instance).delete() models.signals.post_delete.connect(cleanup_queues, sender=Webapp, dispatch_uid='queue-addon-cleanup')
mkt/reviewers/models.py
14,819
Returns common SQL to leaderboard calls. Returns reviewers ordered by highest total points first. Awards points to user based on moderated review. Awards points to user based on an event and the queue. `event` is one of the `REVIEWED_` keys in constants. `status` is one of the `STATUS_` keys in constants. Call the correct post-review function for the queue. Return the review event type constant. This is determined by the app type and the queue the addon is currently in (which is determined from the status). Note: We're not using addon.status because this is called after the status has been updated by the reviewer action. Returns leaderboards with ranking for the past given days. This will return a dict of 3 items:: {'leader_top': [...], 'leader_near: [...], 'user_rank': (int)} If the user is not in the leaderboard, or if the user is in the top 5, 'leader_near' will be an empty list and 'leader_top' will contain 5 elements instead of the normal 3. Returns sum of reviewer points. Returns sum of reviewer points since the given datetime. Returns most recent ReviewerScore records. Returns total points by user. Remove the tarako tag from the app. Add the tarako tag to the app. For automated point rewards. For manual point rewards with a note. Assuming we're invalidating the namespace. Using cache_ns_key so each cache val is invalidated together. If it's not PUBLIC, assume it's a new submission. It's a hosted app. User is in top 5, show top 5. User is last on the leaderboard. Only show level if it changes. TODO: if we ever get rid of ActivityLog for reviewer notes, replace all flag calls to use the comm constant and not have to use ACTION_MAP. TODO: Pull this function from somewhere based on self.queue.
1,747
en
0.8706
_base_ = [ '../../_base_/schedules/schedule_1200e.py', '../../_base_/runtime_10e.py' ] model = dict( type='DBNet', pretrained='torchvision://resnet18', backbone=dict( type='ResNet', depth=18, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=-1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=False, style='caffe'), neck=dict( type='FPNC', in_channels=[64, 128, 256, 512], lateral_channels=256), bbox_head=dict( type='DBHead', text_repr_type='quad', in_channels=256, loss=dict(type='DBLoss', alpha=5.0, beta=10.0, bbce_loss=True)), train_cfg=None, test_cfg=None) dataset_type = 'IcdarDataset' data_root = 'data/icdar2015/' img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) # for visualizing img, pls uncomment it. # img_norm_cfg = dict(mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True) train_pipeline = [ dict(type='LoadImageFromFile'), dict( type='LoadTextAnnotations', with_bbox=True, with_mask=True, poly2mask=False), dict(type='ColorJitter', brightness=32.0 / 255, saturation=0.5), dict(type='Normalize', **img_norm_cfg), # img aug dict( type='ImgAug', args=[['Fliplr', 0.5], dict(cls='Affine', rotate=[-10, 10]), ['Resize', [0.5, 3.0]]]), # random crop dict(type='EastRandomCrop', target_size=(640, 640)), dict(type='DBNetTargets', shrink_ratio=0.4), dict(type='Pad', size_divisor=32), # for visualizing img and gts, pls set visualize = True dict( type='CustomFormatBundle', keys=['gt_shrink', 'gt_shrink_mask', 'gt_thr', 'gt_thr_mask'], visualize=dict(flag=False, boundary_key='gt_shrink')), dict( type='Collect', keys=['img', 'gt_shrink', 'gt_shrink_mask', 'gt_thr', 'gt_thr_mask']) ] test_pipeline = [ dict(type='LoadImageFromFile'), dict( type='MultiScaleFlipAug', img_scale=(1333, 736), flip=False, transforms=[ dict(type='Resize', img_scale=(2944, 736), keep_ratio=True), dict(type='Normalize', **img_norm_cfg), dict(type='Pad', size_divisor=32), dict(type='ImageToTensor', keys=['img']), dict(type='Collect', keys=['img']), ]) ] data = dict( samples_per_gpu=16, workers_per_gpu=8, train=dict( type=dataset_type, ann_file=data_root + '/instances_training.json', # for debugging top k imgs # select_first_k=200, img_prefix=data_root + '/imgs', pipeline=train_pipeline), val=dict( type=dataset_type, ann_file=data_root + '/instances_test.json', img_prefix=data_root + '/imgs', # select_first_k=100, pipeline=test_pipeline), test=dict( type=dataset_type, ann_file=data_root + '/instances_test.json', img_prefix=data_root + '/imgs', # select_first_k=100, pipeline=test_pipeline)) evaluation = dict(interval=100, metric='hmean-iou')
configs/textdet/dbnet/dbnet_r18_fpnc_1200e_icdar2015.py
3,143
for visualizing img, pls uncomment it. img_norm_cfg = dict(mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True) img aug random crop for visualizing img and gts, pls set visualize = True for debugging top k imgs select_first_k=200, select_first_k=100, select_first_k=100,
261
en
0.592518
# Copyright 2013-2022 Lawrence Livermore National Security, LLC and other # Spack Project Developers. See the top-level COPYRIGHT file for details. # # SPDX-License-Identifier: (Apache-2.0 OR MIT) from spack.package import * class RVgam(RPackage): """Vector Generalized Linear and Additive Models. An implementation of about 6 major classes of statistical regression models. The central algorithm is Fisher scoring and iterative reweighted least squares. At the heart of this package are the vector generalized linear and additive model (VGLM/VGAM) classes. VGLMs can be loosely thought of as multivariate GLMs. VGAMs are data-driven VGLMs that use smoothing. The book "Vector Generalized Linear and Additive Models: With an Implementation in R" (Yee, 2015) <DOI:10.1007/978-1-4939-2818-7> gives details of the statistical framework and the package. Currently only fixed-effects models are implemented. Many (150+) models and distributions are estimated by maximum likelihood estimation (MLE) or penalized MLE. The other classes are RR-VGLMs (reduced-rank VGLMs), quadratic RR-VGLMs, reduced-rank VGAMs, RCIMs (row-column interaction models)---these classes perform constrained and unconstrained quadratic ordination (CQO/UQO) models in ecology, as well as constrained additive ordination (CAO). Hauck-Donner effect detection is implemented. Note that these functions are subject to change; see the NEWS and ChangeLog files for latest changes.""" cran = "VGAM" version('1.1-6', sha256='446a61bac5dd4794e05d20c2f3901eec54afac52c6e23ce2787c5575170dd417') version('1.1-5', sha256='30190b150f3e5478137d288a45f575b2654ad7c29254b0a1fe5c954ee010a1bb') version('1.1-1', sha256='de192bd65a7e8818728008de8e60e6dd3b61a13616c887a43e0ccc8147c7da52') version('1.0-6', sha256='121820a167411e847b41bdcb0028b55842d0ccc0c3471755c67449837e0fe3b9') version('1.0-4', sha256='e581985f78ef8b866d0e810b2727061bb9c9bc177b2c9090aebb3a35ae87a964') version('1.0-3', sha256='23bb6690ae15e9ede3198ef55d5d3236c279aa8fa6bd4f7350242379d9d72673') version('1.0-2', sha256='03561bf484f97b616b1979132c759c5faa69c5d5a4cfd7aea2ea6d3612ac0961') version('1.0-1', sha256='c066864e406fcee23f383a28299dba3cf83356e5b68df16324885afac87a05ea') version('1.0-0', sha256='6acdd7db49c0987c565870afe593160ceba72a6ca4a84e6da3cf6f74d1fa02e1') depends_on('r@3.0.0:', type=('build', 'run')) depends_on('r@3.1.0:', type=('build', 'run'), when='@1.0-2:') depends_on('r@3.4.0:', type=('build', 'run'), when='@1.0-4:') depends_on('r@3.5.0:', type=('build', 'run'), when='@1.1-5:')
var/spack/repos/builtin/packages/r-vgam/package.py
2,652
Vector Generalized Linear and Additive Models. An implementation of about 6 major classes of statistical regression models. The central algorithm is Fisher scoring and iterative reweighted least squares. At the heart of this package are the vector generalized linear and additive model (VGLM/VGAM) classes. VGLMs can be loosely thought of as multivariate GLMs. VGAMs are data-driven VGLMs that use smoothing. The book "Vector Generalized Linear and Additive Models: With an Implementation in R" (Yee, 2015) <DOI:10.1007/978-1-4939-2818-7> gives details of the statistical framework and the package. Currently only fixed-effects models are implemented. Many (150+) models and distributions are estimated by maximum likelihood estimation (MLE) or penalized MLE. The other classes are RR-VGLMs (reduced-rank VGLMs), quadratic RR-VGLMs, reduced-rank VGAMs, RCIMs (row-column interaction models)---these classes perform constrained and unconstrained quadratic ordination (CQO/UQO) models in ecology, as well as constrained additive ordination (CAO). Hauck-Donner effect detection is implemented. Note that these functions are subject to change; see the NEWS and ChangeLog files for latest changes. Copyright 2013-2022 Lawrence Livermore National Security, LLC and other Spack Project Developers. See the top-level COPYRIGHT file for details. SPDX-License-Identifier: (Apache-2.0 OR MIT)
1,384
en
0.849033
from django import db from django.conf import settings from django.contrib.contenttypes.models import ContentType from django.contrib.sites.models import Site from django.contrib.contenttypes.views import shortcut from django.core.exceptions import ObjectDoesNotExist from django.http import HttpRequest from django.test import TestCase class ContentTypesTests(TestCase): def setUp(self): # First, let's make sure we're dealing with a blank slate (and that # DEBUG is on so that queries get logged) self.old_DEBUG = settings.DEBUG self.old_Site_meta_installed = Site._meta.installed settings.DEBUG = True ContentType.objects.clear_cache() db.reset_queries() def tearDown(self): settings.DEBUG = self.old_DEBUG Site._meta.installed = self.old_Site_meta_installed def test_lookup_cache(self): """ Make sure that the content type cache (see ContentTypeManager) works correctly. Lookups for a particular content type -- by model or by ID -- should hit the database only on the first lookup. """ # At this point, a lookup for a ContentType should hit the DB ContentType.objects.get_for_model(ContentType) self.assertEqual(1, len(db.connection.queries)) # A second hit, though, won't hit the DB, nor will a lookup by ID ct = ContentType.objects.get_for_model(ContentType) self.assertEqual(1, len(db.connection.queries)) ContentType.objects.get_for_id(ct.id) self.assertEqual(1, len(db.connection.queries)) # Once we clear the cache, another lookup will again hit the DB ContentType.objects.clear_cache() ContentType.objects.get_for_model(ContentType) len(db.connection.queries) self.assertEqual(2, len(db.connection.queries)) def test_shortcut_view(self): """ Check that the shortcut view (used for the admin "view on site" functionality) returns a complete URL regardless of whether the sites framework is installed """ request = HttpRequest() request.META = { "SERVER_NAME": "Example.com", "SERVER_PORT": "80", } from django.contrib.auth.models import User user_ct = ContentType.objects.get_for_model(User) obj = User.objects.create(username="john") if Site._meta.installed: current_site = Site.objects.get_current() response = shortcut(request, user_ct.id, obj.id) self.assertEqual("http://%s/users/john/" % current_site.domain, response._headers.get("location")[1]) Site._meta.installed = False response = shortcut(request, user_ct.id, obj.id) self.assertEqual("http://Example.com/users/john/", response._headers.get("location")[1])
django/contrib/contenttypes/tests.py
2,909
Make sure that the content type cache (see ContentTypeManager) works correctly. Lookups for a particular content type -- by model or by ID -- should hit the database only on the first lookup. Check that the shortcut view (used for the admin "view on site" functionality) returns a complete URL regardless of whether the sites framework is installed First, let's make sure we're dealing with a blank slate (and that DEBUG is on so that queries get logged) At this point, a lookup for a ContentType should hit the DB A second hit, though, won't hit the DB, nor will a lookup by ID Once we clear the cache, another lookup will again hit the DB
642
en
0.874293
""" BasicSR/codes/dataops/common.py (8-Nov-20) https://github.com/victorca25/BasicSR/blob/dev2/codes/dataops/common.py """ import os import math import pickle import random import numpy as np import torch import cv2 import logging import copy from torchvision.utils import make_grid #from dataops.colors import * from .colors import * #from dataops.debug import tmp_vis, describe_numpy, describe_tensor #################### # Files & IO #################### ###################### get image path list ###################### IMG_EXTENSIONS = ['.jpg', '.JPG', '.jpeg', '.JPEG', '.png', '.PNG', '.ppm', '.PPM', '.bmp', '.BMP', '.dng', '.DNG', '.webp','.npy', '.NPY'] def is_image_file(filename): return any(filename.endswith(extension) for extension in IMG_EXTENSIONS) def _get_paths_from_images(path): '''get image path list from image folder''' assert os.path.isdir(path), '{:s} is not a valid directory'.format(path) images = [] for dirpath, _, fnames in sorted(os.walk(path)): for fname in sorted(fnames): if is_image_file(fname): img_path = os.path.join(dirpath, fname) images.append(img_path) assert images, '{:s} has no valid image file'.format(path) return images def _get_paths_from_lmdb(dataroot): '''get image path list from lmdb''' import lmdb env = lmdb.open(dataroot, readonly=True, lock=False, readahead=False, meminit=False) keys_cache_file = os.path.join(dataroot, '_keys_cache.p') logger = logging.getLogger('base') if os.path.isfile(keys_cache_file): logger.info('Read lmdb keys from cache: {}'.format(keys_cache_file)) keys = pickle.load(open(keys_cache_file, "rb")) else: with env.begin(write=False) as txn: logger.info('Creating lmdb keys cache: {}'.format(keys_cache_file)) keys = [key.decode('ascii') for key, _ in txn.cursor()] pickle.dump(keys, open(keys_cache_file, 'wb')) paths = sorted([key for key in keys if not key.endswith('.meta')]) return env, paths def get_image_paths(data_type, dataroot): '''get image path list support lmdb or image files''' env, paths = None, None if dataroot is not None: if data_type == 'lmdb': env, paths = _get_paths_from_lmdb(dataroot) elif data_type == 'img': paths = sorted(_get_paths_from_images(dataroot)) else: raise NotImplementedError('data_type [{:s}] is not recognized.'.format(data_type)) return env, paths ###################### read images ###################### def _read_lmdb_img(env, path): with env.begin(write=False) as txn: buf = txn.get(path.encode('ascii')) buf_meta = txn.get((path + '.meta').encode('ascii')).decode('ascii') img_flat = np.frombuffer(buf, dtype=np.uint8) H, W, C = [int(s) for s in buf_meta.split(',')] img = img_flat.reshape(H, W, C) return img def read_img(env, path, out_nc=3, fix_channels=True): ''' Reads image using cv2 (rawpy if dng) or from lmdb by default (can also use using PIL instead of cv2) Arguments: out_nc: Desired number of channels fix_channels: changes the images to the desired number of channels Output: Numpy uint8, HWC, BGR, [0,255] by default ''' img = None if env is None: # img if(path[-3:].lower() == 'dng'): # if image is a DNG import rawpy with rawpy.imread(path) as raw: img = raw.postprocess() if(path[-3:].lower() == 'npy'): # if image is a NPY numpy array with open(path, 'rb') as f: img = np.load(f) else: # else, if image can be read by cv2 img = cv2.imread(path, cv2.IMREAD_UNCHANGED) #TODO: add variable detecting if cv2 is not available and try PIL instead # elif: # using PIL instead of OpenCV # img = Image.open(path).convert('RGB') # else: # For other images unrecognized by cv2 # import matplotlib.pyplot as plt # img = (255*plt.imread(path)[:,:,:3]).astype('uint8') else: img = _read_lmdb_img(env, path) # if not img: # raise ValueError(f"Failed to read image: {path}") if fix_channels: img = fix_img_channels(img, out_nc) return img def fix_img_channels(img, out_nc): ''' fix image channels to the expected number ''' # if image has only 2 dimensions, add "channel" dimension (1) if img.ndim == 2: #img = img[..., np.newaxis] #alt #img = np.expand_dims(img, axis=2) img = np.tile(np.expand_dims(img, axis=2), (1, 1, 3)) # special case: properly remove alpha channel if out_nc == 3 and img.shape[2] == 4: img = bgra2rgb(img) # remove all extra channels elif img.shape[2] > out_nc: img = img[:, :, :out_nc] # if alpha is expected, add solid alpha channel elif img.shape[2] == 3 and out_nc == 4: img = np.dstack((img, np.full(img.shape[:-1], 255, dtype=np.uint8))) return img #################### # image processing # process on numpy image #################### def bgra2rgb(img): ''' cv2.cvtColor(img, cv2.COLOR_BGRA2BGR) has an issue removing the alpha channel, this gets rid of wrong transparent colors that can harm training ''' if img.shape[2] == 4: #b, g, r, a = cv2.split((img*255).astype(np.uint8)) b, g, r, a = cv2.split((img.astype(np.uint8))) b = cv2.bitwise_and(b, b, mask=a) g = cv2.bitwise_and(g, g, mask=a) r = cv2.bitwise_and(r, r, mask=a) #return cv2.merge([b, g, r]).astype(np.float32)/255. return cv2.merge([b, g, r]) return img def channel_convert(in_c, tar_type, img_list): # conversion among BGR, gray and y # Note: OpenCV uses inverted channels BGR, instead of RGB. # If images are loaded with something other than OpenCV, # check that the channels are in the correct order and use # the alternative conversion functions. #if in_c == 4 and tar_type == 'RGB-A': # BGRA to BGR, remove alpha channel #return [cv2.cvtColor(img, cv2.COLOR_BGRA2BGR) for img in img_list] #return [bgra2rgb(img) for img in img_list] if in_c == 3 and tar_type == 'gray': # BGR to gray gray_list = [cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) for img in img_list] return [np.expand_dims(img, axis=2) for img in gray_list] elif in_c == 3 and tar_type == 'RGB-LAB': #RGB to LAB return [cv2.cvtColor(img, cv2.COLOR_BGR2LAB) for img in img_list] elif in_c == 3 and tar_type == 'LAB-RGB': #RGB to LAB return [cv2.cvtColor(img, cv2.COLOR_LAB2BGR) for img in img_list] elif in_c == 3 and tar_type == 'y': # BGR to y y_list = [bgr2ycbcr(img, only_y=True) for img in img_list] return [np.expand_dims(img, axis=2) for img in y_list] elif in_c == 1 and tar_type == 'RGB': # gray/y to BGR return [cv2.cvtColor(img, cv2.COLOR_GRAY2BGR) for img in img_list] else: return img_list def rgb2ycbcr(img, only_y=True): '''same as matlab rgb2ycbcr only_y: only return Y channel Input: uint8, [0, 255] float, [0, 1] ''' in_img_type = img.dtype img_ = img.astype(np.float32) if in_img_type != np.uint8: img_ *= 255. # convert if only_y: rlt = np.dot(img_ , [65.481, 128.553, 24.966]) / 255.0 + 16.0 else: rlt = np.matmul(img_ , [[65.481, -37.797, 112.0], [128.553, -74.203, -93.786], [24.966, 112.0, -18.214]]) / 255.0 + [16, 128, 128] if in_img_type == np.uint8: rlt = rlt.round() else: rlt /= 255. return rlt.astype(in_img_type) def bgr2ycbcr(img, only_y=True, separate=False): '''bgr version of matlab rgb2ycbcr Python opencv library (cv2) cv2.COLOR_BGR2YCrCb has different parameters with MATLAB color convertion. only_y: only return Y channel separate: if true, will returng the channels as separate images Input: uint8, [0, 255] float, [0, 1] ''' in_img_type = img.dtype img_ = img.astype(np.float32) if in_img_type != np.uint8: img_ *= 255. # convert if only_y: rlt = np.dot(img_ , [24.966, 128.553, 65.481]) / 255.0 + 16.0 else: rlt = np.matmul(img_ , [[24.966, 112.0, -18.214], [128.553, -74.203, -93.786], [65.481, -37.797, 112.0]]) / 255.0 + [16, 128, 128] # to make ycrcb like cv2 # rlt = rlt[:, :, (0, 2, 1)] if in_img_type == np.uint8: rlt = rlt.round() else: rlt /= 255. if separate: rlt = rlt.astype(in_img_type) # y, cb, cr return rlt[:, :, 0], rlt[:, :, 1], rlt[:, :, 2] else: return rlt.astype(in_img_type) ''' def ycbcr2rgb_(img, only_y=True): """same as matlab ycbcr2rgb (Note: this implementation is the original from BasicSR, but appears to be for ycrcb, like cv2) Input: uint8, [0, 255] float, [0, 1] """ in_img_type = img.dtype img_ = img.astype(np.float32) if in_img_type != np.uint8: img_ *= 255. # to make ycrcb like cv2 # rlt = rlt[:, :, (0, 2, 1)] # convert # original (for ycrcb): rlt = np.matmul(img_ , [[0.00456621, 0.00456621, 0.00456621], [0, -0.00153632, 0.00791071], [0.00625893, -0.00318811, 0]]) * 255.0 + [-222.921, 135.576, -276.836] #alternative conversion: # xform = np.array([[1, 0, 1.402], [1, -0.34414, -.71414], [1, 1.772, 0]]) # img_[:, :, [1, 2]] -= 128 # rlt = img_.dot(xform.T) np.putmask(rlt, rlt > 255, 255) np.putmask(rlt, rlt < 0, 0) if in_img_type == np.uint8: rlt = rlt.round() else: rlt /= 255. return rlt.astype(in_img_type) ''' def ycbcr2rgb(img, only_y=True): ''' bgr version of matlab ycbcr2rgb Python opencv library (cv2) cv2.COLOR_YCrCb2BGR has different parameters to MATLAB color convertion. Input: uint8, [0, 255] float, [0, 1] ''' in_img_type = img.dtype img_ = img.astype(np.float32) if in_img_type != np.uint8: img_ *= 255. # to make ycrcb like cv2 # rlt = rlt[:, :, (0, 2, 1)] # convert mat = np.array([[24.966, 128.553, 65.481],[112, -74.203, -37.797], [-18.214, -93.786, 112.0]]) mat = np.linalg.inv(mat.T) * 255 offset = np.array([[[16, 128, 128]]]) rlt = np.dot((img_ - offset), mat) rlt = np.clip(rlt, 0, 255) ## rlt = np.rint(rlt).astype('uint8') if in_img_type == np.uint8: rlt = rlt.round() else: rlt /= 255. return rlt.astype(in_img_type) ''' #TODO: TMP RGB version, to check (PIL) def rgb2ycbcr(img_rgb): ## the range of img_rgb should be (0, 1) img_y = 0.257 * img_rgb[:, :, 0] + 0.504 * img_rgb[:, :, 1] + 0.098 * img_rgb[:, :, 2] + 16 / 255.0 img_cb = -0.148 * img_rgb[:, :, 0] - 0.291 * img_rgb[:, :, 1] + 0.439 * img_rgb[:, :, 2] + 128 / 255.0 img_cr = 0.439 * img_rgb[:, :, 0] - 0.368 * img_rgb[:, :, 1] - 0.071 * img_rgb[:, :, 2] + 128 / 255.0 return img_y, img_cb, img_cr #TODO: TMP RGB version, to check (PIL) def ycbcr2rgb(img_ycbcr): ## the range of img_ycbcr should be (0, 1) img_r = 1.164 * (img_ycbcr[:, :, 0] - 16 / 255.0) + 1.596 * (img_ycbcr[:, :, 2] - 128 / 255.0) img_g = 1.164 * (img_ycbcr[:, :, 0] - 16 / 255.0) - 0.392 * (img_ycbcr[:, :, 1] - 128 / 255.0) - 0.813 * (img_ycbcr[:, :, 2] - 128 / 255.0) img_b = 1.164 * (img_ycbcr[:, :, 0] - 16 / 255.0) + 2.017 * (img_ycbcr[:, :, 1] - 128 / 255.0) img_r = img_r[:, :, np.newaxis] img_g = img_g[:, :, np.newaxis] img_b = img_b[:, :, np.newaxis] img_rgb = np.concatenate((img_r, img_g, img_b), 2) return img_rgb ''' def modcrop(img_in, scale): # img_in: Numpy, HWC or HW img = np.copy(img_in) if img.ndim == 2: H, W = img.shape H_r, W_r = H % scale, W % scale img = img[:H - H_r, :W - W_r] elif img.ndim == 3: H, W, C = img.shape H_r, W_r = H % scale, W % scale img = img[:H - H_r, :W - W_r, :] else: raise ValueError('Wrong img ndim: [{:d}].'.format(img.ndim)) return img #TODO: this should probably be elsewhere (augmentations.py) def augment(img_list, hflip=True, rot=True): # horizontal flip OR rotate hflip = hflip and random.random() < 0.5 vflip = rot and random.random() < 0.5 rot90 = rot and random.random() < 0.5 #rot90n = rot and random.random() < 0.5 def _augment(img): if hflip: img = np.flip(img, axis=1) #img[:, ::-1, :] if vflip: img = np.flip(img, axis=0) #img[::-1, :, :] #if rot90: img = img.transpose(1, 0, 2) if rot90: img = np.rot90(img, 1) #90 degrees # In PIL: img.transpose(Image.ROTATE_90) #if rot90n: img = np.rot90(img, -1) #-90 degrees return img return [_augment(img) for img in img_list] #################### # Normalization functions #################### #TODO: Could also automatically detect the possible range with min and max, like in def ssim() def denorm(x, min_max=(-1.0, 1.0)): ''' Denormalize from [-1,1] range to [0,1] formula: xi' = (xi - mu)/sigma Example: "out = (x + 1.0) / 2.0" for denorm range (-1,1) to (0,1) for use with proper act in Generator output (ie. tanh) ''' out = (x - min_max[0]) / (min_max[1] - min_max[0]) if isinstance(x, torch.Tensor): return out.clamp(0, 1) elif isinstance(x, np.ndarray): return np.clip(out, 0, 1) else: raise TypeError("Got unexpected object type, expected torch.Tensor or \ np.ndarray") def norm(x): #Normalize (z-norm) from [0,1] range to [-1,1] out = (x - 0.5) * 2.0 if isinstance(x, torch.Tensor): return out.clamp(-1, 1) elif isinstance(x, np.ndarray): return np.clip(out, -1, 1) else: raise TypeError("Got unexpected object type, expected torch.Tensor or \ np.ndarray") #################### # np and tensor conversions #################### #2tensor def np2tensor(img, bgr2rgb=True, data_range=1., normalize=False, change_range=True, add_batch=True): """ Converts a numpy image array into a Tensor array. Parameters: img (numpy array): the input image numpy array add_batch (bool): choose if new tensor needs batch dimension added """ if not isinstance(img, np.ndarray): #images expected to be uint8 -> 255 raise TypeError("Got unexpected object type, expected np.ndarray") #check how many channels the image has, then condition, like in my BasicSR. ie. RGB, RGBA, Gray #if bgr2rgb: #img = img[:, :, [2, 1, 0]] #BGR to RGB -> in numpy, if using OpenCV, else not needed. Only if image has colors. if change_range: if np.issubdtype(img.dtype, np.integer): info = np.iinfo elif np.issubdtype(img.dtype, np.floating): info = np.finfo img = img*data_range/info(img.dtype).max #uint8 = /255 img = torch.from_numpy(np.ascontiguousarray(np.transpose(img, (2, 0, 1)))).float() #"HWC to CHW" and "numpy to tensor" if bgr2rgb: if img.shape[0] == 3: #RGB #BGR to RGB -> in tensor, if using OpenCV, else not needed. Only if image has colors. img = bgr_to_rgb(img) elif img.shape[0] == 4: #RGBA #BGR to RGB -> in tensor, if using OpenCV, else not needed. Only if image has colors.) img = bgra_to_rgba(img) if add_batch: img.unsqueeze_(0) # Add fake batch dimension = 1 . squeeze() will remove the dimensions of size 1 if normalize: img = norm(img) return img #2np def tensor2np(img, rgb2bgr=True, remove_batch=True, data_range=255, denormalize=False, change_range=True, imtype=np.uint8): """ Converts a Tensor array into a numpy image array. Parameters: img (tensor): the input image tensor array 4D(B,(3/1),H,W), 3D(C,H,W), or 2D(H,W), any range, RGB channel order remove_batch (bool): choose if tensor of shape BCHW needs to be squeezed denormalize (bool): Used to denormalize from [-1,1] range back to [0,1] imtype (type): the desired type of the converted numpy array (np.uint8 default) Output: img (np array): 3D(H,W,C) or 2D(H,W), [0,255], np.uint8 (default) """ if not isinstance(img, torch.Tensor): raise TypeError("Got unexpected object type, expected torch.Tensor") n_dim = img.dim() #TODO: Check: could denormalize here in tensor form instead, but end result is the same img = img.float().cpu() if n_dim == 4 or n_dim == 3: #if n_dim == 4, has to convert to 3 dimensions, either removing batch or by creating a grid if n_dim == 4 and remove_batch: if img.shape[0] > 1: # leave only the first image in the batch img = img[0,...] else: # remove a fake batch dimension img = img.squeeze() # squeeze removes batch and channel of grayscale images (dimensions = 1) if len(img.shape) < 3: #add back the lost channel dimension img = img.unsqueeze(dim=0) # convert images in batch (BCHW) to a grid of all images (C B*H B*W) else: n_img = len(img) img = make_grid(img, nrow=int(math.sqrt(n_img)), normalize=False) if img.shape[0] == 3 and rgb2bgr: #RGB #RGB to BGR -> in tensor, if using OpenCV, else not needed. Only if image has colors. img_np = rgb_to_bgr(img).numpy() elif img.shape[0] == 4 and rgb2bgr: #RGBA #RGBA to BGRA -> in tensor, if using OpenCV, else not needed. Only if image has colors. img_np = rgba_to_bgra(img).numpy() else: img_np = img.numpy() img_np = np.transpose(img_np, (1, 2, 0)) # "CHW to HWC" -> # HWC, BGR elif n_dim == 2: img_np = img.numpy() else: raise TypeError( 'Only support 4D, 3D and 2D tensor. But received with dimension: {:d}'.format(n_dim)) #if rgb2bgr: #img_np = img_np[[2, 1, 0], :, :] #RGB to BGR -> in numpy, if using OpenCV, else not needed. Only if image has colors. #TODO: Check: could denormalize in the begining in tensor form instead if denormalize: img_np = denorm(img_np) #denormalize if needed if change_range: img_np = np.clip(data_range*img_np,0,data_range).round() #clip to the data_range # Important. Unlike matlab, numpy.unit8() WILL NOT round by default. #has to be in range (0,255) before changing to np.uint8, else np.float32 return img_np.astype(imtype) #################### # Prepare Images #################### # https://github.com/sunreef/BlindSR/blob/master/src/image_utils.py def patchify_tensor(features, patch_size, overlap=10): batch_size, channels, height, width = features.size() effective_patch_size = patch_size - overlap n_patches_height = (height // effective_patch_size) n_patches_width = (width // effective_patch_size) if n_patches_height * effective_patch_size < height: n_patches_height += 1 if n_patches_width * effective_patch_size < width: n_patches_width += 1 patches = [] for b in range(batch_size): for h in range(n_patches_height): for w in range(n_patches_width): patch_start_height = min(h * effective_patch_size, height - patch_size) patch_start_width = min(w * effective_patch_size, width - patch_size) patches.append(features[b:b+1, :, patch_start_height: patch_start_height + patch_size, patch_start_width: patch_start_width + patch_size]) return torch.cat(patches, 0) def recompose_tensor(patches, full_height, full_width, overlap=10): batch_size, channels, patch_size, _ = patches.size() effective_patch_size = patch_size - overlap n_patches_height = (full_height // effective_patch_size) n_patches_width = (full_width // effective_patch_size) if n_patches_height * effective_patch_size < full_height: n_patches_height += 1 if n_patches_width * effective_patch_size < full_width: n_patches_width += 1 n_patches = n_patches_height * n_patches_width if batch_size % n_patches != 0: print("Error: The number of patches provided to the recompose function does not match the number of patches in each image.") final_batch_size = batch_size // n_patches blending_in = torch.linspace(0.1, 1.0, overlap) blending_out = torch.linspace(1.0, 0.1, overlap) middle_part = torch.ones(patch_size - 2 * overlap) blending_profile = torch.cat([blending_in, middle_part, blending_out], 0) horizontal_blending = blending_profile[None].repeat(patch_size, 1) vertical_blending = blending_profile[:, None].repeat(1, patch_size) blending_patch = horizontal_blending * vertical_blending blending_image = torch.zeros(1, channels, full_height, full_width) for h in range(n_patches_height): for w in range(n_patches_width): patch_start_height = min(h * effective_patch_size, full_height - patch_size) patch_start_width = min(w * effective_patch_size, full_width - patch_size) blending_image[0, :, patch_start_height: patch_start_height + patch_size, patch_start_width: patch_start_width + patch_size] += blending_patch[None] recomposed_tensor = torch.zeros(final_batch_size, channels, full_height, full_width) if patches.is_cuda: blending_patch = blending_patch.cuda() blending_image = blending_image.cuda() recomposed_tensor = recomposed_tensor.cuda() patch_index = 0 for b in range(final_batch_size): for h in range(n_patches_height): for w in range(n_patches_width): patch_start_height = min(h * effective_patch_size, full_height - patch_size) patch_start_width = min(w * effective_patch_size, full_width - patch_size) recomposed_tensor[b, :, patch_start_height: patch_start_height + patch_size, patch_start_width: patch_start_width + patch_size] += patches[patch_index] * blending_patch patch_index += 1 recomposed_tensor /= blending_image return recomposed_tensor #TODO: imresize could be an independent file (imresize.py) #################### # Matlab imresize #################### # These next functions are all interpolation methods. x is the distance from the left pixel center def cubic(x): absx = torch.abs(x) absx2 = absx**2 absx3 = absx**3 return (1.5*absx3 - 2.5*absx2 + 1) * ((absx <= 1).type_as(absx)) + \ (-0.5*absx3 + 2.5*absx2 - 4*absx + 2) * (((absx > 1)*(absx <= 2)).type_as(absx)) def box(x): return ((-0.5 <= x) & (x < 0.5)) * 1.0 def linear(x): return (x + 1) * ((-1 <= x) & (x < 0)) + (1 - x) * ((0 <= x) & (x <= 1)) def lanczos2(x): return (((torch.sin(math.pi*x) * torch.sin(math.pi*x/2) + torch.finfo(torch.float32).eps) / ((math.pi**2 * x**2 / 2) + torch.finfo(torch.float32).eps)) * (torch.abs(x) < 2)) def lanczos3(x): return (((torch.sin(math.pi*x) * torch.sin(math.pi*x/3) + torch.finfo(torch.float32).eps) / ((math.pi**2 * x**2 / 3) + torch.finfo(torch.float32).eps)) * (torch.abs(x) < 3)) def calculate_weights_indices(in_length, out_length, scale, kernel, kernel_width, antialiasing): if (scale < 1) and (antialiasing): # Use a modified kernel to simultaneously interpolate and antialias- larger kernel width kernel_width = kernel_width / scale # Output-space coordinates x = torch.linspace(1, out_length, out_length) # Input-space coordinates. Calculate the inverse mapping such that 0.5 # in output space maps to 0.5 in input space, and 0.5+scale in output # space maps to 1.5 in input space. u = x / scale + 0.5 * (1 - 1 / scale) # What is the left-most pixel that can be involved in the computation? left = torch.floor(u - kernel_width / 2) # What is the maximum number of pixels that can be involved in the # computation? Note: it's OK to use an extra pixel here; if the # corresponding weights are all zero, it will be eliminated at the end # of this function. P = math.ceil(kernel_width) + 2 # The indices of the input pixels involved in computing the k-th output # pixel are in row k of the indices matrix. indices = left.view(out_length, 1).expand(out_length, P) + torch.linspace(0, P - 1, P).view( 1, P).expand(out_length, P) # The weights used to compute the k-th output pixel are in row k of the # weights matrix. distance_to_center = u.view(out_length, 1).expand(out_length, P) - indices # apply kernel if (scale < 1) and (antialiasing): weights = scale * kernel(distance_to_center * scale) else: weights = kernel(distance_to_center) # Normalize the weights matrix so that each row sums to 1. weights_sum = torch.sum(weights, 1).view(out_length, 1) weights = weights / weights_sum.expand(out_length, P) # If a column in weights is all zero, get rid of it. only consider the first and last column. weights_zero_tmp = torch.sum((weights == 0), 0) if not math.isclose(weights_zero_tmp[0], 0, rel_tol=1e-6): indices = indices.narrow(1, 1, P - 2) weights = weights.narrow(1, 1, P - 2) if not math.isclose(weights_zero_tmp[-1], 0, rel_tol=1e-6): indices = indices.narrow(1, 0, P - 2) weights = weights.narrow(1, 0, P - 2) weights = weights.contiguous() indices = indices.contiguous() sym_len_s = -indices.min() + 1 sym_len_e = indices.max() - in_length indices = indices + sym_len_s - 1 return weights, indices, int(sym_len_s), int(sym_len_e) def imresize(img, scale, antialiasing=True, interpolation=None): # The scale should be the same for H and W # input: img: CHW RGB [0,1] # output: CHW RGB [0,1] w/o round in_C, in_H, in_W = img.size() out_C, out_H, out_W = in_C, math.ceil(in_H * scale), math.ceil(in_W * scale) # Choose interpolation method, each method has the matching kernel size kernel, kernel_width = { "cubic": (cubic, 4.0), "lanczos2": (lanczos2, 4.0), "lanczos3": (lanczos3, 6.0), "box": (box, 1.0), "linear": (linear, 2.0), None: (cubic, 4.0) # set default interpolation method as cubic }.get(interpolation) # Return the desired dimension order for performing the resize. The # strategy is to perform the resize first along the dimension with the # smallest scale factor. # Now we do not support this. # get weights and indices weights_H, indices_H, sym_len_Hs, sym_len_He = calculate_weights_indices( in_H, out_H, scale, kernel, kernel_width, antialiasing) weights_W, indices_W, sym_len_Ws, sym_len_We = calculate_weights_indices( in_W, out_W, scale, kernel, kernel_width, antialiasing) # process H dimension # symmetric copying img_aug = torch.FloatTensor(in_C, in_H + sym_len_Hs + sym_len_He, in_W) img_aug.narrow(1, sym_len_Hs, in_H).copy_(img) sym_patch = img[:, :sym_len_Hs, :] inv_idx = torch.arange(sym_patch.size(1) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(1, inv_idx) img_aug.narrow(1, 0, sym_len_Hs).copy_(sym_patch_inv) sym_patch = img[:, -sym_len_He:, :] inv_idx = torch.arange(sym_patch.size(1) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(1, inv_idx) img_aug.narrow(1, sym_len_Hs + in_H, sym_len_He).copy_(sym_patch_inv) out_1 = torch.FloatTensor(in_C, out_H, in_W) kernel_width = weights_H.size(1) for i in range(out_H): idx = int(indices_H[i][0]) out_1[0, i, :] = img_aug[0, idx:idx + kernel_width, :].transpose(0, 1).mv(weights_H[i]) out_1[1, i, :] = img_aug[1, idx:idx + kernel_width, :].transpose(0, 1).mv(weights_H[i]) out_1[2, i, :] = img_aug[2, idx:idx + kernel_width, :].transpose(0, 1).mv(weights_H[i]) # process W dimension # symmetric copying out_1_aug = torch.FloatTensor(in_C, out_H, in_W + sym_len_Ws + sym_len_We) out_1_aug.narrow(2, sym_len_Ws, in_W).copy_(out_1) sym_patch = out_1[:, :, :sym_len_Ws] inv_idx = torch.arange(sym_patch.size(2) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(2, inv_idx) out_1_aug.narrow(2, 0, sym_len_Ws).copy_(sym_patch_inv) sym_patch = out_1[:, :, -sym_len_We:] inv_idx = torch.arange(sym_patch.size(2) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(2, inv_idx) out_1_aug.narrow(2, sym_len_Ws + in_W, sym_len_We).copy_(sym_patch_inv) out_2 = torch.FloatTensor(in_C, out_H, out_W) kernel_width = weights_W.size(1) for i in range(out_W): idx = int(indices_W[i][0]) out_2[0, :, i] = out_1_aug[0, :, idx:idx + kernel_width].mv(weights_W[i]) out_2[1, :, i] = out_1_aug[1, :, idx:idx + kernel_width].mv(weights_W[i]) out_2[2, :, i] = out_1_aug[2, :, idx:idx + kernel_width].mv(weights_W[i]) return out_2 def imresize_np(img, scale, antialiasing=True, interpolation=None): # Now the scale should be the same for H and W # input: img: Numpy, HWC BGR [0,1] # output: HWC BGR [0,1] w/o round change_range = False if img.max() > 1: img_type = img.dtype if np.issubdtype(img_type, np.integer): info = np.iinfo elif np.issubdtype(img_type, np.floating): info = np.finfo img = img/info(img_type).max change_range = True img = torch.from_numpy(img) in_H, in_W, in_C = img.size() out_C, out_H, out_W = in_C, math.ceil(in_H * scale), math.ceil(in_W * scale) # Choose interpolation method, each method has the matching kernel size kernel, kernel_width = { "cubic": (cubic, 4.0), "lanczos2": (lanczos2, 4.0), "lanczos3": (lanczos3, 6.0), "box": (box, 1.0), "linear": (linear, 2.0), None: (cubic, 4.0) # set default interpolation method as cubic }.get(interpolation) # Return the desired dimension order for performing the resize. The # strategy is to perform the resize first along the dimension with the # smallest scale factor. # Now we do not support this. # get weights and indices weights_H, indices_H, sym_len_Hs, sym_len_He = calculate_weights_indices( in_H, out_H, scale, kernel, kernel_width, antialiasing) weights_W, indices_W, sym_len_Ws, sym_len_We = calculate_weights_indices( in_W, out_W, scale, kernel, kernel_width, antialiasing) # process H dimension # symmetric copying img_aug = torch.FloatTensor(in_H + sym_len_Hs + sym_len_He, in_W, in_C) img_aug.narrow(0, sym_len_Hs, in_H).copy_(img) sym_patch = img[:sym_len_Hs, :, :] inv_idx = torch.arange(sym_patch.size(0) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(0, inv_idx) img_aug.narrow(0, 0, sym_len_Hs).copy_(sym_patch_inv) sym_patch = img[-sym_len_He:, :, :] inv_idx = torch.arange(sym_patch.size(0) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(0, inv_idx) img_aug.narrow(0, sym_len_Hs + in_H, sym_len_He).copy_(sym_patch_inv) out_1 = torch.FloatTensor(out_H, in_W, in_C) kernel_width = weights_H.size(1) for i in range(out_H): idx = int(indices_H[i][0]) out_1[i, :, 0] = img_aug[idx:idx + kernel_width, :, 0].transpose(0, 1).mv(weights_H[i]) out_1[i, :, 1] = img_aug[idx:idx + kernel_width, :, 1].transpose(0, 1).mv(weights_H[i]) out_1[i, :, 2] = img_aug[idx:idx + kernel_width, :, 2].transpose(0, 1).mv(weights_H[i]) # process W dimension # symmetric copying out_1_aug = torch.FloatTensor(out_H, in_W + sym_len_Ws + sym_len_We, in_C) out_1_aug.narrow(1, sym_len_Ws, in_W).copy_(out_1) sym_patch = out_1[:, :sym_len_Ws, :] inv_idx = torch.arange(sym_patch.size(1) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(1, inv_idx) out_1_aug.narrow(1, 0, sym_len_Ws).copy_(sym_patch_inv) sym_patch = out_1[:, -sym_len_We:, :] inv_idx = torch.arange(sym_patch.size(1) - 1, -1, -1).long() sym_patch_inv = sym_patch.index_select(1, inv_idx) out_1_aug.narrow(1, sym_len_Ws + in_W, sym_len_We).copy_(sym_patch_inv) out_2 = torch.FloatTensor(out_H, out_W, in_C) kernel_width = weights_W.size(1) for i in range(out_W): idx = int(indices_W[i][0]) out_2[:, i, 0] = out_1_aug[:, idx:idx + kernel_width, 0].mv(weights_W[i]) out_2[:, i, 1] = out_1_aug[:, idx:idx + kernel_width, 1].mv(weights_W[i]) out_2[:, i, 2] = out_1_aug[:, idx:idx + kernel_width, 2].mv(weights_W[i]) out_2 = out_2.numpy().clip(0,1) if change_range: out_2 = out_2*info(img_type).max #uint8 = 255 out_2 = out_2.astype(img_type) return out_2 if __name__ == '__main__': # test imresize function # read images img = cv2.imread('test.png') img = img * 1.0 / 255 img = torch.from_numpy(np.transpose(img[:, :, [2, 1, 0]], (2, 0, 1))).float() # imresize scale = 1 / 4 import time total_time = 0 for i in range(10): start_time = time.time() rlt = imresize(img, scale, antialiasing=True) use_time = time.time() - start_time total_time += use_time print('average time: {}'.format(total_time / 10)) import torchvision.utils torchvision.utils.save_image( (rlt * 255).round() / 255, 'rlt.png', nrow=1, padding=0, normalize=False)
mmedit/models/inpaintors/vic/common.py
33,981
get image path list from image folder get image path list from lmdb bgr version of matlab rgb2ycbcr Python opencv library (cv2) cv2.COLOR_BGR2YCrCb has different parameters with MATLAB color convertion. only_y: only return Y channel separate: if true, will returng the channels as separate images Input: uint8, [0, 255] float, [0, 1] cv2.cvtColor(img, cv2.COLOR_BGRA2BGR) has an issue removing the alpha channel, this gets rid of wrong transparent colors that can harm training Denormalize from [-1,1] range to [0,1] formula: xi' = (xi - mu)/sigma Example: "out = (x + 1.0) / 2.0" for denorm range (-1,1) to (0,1) for use with proper act in Generator output (ie. tanh) fix image channels to the expected number get image path list support lmdb or image files Converts a numpy image array into a Tensor array. Parameters: img (numpy array): the input image numpy array add_batch (bool): choose if new tensor needs batch dimension added Reads image using cv2 (rawpy if dng) or from lmdb by default (can also use using PIL instead of cv2) Arguments: out_nc: Desired number of channels fix_channels: changes the images to the desired number of channels Output: Numpy uint8, HWC, BGR, [0,255] by default same as matlab rgb2ycbcr only_y: only return Y channel Input: uint8, [0, 255] float, [0, 1] Converts a Tensor array into a numpy image array. Parameters: img (tensor): the input image tensor array 4D(B,(3/1),H,W), 3D(C,H,W), or 2D(H,W), any range, RGB channel order remove_batch (bool): choose if tensor of shape BCHW needs to be squeezed denormalize (bool): Used to denormalize from [-1,1] range back to [0,1] imtype (type): the desired type of the converted numpy array (np.uint8 default) Output: img (np array): 3D(H,W,C) or 2D(H,W), [0,255], np.uint8 (default) bgr version of matlab ycbcr2rgb Python opencv library (cv2) cv2.COLOR_YCrCb2BGR has different parameters to MATLAB color convertion. Input: uint8, [0, 255] float, [0, 1] BasicSR/codes/dataops/common.py (8-Nov-20) https://github.com/victorca25/BasicSR/blob/dev2/codes/dataops/common.py from dataops.colors import *from dataops.debug import tmp_vis, describe_numpy, describe_tensor Files & IO get image path list read images img if image is a DNG if image is a NPY numpy array else, if image can be read by cv2TODO: add variable detecting if cv2 is not available and try PIL instead elif: using PIL instead of OpenCV img = Image.open(path).convert('RGB') else: For other images unrecognized by cv2 import matplotlib.pyplot as plt img = (255*plt.imread(path)[:,:,:3]).astype('uint8') if not img: raise ValueError(f"Failed to read image: {path}") if image has only 2 dimensions, add "channel" dimension (1)img = img[..., np.newaxis] altimg = np.expand_dims(img, axis=2) special case: properly remove alpha channel remove all extra channels if alpha is expected, add solid alpha channel image processing process on numpy imageb, g, r, a = cv2.split((img*255).astype(np.uint8))return cv2.merge([b, g, r]).astype(np.float32)/255. conversion among BGR, gray and y Note: OpenCV uses inverted channels BGR, instead of RGB. If images are loaded with something other than OpenCV, check that the channels are in the correct order and use the alternative conversion functions.if in_c == 4 and tar_type == 'RGB-A': BGRA to BGR, remove alpha channelreturn [cv2.cvtColor(img, cv2.COLOR_BGRA2BGR) for img in img_list]return [bgra2rgb(img) for img in img_list] BGR to grayRGB to LABRGB to LAB BGR to y gray/y to BGR convert convert to make ycrcb like cv2 rlt = rlt[:, :, (0, 2, 1)] y, cb, cr to make ycrcb like cv2 rlt = rlt[:, :, (0, 2, 1)] convert rlt = np.rint(rlt).astype('uint8') img_in: Numpy, HWC or HWTODO: this should probably be elsewhere (augmentations.py) horizontal flip OR rotaterot90n = rot and random.random() < 0.5img[:, ::-1, :]img[::-1, :, :]if rot90: img = img.transpose(1, 0, 2)90 degrees In PIL: img.transpose(Image.ROTATE_90)if rot90n: img = np.rot90(img, -1) -90 degrees Normalization functionsTODO: Could also automatically detect the possible range with min and max, like in def ssim()Normalize (z-norm) from [0,1] range to [-1,1] np and tensor conversions2tensorimages expected to be uint8 -> 255check how many channels the image has, then condition, like in my BasicSR. ie. RGB, RGBA, Grayif bgr2rgb:img = img[:, :, [2, 1, 0]] BGR to RGB -> in numpy, if using OpenCV, else not needed. Only if image has colors.uint8 = /255"HWC to CHW" and "numpy to tensor"RGBBGR to RGB -> in tensor, if using OpenCV, else not needed. Only if image has colors.RGBABGR to RGB -> in tensor, if using OpenCV, else not needed. Only if image has colors.) Add fake batch dimension = 1 . squeeze() will remove the dimensions of size 12npTODO: Check: could denormalize here in tensor form instead, but end result is the sameif n_dim == 4, has to convert to 3 dimensions, either removing batch or by creating a grid leave only the first image in the batch remove a fake batch dimension squeeze removes batch and channel of grayscale images (dimensions = 1)add back the lost channel dimension convert images in batch (BCHW) to a grid of all images (C B*H B*W)RGBRGB to BGR -> in tensor, if using OpenCV, else not needed. Only if image has colors.RGBARGBA to BGRA -> in tensor, if using OpenCV, else not needed. Only if image has colors. "CHW to HWC" -> HWC, BGRif rgb2bgr:img_np = img_np[[2, 1, 0], :, :] RGB to BGR -> in numpy, if using OpenCV, else not needed. Only if image has colors.TODO: Check: could denormalize in the begining in tensor form insteaddenormalize if neededclip to the data_range Important. Unlike matlab, numpy.unit8() WILL NOT round by default.has to be in range (0,255) before changing to np.uint8, else np.float32 Prepare Images https://github.com/sunreef/BlindSR/blob/master/src/image_utils.pyTODO: imresize could be an independent file (imresize.py) Matlab imresize These next functions are all interpolation methods. x is the distance from the left pixel center Use a modified kernel to simultaneously interpolate and antialias- larger kernel width Output-space coordinates Input-space coordinates. Calculate the inverse mapping such that 0.5 in output space maps to 0.5 in input space, and 0.5+scale in output space maps to 1.5 in input space. What is the left-most pixel that can be involved in the computation? What is the maximum number of pixels that can be involved in the computation? Note: it's OK to use an extra pixel here; if the corresponding weights are all zero, it will be eliminated at the end of this function. The indices of the input pixels involved in computing the k-th output pixel are in row k of the indices matrix. The weights used to compute the k-th output pixel are in row k of the weights matrix. apply kernel Normalize the weights matrix so that each row sums to 1. If a column in weights is all zero, get rid of it. only consider the first and last column. The scale should be the same for H and W input: img: CHW RGB [0,1] output: CHW RGB [0,1] w/o round Choose interpolation method, each method has the matching kernel size set default interpolation method as cubic Return the desired dimension order for performing the resize. The strategy is to perform the resize first along the dimension with the smallest scale factor. Now we do not support this. get weights and indices process H dimension symmetric copying process W dimension symmetric copying Now the scale should be the same for H and W input: img: Numpy, HWC BGR [0,1] output: HWC BGR [0,1] w/o round Choose interpolation method, each method has the matching kernel size set default interpolation method as cubic Return the desired dimension order for performing the resize. The strategy is to perform the resize first along the dimension with the smallest scale factor. Now we do not support this. get weights and indices process H dimension symmetric copying process W dimension symmetric copyinguint8 = 255 test imresize function read images imresize
8,037
en
0.695498
# -*- coding: utf-8 -*- from numpy import NaN as npNaN from pandas import DataFrame, Series # from pandas_ta.overlap.ma import ma from .ma import ma from pandas_ta.utils import get_offset, verify_series def hilo(high, low, close, high_length=None, low_length=None, mamode=None, offset=None, **kwargs): """Indicator: Gann HiLo (HiLo)""" # Validate Arguments high = verify_series(high) low = verify_series(low) close = verify_series(close) high_length = int(high_length) if high_length and high_length > 0 else 13 low_length = int(low_length) if low_length and low_length > 0 else 21 mamode = mamode.lower() if isinstance(mamode, str) else "sma" offset = get_offset(offset) # Calculate Result m = close.size hilo = Series(npNaN, index=close.index) long = Series(npNaN, index=close.index) short = Series(npNaN, index=close.index) high_ma = ma(mamode, high, length=high_length) low_ma = ma(mamode, low, length=low_length) for i in range(1, m): if close.iloc[i] > high_ma.iloc[i - 1]: hilo.iloc[i] = long.iloc[i] = low_ma.iloc[i] elif close.iloc[i] < low_ma.iloc[i - 1]: hilo.iloc[i] = short.iloc[i] = high_ma.iloc[i] else: hilo.iloc[i] = hilo.iloc[i - 1] long.iloc[i] = short.iloc[i] = hilo.iloc[i - 1] # Offset if offset != 0: hilo = hilo.shift(offset) long = long.shift(offset) short = short.shift(offset) # Handle fills if "fillna" in kwargs: hilo.fillna(kwargs["fillna"], inplace=True) long.fillna(kwargs["fillna"], inplace=True) short.fillna(kwargs["fillna"], inplace=True) if "fill_method" in kwargs: hilo.fillna(method=kwargs["fill_method"], inplace=True) long.fillna(method=kwargs["fill_method"], inplace=True) short.fillna(method=kwargs["fill_method"], inplace=True) # Name & Category _props = f"_{high_length}_{low_length}" data = {f"HILO{_props}": hilo, f"HILOl{_props}": long, f"HILOs{_props}": short} df = DataFrame(data, index=close.index) df.name = f"HILO{_props}" df.category = "overlap" return df hilo.__doc__ = \ """Gann HiLo Activator(HiLo) The Gann High Low Activator Indicator was created by Robert Krausz in a 1998 issue of Stocks & Commodities Magazine. It is a moving average based trend indicator consisting of two different simple moving averages. The indicator tracks both curves (of the highs and the lows). The close of the bar defines which of the two gets plotted. Increasing high_length and decreasing low_length better for short trades, vice versa for long positions. Sources: https://www.sierrachart.com/index.php?page=doc/StudiesReference.php&ID=447&Name=Gann_HiLo_Activator https://www.tradingtechnologies.com/help/x-study/technical-indicator-definitions/simple-moving-average-sma/ https://www.tradingview.com/script/XNQSLIYb-Gann-High-Low/ Calculation: Default Inputs: high_length=13, low_length=21, mamode="sma" EMA = Exponential Moving Average HMA = Hull Moving Average SMA = Simple Moving Average # Default if "ema": high_ma = EMA(high, high_length) low_ma = EMA(low, low_length) elif "hma": high_ma = HMA(high, high_length) low_ma = HMA(low, low_length) else: # "sma" high_ma = SMA(high, high_length) low_ma = SMA(low, low_length) # Similar to Supertrend MA selection hilo = Series(npNaN, index=close.index) for i in range(1, m): if close.iloc[i] > high_ma.iloc[i - 1]: hilo.iloc[i] = low_ma.iloc[i] elif close.iloc[i] < low_ma.iloc[i - 1]: hilo.iloc[i] = high_ma.iloc[i] else: hilo.iloc[i] = hilo.iloc[i - 1] Args: high (pd.Series): Series of 'high's low (pd.Series): Series of 'low's close (pd.Series): Series of 'close's high_length (int): It's period. Default: 13 low_length (int): It's period. Default: 21 mamode (str): Options: 'sma' or 'ema'. Default: 'sma' offset (int): How many periods to offset the result. Default: 0 Kwargs: adjust (bool): Default: True presma (bool, optional): If True, uses SMA for initial value. fillna (value, optional): pd.DataFrame.fillna(value) fill_method (value, optional): Type of fill method Returns: pd.DataFrame: HILO (line), HILOl (long), HILOs (short) columns. """
pandas_ta/overlap/hilo.py
4,446
Indicator: Gann HiLo (HiLo) -*- coding: utf-8 -*- from pandas_ta.overlap.ma import ma Validate Arguments Calculate Result Offset Handle fills Name & Category
159
en
0.391075
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import sys if __name__ == '__main__': goods = [] while True: command = input(">>> ").lower() if command == 'exit': break elif command == 'add': name = input("Название товара: ") shop = input("Название магазина: ") price = float(input("Стоимость: ")) good = { 'name': name, 'shop': shop, 'price': price, } goods.append(good) # Отсортировать список в случае необходимости. if len(goods) > 1: goods.sort(key=lambda item: item.get('shop', '')) elif command == 'list': line = '+-{}-+-{}-+-{}-+-{}-+'.format( '-' * 4, '-' * 30, '-' * 20, '-' * 8 ) print(line) print( '| {:^4} | {:^30} | {:^20} | {:^8} |'.format( "№", "Название", "Магазин", "Цена" ) ) print(line) for idx, good in enumerate(goods, 1): print( '| {:>4} | {:<30} | {:<20} | {:>8} |'.format( idx, good.get('name', ''), good.get('shop', ''), good.get('price', 0) ) ) print(line) elif command.startswith('select '): parts = command.split(' ', maxsplit=1) shopName = parts[1] count = 0 for good in goods: if shopName == good.get('shop', shopName): count += 1 print( '{:>4}: {}'.format(count, good.get('name', '')) ) if count == 0: print("Такого магазина не существует либо нет товаров.") elif command == 'help': print("Список команд:\n") print("add - добавить товар;") print("list - вывести список товаров;") print("select <имя магазина> - запросить товары магазина;") print("help - отобразить справку;") print("exit - завершить работу с программой.") else: print(f"Неизвестная команда {command}", file=sys.stderr)
pythonProject/ind.py
2,745
!/usr/bin/env python3 -*- coding: utf-8 -*- Отсортировать список в случае необходимости.
88
ru
0.907529
# -*- coding: utf-8 -*- """ Created on Wed Jun 5 08:32:13 2019 @author: Thiago """ import numpy as np import pylab as pl #%% #Simulação de uma va def va_estoque(): p=np.array([0.1, 0.2, 0.6, 0.1]) x=np.random.rand() if 0 < x <= p[0]: return 1 elif p[0] < x <= p[0]+p[1]: return 2 elif p[0]+p[1] < x <= p[0]+p[1]+p[2]: return 3 elif p[0]+p[1]+p[2] < x <= 1.0: return 4 v = [va_estoque() for i in range(100000)] pl.hist(v,) pl.show() #%% #simulação estoque M, T, estoque, lucro = 3, 3, 10, 0 R = 10000 for i in range(R): Y=va_estoque() lucro += 20*min(estoque, Y) estoque -= max(0, estoque-Y) lucro -= 5*estoque if estoque<M: estoque += T lucro -= 10*T lucro /= R print(M, T, lucro, estoque) #%% #simulação Urna de Ehrenfest N, s = 100, [] for j in range(1000): v = [True for i in range(N)] for i in range(1000): k=np.random.choice(N) v[k] = not v[k] x = sum(v) / N s.append(x) pl.hist(s) #%% #Lei dos grandes números np.random.seed(0) S = [1, 2, 3, 4, 5, 6] n_vals = np.logspace(1, 5, num=200) s=[] for val in n_vals: np.random.seed(0) n = int(val) x = np.random.choice(S,n) p=sum(x==3)/n s.append([n,p]) s=np.array(s) pl.semilogx(s[:,1]) pl.axhline(1./len(S),c='r') #%% #processos ergodicos #%% ''' s = 3000 for n in [1,2,3,5,10,50,100,200,400,1000]: z=np.zeros(s) for k in range(n): x = np.random.uniform(-1, 1, s) z+=x x = z/np.sqrt(n) pl.figure(n) sns.distplot(y, bins=12, rug=True) pl.title('N = ' + str()) '''
aulas/05-06/variaveis_aleatorias.py
1,683
Created on Wed Jun 5 08:32:13 2019 @author: Thiago -*- coding: utf-8 -*-%%Simulação de uma va%%simulação estoque%%simulação Urna de Ehrenfest%%Lei dos grandes números%%processos ergodicos%%
193
pt
0.936932
"""Test GeoTIFF as process output.""" import numpy as np import numpy.ma as ma import os import pytest import rasterio from rasterio.io import MemoryFile from rio_cogeo.cogeo import cog_validate import shutil from tilematrix import Bounds import warnings import mapchete from mapchete.errors import MapcheteConfigError from mapchete.io import path_exists from mapchete.formats.default import gtiff from mapchete.tile import BufferedTilePyramid def test_output_data(mp_tmpdir): """Check GeoTIFF as output data.""" output_params = dict( grid="geodetic", format="GeoTIFF", path=mp_tmpdir, pixelbuffer=0, metatiling=1, bands=1, dtype="int16", delimiters=dict( bounds=Bounds(-180.0, -90.0, 180.0, 90.0), effective_bounds=Bounds(-180.439453125, -90.0, 180.439453125, 90.0), zoom=[5], process_bounds=Bounds(-180.0, -90.0, 180.0, 90.0), ), ) output = gtiff.OutputDataWriter(output_params) assert output.path == mp_tmpdir assert output.file_extension == ".tif" tp = BufferedTilePyramid("geodetic") tile = tp.tile(5, 5, 5) # get_path assert output.get_path(tile) == os.path.join(*[mp_tmpdir, "5", "5", "5" + ".tif"]) # prepare_path try: temp_dir = os.path.join(*[mp_tmpdir, "5", "5"]) output.prepare_path(tile) assert os.path.isdir(temp_dir) finally: shutil.rmtree(temp_dir, ignore_errors=True) # profile assert isinstance(output.profile(tile), dict) # write try: data = np.ones((1,) + tile.shape) * 128 output.write(tile, data) # tiles_exist assert output.tiles_exist(tile) # read data = output.read(tile) assert isinstance(data, np.ndarray) assert not data[0].mask.any() finally: shutil.rmtree(mp_tmpdir, ignore_errors=True) # read empty try: data = output.read(tile) assert isinstance(data, np.ndarray) assert data[0].mask.all() finally: shutil.rmtree(mp_tmpdir, ignore_errors=True) # empty try: empty = output.empty(tile) assert isinstance(empty, ma.MaskedArray) assert not empty.any() finally: shutil.rmtree(mp_tmpdir, ignore_errors=True) # deflate with predictor try: # with pytest.deprecated_call(): output_params.update(compress="deflate", predictor=2) output = gtiff.OutputDataWriter(output_params) assert output.profile(tile)["compress"] == "deflate" assert output.profile(tile)["predictor"] == 2 finally: shutil.rmtree(mp_tmpdir, ignore_errors=True) # using deprecated "compression" property try: with pytest.deprecated_call(): output_params.update(compression="deflate", predictor=2) output = gtiff.OutputDataWriter(output_params) assert output.profile(tile)["compress"] == "deflate" assert output.profile(tile)["predictor"] == 2 finally: shutil.rmtree(mp_tmpdir, ignore_errors=True) def test_for_web(client, mp_tmpdir): """Send GTiff via flask.""" tile_base_url = "/wmts_simple/1.0.0/cleantopo_br/default/WGS84/" for url in ["/"]: response = client.get(url) assert response.status_code == 200 for url in [ tile_base_url + "5/30/62.tif", tile_base_url + "5/30/63.tif", tile_base_url + "5/31/62.tif", tile_base_url + "5/31/63.tif", ]: response = client.get(url) assert response.status_code == 200 img = response.data with warnings.catch_warnings(): warnings.simplefilter("ignore") with MemoryFile(img) as memfile: with memfile.open() as dataset: assert dataset.read().any() def test_input_data(mp_tmpdir, cleantopo_br): """Check GeoTIFF proces output as input data.""" with mapchete.open(cleantopo_br.path) as mp: tp = BufferedTilePyramid("geodetic") # TODO tile with existing but empty data tile = tp.tile(5, 5, 5) output_params = dict( grid="geodetic", format="GeoTIFF", path=mp_tmpdir, pixelbuffer=0, metatiling=1, bands=2, dtype="int16", delimiters=dict( bounds=Bounds(-180.0, -90.0, 180.0, 90.0), effective_bounds=Bounds(-180.439453125, -90.0, 180.439453125, 90.0), zoom=[5], process_bounds=Bounds(-180.0, -90.0, 180.0, 90.0), ), ) output = gtiff.OutputDataWriter(output_params) with output.open(tile, mp) as input_tile: for data in [ input_tile.read(), input_tile.read(1), input_tile.read([1]), # TODO assert valid indexes are passed input_tile.read([1, 2]) ]: assert isinstance(data, ma.masked_array) assert input_tile.is_empty() # open without resampling with output.open(tile, mp) as input_tile: pass def test_write_geotiff_tags(mp_tmpdir, cleantopo_br, write_rasterfile_tags_py): """Pass on metadata tags from user process to rasterio.""" conf = dict(**cleantopo_br.dict) conf.update(process=write_rasterfile_tags_py) with mapchete.open(conf) as mp: for tile in mp.get_process_tiles(): data, tags = mp.execute(tile) assert data.any() assert isinstance(tags, dict) mp.write(process_tile=tile, data=(data, tags)) # read data out_path = mp.config.output.get_path(tile) with rasterio.open(out_path) as src: assert "filewide_tag" in src.tags() assert src.tags()["filewide_tag"] == "value" assert "band_tag" in src.tags(1) assert src.tags(1)["band_tag"] == "True" @pytest.mark.remote def test_s3_write_output_data(gtiff_s3, s3_example_tile, mp_s3_tmpdir): """Write and read output.""" with mapchete.open(gtiff_s3.dict) as mp: process_tile = mp.config.process_pyramid.tile(*s3_example_tile) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(tile=process_tile.id) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() def test_output_single_gtiff(output_single_gtiff): tile_id = (5, 3, 7) with mapchete.open(output_single_gtiff.path) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert os.path.isfile(mp.config.output.path) # error on existing file with pytest.raises(MapcheteConfigError): mapchete.open(output_single_gtiff.path) # overwrite existing file with mapchete.open(output_single_gtiff.path, mode="overwrite") as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(tile=process_tile.id) # check if tile exists assert mp.config.output.tiles_exist(process_tile) assert mp.config.output.tiles_exist( output_tile=mp.config.output_pyramid.intersecting(process_tile)[0] ) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() def test_output_single_gtiff_errors(output_single_gtiff): # single gtiff does not work on multiple zoom levels with pytest.raises(ValueError): mapchete.open(dict(output_single_gtiff.dict, zoom_levels=[5, 6])) # provide either process_tile or output_tile with mapchete.open(output_single_gtiff.path) as mp: tile = mp.config.process_pyramid.tile(5, 3, 7) with pytest.raises(ValueError): mp.config.output.tiles_exist(process_tile=tile, output_tile=tile) def test_output_single_gtiff_pixelbuffer(output_single_gtiff): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff.dict, output=dict(output_single_gtiff.dict["output"], pixelbuffer=5), ), ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(tile=process_tile.id) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() def test_output_single_gtiff_compression(output_single_gtiff): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff.dict, output=dict(output_single_gtiff.dict["output"], compress="deflate"), ), ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) assert "compress" in mp.config.output.profile() assert mp.config.output.profile()["compress"] == "deflate" mp.batch_process(tile=process_tile.id) with rasterio.open(mp.config.output.path) as src: assert src.profile["compress"] == "deflate" def test_output_single_gtiff_overviews(output_single_gtiff): # overwrite existing file with mapchete.open( dict( output_single_gtiff.dict, output=dict( output_single_gtiff.dict["output"], overviews=True, overviews_resampling="bilinear", ), ), ) as mp: tile_id = (5, 3, 7) process_tile = mp.config.process_pyramid.tile(*tile_id) mp.batch_process(tile=process_tile.id) with rasterio.open(mp.config.output.path) as src: assert src.overviews(1) assert src.tags().get("OVR_RESAMPLING_ALG").lower() == "bilinear" for o in [1, 2, 4, 8]: a = src.read( masked=True, out_shape=(1, int(src.height / o), int(src.width / o)) ) assert not a.mask.all() @pytest.mark.remote def test_output_single_gtiff_s3(output_single_gtiff, mp_s3_tmpdir): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff.dict, output=dict( output_single_gtiff.dict["output"], path=os.path.join(mp_s3_tmpdir, "temp.tif"), ), ) ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert path_exists(mp.config.output.path) @pytest.mark.remote def test_output_single_gtiff_s3_tempfile(output_single_gtiff, mp_s3_tmpdir): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff.dict, output=dict( output_single_gtiff.dict["output"], path=os.path.join(mp_s3_tmpdir, "temp.tif"), in_memory=False, ), ) ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert path_exists(mp.config.output.path) def test_output_single_gtiff_cog(output_single_gtiff_cog): tile_id = (5, 3, 7) with mapchete.open(output_single_gtiff_cog.dict) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert path_exists(mp.config.output.path) assert cog_validate(mp.config.output.path, strict=True) def test_output_single_gtiff_cog_tempfile(output_single_gtiff_cog): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff_cog.dict, output=dict(output_single_gtiff_cog.dict["output"], in_memory=False), ) ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert path_exists(mp.config.output.path) assert cog_validate(mp.config.output.path, strict=True) @pytest.mark.remote def test_output_single_gtiff_cog_s3(output_single_gtiff_cog, mp_s3_tmpdir): tile_id = (5, 3, 7) with mapchete.open( dict( output_single_gtiff_cog.dict, output=dict( output_single_gtiff_cog.dict["output"], path=os.path.join(mp_s3_tmpdir, "cog.tif"), ), ) ) as mp: process_tile = mp.config.process_pyramid.tile(*tile_id) # basic functions assert mp.config.output.profile() assert mp.config.output.empty(process_tile).mask.all() assert mp.config.output.get_path(process_tile) # check if tile exists assert not mp.config.output.tiles_exist(process_tile) # write mp.batch_process(multi=2) # check if tile exists assert mp.config.output.tiles_exist(process_tile) # read again, this time with data data = mp.config.output.read(process_tile) assert isinstance(data, np.ndarray) assert not data[0].mask.all() # write empty array data = ma.masked_array( data=np.ones(process_tile.shape), mask=np.ones(process_tile.shape), ) mp.config.output.write(process_tile, data) assert path_exists(mp.config.output.path) assert cog_validate(mp.config.output.path, strict=True)
test/test_formats_geotiff.py
17,980
Send GTiff via flask. Check GeoTIFF proces output as input data. Check GeoTIFF as output data. Write and read output. Pass on metadata tags from user process to rasterio. Test GeoTIFF as process output. get_path prepare_path profile write tiles_exist read read empty empty deflate with predictor with pytest.deprecated_call(): using deprecated "compression" property TODO tile with existing but empty data TODO assert valid indexes are passed input_tile.read([1, 2]) open without resampling read data basic functions check if tile exists write check if tile exists read again, this time with data basic functions check if tile exists write check if tile exists read again, this time with data write empty array error on existing file overwrite existing file write check if tile exists read again, this time with data single gtiff does not work on multiple zoom levels provide either process_tile or output_tile basic functions check if tile exists write check if tile exists read again, this time with data overwrite existing file basic functions check if tile exists write check if tile exists read again, this time with data write empty array basic functions check if tile exists write check if tile exists read again, this time with data write empty array basic functions check if tile exists write check if tile exists read again, this time with data write empty array basic functions check if tile exists write check if tile exists read again, this time with data write empty array basic functions check if tile exists write check if tile exists read again, this time with data write empty array
1,602
en
0.604066
# Copyright (c) 2012, Cloudscaling # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import ast import re import pep8 """ Guidelines for writing new hacking checks - Use only for Manila specific tests. OpenStack general tests should be submitted to the common 'hacking' module. - Pick numbers in the range M3xx. Find the current test with the highest allocated number and then pick the next value. - Keep the test method code in the source file ordered based on the M3xx value. - List the new rule in the top level HACKING.rst file - Add test cases for each new rule to manila/tests/test_hacking.py """ UNDERSCORE_IMPORT_FILES = [] log_translation = re.compile( r"(.)*LOG\.(audit|error|info|critical|exception)\(\s*('|\")") log_translation_LC = re.compile( r"(.)*LOG\.(critical)\(\s*(_\(|'|\")") log_translation_LE = re.compile( r"(.)*LOG\.(error|exception)\(\s*(_\(|'|\")") log_translation_LI = re.compile( r"(.)*LOG\.(info)\(\s*(_\(|'|\")") log_translation_LW = re.compile( r"(.)*LOG\.(warning|warn)\(\s*(_\(|'|\")") translated_log = re.compile( r"(.)*LOG\.(audit|error|info|warn|warning|critical|exception)" "\(\s*_\(\s*('|\")") string_translation = re.compile(r"[^_]*_\(\s*('|\")") underscore_import_check = re.compile(r"(.)*import _(.)*") # We need this for cases where they have created their own _ function. custom_underscore_check = re.compile(r"(.)*_\s*=\s*(.)*") oslo_namespace_imports = re.compile(r"from[\s]*oslo[.](.*)") dict_constructor_with_list_copy_re = re.compile(r".*\bdict\((\[)?(\(|\[)") class BaseASTChecker(ast.NodeVisitor): """Provides a simple framework for writing AST-based checks. Subclasses should implement visit_* methods like any other AST visitor implementation. When they detect an error for a particular node the method should call ``self.add_error(offending_node)``. Details about where in the code the error occurred will be pulled from the node object. Subclasses should also provide a class variable named CHECK_DESC to be used for the human readable error message. """ CHECK_DESC = 'No check message specified' def __init__(self, tree, filename): """This object is created automatically by pep8. :param tree: an AST tree :param filename: name of the file being analyzed (ignored by our checks) """ self._tree = tree self._errors = [] def run(self): """Called automatically by pep8.""" self.visit(self._tree) return self._errors def add_error(self, node, message=None): """Add an error caused by a node to the list of errors for pep8.""" message = message or self.CHECK_DESC error = (node.lineno, node.col_offset, message, self.__class__) self._errors.append(error) def _check_call_names(self, call_node, names): if isinstance(call_node, ast.Call): if isinstance(call_node.func, ast.Name): if call_node.func.id in names: return True return False def no_translate_debug_logs(logical_line, filename): """Check for 'LOG.debug(_(' As per our translation policy, https://wiki.openstack.org/wiki/LoggingStandards#Log_Translation we shouldn't translate debug level logs. * This check assumes that 'LOG' is a logger. * Use filename so we can start enforcing this in specific folders instead of needing to do so all at once. M319 """ if logical_line.startswith("LOG.debug(_("): yield(0, "M319 Don't translate debug level logs") def validate_log_translations(logical_line, physical_line, filename): # Translations are not required in the test and tempest # directories. if ("manila/tests" in filename or "manila_tempest_tests" in filename or "contrib/tempest" in filename): return if pep8.noqa(physical_line): return msg = "M327: LOG.critical messages require translations `_LC()`!" if log_translation_LC.match(logical_line): yield (0, msg) msg = ("M328: LOG.error and LOG.exception messages require translations " "`_LE()`!") if log_translation_LE.match(logical_line): yield (0, msg) msg = "M329: LOG.info messages require translations `_LI()`!" if log_translation_LI.match(logical_line): yield (0, msg) msg = "M330: LOG.warning messages require translations `_LW()`!" if log_translation_LW.match(logical_line): yield (0, msg) msg = "M331: Log messages require translations!" if log_translation.match(logical_line): yield (0, msg) def check_explicit_underscore_import(logical_line, filename): """Check for explicit import of the _ function We need to ensure that any files that are using the _() function to translate logs are explicitly importing the _ function. We can't trust unit test to catch whether the import has been added so we need to check for it here. """ # Build a list of the files that have _ imported. No further # checking needed once it is found. if filename in UNDERSCORE_IMPORT_FILES: pass elif (underscore_import_check.match(logical_line) or custom_underscore_check.match(logical_line)): UNDERSCORE_IMPORT_FILES.append(filename) elif (translated_log.match(logical_line) or string_translation.match(logical_line)): yield(0, "M323: Found use of _() without explicit import of _ !") class CheckForStrExc(BaseASTChecker): """Checks for the use of str() on an exception. This currently only handles the case where str() is used in the scope of an exception handler. If the exception is passed into a function, returned from an assertRaises, or used on an exception created in the same scope, this does not catch it. """ CHECK_DESC = ('M325 str() cannot be used on an exception. ' 'Remove or use six.text_type()') def __init__(self, tree, filename): super(CheckForStrExc, self).__init__(tree, filename) self.name = [] self.already_checked = [] def visit_TryExcept(self, node): for handler in node.handlers: if handler.name: self.name.append(handler.name.id) super(CheckForStrExc, self).generic_visit(node) self.name = self.name[:-1] else: super(CheckForStrExc, self).generic_visit(node) def visit_Call(self, node): if self._check_call_names(node, ['str']): if node not in self.already_checked: self.already_checked.append(node) if isinstance(node.args[0], ast.Name): if node.args[0].id in self.name: self.add_error(node.args[0]) super(CheckForStrExc, self).generic_visit(node) class CheckForTransAdd(BaseASTChecker): """Checks for the use of concatenation on a translated string. Translations should not be concatenated with other strings, but should instead include the string being added to the translated string to give the translators the most information. """ CHECK_DESC = ('M326 Translated messages cannot be concatenated. ' 'String should be included in translated message.') TRANS_FUNC = ['_', '_LI', '_LW', '_LE', '_LC'] def visit_BinOp(self, node): if isinstance(node.op, ast.Add): if self._check_call_names(node.left, self.TRANS_FUNC): self.add_error(node.left) elif self._check_call_names(node.right, self.TRANS_FUNC): self.add_error(node.right) super(CheckForTransAdd, self).generic_visit(node) def check_oslo_namespace_imports(logical_line, physical_line, filename): if pep8.noqa(physical_line): return if re.match(oslo_namespace_imports, logical_line): msg = ("M333: '%s' must be used instead of '%s'.") % ( logical_line.replace('oslo.', 'oslo_'), logical_line) yield(0, msg) def dict_constructor_with_list_copy(logical_line): msg = ("M336: Must use a dict comprehension instead of a dict constructor" " with a sequence of key-value pairs." ) if dict_constructor_with_list_copy_re.match(logical_line): yield (0, msg) def factory(register): register(validate_log_translations) register(check_explicit_underscore_import) register(no_translate_debug_logs) register(CheckForStrExc) register(CheckForTransAdd) register(check_oslo_namespace_imports) register(dict_constructor_with_list_copy)
manila/hacking/checks.py
9,211
Provides a simple framework for writing AST-based checks. Subclasses should implement visit_* methods like any other AST visitor implementation. When they detect an error for a particular node the method should call ``self.add_error(offending_node)``. Details about where in the code the error occurred will be pulled from the node object. Subclasses should also provide a class variable named CHECK_DESC to be used for the human readable error message. Checks for the use of str() on an exception. This currently only handles the case where str() is used in the scope of an exception handler. If the exception is passed into a function, returned from an assertRaises, or used on an exception created in the same scope, this does not catch it. Checks for the use of concatenation on a translated string. Translations should not be concatenated with other strings, but should instead include the string being added to the translated string to give the translators the most information. This object is created automatically by pep8. :param tree: an AST tree :param filename: name of the file being analyzed (ignored by our checks) Add an error caused by a node to the list of errors for pep8. Check for explicit import of the _ function We need to ensure that any files that are using the _() function to translate logs are explicitly importing the _ function. We can't trust unit test to catch whether the import has been added so we need to check for it here. Check for 'LOG.debug(_(' As per our translation policy, https://wiki.openstack.org/wiki/LoggingStandards#Log_Translation we shouldn't translate debug level logs. * This check assumes that 'LOG' is a logger. * Use filename so we can start enforcing this in specific folders instead of needing to do so all at once. M319 Called automatically by pep8. Copyright (c) 2012, Cloudscaling All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. We need this for cases where they have created their own _ function. Translations are not required in the test and tempest directories. Build a list of the files that have _ imported. No further checking needed once it is found.
2,647
en
0.874736
# -*-coding:Utf-8 -* # Copyright (c) 2013 LE GOFF Vincent # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notice, this # list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright notice, # this list of conditions and the following disclaimer in the documentation # and/or other materials provided with the distribution. # * Neither the name of the copyright holder nor the names of its contributors # may be used to endorse or promote products derived from this software # without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT # OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. """Fichier contenant la volonté TenirGouvernail""" import re from secondaires.navigation.equipage.ordres.tenir_gouvernail import \ TenirGouvernail as OrdreTenirGouvernail from secondaires.navigation.equipage.ordres.long_deplacer import LongDeplacer from secondaires.navigation.equipage.volonte import Volonte class TenirGouvernail(Volonte): """Classe représentant une volonté. Cette volonté choisit un matelot pour tenir le gouvernail du navire. """ cle = "tenir_gouvernail" ordre_court = re.compile(r"^tg$", re.I) ordre_long = re.compile(r"^tenir\s+gouvernail?$", re.I) def choisir_matelots(self, exception=None): """Retourne le matelot le plus apte à accomplir la volonté.""" proches = [] matelots = self.navire.equipage.get_matelots_libres(exception) graph = self.navire.graph gouvernail = self.navire.gouvernail if gouvernail is None or gouvernail.tenu is not None: return None for matelot in matelots: origine = matelot.salle.mnemonic destination = gouvernail.parent.mnemonic if origine == destination: proches.append((matelot, [], gouvernail)) else: chemin = graph.get((origine, destination)) if chemin: proches.append((matelot, chemin, gouvernail)) proches = sorted([couple for couple in proches], key=lambda couple: len(couple[1])) if proches: return proches[0] return None def executer(self, sequence): """Exécute la volonté.""" if sequence is None: self.terminer() return matelot, sorties, gouvernail = sequence navire = self.navire ordres = [] if sorties: aller = LongDeplacer(matelot, navire, *sorties) ordres.append(aller) tenir = OrdreTenirGouvernail(matelot, navire) ordres.append(tenir) self.ajouter_ordres(matelot, ordres) def crier_ordres(self, personnage): """On fait crier l'ordre au personnage.""" msg = "{} s'écrie : un homme à la barre !".format( personnage.distinction_audible) self.navire.envoyer(msg) @classmethod def extraire_arguments(cls, navire): """Extrait les arguments de la volonté.""" return ()
src/secondaires/navigation/equipage/volontes/tenir_gouvernail.py
3,986
Classe représentant une volonté. Cette volonté choisit un matelot pour tenir le gouvernail du navire. Retourne le matelot le plus apte à accomplir la volonté. On fait crier l'ordre au personnage. Exécute la volonté. Extrait les arguments de la volonté. Fichier contenant la volonté TenirGouvernail -*-coding:Utf-8 -* Copyright (c) 2013 LE GOFF Vincent All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
1,805
en
0.714943
#https://blog.csdn.net/orangefly0214/article/details/81387077 import MultiTemplate from MultiTemplate import TaskTemplate # https://blog.csdn.net/u013812710/article/details/72886491 # https://blog.csdn.net/ismr_m/article/details/53100896 #https://blog.csdn.net/bcfdsagbfcisbg/article/details/78134172 import kubernetes import os import influxdb import time import yaml def check_path(name): train_dir = os.path.join('/tfdata/k8snfs/', name) print(train_dir) if not os.path.exists(train_dir): os.makedirs(train_dir) return train_dir def check_ns(name): kubernetes.config.load_kube_config() v1 = kubernetes.client.CoreV1Api() # v1.create_namespace() exist_ns = v1.list_namespace() exist_ns_name = [] for i in exist_ns.items: exist_ns_name.append(i.metadata.name) if name in exist_ns_name: return True else: return False class SubTask(): def __init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag): self.template_id = template_id self.ps_replicas = ps_replicas self.worker_replicas = worker_replicas self.training_step = training_step self.interval = interval self.batch_size = batch_size self.task_id = task_id self.tag = tag self.rtimes = rtimes self.influx_client = influxdb.InfluxDBClient(host='192.168.128.10',port=8086,username='admin',password='admin',database="NODEMESSAGE") self.node_list = ['k8s-master','k8s-worker0','k8s-worker2','k8sworker1','k8s-worker3','k8s-worker4','k8s-worker5'] #self.node_list = ['k8s-master','k8s-worker0','k8s-worker2','k8sworker1'] self.node_cpu = {} self.node_cpu['k8s-master'] = 32000 self.node_cpu['k8s-worker0'] = 24000 self.node_cpu['k8s-worker2'] = 24000 self.node_cpu['k8sworker1'] = 16000 self.node_cpu['k8s-worker3'] = 24000 self.node_cpu['k8s-worker4'] = 16000 self.node_cpu['k8s-worker5'] = 24000 self.node_memory = {} self.node_memory['k8s-master'] = float(251*1024) self.node_memory['k8s-worker0'] = float(94*1024) self.node_memory['k8s-worker2'] = float(94*1024) self.node_memory['k8sworker1'] = float(125*1024) self.node_memory['k8s-worker3'] = float(94 * 1024) self.node_memory['k8s-worker4'] = float(125 * 1024) self.node_memory['k8s-worker5'] = float(94 * 1024) self.args = ['--training_step='+str(self.training_step),'--batch_size='+str(self.batch_size),'--interval='+str(self.interval),'--task_id='+str(self.task_id),'--rtimes='+str(self.rtimes),"--tag="+self.tag] class VGGTask(SubTask): def __init__(self,v1,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag,channel1,channel2,channel3,channel4,channel5,num_layer1,num_layer2,num_layer3,num_layer4,num_layer5): SubTask.__init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag) self.channel1 = channel1 self.channel2 = channel2 self.channel3 = channel3 self.channel4 = channel4 self.channel5 = channel5 self.num_layer1 = num_layer1 self.num_layer2 = num_layer2 self.num_layer3 = num_layer3 self.num_layer4 = num_layer4 self.num_layer5 = num_layer5 self.num_layers = num_layer1+num_layer2+num_layer3+num_layer4+num_layer5+3 self.template = TaskTemplate.VGG self.v1 = v1 self.name = 'vgg-'+str(self.task_id)+'-'+str(self.rtimes) def get_node_list(self): node_list = [i.metadata.name for i in self.v1.list_node().items] return node_list def make_args(self): self.args.append('--channel1='+str(self.channel1)) self.args.append('--channel2='+str(self.channel2)) self.args.append('--channel3='+str(self.channel3)) self.args.append('--channel4='+str(self.channel4)) self.args.append('--channel5='+str(self.channel5)) self.args.append('--num_layer1='+str(self.num_layer1)) self.args.append('--num_layer2='+str(self.num_layer2)) self.args.append('--num_layer3='+str(self.num_layer3)) self.args.append('--num_layer4='+str(self.num_layer4)) self.args.append('--num_layer5='+str(self.num_layer5)) self.args.append('--num_layers='+str(self.num_layers)) def create_tf(self): name = 'vgg-'+str(self.task_id)+'-'+str(self.rtimes) ns_body = TaskTemplate.NS ns_body['metadata']['name'] = name if not check_ns(name): self.v1.create_namespace(ns_body) train_dir = check_path(name) time.sleep(12) result = self.influx_client.query("select * from "+"NODEMESSAGE"+" group by nodes order by desc limit 3") node_list = self.get_node_list() result_keys = result.keys() nodes = [i[-1]['nodes'] for i in result_keys] node_mg = [list(result[i]) for i in result_keys] cpu_base = {} memory_base = {} point_base = {} point_base_list = [] for i in range(len(node_mg)): cpu_base[nodes[i]] = 0 memory_base[nodes[i]] = 0 point_base[nodes[i]] = 0.0 for j in range(len(node_mg[0])): cpu_base[nodes[i]] += node_mg[i][j]['cpu'] memory_base[nodes[i]] += node_mg[i][j]['memory'] cpu_base[nodes[i]] = (cpu_base[nodes[i]] / len(node_mg[0]))/self.node_cpu[nodes[i]] memory_base[nodes[i]] = (memory_base[nodes[i]] / len(node_mg[0])) / self.node_memory[nodes[i]] tmp = cpu_base[nodes[i]]*0.6+memory_base[nodes[i]]*0.4 point_base[nodes[i]] = tmp point_base_list.append(tmp) list.sort(point_base_list) for key in nodes: command = 'kubectl label nodes '+key+' woksch-' os.system(command) command2 = 'kubectl label nodes '+key+' wokpro-' os.system(command2) nod_prori = point_base_list.index(point_base[key]) priori = ' wokpro=%d' % nod_prori command3 = 'kubectl label nodes '+key+priori os.system(command3) if cpu_base[key] <= 0.57 and memory_base[key] <= 0.6: command = 'kubectl label nodes '+key+' woksch=true' os.system(command) else: command = 'kubectl label nodes ' + key + ' woksch=false' os.system(command) self.template['metadata']['name'] = name self.template['metadata']['namespace'] = name self.template['spec']['tfReplicaSpecs']['PS']['replicas'] = self.ps_replicas self.template['spec']['tfReplicaSpecs']['Worker']['replicas'] = self.worker_replicas self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.make_args() self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['args'] = self.args[:] self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['args'] = self.args[:] log_dir = '/tfdata/tfcnn/expjob/' # f = open(log_dir+str(name)+'.yaml', "w") f = open(log_dir + str(name) + '.yaml', "w") yaml.dump(self.template, f) f.close() response = os.system('kubectl create -f '+log_dir+str(name)+'.yaml') if response == 0: print('create task sucess') else: print("Error code:"+str(response)) def delete_tf(self): name = 'vgg-'+str(self.task_id)+'-'+str(self.rtimes) log_dir = '/tfdata/tfcnn/expjob/' response = os.system('kubectl delete -f ' + log_dir + str(name) + '.yaml') if response == 0: print('delete task sucess') else: print("Error code:" + str(response)) self.v1.delete_namespace(name=name) class RESTask(SubTask): def __init__(self,v1,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag,bottle,layer1,layer2,layer3,layer4,channel1,channel2,channel3,channel4): SubTask.__init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag) self.channel1 = channel1 self.channel2 = channel2 self.channel3 = channel3 self.channel4 = channel4 self.bottle = bottle self.layer1 = layer1 self.layer2 = layer2 self.layer3 = layer3 self.layer4 = layer4 self.name = 'res-'+str(self.task_id)+'-'+str(self.rtimes) if self.bottle == 1: self.num_layers = 3*(layer1+layer4+layer3+layer2)+2 else: self.num_layers = 2 * (layer1 + layer4 + layer3 + layer2) + 2 self.template = TaskTemplate.RES self.v1 = v1 def get_node_list(self): node_list = [i.metadata.name for i in self.v1.list_node().items] return node_list def make_args(self): self.args.append('--bottle=' + str(self.bottle)) self.args.append('--channel1='+str(self.channel1)) self.args.append('--channel2='+str(self.channel2)) self.args.append('--channel3='+str(self.channel3)) self.args.append('--channel4='+str(self.channel4)) self.args.append('--layer1='+str(self.layer1)) self.args.append('--layer2='+str(self.layer2)) self.args.append('--layer3='+str(self.layer3)) self.args.append('--layer4='+str(self.layer4)) def create_tf(self): name = 'res-'+str(self.task_id)+'-'+str(self.rtimes) ns_body = TaskTemplate.NS ns_body['metadata']['name'] = name if not check_ns(name): self.v1.create_namespace(ns_body) train_dir = check_path(name) time.sleep(12) result = self.influx_client.query("select * from " + "NODEMESSAGE" + " group by nodes order by desc limit 3") node_list = self.get_node_list() result_keys = result.keys() nodes = [i[-1]['nodes'] for i in result_keys] node_mg = [list(result[i]) for i in result_keys] cpu_base = {} memory_base = {} point_base = {} point_base_list = [] for i in range(len(node_mg)): cpu_base[nodes[i]] = 0 memory_base[nodes[i]] = 0 point_base[nodes[i]] = 0.0 for j in range(len(node_mg[0])): cpu_base[nodes[i]] += node_mg[i][j]['cpu'] memory_base[nodes[i]] += node_mg[i][j]['memory'] cpu_base[nodes[i]] = (cpu_base[nodes[i]] / len(node_mg[0])) / self.node_cpu[nodes[i]] memory_base[nodes[i]] = (memory_base[nodes[i]] / len(node_mg[0])) / self.node_memory[nodes[i]] tmp = cpu_base[nodes[i]] * 0.6 + memory_base[nodes[i]] * 0.4 point_base[nodes[i]] = tmp point_base_list.append(tmp) list.sort(point_base_list) for key in nodes: command = 'kubectl label nodes ' + key + ' woksch-' os.system(command) command2 = 'kubectl label nodes ' + key + ' wokpro-' os.system(command2) nod_prori = point_base_list.index(point_base[key]) priori = ' wokpro=%d' % nod_prori command3 = 'kubectl label nodes ' + key + priori os.system(command3) if cpu_base[key] <= 0.6 and memory_base[key] <= 0.6: command = 'kubectl label nodes ' + key + ' woksch=true' os.system(command) else: command = 'kubectl label nodes ' + key + ' woksch=false' os.system(command) self.template['metadata']['name'] = name self.template['metadata']['namespace'] = name self.template['spec']['tfReplicaSpecs']['PS']['replicas'] = self.ps_replicas self.template['spec']['tfReplicaSpecs']['Worker']['replicas'] = self.worker_replicas self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.make_args() self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['args'] = self.args[:] self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['args'] = self.args[:] log_dir = '/tfdata/tfcnn/expjob/' f = open(log_dir+str(name)+'.yaml', "w") yaml.dump(self.template, f) f.close() response = os.system('kubectl create -f '+log_dir+str(name)+'.yaml') if response == 0: print('create task sucess') else: print("Error code:"+str(response)) def delete_tf(self): name = 'res-'+str(self.task_id)+'-'+str(self.rtimes) log_dir = '/tfdata/tfcnn/expjob/' response = os.system('kubectl delete -f ' + log_dir + str(name) + '.yaml') if response == 0: print('delete task sucess') else: print("Error code:" + str(response)) self.v1.delete_namespace(name=name) class RETask(SubTask): def __init__(self,v1,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag,stack,channel1,channel2,channel3,channel4): SubTask.__init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag) self.channel1 = channel1 self.channel2 = channel2 self.channel3 = channel3 self.channel4 = channel4 self.stack = stack self.num_layers = 6*self.stack+2 self.template = TaskTemplate.RE self.name = 're-'+str(self.task_id)+'-'+str(self.rtimes) self.v1 = v1 def get_node_list(self): node_list = [i.metadata.name for i in self.v1.list_node().items] return node_list def make_args(self): self.args.append('--stack='+str(self.stack)) self.args.append('--channel1='+str(self.channel1)) self.args.append('--channel2='+str(self.channel2)) self.args.append('--channel3='+str(self.channel3)) self.args.append('--channel4='+str(self.channel4)) def create_tf(self): name = 're-'+str(self.task_id)+'-'+str(self.rtimes) ns_body = TaskTemplate.NS ns_body['metadata']['name'] = name if not check_ns(name): self.v1.create_namespace(ns_body) train_dir = check_path(name) time.sleep(12) result = self.influx_client.query("select * from " + "NODEMESSAGE" + " group by nodes order by desc limit 3") node_list = self.get_node_list() result_keys = result.keys() nodes = [i[-1]['nodes'] for i in result_keys] node_mg = [list(result[i]) for i in result_keys] cpu_base = {} memory_base = {} point_base = {} point_base_list = [] for i in range(len(node_mg)): cpu_base[nodes[i]] = 0 memory_base[nodes[i]] = 0 point_base[nodes[i]] = 0.0 for j in range(len(node_mg[0])): cpu_base[nodes[i]] += node_mg[i][j]['cpu'] memory_base[nodes[i]] += node_mg[i][j]['memory'] cpu_base[nodes[i]] = (cpu_base[nodes[i]] / len(node_mg[0])) / self.node_cpu[nodes[i]] memory_base[nodes[i]] = (memory_base[nodes[i]] / len(node_mg[0])) / self.node_memory[nodes[i]] tmp = cpu_base[nodes[i]] * 0.6 + memory_base[nodes[i]] * 0.4 point_base[nodes[i]] = tmp point_base_list.append(tmp) list.sort(point_base_list) for key in nodes: command = 'kubectl label nodes ' + key + ' woksch-' os.system(command) command2 = 'kubectl label nodes ' + key + ' wokpro-' os.system(command2) nod_prori = point_base_list.index(point_base[key]) priori = ' wokpro=%d' % nod_prori command3 = 'kubectl label nodes ' + key + priori os.system(command3) if cpu_base[key] <= 0.6 and memory_base[key] <= 0.6: command = 'kubectl label nodes ' + key + ' woksch=true' os.system(command) else: command = 'kubectl label nodes ' + key + ' woksch=false' os.system(command) self.template['metadata']['name'] = name self.template['metadata']['namespace'] = name self.template['spec']['tfReplicaSpecs']['PS']['replicas'] = self.ps_replicas self.template['spec']['tfReplicaSpecs']['Worker']['replicas'] = self.worker_replicas self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.make_args() self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['args'] = self.args[:] self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['args'] = self.args[:] log_dir = '/tfdata/tfcnn/expjob/' f = open(log_dir+str(name)+'.yaml', "w") yaml.dump(self.template, f) f.close() response = os.system('kubectl create -f '+log_dir+str(name)+'.yaml') if response == 0: print('create task sucess') else: print("Error code:"+str(response)) def delete_tf(self): name = 're-'+str(self.task_id)+'-'+str(self.rtimes) log_dir = '/tfdata/tfcnn/expjob/' response = os.system('kubectl delete -f ' + log_dir + str(name) + '.yaml') if response == 0: print('delete task sucess') else: print("Error code:" + str(response)) self.v1.delete_namespace(name=name) class XCETask(SubTask): def __init__(self,v1,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag,repeat,channel1,channel2,channel3,channel4,channel5,channel6,channel7,channel8): SubTask.__init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag) self.channel1 = channel1 self.channel2 = channel2 self.channel3 = channel3 self.channel4 = channel4 self.channel5 = channel5 self.channel6 = channel6 self.channel7 = channel7 self.channel8 = channel8 self.repeat = repeat self.template = TaskTemplate.XCEPTION self.v1 = v1 self.name = 'xception-'+str(self.task_id)+'-'+str(self.rtimes) def get_node_list(self): node_list = [i.metadata.name for i in self.v1.list_node().items] return node_list def make_args(self): self.args.append('--repeat='+str(self.repeat)) self.args.append('--channel1='+str(self.channel1)) self.args.append('--channel2='+str(self.channel2)) self.args.append('--channel3='+str(self.channel3)) self.args.append('--channel4='+str(self.channel4)) self.args.append('--channel5=' + str(self.channel5)) self.args.append('--channel6=' + str(self.channel6)) self.args.append('--channel7=' + str(self.channel7)) self.args.append('--channel8=' + str(self.channel8)) def create_tf(self): name = 'xception-'+str(self.task_id)+'-'+str(self.rtimes) ns_body = TaskTemplate.NS ns_body['metadata']['name'] = name if not check_ns(name): self.v1.create_namespace(ns_body) train_dir = check_path(name) time.sleep(12) result = self.influx_client.query("select * from " + "NODEMESSAGE" + " group by nodes order by desc limit 3") node_list = self.get_node_list() result_keys = result.keys() nodes = [i[-1]['nodes'] for i in result_keys] node_mg = [list(result[i]) for i in result_keys] cpu_base = {} memory_base = {} point_base = {} point_base_list = [] for i in range(len(node_mg)): cpu_base[nodes[i]] = 0 memory_base[nodes[i]] = 0 point_base[nodes[i]] = 0.0 for j in range(len(node_mg[0])): cpu_base[nodes[i]] += node_mg[i][j]['cpu'] memory_base[nodes[i]] += node_mg[i][j]['memory'] cpu_base[nodes[i]] = (cpu_base[nodes[i]] / len(node_mg[0])) / self.node_cpu[nodes[i]] memory_base[nodes[i]] = (memory_base[nodes[i]] / len(node_mg[0])) / self.node_memory[nodes[i]] tmp = cpu_base[nodes[i]] * 0.6 + memory_base[nodes[i]] * 0.4 point_base[nodes[i]] = tmp point_base_list.append(tmp) list.sort(point_base_list) for key in nodes: command = 'kubectl label nodes ' + key + ' woksch-' os.system(command) command2 = 'kubectl label nodes ' + key + ' wokpro-' os.system(command2) nod_prori = point_base_list.index(point_base[key]) priori = ' wokpro=%d' % nod_prori command3 = 'kubectl label nodes ' + key + priori os.system(command3) if cpu_base[key] <= 0.6 and memory_base[key] <= 0.6: command = 'kubectl label nodes ' + key + ' woksch=true' os.system(command) else: command = 'kubectl label nodes ' + key + ' woksch=false' os.system(command) self.template['metadata']['name'] = name self.template['metadata']['namespace'] = name self.template['spec']['tfReplicaSpecs']['PS']['replicas'] = self.ps_replicas self.template['spec']['tfReplicaSpecs']['Worker']['replicas'] = self.worker_replicas self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.make_args() self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['args'] = self.args[:] self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['args'] = self.args[:] log_dir = '/tfdata/tfcnn/expjob/' f = open(log_dir+str(name)+'.yaml', "w") yaml.dump(self.template, f) f.close() response = os.system('kubectl create -f '+log_dir+str(name)+'.yaml') if response == 0: print('create task sucess') else: print("Error code:"+str(response)) def delete_tf(self): name = 'xception-'+str(self.task_id)+'-'+str(self.rtimes) log_dir = '/tfdata/tfcnn/expjob/' response = os.system('kubectl delete -f ' + log_dir + str(name) + '.yaml') if response == 0: print('delete task sucess') else: print("Error code:" + str(response)) self.v1.delete_namespace(name=name) class DENTask(SubTask): def __init__(self,v1,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag,L,k,BC): SubTask.__init__(self,template_id,ps_replicas,worker_replicas,training_step,batch_size,interval,task_id,rtimes,tag) self.L = L self.k = k self.BC = BC self.template = TaskTemplate.DEN self.v1 = v1 self.name = 'den-'+str(self.task_id)+'-'+str(self.rtimes) def get_node_list(self): node_list = [i.metadata.name for i in self.v1.list_node().items] return node_list def make_args(self): self.args.append('--L='+str(self.L)) self.args.append('--k='+str(self.k)) self.args.append('--BC='+str(self.BC)) def create_tf(self): name = 'den-'+str(self.task_id)+'-'+str(self.rtimes) ns_body = TaskTemplate.NS ns_body['metadata']['name'] = name if not check_ns(name): self.v1.create_namespace(ns_body) train_dir = check_path(name) time.sleep(12) result = self.influx_client.query("select * from " + "NODEMESSAGE" + " group by nodes order by desc limit 3") node_list = self.get_node_list() result_keys = result.keys() nodes = [i[-1]['nodes'] for i in result_keys] node_mg = [list(result[i]) for i in result_keys] cpu_base = {} memory_base = {} point_base = {} point_base_list = [] for i in range(len(node_mg)): cpu_base[nodes[i]] = 0 memory_base[nodes[i]] = 0 point_base[nodes[i]] = 0.0 for j in range(len(node_mg[0])): cpu_base[nodes[i]] += node_mg[i][j]['cpu'] memory_base[nodes[i]] += node_mg[i][j]['memory'] cpu_base[nodes[i]] = (cpu_base[nodes[i]] / len(node_mg[0])) / self.node_cpu[nodes[i]] memory_base[nodes[i]] = (memory_base[nodes[i]] / len(node_mg[0])) / self.node_memory[nodes[i]] tmp = cpu_base[nodes[i]] * 0.6 + memory_base[nodes[i]] * 0.4 point_base[nodes[i]] = tmp point_base_list.append(tmp) list.sort(point_base_list) for key in nodes: command = 'kubectl label nodes ' + key + ' woksch-' os.system(command) command2 = 'kubectl label nodes ' + key + ' wokpro-' os.system(command2) nod_prori = point_base_list.index(point_base[key]) priori = ' wokpro=%d' % nod_prori command3 = 'kubectl label nodes ' + key + priori os.system(command3) if cpu_base[key] <= 0.6 and memory_base[key] <= 0.6: command = 'kubectl label nodes ' + key + ' woksch=true' os.system(command) else: command = 'kubectl label nodes ' + key + ' woksch=false' os.system(command) self.template['metadata']['name'] = name self.template['metadata']['namespace'] = name self.template['spec']['tfReplicaSpecs']['PS']['replicas'] = self.ps_replicas self.template['spec']['tfReplicaSpecs']['Worker']['replicas'] = self.worker_replicas self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['volumes'][0]['hostPath']['path'] = train_dir self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['volumeMounts'][0]['name'] = name self.make_args() self.template['spec']['tfReplicaSpecs']['PS']['template']['spec']['containers'][0]['args'] = self.args[:] self.template['spec']['tfReplicaSpecs']['Worker']['template']['spec']['containers'][0]['args'] = self.args[:] log_dir = '/tfdata/tfcnn/expjob/' f = open(log_dir+str(name)+'.yaml', "w") yaml.dump(self.template, f) f.close() response = os.system('kubectl create -f '+log_dir+str(name)+'.yaml') if response == 0: print('create task sucess') else: print("Error code:"+str(response)) def delete_tf(self): name = 'den-'+str(self.task_id)+'-'+str(self.rtimes) log_dir = '/tfdata/tfcnn/expjob/' response = os.system('kubectl delete -f ' + log_dir + str(name) + '.yaml') if response == 0: print('delete task sucess') else: print("Error code:" + str(response)) self.v1.delete_namespace(name=name) if __name__ == '__main__': kubernetes.config.load_kube_config() v1 = kubernetes.client.CoreV1Api() # v1.create_namespace() v1.list_namespace() check_path('ceshi') # vgg = VGGTask(1,2,4,80,1.0,2,1,"ms",32,64,128,256,512,2,3,3,4,4) # vgg.create_tf()
experiment code/CPU Experiments Code/task_submit_save.py
30,278
https://blog.csdn.net/orangefly0214/article/details/81387077 https://blog.csdn.net/u013812710/article/details/72886491 https://blog.csdn.net/ismr_m/article/details/53100896https://blog.csdn.net/bcfdsagbfcisbg/article/details/78134172 v1.create_namespace()self.node_list = ['k8s-master','k8s-worker0','k8s-worker2','k8sworker1'] f = open(log_dir+str(name)+'.yaml', "w") v1.create_namespace() vgg = VGGTask(1,2,4,80,1.0,2,1,"ms",32,64,128,256,512,2,3,3,4,4) vgg.create_tf()
471
en
0.544803
# -*- coding: utf-8 -*- from datetime import datetime import time import unittest from webapp2_caffeine.cache import CacheContainer from webapp2_caffeine.cache import flush class DummyCache(CacheContainer): key = 'dummy_cache' @property def fresh_value(self): return datetime.now() class CacheContainerTest(unittest.TestCase): def setUp(self): flush() def tearDown(self): flush() def test_fresh_value(self): container = CacheContainer() with self.assertRaises(NotImplementedError): container.fresh_value def test_set(self): container = CacheContainer() with self.assertRaises(ValueError): container.set('my value') container = DummyCache() value, expiration = container.set('my value') self.assertEqual(value, 'my value') self.assertTrue(21000 < expiration - time.time() < 21600) self.assertEqual(container.get(), 'my value') def test_get(self): container = DummyCache() self.assertEqual(container.get(), None) container.set('my value', 1000) self.assertEqual(container.get(), None) container.set('my value') self.assertEqual(container.get(), 'my value') def test_delete(self): container = DummyCache() container.set('my value') container.delete() self.assertEqual(container.get(), None) def test_update(self): container = DummyCache() container.update() self.assertTrue(container.get()) def test_value(self): container = DummyCache() old_value = container.value self.assertTrue(old_value) self.assertTrue(container.value, old_value)
tests/test_cache.py
1,750
-*- coding: utf-8 -*-
21
en
0.767281
#!/usr/bin/env python3 # Copyright (c) 2015-2020 The Ttm Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. from test_framework.blocktools import create_block, create_coinbase, get_masternode_payment from test_framework.mininode import P2PDataStore, network_thread_start from test_framework.messages import CTxOut, FromHex, CCbTx, CTransaction, ToHex from test_framework.script import CScript from test_framework.test_framework import TtmTestFramework from test_framework.util import assert_equal, get_bip9_status, hex_str_to_bytes ''' feature_block_reward_reallocation.py Checks block reward reallocation correctness ''' class BlockRewardReallocationTest(TtmTestFramework): def set_test_params(self): self.set_ttm_test_params(2, 1, fast_dip3_enforcement=True) self.set_ttm_dip8_activation(450) # 536870912 == 0x20000000, i.e. not signalling for anything def create_test_block(self, version=536870912): self.bump_mocktime(5) bt = self.nodes[0].getblocktemplate() tip = int(bt['previousblockhash'], 16) nextheight = bt['height'] coinbase = create_coinbase(nextheight) coinbase.nVersion = 3 coinbase.nType = 5 # CbTx coinbase.vout[0].nValue = bt['coinbasevalue'] for mn in bt['masternode']: coinbase.vout.append(CTxOut(mn['amount'], CScript(hex_str_to_bytes(mn['script'])))) coinbase.vout[0].nValue -= mn['amount'] cbtx = FromHex(CCbTx(), bt['coinbase_payload']) coinbase.vExtraPayload = cbtx.serialize() coinbase.rehash() coinbase.calc_sha256() block = create_block(tip, coinbase, self.mocktime) block.nVersion = version # Add quorum commitments from template for tx in bt['transactions']: tx2 = FromHex(CTransaction(), tx['data']) if tx2.nType == 6: block.vtx.append(tx2) block.hashMerkleRoot = block.calc_merkle_root() block.rehash() block.solve() return block def signal(self, num_blocks, expected_lockin): self.log.info("Signal with %d/500 blocks" % (num_blocks)) # create and send non-signalling blocks for i in range(500 - num_blocks): test_block = self.create_test_block() self.nodes[0].submitblock(ToHex(test_block)) # generate at most 10 signaling blocks at a time if num_blocks > 0: for i in range((num_blocks - 1) // 10): self.bump_mocktime(10) self.nodes[0].generate(10) self.nodes[0].generate((num_blocks - 1) % 10) assert_equal(get_bip9_status(self.nodes[0], 'realloc')['status'], 'started') self.nodes[0].generate(1) if expected_lockin: assert_equal(get_bip9_status(self.nodes[0], 'realloc')['status'], 'locked_in') else: assert_equal(get_bip9_status(self.nodes[0], 'realloc')['status'], 'started') def threshold(self, attempt): threshold_calc = 400 - attempt * attempt if threshold_calc < 300: return 300 return threshold_calc def run_test(self): self.log.info("Wait for DIP3 to activate") while get_bip9_status(self.nodes[0], 'dip0003')['status'] != 'active': self.bump_mocktime(10) self.nodes[0].generate(10) self.nodes[0].add_p2p_connection(P2PDataStore()) network_thread_start() self.nodes[0].p2p.wait_for_verack() self.log.info("Mine all but one remaining block in the window") bi = self.nodes[0].getblockchaininfo() for i in range(498 - bi['blocks']): self.bump_mocktime(1) self.nodes[0].generate(1) self.log.info("Initial state is DEFINED") bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 498) assert_equal(bi['bip9_softforks']['realloc']['status'], 'defined') self.log.info("Advance from DEFINED to STARTED at height = 499") self.nodes[0].generate(1) bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 499) assert_equal(bi['bip9_softforks']['realloc']['status'], 'started') assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], self.threshold(0)) self.signal(399, False) # 1 block short self.log.info("Still STARTED but new threshold should be lower at height = 999") bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 999) assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], self.threshold(1)) self.signal(398, False) # 1 block short again self.log.info("Still STARTED but new threshold should be even lower at height = 1499") bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 1499) assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], self.threshold(2)) pre_locked_in_blockhash = bi['bestblockhash'] self.signal(396, True) # just enough to lock in self.log.info("Advanced to LOCKED_IN at height = 1999") for i in range(49): self.bump_mocktime(10) self.nodes[0].generate(10) self.nodes[0].generate(9) self.log.info("Still LOCKED_IN at height = 2498") bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 2498) assert_equal(bi['bip9_softforks']['realloc']['status'], 'locked_in') self.log.info("Advance from LOCKED_IN to ACTIVE at height = 2499") self.nodes[0].generate(1) # activation bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 2499) assert_equal(bi['bip9_softforks']['realloc']['status'], 'active') assert_equal(bi['bip9_softforks']['realloc']['since'], 2500) self.log.info("Reward split should stay ~50/50 before the first superblock after activation") # This applies even if reallocation was activated right at superblock height like it does here bt = self.nodes[0].getblocktemplate() assert_equal(bt['height'], 2500) assert_equal(bt['masternode'][0]['amount'], get_masternode_payment(bt['height'], bt['coinbasevalue'], 2500)) self.nodes[0].generate(9) bt = self.nodes[0].getblocktemplate() assert_equal(bt['masternode'][0]['amount'], get_masternode_payment(bt['height'], bt['coinbasevalue'], 2500)) assert_equal(bt['coinbasevalue'], 13748571607) assert_equal(bt['masternode'][0]['amount'], 6874285801) # 0.4999999998 self.log.info("Reallocation should kick-in with the superblock mined at height = 2010") for period in range(19): # there will be 19 adjustments, 3 superblocks long each for i in range(3): self.bump_mocktime(10) self.nodes[0].generate(10) bt = self.nodes[0].getblocktemplate() assert_equal(bt['masternode'][0]['amount'], get_masternode_payment(bt['height'], bt['coinbasevalue'], 2500)) self.log.info("Reward split should reach ~60/40 after reallocation is done") assert_equal(bt['coinbasevalue'], 10221599170) assert_equal(bt['masternode'][0]['amount'], 6132959502) # 0.6 self.log.info("Reward split should stay ~60/40 after reallocation is done") for period in range(10): # check 10 next superblocks self.bump_mocktime(10) self.nodes[0].generate(10) bt = self.nodes[0].getblocktemplate() assert_equal(bt['masternode'][0]['amount'], get_masternode_payment(bt['height'], bt['coinbasevalue'], 2500)) assert_equal(bt['coinbasevalue'], 9491484944) assert_equal(bt['masternode'][0]['amount'], 5694890966) # 0.6 # make sure all nodes are still synced self.sync_all() self.log.info("Rollback the chain back to the STARTED state") self.mocktime = self.nodes[0].getblock(pre_locked_in_blockhash, 1)['time'] for node in self.nodes: node.invalidateblock(pre_locked_in_blockhash) # create and send non-signalling block test_block = self.create_test_block() self.nodes[0].submitblock(ToHex(test_block)) bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 1499) assert_equal(bi['bip9_softforks']['realloc']['status'], 'started') assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], self.threshold(2)) self.log.info("Check thresholds reach min level and stay there") for i in range(8): # 7 to reach min level and 1 more to check it doesn't go lower than that self.signal(0, False) # no need to signal bi = self.nodes[0].getblockchaininfo() assert_equal(bi['blocks'], 1999 + i * 500) assert_equal(bi['bip9_softforks']['realloc']['status'], 'started') assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], self.threshold(i + 3)) assert_equal(bi['bip9_softforks']['realloc']['statistics']['threshold'], 300) if __name__ == '__main__': BlockRewardReallocationTest().main()
test/functional/feature_block_reward_reallocation.py
9,333
!/usr/bin/env python3 Copyright (c) 2015-2020 The Ttm Core developers Distributed under the MIT software license, see the accompanying file COPYING or http://www.opensource.org/licenses/mit-license.php. 536870912 == 0x20000000, i.e. not signalling for anything CbTx Add quorum commitments from template create and send non-signalling blocks generate at most 10 signaling blocks at a time 1 block short 1 block short again just enough to lock in activation This applies even if reallocation was activated right at superblock height like it does here 0.4999999998 there will be 19 adjustments, 3 superblocks long each 0.6 check 10 next superblocks 0.6 make sure all nodes are still synced create and send non-signalling block 7 to reach min level and 1 more to check it doesn't go lower than that no need to signal
812
en
0.847823
import os from datetime import datetime import numpy as np import xarray as xr from pyoos.collectors.usgs.usgs_rest import UsgsRest from pyoos.parsers.waterml import WaterML11ToPaegan def get_usgs_data(station_id, start_date, end_date, parameter="00060", cache_dir=None): """Get river discharge data from the USGS REST web service. See `U.S. Geological Survey Water Services <https://waterservices.usgs.gov/>`_ (USGS) Parameters ---------- station_id : str The station id to get start_date : str String for start date in the format: 'YYYY-MM-dd', e.g. '1980-01-01' end_date : str String for start date in the format: 'YYYY-MM-dd', e.g. '2018-12-31' parameter : str The parameter code to get, e.g. ('00060') discharge, cubic feet per second cache_dir : str Directory where files retrieved from the web service are cached. If set to None then USGS_DATA_HOME env var will be used as cache directory. Examples -------- >>> from ewatercycle.observation.usgs import get_usgs_data >>> data = get_usgs_data('03109500', '2000-01-01', '2000-12-31', cache_dir='.') >>> data <xarray.Dataset> Dimensions: (time: 8032) Coordinates: * time (time) datetime64[ns] 2000-01-04T05:00:00 ... 2000-12-23T04:00:00 Data variables: Streamflow (time) float32 8.296758 10.420501 ... 10.647034 11.694747 Attributes: title: USGS Data from streamflow data station: Little Beaver Creek near East Liverpool OH stationid: 03109500 location: (40.6758974, -80.5406244) """ # noqa: E501 if cache_dir is None: cache_dir = os.environ["USGS_DATA_HOME"] # Check if we have the netcdf data netcdf = os.path.join( cache_dir, "USGS_" + station_id + "_" + parameter + "_" + start_date + "_" + end_date + ".nc", ) if os.path.exists(netcdf): return xr.open_dataset(netcdf) # Download the data if needed out = os.path.join( cache_dir, "USGS_" + station_id + "_" + parameter + "_" + start_date + "_" + end_date + ".wml", ) if not os.path.exists(out): collector = UsgsRest() collector.filter( start=datetime.strptime(start_date, "%Y-%m-%d"), end=datetime.strptime(end_date, "%Y-%m-%d"), variables=[parameter], features=[station_id], ) data = collector.raw() with open(out, "w") as file: file.write(data) collector.clear() else: with open(out, "r") as file: data = file.read() # Convert the raw data to an xarray data = WaterML11ToPaegan(data).feature # We expect only 1 station if len(data.elements) == 0: raise ValueError("Data does not contain any station data") else: station = data.elements[0] # Unit conversion from cubic feet to cubic meter per second values = np.array( [float(point.members[0]["value"]) / 35.315 for point in station.elements], dtype=np.float32, ) times = [point.time for point in station.elements] attrs = { "units": "cubic meters per second", } # Create the xarray dataset ds = xr.Dataset( {"streamflow": (["time"], values, attrs)}, coords={"time": times} ) # Set some nice attributes ds.attrs["title"] = "USGS Data from streamflow data" ds.attrs["station"] = station.name ds.attrs["stationid"] = station.get_uid() ds.attrs["location"] = (station.location.y, station.location.x) ds.to_netcdf(netcdf) return ds
src/ewatercycle/observation/usgs.py
3,905
Get river discharge data from the USGS REST web service. See `U.S. Geological Survey Water Services <https://waterservices.usgs.gov/>`_ (USGS) Parameters ---------- station_id : str The station id to get start_date : str String for start date in the format: 'YYYY-MM-dd', e.g. '1980-01-01' end_date : str String for start date in the format: 'YYYY-MM-dd', e.g. '2018-12-31' parameter : str The parameter code to get, e.g. ('00060') discharge, cubic feet per second cache_dir : str Directory where files retrieved from the web service are cached. If set to None then USGS_DATA_HOME env var will be used as cache directory. Examples -------- >>> from ewatercycle.observation.usgs import get_usgs_data >>> data = get_usgs_data('03109500', '2000-01-01', '2000-12-31', cache_dir='.') >>> data <xarray.Dataset> Dimensions: (time: 8032) Coordinates: * time (time) datetime64[ns] 2000-01-04T05:00:00 ... 2000-12-23T04:00:00 Data variables: Streamflow (time) float32 8.296758 10.420501 ... 10.647034 11.694747 Attributes: title: USGS Data from streamflow data station: Little Beaver Creek near East Liverpool OH stationid: 03109500 location: (40.6758974, -80.5406244) noqa: E501 Check if we have the netcdf data Download the data if needed Convert the raw data to an xarray We expect only 1 station Unit conversion from cubic feet to cubic meter per second Create the xarray dataset Set some nice attributes
1,516
en
0.448194
#!/usr/bin/python # # Copyright 2018-2020 Polyaxon, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # coding: utf-8 """ Polyaxon SDKs and REST API specification. Polyaxon SDKs and REST API specification. # noqa: E501 The version of the OpenAPI document: 1.1.7 Contact: contact@polyaxon.com Generated by: https://openapi-generator.tech """ from __future__ import absolute_import import re # noqa: F401 # python 2 and python 3 compatibility library import six from polyaxon_sdk.api_client import ApiClient from polyaxon_sdk.exceptions import ApiTypeError, ApiValueError # noqa: F401 class UsersV1Api(object): """NOTE: This class is auto generated by OpenAPI Generator Ref: https://openapi-generator.tech Do not edit the class manually. """ def __init__(self, api_client=None): if api_client is None: api_client = ApiClient() self.api_client = api_client def create_token(self, body, **kwargs): # noqa: E501 """Create token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_token(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.create_token_with_http_info(body, **kwargs) # noqa: E501 def create_token_with_http_info(self, body, **kwargs): # noqa: E501 """Create token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_token_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["body"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method create_token" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'body' is set if self.api_client.client_side_validation and ( "body" not in local_var_params or local_var_params["body"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `body` when calling `create_token`" ) # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if "body" in local_var_params: body_params = local_var_params["body"] # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # HTTP header `Content-Type` header_params[ "Content-Type" ] = self.api_client.select_header_content_type( # noqa: E501 ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens", "POST", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1Token", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def delete_token(self, uuid, **kwargs): # noqa: E501 """Delete token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_token(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: None If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.delete_token_with_http_info(uuid, **kwargs) # noqa: E501 def delete_token_with_http_info(self, uuid, **kwargs): # noqa: E501 """Delete token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_token_with_http_info(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: None If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["uuid"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method delete_token" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'uuid' is set if self.api_client.client_side_validation and ( "uuid" not in local_var_params or local_var_params["uuid"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `uuid` when calling `delete_token`" ) # noqa: E501 collection_formats = {} path_params = {} if "uuid" in local_var_params: path_params["uuid"] = local_var_params["uuid"] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens/{uuid}", "DELETE", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type=None, # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def get_token(self, uuid, **kwargs): # noqa: E501 """Get token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_token(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.get_token_with_http_info(uuid, **kwargs) # noqa: E501 def get_token_with_http_info(self, uuid, **kwargs): # noqa: E501 """Get token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_token_with_http_info(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["uuid"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method get_token" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'uuid' is set if self.api_client.client_side_validation and ( "uuid" not in local_var_params or local_var_params["uuid"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `uuid` when calling `get_token`" ) # noqa: E501 collection_formats = {} path_params = {} if "uuid" in local_var_params: path_params["uuid"] = local_var_params["uuid"] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens/{uuid}", "GET", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1Token", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def get_user(self, **kwargs): # noqa: E501 """Get current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.get_user_with_http_info(**kwargs) # noqa: E501 def get_user_with_http_info(self, **kwargs): # noqa: E501 """Get current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = [] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method get_user" % key ) local_var_params[key] = val del local_var_params["kwargs"] collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users", "GET", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1User", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def list_tokens(self, **kwargs): # noqa: E501 """List tokens # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_tokens(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param int offset: Pagination offset. :param int limit: Limit size. :param str sort: Sort to order the search. :param str query: Query filter the search search. :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1ListTokenResponse If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.list_tokens_with_http_info(**kwargs) # noqa: E501 def list_tokens_with_http_info(self, **kwargs): # noqa: E501 """List tokens # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_tokens_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param int offset: Pagination offset. :param int limit: Limit size. :param str sort: Sort to order the search. :param str query: Query filter the search search. :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1ListTokenResponse, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["offset", "limit", "sort", "query"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method list_tokens" % key ) local_var_params[key] = val del local_var_params["kwargs"] collection_formats = {} path_params = {} query_params = [] if ( "offset" in local_var_params and local_var_params["offset"] is not None ): # noqa: E501 query_params.append(("offset", local_var_params["offset"])) # noqa: E501 if ( "limit" in local_var_params and local_var_params["limit"] is not None ): # noqa: E501 query_params.append(("limit", local_var_params["limit"])) # noqa: E501 if ( "sort" in local_var_params and local_var_params["sort"] is not None ): # noqa: E501 query_params.append(("sort", local_var_params["sort"])) # noqa: E501 if ( "query" in local_var_params and local_var_params["query"] is not None ): # noqa: E501 query_params.append(("query", local_var_params["query"])) # noqa: E501 header_params = {} form_params = [] local_var_files = {} body_params = None # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens", "GET", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1ListTokenResponse", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def patch_token(self, token_uuid, body, **kwargs): # noqa: E501 """Patch token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_token(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.patch_token_with_http_info(token_uuid, body, **kwargs) # noqa: E501 def patch_token_with_http_info(self, token_uuid, body, **kwargs): # noqa: E501 """Patch token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_token_with_http_info(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["token_uuid", "body"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method patch_token" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'token_uuid' is set if self.api_client.client_side_validation and ( "token_uuid" not in local_var_params or local_var_params["token_uuid"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `token_uuid` when calling `patch_token`" ) # noqa: E501 # verify the required parameter 'body' is set if self.api_client.client_side_validation and ( "body" not in local_var_params or local_var_params["body"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `body` when calling `patch_token`" ) # noqa: E501 collection_formats = {} path_params = {} if "token_uuid" in local_var_params: path_params["token.uuid"] = local_var_params["token_uuid"] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if "body" in local_var_params: body_params = local_var_params["body"] # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # HTTP header `Content-Type` header_params[ "Content-Type" ] = self.api_client.select_header_content_type( # noqa: E501 ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens/{token.uuid}", "PATCH", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1Token", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def patch_user(self, body, **kwargs): # noqa: E501 """Patch current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_user(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.patch_user_with_http_info(body, **kwargs) # noqa: E501 def patch_user_with_http_info(self, body, **kwargs): # noqa: E501 """Patch current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_user_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["body"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method patch_user" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'body' is set if self.api_client.client_side_validation and ( "body" not in local_var_params or local_var_params["body"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `body` when calling `patch_user`" ) # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if "body" in local_var_params: body_params = local_var_params["body"] # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # HTTP header `Content-Type` header_params[ "Content-Type" ] = self.api_client.select_header_content_type( # noqa: E501 ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users", "PATCH", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1User", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def update_token(self, token_uuid, body, **kwargs): # noqa: E501 """Update token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_token(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.update_token_with_http_info( token_uuid, body, **kwargs ) # noqa: E501 def update_token_with_http_info(self, token_uuid, body, **kwargs): # noqa: E501 """Update token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_token_with_http_info(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["token_uuid", "body"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method update_token" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'token_uuid' is set if self.api_client.client_side_validation and ( "token_uuid" not in local_var_params or local_var_params["token_uuid"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `token_uuid` when calling `update_token`" ) # noqa: E501 # verify the required parameter 'body' is set if self.api_client.client_side_validation and ( "body" not in local_var_params or local_var_params["body"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `body` when calling `update_token`" ) # noqa: E501 collection_formats = {} path_params = {} if "token_uuid" in local_var_params: path_params["token.uuid"] = local_var_params["token_uuid"] # noqa: E501 query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if "body" in local_var_params: body_params = local_var_params["body"] # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # HTTP header `Content-Type` header_params[ "Content-Type" ] = self.api_client.select_header_content_type( # noqa: E501 ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users/tokens/{token.uuid}", "PUT", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1Token", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, ) def update_user(self, body, **kwargs): # noqa: E501 """Update current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_user(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. """ kwargs["_return_http_data_only"] = True return self.update_user_with_http_info(body, **kwargs) # noqa: E501 def update_user_with_http_info(self, body, **kwargs): # noqa: E501 """Update current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_user_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. """ local_var_params = locals() all_params = ["body"] all_params.extend( [ "async_req", "_return_http_data_only", "_preload_content", "_request_timeout", ] ) for key, val in six.iteritems(local_var_params["kwargs"]): if key not in all_params: raise ApiTypeError( "Got an unexpected keyword argument '%s'" " to method update_user" % key ) local_var_params[key] = val del local_var_params["kwargs"] # verify the required parameter 'body' is set if self.api_client.client_side_validation and ( "body" not in local_var_params or local_var_params["body"] is None # noqa: E501 ): # noqa: E501 raise ApiValueError( "Missing the required parameter `body` when calling `update_user`" ) # noqa: E501 collection_formats = {} path_params = {} query_params = [] header_params = {} form_params = [] local_var_files = {} body_params = None if "body" in local_var_params: body_params = local_var_params["body"] # HTTP header `Accept` header_params["Accept"] = self.api_client.select_header_accept( ["application/json"] ) # noqa: E501 # HTTP header `Content-Type` header_params[ "Content-Type" ] = self.api_client.select_header_content_type( # noqa: E501 ["application/json"] ) # noqa: E501 # Authentication setting auth_settings = ["ApiKey"] # noqa: E501 return self.api_client.call_api( "/api/v1/users", "PUT", path_params, query_params, header_params, body=body_params, post_params=form_params, files=local_var_files, response_type="V1User", # noqa: E501 auth_settings=auth_settings, async_req=local_var_params.get("async_req"), _return_http_data_only=local_var_params.get( "_return_http_data_only" ), # noqa: E501 _preload_content=local_var_params.get("_preload_content", True), _request_timeout=local_var_params.get("_request_timeout"), collection_formats=collection_formats, )
sdks/python/http_client/v1/polyaxon_sdk/api/users_v1_api.py
47,203
NOTE: This class is auto generated by OpenAPI Generator Ref: https://openapi-generator.tech Do not edit the class manually. Create token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_token(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. Create token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.create_token_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Delete token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_token(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: None If the method is called asynchronously, returns the request thread. Delete token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.delete_token_with_http_info(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: None If the method is called asynchronously, returns the request thread. Get token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_token(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. Get token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_token_with_http_info(uuid, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str uuid: UUid of the namespace (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Get current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. Get current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.get_user_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. List tokens # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_tokens(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param int offset: Pagination offset. :param int limit: Limit size. :param str sort: Sort to order the search. :param str query: Query filter the search search. :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1ListTokenResponse If the method is called asynchronously, returns the request thread. List tokens # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.list_tokens_with_http_info(async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param int offset: Pagination offset. :param int limit: Limit size. :param str sort: Sort to order the search. :param str query: Query filter the search search. :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1ListTokenResponse, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Patch token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_token(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. Patch token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_token_with_http_info(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Patch current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_user(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. Patch current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_user_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Update token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_token(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1Token If the method is called asynchronously, returns the request thread. Update token # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_token_with_http_info(token_uuid, body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param str token_uuid: UUID (required) :param V1Token body: Token body (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1Token, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Update current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_user(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: V1User If the method is called asynchronously, returns the request thread. Update current user # noqa: E501 This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.update_user_with_http_info(body, async_req=True) >>> result = thread.get() :param async_req bool: execute request asynchronously :param V1User body: (required) :param _return_http_data_only: response data without head status code and headers :param _preload_content: if False, the urllib3.HTTPResponse object will be returned without reading/decoding response data. Default is True. :param _request_timeout: timeout setting for this request. If one number provided, it will be total request timeout. It can also be a pair (tuple) of (connection, read) timeouts. :return: tuple(V1User, status_code(int), headers(HTTPHeaderDict)) If the method is called asynchronously, returns the request thread. Polyaxon SDKs and REST API specification. Polyaxon SDKs and REST API specification. # noqa: E501 The version of the OpenAPI document: 1.1.7 Contact: contact@polyaxon.com Generated by: https://openapi-generator.tech !/usr/bin/python Copyright 2018-2020 Polyaxon, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. coding: utf-8 noqa: F401 python 2 and python 3 compatibility library noqa: F401 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'body' is set noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 HTTP header `Content-Type` noqa: E501 noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'uuid' is set noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'uuid' is set noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'token_uuid' is set noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'body' is set noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 HTTP header `Content-Type` noqa: E501 noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'body' is set noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 HTTP header `Content-Type` noqa: E501 noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'token_uuid' is set noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'body' is set noqa: E501 noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 HTTP header `Content-Type` noqa: E501 noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 noqa: E501 verify the required parameter 'body' is set noqa: E501 noqa: E501 noqa: E501 HTTP header `Accept` noqa: E501 HTTP header `Content-Type` noqa: E501 noqa: E501 Authentication setting noqa: E501 noqa: E501 noqa: E501
20,775
en
0.639755
#!/home/anitha/Track/virtual/bin/python from django.core import management if __name__ == "__main__": management.execute_from_command_line()
virtual/bin/django-admin.py
146
!/home/anitha/Track/virtual/bin/python
38
en
0.283846
import concurrent.futures import datetime import io import logging import os import random import time import typing as t import discord import discord.ext.commands as commands from PIL import Image, ImageDraw, ImageSequence, ImageFont import bot.extensions as ext from bot.consts import Colors from bot.messaging.events import Events log = logging.getLogger(__name__) MAX_WALDO_GRID_SIZE = 100 CRAB_LINE_LENGTH = 58 CRAB_COMMAND_COOLDOWN = 3 def pillow_process(args, is_rave, lines_in_text, timestamp): # Open crab.gif and add our font with Image.open('bot/cogs/memes_cog/assets/crab.gif') as im: fnt = ImageFont.truetype('bot/cogs/memes_cog/assets/LemonMilk.otf', 11) # Draw text on each frame of the gif # Gonna be honest I don't quite understand how it works but I got it from the Pillow docs/issues frames = [] for frame in ImageSequence.Iterator(im): d = ImageDraw.Draw(frame) w, h = d.textsize(args, fnt) # draws the text on to the frame. Tries to center horizontally and tries to go as close to the bottom as possible d.text((im.size[0] / 2 - w / 2, im.size[1] - h - (5 * lines_in_text)), args, font=fnt, align='center', stroke_width=bool(is_rave), stroke_fill=Colors.ClemsonOrange, spacing=6) del d b = io.BytesIO() frame.save(b, format='GIF') frame = Image.open(b) frames.append(frame) frames[0].save(f'bot/cogs/memes_cog/assets/out_{timestamp}.gif', save_all=True, append_images=frames[1:]) class MemesCog(commands.Cog): def __init__(self, bot): self.bot = bot @ext.command() @ext.long_help( 'A fun command to generate a pseudo bubblewrap effect in discord' ) @ext.short_help('Creates bubblewrap!') @ext.example('bubblewrap') async def bubblewrap(self, ctx): msg = '' for _ in range(0, 5): for _ in range(0, 10): msg += '||pop!|| ' msg += '\n' await ctx.send(msg) @commands.command() @ext.long_help( 'A fun command to generate a wheres waldo effect in discord, see if you can find him first!' 'Optionally takes a size parameter to make it easier or harder' ) @ext.short_help('Can you find him?') @ext.example(('waldo', 'waldo 10')) async def waldo(self, ctx, size=MAX_WALDO_GRID_SIZE): """ Play Where's Waldo! Usage: <prefix>waldo [size = 100] """ random_start_letters = ['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', 'S', 'T', 'U', 'V', 'X', 'Y', 'Z'] max_waldo_line_size = 6 new_line_waldo_chance = 10 msg = '' count = 0 place = random.randint(0, size) for i in range(size + 1): if i == place: msg += '||`WALDO`|| ' count += 1 else: helper = random.randint(0, len(random_start_letters) - 1) letter = random_start_letters[helper] msg += f'||`{letter}ALDO`|| ' count += 1 new_line = random.randint(0, 100) if new_line < new_line_waldo_chance or count > max_waldo_line_size: msg += '\n' count = 0 await ctx.send(msg) @ext.command() @ext.chainable() @ext.long_help( 'A fun command to spongebob meme text in discord' ) @ext.short_help('sO yOu doNt KnOw wHat tHiS Is?') @ext.example('spongebob hello world') async def spongebob(self, ctx, *, args): """ Spongebob Text """ random.seed(time.time()) args = args.replace('"', "'") result = '' for i in args: helper = random.randint(0, 100) if helper > 60: result += str(i).upper() else: result += str(i).lower() await ctx.send(result) @ext.command(aliases=['rave', '🦀']) @commands.cooldown(1, CRAB_COMMAND_COOLDOWN, commands.BucketType.guild) @ext.long_help( 'A fun command to generate a crab rave gif with specified text overlay' ) @ext.short_help('Generates a crab rave gif') @ext.chainable_input() @ext.example('crab hello from crab world') async def crab(self, ctx, is_rave: t.Optional[bool] = True, *, args='Bottom text\n is dead'): """ Create your own crab rave. Usage: <prefix>crab [is_rave=True] [text=Bottom text\\n is dead] Aliases: rave, 🦀 """ # crab.gif dimensions - 352 by 200 # Immediately grab the timestamp incase of multiple calls in a row timestamp = datetime.datetime.utcnow().microsecond wait_msg = await ctx.send('Generating your gif') args = args.replace('\\', '') # Add new lines for when the text would go out of bounds lines_in_text = 1 while len(args) > (CRAB_LINE_LENGTH * lines_in_text): newline_loc = CRAB_LINE_LENGTH * lines_in_text # I didn't want to add a newline in the middle of a word while not args[newline_loc].isspace(): newline_loc -= 1 if newline_loc == CRAB_LINE_LENGTH * (lines_in_text - 1): newline_loc = CRAB_LINE_LENGTH * lines_in_text break args = f'{args[:newline_loc]} \n{args[newline_loc:]}' lines_in_text += 1 loop = self.bot.loop with concurrent.futures.ProcessPoolExecutor() as pool: pil_args = (args, is_rave, lines_in_text, timestamp) await loop.run_in_executor(pool, pillow_process, *pil_args) # Attach, send, and delete created gif attachment = discord.File(filename=f'out_{timestamp}.gif', fp=f'bot/cogs/memes_cog/assets/out_{timestamp}.gif') msg = await ctx.send(file=attachment) await self.bot.messenger.publish(Events.on_set_deletable, msg=msg, author=ctx.author) await wait_msg.delete() os.remove(f'bot/cogs/memes_cog/assets/out_{timestamp}.gif') @ext.command(hidden=True, aliases=['ctray', 'trayforjay']) async def cookouttray(self, ctx, input): """ For those who do finances with cookout trays, we proudly present the command for you Simply type one of the following: cookouttray ctray trayforjay Followed by a monetary value such as (leave off the dollar sign): 20 100 3.14 To have it converted into cookout trays Examples: cookouttray 20 ctray 100 trayforjay 3.14 Clicking the link "Cash to Cookout Tray Converter" in the output will also take you to cookout's website """ money = round(float(input), 2) output = money / 5 embed = discord.Embed( title='Cash to Cookout Tray Converter', description=f'{ctx.message.author.mention} ${money} is approximately {output} cookout trays', url=f"https://www.fastfoodmenuprices.com/cookout-prices/", color=Colors.ClemsonOrange) await ctx.send(embed=embed) def setup(bot): bot.add_cog(MemesCog(bot))
bot/cogs/memes_cog/memes_cog.py
7,666
Open crab.gif and add our font Draw text on each frame of the gif Gonna be honest I don't quite understand how it works but I got it from the Pillow docs/issues draws the text on to the frame. Tries to center horizontally and tries to go as close to the bottom as possible crab.gif dimensions - 352 by 200 Immediately grab the timestamp incase of multiple calls in a row Add new lines for when the text would go out of bounds I didn't want to add a newline in the middle of a word Attach, send, and delete created gif
517
en
0.936863
#! /usr/bin/env python3 # -*- coding: utf8 -*- # Virtual dancers that consumes real GigglePixel packets # # To use, start this up and then bring up a server broadcasting GigglePixel. # When this receives a palette packet, the dancing pair (whose humble wearables # are only capable of displaying one color at a time apiece) will light up # to match the first two elements of the packet received. When an ID packet # is received, they will shout their love of the sender. PORT = 7016 import socket import sys from time import time from x256 import x256 from udp import * WHITE = '\033[0m' RGB1 = None RGB2 = None banner = "Yay" note = u'♪' face = u'(・o・)' # Print without newline def p(s): sys.stdout.write(s) # Return a two-element array showing current arm position, and toggle it for next time arm_phase = False def arms(): global arm_phase arm_phase = not arm_phase if arm_phase: return u'┏┛' else: return u'┗┓' # Take an RGB value and return an ANSI escape sequence to show it in the terminal def color(rgb): if rgb is None: return "" ix = x256.from_rgb(*rgb) return "\033[38;5;%dm" % ix # Draw the dancers def draw(): l, r = arms() p (color(RGB1) + l + face + r + WHITE + ' ' + note + ' ') l, r = arms() p (color(RGB2) + l + face + r + WHITE + " -" + banner + "!") p ("\n\033[1A") # Keep drawing over and over on the same line def handle_packet(gp): global banner global RGB1 global RGB2 if gp is None: return if gp.packet_type == "PALETTE": entries = gp.payload["entries"] if len(entries) < 1: return elif len(entries) == 1: entries.extend(entries) RGB1 = (entries[0]["red"], entries[0]["green"], entries[0]["blue"]) RGB2 = (entries[1]["red"], entries[1]["green"], entries[1]["blue"]) elif gp.packet_type == "ID": banner = "We love " + gp.payload["name"] next_dance = time() listener = GigglePixelListener() try: while True: draw() now = time() time_left = next_dance - now gp = None if time_left > 0: gp = listener.get_packet(time_left) handle_packet(gp) if gp is None: next_dance = time() + 1 arms() # Toggle arm positions except KeyboardInterrupt: print (WHITE)
python-lib/example-consumer.py
2,240
! /usr/bin/env python3 -*- coding: utf8 -*- Virtual dancers that consumes real GigglePixel packets To use, start this up and then bring up a server broadcasting GigglePixel. When this receives a palette packet, the dancing pair (whose humble wearables are only capable of displaying one color at a time apiece) will light up to match the first two elements of the packet received. When an ID packet is received, they will shout their love of the sender. Print without newline Return a two-element array showing current arm position, and toggle it for next time Take an RGB value and return an ANSI escape sequence to show it in the terminal Draw the dancers Keep drawing over and over on the same line Toggle arm positions
722
en
0.865394
from __future__ import unicode_literals import io import os import re import sys from botocore.awsrequest import AWSPreparedRequest from moto.core.utils import ( amzn_request_id, str_to_rfc_1123_datetime, py2_strip_unicode_keys, ) from urllib.parse import ( parse_qs, parse_qsl, urlparse, unquote, urlencode, urlunparse, ) import xmltodict from moto.packages.httpretty.core import HTTPrettyRequest from moto.core.responses import _TemplateEnvironmentMixin, ActionAuthenticatorMixin from moto.core.utils import path_url from moto.core import ACCOUNT_ID from moto.settings import S3_IGNORE_SUBDOMAIN_BUCKETNAME from moto.s3bucket_path.utils import ( bucket_name_from_url as bucketpath_bucket_name_from_url, parse_key_name as bucketpath_parse_key_name, is_delete_keys as bucketpath_is_delete_keys, ) from .exceptions import ( BucketAlreadyExists, BucketMustHaveLockeEnabled, DuplicateTagKeys, InvalidContentMD5, InvalidContinuationToken, S3ClientError, MissingBucket, MissingKey, MissingVersion, InvalidMaxPartArgument, InvalidPartOrder, MalformedXML, MalformedACLError, IllegalLocationConstraintException, InvalidNotificationARN, InvalidNotificationEvent, ObjectNotInActiveTierError, NoSystemTags, PreconditionFailed, InvalidRange, LockNotEnabled, ) from .models import ( s3_backend, get_canned_acl, FakeGrantee, FakeGrant, FakeAcl, FakeKey, ) from .utils import ( bucket_name_from_url, clean_key_name, metadata_from_headers, parse_region_from_url, ) from xml.dom import minidom DEFAULT_REGION_NAME = "us-east-1" ACTION_MAP = { "BUCKET": { "HEAD": {"DEFAULT": "HeadBucket",}, "GET": { "uploads": "ListBucketMultipartUploads", "location": "GetBucketLocation", "lifecycle": "GetLifecycleConfiguration", "versioning": "GetBucketVersioning", "policy": "GetBucketPolicy", "website": "GetBucketWebsite", "acl": "GetBucketAcl", "tagging": "GetBucketTagging", "logging": "GetBucketLogging", "cors": "GetBucketCORS", "notification": "GetBucketNotification", "accelerate": "GetAccelerateConfiguration", "versions": "ListBucketVersions", "public_access_block": "GetPublicAccessBlock", "DEFAULT": "ListBucket", }, "PUT": { "lifecycle": "PutLifecycleConfiguration", "versioning": "PutBucketVersioning", "policy": "PutBucketPolicy", "website": "PutBucketWebsite", "acl": "PutBucketAcl", "tagging": "PutBucketTagging", "logging": "PutBucketLogging", "cors": "PutBucketCORS", "notification": "PutBucketNotification", "accelerate": "PutAccelerateConfiguration", "public_access_block": "PutPublicAccessBlock", "DEFAULT": "CreateBucket", }, "DELETE": { "lifecycle": "PutLifecycleConfiguration", "policy": "DeleteBucketPolicy", "website": "DeleteBucketWebsite", "tagging": "PutBucketTagging", "cors": "PutBucketCORS", "public_access_block": "DeletePublicAccessBlock", "DEFAULT": "DeleteBucket", }, }, "KEY": { "HEAD": {"DEFAULT": "HeadObject",}, "GET": { "uploadId": "ListMultipartUploadParts", "acl": "GetObjectAcl", "tagging": "GetObjectTagging", "versionId": "GetObjectVersion", "DEFAULT": "GetObject", }, "PUT": { "acl": "PutObjectAcl", "tagging": "PutObjectTagging", "DEFAULT": "PutObject", }, "DELETE": { "uploadId": "AbortMultipartUpload", "versionId": "DeleteObjectVersion", "DEFAULT": "DeleteObject", }, "POST": { "uploads": "PutObject", "restore": "RestoreObject", "uploadId": "PutObject", }, }, "CONTROL": { "GET": {"publicAccessBlock": "GetPublicAccessBlock"}, "PUT": {"publicAccessBlock": "PutPublicAccessBlock"}, "DELETE": {"publicAccessBlock": "DeletePublicAccessBlock"}, }, } def parse_key_name(pth): # strip the first '/' left by urlparse return pth[1:] if pth.startswith("/") else pth def is_delete_keys(request, path, bucket_name): # GOlang sends a request as url/?delete= (treating it as a normal key=value, even if the value is empty) # Python sends a request as url/?delete (treating it as a flag) # https://github.com/spulec/moto/issues/2937 return ( path == "/?delete" or path == "/?delete=" or (path == "/" and getattr(request, "query_string", "") == "delete") ) class ResponseObject(_TemplateEnvironmentMixin, ActionAuthenticatorMixin): def __init__(self, backend): super(ResponseObject, self).__init__() self.backend = backend self.method = "" self.path = "" self.data = {} self.headers = {} @property def should_autoescape(self): return True def all_buckets(self): self.data["Action"] = "ListAllMyBuckets" self._authenticate_and_authorize_s3_action() # No bucket specified. Listing all buckets all_buckets = self.backend.list_buckets() template = self.response_template(S3_ALL_BUCKETS) return template.render(buckets=all_buckets) def subdomain_based_buckets(self, request): if S3_IGNORE_SUBDOMAIN_BUCKETNAME: return False host = request.headers.get("host", request.headers.get("Host")) if not host: host = urlparse(request.url).netloc if ( not host or host.startswith("localhost") or host.startswith("localstack") or re.match(r"^[^.]+$", host) or re.match(r"^.*\.svc\.cluster\.local:?\d*$", host) ): # Default to path-based buckets for (1) localhost, (2) localstack hosts (e.g. localstack.dev), # (3) local host names that do not contain a "." (e.g., Docker container host names), or # (4) kubernetes host names return False match = re.match(r"^([^\[\]:]+)(:\d+)?$", host) if match: match = re.match( r"((25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)(\.|$)){4}", match.groups()[0] ) if match: return False match = re.match(r"^\[(.+)\](:\d+)?$", host) if match: match = re.match( r"^(((?=.*(::))(?!.*\3.+\3))\3?|[\dA-F]{1,4}:)([\dA-F]{1,4}(\3|:\b)|\2){5}(([\dA-F]{1,4}(\3|:\b|$)|\2){2}|(((2[0-4]|1\d|[1-9])?\d|25[0-5])\.?\b){4})\Z", match.groups()[0], re.IGNORECASE, ) if match: return False path_based = host == "s3.amazonaws.com" or re.match( r"s3[\.\-]([^.]*)\.amazonaws\.com", host ) return not path_based def is_delete_keys(self, request, path, bucket_name): if self.subdomain_based_buckets(request): return is_delete_keys(request, path, bucket_name) else: return bucketpath_is_delete_keys(request, path, bucket_name) def parse_bucket_name_from_url(self, request, url): if self.subdomain_based_buckets(request): return bucket_name_from_url(url) else: return bucketpath_bucket_name_from_url(url) def parse_key_name(self, request, url): if self.subdomain_based_buckets(request): return parse_key_name(url) else: return bucketpath_parse_key_name(url) def ambiguous_response(self, request, full_url, headers): # Depending on which calling format the client is using, we don't know # if this is a bucket or key request so we have to check if self.subdomain_based_buckets(request): return self.key_or_control_response(request, full_url, headers) else: # Using path-based buckets return self.bucket_response(request, full_url, headers) @amzn_request_id def bucket_response(self, request, full_url, headers): self.method = request.method self.path = self._get_path(request) self.headers = request.headers if "host" not in self.headers: self.headers["host"] = urlparse(full_url).netloc try: response = self._bucket_response(request, full_url, headers) except S3ClientError as s3error: response = s3error.code, {}, s3error.description return self._send_response(response) @staticmethod def _send_response(response): if isinstance(response, str): return 200, {}, response.encode("utf-8") else: status_code, headers, response_content = response if not isinstance(response_content, bytes): response_content = response_content.encode("utf-8") return status_code, headers, response_content def _bucket_response(self, request, full_url, headers): querystring = self._get_querystring(full_url) method = request.method region_name = parse_region_from_url(full_url) bucket_name = self.parse_bucket_name_from_url(request, full_url) if not bucket_name: # If no bucket specified, list all buckets return self.all_buckets() self.data["BucketName"] = bucket_name if hasattr(request, "body"): # Boto body = request.body else: # Flask server body = request.data if body is None: body = b"" if isinstance(body, bytes): body = body.decode("utf-8") body = "{0}".format(body).encode("utf-8") if method == "HEAD": return self._bucket_response_head(bucket_name, querystring) elif method == "GET": return self._bucket_response_get(bucket_name, querystring) elif method == "PUT": return self._bucket_response_put( request, body, region_name, bucket_name, querystring ) elif method == "DELETE": return self._bucket_response_delete(body, bucket_name, querystring) elif method == "POST": return self._bucket_response_post(request, body, bucket_name) else: raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format( method ) ) @staticmethod def _get_querystring(full_url): parsed_url = urlparse(full_url) querystring = parse_qs(parsed_url.query, keep_blank_values=True) return querystring def _bucket_response_head(self, bucket_name, querystring): self._set_action("BUCKET", "HEAD", querystring) self._authenticate_and_authorize_s3_action() try: self.backend.head_bucket(bucket_name) except MissingBucket: # Unless we do this, boto3 does not raise ClientError on # HEAD (which the real API responds with), and instead # raises NoSuchBucket, leading to inconsistency in # error response between real and mocked responses. return 404, {}, "" return 200, {}, "" def _bucket_response_get(self, bucket_name, querystring): self._set_action("BUCKET", "GET", querystring) self._authenticate_and_authorize_s3_action() if "object-lock" in querystring: ( lock_enabled, mode, days, years, ) = self.backend.get_object_lock_configuration(bucket_name) template = self.response_template(S3_BUCKET_LOCK_CONFIGURATION) return template.render( lock_enabled=lock_enabled, mode=mode, days=days, years=years, ) if "uploads" in querystring: for unsup in ("delimiter", "max-uploads"): if unsup in querystring: raise NotImplementedError( "Listing multipart uploads with {} has not been implemented yet.".format( unsup ) ) multiparts = list(self.backend.get_all_multiparts(bucket_name).values()) if "prefix" in querystring: prefix = querystring.get("prefix", [None])[0] multiparts = [ upload for upload in multiparts if upload.key_name.startswith(prefix) ] template = self.response_template(S3_ALL_MULTIPARTS) return template.render(bucket_name=bucket_name, uploads=multiparts) elif "location" in querystring: location = self.backend.get_bucket_location(bucket_name) template = self.response_template(S3_BUCKET_LOCATION) # us-east-1 is different - returns a None location if location == DEFAULT_REGION_NAME: location = None return template.render(location=location) elif "lifecycle" in querystring: rules = self.backend.get_bucket_lifecycle(bucket_name) if not rules: template = self.response_template(S3_NO_LIFECYCLE) return 404, {}, template.render(bucket_name=bucket_name) template = self.response_template(S3_BUCKET_LIFECYCLE_CONFIGURATION) return template.render(rules=rules) elif "versioning" in querystring: versioning = self.backend.get_bucket_versioning(bucket_name) template = self.response_template(S3_BUCKET_GET_VERSIONING) return template.render(status=versioning) elif "policy" in querystring: policy = self.backend.get_bucket_policy(bucket_name) if not policy: template = self.response_template(S3_NO_POLICY) return 404, {}, template.render(bucket_name=bucket_name) return 200, {}, policy elif "website" in querystring: website_configuration = self.backend.get_bucket_website_configuration( bucket_name ) if not website_configuration: template = self.response_template(S3_NO_BUCKET_WEBSITE_CONFIG) return 404, {}, template.render(bucket_name=bucket_name) return 200, {}, website_configuration elif "acl" in querystring: acl = self.backend.get_bucket_acl(bucket_name) template = self.response_template(S3_OBJECT_ACL_RESPONSE) return template.render(acl=acl) elif "tagging" in querystring: tags = self.backend.get_bucket_tagging(bucket_name)["Tags"] # "Special Error" if no tags: if len(tags) == 0: template = self.response_template(S3_NO_BUCKET_TAGGING) return 404, {}, template.render(bucket_name=bucket_name) template = self.response_template(S3_OBJECT_TAGGING_RESPONSE) return template.render(tags=tags) elif "logging" in querystring: logging = self.backend.get_bucket_logging(bucket_name) if not logging: template = self.response_template(S3_NO_LOGGING_CONFIG) return 200, {}, template.render() template = self.response_template(S3_LOGGING_CONFIG) return 200, {}, template.render(logging=logging) elif "cors" in querystring: cors = self.backend.get_bucket_cors(bucket_name) if len(cors) == 0: template = self.response_template(S3_NO_CORS_CONFIG) return 404, {}, template.render(bucket_name=bucket_name) template = self.response_template(S3_BUCKET_CORS_RESPONSE) return template.render(cors=cors) elif "notification" in querystring: notification_configuration = self.backend.get_bucket_notification_configuration( bucket_name ) if not notification_configuration: return 200, {}, "" template = self.response_template(S3_GET_BUCKET_NOTIFICATION_CONFIG) return template.render(config=notification_configuration) elif "accelerate" in querystring: bucket = self.backend.get_bucket(bucket_name) if bucket.accelerate_configuration is None: template = self.response_template(S3_BUCKET_ACCELERATE_NOT_SET) return 200, {}, template.render() template = self.response_template(S3_BUCKET_ACCELERATE) return template.render(bucket=bucket) elif "publicAccessBlock" in querystring: public_block_config = self.backend.get_public_access_block(bucket_name) template = self.response_template(S3_PUBLIC_ACCESS_BLOCK_CONFIGURATION) return template.render(public_block_config=public_block_config) elif "versions" in querystring: delimiter = querystring.get("delimiter", [None])[0] encoding_type = querystring.get("encoding-type", [None])[0] key_marker = querystring.get("key-marker", [None])[0] max_keys = querystring.get("max-keys", [None])[0] prefix = querystring.get("prefix", [""])[0] version_id_marker = querystring.get("version-id-marker", [None])[0] bucket = self.backend.get_bucket(bucket_name) ( versions, common_prefixes, delete_markers, ) = self.backend.list_object_versions( bucket_name, delimiter=delimiter, encoding_type=encoding_type, key_marker=key_marker, max_keys=max_keys, version_id_marker=version_id_marker, prefix=prefix, ) key_list = versions template = self.response_template(S3_BUCKET_GET_VERSIONS) return ( 200, {}, template.render( common_prefixes=common_prefixes, key_list=key_list, delete_marker_list=delete_markers, bucket=bucket, prefix=prefix, max_keys=1000, delimiter=delimiter, key_marker=key_marker, is_truncated="false", ), ) elif "encryption" in querystring: encryption = self.backend.get_bucket_encryption(bucket_name) if not encryption: template = self.response_template(S3_NO_ENCRYPTION) return 404, {}, template.render(bucket_name=bucket_name) template = self.response_template(S3_ENCRYPTION_CONFIG) return 200, {}, template.render(encryption=encryption) elif querystring.get("list-type", [None])[0] == "2": return 200, {}, self._handle_list_objects_v2(bucket_name, querystring) bucket = self.backend.get_bucket(bucket_name) prefix = querystring.get("prefix", [None])[0] if prefix and isinstance(prefix, bytes): prefix = prefix.decode("utf-8") delimiter = querystring.get("delimiter", [None])[0] max_keys = int(querystring.get("max-keys", [1000])[0]) marker = querystring.get("marker", [None])[0] result_keys, result_folders = self.backend.list_objects( bucket, prefix, delimiter ) if marker: result_keys = self._get_results_from_token(result_keys, marker) result_keys, is_truncated, next_marker = self._truncate_result( result_keys, max_keys ) template = self.response_template(S3_BUCKET_GET_RESPONSE) return ( 200, {}, template.render( bucket=bucket, prefix=prefix, delimiter=delimiter, result_keys=result_keys, result_folders=result_folders, is_truncated=is_truncated, next_marker=next_marker, max_keys=max_keys, ), ) def _set_action(self, action_resource_type, method, querystring): action_set = False for action_in_querystring, action in ACTION_MAP[action_resource_type][ method ].items(): if action_in_querystring in querystring: self.data["Action"] = action action_set = True if not action_set: self.data["Action"] = ACTION_MAP[action_resource_type][method]["DEFAULT"] def _handle_list_objects_v2(self, bucket_name, querystring): template = self.response_template(S3_BUCKET_GET_RESPONSE_V2) bucket = self.backend.get_bucket(bucket_name) continuation_token = querystring.get("continuation-token", [None])[0] if continuation_token is not None and continuation_token == "": raise InvalidContinuationToken() prefix = querystring.get("prefix", [None])[0] if prefix and isinstance(prefix, bytes): prefix = prefix.decode("utf-8") delimiter = querystring.get("delimiter", [None])[0] all_keys = self.backend.list_objects_v2(bucket, prefix, delimiter) fetch_owner = querystring.get("fetch-owner", [False])[0] max_keys = int(querystring.get("max-keys", [1000])[0]) start_after = querystring.get("start-after", [None])[0] if continuation_token or start_after: limit = continuation_token or start_after all_keys = self._get_results_from_token(all_keys, limit) truncated_keys, is_truncated, next_continuation_token = self._truncate_result( all_keys, max_keys ) result_keys, result_folders = self._split_truncated_keys(truncated_keys) key_count = len(result_keys) + len(result_folders) return template.render( bucket=bucket, prefix=prefix or "", delimiter=delimiter, key_count=key_count, result_keys=result_keys, result_folders=result_folders, fetch_owner=fetch_owner, max_keys=max_keys, is_truncated=is_truncated, next_continuation_token=next_continuation_token, start_after=None if continuation_token else start_after, ) @staticmethod def _split_truncated_keys(truncated_keys): result_keys = [] result_folders = [] for key in truncated_keys: if isinstance(key, FakeKey): result_keys.append(key) else: result_folders.append(key) return result_keys, result_folders def _get_results_from_token(self, result_keys, token): continuation_index = 0 for key in result_keys: if (key.name if isinstance(key, FakeKey) else key) > token: break continuation_index += 1 return result_keys[continuation_index:] def _truncate_result(self, result_keys, max_keys): if max_keys == 0: result_keys = [] is_truncated = True next_continuation_token = None elif len(result_keys) > max_keys: is_truncated = "true" result_keys = result_keys[:max_keys] item = result_keys[-1] next_continuation_token = item.name if isinstance(item, FakeKey) else item else: is_truncated = "false" next_continuation_token = None return result_keys, is_truncated, next_continuation_token def _body_contains_location_constraint(self, body): if body: try: xmltodict.parse(body)["CreateBucketConfiguration"]["LocationConstraint"] return True except KeyError: pass return False def _create_bucket_configuration_is_empty(self, body): if body: try: create_bucket_configuration = xmltodict.parse(body)[ "CreateBucketConfiguration" ] del create_bucket_configuration["@xmlns"] if len(create_bucket_configuration) == 0: return True except KeyError: pass return False def _parse_pab_config(self, body): parsed_xml = xmltodict.parse(body) parsed_xml["PublicAccessBlockConfiguration"].pop("@xmlns", None) # If Python 2, fix the unicode strings: if sys.version_info[0] < 3: parsed_xml = { "PublicAccessBlockConfiguration": py2_strip_unicode_keys( dict(parsed_xml["PublicAccessBlockConfiguration"]) ) } return parsed_xml def _bucket_response_put( self, request, body, region_name, bucket_name, querystring ): if not request.headers.get("Content-Length"): return 411, {}, "Content-Length required" self._set_action("BUCKET", "PUT", querystring) self._authenticate_and_authorize_s3_action() if "object-lock" in querystring: body_decoded = body.decode() config = self._lock_config_from_xml(body_decoded) if not self.backend.get_bucket(bucket_name).object_lock_enabled: raise BucketMustHaveLockeEnabled self.backend.put_object_lock_configuration( bucket_name, config.get("enabled"), config.get("mode"), config.get("days"), config.get("years"), ) return 200, {}, "" if "versioning" in querystring: ver = re.search("<Status>([A-Za-z]+)</Status>", body.decode()) if ver: self.backend.set_bucket_versioning(bucket_name, ver.group(1)) template = self.response_template(S3_BUCKET_VERSIONING) return template.render(bucket_versioning_status=ver.group(1)) else: return 404, {}, "" elif "lifecycle" in querystring: rules = xmltodict.parse(body)["LifecycleConfiguration"]["Rule"] if not isinstance(rules, list): # If there is only one rule, xmldict returns just the item rules = [rules] self.backend.put_bucket_lifecycle(bucket_name, rules) return "" elif "policy" in querystring: self.backend.put_bucket_policy(bucket_name, body) return "True" elif "acl" in querystring: # Headers are first. If not set, then look at the body (consistent with the documentation): acls = self._acl_from_headers(request.headers) if not acls: acls = self._acl_from_xml(body) self.backend.put_bucket_acl(bucket_name, acls) return "" elif "tagging" in querystring: tagging = self._bucket_tagging_from_xml(body) self.backend.put_bucket_tagging(bucket_name, tagging) return "" elif "website" in querystring: self.backend.set_bucket_website_configuration(bucket_name, body) return "" elif "cors" in querystring: try: self.backend.put_bucket_cors(bucket_name, self._cors_from_xml(body)) return "" except KeyError: raise MalformedXML() elif "logging" in querystring: try: self.backend.put_bucket_logging( bucket_name, self._logging_from_xml(body) ) return "" except KeyError: raise MalformedXML() elif "notification" in querystring: try: self.backend.put_bucket_notification_configuration( bucket_name, self._notification_config_from_xml(body) ) return "" except KeyError: raise MalformedXML() except Exception as e: raise e elif "accelerate" in querystring: try: accelerate_status = self._accelerate_config_from_xml(body) self.backend.put_bucket_accelerate_configuration( bucket_name, accelerate_status ) return "" except KeyError: raise MalformedXML() except Exception as e: raise e elif "publicAccessBlock" in querystring: pab_config = self._parse_pab_config(body) self.backend.put_bucket_public_access_block( bucket_name, pab_config["PublicAccessBlockConfiguration"] ) return "" elif "encryption" in querystring: try: self.backend.put_bucket_encryption( bucket_name, self._encryption_config_from_xml(body) ) return "" except KeyError: raise MalformedXML() except Exception as e: raise e else: # us-east-1, the default AWS region behaves a bit differently # - you should not use it as a location constraint --> it fails # - querying the location constraint returns None # - LocationConstraint has to be specified if outside us-east-1 if ( region_name != DEFAULT_REGION_NAME and not self._body_contains_location_constraint(body) ): raise IllegalLocationConstraintException() if body: if self._create_bucket_configuration_is_empty(body): raise MalformedXML() try: forced_region = xmltodict.parse(body)["CreateBucketConfiguration"][ "LocationConstraint" ] if forced_region == DEFAULT_REGION_NAME: raise S3ClientError( "InvalidLocationConstraint", "The specified location-constraint is not valid", ) else: region_name = forced_region except KeyError: pass try: new_bucket = self.backend.create_bucket(bucket_name, region_name) except BucketAlreadyExists: new_bucket = self.backend.get_bucket(bucket_name) if ( new_bucket.region_name == DEFAULT_REGION_NAME and region_name == DEFAULT_REGION_NAME ): # us-east-1 has different behavior - creating a bucket there is an idempotent operation pass else: template = self.response_template(S3_DUPLICATE_BUCKET_ERROR) return 409, {}, template.render(bucket_name=bucket_name) if "x-amz-acl" in request.headers: # TODO: Support the XML-based ACL format self.backend.put_bucket_acl( bucket_name, self._acl_from_headers(request.headers) ) if ( request.headers.get("x-amz-bucket-object-lock-enabled", "").lower() == "true" ): new_bucket.object_lock_enabled = True new_bucket.versioning_status = "Enabled" template = self.response_template(S3_BUCKET_CREATE_RESPONSE) return 200, {}, template.render(bucket=new_bucket) def _bucket_response_delete(self, body, bucket_name, querystring): self._set_action("BUCKET", "DELETE", querystring) self._authenticate_and_authorize_s3_action() if "policy" in querystring: self.backend.delete_bucket_policy(bucket_name, body) return 204, {}, "" elif "tagging" in querystring: self.backend.delete_bucket_tagging(bucket_name) return 204, {}, "" elif "website" in querystring: self.backend.delete_bucket_website(bucket_name) return 204, {}, "" elif "cors" in querystring: self.backend.delete_bucket_cors(bucket_name) return 204, {}, "" elif "lifecycle" in querystring: self.backend.delete_bucket_lifecycle(bucket_name) return 204, {}, "" elif "publicAccessBlock" in querystring: self.backend.delete_public_access_block(bucket_name) return 204, {}, "" elif "encryption" in querystring: self.backend.delete_bucket_encryption(bucket_name) return 204, {}, "" removed_bucket = self.backend.delete_bucket(bucket_name) if removed_bucket: # Bucket exists template = self.response_template(S3_DELETE_BUCKET_SUCCESS) return 204, {}, template.render(bucket=removed_bucket) else: # Tried to delete a bucket that still has keys template = self.response_template(S3_DELETE_BUCKET_WITH_ITEMS_ERROR) return 409, {}, template.render(bucket=removed_bucket) def _bucket_response_post(self, request, body, bucket_name): response_headers = {} if not request.headers.get("Content-Length"): return 411, {}, "Content-Length required" path = self._get_path(request) if self.is_delete_keys(request, path, bucket_name): self.data["Action"] = "DeleteObject" self._authenticate_and_authorize_s3_action() return self._bucket_response_delete_keys(request, body, bucket_name) self.data["Action"] = "PutObject" self._authenticate_and_authorize_s3_action() # POST to bucket-url should create file from form if hasattr(request, "form"): # Not HTTPretty form = request.form else: # HTTPretty, build new form object body = body.decode() form = dict(parse_qsl(body)) key = form["key"] if "file" in form: f = form["file"] else: fobj = request.files["file"] f = fobj.stream.read() key = key.replace("${filename}", os.path.basename(fobj.filename)) if "success_action_redirect" in form: redirect = form["success_action_redirect"] parts = urlparse(redirect) queryargs = parse_qs(parts.query) queryargs["key"] = key queryargs["bucket"] = bucket_name redirect_queryargs = urlencode(queryargs, doseq=True) newparts = ( parts.scheme, parts.netloc, parts.path, parts.params, redirect_queryargs, parts.fragment, ) fixed_redirect = urlunparse(newparts) response_headers["Location"] = fixed_redirect if "success_action_status" in form: status_code = form["success_action_status"] elif "success_action_redirect" in form: status_code = 303 else: status_code = 204 new_key = self.backend.put_object(bucket_name, key, f) if form.get("acl"): acl = get_canned_acl(form.get("acl")) new_key.set_acl(acl) # Metadata metadata = metadata_from_headers(form) new_key.set_metadata(metadata) return status_code, response_headers, "" @staticmethod def _get_path(request): if isinstance(request, HTTPrettyRequest): path = request.path else: path = ( request.full_path if hasattr(request, "full_path") else path_url(request.url) ) return path def _bucket_response_delete_keys(self, request, body, bucket_name): template = self.response_template(S3_DELETE_KEYS_RESPONSE) body_dict = xmltodict.parse(body, strip_whitespace=False) objects = body_dict["Delete"].get("Object", []) if not isinstance(objects, list): # We expect a list of objects, but when there is a single <Object> node xmltodict does not # return a list. objects = [objects] if len(objects) == 0: raise MalformedXML() deleted_objects = self.backend.delete_objects(bucket_name, objects) error_names = [] return ( 200, {}, template.render(deleted=deleted_objects, delete_errors=error_names), ) def _handle_range_header(self, request, headers, response_content): response_headers = {} length = len(response_content) last = length - 1 _, rspec = request.headers.get("range").split("=") if "," in rspec: raise NotImplementedError("Multiple range specifiers not supported") def toint(i): return int(i) if i else None begin, end = map(toint, rspec.split("-")) if begin is not None: # byte range end = last if end is None else min(end, last) elif end is not None: # suffix byte range begin = length - min(end, length) end = last else: return 400, response_headers, "" if begin < 0 or end > last or begin > min(end, last): raise InvalidRange( actual_size=str(length), range_requested=request.headers.get("range") ) response_headers["content-range"] = "bytes {0}-{1}/{2}".format( begin, end, length ) content = response_content[begin : end + 1] response_headers["content-length"] = len(content) return 206, response_headers, content def _handle_v4_chunk_signatures(self, body, content_length): body_io = io.BytesIO(body) new_body = bytearray(content_length) pos = 0 line = body_io.readline() while line: # https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-streaming.html#sigv4-chunked-body-definition # str(hex(chunk-size)) + ";chunk-signature=" + signature + \r\n + chunk-data + \r\n chunk_size = int(line[: line.find(b";")].decode("utf8"), 16) new_body[pos : pos + chunk_size] = body_io.read(chunk_size) pos = pos + chunk_size body_io.read(2) # skip trailing \r\n line = body_io.readline() return bytes(new_body) @amzn_request_id def key_or_control_response(self, request, full_url, headers): # Key and Control are lumped in because splitting out the regex is too much of a pain :/ self.method = request.method self.path = self._get_path(request) self.headers = request.headers if "host" not in self.headers: self.headers["host"] = urlparse(full_url).netloc response_headers = {} try: # Is this an S3 control response? if isinstance(request, AWSPreparedRequest) and "s3-control" in request.url: response = self._control_response(request, full_url, headers) else: response = self._key_response(request, full_url, self.headers) except S3ClientError as s3error: response = s3error.code, {}, s3error.description if isinstance(response, str): status_code = 200 response_content = response else: status_code, response_headers, response_content = response if ( status_code == 200 and "range" in request.headers and request.headers["range"] != "" ): try: return self._handle_range_header( request, response_headers, response_content ) except S3ClientError as s3error: return s3error.code, {}, s3error.description return status_code, response_headers, response_content def _control_response(self, request, full_url, headers): parsed_url = urlparse(full_url) query = parse_qs(parsed_url.query, keep_blank_values=True) method = request.method if hasattr(request, "body"): # Boto body = request.body if hasattr(body, "read"): body = body.read() else: # Flask server body = request.data if body is None: body = b"" if method == "GET": return self._control_response_get(request, query, headers) elif method == "PUT": return self._control_response_put(request, body, query, headers) elif method == "DELETE": return self._control_response_delete(request, query, headers) else: raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format( method ) ) def _control_response_get(self, request, query, headers): action = self.path.split("?")[0].split("/")[ -1 ] # Gets the action out of the URL sans query params. self._set_action("CONTROL", "GET", action) self._authenticate_and_authorize_s3_action() response_headers = {} if "publicAccessBlock" in action: public_block_config = self.backend.get_account_public_access_block( headers["x-amz-account-id"] ) template = self.response_template(S3_PUBLIC_ACCESS_BLOCK_CONFIGURATION) return ( 200, response_headers, template.render(public_block_config=public_block_config), ) raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format(action) ) def _control_response_put(self, request, body, query, headers): action = self.path.split("?")[0].split("/")[ -1 ] # Gets the action out of the URL sans query params. self._set_action("CONTROL", "PUT", action) self._authenticate_and_authorize_s3_action() response_headers = {} if "publicAccessBlock" in action: pab_config = self._parse_pab_config(body) self.backend.put_account_public_access_block( headers["x-amz-account-id"], pab_config["PublicAccessBlockConfiguration"], ) return 200, response_headers, "" raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format(action) ) def _control_response_delete(self, request, query, headers): action = self.path.split("?")[0].split("/")[ -1 ] # Gets the action out of the URL sans query params. self._set_action("CONTROL", "DELETE", action) self._authenticate_and_authorize_s3_action() response_headers = {} if "publicAccessBlock" in action: self.backend.delete_account_public_access_block(headers["x-amz-account-id"]) return 200, response_headers, "" raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format(action) ) def _key_response(self, request, full_url, headers): parsed_url = urlparse(full_url) query = parse_qs(parsed_url.query, keep_blank_values=True) method = request.method key_name = self.parse_key_name(request, parsed_url.path) bucket_name = self.parse_bucket_name_from_url(request, full_url) # Because we patch the requests library the boto/boto3 API # requests go through this method but so do # `requests.get("https://bucket-name.s3.amazonaws.com/file-name")` # Here we deny public access to private files by checking the # ACL and checking for the mere presence of an Authorization # header. if "Authorization" not in request.headers: if hasattr(request, "url"): signed_url = "Signature=" in request.url elif hasattr(request, "requestline"): signed_url = "Signature=" in request.path key = self.backend.get_object(bucket_name, key_name) if key: if not key.acl.public_read and not signed_url: return 403, {}, "" elif signed_url: # coming in from requests.get(s3.generate_presigned_url()) if self._invalid_headers(request.url, dict(request.headers)): return 403, {}, S3_INVALID_PRESIGNED_PARAMETERS if hasattr(request, "body"): # Boto body = request.body if hasattr(body, "read"): body = body.read() else: # Flask server body = request.data if not body: # when the data is being passed as a file if request.files: for _, value in request.files.items(): body = value.stream.read() elif hasattr(request, "form"): # Body comes through as part of the form, if no content-type is set on the PUT-request # form = ImmutableMultiDict([('some data 123 321', '')]) form = request.form for k, _ in form.items(): body = k if body is None: body = b"" if ( request.headers.get("x-amz-content-sha256", None) == "STREAMING-AWS4-HMAC-SHA256-PAYLOAD" ): body = self._handle_v4_chunk_signatures( body, int(request.headers["x-amz-decoded-content-length"]) ) if method == "GET": return self._key_response_get( bucket_name, query, key_name, headers=request.headers ) elif method == "PUT": return self._key_response_put( request, body, bucket_name, query, key_name, headers ) elif method == "HEAD": return self._key_response_head( bucket_name, query, key_name, headers=request.headers ) elif method == "DELETE": return self._key_response_delete(headers, bucket_name, query, key_name) elif method == "POST": return self._key_response_post(request, body, bucket_name, query, key_name) else: raise NotImplementedError( "Method {0} has not been implemented in the S3 backend yet".format( method ) ) def _key_response_get(self, bucket_name, query, key_name, headers): self._set_action("KEY", "GET", query) self._authenticate_and_authorize_s3_action() response_headers = {} if query.get("uploadId"): upload_id = query["uploadId"][0] # 0 <= PartNumberMarker <= 2,147,483,647 part_number_marker = int(query.get("part-number-marker", [0])[0]) if not (0 <= part_number_marker <= 2147483647): raise InvalidMaxPartArgument("part-number-marker", 0, 2147483647) # 0 <= MaxParts <= 2,147,483,647 (default is 1,000) max_parts = int(query.get("max-parts", [1000])[0]) if not (0 <= max_parts <= 2147483647): raise InvalidMaxPartArgument("max-parts", 0, 2147483647) parts = self.backend.list_parts( bucket_name, upload_id, part_number_marker=part_number_marker, max_parts=max_parts, ) next_part_number_marker = parts[-1].name + 1 if parts else 0 is_truncated = parts and self.backend.is_truncated( bucket_name, upload_id, next_part_number_marker ) template = self.response_template(S3_MULTIPART_LIST_RESPONSE) return ( 200, response_headers, template.render( bucket_name=bucket_name, key_name=key_name, upload_id=upload_id, is_truncated=str(is_truncated).lower(), max_parts=max_parts, next_part_number_marker=next_part_number_marker, parts=parts, part_number_marker=part_number_marker, ), ) version_id = query.get("versionId", [None])[0] if_modified_since = headers.get("If-Modified-Since", None) if_match = headers.get("If-Match", None) if_none_match = headers.get("If-None-Match", None) if_unmodified_since = headers.get("If-Unmodified-Since", None) key = self.backend.get_object(bucket_name, key_name, version_id=version_id) if key is None and version_id is None: raise MissingKey(key_name) elif key is None: raise MissingVersion() if if_unmodified_since: if_unmodified_since = str_to_rfc_1123_datetime(if_unmodified_since) if key.last_modified > if_unmodified_since: raise PreconditionFailed("If-Unmodified-Since") if if_match and key.etag not in [if_match, '"{0}"'.format(if_match)]: raise PreconditionFailed("If-Match") if if_modified_since: if_modified_since = str_to_rfc_1123_datetime(if_modified_since) if key.last_modified < if_modified_since: return 304, response_headers, "Not Modified" if if_none_match and key.etag == if_none_match: return 304, response_headers, "Not Modified" if "acl" in query: acl = s3_backend.get_object_acl(key) template = self.response_template(S3_OBJECT_ACL_RESPONSE) return 200, response_headers, template.render(acl=acl) if "tagging" in query: tags = self.backend.get_object_tagging(key)["Tags"] template = self.response_template(S3_OBJECT_TAGGING_RESPONSE) return 200, response_headers, template.render(tags=tags) if "legal-hold" in query: legal_hold = self.backend.get_object_legal_hold(key) template = self.response_template(S3_OBJECT_LEGAL_HOLD) return 200, response_headers, template.render(legal_hold=legal_hold) response_headers.update(key.metadata) response_headers.update(key.response_dict) return 200, response_headers, key.value def _key_response_put(self, request, body, bucket_name, query, key_name, headers): self._set_action("KEY", "PUT", query) self._authenticate_and_authorize_s3_action() response_headers = {} if query.get("uploadId") and query.get("partNumber"): upload_id = query["uploadId"][0] part_number = int(query["partNumber"][0]) if "x-amz-copy-source" in request.headers: src = unquote(request.headers.get("x-amz-copy-source")).lstrip("/") src_bucket, src_key = src.split("/", 1) src_key, src_version_id = ( src_key.split("?versionId=") if "?versionId=" in src_key else (src_key, None) ) src_range = request.headers.get("x-amz-copy-source-range", "").split( "bytes=" )[-1] try: start_byte, end_byte = src_range.split("-") start_byte, end_byte = int(start_byte), int(end_byte) except ValueError: start_byte, end_byte = None, None if self.backend.get_object( src_bucket, src_key, version_id=src_version_id ): key = self.backend.copy_part( bucket_name, upload_id, part_number, src_bucket, src_key, src_version_id, start_byte, end_byte, ) else: return 404, response_headers, "" template = self.response_template(S3_MULTIPART_UPLOAD_RESPONSE) response = template.render(part=key) else: key = self.backend.upload_part( bucket_name, upload_id, part_number, body ) response = "" response_headers.update(key.response_dict) return 200, response_headers, response storage_class = request.headers.get("x-amz-storage-class", "STANDARD") encryption = request.headers.get("x-amz-server-side-encryption", None) kms_key_id = request.headers.get( "x-amz-server-side-encryption-aws-kms-key-id", None ) bucket_key_enabled = request.headers.get( "x-amz-server-side-encryption-bucket-key-enabled", None ) if bucket_key_enabled is not None: bucket_key_enabled = str(bucket_key_enabled).lower() bucket = self.backend.get_bucket(bucket_name) lock_enabled = bucket.object_lock_enabled lock_mode = request.headers.get("x-amz-object-lock-mode", None) lock_until = request.headers.get("x-amz-object-lock-retain-until-date", None) legal_hold = request.headers.get("x-amz-object-lock-legal-hold", "OFF") if lock_mode or lock_until or legal_hold == "ON": if not request.headers.get("Content-Md5"): raise InvalidContentMD5 if not lock_enabled: raise LockNotEnabled elif lock_enabled and bucket.has_default_lock: if not request.headers.get("Content-Md5"): raise InvalidContentMD5 lock_until = bucket.default_retention() lock_mode = bucket.default_lock_mode acl = self._acl_from_headers(request.headers) if acl is None: acl = self.backend.get_bucket(bucket_name).acl tagging = self._tagging_from_headers(request.headers) if "retention" in query: if not lock_enabled: raise LockNotEnabled version_id = query.get("VersionId") retention = self._mode_until_from_xml(body) self.backend.put_object_retention( bucket_name, key_name, version_id=version_id, retention=retention ) return 200, response_headers, "" if "legal-hold" in query: if not lock_enabled: raise LockNotEnabled version_id = query.get("VersionId") legal_hold_status = self._legal_hold_status_from_xml(body) self.backend.put_object_legal_hold( bucket_name, key_name, version_id, legal_hold_status ) return 200, response_headers, "" if "acl" in query: self.backend.put_object_acl(bucket_name, key_name, acl) return 200, response_headers, "" if "tagging" in query: if "versionId" in query: version_id = query["versionId"][0] else: version_id = None key = self.backend.get_object(bucket_name, key_name, version_id=version_id) tagging = self._tagging_from_xml(body) self.backend.set_key_tags(key, tagging, key_name) return 200, response_headers, "" if "x-amz-copy-source" in request.headers: # Copy key # you can have a quoted ?version=abc with a version Id, so work on # we need to parse the unquoted string first src_key = request.headers.get("x-amz-copy-source") if isinstance(src_key, bytes): src_key = src_key.decode("utf-8") src_key_parsed = urlparse(src_key) src_bucket, src_key = ( clean_key_name(src_key_parsed.path).lstrip("/").split("/", 1) ) src_version_id = parse_qs(src_key_parsed.query).get("versionId", [None])[0] key = self.backend.get_object( src_bucket, src_key, version_id=src_version_id ) if key is not None: if key.storage_class in ["GLACIER", "DEEP_ARCHIVE"]: if key.response_dict.get( "x-amz-restore" ) is None or 'ongoing-request="true"' in key.response_dict.get( "x-amz-restore" ): raise ObjectNotInActiveTierError(key) self.backend.copy_object( src_bucket, src_key, bucket_name, key_name, storage=storage_class, acl=acl, src_version_id=src_version_id, ) else: return 404, response_headers, "" new_key = self.backend.get_object(bucket_name, key_name) mdirective = request.headers.get("x-amz-metadata-directive") if mdirective is not None and mdirective == "REPLACE": metadata = metadata_from_headers(request.headers) new_key.set_metadata(metadata, replace=True) tdirective = request.headers.get("x-amz-tagging-directive") if tdirective == "REPLACE": tagging = self._tagging_from_headers(request.headers) self.backend.set_key_tags(new_key, tagging) template = self.response_template(S3_OBJECT_COPY_RESPONSE) response_headers.update(new_key.response_dict) return 200, response_headers, template.render(key=new_key) streaming_request = hasattr(request, "streaming") and request.streaming closing_connection = headers.get("connection") == "close" if closing_connection and streaming_request: # Closing the connection of a streaming request. No more data new_key = self.backend.get_object(bucket_name, key_name) elif streaming_request: # Streaming request, more data new_key = self.backend.append_to_key(bucket_name, key_name, body) else: # Initial data new_key = self.backend.put_object( bucket_name, key_name, body, storage=storage_class, encryption=encryption, kms_key_id=kms_key_id, bucket_key_enabled=bucket_key_enabled, lock_mode=lock_mode, lock_legal_status=legal_hold, lock_until=lock_until, ) request.streaming = True metadata = metadata_from_headers(request.headers) metadata.update(metadata_from_headers(query)) new_key.set_metadata(metadata) new_key.set_acl(acl) new_key.website_redirect_location = request.headers.get( "x-amz-website-redirect-location" ) self.backend.set_key_tags(new_key, tagging) response_headers.update(new_key.response_dict) return 200, response_headers, "" def _key_response_head(self, bucket_name, query, key_name, headers): self._set_action("KEY", "HEAD", query) self._authenticate_and_authorize_s3_action() response_headers = {} version_id = query.get("versionId", [None])[0] part_number = query.get("partNumber", [None])[0] if part_number: part_number = int(part_number) if_modified_since = headers.get("If-Modified-Since", None) if_match = headers.get("If-Match", None) if_none_match = headers.get("If-None-Match", None) if_unmodified_since = headers.get("If-Unmodified-Since", None) key = self.backend.head_object( bucket_name, key_name, version_id=version_id, part_number=part_number ) if key: response_headers.update(key.metadata) response_headers.update(key.response_dict) if if_unmodified_since: if_unmodified_since = str_to_rfc_1123_datetime(if_unmodified_since) if key.last_modified > if_unmodified_since: return 412, response_headers, "" if if_match and key.etag != if_match: return 412, response_headers, "" if if_modified_since: if_modified_since = str_to_rfc_1123_datetime(if_modified_since) if key.last_modified < if_modified_since: return 304, response_headers, "Not Modified" if if_none_match and key.etag == if_none_match: return 304, response_headers, "Not Modified" return 200, response_headers, "" else: return 404, response_headers, "" def _lock_config_from_xml(self, xml): response_dict = {"enabled": False, "mode": None, "days": None, "years": None} parsed_xml = xmltodict.parse(xml) enabled = ( parsed_xml["ObjectLockConfiguration"]["ObjectLockEnabled"] == "Enabled" ) response_dict["enabled"] = enabled default_retention = parsed_xml.get("ObjectLockConfiguration").get("Rule") if default_retention: default_retention = default_retention.get("DefaultRetention") mode = default_retention["Mode"] days = int(default_retention.get("Days", 0)) years = int(default_retention.get("Years", 0)) if days and years: raise MalformedXML response_dict["mode"] = mode response_dict["days"] = days response_dict["years"] = years return response_dict def _acl_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) if not parsed_xml.get("AccessControlPolicy"): raise MalformedACLError() # The owner is needed for some reason... if not parsed_xml["AccessControlPolicy"].get("Owner"): # TODO: Validate that the Owner is actually correct. raise MalformedACLError() # If empty, then no ACLs: if parsed_xml["AccessControlPolicy"].get("AccessControlList") is None: return [] if not parsed_xml["AccessControlPolicy"]["AccessControlList"].get("Grant"): raise MalformedACLError() permissions = ["READ", "WRITE", "READ_ACP", "WRITE_ACP", "FULL_CONTROL"] if not isinstance( parsed_xml["AccessControlPolicy"]["AccessControlList"]["Grant"], list ): parsed_xml["AccessControlPolicy"]["AccessControlList"]["Grant"] = [ parsed_xml["AccessControlPolicy"]["AccessControlList"]["Grant"] ] grants = self._get_grants_from_xml( parsed_xml["AccessControlPolicy"]["AccessControlList"]["Grant"], MalformedACLError, permissions, ) return FakeAcl(grants) def _get_grants_from_xml(self, grant_list, exception_type, permissions): grants = [] for grant in grant_list: if grant.get("Permission", "") not in permissions: raise exception_type() if grant["Grantee"].get("@xsi:type", "") not in [ "CanonicalUser", "AmazonCustomerByEmail", "Group", ]: raise exception_type() # TODO: Verify that the proper grantee data is supplied based on the type. grants.append( FakeGrant( [ FakeGrantee( id=grant["Grantee"].get("ID", ""), display_name=grant["Grantee"].get("DisplayName", ""), uri=grant["Grantee"].get("URI", ""), ) ], [grant["Permission"]], ) ) return grants def _acl_from_headers(self, headers): canned_acl = headers.get("x-amz-acl", "") if canned_acl: return get_canned_acl(canned_acl) grants = [] for header, value in headers.items(): header = header.lower() if not header.startswith("x-amz-grant-"): continue permission = { "read": "READ", "write": "WRITE", "read-acp": "READ_ACP", "write-acp": "WRITE_ACP", "full-control": "FULL_CONTROL", }[header[len("x-amz-grant-") :]] grantees = [] for key_and_value in value.split(","): key, value = re.match( '([^=]+)="?([^"]+)"?', key_and_value.strip() ).groups() if key.lower() == "id": grantees.append(FakeGrantee(id=value)) else: grantees.append(FakeGrantee(uri=value)) grants.append(FakeGrant(grantees, [permission])) if grants: return FakeAcl(grants) else: return None def _tagging_from_headers(self, headers): tags = {} if headers.get("x-amz-tagging"): parsed_header = parse_qs(headers["x-amz-tagging"], keep_blank_values=True) for tag in parsed_header.items(): tags[tag[0]] = tag[1][0] return tags def _tagging_from_xml(self, xml): parsed_xml = xmltodict.parse(xml, force_list={"Tag": True}) tags = {} for tag in parsed_xml["Tagging"]["TagSet"]["Tag"]: tags[tag["Key"]] = tag["Value"] return tags def _bucket_tagging_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) tags = {} # Optional if no tags are being sent: if parsed_xml["Tagging"].get("TagSet"): # If there is only 1 tag, then it's not a list: if not isinstance(parsed_xml["Tagging"]["TagSet"]["Tag"], list): tags[parsed_xml["Tagging"]["TagSet"]["Tag"]["Key"]] = parsed_xml[ "Tagging" ]["TagSet"]["Tag"]["Value"] else: for tag in parsed_xml["Tagging"]["TagSet"]["Tag"]: if tag["Key"] in tags: raise DuplicateTagKeys() tags[tag["Key"]] = tag["Value"] # Verify that "aws:" is not in the tags. If so, then this is a problem: for key, _ in tags.items(): if key.startswith("aws:"): raise NoSystemTags() return tags def _cors_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) if isinstance(parsed_xml["CORSConfiguration"]["CORSRule"], list): return [cors for cors in parsed_xml["CORSConfiguration"]["CORSRule"]] return [parsed_xml["CORSConfiguration"]["CORSRule"]] def _mode_until_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) return ( parsed_xml["Retention"]["Mode"], parsed_xml["Retention"]["RetainUntilDate"], ) def _legal_hold_status_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) return parsed_xml["LegalHold"]["Status"] def _encryption_config_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) if ( not parsed_xml["ServerSideEncryptionConfiguration"].get("Rule") or not parsed_xml["ServerSideEncryptionConfiguration"]["Rule"].get( "ApplyServerSideEncryptionByDefault" ) or not parsed_xml["ServerSideEncryptionConfiguration"]["Rule"][ "ApplyServerSideEncryptionByDefault" ].get("SSEAlgorithm") ): raise MalformedXML() return [parsed_xml["ServerSideEncryptionConfiguration"]] def _logging_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) if not parsed_xml["BucketLoggingStatus"].get("LoggingEnabled"): return {} if not parsed_xml["BucketLoggingStatus"]["LoggingEnabled"].get("TargetBucket"): raise MalformedXML() if not parsed_xml["BucketLoggingStatus"]["LoggingEnabled"].get("TargetPrefix"): parsed_xml["BucketLoggingStatus"]["LoggingEnabled"]["TargetPrefix"] = "" # Get the ACLs: if parsed_xml["BucketLoggingStatus"]["LoggingEnabled"].get("TargetGrants"): permissions = ["READ", "WRITE", "FULL_CONTROL"] if not isinstance( parsed_xml["BucketLoggingStatus"]["LoggingEnabled"]["TargetGrants"][ "Grant" ], list, ): target_grants = self._get_grants_from_xml( [ parsed_xml["BucketLoggingStatus"]["LoggingEnabled"][ "TargetGrants" ]["Grant"] ], MalformedXML, permissions, ) else: target_grants = self._get_grants_from_xml( parsed_xml["BucketLoggingStatus"]["LoggingEnabled"]["TargetGrants"][ "Grant" ], MalformedXML, permissions, ) parsed_xml["BucketLoggingStatus"]["LoggingEnabled"][ "TargetGrants" ] = target_grants return parsed_xml["BucketLoggingStatus"]["LoggingEnabled"] def _notification_config_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) if not len(parsed_xml["NotificationConfiguration"]): return {} # The types of notifications, and their required fields (apparently lambda is categorized by the API as # "CloudFunction"): notification_fields = [ ("Topic", "sns"), ("Queue", "sqs"), ("CloudFunction", "lambda"), ] event_names = [ "s3:ReducedRedundancyLostObject", "s3:ObjectCreated:*", "s3:ObjectCreated:Put", "s3:ObjectCreated:Post", "s3:ObjectCreated:Copy", "s3:ObjectCreated:CompleteMultipartUpload", "s3:ObjectRemoved:*", "s3:ObjectRemoved:Delete", "s3:ObjectRemoved:DeleteMarkerCreated", ] found_notifications = ( 0 # Tripwire -- if this is not ever set, then there were no notifications ) for name, arn_string in notification_fields: # 1st verify that the proper notification configuration has been passed in (with an ARN that is close # to being correct -- nothing too complex in the ARN logic): the_notification = parsed_xml["NotificationConfiguration"].get( "{}Configuration".format(name) ) if the_notification: found_notifications += 1 if not isinstance(the_notification, list): the_notification = parsed_xml["NotificationConfiguration"][ "{}Configuration".format(name) ] = [the_notification] for n in the_notification: if not n[name].startswith("arn:aws:{}:".format(arn_string)): raise InvalidNotificationARN() # 2nd, verify that the Events list is correct: assert n["Event"] if not isinstance(n["Event"], list): n["Event"] = [n["Event"]] for event in n["Event"]: if event not in event_names: raise InvalidNotificationEvent() # Parse out the filters: if n.get("Filter"): # Error if S3Key is blank: if not n["Filter"]["S3Key"]: raise KeyError() if not isinstance(n["Filter"]["S3Key"]["FilterRule"], list): n["Filter"]["S3Key"]["FilterRule"] = [ n["Filter"]["S3Key"]["FilterRule"] ] for filter_rule in n["Filter"]["S3Key"]["FilterRule"]: assert filter_rule["Name"] in ["suffix", "prefix"] assert filter_rule["Value"] if not found_notifications: return {} return parsed_xml["NotificationConfiguration"] def _accelerate_config_from_xml(self, xml): parsed_xml = xmltodict.parse(xml) config = parsed_xml["AccelerateConfiguration"] return config["Status"] def _key_response_delete(self, headers, bucket_name, query, key_name): self._set_action("KEY", "DELETE", query) self._authenticate_and_authorize_s3_action() if query.get("uploadId"): upload_id = query["uploadId"][0] self.backend.abort_multipart_upload(bucket_name, upload_id) return 204, {}, "" version_id = query.get("versionId", [None])[0] if "tagging" in query: self.backend.delete_object_tagging( bucket_name, key_name, version_id=version_id ) template = self.response_template(S3_DELETE_KEY_TAGGING_RESPONSE) return 204, {}, template.render(version_id=version_id) bypass = headers.get("X-Amz-Bypass-Governance-Retention") success, response_meta = self.backend.delete_object( bucket_name, key_name, version_id=version_id, bypass=bypass ) response_headers = {} if response_meta is not None: for k in response_meta: response_headers["x-amz-{}".format(k)] = response_meta[k] return 204, response_headers, "" def _complete_multipart_body(self, body): ps = minidom.parseString(body).getElementsByTagName("Part") prev = 0 for p in ps: pn = int(p.getElementsByTagName("PartNumber")[0].firstChild.wholeText) if pn <= prev: raise InvalidPartOrder() yield (pn, p.getElementsByTagName("ETag")[0].firstChild.wholeText) def _key_response_post(self, request, body, bucket_name, query, key_name): self._set_action("KEY", "POST", query) self._authenticate_and_authorize_s3_action() if body == b"" and "uploads" in query: metadata = metadata_from_headers(request.headers) storage_type = request.headers.get("x-amz-storage-class", "STANDARD") multipart_id = self.backend.create_multipart_upload( bucket_name, key_name, metadata, storage_type ) template = self.response_template(S3_MULTIPART_INITIATE_RESPONSE) response = template.render( bucket_name=bucket_name, key_name=key_name, upload_id=multipart_id ) return 200, {}, response if query.get("uploadId"): body = self._complete_multipart_body(body) multipart_id = query["uploadId"][0] multipart, value, etag = self.backend.complete_multipart_upload( bucket_name, multipart_id, body ) if value is None: return 400, {}, "" key = self.backend.put_object( bucket_name, multipart.key_name, value, storage=multipart.storage, etag=etag, multipart=multipart, ) key.set_metadata(multipart.metadata) template = self.response_template(S3_MULTIPART_COMPLETE_RESPONSE) headers = {} if key.version_id: headers["x-amz-version-id"] = key.version_id return ( 200, headers, template.render( bucket_name=bucket_name, key_name=key.name, etag=key.etag ), ) elif "restore" in query: es = minidom.parseString(body).getElementsByTagName("Days") days = es[0].childNodes[0].wholeText key = self.backend.get_object(bucket_name, key_name) r = 202 if key.expiry_date is not None: r = 200 key.restore(int(days)) return r, {}, "" else: raise NotImplementedError( "Method POST had only been implemented for multipart uploads and restore operations, so far" ) def _invalid_headers(self, url, headers): """ Verify whether the provided metadata in the URL is also present in the headers :param url: .../file.txt&content-type=app%2Fjson&Signature=.. :param headers: Content-Type=app/json :return: True or False """ metadata_to_check = { "content-disposition": "Content-Disposition", "content-encoding": "Content-Encoding", "content-language": "Content-Language", "content-length": "Content-Length", "content-md5": "Content-MD5", "content-type": "Content-Type", } for url_key, header_key in metadata_to_check.items(): metadata_in_url = re.search(url_key + "=(.+?)(&.+$|$)", url) if metadata_in_url: url_value = unquote(metadata_in_url.group(1)) if header_key not in headers or (url_value != headers[header_key]): return True return False S3ResponseInstance = ResponseObject(s3_backend) S3_ALL_BUCKETS = """<ListAllMyBucketsResult xmlns="http://s3.amazonaws.com/doc/2006-03-01"> <Owner> <ID>bcaf1ffd86f41161ca5fb16fd081034f</ID> <DisplayName>webfile</DisplayName> </Owner> <Buckets> {% for bucket in buckets %} <Bucket> <Name>{{ bucket.name }}</Name> <CreationDate>{{ bucket.creation_date_ISO8601 }}</CreationDate> </Bucket> {% endfor %} </Buckets> </ListAllMyBucketsResult>""" S3_BUCKET_GET_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <ListBucketResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Name>{{ bucket.name }}</Name> {% if prefix != None %} <Prefix>{{ prefix }}</Prefix> {% endif %} <MaxKeys>{{ max_keys }}</MaxKeys> {% if delimiter %} <Delimiter>{{ delimiter }}</Delimiter> {% endif %} <IsTruncated>{{ is_truncated }}</IsTruncated> {% if next_marker %} <NextMarker>{{ next_marker }}</NextMarker> {% endif %} {% for key in result_keys %} <Contents> <Key>{{ key.name }}</Key> <LastModified>{{ key.last_modified_ISO8601 }}</LastModified> <ETag>{{ key.etag }}</ETag> <Size>{{ key.size }}</Size> <StorageClass>{{ key.storage_class }}</StorageClass> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> </Contents> {% endfor %} {% if delimiter %} {% for folder in result_folders %} <CommonPrefixes> <Prefix>{{ folder }}</Prefix> </CommonPrefixes> {% endfor %} {% endif %} </ListBucketResult>""" S3_BUCKET_GET_RESPONSE_V2 = """<?xml version="1.0" encoding="UTF-8"?> <ListBucketResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Name>{{ bucket.name }}</Name> {% if prefix != None %} <Prefix>{{ prefix }}</Prefix> {% endif %} <MaxKeys>{{ max_keys }}</MaxKeys> <KeyCount>{{ key_count }}</KeyCount> {% if delimiter %} <Delimiter>{{ delimiter }}</Delimiter> {% endif %} <IsTruncated>{{ is_truncated }}</IsTruncated> {% if next_continuation_token %} <NextContinuationToken>{{ next_continuation_token }}</NextContinuationToken> {% endif %} {% if start_after %} <StartAfter>{{ start_after }}</StartAfter> {% endif %} {% for key in result_keys %} <Contents> <Key>{{ key.name }}</Key> <LastModified>{{ key.last_modified_ISO8601 }}</LastModified> <ETag>{{ key.etag }}</ETag> <Size>{{ key.size }}</Size> <StorageClass>{{ key.storage_class }}</StorageClass> {% if fetch_owner %} <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> {% endif %} </Contents> {% endfor %} {% if delimiter %} {% for folder in result_folders %} <CommonPrefixes> <Prefix>{{ folder }}</Prefix> </CommonPrefixes> {% endfor %} {% endif %} </ListBucketResult>""" S3_BUCKET_CREATE_RESPONSE = """<CreateBucketResponse xmlns="http://s3.amazonaws.com/doc/2006-03-01"> <CreateBucketResponse> <Bucket>{{ bucket.name }}</Bucket> </CreateBucketResponse> </CreateBucketResponse>""" S3_DELETE_BUCKET_SUCCESS = """<DeleteBucketResponse xmlns="http://s3.amazonaws.com/doc/2006-03-01"> <DeleteBucketResponse> <Code>204</Code> <Description>No Content</Description> </DeleteBucketResponse> </DeleteBucketResponse>""" S3_DELETE_BUCKET_WITH_ITEMS_ERROR = """<?xml version="1.0" encoding="UTF-8"?> <Error><Code>BucketNotEmpty</Code> <Message>The bucket you tried to delete is not empty</Message> <BucketName>{{ bucket.name }}</BucketName> <RequestId>asdfasdfsdafds</RequestId> <HostId>sdfgdsfgdsfgdfsdsfgdfs</HostId> </Error>""" S3_BUCKET_LOCATION = """<?xml version="1.0" encoding="UTF-8"?> <LocationConstraint xmlns="http://s3.amazonaws.com/doc/2006-03-01/">{% if location != None %}{{ location }}{% endif %}</LocationConstraint>""" S3_BUCKET_LIFECYCLE_CONFIGURATION = """<?xml version="1.0" encoding="UTF-8"?> <LifecycleConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> {% for rule in rules %} <Rule> <ID>{{ rule.id }}</ID> {% if rule.filter %} <Filter> {% if rule.filter.prefix != None %} <Prefix>{{ rule.filter.prefix }}</Prefix> {% endif %} {% if rule.filter.tag_key %} <Tag> <Key>{{ rule.filter.tag_key }}</Key> <Value>{{ rule.filter.tag_value }}</Value> </Tag> {% endif %} {% if rule.filter.and_filter %} <And> {% if rule.filter.and_filter.prefix != None %} <Prefix>{{ rule.filter.and_filter.prefix }}</Prefix> {% endif %} {% for key, value in rule.filter.and_filter.tags.items() %} <Tag> <Key>{{ key }}</Key> <Value>{{ value }}</Value> </Tag> {% endfor %} </And> {% endif %} </Filter> {% else %} {% if rule.prefix != None %} <Prefix>{{ rule.prefix }}</Prefix> {% endif %} {% endif %} <Status>{{ rule.status }}</Status> {% if rule.storage_class %} <Transition> {% if rule.transition_days %} <Days>{{ rule.transition_days }}</Days> {% endif %} {% if rule.transition_date %} <Date>{{ rule.transition_date }}</Date> {% endif %} <StorageClass>{{ rule.storage_class }}</StorageClass> </Transition> {% endif %} {% if rule.expiration_days or rule.expiration_date or rule.expired_object_delete_marker %} <Expiration> {% if rule.expiration_days %} <Days>{{ rule.expiration_days }}</Days> {% endif %} {% if rule.expiration_date %} <Date>{{ rule.expiration_date }}</Date> {% endif %} {% if rule.expired_object_delete_marker %} <ExpiredObjectDeleteMarker>{{ rule.expired_object_delete_marker }}</ExpiredObjectDeleteMarker> {% endif %} </Expiration> {% endif %} {% if rule.nvt_noncurrent_days and rule.nvt_storage_class %} <NoncurrentVersionTransition> <NoncurrentDays>{{ rule.nvt_noncurrent_days }}</NoncurrentDays> <StorageClass>{{ rule.nvt_storage_class }}</StorageClass> </NoncurrentVersionTransition> {% endif %} {% if rule.nve_noncurrent_days %} <NoncurrentVersionExpiration> <NoncurrentDays>{{ rule.nve_noncurrent_days }}</NoncurrentDays> </NoncurrentVersionExpiration> {% endif %} {% if rule.aimu_days %} <AbortIncompleteMultipartUpload> <DaysAfterInitiation>{{ rule.aimu_days }}</DaysAfterInitiation> </AbortIncompleteMultipartUpload> {% endif %} </Rule> {% endfor %} </LifecycleConfiguration> """ S3_BUCKET_VERSIONING = """<?xml version="1.0" encoding="UTF-8"?> <VersioningConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Status>{{ bucket_versioning_status }}</Status> </VersioningConfiguration> """ S3_BUCKET_GET_VERSIONING = """<?xml version="1.0" encoding="UTF-8"?> {% if status is none %} <VersioningConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"/> {% else %} <VersioningConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Status>{{ status }}</Status> </VersioningConfiguration> {% endif %} """ S3_BUCKET_GET_VERSIONS = """<?xml version="1.0" encoding="UTF-8"?> <ListVersionsResult xmlns="http://s3.amazonaws.com/doc/2006-03-01"> <Name>{{ bucket.name }}</Name> {% if prefix != None %} <Prefix>{{ prefix }}</Prefix> {% endif %} {% if common_prefixes %} {% for prefix in common_prefixes %} <CommonPrefixes> <Prefix>{{ prefix }}</Prefix> </CommonPrefixes> {% endfor %} {% endif %} <Delimiter>{{ delimiter }}</Delimiter> <KeyMarker>{{ key_marker or "" }}</KeyMarker> <MaxKeys>{{ max_keys }}</MaxKeys> <IsTruncated>{{ is_truncated }}</IsTruncated> {% for key in key_list %} <Version> <Key>{{ key.name }}</Key> <VersionId>{% if key.version_id is none %}null{% else %}{{ key.version_id }}{% endif %}</VersionId> <IsLatest>{{ 'true' if key.is_latest else 'false' }}</IsLatest> <LastModified>{{ key.last_modified_ISO8601 }}</LastModified> <ETag>{{ key.etag }}</ETag> <Size>{{ key.size }}</Size> <StorageClass>{{ key.storage_class }}</StorageClass> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> </Version> {% endfor %} {% for marker in delete_marker_list %} <DeleteMarker> <Key>{{ marker.name }}</Key> <VersionId>{{ marker.version_id }}</VersionId> <IsLatest>{{ 'true' if marker.is_latest else 'false' }}</IsLatest> <LastModified>{{ marker.last_modified_ISO8601 }}</LastModified> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> </DeleteMarker> {% endfor %} </ListVersionsResult> """ S3_DELETE_KEYS_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <DeleteResult xmlns="http://s3.amazonaws.com/doc/2006-03-01"> {% for k, v in deleted %} <Deleted> <Key>{{k}}</Key> {% if v %}<VersionId>{{v}}</VersionId>{% endif %} </Deleted> {% endfor %} {% for k in delete_errors %} <Error> <Key>{{k}}</Key> </Error> {% endfor %} </DeleteResult>""" S3_DELETE_KEY_TAGGING_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <DeleteObjectTaggingResult xmlns="http://s3.amazonaws.com/doc/2006-03-01"> <VersionId>{{version_id}}</VersionId> </DeleteObjectTaggingResult> """ S3_OBJECT_ACL_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <AccessControlPolicy xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> <AccessControlList> {% for grant in acl.grants %} <Grant> {% for grantee in grant.grantees %} <Grantee xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:type="{{ grantee.type }}"> {% if grantee.uri %} <URI>{{ grantee.uri }}</URI> {% endif %} {% if grantee.id %} <ID>{{ grantee.id }}</ID> {% endif %} {% if grantee.display_name %} <DisplayName>{{ grantee.display_name }}</DisplayName> {% endif %} </Grantee> {% endfor %} {% for permission in grant.permissions %} <Permission>{{ permission }}</Permission> {% endfor %} </Grant> {% endfor %} </AccessControlList> </AccessControlPolicy>""" S3_OBJECT_LEGAL_HOLD = """<?xml version="1.0" encoding="UTF-8"?> <LegalHold> <Status>{{ legal_hold }}</Status> </LegalHold> """ S3_OBJECT_TAGGING_RESPONSE = """\ <?xml version="1.0" encoding="UTF-8"?> <Tagging xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <TagSet> {% for tag in tags %} <Tag> <Key>{{ tag.Key }}</Key> <Value>{{ tag.Value }}</Value> </Tag> {% endfor %} </TagSet> </Tagging>""" S3_BUCKET_CORS_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <CORSConfiguration> {% for cors in cors %} <CORSRule> {% for origin in cors.allowed_origins %} <AllowedOrigin>{{ origin }}</AllowedOrigin> {% endfor %} {% for method in cors.allowed_methods %} <AllowedMethod>{{ method }}</AllowedMethod> {% endfor %} {% if cors.allowed_headers is not none %} {% for header in cors.allowed_headers %} <AllowedHeader>{{ header }}</AllowedHeader> {% endfor %} {% endif %} {% if cors.exposed_headers is not none %} {% for header in cors.exposed_headers %} <ExposedHeader>{{ header }}</ExposedHeader> {% endfor %} {% endif %} {% if cors.max_age_seconds is not none %} <MaxAgeSeconds>{{ cors.max_age_seconds }}</MaxAgeSeconds> {% endif %} </CORSRule> {% endfor %} </CORSConfiguration> """ S3_OBJECT_COPY_RESPONSE = """\ <CopyObjectResult xmlns="http://doc.s3.amazonaws.com/2006-03-01"> <ETag>{{ key.etag }}</ETag> <LastModified>{{ key.last_modified_ISO8601 }}</LastModified> </CopyObjectResult>""" S3_MULTIPART_INITIATE_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <InitiateMultipartUploadResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Bucket>{{ bucket_name }}</Bucket> <Key>{{ key_name }}</Key> <UploadId>{{ upload_id }}</UploadId> </InitiateMultipartUploadResult>""" S3_MULTIPART_UPLOAD_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <CopyPartResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <LastModified>{{ part.last_modified_ISO8601 }}</LastModified> <ETag>{{ part.etag }}</ETag> </CopyPartResult>""" S3_MULTIPART_LIST_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <ListPartsResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Bucket>{{ bucket_name }}</Bucket> <Key>{{ key_name }}</Key> <UploadId>{{ upload_id }}</UploadId> <StorageClass>STANDARD</StorageClass> <Initiator> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Initiator> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> <PartNumberMarker>{{ part_number_marker }}</PartNumberMarker> <NextPartNumberMarker>{{ next_part_number_marker }}</NextPartNumberMarker> <MaxParts>{{ max_parts }}</MaxParts> <IsTruncated>{{ is_truncated }}</IsTruncated> {% for part in parts %} <Part> <PartNumber>{{ part.name }}</PartNumber> <LastModified>{{ part.last_modified_ISO8601 }}</LastModified> <ETag>{{ part.etag }}</ETag> <Size>{{ part.size }}</Size> </Part> {% endfor %} </ListPartsResult>""" S3_MULTIPART_COMPLETE_RESPONSE = """<?xml version="1.0" encoding="UTF-8"?> <CompleteMultipartUploadResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Location>http://{{ bucket_name }}.s3.amazonaws.com/{{ key_name }}</Location> <Bucket>{{ bucket_name }}</Bucket> <Key>{{ key_name }}</Key> <ETag>{{ etag }}</ETag> </CompleteMultipartUploadResult> """ S3_ALL_MULTIPARTS = ( """<?xml version="1.0" encoding="UTF-8"?> <ListMultipartUploadsResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Bucket>{{ bucket_name }}</Bucket> <KeyMarker></KeyMarker> <UploadIdMarker></UploadIdMarker> <MaxUploads>1000</MaxUploads> <IsTruncated>false</IsTruncated> {% for upload in uploads %} <Upload> <Key>{{ upload.key_name }}</Key> <UploadId>{{ upload.id }}</UploadId> <Initiator> <ID>arn:aws:iam::""" + ACCOUNT_ID + """:user/user1-11111a31-17b5-4fb7-9df5-b111111f13de</ID> <DisplayName>user1-11111a31-17b5-4fb7-9df5-b111111f13de</DisplayName> </Initiator> <Owner> <ID>75aa57f09aa0c8caeab4f8c24e99d10f8e7faeebf76c078efc7c6caea54ba06a</ID> <DisplayName>webfile</DisplayName> </Owner> <StorageClass>STANDARD</StorageClass> <Initiated>2010-11-10T20:48:33.000Z</Initiated> </Upload> {% endfor %} </ListMultipartUploadsResult> """ ) S3_NO_POLICY = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchBucketPolicy</Code> <Message>The bucket policy does not exist</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>0D68A23BB2E2215B</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_NO_LIFECYCLE = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchLifecycleConfiguration</Code> <Message>The lifecycle configuration does not exist</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_NO_BUCKET_TAGGING = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchTagSet</Code> <Message>The TagSet does not exist</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_NO_BUCKET_WEBSITE_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchWebsiteConfiguration</Code> <Message>The specified bucket does not have a website configuration</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_INVALID_CORS_REQUEST = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchWebsiteConfiguration</Code> <Message>The specified bucket does not have a website configuration</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_NO_CORS_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>NoSuchCORSConfiguration</Code> <Message>The CORS configuration does not exist</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_LOGGING_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <BucketLoggingStatus xmlns="http://doc.s3.amazonaws.com/2006-03-01"> <LoggingEnabled> <TargetBucket>{{ logging["TargetBucket"] }}</TargetBucket> <TargetPrefix>{{ logging["TargetPrefix"] }}</TargetPrefix> {% if logging.get("TargetGrants") %} <TargetGrants> {% for grant in logging["TargetGrants"] %} <Grant> <Grantee xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:type="{{ grant.grantees[0].type }}"> {% if grant.grantees[0].uri %} <URI>{{ grant.grantees[0].uri }}</URI> {% endif %} {% if grant.grantees[0].id %} <ID>{{ grant.grantees[0].id }}</ID> {% endif %} {% if grant.grantees[0].display_name %} <DisplayName>{{ grant.grantees[0].display_name }}</DisplayName> {% endif %} </Grantee> <Permission>{{ grant.permissions[0] }}</Permission> </Grant> {% endfor %} </TargetGrants> {% endif %} </LoggingEnabled> </BucketLoggingStatus> """ S3_NO_LOGGING_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <BucketLoggingStatus xmlns="http://doc.s3.amazonaws.com/2006-03-01" /> """ S3_ENCRYPTION_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <ServerSideEncryptionConfiguration xmlns="http://doc.s3.amazonaws.com/2006-03-01"> {% for entry in encryption %} <Rule> <ApplyServerSideEncryptionByDefault> <SSEAlgorithm>{{ entry["Rule"]["ApplyServerSideEncryptionByDefault"]["SSEAlgorithm"] }}</SSEAlgorithm> {% if entry["Rule"]["ApplyServerSideEncryptionByDefault"].get("KMSMasterKeyID") %} <KMSMasterKeyID>{{ entry["Rule"]["ApplyServerSideEncryptionByDefault"]["KMSMasterKeyID"] }}</KMSMasterKeyID> {% endif %} </ApplyServerSideEncryptionByDefault> <BucketKeyEnabled>{{ 'true' if entry["Rule"].get("BucketKeyEnabled") == 'true' else 'false' }}</BucketKeyEnabled> </Rule> {% endfor %} </ServerSideEncryptionConfiguration> """ S3_INVALID_PRESIGNED_PARAMETERS = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>SignatureDoesNotMatch</Code> <Message>The request signature we calculated does not match the signature you provided. Check your key and signing method.</Message> <RequestId>0D68A23BB2E2215B</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_NO_ENCRYPTION = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>ServerSideEncryptionConfigurationNotFoundError</Code> <Message>The server side encryption configuration was not found</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>0D68A23BB2E2215B</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """ S3_GET_BUCKET_NOTIFICATION_CONFIG = """<?xml version="1.0" encoding="UTF-8"?> <NotificationConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> {% for topic in config.topic %} <TopicConfiguration> <Id>{{ topic.id }}</Id> <Topic>{{ topic.arn }}</Topic> {% for event in topic.events %} <Event>{{ event }}</Event> {% endfor %} {% if topic.filters %} <Filter> <S3Key> {% for rule in topic.filters["S3Key"]["FilterRule"] %} <FilterRule> <Name>{{ rule["Name"] }}</Name> <Value>{{ rule["Value"] }}</Value> </FilterRule> {% endfor %} </S3Key> </Filter> {% endif %} </TopicConfiguration> {% endfor %} {% for queue in config.queue %} <QueueConfiguration> <Id>{{ queue.id }}</Id> <Queue>{{ queue.arn }}</Queue> {% for event in queue.events %} <Event>{{ event }}</Event> {% endfor %} {% if queue.filters %} <Filter> <S3Key> {% for rule in queue.filters["S3Key"]["FilterRule"] %} <FilterRule> <Name>{{ rule["Name"] }}</Name> <Value>{{ rule["Value"] }}</Value> </FilterRule> {% endfor %} </S3Key> </Filter> {% endif %} </QueueConfiguration> {% endfor %} {% for cf in config.cloud_function %} <CloudFunctionConfiguration> <Id>{{ cf.id }}</Id> <CloudFunction>{{ cf.arn }}</CloudFunction> {% for event in cf.events %} <Event>{{ event }}</Event> {% endfor %} {% if cf.filters %} <Filter> <S3Key> {% for rule in cf.filters["S3Key"]["FilterRule"] %} <FilterRule> <Name>{{ rule["Name"] }}</Name> <Value>{{ rule["Value"] }}</Value> </FilterRule> {% endfor %} </S3Key> </Filter> {% endif %} </CloudFunctionConfiguration> {% endfor %} </NotificationConfiguration> """ S3_BUCKET_ACCELERATE = """ <AccelerateConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> <Status>{{ bucket.accelerate_configuration }}</Status> </AccelerateConfiguration> """ S3_BUCKET_ACCELERATE_NOT_SET = """ <AccelerateConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"/> """ S3_PUBLIC_ACCESS_BLOCK_CONFIGURATION = """ <PublicAccessBlockConfiguration> <BlockPublicAcls>{{public_block_config.block_public_acls}}</BlockPublicAcls> <IgnorePublicAcls>{{public_block_config.ignore_public_acls}}</IgnorePublicAcls> <BlockPublicPolicy>{{public_block_config.block_public_policy}}</BlockPublicPolicy> <RestrictPublicBuckets>{{public_block_config.restrict_public_buckets}}</RestrictPublicBuckets> </PublicAccessBlockConfiguration> """ S3_BUCKET_LOCK_CONFIGURATION = """ <ObjectLockConfiguration xmlns="http://s3.amazonaws.com/doc/2006-03-01/"> {%if lock_enabled %} <ObjectLockEnabled>Enabled</ObjectLockEnabled> {% else %} <ObjectLockEnabled>Disabled</ObjectLockEnabled> {% endif %} {% if mode %} <Rule> <DefaultRetention> <Mode>{{mode}}</Mode> <Days>{{days}}</Days> <Years>{{years}}</Years> </DefaultRetention> </Rule> {% endif %} </ObjectLockConfiguration> """ S3_DUPLICATE_BUCKET_ERROR = """<?xml version="1.0" encoding="UTF-8"?> <Error> <Code>BucketAlreadyOwnedByYou</Code> <Message>Your previous request to create the named bucket succeeded and you already own it.</Message> <BucketName>{{ bucket_name }}</BucketName> <RequestId>44425877V1D0A2F9</RequestId> <HostId>9Gjjt1m+cjU4OPvX9O9/8RuvnG41MRb/18Oux2o5H5MY7ISNTlXN+Dz9IG62/ILVxhAGI0qyPfg=</HostId> </Error> """
moto/s3/responses.py
103,140
Verify whether the provided metadata in the URL is also present in the headers :param url: .../file.txt&content-type=app%2Fjson&Signature=.. :param headers: Content-Type=app/json :return: True or False strip the first '/' left by urlparse GOlang sends a request as url/?delete= (treating it as a normal key=value, even if the value is empty) Python sends a request as url/?delete (treating it as a flag) https://github.com/spulec/moto/issues/2937 No bucket specified. Listing all buckets Default to path-based buckets for (1) localhost, (2) localstack hosts (e.g. localstack.dev), (3) local host names that do not contain a "." (e.g., Docker container host names), or (4) kubernetes host names Depending on which calling format the client is using, we don't know if this is a bucket or key request so we have to check Using path-based buckets If no bucket specified, list all buckets Boto Flask server Unless we do this, boto3 does not raise ClientError on HEAD (which the real API responds with), and instead raises NoSuchBucket, leading to inconsistency in error response between real and mocked responses. us-east-1 is different - returns a None location "Special Error" if no tags: If Python 2, fix the unicode strings: If there is only one rule, xmldict returns just the item Headers are first. If not set, then look at the body (consistent with the documentation): us-east-1, the default AWS region behaves a bit differently - you should not use it as a location constraint --> it fails - querying the location constraint returns None - LocationConstraint has to be specified if outside us-east-1 us-east-1 has different behavior - creating a bucket there is an idempotent operation TODO: Support the XML-based ACL format Bucket exists Tried to delete a bucket that still has keys POST to bucket-url should create file from form Not HTTPretty HTTPretty, build new form object Metadata We expect a list of objects, but when there is a single <Object> node xmltodict does not return a list. byte range suffix byte range https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-streaming.htmlsigv4-chunked-body-definition str(hex(chunk-size)) + ";chunk-signature=" + signature + \r\n + chunk-data + \r\n skip trailing \r\n Key and Control are lumped in because splitting out the regex is too much of a pain :/ Is this an S3 control response? Boto Flask server Gets the action out of the URL sans query params. Gets the action out of the URL sans query params. Gets the action out of the URL sans query params. Because we patch the requests library the boto/boto3 API requests go through this method but so do `requests.get("https://bucket-name.s3.amazonaws.com/file-name")` Here we deny public access to private files by checking the ACL and checking for the mere presence of an Authorization header. coming in from requests.get(s3.generate_presigned_url()) Boto Flask server when the data is being passed as a file Body comes through as part of the form, if no content-type is set on the PUT-request form = ImmutableMultiDict([('some data 123 321', '')]) 0 <= PartNumberMarker <= 2,147,483,647 0 <= MaxParts <= 2,147,483,647 (default is 1,000) Copy key you can have a quoted ?version=abc with a version Id, so work on we need to parse the unquoted string first Closing the connection of a streaming request. No more data Streaming request, more data Initial data The owner is needed for some reason... TODO: Validate that the Owner is actually correct. If empty, then no ACLs: TODO: Verify that the proper grantee data is supplied based on the type. Optional if no tags are being sent: If there is only 1 tag, then it's not a list: Verify that "aws:" is not in the tags. If so, then this is a problem: Get the ACLs: The types of notifications, and their required fields (apparently lambda is categorized by the API as "CloudFunction"): Tripwire -- if this is not ever set, then there were no notifications 1st verify that the proper notification configuration has been passed in (with an ARN that is close to being correct -- nothing too complex in the ARN logic): 2nd, verify that the Events list is correct: Parse out the filters: Error if S3Key is blank:
4,160
en
0.857884
# -*- coding: utf-8 -*- # # Python Github documentation build configuration file, created by # sphinx-quickstart on Tue Feb 3 23:23:15 2015. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys import os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.doctest', 'sphinx.ext.intersphinx', 'sphinx.ext.todo', 'sphinx.ext.viewcode', ] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'Python Github' copyright = u'2015, Nicolas Mendoza' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '0.1.0' # The full version, including alpha/beta/rc tags. release = '0.1.0' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build'] # The reST default role (used for this markup: `text`) to use for all # documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. #keep_warnings = False # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'default' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. #html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. #html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. #html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'PythonGithubdoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ ('index', 'PythonGithub.tex', u'Python Github Documentation', u'Nicolas Mendoza', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'pythongithub', u'Python Github Documentation', [u'Nicolas Mendoza'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'PythonGithub', u'Python Github Documentation', u'Nicolas Mendoza', 'PythonGithub', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. #texinfo_no_detailmenu = False # Example configuration for intersphinx: refer to the Python standard library. intersphinx_mapping = {'http://docs.python.org/': None}
docs/conf.py
8,480
-*- coding: utf-8 -*- Python Github documentation build configuration file, created by sphinx-quickstart on Tue Feb 3 23:23:15 2015. This file is execfile()d with the current directory set to its containing dir. Note that not all possible configuration values are present in this autogenerated file. All configuration values have a default; values that are commented out serve to show the default. If extensions (or modules to document with autodoc) are in another directory, add these directories to sys.path here. If the directory is relative to the documentation root, use os.path.abspath to make it absolute, like shown here.sys.path.insert(0, os.path.abspath('.')) -- General configuration ------------------------------------------------ If your documentation needs a minimal Sphinx version, state it here.needs_sphinx = '1.0' Add any Sphinx extension module names here, as strings. They can be extensions coming with Sphinx (named 'sphinx.ext.*') or your custom ones. Add any paths that contain templates here, relative to this directory. The suffix of source filenames. The encoding of source files.source_encoding = 'utf-8-sig' The master toctree document. General information about the project. The version info for the project you're documenting, acts as replacement for |version| and |release|, also used in various other places throughout the built documents. The short X.Y version. The full version, including alpha/beta/rc tags. The language for content autogenerated by Sphinx. Refer to documentation for a list of supported languages.language = None There are two options for replacing |today|: either, you set today to some non-false value, then it is used:today = '' Else, today_fmt is used as the format for a strftime call.today_fmt = '%B %d, %Y' List of patterns, relative to source directory, that match files and directories to ignore when looking for source files. The reST default role (used for this markup: `text`) to use for all documents.default_role = None If true, '()' will be appended to :func: etc. cross-reference text.add_function_parentheses = True If true, the current module name will be prepended to all description unit titles (such as .. function::).add_module_names = True If true, sectionauthor and moduleauthor directives will be shown in the output. They are ignored by default.show_authors = False The name of the Pygments (syntax highlighting) style to use. A list of ignored prefixes for module index sorting.modindex_common_prefix = [] If true, keep warnings as "system message" paragraphs in the built documents.keep_warnings = False -- Options for HTML output ---------------------------------------------- The theme to use for HTML and HTML Help pages. See the documentation for a list of builtin themes. Theme options are theme-specific and customize the look and feel of a theme further. For a list of options available for each theme, see the documentation.html_theme_options = {} Add any paths that contain custom themes here, relative to this directory.html_theme_path = [] The name for this set of Sphinx documents. If None, it defaults to "<project> v<release> documentation".html_title = None A shorter title for the navigation bar. Default is the same as html_title.html_short_title = None The name of an image file (relative to this directory) to place at the top of the sidebar.html_logo = None The name of an image file (within the static path) to use as favicon of the docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 pixels large.html_favicon = None Add any paths that contain custom static files (such as style sheets) here, relative to this directory. They are copied after the builtin static files, so a file named "default.css" will overwrite the builtin "default.css". Add any extra paths that contain custom files (such as robots.txt or .htaccess) here, relative to this directory. These files are copied directly to the root of the documentation.html_extra_path = [] If not '', a 'Last updated on:' timestamp is inserted at every page bottom, using the given strftime format.html_last_updated_fmt = '%b %d, %Y' If true, SmartyPants will be used to convert quotes and dashes to typographically correct entities.html_use_smartypants = True Custom sidebar templates, maps document names to template names.html_sidebars = {} Additional templates that should be rendered to pages, maps page names to template names.html_additional_pages = {} If false, no module index is generated.html_domain_indices = True If false, no index is generated.html_use_index = True If true, the index is split into individual pages for each letter.html_split_index = False If true, links to the reST sources are added to the pages.html_show_sourcelink = True If true, "Created using Sphinx" is shown in the HTML footer. Default is True.html_show_sphinx = True If true, "(C) Copyright ..." is shown in the HTML footer. Default is True.html_show_copyright = True If true, an OpenSearch description file will be output, and all pages will contain a <link> tag referring to it. The value of this option must be the base URL from which the finished HTML is served.html_use_opensearch = '' This is the file name suffix for HTML files (e.g. ".xhtml").html_file_suffix = None Output file base name for HTML help builder. -- Options for LaTeX output --------------------------------------------- The paper size ('letterpaper' or 'a4paper').'papersize': 'letterpaper', The font size ('10pt', '11pt' or '12pt').'pointsize': '10pt', Additional stuff for the LaTeX preamble.'preamble': '', Grouping the document tree into LaTeX files. List of tuples (source start file, target name, title, author, documentclass [howto, manual, or own class]). The name of an image file (relative to this directory) to place at the top of the title page.latex_logo = None For "manual" documents, if this is true, then toplevel headings are parts, not chapters.latex_use_parts = False If true, show page references after internal links.latex_show_pagerefs = False If true, show URL addresses after external links.latex_show_urls = False Documents to append as an appendix to all manuals.latex_appendices = [] If false, no module index is generated.latex_domain_indices = True -- Options for manual page output --------------------------------------- One entry per manual page. List of tuples (source start file, name, description, authors, manual section). If true, show URL addresses after external links.man_show_urls = False -- Options for Texinfo output ------------------------------------------- Grouping the document tree into Texinfo files. List of tuples (source start file, target name, title, author, dir menu entry, description, category) Documents to append as an appendix to all manuals.texinfo_appendices = [] If false, no module index is generated.texinfo_domain_indices = True How to display URL addresses: 'footnote', 'no', or 'inline'.texinfo_show_urls = 'footnote' If true, do not generate a @detailmenu in the "Top" node's menu.texinfo_no_detailmenu = False Example configuration for intersphinx: refer to the Python standard library.
7,108
en
0.660438
from setuptools import setup setup( name='dst', version='0.1.5', author='Jeroen Janssens', author_email='jeroen@jeroenjanssens.com', packages=['dst'], url='http://datasciencetoolbox.org', license='BSD', description='Data Science Toolbox -- Start doing data science in minutes.', long_description=open('README.txt').read(), install_requires=[ "ansible >= 1.5", ], entry_points={ 'console_scripts': ['dst = dst.dst:main'] }, classifiers=[ # https://pypi.python.org/pypi?:action=list_classifiers 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Science/Research', 'Intended Audience :: Developers', 'Intended Audience :: System Administrators', 'Intended Audience :: End Users/Desktop', 'License :: OSI Approved :: BSD License', 'Operating System :: POSIX :: Linux', 'Topic :: Scientific/Engineering', 'Topic :: Utilities', 'Topic :: System :: Software Distribution', 'Topic :: System :: Systems Administration', 'Programming Language :: Python :: 2.7', 'Programming Language :: Unix Shell', ], )
manager/setup.py
1,216
https://pypi.python.org/pypi?:action=list_classifiers
53
en
0.278693
#!coding:utf8 #author:yqq #date:2020/4/30 0030 17:11 #description: import os import pymysql SQL_PASSWD = os.environ.get('SQL_PWD') def open(host : str,usr : str, passwd : str,db_name : str): conn = pymysql.connect(host=host, user=usr, password=passwd, db=db_name, charset='utf8', cursorclass=pymysql.cursors.DictCursor) return conn def close(conn): conn.close() def execute(conn,cmd): cur = conn.cursor() cur.execute(cmd) conn.commit() #fixed bug by yqq 2019-05-01 return cur.fetchall() def run(cmd): conn = open() result = execute(conn,cmd) close(conn) return result def get_column_values(conn,table_name,column_name): cmd = "SELECT {0} FROM {1}".format(column_name,table_name) return execute(conn,cmd) def main(): host = '192.168.10.29' usr = 'root' passwd = 'eWFuZ3FpbmdxaW5n' dbname = 'test_1' conn = open(host=host, usr=usr, passwd=passwd, db_name=dbname ) print(get_column_values(conn,'t_test_student','name')) close(conn) if __name__ == "__main__": main()
Python3/Tornado/apps/pg/PG_Admin/lib/sql.py
1,105
!coding:utf8author:yqqdate:2020/4/30 0030 17:11description:fixed bug by yqq 2019-05-01
86
en
0.434222
# Copyright 2018 Open Source Robotics Foundation, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # This file is borrowed from ros2/rosidl: # https://github.com/ros2/rosidl/blob/master/rosidl_adapter/rosidl_adapter/resource/__init__.py # Slight modifications were made, so proper paths to files are accepted. from io import StringIO import os import sys import em def expand_template(template_name, data, output_file, encoding='utf-8'): content = evaluate_template(template_name, data) if output_file.exists(): existing_content = output_file.read_text(encoding=encoding) if existing_content == content: return elif output_file.parent: os.makedirs(str(output_file.parent), exist_ok=True) output_file.write_text(content, encoding=encoding) _interpreter = None def evaluate_template(template_name, data): global _interpreter # create copy before manipulating data = dict(data) data['TEMPLATE'] = _evaluate_template #template_path = os.path.join(os.path.dirname(__file__), template_name) template_path = template_name output = StringIO() try: _interpreter = em.Interpreter( output=output, options={ em.BUFFERED_OPT: True, em.RAW_OPT: True, }) with open(template_path, 'r') as h: content = h.read() _interpreter.invoke( 'beforeFile', name=template_name, file=h, locals=data) _interpreter.string(content, template_path, locals=data) _interpreter.invoke('afterFile') return output.getvalue() except Exception as e: # noqa: F841 print( f"{e.__class__.__name__} processing template '{template_name}'", file=sys.stderr) raise finally: _interpreter.shutdown() _interpreter = None def _evaluate_template(template_name, **kwargs): global _interpreter #template_path = os.path.join(os.path.dirname(__file__), template_name) template_path = template_name with open(template_path, 'r') as h: _interpreter.invoke( 'beforeInclude', name=template_path, file=h, locals=kwargs) content = h.read() try: _interpreter.string(content, template_path, kwargs) except Exception as e: # noqa: F841 print( f"{e.__class__.__name__} processing template '{template_name}': " f'{e}', file=sys.stderr) sys.exit(1) _interpreter.invoke('afterInclude')
doc/extensions/empy_helpers/__init__.py
3,028
Copyright 2018 Open Source Robotics Foundation, Inc. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. This file is borrowed from ros2/rosidl: https://github.com/ros2/rosidl/blob/master/rosidl_adapter/rosidl_adapter/resource/__init__.py Slight modifications were made, so proper paths to files are accepted. create copy before manipulatingtemplate_path = os.path.join(os.path.dirname(__file__), template_name) noqa: F841template_path = os.path.join(os.path.dirname(__file__), template_name) noqa: F841
973
en
0.822695
""" This magical module will rewrite all public methods in the public interface of the library so they can run the loop on their own if it's not already running. This rewrite may not be desirable if the end user always uses the methods they way they should be ran, but it's incredibly useful for quick scripts and the runtime overhead is relatively low. Some really common methods which are hardly used offer this ability by default, such as ``.start()`` and ``.run_until_disconnected()`` (since you may want to start, and then run until disconnected while using async event handlers). """ import asyncio import functools import inspect from . import connection from .client.account import _TakeoutClient from .client.telegramclient import TelegramClient from .tl import types, functions, custom from .tl.custom import ( Draft, Dialog, MessageButton, Forward, Button, Message, InlineResult, Conversation ) from .tl.custom.chatgetter import ChatGetter from .tl.custom.sendergetter import SenderGetter def _syncify_wrap(t, method_name): method = getattr(t, method_name) @functools.wraps(method) def syncified(*args, **kwargs): coro = method(*args, **kwargs) loop = asyncio.get_event_loop() if loop.is_running(): return coro else: return loop.run_until_complete(coro) # Save an accessible reference to the original method setattr(syncified, '__tl.sync', method) setattr(t, method_name, syncified) def syncify(*types): """ Converts all the methods in the given types (class definitions) into synchronous, which return either the coroutine or the result based on whether ``asyncio's`` event loop is running. """ # Our asynchronous generators all are `RequestIter`, which already # provide a synchronous iterator variant, so we don't need to worry # about asyncgenfunction's here. for t in types: for name in dir(t): if not name.startswith('_') or name == '__call__': if inspect.iscoroutinefunction(getattr(t, name)): _syncify_wrap(t, name) syncify(TelegramClient, _TakeoutClient, Draft, Dialog, MessageButton, ChatGetter, SenderGetter, Forward, Message, InlineResult, Conversation) __all__ = [ 'TelegramClient', 'Button', 'types', 'functions', 'custom', 'errors', 'events', 'utils', 'connection' ]
telethon/sync.py
2,405
Converts all the methods in the given types (class definitions) into synchronous, which return either the coroutine or the result based on whether ``asyncio's`` event loop is running. This magical module will rewrite all public methods in the public interface of the library so they can run the loop on their own if it's not already running. This rewrite may not be desirable if the end user always uses the methods they way they should be ran, but it's incredibly useful for quick scripts and the runtime overhead is relatively low. Some really common methods which are hardly used offer this ability by default, such as ``.start()`` and ``.run_until_disconnected()`` (since you may want to start, and then run until disconnected while using async event handlers). Save an accessible reference to the original method Our asynchronous generators all are `RequestIter`, which already provide a synchronous iterator variant, so we don't need to worry about asyncgenfunction's here.
982
en
0.933085
""" Plotting code for nilearn """ # Original Authors: Chris Filo Gorgolewski, Gael Varoquaux import os import sys import importlib ############################################################################### # Make sure that we don't get DISPLAY problems when running without X on # unices def _set_mpl_backend(): # We are doing local imports here to avoid polluting our namespace try: import matplotlib except ImportError: if importlib.util.find_spec("pytest") is not None: from .._utils.testing import skip_if_running_tests # No need to fail when running tests skip_if_running_tests('matplotlib not installed') raise else: from ..version import (_import_module_with_version_check, OPTIONAL_MATPLOTLIB_MIN_VERSION) # When matplotlib was successfully imported we need to check # that the version is greater that the minimum required one _import_module_with_version_check('matplotlib', OPTIONAL_MATPLOTLIB_MIN_VERSION) current_backend = matplotlib.get_backend().lower() if 'inline' in current_backend or 'nbagg' in current_backend: return # Set the backend to a non-interactive one for unices without X # (see gh-2560) if (sys.platform not in ('darwin', 'win32') and 'DISPLAY' not in os.environ): matplotlib.use('Agg') _set_mpl_backend() ############################################################################### from . import cm from .img_plotting import ( plot_img, plot_anat, plot_epi, plot_roi, plot_stat_map, plot_glass_brain, plot_connectome, plot_connectome_strength, plot_markers, plot_prob_atlas, plot_carpet, plot_img_comparison, show) from .find_cuts import find_xyz_cut_coords, find_cut_slices, \ find_parcellation_cut_coords, find_probabilistic_atlas_cut_coords from .matrix_plotting import (plot_matrix, plot_contrast_matrix, plot_design_matrix, plot_event) from .html_surface import view_surf, view_img_on_surf from .html_stat_map import view_img from .html_connectome import view_connectome, view_markers from .surf_plotting import (plot_surf, plot_surf_stat_map, plot_surf_roi, plot_img_on_surf, plot_surf_contours) __all__ = ['cm', 'plot_img', 'plot_anat', 'plot_epi', 'plot_roi', 'plot_stat_map', 'plot_glass_brain', 'plot_markers', 'plot_connectome', 'plot_prob_atlas', 'find_xyz_cut_coords', 'find_cut_slices', 'plot_img_comparison', 'show', 'plot_matrix', 'plot_design_matrix', 'plot_contrast_matrix', 'plot_event', 'view_surf', 'view_img_on_surf', 'view_img', 'view_connectome', 'view_markers', 'find_parcellation_cut_coords', 'find_probabilistic_atlas_cut_coords', 'plot_surf', 'plot_surf_stat_map', 'plot_surf_roi', 'plot_img_on_surf', 'plot_connectome_strength', 'plot_carpet', 'plot_surf_contours']
nilearn/plotting/__init__.py
3,118
Plotting code for nilearn Original Authors: Chris Filo Gorgolewski, Gael Varoquaux Make sure that we don't get DISPLAY problems when running without X on unices We are doing local imports here to avoid polluting our namespace No need to fail when running tests When matplotlib was successfully imported we need to check that the version is greater that the minimum required one Set the backend to a non-interactive one for unices without X (see gh-2560)
455
en
0.855367
import asyncio import logging import voluptuous as vol from homeassistant.components.system_log import CONF_LOGGER from homeassistant.config_entries import ConfigEntry from homeassistant.const import EVENT_HOMEASSISTANT_STOP from homeassistant.core import HomeAssistant, Event from homeassistant.helpers import config_validation as cv from homeassistant.helpers.aiohttp_client import async_create_clientsession from homeassistant.helpers.entity_registry import EntityRegistry from homeassistant.helpers.storage import Store from .core import logger from .core.gateway3 import Gateway3 from .core.helpers import DevicesRegistry from .core.utils import DOMAIN, XiaomiGateway3Debug from .core.xiaomi_cloud import MiCloud _LOGGER = logging.getLogger(__name__) DOMAINS = ['binary_sensor', 'climate', 'cover', 'light', 'remote', 'sensor', 'switch', 'alarm_control_panel'] CONF_DEVICES = 'devices' CONF_ATTRIBUTES_TEMPLATE = 'attributes_template' CONFIG_SCHEMA = vol.Schema({ DOMAIN: vol.Schema({ vol.Optional(CONF_DEVICES): { cv.string: vol.Schema({ vol.Optional('occupancy_timeout'): cv.positive_int, }, extra=vol.ALLOW_EXTRA), }, CONF_LOGGER: logger.CONFIG_SCHEMA, vol.Optional(CONF_ATTRIBUTES_TEMPLATE): cv.template }, extra=vol.ALLOW_EXTRA), }, extra=vol.ALLOW_EXTRA) async def async_setup(hass: HomeAssistant, hass_config: dict): config = hass_config.get(DOMAIN) or {} if CONF_LOGGER in config: logger.init(__name__, config[CONF_LOGGER], hass.config.config_dir) info = await hass.helpers.system_info.async_get_system_info() _LOGGER.debug(f"SysInfo: {info}") # update global debug_mode for all gateways if 'debug_mode' in config[CONF_LOGGER]: setattr(Gateway3, 'debug_mode', config[CONF_LOGGER]['debug_mode']) if CONF_DEVICES in config: for k, v in config[CONF_DEVICES].items(): # AA:BB:CC:DD:EE:FF => aabbccddeeff k = k.replace(':', '').lower() DevicesRegistry.defaults[k] = v hass.data[DOMAIN] = { CONF_ATTRIBUTES_TEMPLATE: config.get(CONF_ATTRIBUTES_TEMPLATE) } await _handle_device_remove(hass) # utils.migrate_unique_id(hass) return True async def async_setup_entry(hass: HomeAssistant, entry: ConfigEntry): """Support two kind of enties - MiCloud and Gateway.""" # entry for MiCloud login if 'servers' in entry.data: return await _setup_micloud_entry(hass, entry) # migrate data (also after first setup) to options if entry.data: hass.config_entries.async_update_entry(entry, data={}, options=entry.data) await _setup_logger(hass) # add options handler if not entry.update_listeners: entry.add_update_listener(async_update_options) hass.data[DOMAIN][entry.entry_id] = Gateway3(**entry.options) hass.async_create_task(_setup_domains(hass, entry)) return True async def async_update_options(hass: HomeAssistant, entry: ConfigEntry): await hass.config_entries.async_reload(entry.entry_id) async def async_unload_entry(hass: HomeAssistant, entry: ConfigEntry): # check unload cloud integration if entry.entry_id not in hass.data[DOMAIN]: return # remove all stats entities if disable stats if not entry.options.get('stats'): suffix = ('_gateway', '_zigbee', '_ble') registry: EntityRegistry = hass.data['entity_registry'] remove = [ entity.entity_id for entity in list(registry.entities.values()) if (entity.config_entry_id == entry.entry_id and entity.unique_id.endswith(suffix)) ] for entity_id in remove: registry.async_remove(entity_id) gw: Gateway3 = hass.data[DOMAIN][entry.entry_id] await gw.stop() await asyncio.gather(*[ hass.config_entries.async_forward_entry_unload(entry, domain) for domain in DOMAINS ]) return True async def _setup_domains(hass: HomeAssistant, entry: ConfigEntry): # init setup for each supported domains await asyncio.gather(*[ hass.config_entries.async_forward_entry_setup(entry, domain) for domain in DOMAINS ]) gw: Gateway3 = hass.data[DOMAIN][entry.entry_id] gw.start() entry.async_on_unload( hass.bus.async_listen_once(EVENT_HOMEASSISTANT_STOP, gw.stop) ) async def _setup_micloud_entry(hass: HomeAssistant, config_entry): data: dict = config_entry.data.copy() session = async_create_clientsession(hass) hass.data[DOMAIN]['cloud'] = cloud = MiCloud(session, data['servers']) if 'service_token' in data: # load devices with saved MiCloud auth cloud.auth = data devices = await cloud.get_devices() else: devices = None if devices is None: _LOGGER.debug(f"Login to MiCloud for {config_entry.title}") if await cloud.login(data['username'], data['password']): # update MiCloud auth in .storage data.update(cloud.auth) hass.config_entries.async_update_entry(config_entry, data=data) devices = await cloud.get_devices() if devices is None: _LOGGER.error("Can't load devices from MiCloud") else: _LOGGER.error("Can't login to MiCloud") # load devices from or save to .storage store = Store(hass, 1, f"{DOMAIN}/{data['username']}.json") if devices is None: _LOGGER.debug("Loading a list of devices from the .storage") devices = await store.async_load() else: _LOGGER.debug(f"Loaded from MiCloud {len(devices)} devices") await store.async_save(devices) if devices is None: _LOGGER.debug("No devices in .storage") return False # TODO: Think about a bunch of devices if 'devices' not in hass.data[DOMAIN]: hass.data[DOMAIN]['devices'] = devices else: hass.data[DOMAIN]['devices'] += devices for device in devices: # key - mac for BLE, and did for others did = device['did'] if device['pid'] not in '6' else \ device['mac'].replace(':', '').lower() DevicesRegistry.defaults.setdefault(did, {}) # don't override name if exists DevicesRegistry.defaults[did].setdefault('device_name', device['name']) return True async def _handle_device_remove(hass: HomeAssistant): """Remove device from Hass and Mi Home if the device is renamed to `delete`. """ async def device_registry_updated(event: Event): if event.data['action'] != 'update': return registry = hass.data['device_registry'] hass_device = registry.async_get(event.data['device_id']) # check empty identifiers if not hass_device or not hass_device.identifiers: return # handle only our devices for hass_did in hass_device.identifiers: if hass_did[0] == DOMAIN and hass_device.name_by_user == 'delete': break else: return # remove from Mi Home for gw in hass.data[DOMAIN].values(): if not isinstance(gw, Gateway3): continue gw_device = gw.get_device(hass_did[1]) if not gw_device: continue if gw_device['type'] == 'zigbee': gw.debug(f"Remove device: {gw_device['did']}") await gw.miio.send('remove_device', [gw_device['did']]) break # remove from Hass registry.async_remove_device(hass_device.id) hass.bus.async_listen('device_registry_updated', device_registry_updated) async def _setup_logger(hass: HomeAssistant): if not hasattr(_LOGGER, 'defaul_level'): # default level from Hass config _LOGGER.defaul_level = _LOGGER.level entries = hass.config_entries.async_entries(DOMAIN) web_logs = any(e.options.get('debug') for e in entries) # only if global logging don't set if _LOGGER.defaul_level == logging.NOTSET: # disable log to console _LOGGER.propagate = web_logs is False # set debug if any of integrations has debug _LOGGER.setLevel(logging.DEBUG if web_logs else logging.NOTSET) # if don't set handler yet if web_logs: # skip if already added if any(isinstance(h, XiaomiGateway3Debug) for h in _LOGGER.handlers): return handler = XiaomiGateway3Debug(hass) _LOGGER.addHandler(handler) if _LOGGER.defaul_level == logging.NOTSET: info = await hass.helpers.system_info.async_get_system_info() _LOGGER.debug(f"SysInfo: {info}")
custom_components/xiaomi_gateway3/__init__.py
8,812
update global debug_mode for all gateways AA:BB:CC:DD:EE:FF => aabbccddeeff utils.migrate_unique_id(hass) entry for MiCloud login migrate data (also after first setup) to options add options handler check unload cloud integration remove all stats entities if disable stats init setup for each supported domains load devices with saved MiCloud auth update MiCloud auth in .storage load devices from or save to .storage TODO: Think about a bunch of devices key - mac for BLE, and did for others don't override name if exists check empty identifiers handle only our devices remove from Mi Home remove from Hass default level from Hass config only if global logging don't set disable log to console set debug if any of integrations has debug if don't set handler yet skip if already added
784
en
0.70239
# User class to hold name and __data class User: ### Instance Variables ### __userName = "" __validUser = None __data = [] __weights = [] __notes = [] __dates = [] __intWeights = [] __avgWeight = 0 __minWeight = 0 __maxWeight = 0 ########################## ### Getters ### def getUserName(self): return self.__userName def getData(self): return self.__data def getValidUser(self): return self.__validUser def getWeights(self): return self.__weights def getNotes(self): return self.__notes def getDates(self): return self.__dates def getAvgWeight(self): return str(self.__avgWeight) def getMinWeight(self): return str(self.__minWeight) def getMaxWeight(self): return str(self.__maxWeight) ################ ### Setters ### def setUserName(self, name): self.__userName = name def setData(self, data): self.__data = data def setValidUser(self, valid): self.__validUser = valid def setWeights(self, weights): self.__weights = weights def setNotes(self, notes): self.__notes = notes def setDates(self, dates): self.__dates = dates ################ def addData(self, data): self.__data.append(data) def addWeight(self, weight): self.__weights.append(weight) def addNote(self, note): self.__notes.append(note) def addDate(self, date): self.__dates.append(date) def calcAvg(self): self.__avgWeight = int(sum(self.__intWeights)/len(self.__intWeights)) def calcMaxWeight(self): self.__maxWeight = max(self.__intWeights) def calacMinWeight(self): self.__minWeight = min(self.__intWeights) def averageWeightDelta(self, weightData): pass def convertWeightList(self, weightData): for i in range(len(weightData)): weightData[i] = int(weightData[i]) self.__intWeights = weightData
User.py
2,065
User class to hold name and __data Instance Variables Getters Setters
71
en
0.596517
# Generated by Django 3.2.4 on 2021-07-04 11:51 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ('auth', '0012_alter_user_first_name_max_length'), ] operations = [ migrations.CreateModel( name='User', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('password', models.CharField(max_length=128, verbose_name='password')), ('last_login', models.DateTimeField(blank=True, null=True, verbose_name='last login')), ('is_superuser', models.BooleanField(default=False, help_text='Designates that this user has all permissions without explicitly assigning them.', verbose_name='superuser status')), ('email', models.EmailField(max_length=255, unique=True)), ('name', models.CharField(max_length=255)), ('is_active', models.BooleanField(default=True)), ('is_staff', models.BooleanField(default=False)), ('groups', models.ManyToManyField(blank=True, help_text='The groups this user belongs to. A user will get all permissions granted to each of their groups.', related_name='user_set', related_query_name='user', to='auth.Group', verbose_name='groups')), ('user_permissions', models.ManyToManyField(blank=True, help_text='Specific permissions for this user.', related_name='user_set', related_query_name='user', to='auth.Permission', verbose_name='user permissions')), ], options={ 'abstract': False, }, ), migrations.CreateModel( name='Game', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('tower_blocks_score', models.IntegerField(default=0)), ('bounce_score', models.IntegerField(default=0)), ('kill_birds_score', models.IntegerField(default=0)), ('snake_score', models.IntegerField(default=0)), ('last_updated', models.DateTimeField(auto_now_add=True)), ('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), ]
app/core/migrations/0001_initial.py
2,470
Generated by Django 3.2.4 on 2021-07-04 11:51
45
en
0.759579
from simplecoremidi import send_midi from time import sleep def play_a_scale(): root_note = 60 # This is middle C channel = 1 # This is MIDI channel 1 note_on_action = 0x90 major_steps = [2, 2, 1, 2, 2, 2, 1, 0] velocity = 127 note = root_note for step in major_steps: send_midi((note_on_action | channel, note, velocity)) sleep(0.1) send_midi((note_on_action | channel, note, 0)) # A note-off is just a note-on with velocity 0 note += step sleep(0.2) if __name__=='__main__': while True: play_a_scale()
simplecoremidi/examples/play_a_scale.py
667
This is middle C This is MIDI channel 1 A note-off is just a note-on with velocity 0
84
en
0.973469
# if funktioniert (fast) wie in allen anderen Sprachen # - Einrückungen ersetzen { } Gilt für Python generell! # - Es gibt ein elif statt einem else if weight = 50 # kg height = 190 # cm bmi = weight / (height/100)**2 # bmi < 18.5 : Untergewicht # bmi > 25 : Übergewicht # sonst : Normalgewicht if bmi < 18.5: print("Untergewicht") print("Mehr essen!") elif bmi > 25: print("Übergewicht") else: print("Normalgewicht")
Crashkurs Python/03_if.py
449
if funktioniert (fast) wie in allen anderen Sprachen - Einrückungen ersetzen { } Gilt für Python generell! - Es gibt ein elif statt einem else if kg cm bmi < 18.5 : Untergewicht bmi > 25 : Übergewicht sonst : Normalgewicht
229
de
0.978166
import sys import os import math import shutil import disk_sort import struct import operator import logging from decimal import Decimal from fractions import Fraction import numpy from scipy.linalg import eig import scipy.ndimage import cProfile import pstats from osgeo import gdal, ogr import pygeoprocessing.geoprocessing import shutil logging.basicConfig(format='%(asctime)s %(name)-20s %(levelname)-8s \ %(message)s', level=logging.DEBUG, datefmt='%m/%d/%Y %H:%M:%S ') LOGGER = logging.getLogger('invest_natcap.scenario_generator.scenario_generator') def calculate_weights(arr, rounding=4): PLACES = Decimal(10) ** -(rounding) # get eigenvalues and vectors evas, eves = eig(arr) # get primary eigenvalue and vector eva = max(evas) eva_idx = evas.tolist().index(eva) eve = eves.take((eva_idx,), axis=1) # priority vector = normalized primary eigenvector normalized = eve / sum(eve) # turn into list of real part values vector = [abs(e[0]) for e in normalized] # return nice rounded Decimal values with labels return [ Decimal( str(v) ).quantize(PLACES) for v in vector ] def calculate_priority(table_uri): table = [line.strip().split(",") for line in open(table_uri).readlines()] id_index = table[0].index("Id") cover_id_list = [row[id_index] for row in table] cover_id_list.pop(0) cover_id_index_list = [table[0].index(cover_id) for cover_id in cover_id_list] matrix = numpy.zeros((len(cover_id_list),len(cover_id_list))) for row in range(len(cover_id_list)): for col in range(row+1): matrix[row][col] = float(table[row+1][cover_id_index_list[col]]) matrix[col][row] = 1 / matrix[row][col] cover_id_list = [int(cover_id) for cover_id in cover_id_list] return dict(zip(cover_id_list, calculate_weights(matrix, 4))) def calculate_distance_raster_uri(dataset_in_uri, dataset_out_uri): # Compute pixel distance pygeoprocessing.geoprocessing.distance_transform_edt(dataset_in_uri, dataset_out_uri) # Convert to meters def pixel_to_meters_op(x): x[x != nodata] *= cell_size return x cell_size = pygeoprocessing.geoprocessing.get_cell_size_from_uri(dataset_in_uri) nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(dataset_out_uri) tmp = pygeoprocessing.geoprocessing.temporary_filename() pygeoprocessing.geoprocessing.vectorize_datasets( [dataset_out_uri], \ pixel_to_meters_op, \ tmp, \ gdal.GDT_Float64, \ nodata, \ cell_size, \ 'union', \ vectorize_op = False) def identity_op(x): return x pygeoprocessing.geoprocessing.vectorize_datasets( [tmp], \ identity_op, \ dataset_out_uri, \ gdal.GDT_Float64, \ nodata, \ cell_size, \ 'union', \ vectorize_op = False) # Compute raster stats so the raster is viewable in QGIS and Arc pygeoprocessing.geoprocessing.calculate_raster_stats_uri(dataset_out_uri) ##def calculate_distance_raster_uri(dataset_in_uri, dataset_out_uri, cell_size = None, max_distance = None): ## if cell_size == None: ## cell_size = pygeoprocessing.geoprocessing.get_cell_size_from_uri(dataset_in_uri) ## ## memory_array = pygeoprocessing.geoprocessing.load_memory_mapped_array(dataset_in_uri, pygeoprocessing.geoprocessing.temporary_filename()) ## ## memory_array = scipy.ndimage.morphology.distance_transform_edt(memory_array) * cell_size ## ## nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(dataset_in_uri) ## #### if max_distance != None: #### memory_array[memory_array > max_distance] = nodata ## ## pygeoprocessing.geoprocessing.new_raster_from_base_uri(dataset_in_uri, dataset_out_uri, 'GTiff', nodata, gdal.GDT_Float32) ## ## dataset_out = gdal.Open(dataset_out_uri, 1) ## band = dataset_out.GetRasterBand(1) ## band.WriteArray(memory_array) ## ## band = None ## dataset_out = None shapeTypes= {0: "Null Shape", 1: "Point", 3: "PolyLine", 5: "Polygon", 8: "MultiPoint", 11: "PointZ", 13: "PolyLineZ", 15: "PolygonZ", 18: "MultiPointZ", 21: "PointM", 23: "PolyLineM", 25: "PolygonM", 28: "MultiPointM", 31: "MultiPatch"} def get_geometry_type_from_uri(datasource_uri): datasource = open(datasource_uri, 'r') datasource.seek(32) shape_type ,= struct.unpack('<i',datasource.read(4)) datasource.close() return shape_type def get_transition_set_count_from_uri(dataset_uri_list): cell_size = pygeoprocessing.geoprocessing.get_cell_size_from_uri(dataset_uri_list[0]) lulc_nodata = int(pygeoprocessing.geoprocessing.get_nodata_from_uri(dataset_uri_list[0])) nodata = 0 #reclass rasters to compact bit space lulc_codes = set() unique_raster_values_count = {} for dataset_uri in dataset_uri_list: unique_raster_values_count[dataset_uri] = pygeoprocessing.geoprocessing.unique_raster_values_count(dataset_uri) lulc_codes.update(unique_raster_values_count[dataset_uri].keys()) lulc_codes = list(lulc_codes) lulc_codes.sort() if len(lulc_codes) < 2 ** 8: data_type = gdal.GDT_UInt16 shift = 8 elif len(lulc_codes) < 2 ** 16: data_type = gdal.GDT_UInt32 shift = 16 else: raise ValueError, "Too many LULC codes." #renumber and reclass rasters reclass_orig_dict = dict(zip(lulc_codes,range(1,len(lulc_codes)+1))) reclass_dest_dict = {} for key in reclass_orig_dict: reclass_dest_dict[key] = reclass_orig_dict[key] << shift def add_op(orig, dest): return orig + dest counts={} for i in range(len(dataset_uri_list)-1): orig_uri = pygeoprocessing.geoprocessing.temporary_filename() dest_uri = pygeoprocessing.geoprocessing.temporary_filename() multi_uri = pygeoprocessing.geoprocessing.temporary_filename() #reclass orig values pygeoprocessing.geoprocessing.reclassify_dataset_uri(dataset_uri_list[i], reclass_orig_dict, orig_uri, data_type, nodata, exception_flag="values_required") #reclass dest values pygeoprocessing.geoprocessing.reclassify_dataset_uri(dataset_uri_list[i+1], reclass_dest_dict, dest_uri, data_type, nodata, exception_flag="values_required") #multiplex orig with dest pygeoprocessing.geoprocessing.vectorize_datasets([orig_uri, dest_uri], add_op, multi_uri, data_type, nodata, cell_size, "union") #get unique counts counts[i]=pygeoprocessing.geoprocessing.unique_raster_values_count(multi_uri, False) restore_classes = {} for key in reclass_orig_dict: restore_classes[reclass_orig_dict[key]] = key restore_classes[nodata] = lulc_nodata LOGGER.debug("Decoding transition table.") transitions = {} for key in counts: transitions[key]={} for k in counts[key]: try: orig = restore_classes[k % (2**shift)] except KeyError: orig = lulc_nodata try: dest = restore_classes[k >> shift] except KeyError: dest = lulc_nodata try: transitions[key][orig][dest] = counts[key][k] except KeyError: transitions[key][orig] = {dest : counts[key][k]} return unique_raster_values_count, transitions def generate_chart_html(cover_dict, cover_names_dict, workspace_dir): html = "\n<table BORDER=1>" html += "\n<TR><td>Id</td><td>% Before</td><td>% After</td></TR>" cover_id_list = cover_dict.keys() cover_id_list.sort() cover_id_list_chart = cover_names_dict.keys() cover_id_list_chart.sort() pixcount = 0 for cover_id in cover_id_list: pixcount += cover_dict[cover_id][0] pixcount = float(pixcount) for cover_id in cover_id_list: html += "\n<TR><td>%i</td><td>%i</td><td>%i</td></TR>" % (cover_id, (cover_dict[cover_id][0] / pixcount) * 100, (cover_dict[cover_id][1] / pixcount) * 100 ) html += "\n<table>" #create three charts for original, final and change thecharts = [ ['Original',0], ['Final',1], ['Change',2] ] hainitial = "" hainitialnegative = "" hainitiallist = [] hafinal = "" hafinalnegative = "" hafinallist = [] hachange = "" hachangelist = [] haall = [] initialcover = [] finalcover = [] for cover_id in cover_id_list_chart: try: initialcover.append((cover_dict[cover_id][0] / pixcount) * 100) except KeyError: initialcover.append(0) try: finalcover.append((cover_dict[cover_id][1] / pixcount) * 100) except KeyError: finalcover.append(0) #return html html += "<style type='text/css'>" html += "body {font-family: Arial, Helvetica, sans-serif; font-size: 0.9em;}" html += "table#results {margin: 20px auto}" html += "table#results th {text-align: left}" html += "</style>" html += "<script type='text/javascript'>\n" html += "var chart,\n" categories = [] html += "categories = [" for cover_id in cover_id_list_chart: #pass categories.append("'"+cover_names_dict[cover_id]+"'") html += ",".join(categories) html += "]\n" html +="$(document).ready(function() {\n" for x in initialcover: hainitial = hainitial +str(x)+"," hainitialnegative = hainitialnegative + "0," hainitiallist.append(float(x)) temp = [] temp.append(hainitial) temp.append(hainitialnegative) haall.append(temp) thecharts[0].append(max(hainitiallist)) thecharts[0].append(min(hainitiallist)) for x in finalcover: hafinal = hafinal +str(x)+"," hafinalnegative = hafinalnegative + "0," hafinallist.append(float(x)) temp = [] temp.append(hafinal) temp.append(hafinalnegative) haall.append(temp) thecharts[1].append(max(hafinallist)) thecharts[1].append(min(hafinallist)) for x in range(len(initialcover)): hachange = hachange + str(float(finalcover[x]) - float(initialcover[x]))+"," hachangelist.append(float(finalcover[x]) - float(initialcover[x])) #split the change values hachangelistnegative = "" hachangelistpositive = "" for item in hachangelist: if item < 0: hachangelistnegative=hachangelistnegative+str(item)+"," hachangelistpositive=hachangelistpositive+"0," else: hachangelistpositive=hachangelistpositive+str(item)+"," hachangelistnegative=hachangelistnegative+"0," temp = [] temp.append(hachangelistpositive) temp.append(hachangelistnegative) haall.append(temp) thecharts[2].append(max(hachangelist)) thecharts[2].append(min(hachangelist)) if thecharts[0][2] > thecharts[1][2]: thecharts[1][2] = thecharts[0][2] thecharts[2][2] = thecharts[0][2] else: thecharts[0][2] = thecharts[1][2] thecharts[2][2] = thecharts[1][2] for x in thecharts: if x[0] == 'Change': themin = x[3] else: themin = 0 html += "chart = new Highcharts.Chart({\n" html += "chart: {renderTo: '"+x[0]+"container',defaultSeriesType: 'bar'}," html += "title: {text: '"+x[0]+" Landcover'}," html += "subtitle: {text: ''}," html += "xAxis: [{categories: categories,reversed: false}, {opposite: true, reversed: false,categories: categories,linkedTo: 0}]," html += "yAxis: {title: {text: null},labels: {formatter: function(){return Math.abs(this.value)}},min: "+str(themin)+",max: "+str(x[2])+"}," html += "plotOptions: {series: { stacking: 'normal', showInLegend: false } }," html += "tooltip: { formatter: function(){return '<b>'+ this.point.category +'</b><br/>'+'Area: '+ Highcharts.numberFormat(Math.abs(this.point.y), 0)+'%';}}," html += "series: [{name: ''," html += "data: ["+haall[x[1]][0]+"]}, {" html += "name: ''," html += "data: ["+haall[x[1]][1]+"]}]});\n" html += "});\n" html += "</script>\n" for x in thecharts: html += "<div id='"+x[0]+"container' style='width: 800px; height: 400px; margin: 20px 0'></div>\n" return html def filter_fragments(input_uri, size, output_uri): #clump and sieve LOGGER.debug("Filtering patches smaller than %i from %s.", size, input_uri) src_ds = gdal.Open(input_uri) src_band = src_ds.GetRasterBand(1) src_array = src_band.ReadAsArray() driver = gdal.GetDriverByName("GTiff") driver.CreateCopy(output_uri, src_ds, 0 ) dst_ds = gdal.Open(output_uri, 1) dst_band = dst_ds.GetRasterBand(1) dst_array = numpy.copy(src_array) suitability_values = numpy.unique(src_array) if suitability_values[0] == 0: suitability_values = suitability_values[1:] #8 connectedness preferred, 4 connectedness allowed eight_connectedness = numpy.array([[1, 1, 1], [1, 1, 1], [1, 1, 1]]) four_connectedness = numpy.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) suitability_values_count = suitability_values.size for v in range(1, suitability_values_count): LOGGER.debug('Processing suitability value ' + \ str(suitability_values.size - v)) value = suitability_values[v] # Pixels of interest set to 1, 0 otherwise mask = src_array == value # Number of pixels to process ones_in_mask = numpy.sum(mask) # Label and count disconnected components (fragments) label_im, nb_labels = scipy.ndimage.label(mask, four_connectedness) # Compute fragment sizes fragment_sizes = \ scipy.ndimage.sum(mask, label_im, range(nb_labels + 1)) # List fragments fragment_labels = numpy.array(range(nb_labels + 1)) # Discard large fragments small_fragment_mask = numpy.where(fragment_sizes <= size) # Gather small fragment information small_fragment_sizes = fragment_sizes[small_fragment_mask] small_fragment_labels = fragment_labels[small_fragment_mask] combined_small_fragment_size = numpy.sum(small_fragment_sizes) # Find each fragment fragments_location = scipy.ndimage.find_objects(label_im, nb_labels) removed_pixels = 0 small_fragment_labels_count = small_fragment_labels.size for l in range(small_fragment_labels_count-1): label = small_fragment_labels[l+1] last_label = small_fragment_labels[l] size = small_fragment_sizes[l+1] source = label_im[fragments_location[last_label]] target = dst_array[fragments_location[last_label]] pixels_to_remove = numpy.where(source == label) target[pixels_to_remove] = 0 dst_band.WriteArray(dst_array) def sum_uri(dataset_uri, datasource_uri): """Wrapper call to pygeoprocessing.geoprocessing.aggregate_raster_values_uri to extract total :param dataset_uri: The uri for the input raster. :type dataset_uri: str :return: None :rtype: None """ total = pygeoprocessing.geoprocessing.aggregate_raster_values_uri(dataset_uri, datasource_uri).total return total.__getitem__(total.keys().pop()) def execute(args): ### #overiding, non-standard field names ### # Preliminary tests if ('transition' in args) and ('suitability' in args): assert args['transition'] != args['suitability'], \ 'Transition and suitability tables are the same: ' + \ args['transition'] + '. The model expects different tables.' #transition table fields args["transition_id"] = "Id" args["percent_field"] = "Percent Change" args["area_field"] = "Area Change" args["priority_field"] = "Priority" args["proximity_field"] = "Proximity" args["proximity_weight"] = "0.3" args["patch_field"] = "Patch ha" #factors table fields args["suitability_id"] = "Id" args["suitability_layer"] = "Layer" args["suitability_weight"] = "Wt" args["suitability_field"] = "Suitfield" args["distance_field"] = "Dist" args["suitability_cover_id"] = "Cover ID" #exercise fields args["returns_cover_id"] = "Cover ID" args["returns_layer"] = "/Users/olwero/Dropbox/Work/Ecosystem_Services/NatCap/Olympics/2014/Scenarios/Exercise/inputtest/returns.csv" ### #get parameters, set outputs ### workspace = args["workspace_dir"] if not os.path.exists(workspace): os.makedirs(workspace) landcover_uri = args["landcover"] if len(args["suffix"]) > 0: suffix = "_" + args["suffix"].strip("_") else: suffix = "" intermediate_dir = "intermediate" if not os.path.exists(os.path.join(workspace, intermediate_dir)): os.makedirs(os.path.join(workspace, intermediate_dir)) proximity_weight = float(args["proximity_weight"]) #it might be better to just check if factors being used try: physical_suitability_weight = float(args["weight"]) except KeyError: physical_suitability_weight = 0.5 ##output file names #absolute paths landcover_resample_uri = os.path.join(workspace, "resample" + suffix + ".tif") landcover_transition_uri = os.path.join(workspace,"transitioned" + suffix + ".tif") override_dataset_uri = os.path.join(workspace,"override" + suffix + ".tif") landcover_htm_uri = os.path.join(workspace,"scenario-output-summary" + suffix + ".html") pygeoprocessing.geoprocessing.create_directories([workspace]) #relative paths, or with patterned name transition_name = os.path.join(intermediate_dir, "transition_%i" + suffix + ".tif") suitability_name = os.path.join(intermediate_dir, "%s_%s" + suffix + ".tif") normalized_name = os.path.join(intermediate_dir, "%s_%s_norm" + suffix + ".tif") combined_name = os.path.join(intermediate_dir, "factors_%s" + suffix + ".tif") constraints_name = os.path.join(intermediate_dir, "constraints" + suffix + ".tif") filter_name = os.path.join(intermediate_dir, "filter_%i" + suffix + ".tif") factors_name = os.path.join(intermediate_dir, "suitability_%s" + suffix + ".tif") cover_name = os.path.join(intermediate_dir, "cover_%i" + suffix + ".tif") proximity_name = os.path.join(intermediate_dir, "proximity_%s" + suffix + ".tif") normalized_proximity_name = os.path.join(intermediate_dir, "proximity_norm_%s" + suffix + ".tif") adjusted_suitability_name = os.path.join(intermediate_dir, "adjusted_suitability_%s" + suffix + ".tif") scenario_name = "scenario" + suffix + ".tif" ### #constants ### raster_format = "GTiff" transition_type = gdal.GDT_Int16 transition_nodata = -1 change_nodata = -9999 #value to multiply transition matrix entries (ie covert 10 point scale to 100 point scale) transition_scale = 10 distance_scale = 100 suitability_nodata = 0 suitability_type = gdal.GDT_Int16 def suitability_op(trans, suit): if trans == 0: return 0 return round(((1 - physical_suitability_weight) * trans)\ + (physical_suitability_weight * suit)) ds_type = "GTiff" driver = gdal.GetDriverByName(ds_type) ### #validate data ### #raise warning if nothing is going to happen if not any([args["calculate_transition"], args["calculate_factors"], args["override_layer"]]): msg = "You must select at least one of the following: specify transitions, use factors, or override layer." LOGGER.error(msg) raise ValueError, msg ##transition table validation #raise error if transition table provided, but not used if args["transition"] and not(args["calculate_transition"] or args["calculate_factors"]): msg = "Transition table provided but not used." LOGGER.warn(msg) #raise ValueError, msg transition_dict = {} if args["calculate_transition"] or args["calculate_factors"]: #load transition table transition_dict = pygeoprocessing.geoprocessing.get_lookup_from_csv(args["transition"], args["transition_id"]) #raise error if LULC contains cover id's not in transition table landcover_count_dict = pygeoprocessing.geoprocessing.unique_raster_values_count(landcover_uri) missing_lulc = set(landcover_count_dict).difference(transition_dict.keys()) if len(missing_lulc) > 0 : missing_lulc = list(missing_lulc) missing_lulc.sort() mising_lulc = ", ".join([str(l) for l in missing_lulc]) msg = "Missing suitability information for cover(s) %s." % missing_lulc LOGGER.error(msg) raise ValueError, msg for cover_id in transition_dict: #raise error if percent change for new LULC if (transition_dict[cover_id][args["percent_field"]] > 0) and not (cover_id in landcover_count_dict): msg = "Cover %i does not exist in LULC and therefore cannot have a percent change." % cover_id LOGGER.error(msg) raise ValueError, msg #raise error if change by percent and area both specified if (transition_dict[cover_id][args["percent_field"]] > 0) and (transition_dict[cover_id][args["area_field"]] > 0): msg = "Cover %i cannot have both an increase by percent and area." % cover_id LOGGER.error(msg) raise ValueError, msg ##factor parameters validation if args["calculate_factors"]: pass #error if overall physical weight not in [0, 1] range ##factor table validation #if polygon no distance field allowed #if point or line, integer distance field only #error if same factor twice for same coverage ### #resample, align and rasterize data ### if args["calculate_priorities"]: LOGGER.info("Calculating priorities.") priorities_dict = calculate_priority(args["priorities_csv_uri"]) #check geographic extents, projections ## #validate resampling size ## if args["resolution"] != "": ## if args["resolution"] < pygeoprocessing.geoprocessing.get_cell_size_from_uri(landcover_uri): ## msg = "The analysis resolution cannot be smaller than the input." ## LOGGER.error(msg) ## raise ValueError, msg ## ## else: ## LOGGER.info("Resampling land cover.") ## #gdal.GRA_Mode might be a better resample method, but requires GDAL >= 1.10.0 ## bounding_box = pygeoprocessing.geoprocessing.get_bounding_box(landcover_uri) ## pygeoprocessing.geoprocessing.resize_and_resample_dataset_uri(landcover_uri, ## bounding_box, ## args["resolution"], ## landcover_resample_uri, ## "nearest") ## LOGGER.debug("Changing landcover uri to resampled uri.") ## landcover_uri = landcover_resample_uri cell_size = pygeoprocessing.geoprocessing.get_cell_size_from_uri(landcover_uri) suitability_transition_dict = {} if args["calculate_transition"]: for next_lulc in transition_dict: this_uri = os.path.join(workspace, transition_name % next_lulc) #construct reclass dictionary reclass_dict = {} all_zeros = True for this_lulc in transition_dict: value = int(transition_dict[this_lulc][str(next_lulc)]) reclass_dict[this_lulc] = value * transition_scale all_zeros = all_zeros and (value == 0) if not all_zeros: #reclass lulc by reclass_dict pygeoprocessing.geoprocessing.reclassify_dataset_uri(landcover_uri, reclass_dict, this_uri, transition_type, suitability_nodata, exception_flag = "values_required") #changing nodata value so 0's no longer nodata dataset = gdal.Open(this_uri, 1) band = dataset.GetRasterBand(1) nodata = band.SetNoDataValue(transition_nodata) dataset = None suitability_transition_dict[next_lulc] = this_uri suitability_factors_dict = {} if args["calculate_factors"]: factor_dict = pygeoprocessing.geoprocessing.get_lookup_from_csv(args["suitability"], args["suitability_id"]) factor_uri_dict = {} factor_folder = args["suitability_folder"] if not args["factor_inclusion"]: option_list=["ALL_TOUCHED=TRUE"] else: option_list = ["ALL_TOUCHED=FALSE"] for factor_id in factor_dict: factor = factor_dict[factor_id][args["suitability_layer"]] factor_stem, _ = os.path.splitext(factor) suitability_field_name = factor_dict[factor_id][args["suitability_field"]] distance = factor_dict[factor_id][args["distance_field"]] cover_id = int(factor_dict[factor_id][args["suitability_cover_id"]]) weight = int(factor_dict[factor_id][args["suitability_weight"]]) LOGGER.debug("Found reference to factor (%s, %s, %s) for cover %i.", factor_stem, suitability_field_name, distance, cover_id) if not (factor_stem, suitability_field_name, distance) in factor_uri_dict: factor_uri = os.path.join(factor_folder, factor) if not os.path.exists(factor_uri): msg = "Missing file %s." % factor_uri LOGGER.error(msg) raise ValueError, msg shape_type = get_geometry_type_from_uri(factor_uri) LOGGER.debug("Processing %s.", shapeTypes[shape_type]) if shape_type in [5, 15, 25, 31]: #polygon LOGGER.info("Rasterizing %s using sutibality field %s.", factor_stem, suitability_field_name) ds_uri = os.path.join(workspace, suitability_name % (factor_stem, suitability_field_name)) burn_value = [1] suitability_field = ["ATTRIBUTE=%s" % suitability_field_name] gdal_format = gdal.GDT_Float64 pygeoprocessing.geoprocessing.new_raster_from_base_uri(landcover_uri, ds_uri, raster_format, transition_nodata, gdal_format, fill_value = 0) pygeoprocessing.geoprocessing.rasterize_layer_uri(ds_uri, factor_uri, burn_value, option_list=option_list + suitability_field) factor_uri_dict[(factor_stem, suitability_field_name, distance)] = ds_uri elif shape_type in [1, 3, 8, 11, 13, 18, 21, 23, 28]: #point or line # For features with no area, it's (almost) impossible to # hit the center pixel, so we use ALL_TOUCHED=TRUE option_list=["ALL_TOUCHED=TRUE"] distance = int(distance) ds_uri = os.path.join(workspace, suitability_name % (factor_stem, str(distance) + '_raw_raster')) distance_uri = os.path.join(workspace, suitability_name % (factor_stem, str(distance) + '_raw_distance')) fdistance_uri = os.path.join(workspace, suitability_name % (factor_stem, distance)) normalized_uri = os.path.join(workspace, normalized_name % (factor_stem, distance)) burn_value = [1] LOGGER.info("Buffering rasterization of %s to distance of %i.", factor_stem, distance) gdal_format = gdal.GDT_Byte pygeoprocessing.geoprocessing.new_raster_from_base_uri(landcover_uri, ds_uri, raster_format, -1, gdal_format) landcover_nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(landcover_uri) ds_nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(ds_uri) pygeoprocessing.geoprocessing.vectorize_datasets([landcover_uri], \ lambda x: 0 if x != landcover_nodata else -1, \ ds_uri, \ pygeoprocessing.geoprocessing.get_datatype_from_uri(ds_uri), \ ds_nodata, \ pygeoprocessing.geoprocessing.get_cell_size_from_uri(ds_uri), \ 'intersection') pygeoprocessing.geoprocessing.rasterize_layer_uri(ds_uri, factor_uri, burn_value, option_list) calculate_distance_raster_uri(ds_uri, distance_uri) def threshold(value): result = numpy.where(value > distance, transition_nodata, value) return numpy.where(value == transition_nodata, transition_nodata, result) pygeoprocessing.geoprocessing.vectorize_datasets([distance_uri], threshold, fdistance_uri, pygeoprocessing.geoprocessing.get_datatype_from_uri(distance_uri), transition_nodata, cell_size, "union", vectorize_op = False) pygeoprocessing.geoprocessing.calculate_raster_stats_uri(fdistance_uri) minimum, maximum, _, _ = pygeoprocessing.geoprocessing.get_statistics_from_uri(fdistance_uri) def normalize_op(value): diff = float(maximum - minimum) return numpy.where( value == transition_nodata, suitability_nodata, ((distance_scale - 1) - (((value - minimum) / \ diff) * (distance_scale - 1))) + 1) pygeoprocessing.geoprocessing.vectorize_datasets([fdistance_uri], normalize_op, normalized_uri, transition_type, transition_nodata, cell_size, "union", vectorize_op = False) factor_uri_dict[(factor_stem, suitability_field_name, distance)] = normalized_uri else: raise ValueError, "Invalid geometry type %i." % shape_type # Apply nodata to the factors raster landcover_nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(landcover_uri) temp_uri = pygeoprocessing.geoprocessing.temporary_filename() def apply_nodata_op(landcover, value): return numpy.where(landcover == landcover_uri, 0, value) pygeoprocessing.geoprocessing.vectorize_datasets( \ [landcover_uri, factor_uri_dict[(factor_stem, suitability_field_name, distance)]], apply_nodata_op, temp_uri, transition_type, transition_nodata, cell_size, "union", vectorize_op = False) def identity_op(x): return x pygeoprocessing.geoprocessing.vectorize_datasets( \ [temp_uri], identity_op, factor_uri_dict[(factor_stem, suitability_field_name, distance)], transition_type, transition_nodata, cell_size, "union", vectorize_op = False) else: LOGGER.debug("Skipping already processed suitability layer.") LOGGER.debug("Adding factor (%s, %s, %s) to cover %i suitability list.", factor_stem, suitability_field_name, distance, cover_id) if cover_id in suitability_factors_dict: suitability_factors_dict[cover_id].append((factor_uri_dict[(factor_stem, suitability_field_name, distance)], weight)) else: suitability_factors_dict[cover_id] = [(factor_uri_dict[(factor_stem, suitability_field_name, distance)], weight)] for cover_id in suitability_factors_dict: if len(suitability_factors_dict[cover_id]) > 1: LOGGER.info("Combining factors for cover type %i.", cover_id) ds_uri = os.path.join(workspace, combined_name % cover_id) uri_list, weights_list = apply(zip, suitability_factors_dict[cover_id]) total = float(sum(weights_list)) weights_list = [weight / total for weight in weights_list] def weighted_op(*values): result = (values[0] * weights_list[0]).astype(float) for v, w in zip(values[1:], weights_list[1:]): result += v * w return result # print('------files:', uri_list, weights_list) pygeoprocessing.geoprocessing.vectorize_datasets(list(uri_list), weighted_op, ds_uri, suitability_type, transition_nodata, cell_size, "union", vectorize_op = False) suitability_factors_dict[cover_id] = ds_uri else: suitability_factors_dict[cover_id] = suitability_factors_dict[cover_id][0][0] suitability_dict = {} if args["calculate_transition"]: suitability_dict = suitability_transition_dict if args["calculate_factors"]: for cover_id in suitability_factors_dict: if cover_id in suitability_dict: LOGGER.info("Combining suitability for cover %i.", cover_id) ds_uri = os.path.join(workspace, factors_name % cover_id) print('cover_ids', suitability_dict.keys()) pygeoprocessing.geoprocessing.vectorize_datasets([suitability_transition_dict[cover_id], suitability_factors_dict[cover_id]], suitability_op, ds_uri, transition_type, transition_nodata, cell_size, "union") suitability_dict[cover_id] = ds_uri else: suitability_dict[cover_id] = suitability_factors_dict[cover_id] elif args["calculate_factors"]: suitability_dict = suitability_factors_dict #clump and sieve for cover_id in transition_dict: if (transition_dict[cover_id][args["patch_field"]] > 0) and (cover_id in suitability_dict): LOGGER.info("Filtering patches from %i.", cover_id) size = 10000 * int(math.ceil( \ transition_dict[cover_id][args["patch_field"]] / \ (cell_size ** 2))) output_uri = os.path.join(workspace, filter_name % cover_id) filter_fragments(suitability_dict[cover_id], size, output_uri) suitability_dict[cover_id] = output_uri ### #compute intermediate data if needed ### #contraints raster (reclass using permability values, filters on clump size) if args["calculate_constraints"]: LOGGER.info("Rasterizing constraints.") constraints_uri = args["constraints"] constraints_field_name = args["constraints_field"] constraints_ds_uri = os.path.join(workspace, constraints_name) option_list = ["ALL_TOUCHED=FALSE"] burn_value = [0] constraints_field = ["ATTRIBUTE=%s" % constraints_field_name] gdal_format = gdal.GDT_Float64 pygeoprocessing.geoprocessing.new_raster_from_base_uri(landcover_uri, constraints_ds_uri, raster_format, transition_nodata, gdal_format, fill_value = 1) pygeoprocessing.geoprocessing.rasterize_layer_uri(constraints_ds_uri, constraints_uri, burn_value, option_list=option_list + constraints_field) # Check that the values make sense raster = gdal.Open(constraints_ds_uri) band = raster.GetRasterBand(1) array = band.ReadAsArray() unique = numpy.unique(array) assert (unique[0] >= 0.0) and (unique[-1] <= 1.0), \ 'Invalid raster value in field ' + constraints_field_name + ' in ' \ + constraints_uri else: LOGGER.info("Constraints not included.") proximity_dict = {} if args["calculate_proximity"]: LOGGER.info("Calculating proximity.") cover_types = transition_dict.keys() for cover_id in transition_dict: if transition_dict[cover_id][args["proximity_field"]] > 0 and cover_id in suitability_dict: distance = int(transition_dict[cover_id][args["proximity_field"]]) LOGGER.info("Calculating proximity for %i.", cover_id) reclass_dict = dict(zip(cover_types, [1] * len(cover_types))) reclass_dict[cover_id] = 0 ds_uri = os.path.join(workspace, cover_name % cover_id) distance_uri = pygeoprocessing.geoprocessing.temporary_filename() fdistance_uri = os.path.join(workspace, proximity_name % cover_id) normalized_uri = os.path.join(workspace, normalized_proximity_name % cover_id) pygeoprocessing.geoprocessing.reclassify_dataset_uri(landcover_uri, reclass_dict, ds_uri, transition_type, transition_nodata, exception_flag = "values_required") calculate_distance_raster_uri(ds_uri, distance_uri) def threshold(value): if value > distance: return transition_nodata return value pygeoprocessing.geoprocessing.vectorize_datasets([distance_uri], threshold, fdistance_uri, pygeoprocessing.geoprocessing.get_datatype_from_uri(distance_uri), transition_nodata, cell_size, "union") minimum, maximum, _, _ = pygeoprocessing.geoprocessing.get_statistics_from_uri(fdistance_uri) assert minimum < maximum, "Wrong distance (min, max) = (" + \ str(minimum) + ", " + str(maximum) + ") in " + fdistance_uri def normalize_op(value): if value == transition_nodata: return suitability_nodata else: return ((distance_scale - 1) \ - (((value - minimum) \ / float(maximum - minimum)) \ * (distance_scale - 1))) \ + 1 pygeoprocessing.geoprocessing.vectorize_datasets([fdistance_uri], normalize_op, normalized_uri, transition_type, transition_nodata, cell_size, "union") proximity_dict[cover_id] = normalized_uri def es_change_op(final_es ,initial_es): return final_es - initial_es def constraint_op(suit, cons): return suit * cons def proximity_op(suit, prox): v = suit + (prox * proximity_weight) if v > 100: return 100 else: return v def constraint_proximity_op(suit, cons, prox): v = (cons * suit) + (prox * proximity_weight) if v > 100: return 100 else: return v for cover_id in suitability_dict: suitability_uri = os.path.join(workspace, adjusted_suitability_name % cover_id) if args["calculate_constraints"]: if cover_id in proximity_dict: LOGGER.info("Combining suitability, proximity, and constraints for %i.", cover_id) uri_list = [suitability_dict[cover_id], constraints_ds_uri, proximity_dict[cover_id]] LOGGER.info("Vectorizing: %s", ", ".join(uri_list)) pygeoprocessing.geoprocessing.vectorize_datasets(uri_list, constraint_proximity_op, suitability_uri, transition_type, transition_nodata, cell_size, "union") suitability_dict[cover_id] = suitability_uri else: LOGGER.info("Combining suitability and constraints for %i.", cover_id) uri_list = [suitability_dict[cover_id], constraints_ds_uri] # print('------suitability and constraint files:', uri_list) LOGGER.info("Vectorizing: %s", ", ".join(uri_list)) pygeoprocessing.geoprocessing.vectorize_datasets(uri_list, constraint_op, suitability_uri, transition_type, transition_nodata, cell_size, "union") suitability_dict[cover_id] = suitability_uri elif cover_id in proximity_dict: LOGGER.info("Combining suitability and proximity for %i.", cover_id) uri_list = [suitability_dict[cover_id], proximity_dict[cover_id]] LOGGER.info("Vectorizing: %s", ", ".join(uri_list)) pygeoprocessing.geoprocessing.vectorize_datasets(uri_list, proximity_op, suitability_uri, transition_type, transition_nodata, cell_size, "union") suitability_dict[cover_id] = suitability_uri #normalize probabilities to be on a 10 point scale #probability raster (reclass using probability matrix) #proximity raster (gaussian for each landcover type, using max distance) #InVEST 2 uses 4-connectedness? #combine rasters for weighting into sutibility raster, multiply proximity by 0.3 #[suitability * (1-factor weight)] + (factors * factor weight) or only single raster ### #reallocate pixels (disk heap sort, randomly reassign equal value pixels, applied in order) ### #copy initial LULC scenario_uri = os.path.join(workspace, scenario_name) src_ds = gdal.Open(landcover_uri) n_cols = src_ds.RasterXSize n_rows = src_ds.RasterYSize dst_ds = driver.CreateCopy(scenario_uri, src_ds, 0) dst_ds = None src_ds = None #identify LULC types undergoing change change_list = [] if args["calculate_priorities"]: for cover_id in transition_dict: percent_change = transition_dict[cover_id][args["percent_field"]] area_change = transition_dict[cover_id][args["area_field"]] if percent_change > 0: change_list.append((priorities_dict[cover_id], cover_id, int((percent_change / 100.0) \ * landcover_count_dict[cover_id]))) elif area_change > 0: change_list.append((priorities_dict[cover_id], cover_id, 10000 * int(math.ceil(area_change \ / (cell_size**2))))) else: LOGGER.warn("Cover %i suitability specified, but no change indicated.", cover_id) else: for cover_id in transition_dict: percent_change = transition_dict[cover_id][args["percent_field"]] area_change = transition_dict[cover_id][args["area_field"]] if percent_change > 0: change_list.append((transition_dict[cover_id][args["priority_field"]], cover_id, int((percent_change / 100.0) \ * landcover_count_dict[cover_id]))) elif area_change > 0: change_list.append((transition_dict[cover_id][args["priority_field"]], cover_id, 10000 * int(math.ceil(area_change \ / (cell_size**2))))) else: LOGGER.warn("Cover %i suitability specified, but no change indicated.", cover_id) change_list.sort(reverse=True) #change pixels scenario_ds = gdal.Open(scenario_uri, 1) scenario_band = scenario_ds.GetRasterBand(1) scenario_array = scenario_band.ReadAsArray() unconverted_pixels = {} for index, (priority, cover_id, count) in enumerate(change_list): LOGGER.debug("Increasing cover %i by %i pixels.", cover_id, count) #open all lower priority suitability rasters and assign changed pixels value of 0 update_ds = {} update_bands = {} update_arrays = {} for _, update_id, _ in change_list[index+1:]: update_ds[update_id] = gdal.Open(suitability_dict[update_id], 1) update_bands[update_id] = update_ds[update_id].GetRasterBand(1) update_arrays[update_id] = update_bands[update_id].ReadAsArray() ##select pixels #open suitability raster src_ds = gdal.Open(suitability_dict[cover_id], 1) src_band = src_ds.GetRasterBand(1) src_array = src_band.ReadAsArray() pixels_changed = 0 suitability_values = list(numpy.unique(src_array)) suitability_values.sort(reverse=True) if suitability_values[-1]==0: suitability_values.pop(-1) for suitability_score in suitability_values: # Check if suitsbility is between 0 and 100 inclusive if abs(suitability_score - 50) > 50: print('suitability_values:', suitability_dict[cover_id]) for v in suitability_values: print v, ' ', assert abs(suitability_score - 50) <= 50, \ 'Invalid suitability score ' + str(suitability_score) if pixels_changed == count: LOGGER.debug("All necessay pixels converted.") break LOGGER.debug("Checking pixels with suitability of %i.", suitability_score) #mask out everything except the current suitability score mask = src_array == suitability_score #label patches label_im, nb_labels = scipy.ndimage.label(mask) #get patch sizes patch_sizes = scipy.ndimage.sum(mask, label_im, range(1, nb_labels + 1)) patch_labels = numpy.array(range(1, nb_labels + 1)) patch_locations = scipy.ndimage.find_objects(label_im, nb_labels) #randomize patch order numpy.random.shuffle(patch_labels) #check patches for conversion patch_label_count = patch_labels.size for l in range(patch_label_count): label = patch_labels[l] source = label_im[patch_locations[label-1]] target = scenario_array[patch_locations[label-1]] pixels_to_change = numpy.where(source == label) assert pixels_to_change[0].size == patch_sizes[label-1] if patch_sizes[label-1] + pixels_changed > count: #mask out everything except the current patch #patch = numpy.where(label_im == label) #patch_mask = numpy.zeros_like(scenario_array) patch_mask = numpy.zeros_like(target) #patch_mask[patch] = 1 patch_mask[pixels_to_change] = 1 #calculate the distance to exit the patch #tmp_array = scipy.ndimage.morphology.distance_transform_edt(patch_mask) tmp_array = scipy.ndimage.morphology.distance_transform_edt(patch_mask) #tmp_array = tmp_array[patch] tmp_array = tmp_array[pixels_to_change] #select the number of pixels that need to be converted tmp_index = numpy.argsort(tmp_array) tmp_index = tmp_index[:count - pixels_changed] #convert the selected pixels into coordinates #pixels_to_change = numpy.array(zip(patch[0], patch[1])) pixels_to_change = numpy.array(zip(pixels_to_change[0], pixels_to_change[1])) pixels_to_change = pixels_to_change[tmp_index] pixels_to_change = apply(zip, pixels_to_change) #change the pixels in the scenario #scenario_array[pixels_to_change] = cover_id target[pixels_to_change] = cover_id pixels_changed = count #alter other suitability rasters to prevent double conversion for _, update_id, _ in change_list[index+1:]: #update_arrays[update_id][pixels_to_change] = 0 target = update_arrays[update_id][patch_locations[label-1]] target[pixels_to_change] = 0 break else: #convert patch, increase count of changes target[pixels_to_change] = cover_id pixels_changed += patch_sizes[label-1] #alter other suitability rasters to prevent double conversion for _, update_id, _ in change_list[index+1:]: target = update_arrays[update_id][patch_locations[label-1]] target[pixels_to_change] = 0 #report and record unchanged pixels if pixels_changed < count: LOGGER.warn("Not all pixels converted.") unconverted_pixels[cover_id] = count - pixels_changed #write new suitability arrays for _, update_id, _ in change_list[index+1:]: update_bands[update_id].WriteArray(update_arrays[update_id]) update_arrays[update_id] = None update_bands[update_id] = None update_ds[update_id] = None scenario_band.WriteArray(scenario_array) scenario_array = None scenario_band = None scenario_ds = None #apply override if args["override_layer"]: LOGGER.info("Overriding pixels using values from field %s.", args["override_field"]) datasource = ogr.Open(args["override"]) layer = datasource.GetLayer() dataset = gdal.Open(scenario_uri, 1) if dataset == None: msg = "Could not open landcover transition raster." LOGGER.error(msg) raise IOError, msg if datasource == None: msg = "Could not open override vector." LOGGER.error(msg) raise IOError, msg if not bool(args["override_inclusion"]): LOGGER.debug("Overriding all touched pixels.") options = ["ALL_TOUCHED=TRUE", "ATTRIBUTE=%s" % args["override_field"]] else: LOGGER.debug("Overriding only pixels with covered center points.") options = ["ATTRIBUTE=%s" % args["override_field"]] gdal.RasterizeLayer(dataset, [1], layer, options=options) dataset.FlushCache() datasource = None dataset = None ### #tabulate coverages ### unique_raster_values_count, transitions = get_transition_set_count_from_uri([landcover_uri, scenario_uri]) htm = open(landcover_htm_uri,'w') htm.write("<html><head><title>Scenario Generator Report</title>") htm.write("<style type='text/css'>") htm.write("table {border-collapse: collapse; font-size: 1em;}") htm.write("td {padding: 10px;}") htm.write('body {font-family: Arial, Helvetica, sans-serif; font-size: 1em;}') htm.write('h2 {background: #DDDDDD; padding: 10px;}') htm.write("</style>") jquery_uri = os.path.join(os.path.dirname(os.path.abspath(__file__)), "jquery-1.6.2.min.js") htm.write("<script>\n" + open(jquery_uri).read() + "\n</script>") highcharts_uri = os.path.join(os.path.dirname(os.path.abspath(__file__)), "highcharts.js") htm.write("<script>\n" + open(highcharts_uri).read() + "\n</script>") htm.write("</head><body>") htm.write("<div style=''>") htm.write("<h1>Scenario Output Summary</h1>") htm.write("<h2>Initial Landscape</h2>") htm.write("\n<table BORDER=1>") initial_cover_id_list = unique_raster_values_count[landcover_uri].keys() initial_cover_id_list.sort() htm.write("\n<tr><td>ID</td><td>") htm.write("</td><td>".join([str(cover_id) for cover_id in initial_cover_id_list])) htm.write("\n</td></tr>") htm.write("\n<tr><td>Count</td><td>") htm.write("</td><td>".join([str(unique_raster_values_count[landcover_uri][cover_id]) for cover_id in initial_cover_id_list])) htm.write("\n</td></tr>") htm.write("\n</table>") htm.write("<h2>Scenario Landscape</h2>") htm.write("\n<table BORDER=1>") scenario_cover_id_list = unique_raster_values_count[scenario_uri].keys() scenario_cover_id_list.sort() htm.write("\n<tr><td>ID</td><td>") htm.write("</td><td>".join([str(cover_id) for cover_id in scenario_cover_id_list])) htm.write("\n</td></tr>") htm.write("\n<tr><td>Count</td><td>") htm.write("</td><td>".join([str(unique_raster_values_count[scenario_uri][cover_id]) for cover_id in scenario_cover_id_list])) htm.write("\n</td></tr>") htm.write("\n</table>") cover_dict = {} for cover_id in set(unique_raster_values_count[landcover_uri].keys()).union(set(unique_raster_values_count[scenario_uri].keys())): try: before = unique_raster_values_count[landcover_uri][cover_id] except KeyError: before = 0 try: after =unique_raster_values_count[scenario_uri][cover_id] except KeyError: after = 0 cover_dict[cover_id] = (before, after) htm.write("<h2>Change Table</h2>") cover_names_dict = {} transition_dict = pygeoprocessing.geoprocessing.get_lookup_from_csv(args["transition"], args["transition_id"]) cover_names_dict = {} for cover in transition_dict: cover_names_dict[cover] = transition_dict[cover]["Name"] htm.write(generate_chart_html(cover_dict, cover_names_dict, workspace)) htm.write("<h2>Transition Matrix</h2>") htm.write("\n<table BORDER=1>") htm.write("\n<tr><td>ID</td><td>") htm.write("</td><td>".join([str(cover_id) for cover_id in scenario_cover_id_list])) htm.write("\n</td></tr>") for initial_cover_id in initial_cover_id_list: htm.write("\n<tr><td>%i</td>" % initial_cover_id) for scenario_cover_id in scenario_cover_id_list: try: htm.write("<td>%i</td>" % transitions[0][initial_cover_id][scenario_cover_id]) except KeyError: htm.write("<td><FONT COLOR=lightgray>%i</FONT></td>" % 0) htm.write("\n</tr>") htm.write("\n</table>") unconverted_cover_id_list = unconverted_pixels.keys() unconverted_cover_id_list.sort() if len(unconverted_cover_id_list) > 0: htm.write("<h2>Unconverted Pixels</h2>") htm.write("\n<table BORDER=1>") htm.write("<tr><td>ID</td><td>Count</td></tr>") for cover_id in unconverted_cover_id_list: htm.write("<tr><td>%i</td><td>%i</td></tr>" % (cover_id, unconverted_pixels[cover_id])) htm.write("\n</table>") else: htm.write("<p><i>All target pixels converted.</i></p>") htm.write("\n</html>") #input CSVs input_csv_list = [] if args["calculate_priorities"]: input_csv_list.append((args["priorities_csv_uri"], "Priorities Table")) if args["calculate_transition"] or args["calculate_factors"]: input_csv_list.append((args["transition"], "Transition Table")) if args["calculate_factors"]: input_csv_list.append((args["suitability"], "Factors Table")) htm.write("<h1>Input Tables</h1>") for csv_uri, name in input_csv_list: table = "\n<table BORDER=1><tr><td>" + open(csv_uri).read().strip().replace(",","</td><td>").replace("\n","</td></tr><tr><td>") + "</td></tr></table>" htm.write("<h2>%s</h2>" % name) htm.write(table) htm.write("\n</div>\n</body>\n</html>") htm.close()
invest_natcap/scenario_generator/scenario_generator.py
59,952
get eigenvalues and vectors get primary eigenvalue and vector priority vector = normalized primary eigenvector turn into list of real part values return nice rounded Decimal values with labels Compute pixel distance Convert to meters Compute raster stats so the raster is viewable in QGIS and Arcdef calculate_distance_raster_uri(dataset_in_uri, dataset_out_uri, cell_size = None, max_distance = None): if cell_size == None: cell_size = pygeoprocessing.geoprocessing.get_cell_size_from_uri(dataset_in_uri) memory_array = pygeoprocessing.geoprocessing.load_memory_mapped_array(dataset_in_uri, pygeoprocessing.geoprocessing.temporary_filename()) memory_array = scipy.ndimage.morphology.distance_transform_edt(memory_array) * cell_size nodata = pygeoprocessing.geoprocessing.get_nodata_from_uri(dataset_in_uri) if max_distance != None: memory_array[memory_array > max_distance] = nodata pygeoprocessing.geoprocessing.new_raster_from_base_uri(dataset_in_uri, dataset_out_uri, 'GTiff', nodata, gdal.GDT_Float32) dataset_out = gdal.Open(dataset_out_uri, 1) band = dataset_out.GetRasterBand(1) band.WriteArray(memory_array) band = None dataset_out = Nonereclass rasters to compact bit spacerenumber and reclass rastersreclass orig valuesreclass dest valuesmultiplex orig with destget unique countscreate three charts for original, final and changereturn htmlpasssplit the change valuesclump and sieve8 connectedness preferred, 4 connectedness allowed Pixels of interest set to 1, 0 otherwise Number of pixels to process Label and count disconnected components (fragments) Compute fragment sizes List fragments Discard large fragments Gather small fragment information Find each fragmentoveriding, non-standard field names Preliminary teststransition table fieldsfactors table fieldsexercise fieldsget parameters, set outputsit might be better to just check if factors being usedoutput file namesabsolute pathsrelative paths, or with patterned nameconstantsvalue to multiply transition matrix entries (ie covert 10 point scale to 100 point scale)validate dataraise warning if nothing is going to happentransition table validationraise error if transition table provided, but not usedraise ValueError, msgload transition tableraise error if LULC contains cover id's not in transition tableraise error if percent change for new LULCraise error if change by percent and area both specifiedfactor parameters validationerror if overall physical weight not in [0, 1] rangefactor table validationif polygon no distance field allowedif point or line, integer distance field onlyerror if same factor twice for same coverageresample, align and rasterize datacheck geographic extents, projections validate resampling size if args["resolution"] != "": if args["resolution"] < pygeoprocessing.geoprocessing.get_cell_size_from_uri(landcover_uri): msg = "The analysis resolution cannot be smaller than the input." LOGGER.error(msg) raise ValueError, msg else: LOGGER.info("Resampling land cover.") gdal.GRA_Mode might be a better resample method, but requires GDAL >= 1.10.0 bounding_box = pygeoprocessing.geoprocessing.get_bounding_box(landcover_uri) pygeoprocessing.geoprocessing.resize_and_resample_dataset_uri(landcover_uri, bounding_box, args["resolution"], landcover_resample_uri, "nearest") LOGGER.debug("Changing landcover uri to resampled uri.") landcover_uri = landcover_resample_uriconstruct reclass dictionaryreclass lulc by reclass_dictchanging nodata value so 0's no longer nodatapolygonpoint or line For features with no area, it's (almost) impossible to hit the center pixel, so we use ALL_TOUCHED=TRUE Apply nodata to the factors raster print('------files:', uri_list, weights_list)clump and sievecompute intermediate data if neededcontraints raster (reclass using permability values, filters on clump size) Check that the values make sense print('------suitability and constraint files:', uri_list)normalize probabilities to be on a 10 point scaleprobability raster (reclass using probability matrix)proximity raster (gaussian for each landcover type, using max distance)InVEST 2 uses 4-connectedness?combine rasters for weighting into sutibility raster, multiply proximity by 0.3[suitability * (1-factor weight)] + (factors * factor weight) or only single rasterreallocate pixels (disk heap sort, randomly reassign equal value pixels, applied in order)copy initial LULCidentify LULC types undergoing changechange pixelsopen all lower priority suitability rasters and assign changed pixels value of 0select pixelsopen suitability raster Check if suitsbility is between 0 and 100 inclusivemask out everything except the current suitability scorelabel patchesget patch sizesrandomize patch ordercheck patches for conversionmask out everything except the current patchpatch = numpy.where(label_im == label)patch_mask = numpy.zeros_like(scenario_array)patch_mask[patch] = 1calculate the distance to exit the patchtmp_array = scipy.ndimage.morphology.distance_transform_edt(patch_mask)tmp_array = tmp_array[patch]select the number of pixels that need to be convertedconvert the selected pixels into coordinatespixels_to_change = numpy.array(zip(patch[0], patch[1]))change the pixels in the scenarioscenario_array[pixels_to_change] = cover_idalter other suitability rasters to prevent double conversionupdate_arrays[update_id][pixels_to_change] = 0convert patch, increase count of changesalter other suitability rasters to prevent double conversionreport and record unchanged pixelswrite new suitability arraysapply overridetabulate coveragesinput CSVs
5,961
en
0.574289
# Script that uses meshgrid to get map coordinates and then plots # the DEM in 3d. import numpy as np import matplotlib.pyplot as plt from mpl_toolkits.mplot3d import Axes3D from osgeo import gdal ds = gdal.Open(r'D:\osgeopy-data\Washington\dem\sthelens_utm.tif') band = ds.GetRasterBand(1) ov_band = band.GetOverview(band.GetOverviewCount() - 3) data = ov_band.ReadAsArray() # Calculate bounding coordinates. geotransform = ds.GetGeoTransform() minx = geotransform[0] maxy = geotransform[3] maxx = minx + ov_band.XSize * geotransform[1] miny = maxy + ov_band.YSize * geotransform[5] # Get the x and y arrays. x = np.arange(minx, maxx, geotransform[1]) y = np.arange(maxy, miny, geotransform[5]) x, y = np.meshgrid(x[:ov_band.XSize], y[:ov_band.YSize]) # Make the 3D plot. fig = plt.figure() ax = fig.gca(projection='3d') ax.plot_surface(x, y, data, cmap='gist_earth', lw=0) plt.axis('equal') # # Change the viewpoint and turn the ticks off. # ax.view_init(elev=55, azim=60) # plt.axis('off') # # Create an animation. # import matplotlib.animation as animation # def animate(i): # ax.view_init(elev=65, azim=i) # anim = animation.FuncAnimation( # fig, animate, frames=range(0, 360, 10), interval=100) # plt.axis('off') # # If you have FFmpeg and it's in your path, you can save the # # animation. # anim.save('d:/temp/helens.mp4', 'ffmpeg') plt.show()
Chapter13/listing13_7.py
1,370
Script that uses meshgrid to get map coordinates and then plots the DEM in 3d. Calculate bounding coordinates. Get the x and y arrays. Make the 3D plot. Change the viewpoint and turn the ticks off. ax.view_init(elev=55, azim=60) plt.axis('off') Create an animation. import matplotlib.animation as animation def animate(i): ax.view_init(elev=65, azim=i) anim = animation.FuncAnimation( fig, animate, frames=range(0, 360, 10), interval=100) plt.axis('off') If you have FFmpeg and it's in your path, you can save the animation. anim.save('d:/temp/helens.mp4', 'ffmpeg')
578
en
0.634263