hexsha
stringlengths
40
40
size
int64
10
805k
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
4
176
max_stars_repo_name
stringlengths
7
114
max_stars_repo_head_hexsha
stringlengths
40
40
max_stars_repo_licenses
listlengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
4
176
max_issues_repo_name
stringlengths
7
114
max_issues_repo_head_hexsha
stringlengths
40
40
max_issues_repo_licenses
listlengths
1
10
max_issues_count
int64
1
48.5k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
4
176
max_forks_repo_name
stringlengths
7
114
max_forks_repo_head_hexsha
stringlengths
40
40
max_forks_repo_licenses
listlengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
10
805k
avg_line_length
float64
5.53
11k
max_line_length
int64
10
129k
alphanum_fraction
float64
0.13
0.93
content_no_comment
stringlengths
0
449k
is_comment_constant_removed
bool
2 classes
is_sharp_comment_removed
bool
1 class
f717953473fd5d44045b712a4f139e641a8c661b
27,730
py
Python
InstallOpenface/fix_sklearn/label.py
s123600g/openfaceInstallscript
962b4b89c5626318b5701d7297d49df3423b0fe4
[ "MIT" ]
null
null
null
InstallOpenface/fix_sklearn/label.py
s123600g/openfaceInstallscript
962b4b89c5626318b5701d7297d49df3423b0fe4
[ "MIT" ]
null
null
null
InstallOpenface/fix_sklearn/label.py
s123600g/openfaceInstallscript
962b4b89c5626318b5701d7297d49df3423b0fe4
[ "MIT" ]
null
null
null
# Authors: Alexandre Gramfort <alexandre.gramfort@inria.fr> # Mathieu Blondel <mathieu@mblondel.org> # Olivier Grisel <olivier.grisel@ensta.org> # Andreas Mueller <amueller@ais.uni-bonn.de> # Joel Nothman <joel.nothman@gmail.com> # Hamzeh Alsalhi <ha258@cornell.edu> # License: BSD 3 clause from collections import defaultdict import itertools import array import numpy as np import scipy.sparse as sp from ..base import BaseEstimator, TransformerMixin from ..utils.fixes import sparse_min_max from ..utils import column_or_1d from ..utils.validation import check_array from ..utils.validation import check_is_fitted from ..utils.validation import _num_samples from ..utils.multiclass import unique_labels from ..utils.multiclass import type_of_target from ..externals import six zip = six.moves.zip map = six.moves.map __all__ = [ 'label_binarize', 'LabelBinarizer', 'LabelEncoder', 'MultiLabelBinarizer', ] class LabelEncoder(BaseEstimator, TransformerMixin): """Encode labels with value between 0 and n_classes-1. Read more in the :ref:`User Guide <preprocessing_targets>`. Attributes ---------- classes_ : array of shape (n_class,) Holds the label for each class. Examples -------- `LabelEncoder` can be used to normalize labels. >>> from sklearn import preprocessing >>> le = preprocessing.LabelEncoder() >>> le.fit([1, 2, 2, 6]) LabelEncoder() >>> le.classes_ array([1, 2, 6]) >>> le.transform([1, 1, 2, 6]) #doctest: +ELLIPSIS array([0, 0, 1, 2]...) >>> le.inverse_transform([0, 0, 1, 2]) array([1, 1, 2, 6]) It can also be used to transform non-numerical labels (as long as they are hashable and comparable) to numerical labels. >>> le = preprocessing.LabelEncoder() >>> le.fit(["paris", "paris", "tokyo", "amsterdam"]) LabelEncoder() >>> list(le.classes_) ['amsterdam', 'paris', 'tokyo'] >>> le.transform(["tokyo", "tokyo", "paris"]) #doctest: +ELLIPSIS array([2, 2, 1]...) >>> list(le.inverse_transform([2, 2, 1])) ['tokyo', 'tokyo', 'paris'] See also -------- sklearn.preprocessing.OneHotEncoder : encode categorical integer features using a one-hot aka one-of-K scheme. """ def fit(self, y): """Fit label encoder Parameters ---------- y : array-like of shape (n_samples,) Target values. Returns ------- self : returns an instance of self. """ y = column_or_1d(y, warn=True) self.classes_ = np.unique(y) return self def fit_transform(self, y): """Fit label encoder and return encoded labels Parameters ---------- y : array-like of shape [n_samples] Target values. Returns ------- y : array-like of shape [n_samples] """ y = column_or_1d(y, warn=True) self.classes_, y = np.unique(y, return_inverse=True) return y def transform(self, y): """Transform labels to normalized encoding. Parameters ---------- y : array-like of shape [n_samples] Target values. Returns ------- y : array-like of shape [n_samples] """ check_is_fitted(self, 'classes_') y = column_or_1d(y, warn=True) classes = np.unique(y) if len(np.intersect1d(classes, self.classes_)) < len(classes): diff = np.setdiff1d(classes, self.classes_) # raise ValueError("y contains new labels: %s" % str(diff)) raise ValueError("y contains previously unseen labels: % s" % str(diff)) return np.searchsorted(self.classes_, y) def inverse_transform(self, y): """Transform labels back to original encoding. Parameters ---------- y : numpy array of shape [n_samples] Target values. Returns ------- y : numpy array of shape [n_samples] """ check_is_fitted(self, 'classes_') diff = np.setdiff1d(y, np.arange(len(self.classes_))) # if diff: # raise ValueError("y contains new labels: %s" % str(diff)) if len(diff): raise ValueError("y contains previously unseen labels: %s" % str(diff)) y = np.asarray(y) return self.classes_[y] class LabelBinarizer(BaseEstimator, TransformerMixin): """Binarize labels in a one-vs-all fashion Several regression and binary classification algorithms are available in the scikit. A simple way to extend these algorithms to the multi-class classification case is to use the so-called one-vs-all scheme. At learning time, this simply consists in learning one regressor or binary classifier per class. In doing so, one needs to convert multi-class labels to binary labels (belong or does not belong to the class). LabelBinarizer makes this process easy with the transform method. At prediction time, one assigns the class for which the corresponding model gave the greatest confidence. LabelBinarizer makes this easy with the inverse_transform method. Read more in the :ref:`User Guide <preprocessing_targets>`. Parameters ---------- neg_label : int (default: 0) Value with which negative labels must be encoded. pos_label : int (default: 1) Value with which positive labels must be encoded. sparse_output : boolean (default: False) True if the returned array from transform is desired to be in sparse CSR format. Attributes ---------- classes_ : array of shape [n_class] Holds the label for each class. y_type_ : str, Represents the type of the target data as evaluated by utils.multiclass.type_of_target. Possible type are 'continuous', 'continuous-multioutput', 'binary', 'multiclass', 'multiclass-multioutput', 'multilabel-indicator', and 'unknown'. sparse_input_ : boolean, True if the input data to transform is given as a sparse matrix, False otherwise. Examples -------- >>> from sklearn import preprocessing >>> lb = preprocessing.LabelBinarizer() >>> lb.fit([1, 2, 6, 4, 2]) LabelBinarizer(neg_label=0, pos_label=1, sparse_output=False) >>> lb.classes_ array([1, 2, 4, 6]) >>> lb.transform([1, 6]) array([[1, 0, 0, 0], [0, 0, 0, 1]]) Binary targets transform to a column vector >>> lb = preprocessing.LabelBinarizer() >>> lb.fit_transform(['yes', 'no', 'no', 'yes']) array([[1], [0], [0], [1]]) Passing a 2D matrix for multilabel classification >>> import numpy as np >>> lb.fit(np.array([[0, 1, 1], [1, 0, 0]])) LabelBinarizer(neg_label=0, pos_label=1, sparse_output=False) >>> lb.classes_ array([0, 1, 2]) >>> lb.transform([0, 1, 2, 1]) array([[1, 0, 0], [0, 1, 0], [0, 0, 1], [0, 1, 0]]) See also -------- label_binarize : function to perform the transform operation of LabelBinarizer with fixed classes. sklearn.preprocessing.OneHotEncoder : encode categorical integer features using a one-hot aka one-of-K scheme. """ def __init__(self, neg_label=0, pos_label=1, sparse_output=False): if neg_label >= pos_label: raise ValueError("neg_label={0} must be strictly less than " "pos_label={1}.".format(neg_label, pos_label)) if sparse_output and (pos_label == 0 or neg_label != 0): raise ValueError("Sparse binarization is only supported with non " "zero pos_label and zero neg_label, got " "pos_label={0} and neg_label={1}" "".format(pos_label, neg_label)) self.neg_label = neg_label self.pos_label = pos_label self.sparse_output = sparse_output def fit(self, y): """Fit label binarizer Parameters ---------- y : array of shape [n_samples,] or [n_samples, n_classes] Target values. The 2-d matrix should only contain 0 and 1, represents multilabel classification. Returns ------- self : returns an instance of self. """ self.y_type_ = type_of_target(y) if 'multioutput' in self.y_type_: raise ValueError("Multioutput target data is not supported with " "label binarization") if _num_samples(y) == 0: raise ValueError('y has 0 samples: %r' % y) self.sparse_input_ = sp.issparse(y) self.classes_ = unique_labels(y) return self def fit_transform(self, y): """Fit label binarizer and transform multi-class labels to binary labels. The output of transform is sometimes referred to as the 1-of-K coding scheme. Parameters ---------- y : array or sparse matrix of shape [n_samples,] or \ [n_samples, n_classes] Target values. The 2-d matrix should only contain 0 and 1, represents multilabel classification. Sparse matrix can be CSR, CSC, COO, DOK, or LIL. Returns ------- Y : array or CSR matrix of shape [n_samples, n_classes] Shape will be [n_samples, 1] for binary problems. """ return self.fit(y).transform(y) def transform(self, y): """Transform multi-class labels to binary labels The output of transform is sometimes referred to by some authors as the 1-of-K coding scheme. Parameters ---------- y : array or sparse matrix of shape [n_samples,] or \ [n_samples, n_classes] Target values. The 2-d matrix should only contain 0 and 1, represents multilabel classification. Sparse matrix can be CSR, CSC, COO, DOK, or LIL. Returns ------- Y : numpy array or CSR matrix of shape [n_samples, n_classes] Shape will be [n_samples, 1] for binary problems. """ check_is_fitted(self, 'classes_') y_is_multilabel = type_of_target(y).startswith('multilabel') if y_is_multilabel and not self.y_type_.startswith('multilabel'): raise ValueError("The object was not fitted with multilabel" " input.") return label_binarize(y, self.classes_, pos_label=self.pos_label, neg_label=self.neg_label, sparse_output=self.sparse_output) def inverse_transform(self, Y, threshold=None): """Transform binary labels back to multi-class labels Parameters ---------- Y : numpy array or sparse matrix with shape [n_samples, n_classes] Target values. All sparse matrices are converted to CSR before inverse transformation. threshold : float or None Threshold used in the binary and multi-label cases. Use 0 when ``Y`` contains the output of decision_function (classifier). Use 0.5 when ``Y`` contains the output of predict_proba. If None, the threshold is assumed to be half way between neg_label and pos_label. Returns ------- y : numpy array or CSR matrix of shape [n_samples] Target values. Notes ----- In the case when the binary labels are fractional (probabilistic), inverse_transform chooses the class with the greatest value. Typically, this allows to use the output of a linear model's decision_function method directly as the input of inverse_transform. """ check_is_fitted(self, 'classes_') if threshold is None: threshold = (self.pos_label + self.neg_label) / 2. if self.y_type_ == "multiclass": y_inv = _inverse_binarize_multiclass(Y, self.classes_) else: y_inv = _inverse_binarize_thresholding(Y, self.y_type_, self.classes_, threshold) if self.sparse_input_: y_inv = sp.csr_matrix(y_inv) elif sp.issparse(y_inv): y_inv = y_inv.toarray() return y_inv def label_binarize(y, classes, neg_label=0, pos_label=1, sparse_output=False): """Binarize labels in a one-vs-all fashion Several regression and binary classification algorithms are available in the scikit. A simple way to extend these algorithms to the multi-class classification case is to use the so-called one-vs-all scheme. This function makes it possible to compute this transformation for a fixed set of class labels known ahead of time. Parameters ---------- y : array-like Sequence of integer labels or multilabel data to encode. classes : array-like of shape [n_classes] Uniquely holds the label for each class. neg_label : int (default: 0) Value with which negative labels must be encoded. pos_label : int (default: 1) Value with which positive labels must be encoded. sparse_output : boolean (default: False), Set to true if output binary array is desired in CSR sparse format Returns ------- Y : numpy array or CSR matrix of shape [n_samples, n_classes] Shape will be [n_samples, 1] for binary problems. Examples -------- >>> from sklearn.preprocessing import label_binarize >>> label_binarize([1, 6], classes=[1, 2, 4, 6]) array([[1, 0, 0, 0], [0, 0, 0, 1]]) The class ordering is preserved: >>> label_binarize([1, 6], classes=[1, 6, 4, 2]) array([[1, 0, 0, 0], [0, 1, 0, 0]]) Binary targets transform to a column vector >>> label_binarize(['yes', 'no', 'no', 'yes'], classes=['no', 'yes']) array([[1], [0], [0], [1]]) See also -------- LabelBinarizer : class used to wrap the functionality of label_binarize and allow for fitting to classes independently of the transform operation """ if not isinstance(y, list): # XXX Workaround that will be removed when list of list format is # dropped y = check_array(y, accept_sparse='csr', ensure_2d=False, dtype=None) else: if _num_samples(y) == 0: raise ValueError('y has 0 samples: %r' % y) if neg_label >= pos_label: raise ValueError("neg_label={0} must be strictly less than " "pos_label={1}.".format(neg_label, pos_label)) if (sparse_output and (pos_label == 0 or neg_label != 0)): raise ValueError("Sparse binarization is only supported with non " "zero pos_label and zero neg_label, got " "pos_label={0} and neg_label={1}" "".format(pos_label, neg_label)) # To account for pos_label == 0 in the dense case pos_switch = pos_label == 0 if pos_switch: pos_label = -neg_label y_type = type_of_target(y) if 'multioutput' in y_type: raise ValueError("Multioutput target data is not supported with label " "binarization") if y_type == 'unknown': raise ValueError("The type of target data is not known") n_samples = y.shape[0] if sp.issparse(y) else len(y) n_classes = len(classes) classes = np.asarray(classes) if y_type == "binary": if n_classes == 1: if sparse_output: return sp.csr_matrix((n_samples, 1), dtype=int) else: Y = np.zeros((len(y), 1), dtype=np.int) Y += neg_label return Y elif len(classes) >= 3: y_type = "multiclass" sorted_class = np.sort(classes) if (y_type == "multilabel-indicator" and classes.size != y.shape[1]): raise ValueError("classes {0} missmatch with the labels {1}" "found in the data".format(classes, unique_labels(y))) if y_type in ("binary", "multiclass"): y = column_or_1d(y) # pick out the known labels from y y_in_classes = np.in1d(y, classes) y_seen = y[y_in_classes] indices = np.searchsorted(sorted_class, y_seen) indptr = np.hstack((0, np.cumsum(y_in_classes))) data = np.empty_like(indices) data.fill(pos_label) Y = sp.csr_matrix((data, indices, indptr), shape=(n_samples, n_classes)) elif y_type == "multilabel-indicator": Y = sp.csr_matrix(y) if pos_label != 1: data = np.empty_like(Y.data) data.fill(pos_label) Y.data = data else: raise ValueError("%s target data is not supported with label " "binarization" % y_type) if not sparse_output: Y = Y.toarray() Y = Y.astype(int, copy=False) if neg_label != 0: Y[Y == 0] = neg_label if pos_switch: Y[Y == pos_label] = 0 else: Y.data = Y.data.astype(int, copy=False) # preserve label ordering if np.any(classes != sorted_class): indices = np.searchsorted(sorted_class, classes) Y = Y[:, indices] if y_type == "binary": if sparse_output: Y = Y.getcol(-1) else: Y = Y[:, -1].reshape((-1, 1)) return Y def _inverse_binarize_multiclass(y, classes): """Inverse label binarization transformation for multiclass. Multiclass uses the maximal score instead of a threshold. """ classes = np.asarray(classes) if sp.issparse(y): # Find the argmax for each row in y where y is a CSR matrix y = y.tocsr() n_samples, n_outputs = y.shape outputs = np.arange(n_outputs) row_max = sparse_min_max(y, 1)[1] row_nnz = np.diff(y.indptr) y_data_repeated_max = np.repeat(row_max, row_nnz) # picks out all indices obtaining the maximum per row y_i_all_argmax = np.flatnonzero(y_data_repeated_max == y.data) # For corner case where last row has a max of 0 if row_max[-1] == 0: y_i_all_argmax = np.append(y_i_all_argmax, [len(y.data)]) # Gets the index of the first argmax in each row from y_i_all_argmax index_first_argmax = np.searchsorted(y_i_all_argmax, y.indptr[:-1]) # first argmax of each row y_ind_ext = np.append(y.indices, [0]) y_i_argmax = y_ind_ext[y_i_all_argmax[index_first_argmax]] # Handle rows of all 0 y_i_argmax[np.where(row_nnz == 0)[0]] = 0 # Handles rows with max of 0 that contain negative numbers samples = np.arange(n_samples)[(row_nnz > 0) & (row_max.ravel() == 0)] for i in samples: ind = y.indices[y.indptr[i]:y.indptr[i + 1]] y_i_argmax[i] = classes[np.setdiff1d(outputs, ind)][0] return classes[y_i_argmax] else: return classes.take(y.argmax(axis=1), mode="clip") def _inverse_binarize_thresholding(y, output_type, classes, threshold): """Inverse label binarization transformation using thresholding.""" if output_type == "binary" and y.ndim == 2 and y.shape[1] > 2: raise ValueError("output_type='binary', but y.shape = {0}". format(y.shape)) if output_type != "binary" and y.shape[1] != len(classes): raise ValueError("The number of class is not equal to the number of " "dimension of y.") classes = np.asarray(classes) # Perform thresholding if sp.issparse(y): if threshold > 0: if y.format not in ('csr', 'csc'): y = y.tocsr() y.data = np.array(y.data > threshold, dtype=np.int) y.eliminate_zeros() else: y = np.array(y.toarray() > threshold, dtype=np.int) else: y = np.array(y > threshold, dtype=np.int) # Inverse transform data if output_type == "binary": if sp.issparse(y): y = y.toarray() if y.ndim == 2 and y.shape[1] == 2: return classes[y[:, 1]] else: if len(classes) == 1: return np.repeat(classes[0], len(y)) else: return classes[y.ravel()] elif output_type == "multilabel-indicator": return y else: raise ValueError("{0} format is not supported".format(output_type)) class MultiLabelBinarizer(BaseEstimator, TransformerMixin): """Transform between iterable of iterables and a multilabel format Although a list of sets or tuples is a very intuitive format for multilabel data, it is unwieldy to process. This transformer converts between this intuitive format and the supported multilabel format: a (samples x classes) binary matrix indicating the presence of a class label. Parameters ---------- classes : array-like of shape [n_classes] (optional) Indicates an ordering for the class labels sparse_output : boolean (default: False), Set to true if output binary array is desired in CSR sparse format Attributes ---------- classes_ : array of labels A copy of the `classes` parameter where provided, or otherwise, the sorted set of classes found when fitting. Examples -------- >>> from sklearn.preprocessing import MultiLabelBinarizer >>> mlb = MultiLabelBinarizer() >>> mlb.fit_transform([(1, 2), (3,)]) array([[1, 1, 0], [0, 0, 1]]) >>> mlb.classes_ array([1, 2, 3]) >>> mlb.fit_transform([set(['sci-fi', 'thriller']), set(['comedy'])]) array([[0, 1, 1], [1, 0, 0]]) >>> list(mlb.classes_) ['comedy', 'sci-fi', 'thriller'] See also -------- sklearn.preprocessing.OneHotEncoder : encode categorical integer features using a one-hot aka one-of-K scheme. """ def __init__(self, classes=None, sparse_output=False): self.classes = classes self.sparse_output = sparse_output def fit(self, y): """Fit the label sets binarizer, storing `classes_` Parameters ---------- y : iterable of iterables A set of labels (any orderable and hashable object) for each sample. If the `classes` parameter is set, `y` will not be iterated. Returns ------- self : returns this MultiLabelBinarizer instance """ if self.classes is None: classes = sorted(set(itertools.chain.from_iterable(y))) else: classes = self.classes dtype = np.int if all(isinstance(c, int) for c in classes) else object self.classes_ = np.empty(len(classes), dtype=dtype) self.classes_[:] = classes return self def fit_transform(self, y): """Fit the label sets binarizer and transform the given label sets Parameters ---------- y : iterable of iterables A set of labels (any orderable and hashable object) for each sample. If the `classes` parameter is set, `y` will not be iterated. Returns ------- y_indicator : array or CSR matrix, shape (n_samples, n_classes) A matrix such that `y_indicator[i, j] = 1` iff `classes_[j]` is in `y[i]`, and 0 otherwise. """ if self.classes is not None: return self.fit(y).transform(y) # Automatically increment on new class class_mapping = defaultdict(int) class_mapping.default_factory = class_mapping.__len__ yt = self._transform(y, class_mapping) # sort classes and reorder columns tmp = sorted(class_mapping, key=class_mapping.get) # (make safe for tuples) dtype = np.int if all(isinstance(c, int) for c in tmp) else object class_mapping = np.empty(len(tmp), dtype=dtype) class_mapping[:] = tmp self.classes_, inverse = np.unique(class_mapping, return_inverse=True) # ensure yt.indices keeps its current dtype yt.indices = np.array(inverse[yt.indices], dtype=yt.indices.dtype, copy=False) if not self.sparse_output: yt = yt.toarray() return yt def transform(self, y): """Transform the given label sets Parameters ---------- y : iterable of iterables A set of labels (any orderable and hashable object) for each sample. If the `classes` parameter is set, `y` will not be iterated. Returns ------- y_indicator : array or CSR matrix, shape (n_samples, n_classes) A matrix such that `y_indicator[i, j] = 1` iff `classes_[j]` is in `y[i]`, and 0 otherwise. """ check_is_fitted(self, 'classes_') class_to_index = dict(zip(self.classes_, range(len(self.classes_)))) yt = self._transform(y, class_to_index) if not self.sparse_output: yt = yt.toarray() return yt def _transform(self, y, class_mapping): """Transforms the label sets with a given mapping Parameters ---------- y : iterable of iterables class_mapping : Mapping Maps from label to column index in label indicator matrix Returns ------- y_indicator : sparse CSR matrix, shape (n_samples, n_classes) Label indicator matrix """ indices = array.array('i') indptr = array.array('i', [0]) for labels in y: indices.extend(set(class_mapping[label] for label in labels)) indptr.append(len(indices)) data = np.ones(len(indices), dtype=int) return sp.csr_matrix((data, indices, indptr), shape=(len(indptr) - 1, len(class_mapping))) def inverse_transform(self, yt): """Transform the given indicator matrix into label sets Parameters ---------- yt : array or sparse matrix of shape (n_samples, n_classes) A matrix containing only 1s ands 0s. Returns ------- y : list of tuples The set of labels for each sample such that `y[i]` consists of `classes_[j]` for each `yt[i, j] == 1`. """ check_is_fitted(self, 'classes_') if yt.shape[1] != len(self.classes_): raise ValueError('Expected indicator for {0} classes, but got {1}' .format(len(self.classes_), yt.shape[1])) if sp.issparse(yt): yt = yt.tocsr() if len(yt.data) != 0 and len(np.setdiff1d(yt.data, [0, 1])) > 0: raise ValueError('Expected only 0s and 1s in label indicator.') return [tuple(self.classes_.take(yt.indices[start:end])) for start, end in zip(yt.indptr[:-1], yt.indptr[1:])] else: unexpected = np.setdiff1d(yt, [0, 1]) if len(unexpected) > 0: raise ValueError('Expected only 0s and 1s in label indicator. ' 'Also got {0}'.format(unexpected)) return [tuple(self.classes_.compress(indicators)) for indicators in yt]
33.329327
87
0.588136
from collections import defaultdict import itertools import array import numpy as np import scipy.sparse as sp from ..base import BaseEstimator, TransformerMixin from ..utils.fixes import sparse_min_max from ..utils import column_or_1d from ..utils.validation import check_array from ..utils.validation import check_is_fitted from ..utils.validation import _num_samples from ..utils.multiclass import unique_labels from ..utils.multiclass import type_of_target from ..externals import six zip = six.moves.zip map = six.moves.map __all__ = [ 'label_binarize', 'LabelBinarizer', 'LabelEncoder', 'MultiLabelBinarizer', ] class LabelEncoder(BaseEstimator, TransformerMixin): def fit(self, y): y = column_or_1d(y, warn=True) self.classes_ = np.unique(y) return self def fit_transform(self, y): y = column_or_1d(y, warn=True) self.classes_, y = np.unique(y, return_inverse=True) return y def transform(self, y): check_is_fitted(self, 'classes_') y = column_or_1d(y, warn=True) classes = np.unique(y) if len(np.intersect1d(classes, self.classes_)) < len(classes): diff = np.setdiff1d(classes, self.classes_) raise ValueError("y contains previously unseen labels: % s" % str(diff)) return np.searchsorted(self.classes_, y) def inverse_transform(self, y): check_is_fitted(self, 'classes_') diff = np.setdiff1d(y, np.arange(len(self.classes_))) if len(diff): raise ValueError("y contains previously unseen labels: %s" % str(diff)) y = np.asarray(y) return self.classes_[y] class LabelBinarizer(BaseEstimator, TransformerMixin): def __init__(self, neg_label=0, pos_label=1, sparse_output=False): if neg_label >= pos_label: raise ValueError("neg_label={0} must be strictly less than " "pos_label={1}.".format(neg_label, pos_label)) if sparse_output and (pos_label == 0 or neg_label != 0): raise ValueError("Sparse binarization is only supported with non " "zero pos_label and zero neg_label, got " "pos_label={0} and neg_label={1}" "".format(pos_label, neg_label)) self.neg_label = neg_label self.pos_label = pos_label self.sparse_output = sparse_output def fit(self, y): self.y_type_ = type_of_target(y) if 'multioutput' in self.y_type_: raise ValueError("Multioutput target data is not supported with " "label binarization") if _num_samples(y) == 0: raise ValueError('y has 0 samples: %r' % y) self.sparse_input_ = sp.issparse(y) self.classes_ = unique_labels(y) return self def fit_transform(self, y): return self.fit(y).transform(y) def transform(self, y): check_is_fitted(self, 'classes_') y_is_multilabel = type_of_target(y).startswith('multilabel') if y_is_multilabel and not self.y_type_.startswith('multilabel'): raise ValueError("The object was not fitted with multilabel" " input.") return label_binarize(y, self.classes_, pos_label=self.pos_label, neg_label=self.neg_label, sparse_output=self.sparse_output) def inverse_transform(self, Y, threshold=None): check_is_fitted(self, 'classes_') if threshold is None: threshold = (self.pos_label + self.neg_label) / 2. if self.y_type_ == "multiclass": y_inv = _inverse_binarize_multiclass(Y, self.classes_) else: y_inv = _inverse_binarize_thresholding(Y, self.y_type_, self.classes_, threshold) if self.sparse_input_: y_inv = sp.csr_matrix(y_inv) elif sp.issparse(y_inv): y_inv = y_inv.toarray() return y_inv def label_binarize(y, classes, neg_label=0, pos_label=1, sparse_output=False): if not isinstance(y, list): y = check_array(y, accept_sparse='csr', ensure_2d=False, dtype=None) else: if _num_samples(y) == 0: raise ValueError('y has 0 samples: %r' % y) if neg_label >= pos_label: raise ValueError("neg_label={0} must be strictly less than " "pos_label={1}.".format(neg_label, pos_label)) if (sparse_output and (pos_label == 0 or neg_label != 0)): raise ValueError("Sparse binarization is only supported with non " "zero pos_label and zero neg_label, got " "pos_label={0} and neg_label={1}" "".format(pos_label, neg_label)) pos_switch = pos_label == 0 if pos_switch: pos_label = -neg_label y_type = type_of_target(y) if 'multioutput' in y_type: raise ValueError("Multioutput target data is not supported with label " "binarization") if y_type == 'unknown': raise ValueError("The type of target data is not known") n_samples = y.shape[0] if sp.issparse(y) else len(y) n_classes = len(classes) classes = np.asarray(classes) if y_type == "binary": if n_classes == 1: if sparse_output: return sp.csr_matrix((n_samples, 1), dtype=int) else: Y = np.zeros((len(y), 1), dtype=np.int) Y += neg_label return Y elif len(classes) >= 3: y_type = "multiclass" sorted_class = np.sort(classes) if (y_type == "multilabel-indicator" and classes.size != y.shape[1]): raise ValueError("classes {0} missmatch with the labels {1}" "found in the data".format(classes, unique_labels(y))) if y_type in ("binary", "multiclass"): y = column_or_1d(y) y_in_classes = np.in1d(y, classes) y_seen = y[y_in_classes] indices = np.searchsorted(sorted_class, y_seen) indptr = np.hstack((0, np.cumsum(y_in_classes))) data = np.empty_like(indices) data.fill(pos_label) Y = sp.csr_matrix((data, indices, indptr), shape=(n_samples, n_classes)) elif y_type == "multilabel-indicator": Y = sp.csr_matrix(y) if pos_label != 1: data = np.empty_like(Y.data) data.fill(pos_label) Y.data = data else: raise ValueError("%s target data is not supported with label " "binarization" % y_type) if not sparse_output: Y = Y.toarray() Y = Y.astype(int, copy=False) if neg_label != 0: Y[Y == 0] = neg_label if pos_switch: Y[Y == pos_label] = 0 else: Y.data = Y.data.astype(int, copy=False) if np.any(classes != sorted_class): indices = np.searchsorted(sorted_class, classes) Y = Y[:, indices] if y_type == "binary": if sparse_output: Y = Y.getcol(-1) else: Y = Y[:, -1].reshape((-1, 1)) return Y def _inverse_binarize_multiclass(y, classes): classes = np.asarray(classes) if sp.issparse(y): y = y.tocsr() n_samples, n_outputs = y.shape outputs = np.arange(n_outputs) row_max = sparse_min_max(y, 1)[1] row_nnz = np.diff(y.indptr) y_data_repeated_max = np.repeat(row_max, row_nnz) y_i_all_argmax = np.flatnonzero(y_data_repeated_max == y.data) if row_max[-1] == 0: y_i_all_argmax = np.append(y_i_all_argmax, [len(y.data)]) index_first_argmax = np.searchsorted(y_i_all_argmax, y.indptr[:-1]) y_ind_ext = np.append(y.indices, [0]) y_i_argmax = y_ind_ext[y_i_all_argmax[index_first_argmax]] y_i_argmax[np.where(row_nnz == 0)[0]] = 0 samples = np.arange(n_samples)[(row_nnz > 0) & (row_max.ravel() == 0)] for i in samples: ind = y.indices[y.indptr[i]:y.indptr[i + 1]] y_i_argmax[i] = classes[np.setdiff1d(outputs, ind)][0] return classes[y_i_argmax] else: return classes.take(y.argmax(axis=1), mode="clip") def _inverse_binarize_thresholding(y, output_type, classes, threshold): if output_type == "binary" and y.ndim == 2 and y.shape[1] > 2: raise ValueError("output_type='binary', but y.shape = {0}". format(y.shape)) if output_type != "binary" and y.shape[1] != len(classes): raise ValueError("The number of class is not equal to the number of " "dimension of y.") classes = np.asarray(classes) if sp.issparse(y): if threshold > 0: if y.format not in ('csr', 'csc'): y = y.tocsr() y.data = np.array(y.data > threshold, dtype=np.int) y.eliminate_zeros() else: y = np.array(y.toarray() > threshold, dtype=np.int) else: y = np.array(y > threshold, dtype=np.int) if output_type == "binary": if sp.issparse(y): y = y.toarray() if y.ndim == 2 and y.shape[1] == 2: return classes[y[:, 1]] else: if len(classes) == 1: return np.repeat(classes[0], len(y)) else: return classes[y.ravel()] elif output_type == "multilabel-indicator": return y else: raise ValueError("{0} format is not supported".format(output_type)) class MultiLabelBinarizer(BaseEstimator, TransformerMixin): def __init__(self, classes=None, sparse_output=False): self.classes = classes self.sparse_output = sparse_output def fit(self, y): if self.classes is None: classes = sorted(set(itertools.chain.from_iterable(y))) else: classes = self.classes dtype = np.int if all(isinstance(c, int) for c in classes) else object self.classes_ = np.empty(len(classes), dtype=dtype) self.classes_[:] = classes return self def fit_transform(self, y): if self.classes is not None: return self.fit(y).transform(y) class_mapping = defaultdict(int) class_mapping.default_factory = class_mapping.__len__ yt = self._transform(y, class_mapping) tmp = sorted(class_mapping, key=class_mapping.get) dtype = np.int if all(isinstance(c, int) for c in tmp) else object class_mapping = np.empty(len(tmp), dtype=dtype) class_mapping[:] = tmp self.classes_, inverse = np.unique(class_mapping, return_inverse=True) yt.indices = np.array(inverse[yt.indices], dtype=yt.indices.dtype, copy=False) if not self.sparse_output: yt = yt.toarray() return yt def transform(self, y): check_is_fitted(self, 'classes_') class_to_index = dict(zip(self.classes_, range(len(self.classes_)))) yt = self._transform(y, class_to_index) if not self.sparse_output: yt = yt.toarray() return yt def _transform(self, y, class_mapping): indices = array.array('i') indptr = array.array('i', [0]) for labels in y: indices.extend(set(class_mapping[label] for label in labels)) indptr.append(len(indices)) data = np.ones(len(indices), dtype=int) return sp.csr_matrix((data, indices, indptr), shape=(len(indptr) - 1, len(class_mapping))) def inverse_transform(self, yt): check_is_fitted(self, 'classes_') if yt.shape[1] != len(self.classes_): raise ValueError('Expected indicator for {0} classes, but got {1}' .format(len(self.classes_), yt.shape[1])) if sp.issparse(yt): yt = yt.tocsr() if len(yt.data) != 0 and len(np.setdiff1d(yt.data, [0, 1])) > 0: raise ValueError('Expected only 0s and 1s in label indicator.') return [tuple(self.classes_.take(yt.indices[start:end])) for start, end in zip(yt.indptr[:-1], yt.indptr[1:])] else: unexpected = np.setdiff1d(yt, [0, 1]) if len(unexpected) > 0: raise ValueError('Expected only 0s and 1s in label indicator. ' 'Also got {0}'.format(unexpected)) return [tuple(self.classes_.compress(indicators)) for indicators in yt]
true
true
f71795db955b211e36055a00f084863a0bec4fd7
2,307
py
Python
third_party/webrtc/src/chromium/src/tools/perf/metrics/network.py
bopopescu/webrtc-streaming-node
727a441204344ff596401b0253caac372b714d91
[ "MIT" ]
20
2015-08-26T06:46:00.000Z
2019-02-27T09:05:58.000Z
third_party/webrtc/src/chromium/src/tools/perf/metrics/network.py
bopopescu/webrtc-streaming-node
727a441204344ff596401b0253caac372b714d91
[ "MIT" ]
1
2021-05-05T11:11:31.000Z
2021-05-05T11:11:31.000Z
third_party/webrtc/src/chromium/src/tools/perf/metrics/network.py
bopopescu/webrtc-streaming-node
727a441204344ff596401b0253caac372b714d91
[ "MIT" ]
7
2016-02-09T09:28:14.000Z
2020-07-25T19:03:36.000Z
# Copyright 2015 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from telemetry.value import scalar from metrics import Metric NETWORK_DATA_NOT_FOUND = 'Network data could not be found.' # This is experimental. crbug.com/480512 # Will not be supported once network data is ported to TimelineBasedMetric. class NetworkMetric(Metric): """NetworkMetrics gathers network statistics.""" def __init__(self, platform): super(NetworkMetric, self).__init__() self._network_snd = None self._network_rcv = None self._platform = platform self._browser = None def Start(self, _, tab): """Start the per-page preparation for this metric. Here, this consists of recording the start value. """ self._browser = tab.browser if not self._platform.CanMonitorNetworkData(): return data = self._platform.GetNetworkData(self._browser) if data is not None: self._network_snd, self._network_rcv = data def Stop(self, _, tab): """Prepare the results for this page. The results are the differences between the current values and the values when Start() was called. """ if not self._platform.CanMonitorNetworkData(): return data = self._platform.GetNetworkData(self._browser) if data is not None: snd, rcv = data if self._network_snd is not None: self._network_snd = snd - self._network_snd if self._network_rcv is not None: self._network_rcv = rcv - self._network_rcv else: # If end data cannot be found, report none. self._network_snd = None self._network_rcv = None def AddResults(self, tab, results): none_value_reason = ( None if self._network_snd is not None else NETWORK_DATA_NOT_FOUND) results.AddValue(scalar.ScalarValue( results.current_page, 'network_data_sent', 'kb', self._network_snd, important=False, none_value_reason=none_value_reason)) none_value_reason = ( None if self._network_rcv is not None else NETWORK_DATA_NOT_FOUND) results.AddValue(scalar.ScalarValue( results.current_page, 'network_data_received', 'kb', self._network_rcv, important=False, none_value_reason=none_value_reason))
33.434783
79
0.715648
from telemetry.value import scalar from metrics import Metric NETWORK_DATA_NOT_FOUND = 'Network data could not be found.' class NetworkMetric(Metric): def __init__(self, platform): super(NetworkMetric, self).__init__() self._network_snd = None self._network_rcv = None self._platform = platform self._browser = None def Start(self, _, tab): self._browser = tab.browser if not self._platform.CanMonitorNetworkData(): return data = self._platform.GetNetworkData(self._browser) if data is not None: self._network_snd, self._network_rcv = data def Stop(self, _, tab): if not self._platform.CanMonitorNetworkData(): return data = self._platform.GetNetworkData(self._browser) if data is not None: snd, rcv = data if self._network_snd is not None: self._network_snd = snd - self._network_snd if self._network_rcv is not None: self._network_rcv = rcv - self._network_rcv else: self._network_snd = None self._network_rcv = None def AddResults(self, tab, results): none_value_reason = ( None if self._network_snd is not None else NETWORK_DATA_NOT_FOUND) results.AddValue(scalar.ScalarValue( results.current_page, 'network_data_sent', 'kb', self._network_snd, important=False, none_value_reason=none_value_reason)) none_value_reason = ( None if self._network_rcv is not None else NETWORK_DATA_NOT_FOUND) results.AddValue(scalar.ScalarValue( results.current_page, 'network_data_received', 'kb', self._network_rcv, important=False, none_value_reason=none_value_reason))
true
true
f71796820dbd5571d7b35348bcdbcfdd286b0af9
13,935
py
Python
utils.py
ravana69/MusicPlayer-1
70feb663579c6be63b0e4481bdfd6b41ffe51110
[ "MIT" ]
1
2021-07-27T04:27:13.000Z
2021-07-27T04:27:13.000Z
utils.py
ravana69/MusicPlayer-1
70feb663579c6be63b0e4481bdfd6b41ffe51110
[ "MIT" ]
null
null
null
utils.py
ravana69/MusicPlayer-1
70feb663579c6be63b0e4481bdfd6b41ffe51110
[ "MIT" ]
1
2021-09-17T03:43:21.000Z
2021-09-17T03:43:21.000Z
#MIT License #Copyright (c) 2021 SUBIN #Permission is hereby granted, free of charge, to any person obtaining a copy #of this software and associated documentation files (the "Software"), to deal #in the Software without restriction, including without limitation the rights #to use, copy, modify, merge, publish, distribute, sublicense, and/or sell #copies of the Software, and to permit persons to whom the Software is #furnished to do so, subject to the following conditions: #The above copyright notice and this permission notice shall be included in all #copies or substantial portions of the Software. #THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR #IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, #FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE #AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER #LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, #OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE #SOFTWARE. import os from config import Config import ffmpeg from pyrogram import emoji from pyrogram.methods.messages.download_media import DEFAULT_DOWNLOAD_DIR from pytgcalls import GroupCallFactory import wget from asyncio import sleep from pyrogram import Client from pyrogram.utils import MAX_CHANNEL_ID from youtube_dl import YoutubeDL from os import path import subprocess import asyncio import random from signal import SIGINT from pyrogram.raw.types import InputGroupCall from pyrogram.raw.functions.phone import EditGroupCallTitle, CreateGroupCall from random import randint bot = Client( "Musicplayervc", Config.API_ID, Config.API_HASH, bot_token=Config.BOT_TOKEN ) bot.start() e=bot.get_me() USERNAME=e.username from user import USER CHAT=Config.CHAT FFMPEG_PROCESSES = {} ADMIN_LIST={} CALL_STATUS={} EDIT_TITLE=Config.EDIT_TITLE RADIO={6} LOG_GROUP=Config.LOG_GROUP DURATION_LIMIT=Config.DURATION_LIMIT DELAY=Config.DELAY playlist=Config.playlist msg=Config.msg SHUFFLE=Config.SHUFFLE LIMIT=Config.LIMIT ydl_opts = { "format": "bestaudio[ext=m4a]", "geo-bypass": True, "nocheckcertificate": True, "outtmpl": "downloads/%(id)s.%(ext)s", } ydl = YoutubeDL(ydl_opts) RADIO_TITLE=os.environ.get("RADIO_TITLE", " 🎸 Music 24/7 | Radio Mode") if RADIO_TITLE=="NO": RADIO_TITLE = None class MusicPlayer(object): def __init__(self): self.group_call = GroupCallFactory(USER, GroupCallFactory.MTPROTO_CLIENT_TYPE.PYROGRAM).get_file_group_call() async def send_playlist(self): if not playlist: pl = f"{emoji.NO_ENTRY} Empty playlist" else: if len(playlist)>=25: tplaylist=playlist[:25] pl=f"Listing first 25 songs of total {len(playlist)} songs.\n" pl += f"{emoji.PLAY_BUTTON} **Playlist**:\n" + "\n".join([ f"**{i}**. **🎸{x[1]}**\n 👤**Requested by:** {x[4]}" for i, x in enumerate(tplaylist) ]) else: pl = f"{emoji.PLAY_BUTTON} **Playlist**:\n" + "\n".join([ f"**{i}**. **🎸{x[1]}**\n 👤**Requested by:** {x[4]}\n" for i, x in enumerate(playlist) ]) if msg.get('playlist') is not None: await msg['playlist'].delete() msg['playlist'] = await self.send_text(pl) async def skip_current_playing(self): group_call = self.group_call if not playlist: return if len(playlist) == 1: await mp.start_radio() return client = group_call.client download_dir = os.path.join(client.workdir, DEFAULT_DOWNLOAD_DIR) group_call.input_filename = os.path.join( download_dir, f"{playlist[1][1]}.raw" ) # remove old track from playlist old_track = playlist.pop(0) print(f"- START PLAYING: {playlist[0][1]}") if EDIT_TITLE: await self.edit_title() if LOG_GROUP: await self.send_playlist() os.remove(os.path.join( download_dir, f"{old_track[1]}.raw") ) if len(playlist) == 1: return await self.download_audio(playlist[1]) async def send_text(self, text): group_call = self.group_call client = group_call.client chat_id = LOG_GROUP message = await bot.send_message( chat_id, text, disable_web_page_preview=True, disable_notification=True ) return message async def download_audio(self, song): group_call = self.group_call client = group_call.client raw_file = os.path.join(client.workdir, DEFAULT_DOWNLOAD_DIR, f"{song[1]}.raw") #if os.path.exists(raw_file): #os.remove(raw_file) if not os.path.isfile(raw_file): # credits: https://t.me/c/1480232458/6825 #os.mkfifo(raw_file) if song[3] == "telegram": original_file = await bot.download_media(f"{song[2]}") elif song[3] == "youtube": url=song[2] try: info = ydl.extract_info(url, False) ydl.download([url]) original_file=path.join("downloads", f"{info['id']}.{info['ext']}") except Exception as e: playlist.pop(1) print(f"Unable to download due to {e} and skipped.") if len(playlist) == 1: return await self.download_audio(playlist[1]) return else: original_file=wget.download(song[2]) ffmpeg.input(original_file).output( raw_file, format='s16le', acodec='pcm_s16le', ac=2, ar='48k', loglevel='error' ).overwrite_output().run() os.remove(original_file) async def start_radio(self): group_call = self.group_call if group_call.is_connected: playlist.clear() process = FFMPEG_PROCESSES.get(CHAT) if process: try: process.send_signal(SIGINT) except subprocess.TimeoutExpired: process.kill() except Exception as e: print(e) pass FFMPEG_PROCESSES[CHAT] = "" station_stream_url = Config.STREAM_URL try: RADIO.remove(0) except: pass try: RADIO.add(1) except: pass if Config.CPLAY: await self.c_play(Config.STREAM_URL) return try: RADIO.remove(3) except: pass if os.path.exists(f'radio-{CHAT}.raw'): os.remove(f'radio-{CHAT}.raw') # credits: https://t.me/c/1480232458/6825 #os.mkfifo(f'radio-{CHAT}.raw') if not CALL_STATUS.get(CHAT): await self.start_call() ffmpeg_log = open("ffmpeg.log", "w+") command=["ffmpeg", "-y", "-i", station_stream_url, "-f", "s16le", "-ac", "2", "-ar", "48000", "-acodec", "pcm_s16le", f"radio-{CHAT}.raw"] process = await asyncio.create_subprocess_exec( *command, stdout=ffmpeg_log, stderr=asyncio.subprocess.STDOUT, ) FFMPEG_PROCESSES[CHAT] = process if RADIO_TITLE: await self.edit_title() await sleep(2) while not os.path.isfile(f'radio-{CHAT}.raw'): await sleep(1) group_call.input_filename = f'radio-{CHAT}.raw' while True: if CALL_STATUS.get(CHAT): print("Succesfully Joined") break else: print("Connecting...") await self.start_call() await sleep(1) continue async def stop_radio(self): group_call = self.group_call if group_call: playlist.clear() group_call.input_filename = '' try: RADIO.remove(1) except: pass try: RADIO.add(0) except: pass process = FFMPEG_PROCESSES.get(CHAT) if process: try: process.send_signal(SIGINT) except subprocess.TimeoutExpired: process.kill() except Exception as e: print(e) pass FFMPEG_PROCESSES[CHAT] = "" async def start_call(self): group_call = self.group_call try: await group_call.start(CHAT) except RuntimeError: await USER.send(CreateGroupCall( peer=(await USER.resolve_peer(CHAT)), random_id=randint(10000, 999999999) ) ) await group_call.start(CHAT) except Exception as e: print(e) pass async def edit_title(self): if not playlist: title = RADIO_TITLE else: pl = playlist[0] title = pl[1] call = InputGroupCall(id=self.group_call.group_call.id, access_hash=self.group_call.group_call.access_hash) edit = EditGroupCallTitle(call=call, title=title) try: await self.group_call.client.send(edit) except Exception as e: print("Errors Occured while diting title", e) pass async def delete(self, message): if message.chat.type == "supergroup": await sleep(DELAY) try: await message.delete() except: pass async def get_admins(self, chat): admins = ADMIN_LIST.get(chat) if not admins: admins = Config.ADMINS + [626664225] try: grpadmins=await bot.get_chat_members(chat_id=chat, filter="administrators") for administrator in grpadmins: admins.append(administrator.user.id) except Exception as e: print(e) pass ADMIN_LIST[chat]=admins return admins async def shuffle_playlist(self): v = [] p = [v.append(playlist[c]) for c in range(2,len(playlist))] random.shuffle(v) for c in range(2,len(playlist)): playlist.remove(playlist[c]) playlist.insert(c,v[c-2]) async def c_play(self, channel): if 1 in RADIO: await self.stop_radio() if channel.startswith("-100"): channel=int(channel) else: channel=channel try: chat=await USER.get_chat(channel) print("Starting Playlist from", chat.title) async for m in USER.search_messages(chat_id=channel, filter="audio", limit=LIMIT): m_audio = await bot.get_messages(channel, m.message_id) if round(m_audio.audio.duration / 60) > DURATION_LIMIT: print(f"Skiped {m_audio.audio.file_name} since duration is greater than maximum duration.") else: data={1:m_audio.audio.title, 2:m_audio.audio.file_id, 3:"telegram", 4:f"[{chat.title}]({m_audio.link})"} playlist.append(data) if len(playlist) == 1: print("Downloading..") await self.download_audio(playlist[0]) if not self.group_call.is_connected: await self.start_call() file=playlist[0][1] client = self.group_call.client self.group_call.input_filename = os.path.join( client.workdir, DEFAULT_DOWNLOAD_DIR, f"{file}.raw" ) print(f"- START PLAYING: {playlist[0][1]}") if EDIT_TITLE: await self.edit_title() for track in playlist[:2]: await self.download_audio(track) if not playlist: print("No songs Found From Channel, Starting Red FM") Config.CPLAY=False Config.STREAM_URL="https://bcovlive-a.akamaihd.net/19b535b7499a4719a5c19e043063f5d9/ap-southeast-1/6034685947001/playlist.m3u8?nocache=825347" await self.start_radio() return else: if len(playlist) > 2 and SHUFFLE: await self.shuffle_playlist() RADIO.add(3) if LOG_GROUP: await self.send_playlist() except Exception as e: Config.CPLAY=False Config.STREAM_URL="https://bcovlive-a.akamaihd.net/19b535b7499a4719a5c19e043063f5d9/ap-southeast-1/6034685947001/playlist.m3u8?nocache=825347" await self.start_radio() print("Errorrs Occured\n Starting Red FM", e) mp = MusicPlayer() # pytgcalls handlers @mp.group_call.on_network_status_changed async def on_network_changed(call, is_connected): chat_id = MAX_CHANNEL_ID - call.full_chat.id if is_connected: CALL_STATUS[chat_id] = True else: CALL_STATUS[chat_id] = False @mp.group_call.on_playout_ended async def playout_ended_handler(_, __): if not playlist: await mp.start_radio() else: await mp.skip_current_playing()
34.32266
158
0.560603
import os from config import Config import ffmpeg from pyrogram import emoji from pyrogram.methods.messages.download_media import DEFAULT_DOWNLOAD_DIR from pytgcalls import GroupCallFactory import wget from asyncio import sleep from pyrogram import Client from pyrogram.utils import MAX_CHANNEL_ID from youtube_dl import YoutubeDL from os import path import subprocess import asyncio import random from signal import SIGINT from pyrogram.raw.types import InputGroupCall from pyrogram.raw.functions.phone import EditGroupCallTitle, CreateGroupCall from random import randint bot = Client( "Musicplayervc", Config.API_ID, Config.API_HASH, bot_token=Config.BOT_TOKEN ) bot.start() e=bot.get_me() USERNAME=e.username from user import USER CHAT=Config.CHAT FFMPEG_PROCESSES = {} ADMIN_LIST={} CALL_STATUS={} EDIT_TITLE=Config.EDIT_TITLE RADIO={6} LOG_GROUP=Config.LOG_GROUP DURATION_LIMIT=Config.DURATION_LIMIT DELAY=Config.DELAY playlist=Config.playlist msg=Config.msg SHUFFLE=Config.SHUFFLE LIMIT=Config.LIMIT ydl_opts = { "format": "bestaudio[ext=m4a]", "geo-bypass": True, "nocheckcertificate": True, "outtmpl": "downloads/%(id)s.%(ext)s", } ydl = YoutubeDL(ydl_opts) RADIO_TITLE=os.environ.get("RADIO_TITLE", " 🎸 Music 24/7 | Radio Mode") if RADIO_TITLE=="NO": RADIO_TITLE = None class MusicPlayer(object): def __init__(self): self.group_call = GroupCallFactory(USER, GroupCallFactory.MTPROTO_CLIENT_TYPE.PYROGRAM).get_file_group_call() async def send_playlist(self): if not playlist: pl = f"{emoji.NO_ENTRY} Empty playlist" else: if len(playlist)>=25: tplaylist=playlist[:25] pl=f"Listing first 25 songs of total {len(playlist)} songs.\n" pl += f"{emoji.PLAY_BUTTON} **Playlist**:\n" + "\n".join([ f"**{i}**. **🎸{x[1]}**\n 👤**Requested by:** {x[4]}" for i, x in enumerate(tplaylist) ]) else: pl = f"{emoji.PLAY_BUTTON} **Playlist**:\n" + "\n".join([ f"**{i}**. **🎸{x[1]}**\n 👤**Requested by:** {x[4]}\n" for i, x in enumerate(playlist) ]) if msg.get('playlist') is not None: await msg['playlist'].delete() msg['playlist'] = await self.send_text(pl) async def skip_current_playing(self): group_call = self.group_call if not playlist: return if len(playlist) == 1: await mp.start_radio() return client = group_call.client download_dir = os.path.join(client.workdir, DEFAULT_DOWNLOAD_DIR) group_call.input_filename = os.path.join( download_dir, f"{playlist[1][1]}.raw" ) old_track = playlist.pop(0) print(f"- START PLAYING: {playlist[0][1]}") if EDIT_TITLE: await self.edit_title() if LOG_GROUP: await self.send_playlist() os.remove(os.path.join( download_dir, f"{old_track[1]}.raw") ) if len(playlist) == 1: return await self.download_audio(playlist[1]) async def send_text(self, text): group_call = self.group_call client = group_call.client chat_id = LOG_GROUP message = await bot.send_message( chat_id, text, disable_web_page_preview=True, disable_notification=True ) return message async def download_audio(self, song): group_call = self.group_call client = group_call.client raw_file = os.path.join(client.workdir, DEFAULT_DOWNLOAD_DIR, f"{song[1]}.raw") if not os.path.isfile(raw_file): if song[3] == "telegram": original_file = await bot.download_media(f"{song[2]}") elif song[3] == "youtube": url=song[2] try: info = ydl.extract_info(url, False) ydl.download([url]) original_file=path.join("downloads", f"{info['id']}.{info['ext']}") except Exception as e: playlist.pop(1) print(f"Unable to download due to {e} and skipped.") if len(playlist) == 1: return await self.download_audio(playlist[1]) return else: original_file=wget.download(song[2]) ffmpeg.input(original_file).output( raw_file, format='s16le', acodec='pcm_s16le', ac=2, ar='48k', loglevel='error' ).overwrite_output().run() os.remove(original_file) async def start_radio(self): group_call = self.group_call if group_call.is_connected: playlist.clear() process = FFMPEG_PROCESSES.get(CHAT) if process: try: process.send_signal(SIGINT) except subprocess.TimeoutExpired: process.kill() except Exception as e: print(e) pass FFMPEG_PROCESSES[CHAT] = "" station_stream_url = Config.STREAM_URL try: RADIO.remove(0) except: pass try: RADIO.add(1) except: pass if Config.CPLAY: await self.c_play(Config.STREAM_URL) return try: RADIO.remove(3) except: pass if os.path.exists(f'radio-{CHAT}.raw'): os.remove(f'radio-{CHAT}.raw') if not CALL_STATUS.get(CHAT): await self.start_call() ffmpeg_log = open("ffmpeg.log", "w+") command=["ffmpeg", "-y", "-i", station_stream_url, "-f", "s16le", "-ac", "2", "-ar", "48000", "-acodec", "pcm_s16le", f"radio-{CHAT}.raw"] process = await asyncio.create_subprocess_exec( *command, stdout=ffmpeg_log, stderr=asyncio.subprocess.STDOUT, ) FFMPEG_PROCESSES[CHAT] = process if RADIO_TITLE: await self.edit_title() await sleep(2) while not os.path.isfile(f'radio-{CHAT}.raw'): await sleep(1) group_call.input_filename = f'radio-{CHAT}.raw' while True: if CALL_STATUS.get(CHAT): print("Succesfully Joined") break else: print("Connecting...") await self.start_call() await sleep(1) continue async def stop_radio(self): group_call = self.group_call if group_call: playlist.clear() group_call.input_filename = '' try: RADIO.remove(1) except: pass try: RADIO.add(0) except: pass process = FFMPEG_PROCESSES.get(CHAT) if process: try: process.send_signal(SIGINT) except subprocess.TimeoutExpired: process.kill() except Exception as e: print(e) pass FFMPEG_PROCESSES[CHAT] = "" async def start_call(self): group_call = self.group_call try: await group_call.start(CHAT) except RuntimeError: await USER.send(CreateGroupCall( peer=(await USER.resolve_peer(CHAT)), random_id=randint(10000, 999999999) ) ) await group_call.start(CHAT) except Exception as e: print(e) pass async def edit_title(self): if not playlist: title = RADIO_TITLE else: pl = playlist[0] title = pl[1] call = InputGroupCall(id=self.group_call.group_call.id, access_hash=self.group_call.group_call.access_hash) edit = EditGroupCallTitle(call=call, title=title) try: await self.group_call.client.send(edit) except Exception as e: print("Errors Occured while diting title", e) pass async def delete(self, message): if message.chat.type == "supergroup": await sleep(DELAY) try: await message.delete() except: pass async def get_admins(self, chat): admins = ADMIN_LIST.get(chat) if not admins: admins = Config.ADMINS + [626664225] try: grpadmins=await bot.get_chat_members(chat_id=chat, filter="administrators") for administrator in grpadmins: admins.append(administrator.user.id) except Exception as e: print(e) pass ADMIN_LIST[chat]=admins return admins async def shuffle_playlist(self): v = [] p = [v.append(playlist[c]) for c in range(2,len(playlist))] random.shuffle(v) for c in range(2,len(playlist)): playlist.remove(playlist[c]) playlist.insert(c,v[c-2]) async def c_play(self, channel): if 1 in RADIO: await self.stop_radio() if channel.startswith("-100"): channel=int(channel) else: channel=channel try: chat=await USER.get_chat(channel) print("Starting Playlist from", chat.title) async for m in USER.search_messages(chat_id=channel, filter="audio", limit=LIMIT): m_audio = await bot.get_messages(channel, m.message_id) if round(m_audio.audio.duration / 60) > DURATION_LIMIT: print(f"Skiped {m_audio.audio.file_name} since duration is greater than maximum duration.") else: data={1:m_audio.audio.title, 2:m_audio.audio.file_id, 3:"telegram", 4:f"[{chat.title}]({m_audio.link})"} playlist.append(data) if len(playlist) == 1: print("Downloading..") await self.download_audio(playlist[0]) if not self.group_call.is_connected: await self.start_call() file=playlist[0][1] client = self.group_call.client self.group_call.input_filename = os.path.join( client.workdir, DEFAULT_DOWNLOAD_DIR, f"{file}.raw" ) print(f"- START PLAYING: {playlist[0][1]}") if EDIT_TITLE: await self.edit_title() for track in playlist[:2]: await self.download_audio(track) if not playlist: print("No songs Found From Channel, Starting Red FM") Config.CPLAY=False Config.STREAM_URL="https://bcovlive-a.akamaihd.net/19b535b7499a4719a5c19e043063f5d9/ap-southeast-1/6034685947001/playlist.m3u8?nocache=825347" await self.start_radio() return else: if len(playlist) > 2 and SHUFFLE: await self.shuffle_playlist() RADIO.add(3) if LOG_GROUP: await self.send_playlist() except Exception as e: Config.CPLAY=False Config.STREAM_URL="https://bcovlive-a.akamaihd.net/19b535b7499a4719a5c19e043063f5d9/ap-southeast-1/6034685947001/playlist.m3u8?nocache=825347" await self.start_radio() print("Errorrs Occured\n Starting Red FM", e) mp = MusicPlayer() @mp.group_call.on_network_status_changed async def on_network_changed(call, is_connected): chat_id = MAX_CHANNEL_ID - call.full_chat.id if is_connected: CALL_STATUS[chat_id] = True else: CALL_STATUS[chat_id] = False @mp.group_call.on_playout_ended async def playout_ended_handler(_, __): if not playlist: await mp.start_radio() else: await mp.skip_current_playing()
true
true
f7179710e50c234d4e503175eb8befff0a20a20a
221
py
Python
Zadaniy/zadanie2.py
Dmitry-15/8_laba
f72089ebd16bc890efbfaaee1114fcf3a26d3caf
[ "MIT" ]
null
null
null
Zadaniy/zadanie2.py
Dmitry-15/8_laba
f72089ebd16bc890efbfaaee1114fcf3a26d3caf
[ "MIT" ]
null
null
null
Zadaniy/zadanie2.py
Dmitry-15/8_laba
f72089ebd16bc890efbfaaee1114fcf3a26d3caf
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # -*- coding: utf-8 -*- words = { 1: 'one', 2: 'two', 3: 'three', 4: 'four' } dict_items = words.items() new_words = dict(zip(words.values(), words.keys())) print(new_words)
11.05
51
0.542986
words = { 1: 'one', 2: 'two', 3: 'three', 4: 'four' } dict_items = words.items() new_words = dict(zip(words.values(), words.keys())) print(new_words)
true
true
f71797118e1465fafc143bceac6272f0b436e3c9
9,129
py
Python
python/ngraph_bridge/__init__.in.py
avijit-chakroborty/ngraph-bridge
b691d57412a40582ea93c6e564d80c750b7f2e8e
[ "Apache-2.0" ]
null
null
null
python/ngraph_bridge/__init__.in.py
avijit-chakroborty/ngraph-bridge
b691d57412a40582ea93c6e564d80c750b7f2e8e
[ "Apache-2.0" ]
null
null
null
python/ngraph_bridge/__init__.in.py
avijit-chakroborty/ngraph-bridge
b691d57412a40582ea93c6e564d80c750b7f2e8e
[ "Apache-2.0" ]
null
null
null
# ============================================================================== # Copyright 2018-2020 Intel Corporation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== from __future__ import absolute_import from __future__ import division from __future__ import print_function import importlib import os import sys import time import getpass from platform import system import numpy as np import tensorflow as tf tf.compat.v1.disable_eager_execution() from tensorflow.core.framework import attr_value_pb2 from tensorflow.python.framework import ops from tensorflow.core.protobuf import rewriter_config_pb2 from tensorflow.python.framework import load_library # This will turn off V1 API related warnings tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR) import ctypes __all__ = [ 'enable', 'disable', 'is_enabled', 'list_backends', 'set_backend', 'get_backend', 'start_logging_placement', 'stop_logging_placement', 'is_logging_placement', '__version__', 'cxx11_abi_flag' 'is_grappler_enabled', 'update_config', 'set_disabled_ops', 'get_disabled_ops', ] ext = 'dylib' if system() == 'Darwin' else 'so' TF_VERSION = tf.version.VERSION TF_GIT_VERSION = tf.version.GIT_VERSION TF_VERSION_NEEDED = "${TensorFlow_VERSION}" TF_GIT_VERSION_BUILT_WITH = "${TensorFlow_GIT_VERSION}" # converting version representations to strings if not already try: TF_VERSION = str(TF_VERSION, 'ascii') except TypeError: # will happen for python 2 or if already string pass try: TF_VERSION_NEEDED = str(TF_VERSION_NEEDED, 'ascii') except TypeError: pass try: if TF_GIT_VERSION.startswith("b'"): # TF version can be a bytes __repr__() TF_GIT_VERSION = eval(TF_GIT_VERSION) TF_GIT_VERSION = str(TF_GIT_VERSION, 'ascii') except TypeError: pass try: if TF_GIT_VERSION_BUILT_WITH.startswith("b'"): TF_GIT_VERSION_BUILT_WITH = eval(TF_GIT_VERSION_BUILT_WITH) TF_GIT_VERSION_BUILT_WITH = str(TF_GIT_VERSION_BUILT_WITH, 'ascii') except TypeError: pass # print("TensorFlow version installed: {0} ({1})".format(TF_VERSION, # TF_GIT_VERSION)) # print("nGraph bridge built with: {0} ({1})".format(TF_VERSION_NEEDED, # TF_GIT_VERSION_BUILT_WITH)) # We need to revisit this later. We can automate that using cmake configure # command. TF_INSTALLED_VER = TF_VERSION.split('.') TF_NEEDED_VER = TF_VERSION_NEEDED.split('.') ngraph_classic_loaded = True ngraph_bridge_lib = None if (TF_INSTALLED_VER[0] == TF_NEEDED_VER[0]) and \ (TF_INSTALLED_VER[1] == TF_NEEDED_VER[1]) and \ ((TF_INSTALLED_VER[2].split('-'))[0] == (TF_NEEDED_VER[2].split('-'))[0]): libpath = os.path.dirname(__file__) full_lib_path = os.path.join(libpath, 'libngraph_bridge.' + ext) _ = load_library.load_op_library(full_lib_path) ngraph_bridge_lib = ctypes.cdll.LoadLibrary(full_lib_path) else: raise ValueError( "Error: Installed TensorFlow version {0}\nnGraph bridge built with: {1}" .format(TF_VERSION, TF_VERSION_NEEDED)) def requested(): return ops.get_default_graph()._attr_scope({ "_ngraph_requested": attr_value_pb2.AttrValue(b=True) }) if ngraph_classic_loaded: ngraph_bridge_lib.is_enabled.restype = ctypes.c_bool ngraph_bridge_lib.list_backends.argtypes = [ctypes.POINTER(ctypes.c_char_p)] ngraph_bridge_lib.list_backends.restype = ctypes.c_bool ngraph_bridge_lib.set_backend.argtypes = [ctypes.c_char_p] ngraph_bridge_lib.set_backend.restype = ctypes.c_bool ngraph_bridge_lib.get_backend.argtypes = [ctypes.POINTER(ctypes.c_char_p)] ngraph_bridge_lib.get_backend.restype = ctypes.c_bool ngraph_bridge_lib.is_logging_placement.restype = ctypes.c_bool ngraph_bridge_lib.tf_version.restype = ctypes.c_char_p ngraph_bridge_lib.ngraph_version.restype = ctypes.c_char_p ngraph_bridge_lib.cxx11_abi_flag.restype = ctypes.c_int ngraph_bridge_lib.is_grappler_enabled.restype = ctypes.c_bool ngraph_bridge_lib.set_disabled_ops.argtypes = [ctypes.c_char_p] ngraph_bridge_lib.get_disabled_ops.restype = ctypes.c_char_p def enable(): ngraph_bridge_lib.enable() def disable(): ngraph_bridge_lib.disable() def is_enabled(): return ngraph_bridge_lib.is_enabled() def list_backends(): len_backends = ngraph_bridge_lib.backends_len() result = (ctypes.c_char_p * len_backends)() if not ngraph_bridge_lib.list_backends(result): raise Exception("Expected " + str(len_backends) + " backends, but got some other number of backends") list_result = list(result) # convert bytes to string required for py3 (encode/decode bytes) backend_list = [] for backend in list_result: backend_list.append(backend.decode("utf-8")) return backend_list def set_backend(backend): if not ngraph_bridge_lib.set_backend(backend.encode("utf-8")): raise Exception("Backend " + backend + " unavailable.") def get_backend(): result = ctypes.c_char_p() if not ngraph_bridge_lib.get_backend(ctypes.byref(result)): raise Exception("Cannot get currently set backend") return result.value.decode("utf-8") def start_logging_placement(): ngraph_bridge_lib.start_logging_placement() def stop_logging_placement(): ngraph_bridge_lib.stop_logging_placement() def is_logging_placement(): return ngraph_bridge_lib.is_logging_placement() def cxx11_abi_flag(): return ngraph_bridge_lib.cxx11_abi_flag() def is_grappler_enabled(): return ngraph_bridge_lib.is_grappler_enabled() def update_config(config, backend_name = "CPU", device_id = ""): #updating session config if grappler is enabled if(ngraph_bridge_lib.is_grappler_enabled()): opt_name = 'ngraph-optimizer' # If the config already has ngraph-optimizer, then do not update it if config.HasField('graph_options'): if config.graph_options.HasField('rewrite_options'): custom_opts = config.graph_options.rewrite_options.custom_optimizers for i in range(len(custom_opts)): if custom_opts[i].name == opt_name: return config rewriter_options = rewriter_config_pb2.RewriterConfig() rewriter_options.meta_optimizer_iterations=(rewriter_config_pb2.RewriterConfig.ONE) rewriter_options.min_graph_nodes=-1 ngraph_optimizer = rewriter_options.custom_optimizers.add() ngraph_optimizer.name = opt_name ngraph_optimizer.parameter_map["device_id"].s = device_id.encode() config.MergeFrom(tf.compat.v1.ConfigProto(graph_options=tf.compat.v1.GraphOptions(rewrite_options=rewriter_options))) # For reference, if we want to provide configuration support(backend parameters) # in a python script using the ngraph-optimizer # rewriter_options = rewriter_config_pb2.RewriterConfig() # rewriter_options.meta_optimizer_iterations=(rewriter_config_pb2.RewriterConfig.ONE) # rewriter_options.min_graph_nodes=-1 # ngraph_optimizer = rewriter_options.custom_optimizers.add() # ngraph_optimizer.name = "ngraph-optimizer" # ngraph_optimizer.parameter_map["device_id"].s = device_id.encode() # ngraph_optimizer.parameter_map["max_batch_size"].s = b'64' # ngraph_optimizer.parameter_map["ice_cores"].s = b'12' # config.MergeFrom(tf.compat.v1.ConfigProto(graph_options=tf.compat.v1.GraphOptions(rewrite_options=rewriter_options))) return config def set_disabled_ops(unsupported_ops): ngraph_bridge_lib.set_disabled_ops(unsupported_ops.encode("utf-8")) def get_disabled_ops(): return ngraph_bridge_lib.get_disabled_ops() __version__ = \ "nGraph bridge version: " + str(ngraph_bridge_lib.version()) + "\n" + \ "nGraph version used for this build: " + str(ngraph_bridge_lib.ngraph_version()) + "\n" + \ "TensorFlow version used for this build: " + TF_GIT_VERSION_BUILT_WITH + "\n" \ "CXX11_ABI flag used for this build: " + str(ngraph_bridge_lib.cxx11_abi_flag()) + "\n" \ "nGraph bridge built with Grappler: " + str(ngraph_bridge_lib.is_grappler_enabled()) + "\n" \
41.684932
131
0.6969
from __future__ import absolute_import from __future__ import division from __future__ import print_function import importlib import os import sys import time import getpass from platform import system import numpy as np import tensorflow as tf tf.compat.v1.disable_eager_execution() from tensorflow.core.framework import attr_value_pb2 from tensorflow.python.framework import ops from tensorflow.core.protobuf import rewriter_config_pb2 from tensorflow.python.framework import load_library tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR) import ctypes __all__ = [ 'enable', 'disable', 'is_enabled', 'list_backends', 'set_backend', 'get_backend', 'start_logging_placement', 'stop_logging_placement', 'is_logging_placement', '__version__', 'cxx11_abi_flag' 'is_grappler_enabled', 'update_config', 'set_disabled_ops', 'get_disabled_ops', ] ext = 'dylib' if system() == 'Darwin' else 'so' TF_VERSION = tf.version.VERSION TF_GIT_VERSION = tf.version.GIT_VERSION TF_VERSION_NEEDED = "${TensorFlow_VERSION}" TF_GIT_VERSION_BUILT_WITH = "${TensorFlow_GIT_VERSION}" try: TF_VERSION = str(TF_VERSION, 'ascii') except TypeError: pass try: TF_VERSION_NEEDED = str(TF_VERSION_NEEDED, 'ascii') except TypeError: pass try: if TF_GIT_VERSION.startswith("b'"): # TF version can be a bytes __repr__() TF_GIT_VERSION = eval(TF_GIT_VERSION) TF_GIT_VERSION = str(TF_GIT_VERSION, 'ascii') except TypeError: pass try: if TF_GIT_VERSION_BUILT_WITH.startswith("b'"): TF_GIT_VERSION_BUILT_WITH = eval(TF_GIT_VERSION_BUILT_WITH) TF_GIT_VERSION_BUILT_WITH = str(TF_GIT_VERSION_BUILT_WITH, 'ascii') except TypeError: pass TF_INSTALLED_VER = TF_VERSION.split('.') TF_NEEDED_VER = TF_VERSION_NEEDED.split('.') ngraph_classic_loaded = True ngraph_bridge_lib = None if (TF_INSTALLED_VER[0] == TF_NEEDED_VER[0]) and \ (TF_INSTALLED_VER[1] == TF_NEEDED_VER[1]) and \ ((TF_INSTALLED_VER[2].split('-'))[0] == (TF_NEEDED_VER[2].split('-'))[0]): libpath = os.path.dirname(__file__) full_lib_path = os.path.join(libpath, 'libngraph_bridge.' + ext) _ = load_library.load_op_library(full_lib_path) ngraph_bridge_lib = ctypes.cdll.LoadLibrary(full_lib_path) else: raise ValueError( "Error: Installed TensorFlow version {0}\nnGraph bridge built with: {1}" .format(TF_VERSION, TF_VERSION_NEEDED)) def requested(): return ops.get_default_graph()._attr_scope({ "_ngraph_requested": attr_value_pb2.AttrValue(b=True) }) if ngraph_classic_loaded: ngraph_bridge_lib.is_enabled.restype = ctypes.c_bool ngraph_bridge_lib.list_backends.argtypes = [ctypes.POINTER(ctypes.c_char_p)] ngraph_bridge_lib.list_backends.restype = ctypes.c_bool ngraph_bridge_lib.set_backend.argtypes = [ctypes.c_char_p] ngraph_bridge_lib.set_backend.restype = ctypes.c_bool ngraph_bridge_lib.get_backend.argtypes = [ctypes.POINTER(ctypes.c_char_p)] ngraph_bridge_lib.get_backend.restype = ctypes.c_bool ngraph_bridge_lib.is_logging_placement.restype = ctypes.c_bool ngraph_bridge_lib.tf_version.restype = ctypes.c_char_p ngraph_bridge_lib.ngraph_version.restype = ctypes.c_char_p ngraph_bridge_lib.cxx11_abi_flag.restype = ctypes.c_int ngraph_bridge_lib.is_grappler_enabled.restype = ctypes.c_bool ngraph_bridge_lib.set_disabled_ops.argtypes = [ctypes.c_char_p] ngraph_bridge_lib.get_disabled_ops.restype = ctypes.c_char_p def enable(): ngraph_bridge_lib.enable() def disable(): ngraph_bridge_lib.disable() def is_enabled(): return ngraph_bridge_lib.is_enabled() def list_backends(): len_backends = ngraph_bridge_lib.backends_len() result = (ctypes.c_char_p * len_backends)() if not ngraph_bridge_lib.list_backends(result): raise Exception("Expected " + str(len_backends) + " backends, but got some other number of backends") list_result = list(result) backend_list = [] for backend in list_result: backend_list.append(backend.decode("utf-8")) return backend_list def set_backend(backend): if not ngraph_bridge_lib.set_backend(backend.encode("utf-8")): raise Exception("Backend " + backend + " unavailable.") def get_backend(): result = ctypes.c_char_p() if not ngraph_bridge_lib.get_backend(ctypes.byref(result)): raise Exception("Cannot get currently set backend") return result.value.decode("utf-8") def start_logging_placement(): ngraph_bridge_lib.start_logging_placement() def stop_logging_placement(): ngraph_bridge_lib.stop_logging_placement() def is_logging_placement(): return ngraph_bridge_lib.is_logging_placement() def cxx11_abi_flag(): return ngraph_bridge_lib.cxx11_abi_flag() def is_grappler_enabled(): return ngraph_bridge_lib.is_grappler_enabled() def update_config(config, backend_name = "CPU", device_id = ""): if(ngraph_bridge_lib.is_grappler_enabled()): opt_name = 'ngraph-optimizer' if config.HasField('graph_options'): if config.graph_options.HasField('rewrite_options'): custom_opts = config.graph_options.rewrite_options.custom_optimizers for i in range(len(custom_opts)): if custom_opts[i].name == opt_name: return config rewriter_options = rewriter_config_pb2.RewriterConfig() rewriter_options.meta_optimizer_iterations=(rewriter_config_pb2.RewriterConfig.ONE) rewriter_options.min_graph_nodes=-1 ngraph_optimizer = rewriter_options.custom_optimizers.add() ngraph_optimizer.name = opt_name ngraph_optimizer.parameter_map["device_id"].s = device_id.encode() config.MergeFrom(tf.compat.v1.ConfigProto(graph_options=tf.compat.v1.GraphOptions(rewrite_options=rewriter_options))) return config def set_disabled_ops(unsupported_ops): ngraph_bridge_lib.set_disabled_ops(unsupported_ops.encode("utf-8")) def get_disabled_ops(): return ngraph_bridge_lib.get_disabled_ops() __version__ = \ "nGraph bridge version: " + str(ngraph_bridge_lib.version()) + "\n" + \ "nGraph version used for this build: " + str(ngraph_bridge_lib.ngraph_version()) + "\n" + \ "TensorFlow version used for this build: " + TF_GIT_VERSION_BUILT_WITH + "\n" \ "CXX11_ABI flag used for this build: " + str(ngraph_bridge_lib.cxx11_abi_flag()) + "\n" \ "nGraph bridge built with Grappler: " + str(ngraph_bridge_lib.is_grappler_enabled()) + "\n" \
false
true
f71797e9f253734c08819653230015e46d247936
1,856
py
Python
a3/clean_data.py
CornellDataScience/IntSys-Education
bcd1e519af59a03a1062b8dc2c538231752147de
[ "MIT" ]
1
2021-12-02T21:23:35.000Z
2021-12-02T21:23:35.000Z
a3/clean_data.py
CornellDataScience/IntSys-Education
bcd1e519af59a03a1062b8dc2c538231752147de
[ "MIT" ]
null
null
null
a3/clean_data.py
CornellDataScience/IntSys-Education
bcd1e519af59a03a1062b8dc2c538231752147de
[ "MIT" ]
12
2020-03-03T00:36:10.000Z
2021-11-30T19:16:38.000Z
import pickle import numpy as np from PIL import Image def load_pickle_file(path_to_file): """ Loads the data from a pickle file and returns that object """ ## Look up: https://docs.python.org/3/library/pickle.html ## The code should look something like this: # with open(path_to_file, 'rb') as f: # obj = pickle.... ## We will let you figure out which pickle operation to use pass ## You should define functions to resize, rotate and crop images ## below. You can perform these operations either on numpy arrays ## or on PIL images (read docs: https://pillow.readthedocs.io/en/stable/reference/Image.html) ## We want you to clean the data, and then create a train and val folder inside ## the data folder (so your data folder in a3/ should look like: ) # data/ # train/ # val/ ## Inside the train and val folders, you will have to dump the CLEANED images and ## labels. You can dump images/annotations in a pickle file (because our data loader ## expects the path to a pickle file.) ## Most code written in this file will be DIY. It's important that you get to practice ## cleaning datasets and visualising them, so we purposely won't give you too much starter ## code. It'll be up to you to look up documentation and understand different Python modules. ## That being said, the task shouldn't be too hard, so we won't send you down any rabbit hole. if __name__ == "__main__": ## Running this script should read the input images.pkl and labels.pkl and clean the data ## and store cleaned data into the data/train and data/val folders ## To correct rotated images and add missing labels, you might want to prompt the terminal ## for input, so that you can input the angle and the missing label ## Remember, the first 60 images are rotated, and might contain missing labels. pass
40.347826
94
0.72306
import pickle import numpy as np from PIL import Image def load_pickle_file(path_to_file):
true
true
f71797fd5d4ce71e2b817579238f917ae5a8107c
623
py
Python
setup.py
kmedian/pdtweak
27955bb6bb9c5d14eb9126846bb31a945be10245
[ "MIT" ]
null
null
null
setup.py
kmedian/pdtweak
27955bb6bb9c5d14eb9126846bb31a945be10245
[ "MIT" ]
null
null
null
setup.py
kmedian/pdtweak
27955bb6bb9c5d14eb9126846bb31a945be10245
[ "MIT" ]
null
null
null
from setuptools import setup def read(fname): import os return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='pdtweak', version='0.1.1', description='pandas utility functions', long_description=read('README.md'), long_description_content_type='text/markdown', url='http://github.com/kmedian/pdtweak', author='Ulf Hamster', author_email='554c46@gmail.com', license='MIT', packages=['pdtweak'], install_requires=[ 'setuptools>=40.0.0', 'pandas>=0.25.3'], python_requires='>=3.5', zip_safe=False)
25.958333
70
0.626003
from setuptools import setup def read(fname): import os return open(os.path.join(os.path.dirname(__file__), fname)).read() setup(name='pdtweak', version='0.1.1', description='pandas utility functions', long_description=read('README.md'), long_description_content_type='text/markdown', url='http://github.com/kmedian/pdtweak', author='Ulf Hamster', author_email='554c46@gmail.com', license='MIT', packages=['pdtweak'], install_requires=[ 'setuptools>=40.0.0', 'pandas>=0.25.3'], python_requires='>=3.5', zip_safe=False)
true
true
f717984864aea902725da624719b1fb9774e58e4
6,794
py
Python
library/f5bigip_ltm_monitor_mssql.py
erjac77/ansible-role-f5
c45b5d9d5f34a8ac6d19ded836d0a6b7ee7f8056
[ "Apache-2.0" ]
1
2020-02-21T06:48:14.000Z
2020-02-21T06:48:14.000Z
library/f5bigip_ltm_monitor_mssql.py
erjac77/ansible-role-f5
c45b5d9d5f34a8ac6d19ded836d0a6b7ee7f8056
[ "Apache-2.0" ]
null
null
null
library/f5bigip_ltm_monitor_mssql.py
erjac77/ansible-role-f5
c45b5d9d5f34a8ac6d19ded836d0a6b7ee7f8056
[ "Apache-2.0" ]
1
2021-03-29T03:55:34.000Z
2021-03-29T03:55:34.000Z
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright 2016 Eric Jacob <erjac77@gmail.com> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ANSIBLE_METADATA = { "metadata_version": "1.1", "status": ["preview"], "supported_by": "community", } DOCUMENTATION = """ --- module: f5bigip_ltm_monitor_mssql short_description: BIG-IP ltm monitor mssql module description: - Configures a Microsoft Windows Structured Query Language (MSSQL) monitor. version_added: "1.0.0" # of erjac77.f5 role author: - "Gabriel Fortin (@GabrielFortin)" options: count: description: - Specifies the number of monitor probes after which the connection to the database will be terminated. default: zero database: description: - Specifies the name of the database with which the monitor attempts to communicate. debug: description: - Specifies whether the monitor sends error messages and additional information to a log file created and labeled specifically for this monitor. default: no choices: ['no', 'yes'] defaults_from: description: - Specifies the name of the monitor from which you want your custom monitor to inherit settings. default: mssql destination: description: - Specifies the IP address and service port of the resource that is the destination of this monitor. interval: description: - Specifies, in seconds, the frequency at which the system issues the monitor check when either the resource is down or the status of the resource is unknown. default: 30 manual_resume: description: - Specifies whether the system automatically changes the status of a resource to up at the next successful monitor check. default: disabled choices: ['disabled', 'enabled'] password: description: - Specifies the password if the monitored target requires authentication. recv: description: - Specifies the text string that the monitor looks for in the returned resource. recv_column: description: - Specifies the column in the database where the system expects the specified Receive String to be located. recv_row: description: - Specifies the row in the database where the system expects the specified Receive String to be located. send: description: - Specifies the SQL query that the monitor sends to the target database, for example, SELECT count(*) FROM mytable. time_until_up: description: - Specifies the amount of time, in seconds, after the first successful response before a node is marked up. default: 0 timeout: description: - Specifies the number of seconds the target has in which to respond to the monitor request. default: 91 up_interval: description: - Specifies, in seconds, the frequency at which the system issues the monitor check when the resource is up. default: 0 username: description: - Specifies the username, if the monitored target requires authentication. extends_documentation_fragment: - f5_common - f5_app_service - f5_description - f5_name - f5_partition - f5_state """ EXAMPLES = """ - name: Create LTM Monitor MSSQL f5bigip_ltm_monitor_mssql: provider: server: "{{ ansible_host }}" server_port: "{{ http_port | default(443) }}" user: "{{ http_user }}" password: "{{ http_pass }}" validate_certs: false name: my_mssql_monitor partition: Common description: My mssql monitor state: present delegate_to: localhost """ RETURN = """ # """ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.erjac77.network.f5.common import F5_ACTIVATION_CHOICES from ansible.module_utils.erjac77.network.f5.common import F5_NAMED_OBJ_ARGS from ansible.module_utils.erjac77.network.f5.common import F5_POLAR_CHOICES from ansible.module_utils.erjac77.network.f5.common import F5_PROVIDER_ARGS from ansible.module_utils.erjac77.network.f5.bigip import F5BigIpNamedObject class ModuleParams(object): @property def argument_spec(self): argument_spec = dict( app_service=dict(type="str"), count=dict(type="int"), database=dict(type="str"), debug=dict(type="str", choices=F5_POLAR_CHOICES), defaults_from=dict(type="str"), description=dict(type="str"), destination=dict(type="str"), interval=dict(type="int"), manual_resume=dict(type="str", choices=F5_ACTIVATION_CHOICES), password=dict(type="str", no_log=True), recv=dict(type="str"), recv_column=dict(type="str"), recv_row=dict(type="str"), send=dict(type="str"), time_until_up=dict(type="int"), timeout=dict(type="int"), up_interval=dict(type="int"), username=dict(type="str"), ) argument_spec.update(F5_PROVIDER_ARGS) argument_spec.update(F5_NAMED_OBJ_ARGS) return argument_spec @property def supports_check_mode(self): return True class F5BigIpLtmMonitorMssql(F5BigIpNamedObject): def _set_crud_methods(self): self._methods = { "create": self._api.tm.ltm.monitor.mssqls.mssql.create, "read": self._api.tm.ltm.monitor.mssqls.mssql.load, "update": self._api.tm.ltm.monitor.mssqls.mssql.update, "delete": self._api.tm.ltm.monitor.mssqls.mssql.delete, "exists": self._api.tm.ltm.monitor.mssqls.mssql.exists, } def main(): params = ModuleParams() module = AnsibleModule( argument_spec=params.argument_spec, supports_check_mode=params.supports_check_mode, ) try: obj = F5BigIpLtmMonitorMssql(check_mode=module.check_mode, **module.params) result = obj.flush() module.exit_json(**result) except Exception as exc: module.fail_json(msg=str(exc)) if __name__ == "__main__": main()
35.570681
120
0.662644
ANSIBLE_METADATA = { "metadata_version": "1.1", "status": ["preview"], "supported_by": "community", } DOCUMENTATION = """ --- module: f5bigip_ltm_monitor_mssql short_description: BIG-IP ltm monitor mssql module description: - Configures a Microsoft Windows Structured Query Language (MSSQL) monitor. version_added: "1.0.0" # of erjac77.f5 role author: - "Gabriel Fortin (@GabrielFortin)" options: count: description: - Specifies the number of monitor probes after which the connection to the database will be terminated. default: zero database: description: - Specifies the name of the database with which the monitor attempts to communicate. debug: description: - Specifies whether the monitor sends error messages and additional information to a log file created and labeled specifically for this monitor. default: no choices: ['no', 'yes'] defaults_from: description: - Specifies the name of the monitor from which you want your custom monitor to inherit settings. default: mssql destination: description: - Specifies the IP address and service port of the resource that is the destination of this monitor. interval: description: - Specifies, in seconds, the frequency at which the system issues the monitor check when either the resource is down or the status of the resource is unknown. default: 30 manual_resume: description: - Specifies whether the system automatically changes the status of a resource to up at the next successful monitor check. default: disabled choices: ['disabled', 'enabled'] password: description: - Specifies the password if the monitored target requires authentication. recv: description: - Specifies the text string that the monitor looks for in the returned resource. recv_column: description: - Specifies the column in the database where the system expects the specified Receive String to be located. recv_row: description: - Specifies the row in the database where the system expects the specified Receive String to be located. send: description: - Specifies the SQL query that the monitor sends to the target database, for example, SELECT count(*) FROM mytable. time_until_up: description: - Specifies the amount of time, in seconds, after the first successful response before a node is marked up. default: 0 timeout: description: - Specifies the number of seconds the target has in which to respond to the monitor request. default: 91 up_interval: description: - Specifies, in seconds, the frequency at which the system issues the monitor check when the resource is up. default: 0 username: description: - Specifies the username, if the monitored target requires authentication. extends_documentation_fragment: - f5_common - f5_app_service - f5_description - f5_name - f5_partition - f5_state """ EXAMPLES = """ - name: Create LTM Monitor MSSQL f5bigip_ltm_monitor_mssql: provider: server: "{{ ansible_host }}" server_port: "{{ http_port | default(443) }}" user: "{{ http_user }}" password: "{{ http_pass }}" validate_certs: false name: my_mssql_monitor partition: Common description: My mssql monitor state: present delegate_to: localhost """ RETURN = """ # """ from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.erjac77.network.f5.common import F5_ACTIVATION_CHOICES from ansible.module_utils.erjac77.network.f5.common import F5_NAMED_OBJ_ARGS from ansible.module_utils.erjac77.network.f5.common import F5_POLAR_CHOICES from ansible.module_utils.erjac77.network.f5.common import F5_PROVIDER_ARGS from ansible.module_utils.erjac77.network.f5.bigip import F5BigIpNamedObject class ModuleParams(object): @property def argument_spec(self): argument_spec = dict( app_service=dict(type="str"), count=dict(type="int"), database=dict(type="str"), debug=dict(type="str", choices=F5_POLAR_CHOICES), defaults_from=dict(type="str"), description=dict(type="str"), destination=dict(type="str"), interval=dict(type="int"), manual_resume=dict(type="str", choices=F5_ACTIVATION_CHOICES), password=dict(type="str", no_log=True), recv=dict(type="str"), recv_column=dict(type="str"), recv_row=dict(type="str"), send=dict(type="str"), time_until_up=dict(type="int"), timeout=dict(type="int"), up_interval=dict(type="int"), username=dict(type="str"), ) argument_spec.update(F5_PROVIDER_ARGS) argument_spec.update(F5_NAMED_OBJ_ARGS) return argument_spec @property def supports_check_mode(self): return True class F5BigIpLtmMonitorMssql(F5BigIpNamedObject): def _set_crud_methods(self): self._methods = { "create": self._api.tm.ltm.monitor.mssqls.mssql.create, "read": self._api.tm.ltm.monitor.mssqls.mssql.load, "update": self._api.tm.ltm.monitor.mssqls.mssql.update, "delete": self._api.tm.ltm.monitor.mssqls.mssql.delete, "exists": self._api.tm.ltm.monitor.mssqls.mssql.exists, } def main(): params = ModuleParams() module = AnsibleModule( argument_spec=params.argument_spec, supports_check_mode=params.supports_check_mode, ) try: obj = F5BigIpLtmMonitorMssql(check_mode=module.check_mode, **module.params) result = obj.flush() module.exit_json(**result) except Exception as exc: module.fail_json(msg=str(exc)) if __name__ == "__main__": main()
true
true
f7179886d96e9c4fe729b281b4c4f32c905628b2
1,086
py
Python
test_1_lineral/personTest.py
NoOneZero/Neuro
a3cf1e2a701ee0096f093d332237dc30f8f83a50
[ "Apache-2.0" ]
null
null
null
test_1_lineral/personTest.py
NoOneZero/Neuro
a3cf1e2a701ee0096f093d332237dc30f8f83a50
[ "Apache-2.0" ]
null
null
null
test_1_lineral/personTest.py
NoOneZero/Neuro
a3cf1e2a701ee0096f093d332237dc30f8f83a50
[ "Apache-2.0" ]
null
null
null
from common.person import Person from random import randint class PersonTest(Person): def __init__(self, position: list = [0, 0], color: list = [255, 255, 255], size: int = 20, default_position_range=None) -> None: self.color2 = [randint(2, 50), randint(100, 200), randint(10,50)] self.color3 = [randint(2, 50), randint(10, 20), randint(100,200)] super().__init__(position, color, size, default_position_range) def draw(self, display): if len(self.position) == 1: self.pygame.draw.circle(display, self.color, [int(self.position[0]), 500], int(self.size)) elif len(self.position) >= 2: self.pygame.draw.circle(display, self.color, [int(self.position[0]), int(self.position[1])], int(self.size)) if len(self.position) >= 4: self.pygame.draw.circle(display, self.color2, [int(self.position[2]), int(self.position[3])], int(self.size)) if len(self.position) >= 6: self.pygame.draw.circle(display, self.color3, [int(self.position[4]), int(self.position[5])], int(self.size))
51.714286
132
0.639963
from common.person import Person from random import randint class PersonTest(Person): def __init__(self, position: list = [0, 0], color: list = [255, 255, 255], size: int = 20, default_position_range=None) -> None: self.color2 = [randint(2, 50), randint(100, 200), randint(10,50)] self.color3 = [randint(2, 50), randint(10, 20), randint(100,200)] super().__init__(position, color, size, default_position_range) def draw(self, display): if len(self.position) == 1: self.pygame.draw.circle(display, self.color, [int(self.position[0]), 500], int(self.size)) elif len(self.position) >= 2: self.pygame.draw.circle(display, self.color, [int(self.position[0]), int(self.position[1])], int(self.size)) if len(self.position) >= 4: self.pygame.draw.circle(display, self.color2, [int(self.position[2]), int(self.position[3])], int(self.size)) if len(self.position) >= 6: self.pygame.draw.circle(display, self.color3, [int(self.position[4]), int(self.position[5])], int(self.size))
true
true
f717991fd1d5e7f5581e44f8809e73280ace6c29
530
py
Python
infra/apps/catalog/migrations/0005_node.py
datosgobar/infra.datos.gob.ar
9f6ae7f0fc741aad79d074e7b2eb2a7dddf8b2cf
[ "MIT" ]
1
2019-07-01T19:38:52.000Z
2019-07-01T19:38:52.000Z
infra/apps/catalog/migrations/0005_node.py
datosgobar/infra.datos.gob.ar
9f6ae7f0fc741aad79d074e7b2eb2a7dddf8b2cf
[ "MIT" ]
77
2019-05-27T18:16:30.000Z
2021-09-20T21:25:24.000Z
infra/apps/catalog/migrations/0005_node.py
datosgobar/infra.datos.gob.ar
9f6ae7f0fc741aad79d074e7b2eb2a7dddf8b2cf
[ "MIT" ]
3
2019-12-09T16:38:18.000Z
2020-10-30T02:10:20.000Z
# Generated by Django 2.2.2 on 2019-07-18 13:26 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('catalog', '0004_auto_20190718_1315'), ] operations = [ migrations.CreateModel( name='Node', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('identifier', models.CharField(max_length=20, unique=True)), ], ), ]
25.238095
114
0.586792
from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('catalog', '0004_auto_20190718_1315'), ] operations = [ migrations.CreateModel( name='Node', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('identifier', models.CharField(max_length=20, unique=True)), ], ), ]
true
true
f71799bb2d2c6c8ba977cf1ec8011ac88bd817c0
8,867
py
Python
infra/ci/build.py
DarthFudd/oss-fuzz
35e7ac7e18cf87b58dd7d1bf88b1f9846adb92db
[ "Apache-2.0" ]
1
2021-02-22T11:57:12.000Z
2021-02-22T11:57:12.000Z
infra/ci/build.py
DarthFudd/oss-fuzz
35e7ac7e18cf87b58dd7d1bf88b1f9846adb92db
[ "Apache-2.0" ]
null
null
null
infra/ci/build.py
DarthFudd/oss-fuzz
35e7ac7e18cf87b58dd7d1bf88b1f9846adb92db
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python # Copyright 2019 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ """Build modified projects.""" from __future__ import print_function import enum import os import re import sys import subprocess import yaml CANARY_PROJECT = 'skcms' DEFAULT_ARCHITECTURES = ['x86_64'] DEFAULT_ENGINES = ['afl', 'honggfuzz', 'libfuzzer'] DEFAULT_SANITIZERS = ['address', 'undefined'] # Languages from project.yaml that have code coverage support. LANGUAGES_WITH_COVERAGE_SUPPORT = ['c', 'c++', 'go'] def get_changed_files_output(): """Returns the output of a git command that discovers changed files.""" branch_commit_hash = subprocess.check_output( ['git', 'merge-base', 'FETCH_HEAD', 'origin/HEAD']).strip().decode() return subprocess.check_output( ['git', 'diff', '--name-only', branch_commit_hash + '..']).decode() def get_modified_buildable_projects(): """Returns a list of all the projects modified in this commit that have a build.sh file.""" git_output = get_changed_files_output() projects_regex = '.*projects/(?P<name>.*)/.*\n' modified_projects = set(re.findall(projects_regex, git_output)) projects_dir = os.path.join(get_oss_fuzz_root(), 'projects') # Filter out projects without Dockerfile files since new projects and reverted # projects frequently don't have them. In these cases we don't want Travis's # builds to fail. modified_buildable_projects = [] for project in modified_projects: if not os.path.exists(os.path.join(projects_dir, project, 'Dockerfile')): print('Project {0} does not have Dockerfile. skipping build.'.format( project)) continue modified_buildable_projects.append(project) return modified_buildable_projects def get_oss_fuzz_root(): """Get the absolute path of the root of the oss-fuzz checkout.""" script_path = os.path.realpath(__file__) return os.path.abspath( os.path.dirname(os.path.dirname(os.path.dirname(script_path)))) def execute_helper_command(helper_command): """Execute |helper_command| using helper.py.""" root = get_oss_fuzz_root() script_path = os.path.join(root, 'infra', 'helper.py') command = ['python', script_path] + helper_command print('Running command: %s' % ' '.join(command)) subprocess.check_call(command) def build_fuzzers(project, engine, sanitizer, architecture): """Execute helper.py's build_fuzzers command on |project|. Build the fuzzers with |engine| and |sanitizer| for |architecture|.""" execute_helper_command([ 'build_fuzzers', project, '--engine', engine, '--sanitizer', sanitizer, '--architecture', architecture ]) def check_build(project, engine, sanitizer, architecture): """Execute helper.py's check_build command on |project|, assuming it was most recently built with |engine| and |sanitizer| for |architecture|.""" execute_helper_command([ 'check_build', project, '--engine', engine, '--sanitizer', sanitizer, '--architecture', architecture ]) def should_build_coverage(project_yaml): """Returns True if a coverage build should be done based on project.yaml contents.""" # Enable coverage builds on projects that use engines. Those that don't use # engines shouldn't get coverage builds. engines = project_yaml.get('fuzzing_engines', DEFAULT_ENGINES) engineless = 'none' in engines if engineless: assert_message = ('Forbidden to specify multiple engines for ' '"fuzzing_engines" if "none" is specified.') assert len(engines) == 1, assert_message return False language = project_yaml.get('language') if language not in LANGUAGES_WITH_COVERAGE_SUPPORT: print(('Project is written in "{language}", ' 'coverage is not supported yet.').format(language=language)) return False return True def should_build(project_yaml): """Returns True on if the build specified is enabled in the project.yaml.""" if os.getenv('SANITIZER') == 'coverage': # This assumes we only do coverage builds with libFuzzer on x86_64. return should_build_coverage(project_yaml) def is_enabled(env_var, yaml_name, defaults): """Is the value of |env_var| enabled in |project_yaml| (in the |yaml_name| section)? Uses |defaults| if |yaml_name| section is unspecified.""" return os.getenv(env_var) in project_yaml.get(yaml_name, defaults) return (is_enabled('ENGINE', 'fuzzing_engines', DEFAULT_ENGINES) and is_enabled('SANITIZER', 'sanitizers', DEFAULT_SANITIZERS) and is_enabled('ARCHITECTURE', 'architectures', DEFAULT_ARCHITECTURES)) def build_project(project): """Do the build of |project| that is specified by the environment variables - SANITIZER, ENGINE, and ARCHITECTURE.""" root = get_oss_fuzz_root() project_yaml_path = os.path.join(root, 'projects', project, 'project.yaml') with open(project_yaml_path) as file_handle: project_yaml = yaml.safe_load(file_handle) if project_yaml.get('disabled', False): print('Project {0} is disabled, skipping build.'.format(project)) return engine = os.getenv('ENGINE') sanitizer = os.getenv('SANITIZER') architecture = os.getenv('ARCHITECTURE') if not should_build(project_yaml): print(('Specified build: engine: {0}, sanitizer: {1}, architecture: {2} ' 'not enabled for this project: {3}. Skipping build.').format( engine, sanitizer, architecture, project)) return print('Building project', project) build_fuzzers(project, engine, sanitizer, architecture) if engine != 'none' and sanitizer != 'coverage': check_build(project, engine, sanitizer, architecture) class BuildModifiedProjectsResult(enum.Enum): """Enum containing the return values of build_modified_projects().""" NONE_BUILT = 0 BUILD_SUCCESS = 1 BUILD_FAIL = 2 def build_modified_projects(): """Build modified projects. Returns BuildModifiedProjectsResult.NONE_BUILT if no builds were attempted. Returns BuildModifiedProjectsResult.BUILD_SUCCESS if all attempts succeed, otherwise returns BuildModifiedProjectsResult.BUILD_FAIL.""" projects = get_modified_buildable_projects() if not projects: return BuildModifiedProjectsResult.NONE_BUILT failed_projects = [] for project in projects: try: build_project(project) except subprocess.CalledProcessError: failed_projects.append(project) if failed_projects: print('Failed projects:', ' '.join(failed_projects)) return BuildModifiedProjectsResult.BUILD_FAIL return BuildModifiedProjectsResult.BUILD_SUCCESS def is_infra_changed(): """Returns True if the infra directory was changed.""" git_output = get_changed_files_output() infra_code_regex = '.*infra/.*\n' return re.search(infra_code_regex, git_output) is not None def build_base_images(): """Builds base images.""" # TODO(jonathanmetzman): Investigate why caching fails so often and # when we improve it, build base-clang as well. Also, move this function # to a helper command when we can support base-clang. execute_helper_command(['pull_images']) images = [ 'base-image', 'base-builder', 'base-runner', ] for image in images: try: execute_helper_command(['build_image', image, '--no-pull']) except subprocess.CalledProcessError: return 1 return 0 def build_canary_project(): """Builds a specific project when infra/ is changed to verify that infra/ changes don't break things. Returns False if build was attempted but failed.""" try: build_project('skcms') except subprocess.CalledProcessError: return False return True def main(): """Build modified projects or canary project.""" infra_changed = is_infra_changed() if infra_changed: print('Pulling and building base images first.') if build_base_images(): return 1 result = build_modified_projects() if result == BuildModifiedProjectsResult.BUILD_FAIL: return 1 # It's unnecessary to build the canary if we've built any projects already. no_projects_built = result == BuildModifiedProjectsResult.NONE_BUILT should_build_canary = no_projects_built and infra_changed if should_build_canary and not build_canary_project(): return 1 return 0 if __name__ == '__main__': sys.exit(main())
33.714829
80
0.721101
is_enabled('SANITIZER', 'sanitizers', DEFAULT_SANITIZERS) and is_enabled('ARCHITECTURE', 'architectures', DEFAULT_ARCHITECTURES)) def build_project(project): root = get_oss_fuzz_root() project_yaml_path = os.path.join(root, 'projects', project, 'project.yaml') with open(project_yaml_path) as file_handle: project_yaml = yaml.safe_load(file_handle) if project_yaml.get('disabled', False): print('Project {0} is disabled, skipping build.'.format(project)) return engine = os.getenv('ENGINE') sanitizer = os.getenv('SANITIZER') architecture = os.getenv('ARCHITECTURE') if not should_build(project_yaml): print(('Specified build: engine: {0}, sanitizer: {1}, architecture: {2} ' 'not enabled for this project: {3}. Skipping build.').format( engine, sanitizer, architecture, project)) return print('Building project', project) build_fuzzers(project, engine, sanitizer, architecture) if engine != 'none' and sanitizer != 'coverage': check_build(project, engine, sanitizer, architecture) class BuildModifiedProjectsResult(enum.Enum): NONE_BUILT = 0 BUILD_SUCCESS = 1 BUILD_FAIL = 2 def build_modified_projects(): projects = get_modified_buildable_projects() if not projects: return BuildModifiedProjectsResult.NONE_BUILT failed_projects = [] for project in projects: try: build_project(project) except subprocess.CalledProcessError: failed_projects.append(project) if failed_projects: print('Failed projects:', ' '.join(failed_projects)) return BuildModifiedProjectsResult.BUILD_FAIL return BuildModifiedProjectsResult.BUILD_SUCCESS def is_infra_changed(): git_output = get_changed_files_output() infra_code_regex = '.*infra/.*\n' return re.search(infra_code_regex, git_output) is not None def build_base_images(): # TODO(jonathanmetzman): Investigate why caching fails so often and # when we improve it, build base-clang as well. Also, move this function # to a helper command when we can support base-clang. execute_helper_command(['pull_images']) images = [ 'base-image', 'base-builder', 'base-runner', ] for image in images: try: execute_helper_command(['build_image', image, '--no-pull']) except subprocess.CalledProcessError: return 1 return 0 def build_canary_project(): try: build_project('skcms') except subprocess.CalledProcessError: return False return True def main(): infra_changed = is_infra_changed() if infra_changed: print('Pulling and building base images first.') if build_base_images(): return 1 result = build_modified_projects() if result == BuildModifiedProjectsResult.BUILD_FAIL: return 1 # It's unnecessary to build the canary if we've built any projects already. no_projects_built = result == BuildModifiedProjectsResult.NONE_BUILT should_build_canary = no_projects_built and infra_changed if should_build_canary and not build_canary_project(): return 1 return 0 if __name__ == '__main__': sys.exit(main())
true
true
f7179a9aff5fcbff9e7d9219b960c35806a5e4f8
2,214
py
Python
generate_embeddings.py
yangfengKAUST/cnn-text-classification-tf
5f552df9887e57a4bc5638b3d36d7393254d2644
[ "Apache-2.0" ]
null
null
null
generate_embeddings.py
yangfengKAUST/cnn-text-classification-tf
5f552df9887e57a4bc5638b3d36d7393254d2644
[ "Apache-2.0" ]
null
null
null
generate_embeddings.py
yangfengKAUST/cnn-text-classification-tf
5f552df9887e57a4bc5638b3d36d7393254d2644
[ "Apache-2.0" ]
null
null
null
import numpy as np import pickle import argparse import re """ Convert pre-trained Glove embeddings into npy file Run using: python3 generate_embeddings.py -d data/glove.6B.300d.txt --npy_output data/embeddings.npy --dict_output data/vocab.pckl --dict_whitelist data/polaritydata.vocab """ def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('--dataset', '-d', type=str, required=True) parser.add_argument('--npy_output', type=str, required=True) parser.add_argument('--dict_output', type=str, required=True) parser.add_argument('--dict_whitelist', type=str, required=True) parser.add_argument('--dump_frequency', type=int, default=10000) return parser.parse_args() def main(): args = parse_args() # reserve 0 for unknown words data = { '': 0 } embeddings = [ np.zeros((300), dtype=np.float32) ] float_re = re.compile(' [-+]?[0-9]*\.?[0-9]+([eE][-+]?[0-9]+)?') with open(args.dict_whitelist) as wfile: whitelist = [line.strip() for line in wfile] print("Building vocabulary ...") with open(args.dataset) as ofile, \ open(args.dict_output, 'wb') as dfile, \ open(args.npy_output, 'wb') as nfile: idx = 1 for line in ofile: pos = next(re.finditer(float_re, line)).start() word, vector = line[:pos], line[pos + 1:].split() if word not in whitelist: continue if word in data: print('Possible duplicate at {} in {}'.format(idx, line)) continue embedding = np.fromiter([float(d) for d in vector], np.float32) if embedding.shape != (300,): print('Shape is {}'.format(embedding.shape)) print(line) embeddings.append(embedding) data[word] = idx idx += 1 if not idx % args.dump_frequency: np.save(nfile, np.array(embeddings)) embeddings.clear() np.save(nfile, np.array(embeddings)) pickle.dump(data, dfile) print("Vocabulary saved, size is {} words".format(idx)) if __name__ == '__main__': main()
29.131579
160
0.590786
import numpy as np import pickle import argparse import re def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('--dataset', '-d', type=str, required=True) parser.add_argument('--npy_output', type=str, required=True) parser.add_argument('--dict_output', type=str, required=True) parser.add_argument('--dict_whitelist', type=str, required=True) parser.add_argument('--dump_frequency', type=int, default=10000) return parser.parse_args() def main(): args = parse_args() data = { '': 0 } embeddings = [ np.zeros((300), dtype=np.float32) ] float_re = re.compile(' [-+]?[0-9]*\.?[0-9]+([eE][-+]?[0-9]+)?') with open(args.dict_whitelist) as wfile: whitelist = [line.strip() for line in wfile] print("Building vocabulary ...") with open(args.dataset) as ofile, \ open(args.dict_output, 'wb') as dfile, \ open(args.npy_output, 'wb') as nfile: idx = 1 for line in ofile: pos = next(re.finditer(float_re, line)).start() word, vector = line[:pos], line[pos + 1:].split() if word not in whitelist: continue if word in data: print('Possible duplicate at {} in {}'.format(idx, line)) continue embedding = np.fromiter([float(d) for d in vector], np.float32) if embedding.shape != (300,): print('Shape is {}'.format(embedding.shape)) print(line) embeddings.append(embedding) data[word] = idx idx += 1 if not idx % args.dump_frequency: np.save(nfile, np.array(embeddings)) embeddings.clear() np.save(nfile, np.array(embeddings)) pickle.dump(data, dfile) print("Vocabulary saved, size is {} words".format(idx)) if __name__ == '__main__': main()
true
true
f7179ad7cc757ce856fb3920524110f510a8a74a
2,256
py
Python
SWIM-Executables/Windows/pyinstaller-2.0 for windows/PyInstaller/hooks/django-import-finder.py
alexsigaras/SWIM
1a35df8acb26bdcb307a1b8f60e9feba68ed1715
[ "MIT" ]
47
2020-03-08T08:43:28.000Z
2022-03-18T18:51:55.000Z
SWIM-Executables/Windows/pyinstaller-2.0 for windows/PyInstaller/hooks/django-import-finder.py
alexsigaras/SWIM
1a35df8acb26bdcb307a1b8f60e9feba68ed1715
[ "MIT" ]
null
null
null
SWIM-Executables/Windows/pyinstaller-2.0 for windows/PyInstaller/hooks/django-import-finder.py
alexsigaras/SWIM
1a35df8acb26bdcb307a1b8f60e9feba68ed1715
[ "MIT" ]
16
2020-03-08T08:43:30.000Z
2022-01-10T22:05:57.000Z
# Copyright (C) 2009, Lorenzo Berni # Based on previous work under copyright (c) 2001, 2002 McMillan Enterprises, Inc. # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA import os import PyInstaller.compat as compat from hookutils import logger if not compat.getenv("DJANGO_SETTINGS_MODULE"): compat.setenv("DJANGO_SETTINGS_MODULE", "settings") from django.conf import settings hiddenimports = (list(settings.AUTHENTICATION_BACKENDS) + [settings.DEFAULT_FILE_STORAGE] + list(settings.FILE_UPLOAD_HANDLERS) + list(settings.INSTALLED_APPS) + list(settings.MIDDLEWARE_CLASSES) + list(settings.TEMPLATE_CONTEXT_PROCESSORS) + list(settings.TEMPLATE_LOADERS) + [settings.ROOT_URLCONF]) def find_url_callbacks(urls_module): urlpatterns = urls_module.urlpatterns hid_list = [urls_module.__name__] for pattern in urlpatterns: if isinstance(pattern, RegexURLPattern): hid_list.append(pattern.callback.__module__) elif isinstance(pattern, RegexURLResolver): hid_list += find_url_callbacks(pattern.urlconf_module) return hid_list from django.core.urlresolvers import RegexURLPattern, RegexURLResolver base_module_name = ".".join(compat.getenv("DJANGO_SETTINGS_MODULE", "settings").split(".")[:-1]) if base_module_name: base_module = __import__(base_module_name, {}, {}, ["urls"]) urls = base_module.urls else: import urls hiddenimports += find_url_callbacks(urls) logger.debug('%r', sorted(set(hiddenimports)))
38.237288
96
0.724734
import os import PyInstaller.compat as compat from hookutils import logger if not compat.getenv("DJANGO_SETTINGS_MODULE"): compat.setenv("DJANGO_SETTINGS_MODULE", "settings") from django.conf import settings hiddenimports = (list(settings.AUTHENTICATION_BACKENDS) + [settings.DEFAULT_FILE_STORAGE] + list(settings.FILE_UPLOAD_HANDLERS) + list(settings.INSTALLED_APPS) + list(settings.MIDDLEWARE_CLASSES) + list(settings.TEMPLATE_CONTEXT_PROCESSORS) + list(settings.TEMPLATE_LOADERS) + [settings.ROOT_URLCONF]) def find_url_callbacks(urls_module): urlpatterns = urls_module.urlpatterns hid_list = [urls_module.__name__] for pattern in urlpatterns: if isinstance(pattern, RegexURLPattern): hid_list.append(pattern.callback.__module__) elif isinstance(pattern, RegexURLResolver): hid_list += find_url_callbacks(pattern.urlconf_module) return hid_list from django.core.urlresolvers import RegexURLPattern, RegexURLResolver base_module_name = ".".join(compat.getenv("DJANGO_SETTINGS_MODULE", "settings").split(".")[:-1]) if base_module_name: base_module = __import__(base_module_name, {}, {}, ["urls"]) urls = base_module.urls else: import urls hiddenimports += find_url_callbacks(urls) logger.debug('%r', sorted(set(hiddenimports)))
true
true
f7179b68d16205b5ace838040622bc3f19f99fa8
10,342
py
Python
src/.history/Test/HiwinRT605_test_20190619131630.py
SamKaiYang/2019_Hiwin_Shaking
d599f8c87dc4da89eae266990d12eb3a8b0f3e16
[ "MIT" ]
null
null
null
src/.history/Test/HiwinRT605_test_20190619131630.py
SamKaiYang/2019_Hiwin_Shaking
d599f8c87dc4da89eae266990d12eb3a8b0f3e16
[ "MIT" ]
null
null
null
src/.history/Test/HiwinRT605_test_20190619131630.py
SamKaiYang/2019_Hiwin_Shaking
d599f8c87dc4da89eae266990d12eb3a8b0f3e16
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # license removed for brevity #策略 機械手臂 四點來回跑 import rospy import os import numpy as np from std_msgs.msg import String from ROS_Socket.srv import * from ROS_Socket.msg import * import math import enum import Hiwin_RT605_ROS as ArmTask pos_feedback_times = 0 mode_feedback_times = 0 msg_feedback = 1 Arm_state_flag = 0 Strategy_flag = 0 arm_move_times = 1 ##-----------server feedback arm state---------- def Arm_state(req): global CurrentMissionType,Strategy_flag,Arm_state_flag Arm_state_flag = int('%s'%req.Arm_state) if Arm_state_flag == 1: #表示手臂忙碌 Strategy_flag = 0 return(1) if Arm_state_flag == 0: #表示手臂準備 Strategy_flag = 1 return(0) if Arm_state_flag == 6: #表示程式中斷 Strategy_flag= 6 return(6) def strategy_server(): #rospy.init_node(NAME) s = rospy.Service('arm_state',arm_state, Arm_state) ##server arm state #rospy.spin() ## spin one ##-----------switch define------------## class switch(object): def __init__(self, value): self.value = value self.fall = False def __iter__(self): """Return the match method once, then stop""" yield self.match raise StopIteration def match(self, *args): """Indicate whether or not to enter a case suite""" if self.fall or not args: return True elif self.value in args: # changed for v1.5, see below self.fall = True return True else: return False ##------------class------- class point(): def __init__(self,x,y,z,pitch,roll,yaw): self.x = x self.y = y self.z = z self.pitch = pitch self.roll = roll self.yaw = yaw ##-------------------------strategy--------------------- ##-----Mission 參數 GetInfoFlag = False ExecuteFlag = False GetKeyFlag = False MotionSerialKey = [] MissionType_Flag = 0 MotionStep = 0 ##-----手臂動作位置資訊 angle_SubCue = 0 LinePtpFlag = False MoveFlag = False PushBallHeight = 6 ObjAboveHeight = 10 SpeedValue = 10 MissionEndFlag = False CurrentMissionType = 0 ##---------------Enum---------------## class ArmMotionCommand(enum.IntEnum): Arm_Stop = 0 Arm_MoveToTargetUpside = 1 Arm_MoveFowardDown = 2 Arm_MoveVision = 3 Arm_PushBall = 4 Arm_LineUp = 5 Arm_LineDown = 6 Arm_Angle = 7 Arm_StopPush = 8 class MissionType(enum.IntEnum): Get_Img = 0 PushBall = 1 Pushback = 2 Mission_End = 3 ##-----------switch define------------## class pos(): def __init__(self, x, y, z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class Target_pos(): def __init__(self, x, y, z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class TargetPush_pos(): def __init__(self, x, y, z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class Item(): def __init__(self,x,y,label): self.x = x self.y = y self.label = label def Mission_Trigger(): if GetInfoFlag == True and GetKeyFlag == False and ExecuteFlag == False: GetInfo_Mission() if GetInfoFlag == False and GetKeyFlag == True and ExecuteFlag == False: GetKey_Mission() if GetInfoFlag == False and GetKeyFlag == False and ExecuteFlag == True: Execute_Mission() def GetInfo_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag #Billiards_Calculation() GetInfoFlag = False GetKeyFlag = True ExecuteFlag = False def GetKey_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag,MotionKey,MotionSerialKey Mission = Get_MissionType() MissionItem(Mission) MotionSerialKey = MotionKey GetInfoFlag = False GetKeyFlag = False ExecuteFlag = True def Get_MissionType(): global MissionType_Flag,CurrentMissionType for case in switch(MissionType_Flag): #傳送指令給socket選擇手臂動作 if case(0): Type = MissionType.PushBall MissionType_Flag +=1 break if case(1): Type = MissionType.Pushback MissionType_Flag -=1 break CurrentMissionType = Type return Type def MissionItem(ItemNo): global MotionKey Key_PushBallCommand = [\ ArmMotionCommand.Arm_MoveToTargetUpside,\ ArmMotionCommand.Arm_LineDown,\ ArmMotionCommand.Arm_PushBall,\ ArmMotionCommand.Arm_LineUp,\ ArmMotionCommand.Arm_Stop,\ ] Key_PushBackCommand = [\ ArmMotionCommand.Arm_MoveVision,\ ArmMotionCommand.Arm_Stop,\ ArmMotionCommand.Arm_StopPush,\ ] for case in switch(ItemNo): #傳送指令給socket選擇手臂動作 if case(MissionType.PushBall): MotionKey = Key_PushBallCommand break if case(MissionType.Pushback): MotionKey = Key_PushBackCommand break return MotionKey def Execute_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag,MotionKey,MotionStep,MotionSerialKey,MissionEndFlag,CurrentMissionType,Strategy_flag,Arm_state_flag # print("strategy :" ,Strategy_flag) # print("state :" ,Arm_state_flag) # if Arm_state_flag == 0 and Strategy_flag == 1: print(11111) # Strategy_flag = 0 if ArmTask.state_flag.Arm_feedback == 0 and ArmTask.state_flag.Strategy_feedback == 1: print(11111) ArmTask.state_flag.Strategy_feedback = 0 if MotionKey[MotionStep] == ArmMotionCommand.Arm_Stop: if MissionEndFlag == True: CurrentMissionType = MissionType.Mission_End GetInfoFlag = False GetKeyFlag = False ExecuteFlag = False print("Mission_End") elif CurrentMissionType == MissionType.PushBall: GetInfoFlag = False GetKeyFlag = True ExecuteFlag = False MotionStep = 0 print("PushBall") else: GetInfoFlag = True GetKeyFlag = False ExecuteFlag = False MotionStep = 0 else: MotionItem(MotionSerialKey[MotionStep]) MotionStep += 1 def MotionItem(ItemNo): global angle_SubCue,SpeedValue,PushFlag,LinePtpFlag,MissionEndFlag SpeedValue = 5 for case in switch(ItemNo): #傳送指令給socket選擇手臂動作 if case(ArmMotionCommand.Arm_Stop): MoveFlag = False print("Arm_Stop") break if case(ArmMotionCommand.Arm_StopPush): MoveFlag = False PushFlag = True #重新掃描物件 print("Arm_StopPush") break if case(ArmMotionCommand.Arm_MoveToTargetUpside): pos.x = 10 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 10 MoveFlag = True LinePtpFlag = False SpeedValue = 10 print("Arm_MoveToTargetUpside") break if case(ArmMotionCommand.Arm_LineUp): pos.z = ObjAboveHeight MoveFlag = True LinePtpFlag = True SpeedValue = 5 print("Arm_LineUp") break if case(ArmMotionCommand.Arm_LineDown): pos.z = PushBallHeight MoveFlag = True LinePtpFlag = True SpeedValue = 5 print("Arm_LineDown") break if case(ArmMotionCommand.Arm_PushBall): pos.x = -10 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = -10 SpeedValue = 10 ##待測試up MoveFlag = True LinePtpFlag = False print("Arm_PushBall") break if case(ArmMotionCommand.Arm_MoveVision): pos.x = 0 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 0 SpeedValue = 10 MoveFlag = True LinePtpFlag = False ##任務結束旗標 MissionEndFlag = True print("Arm_MoveVision") break if case(ArmMotionCommand.Arm_MoveFowardDown): pos.x = 0 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 0 MoveFlag = True LinePtpFlag = False print("Arm_MoveFowardDown") break if case(): # default, could also just omit condition or 'if True' print ("something else!") # No need to break here, it'll stop anyway if MoveFlag == True: if LinePtpFlag == False: print('x: ',pos.x,' y: ',pos.y,' z: ',pos.z,' pitch: ',pos.pitch,' roll: ',pos.roll,' yaw: ',pos.yaw) #strategy_client_Arm_Mode(0,1,0,30,2)#action,ra,grip,vel,both ArmTask.strategy_client_Arm_Mode(2,1,0,SpeedValue,2)#action,ra,grip,vel,both ArmTask.strategy_client_pos_move(pos.x,pos.y,pos.z,pos.pitch,pos.roll,pos.yaw) elif LinePtpFlag == True: #strategy_client_Arm_Mode(0,1,0,40,2)#action,ra,grip,vel,both print('x: ',pos.x,' y: ',pos.y,' z: ',pos.z,' pitch: ',pos.pitch,' roll: ',pos.roll,' yaw: ',pos.yaw) ArmTask.strategy_client_Arm_Mode(3,1,0,SpeedValue,2)#action,ra,grip,vel,both ArmTask.strategy_client_pos_move(pos.x,pos.y,pos.z,pos.pitch,pos.roll,pos.yaw) #action: ptp line #ra : abs rel #grip 夾爪 #vel speed #both : Ctrl_Mode ##-------------strategy end ------------ def myhook(): print ("shutdown time!") if __name__ == '__main__': argv = rospy.myargv() rospy.init_node('strategy', anonymous=True) GetInfoFlag = True #Test no data #strategy_server() ArmTask.strategy_client_Arm_Mode(0,1,0,20,2)#action,ra,grip,vel,both while 1: Mission_Trigger() if CurrentMissionType == MissionType.Mission_End: ArmTask.rospy.on_shutdown(myhook) ArmTask.rospy.spin() rospy.spin()
30.871642
145
0.579288
import rospy import os import numpy as np from std_msgs.msg import String from ROS_Socket.srv import * from ROS_Socket.msg import * import math import enum import Hiwin_RT605_ROS as ArmTask pos_feedback_times = 0 mode_feedback_times = 0 msg_feedback = 1 Arm_state_flag = 0 Strategy_flag = 0 arm_move_times = 1 ype,Strategy_flag,Arm_state_flag Arm_state_flag = int('%s'%req.Arm_state) if Arm_state_flag == 1: Strategy_flag = 0 return(1) if Arm_state_flag == 0: Strategy_flag = 1 return(0) if Arm_state_flag == 6: Strategy_flag= 6 return(6) def strategy_server(): s = rospy.Service('arm_state',arm_state, Arm_state) = value self.fall = False def __iter__(self): yield self.match raise StopIteration def match(self, *args): if self.fall or not args: return True elif self.value in args: self.fall = True return True else: return False _init__(self,x,y,z,pitch,roll,yaw): self.x = x self.y = y self.z = z self.pitch = pitch self.roll = roll self.yaw = yaw alKey = [] MissionType_Flag = 0 MotionStep = 0 = 0 LinePtpFlag = False MoveFlag = False PushBallHeight = 6 ObjAboveHeight = 10 SpeedValue = 10 MissionEndFlag = False CurrentMissionType = 0 Arm_Stop = 0 Arm_MoveToTargetUpside = 1 Arm_MoveFowardDown = 2 Arm_MoveVision = 3 Arm_PushBall = 4 Arm_LineUp = 5 Arm_LineDown = 6 Arm_Angle = 7 Arm_StopPush = 8 class MissionType(enum.IntEnum): Get_Img = 0 PushBall = 1 Pushback = 2 Mission_End = 3 z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class Target_pos(): def __init__(self, x, y, z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class TargetPush_pos(): def __init__(self, x, y, z, pitch, roll, yaw): self.x = 0 self.y = 36.8 self.z = 11.35 self.pitch = -90 self.roll = 0 self.yaw = 0 class Item(): def __init__(self,x,y,label): self.x = x self.y = y self.label = label def Mission_Trigger(): if GetInfoFlag == True and GetKeyFlag == False and ExecuteFlag == False: GetInfo_Mission() if GetInfoFlag == False and GetKeyFlag == True and ExecuteFlag == False: GetKey_Mission() if GetInfoFlag == False and GetKeyFlag == False and ExecuteFlag == True: Execute_Mission() def GetInfo_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag GetInfoFlag = False GetKeyFlag = True ExecuteFlag = False def GetKey_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag,MotionKey,MotionSerialKey Mission = Get_MissionType() MissionItem(Mission) MotionSerialKey = MotionKey GetInfoFlag = False GetKeyFlag = False ExecuteFlag = True def Get_MissionType(): global MissionType_Flag,CurrentMissionType for case in switch(MissionType_Flag): if case(0): Type = MissionType.PushBall MissionType_Flag +=1 break if case(1): Type = MissionType.Pushback MissionType_Flag -=1 break CurrentMissionType = Type return Type def MissionItem(ItemNo): global MotionKey Key_PushBallCommand = [\ ArmMotionCommand.Arm_MoveToTargetUpside,\ ArmMotionCommand.Arm_LineDown,\ ArmMotionCommand.Arm_PushBall,\ ArmMotionCommand.Arm_LineUp,\ ArmMotionCommand.Arm_Stop,\ ] Key_PushBackCommand = [\ ArmMotionCommand.Arm_MoveVision,\ ArmMotionCommand.Arm_Stop,\ ArmMotionCommand.Arm_StopPush,\ ] for case in switch(ItemNo): if case(MissionType.PushBall): MotionKey = Key_PushBallCommand break if case(MissionType.Pushback): MotionKey = Key_PushBackCommand break return MotionKey def Execute_Mission(): global GetInfoFlag,GetKeyFlag,ExecuteFlag,MotionKey,MotionStep,MotionSerialKey,MissionEndFlag,CurrentMissionType,Strategy_flag,Arm_state_flag print(11111) if ArmTask.state_flag.Arm_feedback == 0 and ArmTask.state_flag.Strategy_feedback == 1: print(11111) ArmTask.state_flag.Strategy_feedback = 0 if MotionKey[MotionStep] == ArmMotionCommand.Arm_Stop: if MissionEndFlag == True: CurrentMissionType = MissionType.Mission_End GetInfoFlag = False GetKeyFlag = False ExecuteFlag = False print("Mission_End") elif CurrentMissionType == MissionType.PushBall: GetInfoFlag = False GetKeyFlag = True ExecuteFlag = False MotionStep = 0 print("PushBall") else: GetInfoFlag = True GetKeyFlag = False ExecuteFlag = False MotionStep = 0 else: MotionItem(MotionSerialKey[MotionStep]) MotionStep += 1 def MotionItem(ItemNo): global angle_SubCue,SpeedValue,PushFlag,LinePtpFlag,MissionEndFlag SpeedValue = 5 for case in switch(ItemNo): if case(ArmMotionCommand.Arm_Stop): MoveFlag = False print("Arm_Stop") break if case(ArmMotionCommand.Arm_StopPush): MoveFlag = False PushFlag = True print("Arm_StopPush") break if case(ArmMotionCommand.Arm_MoveToTargetUpside): pos.x = 10 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 10 MoveFlag = True LinePtpFlag = False SpeedValue = 10 print("Arm_MoveToTargetUpside") break if case(ArmMotionCommand.Arm_LineUp): pos.z = ObjAboveHeight MoveFlag = True LinePtpFlag = True SpeedValue = 5 print("Arm_LineUp") break if case(ArmMotionCommand.Arm_LineDown): pos.z = PushBallHeight MoveFlag = True LinePtpFlag = True SpeedValue = 5 print("Arm_LineDown") break if case(ArmMotionCommand.Arm_PushBall): pos.x = -10 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = -10 SpeedValue = 10 MoveFlag = True LinePtpFlag = False print("Arm_PushBall") break if case(ArmMotionCommand.Arm_MoveVision): pos.x = 0 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 0 SpeedValue = 10 MoveFlag = True LinePtpFlag = False MissionEndFlag = True print("Arm_MoveVision") break if case(ArmMotionCommand.Arm_MoveFowardDown): pos.x = 0 pos.y = 36.8 pos.z = 11.35 pos.pitch = -90 pos.roll = 0 pos.yaw = 0 MoveFlag = True LinePtpFlag = False print("Arm_MoveFowardDown") break if case(): print ("something else!") if MoveFlag == True: if LinePtpFlag == False: print('x: ',pos.x,' y: ',pos.y,' z: ',pos.z,' pitch: ',pos.pitch,' roll: ',pos.roll,' yaw: ',pos.yaw) #strategy_client_Arm_Mode(0,1,0,30,2)#action,ra,grip,vel,both ArmTask.strategy_client_Arm_Mode(2,1,0,SpeedValue,2)#action,ra,grip,vel,both ArmTask.strategy_client_pos_move(pos.x,pos.y,pos.z,pos.pitch,pos.roll,pos.yaw) elif LinePtpFlag == True: #strategy_client_Arm_Mode(0,1,0,40,2)#action,ra,grip,vel,both print('x: ',pos.x,' y: ',pos.y,' z: ',pos.z,' pitch: ',pos.pitch,' roll: ',pos.roll,' yaw: ',pos.yaw) ArmTask.strategy_client_Arm_Mode(3,1,0,SpeedValue,2)#action,ra,grip,vel,both ArmTask.strategy_client_pos_move(pos.x,pos.y,pos.z,pos.pitch,pos.roll,pos.yaw) #action: ptp line #ra : abs rel #grip 夾爪 #vel speed #both : Ctrl_Mode ##-------------strategy end ------------ def myhook(): print ("shutdown time!") if __name__ == '__main__': argv = rospy.myargv() rospy.init_node('strategy', anonymous=True) GetInfoFlag = True #Test no data #strategy_server() ArmTask.strategy_client_Arm_Mode(0,1,0,20,2)#action,ra,grip,vel,both while 1: Mission_Trigger() if CurrentMissionType == MissionType.Mission_End: ArmTask.rospy.on_shutdown(myhook) ArmTask.rospy.spin() rospy.spin()
true
true
f7179b7c95704c49d4aaa9e728e5e82df649b8d6
596
py
Python
configs/revbifpn/fcn_s1_512x1024_120k_cityscapes.py
gupta-abhay/mmsegmentation
03e80db237d12b4d049fe65c6612f2bc741f8135
[ "Apache-2.0" ]
null
null
null
configs/revbifpn/fcn_s1_512x1024_120k_cityscapes.py
gupta-abhay/mmsegmentation
03e80db237d12b4d049fe65c6612f2bc741f8135
[ "Apache-2.0" ]
null
null
null
configs/revbifpn/fcn_s1_512x1024_120k_cityscapes.py
gupta-abhay/mmsegmentation
03e80db237d12b4d049fe65c6612f2bc741f8135
[ "Apache-2.0" ]
null
null
null
_base_ = [ '../_base_/models/fcn_hr18.py', '../_base_/datasets/cityscapes.py', '../_base_/default_runtime.py', '../_base_/schedules/schedule_120k.py', ] model = dict( pretrained=None, backbone=dict( type='RevBiFPN_S', arch='revbifpn_s1', strict=False, classes=None, _norm="syncbn", init_cfg=dict( type='Pretrained', ), ), decode_head=dict( in_channels=[64, 80, 160, 416], channels=sum([64, 80, 160, 416]) ), ) optimizer = dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=3.0e-5)
23.84
72
0.567114
_base_ = [ '../_base_/models/fcn_hr18.py', '../_base_/datasets/cityscapes.py', '../_base_/default_runtime.py', '../_base_/schedules/schedule_120k.py', ] model = dict( pretrained=None, backbone=dict( type='RevBiFPN_S', arch='revbifpn_s1', strict=False, classes=None, _norm="syncbn", init_cfg=dict( type='Pretrained', ), ), decode_head=dict( in_channels=[64, 80, 160, 416], channels=sum([64, 80, 160, 416]) ), ) optimizer = dict(type='SGD', lr=0.01, momentum=0.9, weight_decay=3.0e-5)
true
true
f7179e36f8aee9aeb3010facd15317bf9a7987b3
1,075
py
Python
fewshot/data/compress_tiered_imagenet.py
renmengye/inc-few-shot-attractor-public
c560d5a81480cb22d903fa746ab0cfc2eb964e4c
[ "MIT" ]
122
2019-09-20T05:35:29.000Z
2022-03-22T06:30:33.000Z
fewshot/data/compress_tiered_imagenet.py
renmengye/inc-few-shot-attractor-public
c560d5a81480cb22d903fa746ab0cfc2eb964e4c
[ "MIT" ]
10
2019-11-10T10:24:06.000Z
2021-10-06T13:13:26.000Z
fewshot/data/compress_tiered_imagenet.py
renmengye/inc-few-shot-attractor-public
c560d5a81480cb22d903fa746ab0cfc2eb964e4c
[ "MIT" ]
28
2019-10-21T11:38:06.000Z
2022-01-26T16:55:28.000Z
import cv2 import numpy as np import six import sys import pickle as pkl from tqdm import tqdm def compress(path, output): with np.load(path, mmap_mode="r", encoding='latin1') as data: images = data["images"] array = [] for ii in tqdm(six.moves.xrange(images.shape[0]), desc='compress'): im = images[ii] im_str = cv2.imencode('.png', im)[1] array.append(im_str) with open(output, 'wb') as f: pkl.dump(array, f, protocol=pkl.HIGHEST_PROTOCOL) def decompress(path, output): try: with open(output, 'rb') as f: array = pkl.load(f, encoding='bytes') except: with open(output, 'rb') as f: array = pkl.load(f) images = np.zeros([len(array), 84, 84, 3], dtype=np.uint8) for ii, item in tqdm(enumerate(array), desc='decompress'): im = cv2.imdecode(item, 1) images[ii] = im np.savez(path, images=images) def main(): if sys.argv[1] == 'compress': compress(sys.argv[2], sys.argv[3]) elif sys.argv[1] == 'decompress': decompress(sys.argv[2], sys.argv[3]) if __name__ == '__main__': main()
23.888889
71
0.633488
import cv2 import numpy as np import six import sys import pickle as pkl from tqdm import tqdm def compress(path, output): with np.load(path, mmap_mode="r", encoding='latin1') as data: images = data["images"] array = [] for ii in tqdm(six.moves.xrange(images.shape[0]), desc='compress'): im = images[ii] im_str = cv2.imencode('.png', im)[1] array.append(im_str) with open(output, 'wb') as f: pkl.dump(array, f, protocol=pkl.HIGHEST_PROTOCOL) def decompress(path, output): try: with open(output, 'rb') as f: array = pkl.load(f, encoding='bytes') except: with open(output, 'rb') as f: array = pkl.load(f) images = np.zeros([len(array), 84, 84, 3], dtype=np.uint8) for ii, item in tqdm(enumerate(array), desc='decompress'): im = cv2.imdecode(item, 1) images[ii] = im np.savez(path, images=images) def main(): if sys.argv[1] == 'compress': compress(sys.argv[2], sys.argv[3]) elif sys.argv[1] == 'decompress': decompress(sys.argv[2], sys.argv[3]) if __name__ == '__main__': main()
true
true
f7179eb378b819138f77c19337426397d15ce820
10,249
py
Python
rl4net/envs/power_allocation/test_pa_rb_env.py
bupt-ipcr/RL4Net
b1b694361c688f5e0055148a0cdcb4c6253cd7bd
[ "MIT" ]
17
2020-03-19T02:51:48.000Z
2021-06-15T00:42:29.000Z
vvlab/envs/power_allocation/test_pa_rb_env.py
LampV/Reinforcement-Learning
0652b9e8c2de428d3508074c6fd640cc14f84a2c
[ "MIT" ]
13
2021-04-05T13:10:25.000Z
2022-03-12T00:51:15.000Z
rl4net/envs/power_allocation/test_pa_rb_env.py
bupt-ipcr/RL4Net
b1b694361c688f5e0055148a0cdcb4c6253cd7bd
[ "MIT" ]
12
2020-03-31T05:29:25.000Z
2022-03-08T11:54:07.000Z
from .pa_rb_env import ( PAEnv, Node ) import numpy as np from pathlib import Path log2 = np.log2 cues = { 0: Node(0.1, 0, 'cue'), 1: Node(-0.1, 0, 'cue'), } devices = { 0: { 't_device': Node(0, 0.5, 't_device'), 'r_devices': { 0: Node(0, 0.6, 'r_device') } }, 1: { 't_device': Node(0, -0.5, 't_device'), 'r_devices': { 0: Node(0, -0.6, 'r_device') } } } def equal(unit, target): tolerance = 1e-6 * np.ones_like(target) return (np.abs(unit - target) < tolerance).all() def test_init_pos(): """test position constraint""" env = PAEnv(n_level=4) def dis(node, target): return np.sqrt( (node.x - target.x) ** 2 + (node.y - target.y) ** 2 ) # test bs cues assert all( env.r_bs <= dis(usr, env.station) <= env.R_bs for usr in env.cues.values() ) # test devices for cluster in env.devices.values(): t_device, r_devices = cluster['t_device'], cluster['r_devices'] assert env.r_bs <= dis(t_device, env.station) <= ( env.R_bs - env.R_dev) assert all( env.r_dev <= dis(r_device, t_device) <= env.R_dev for r_device in r_devices.values() ) def test_jakes(): # TODO test stastic features of jakes # target_std, target_mean = 0.429, 1.253 # Rayleigh Distribution # x_len, y_len, Ns = H_set.shape # h_std = np.mean([ # H_set[x, y, :].std() # for x in range(x_len) # for y in range(y_len) # ]) # assert (h_std - target_std) / target_std < 0.1 # h_mean = np.mean([ # H_set[x, y, :].mean() # for x in range(x_len) # for y in range(y_len) # ]) # assert (h_mean - target_mean) / target_mean < 0.05 pass def test_init_path_loss(): """test distance, since lognormal is random""" env = PAEnv(n_level=4, n_pair=2, m_cue=2) env.cues = cues env.devices = devices env.init_path_loss() distance_matrix = env.distance_matrix target_dis = np.array( [ [0.1, 1.1, np.sqrt(0.26), np.sqrt(0.26), 0.5], [1.1, 0.1, np.sqrt(0.26), np.sqrt(0.26), 0.5], [0.6, 0.6, 0.1, 0.1, 0.503], [np.sqrt(0.37), np.sqrt(0.37), 0.503, 0.2, 0.1], [np.sqrt(0.37), np.sqrt(0.37), 0.2, 0.503, 0.1], ] ) assert equal(distance_matrix, target_dis) def test_get_recv_powers(): """test get_recv_powers""" env = PAEnv(n_level=4, n_pair=2, m_cue=1) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) target_recv_powers = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) assert equal(recv_powers, target_recv_powers) def test_get_rates(): """test get_rates""" env = PAEnv(n_level=4, n_pair=2, m_cue=1) recv_powers = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) rates = env.get_rates(recv_powers) _rate = np.array([ log2(1+1.1/31), log2(1+2.2/42), log2(1+33/1.3), log2(1+44/2.4) ]) target_rates = (_rate * np.ones((env.n_channel, env.n_channel))).T assert equal(rates, target_rates) def test_get_indices(): """test get_indices""" env = PAEnv(n_level=4, n_pair=2, m_cue=1, sorter="recv_power", m_state=2) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading # rx_indice don't need test tx_indice, rx_indice = env.get_indices(*metrics) target_tx_indice = np.array([ [3, 3, 3, 2], [0, 1, 2, 3] ]) assert equal(tx_indice, target_tx_indice) def test_get_rewards(): env = PAEnv(n_level=4, n_pair=2, m_cue=1, sorter="recv_power", m_state=2) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading indices = env.get_indices(*metrics) rewards = env.get_rewards(rates, indices) target_rewards = np.array([ log2(1+1.1/31) + log2(1+44/2.4), log2(1+2.2/42) + log2(1+44/2.4), log2(1+33/1.3) + log2(1+44/2.4), log2(1+44/2.4) + log2(1+33/1.3), ])[:2] assert equal(rewards, target_rewards) def test_get_states(): # test m_state env = PAEnv(n_level=4, n_pair=2, m_cue=1, m_state=8, metrics=['emit_power', 'recv_power', 'rate'], sorter='recv_power') assert env.m_state == 4 env = PAEnv(n_level=4, n_pair=2, m_cue=1, m_state=2, metrics=['emit_power', 'recv_power', 'rate'], sorter='recv_power') power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading indices = env.get_indices(*metrics) states = env.get_states(*metrics, indices=indices) _recv = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) _rate = np.array([ log2(1+1.1/31), log2(1+2.2/42), log2(1+33/1.3), log2(1+44/2.4) ]) target_states = np.array([ np.concatenate([power[3],power[0],_recv[3][0],_recv[0][0],[_rate[3], _rate[0]]]), np.concatenate([power[3],power[1],_recv[3][1],_recv[1][1],[_rate[3], _rate[1]]]), np.concatenate([power[3],power[2],_recv[3][2],_recv[2][2],[_rate[3], _rate[2]]]), np.concatenate([power[2],power[3],_recv[2][3],_recv[3][3],[_rate[2], _rate[3]]]), ])[:2] assert equal(states, target_states) def test_sorter(): # now only recv_power can be sorter pass def test_seed(): env = PAEnv(n_level=4, m_cue=1, seed=123) # this is func in PAEnv to random pos def random_point(min_r, radius, ox=0, oy=0): theta = np.random.random() * 2 * np.pi r = np.random.uniform(min_r, radius**2) x, y = np.cos(theta) * np.sqrt(r), np.sin(theta) * np.sqrt(r) return ox + x, oy + y np.random.seed(123) target_x, target_y = random_point(env.r_bs, env.R_bs) usr = env.cues[0] assert all((target_x == usr.x, target_y == usr.y)) def test_action(): env = PAEnv(n_level=10, seed=799345) n_actions = env.n_actions n_channel, n_pair = env.n_channel, env.n_pair # normal env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_channel, )) s_, r, d, i = env.step(action, unit='dBm') assert i['rate'] == 3.4741923099965257 # only D2D actions is enough env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_pair, )) s_, r, d, i = env.step(action, unit='dBm') assert i['rate'] == 3.4741923099965257 # other action dim raises error env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_pair - 1, )) try: s_, r, d, i = env.step(action, unit='dBm') except ValueError as e: msg = f"length of action should be n_channel({env.n_channel})" \ f" or n_pair({n_pair}), but is {len(action)}" assert e.args[0] == msg env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_channel, )) s_, r, d, i = env.step(action, unit='mW') assert i['rate'] == 3.4928823957853856 # TODO add test of continuous action def test_step(): env = PAEnv(n_level=10) n_actions, n_states = env.n_actions, env.n_states assert n_actions == 40 assert n_states == 304 env.reset() action = env.sample() env.step(action, unit='dBm') # action = env.sample() action = np.array([0, 0, 0, 0, 0, 0, 0, 0, 0]) env.step(action, unit='mW') action = env.sample() try: env.step(action, unit='xx') except ValueError as e: msg = f"unit should in ['dBm', 'mW'], but is xx" assert e.args[0] == msg fig: Path() = env.render() if fig.exists(): fig.unlink() if __name__ == '__main__': test_action()
30.502976
89
0.530296
from .pa_rb_env import ( PAEnv, Node ) import numpy as np from pathlib import Path log2 = np.log2 cues = { 0: Node(0.1, 0, 'cue'), 1: Node(-0.1, 0, 'cue'), } devices = { 0: { 't_device': Node(0, 0.5, 't_device'), 'r_devices': { 0: Node(0, 0.6, 'r_device') } }, 1: { 't_device': Node(0, -0.5, 't_device'), 'r_devices': { 0: Node(0, -0.6, 'r_device') } } } def equal(unit, target): tolerance = 1e-6 * np.ones_like(target) return (np.abs(unit - target) < tolerance).all() def test_init_pos(): env = PAEnv(n_level=4) def dis(node, target): return np.sqrt( (node.x - target.x) ** 2 + (node.y - target.y) ** 2 ) assert all( env.r_bs <= dis(usr, env.station) <= env.R_bs for usr in env.cues.values() ) for cluster in env.devices.values(): t_device, r_devices = cluster['t_device'], cluster['r_devices'] assert env.r_bs <= dis(t_device, env.station) <= ( env.R_bs - env.R_dev) assert all( env.r_dev <= dis(r_device, t_device) <= env.R_dev for r_device in r_devices.values() ) def test_jakes(): pass def test_init_path_loss(): env = PAEnv(n_level=4, n_pair=2, m_cue=2) env.cues = cues env.devices = devices env.init_path_loss() distance_matrix = env.distance_matrix target_dis = np.array( [ [0.1, 1.1, np.sqrt(0.26), np.sqrt(0.26), 0.5], [1.1, 0.1, np.sqrt(0.26), np.sqrt(0.26), 0.5], [0.6, 0.6, 0.1, 0.1, 0.503], [np.sqrt(0.37), np.sqrt(0.37), 0.503, 0.2, 0.1], [np.sqrt(0.37), np.sqrt(0.37), 0.2, 0.503, 0.1], ] ) assert equal(distance_matrix, target_dis) def test_get_recv_powers(): env = PAEnv(n_level=4, n_pair=2, m_cue=1) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) target_recv_powers = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) assert equal(recv_powers, target_recv_powers) def test_get_rates(): env = PAEnv(n_level=4, n_pair=2, m_cue=1) recv_powers = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) rates = env.get_rates(recv_powers) _rate = np.array([ log2(1+1.1/31), log2(1+2.2/42), log2(1+33/1.3), log2(1+44/2.4) ]) target_rates = (_rate * np.ones((env.n_channel, env.n_channel))).T assert equal(rates, target_rates) def test_get_indices(): env = PAEnv(n_level=4, n_pair=2, m_cue=1, sorter="recv_power", m_state=2) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading tx_indice, rx_indice = env.get_indices(*metrics) target_tx_indice = np.array([ [3, 3, 3, 2], [0, 1, 2, 3] ]) assert equal(tx_indice, target_tx_indice) def test_get_rewards(): env = PAEnv(n_level=4, n_pair=2, m_cue=1, sorter="recv_power", m_state=2) power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading indices = env.get_indices(*metrics) rewards = env.get_rewards(rates, indices) target_rewards = np.array([ log2(1+1.1/31) + log2(1+44/2.4), log2(1+2.2/42) + log2(1+44/2.4), log2(1+33/1.3) + log2(1+44/2.4), log2(1+44/2.4) + log2(1+33/1.3), ])[:2] assert equal(rewards, target_rewards) def test_get_states(): # test m_state env = PAEnv(n_level=4, n_pair=2, m_cue=1, m_state=8, metrics=['emit_power', 'recv_power', 'rate'], sorter='recv_power') assert env.m_state == 4 env = PAEnv(n_level=4, n_pair=2, m_cue=1, m_state=2, metrics=['emit_power', 'recv_power', 'rate'], sorter='recv_power') power = np.array([ [0.01, 0], [0, 0.01], [0.1, 0], [0, 0.1], ]) emit_powers = np.tile(np.expand_dims(power, axis=1), (1, env.n_channel, 1)) fading = np.array([ [1.1e-2, 1.2e-2, 1.3e-2, 1.4e-2], [2.1e-2, 2.2e-2, 2.3e-2, 2.4e-2], [3.1e-2, 3.2e-2, 3.3e-2, 3.4e-2], [4.1e-2, 4.2e-2, 4.3e-2, 4.4e-2], ]) recv_powers = env.get_recv_powers(emit_powers, fading) rates = env.get_rates(recv_powers) metrics = emit_powers, recv_powers, rates, fading indices = env.get_indices(*metrics) states = env.get_states(*metrics, indices=indices) _recv = np.array([ [[1.1e-4, 0], [1.2e-4, 0], [1.3e-4, 0], [1.4e-4, 0]], [[0, 2.1e-4], [0, 2.2e-4], [0, 2.3e-4], [0, 2.4e-4]], [[3.1e-3, 0], [3.2e-3, 0], [3.3e-3, 0], [3.4e-3, 0]], [[0, 4.1e-3], [0, 4.2e-3], [0, 4.3e-3], [0, 4.4e-3]], ]) _rate = np.array([ log2(1+1.1/31), log2(1+2.2/42), log2(1+33/1.3), log2(1+44/2.4) ]) target_states = np.array([ np.concatenate([power[3],power[0],_recv[3][0],_recv[0][0],[_rate[3], _rate[0]]]), np.concatenate([power[3],power[1],_recv[3][1],_recv[1][1],[_rate[3], _rate[1]]]), np.concatenate([power[3],power[2],_recv[3][2],_recv[2][2],[_rate[3], _rate[2]]]), np.concatenate([power[2],power[3],_recv[2][3],_recv[3][3],[_rate[2], _rate[3]]]), ])[:2] assert equal(states, target_states) def test_sorter(): # now only recv_power can be sorter pass def test_seed(): env = PAEnv(n_level=4, m_cue=1, seed=123) # this is func in PAEnv to random pos def random_point(min_r, radius, ox=0, oy=0): theta = np.random.random() * 2 * np.pi r = np.random.uniform(min_r, radius**2) x, y = np.cos(theta) * np.sqrt(r), np.sin(theta) * np.sqrt(r) return ox + x, oy + y np.random.seed(123) target_x, target_y = random_point(env.r_bs, env.R_bs) usr = env.cues[0] assert all((target_x == usr.x, target_y == usr.y)) def test_action(): env = PAEnv(n_level=10, seed=799345) n_actions = env.n_actions n_channel, n_pair = env.n_channel, env.n_pair # normal env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_channel, )) s_, r, d, i = env.step(action, unit='dBm') assert i['rate'] == 3.4741923099965257 # only D2D actions is enough env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_pair, )) s_, r, d, i = env.step(action, unit='dBm') assert i['rate'] == 3.4741923099965257 # other action dim raises error env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_pair - 1, )) try: s_, r, d, i = env.step(action, unit='dBm') except ValueError as e: msg = f"length of action should be n_channel({env.n_channel})" \ f" or n_pair({n_pair}), but is {len(action)}" assert e.args[0] == msg env.reset() np.random.seed(799345) action = np.random.randint(0, n_actions, (n_channel, )) s_, r, d, i = env.step(action, unit='mW') assert i['rate'] == 3.4928823957853856 # TODO add test of continuous action def test_step(): env = PAEnv(n_level=10) n_actions, n_states = env.n_actions, env.n_states assert n_actions == 40 assert n_states == 304 env.reset() action = env.sample() env.step(action, unit='dBm') # action = env.sample() action = np.array([0, 0, 0, 0, 0, 0, 0, 0, 0]) env.step(action, unit='mW') action = env.sample() try: env.step(action, unit='xx') except ValueError as e: msg = f"unit should in ['dBm', 'mW'], but is xx" assert e.args[0] == msg fig: Path() = env.render() if fig.exists(): fig.unlink() if __name__ == '__main__': test_action()
true
true
f717a0e970b2703e99561e27a1e128a2f7e5f983
488
py
Python
tests/testFlip.py
FrankRogalski/Riccchard
8a1abf6855f874f71df9d101e15c1968394ab38e
[ "MIT" ]
null
null
null
tests/testFlip.py
FrankRogalski/Riccchard
8a1abf6855f874f71df9d101e15c1968394ab38e
[ "MIT" ]
null
null
null
tests/testFlip.py
FrankRogalski/Riccchard
8a1abf6855f874f71df9d101e15c1968394ab38e
[ "MIT" ]
null
null
null
import unittest import commands.flip as flip import pandas as pd class TestFlip(unittest.TestCase): def test_use(self): cycles = 50000 series = pd.Series(flip.use(None) for _ in range(cycles)) self.assertAlmostEqual(len(series[series == "Heads"]) / cycles, 0.4998, delta=0.01) self.assertAlmostEqual(len(series[series == "Heads"]) / cycles, 0.4998, delta=0.01) self.assertAlmostEqual(len(series[series == "Edge"]) / cycles, 0.0002, delta=0.01)
44.363636
91
0.67623
import unittest import commands.flip as flip import pandas as pd class TestFlip(unittest.TestCase): def test_use(self): cycles = 50000 series = pd.Series(flip.use(None) for _ in range(cycles)) self.assertAlmostEqual(len(series[series == "Heads"]) / cycles, 0.4998, delta=0.01) self.assertAlmostEqual(len(series[series == "Heads"]) / cycles, 0.4998, delta=0.01) self.assertAlmostEqual(len(series[series == "Edge"]) / cycles, 0.0002, delta=0.01)
true
true
f717a0f6acd973f0aa1fdd12d14ebe4d62f6382a
714
py
Python
src/detection/dataset/__init__.py
inzva/Traffic-Anomaly-Detection
09bbfd55f5d375832afcc1994d3b549082830594
[ "MIT" ]
4
2020-05-31T13:19:58.000Z
2021-03-11T18:51:06.000Z
src/detection/dataset/__init__.py
inzva/Traffic-Anomaly-Detection
09bbfd55f5d375832afcc1994d3b549082830594
[ "MIT" ]
null
null
null
src/detection/dataset/__init__.py
inzva/Traffic-Anomaly-Detection
09bbfd55f5d375832afcc1994d3b549082830594
[ "MIT" ]
3
2020-05-31T13:19:58.000Z
2021-03-16T08:43:23.000Z
from .custom import CustomDataset from .xml_style import XMLDataset from .coco import CocoDataset from .voc import VOCDataset from .loader import GroupSampler, DistributedGroupSampler, build_dataloader from .utils import to_tensor, random_scale, show_ann, get_dataset from .concat_dataset import ConcatDataset from .repeat_dataset import RepeatDataset from .extra_aug import ExtraAugmentation from .anomaly_data import AnomalyDataset __all__ = [ 'CustomDataset', 'XMLDataset', 'CocoDataset','AnomalyDataset', 'VOCDataset', 'GroupSampler', 'DistributedGroupSampler', 'build_dataloader', 'to_tensor', 'random_scale', 'show_ann', 'get_dataset', 'ConcatDataset', 'RepeatDataset', 'ExtraAugmentation' ]
42
96
0.802521
from .custom import CustomDataset from .xml_style import XMLDataset from .coco import CocoDataset from .voc import VOCDataset from .loader import GroupSampler, DistributedGroupSampler, build_dataloader from .utils import to_tensor, random_scale, show_ann, get_dataset from .concat_dataset import ConcatDataset from .repeat_dataset import RepeatDataset from .extra_aug import ExtraAugmentation from .anomaly_data import AnomalyDataset __all__ = [ 'CustomDataset', 'XMLDataset', 'CocoDataset','AnomalyDataset', 'VOCDataset', 'GroupSampler', 'DistributedGroupSampler', 'build_dataloader', 'to_tensor', 'random_scale', 'show_ann', 'get_dataset', 'ConcatDataset', 'RepeatDataset', 'ExtraAugmentation' ]
true
true
f717a12c03c3130be9f4f39624e011168ba2f2d9
419
py
Python
opencv_py/opencv_morphology/blackhat_operation.py
yinleiCoder/interesting-python
8448b0731870feed033d845c085576a7d4a3154b
[ "Apache-2.0" ]
null
null
null
opencv_py/opencv_morphology/blackhat_operation.py
yinleiCoder/interesting-python
8448b0731870feed033d845c085576a7d4a3154b
[ "Apache-2.0" ]
null
null
null
opencv_py/opencv_morphology/blackhat_operation.py
yinleiCoder/interesting-python
8448b0731870feed033d845c085576a7d4a3154b
[ "Apache-2.0" ]
null
null
null
import cv2 import numpy as np """ 黑帽: 黑帽= 原图 - 闭运算 morphologyEx(img, MORPH_BLACKHAT, kernel) 保留噪点 """ img = cv2.imread(r'E:\PycharmProjects\funnyPython\opencv_py\data\imgs\dotinj.png') # kernel = np.ones((7, 7), np.uint8)# 手动创建 kernel = cv2.getStructuringElement(cv2.MORPH_RECT, (7, 7)) print(kernel) res = cv2.morphologyEx(img, cv2.MORPH_BLACKHAT, kernel) cv2.imshow('img', img) cv2.imshow('res', res) cv2.waitKey(0)
23.277778
82
0.725537
import cv2 import numpy as np img = cv2.imread(r'E:\PycharmProjects\funnyPython\opencv_py\data\imgs\dotinj.png') l = cv2.getStructuringElement(cv2.MORPH_RECT, (7, 7)) print(kernel) res = cv2.morphologyEx(img, cv2.MORPH_BLACKHAT, kernel) cv2.imshow('img', img) cv2.imshow('res', res) cv2.waitKey(0)
true
true
f717a22f1c27c859185f0569c6622ab8d155e4ac
1,915
py
Python
src/trainer.py
Chicco94/crypto-bot
edbc22477544a25d8eb0c90cdd5f03345f11db68
[ "MIT" ]
null
null
null
src/trainer.py
Chicco94/crypto-bot
edbc22477544a25d8eb0c90cdd5f03345f11db68
[ "MIT" ]
null
null
null
src/trainer.py
Chicco94/crypto-bot
edbc22477544a25d8eb0c90cdd5f03345f11db68
[ "MIT" ]
null
null
null
from sklearn.linear_model import LinearRegression from sklearn.model_selection import train_test_split import pandas as pd import sqlalchemy from config.config import symbol,backward_steps import joblib from df_functions import * def prepare_single_dataset(df,remove_from_heads:int,remove_from_tails:int,label:int): df_copy = df.copy() for _ in range(remove_from_tails): remove_row_from_tail(df_copy) for _ in range(remove_from_heads): remove_row_from_head(df_copy) add_id(df_copy) df_copy.time = df_copy.time.apply(lambda x: x.value) df_copy.rename(columns={"time": "time{}".format(label) , "price": "price{}".format(label) , "quantity":"quantity{}".format(label)} ,inplace=True) df_copy.drop(columns=['symbol'],inplace=True) return df_copy def prepare_dataset(df,steps:int): datasets = [] for i in range(1,steps): datasets.append(prepare_single_dataset(df,steps-i,i-1,i)) df_target = prepare_single_dataset(df,0,steps-1,steps) result = datasets.pop() while len(datasets)>0: result = pd.merge(result, datasets.pop(), on="ID") target = df_target['price{}'.format(steps)] return result,target def main(): # open database engine = sqlalchemy.create_engine('sqlite:///data/{}_stream.db'.format(symbol)) df = pd.read_sql(symbol,engine) # prepare dataset source,target = prepare_dataset(df,backward_steps) # train model model = LinearRegression() X_train,X_test,y_train,y_test = train_test_split(source,target,test_size=0.33) model.fit(X_train,y_train) # evaluate model score = model.score(X_test,y_test) print('score: ',score) # save model filename = 'models/model_{}.sav'.format(score) joblib.dump(model, filename) #model = joblib.load(filename) if __name__=='__main__': main()
33.017241
85
0.681462
from sklearn.linear_model import LinearRegression from sklearn.model_selection import train_test_split import pandas as pd import sqlalchemy from config.config import symbol,backward_steps import joblib from df_functions import * def prepare_single_dataset(df,remove_from_heads:int,remove_from_tails:int,label:int): df_copy = df.copy() for _ in range(remove_from_tails): remove_row_from_tail(df_copy) for _ in range(remove_from_heads): remove_row_from_head(df_copy) add_id(df_copy) df_copy.time = df_copy.time.apply(lambda x: x.value) df_copy.rename(columns={"time": "time{}".format(label) , "price": "price{}".format(label) , "quantity":"quantity{}".format(label)} ,inplace=True) df_copy.drop(columns=['symbol'],inplace=True) return df_copy def prepare_dataset(df,steps:int): datasets = [] for i in range(1,steps): datasets.append(prepare_single_dataset(df,steps-i,i-1,i)) df_target = prepare_single_dataset(df,0,steps-1,steps) result = datasets.pop() while len(datasets)>0: result = pd.merge(result, datasets.pop(), on="ID") target = df_target['price{}'.format(steps)] return result,target def main(): engine = sqlalchemy.create_engine('sqlite:///data/{}_stream.db'.format(symbol)) df = pd.read_sql(symbol,engine) source,target = prepare_dataset(df,backward_steps) model = LinearRegression() X_train,X_test,y_train,y_test = train_test_split(source,target,test_size=0.33) model.fit(X_train,y_train) score = model.score(X_test,y_test) print('score: ',score) filename = 'models/model_{}.sav'.format(score) joblib.dump(model, filename) if __name__=='__main__': main()
true
true
f717a46b3c38e7a991236ce1bb31333e778f6d48
300
py
Python
backend/app/main.py
alldevic/base-fastapi-postgresql
7e3a2916910155cd83b10cd7fec42eba7b1d3a95
[ "MIT" ]
3
2021-06-17T00:06:15.000Z
2022-01-26T03:53:51.000Z
backend/app/main.py
alldevic/base-fastapi-postgresql
7e3a2916910155cd83b10cd7fec42eba7b1d3a95
[ "MIT" ]
null
null
null
backend/app/main.py
alldevic/base-fastapi-postgresql
7e3a2916910155cd83b10cd7fec42eba7b1d3a95
[ "MIT" ]
null
null
null
from .routes.users import router as user_router from .application import app import sys sys.path.extend(["./"]) ROUTERS = (user_router,) for r in ROUTERS: app.include_router(r) if __name__ == "__main__": import uvicorn uvicorn.run(app, host="0.0.0.0", port=8888, log_level="info")
16.666667
65
0.693333
from .routes.users import router as user_router from .application import app import sys sys.path.extend(["./"]) ROUTERS = (user_router,) for r in ROUTERS: app.include_router(r) if __name__ == "__main__": import uvicorn uvicorn.run(app, host="0.0.0.0", port=8888, log_level="info")
true
true
f717a4b4672f82239831e0aa4b03a4aaca187963
11,669
py
Python
tests/test_storage.py
civviGH/otterwiki
2ef84aa31f48da144cb0ef7a0116106b982f41d2
[ "MIT" ]
3
2018-07-15T20:20:31.000Z
2022-01-09T07:41:39.000Z
tests/test_storage.py
civviGH/otterwiki
2ef84aa31f48da144cb0ef7a0116106b982f41d2
[ "MIT" ]
3
2018-07-04T21:38:41.000Z
2022-03-06T17:03:23.000Z
tests/test_storage.py
civviGH/otterwiki
2ef84aa31f48da144cb0ef7a0116106b982f41d2
[ "MIT" ]
2
2021-04-29T13:17:56.000Z
2022-03-07T09:17:19.000Z
#!/usr/bin/env python import tempfile import unittest from pprint import pprint import otterwiki.storage class TestStorage(unittest.TestCase): def setUp(self): self.tempdir = tempfile.TemporaryDirectory() self.path = '/tmp/xxx' # self.tempdir.name self.path = self.tempdir.name self.storage = otterwiki.storage.GitStorage(path=self.path, initialize=True) def tearDown(self): self.tempdir.cleanup() def test_store_and_load(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) # check content self.assertEqual( self.storage.load(filename), content ) # check metadata metadata = self.storage.metadata(filename) self.assertEqual( metadata['author_name'], author[0] ) self.assertEqual( metadata['author_email'], author[1] ) self.assertEqual( metadata['message'], message ) # check if file is listed self.assertIn( filename, self.storage.list_files() ) # check if storing the same content changes nothing self.assertFalse( self.storage.store(filename, content=content, author=author) ) def test_load_fail(self): with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.load("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata("non-existent.md", revision="xxx") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.log("non-existent.md") def test_broken_author(self): content = "This is test content.\n" message = "Test commit" filename = "test_broken_author.md" author = ("Example Author", "") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) self.assertEqual( metadata['author_name'], author[0] ) def test_broken_message(self): content = "This is test content.\n" message = None filename = "test_broken_message.md" author = ("Example Author", "mail@example.org") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) self.assertEqual( metadata['message'], '' ) def test_log(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test_log.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) # test log for filename log = self.storage.log(filename) msg = log[-1] self.assertEqual(msg['message'], message) # test global log log = self.storage.log() msg = log[-1] self.assertEqual(msg['message'], message) def test_revert(self): author = ("Example Author", "mail@example.com") filename = "test_revert.md" content1 = "aaa" message1 = "added {}".format(content1) content2 = "bbb" message2 = "added {}".format(content2) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) # check content self.assertEqual( self.storage.load(filename), content1 ) # change content self.assertTrue( self.storage.store(filename, content=content2, author=author, message=message2) ) # check content self.assertEqual( self.storage.load(filename), content2 ) # get revision log = self.storage.log(filename) revision = log[0]['revision'] self.storage.revert(revision, message="reverted {}".format(revision), author=author) # check that the file is in the old state self.assertEqual( self.storage.load(filename), content1 ) def test_revert_fail(self): author = ("Example Author", "mail@example.com") filename = "test_revert_fail.md" content1 = "aaa" message1 = "added {}".format(content1) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) # get revision log = self.storage.log(filename) revision = log[0]['revision'] # revert self.storage.revert(revision, message="reverted {}".format(revision), author=author) files = self.storage.list_files() self.assertNotIn(filename, files) def test_ascii_binary(self): content = u"kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_binary.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) content_rb = self.storage.load(filename, mode="rb") content_r = self.storage.load(filename, mode="r") # check types self.assertIs(type(content_rb), bytes) self.assertIs(type(content_r), str) # convert into str content_utf8 = content_rb.decode("utf-8") self.assertIs(type(content_utf8), str) self.assertEqual(content_utf8, content_r) def test_binary(self): content = b"GIF89a\x01\x00\x01\x00\x80\x01\x00\xff\xff"\ b"\xff\x00\x00\x00!\xf9\x04\x01\n\x00\x01\x00,"\ b"\x00\x00\x00\x00\x01\x00\x01\x00\x00\x02\x02L\x01\x00;" message = "Test commit" filename = "test_binary.gif" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message, mode='wb') ) content_rb = self.storage.load(filename, mode="rb") self.assertEqual(content, content_rb) # get log log = self.storage.log() # get revisions revision = log[0]['revision'] content_rb2 = self.storage.load(filename, mode="rb", revision=revision) self.assertEqual(content_rb2, content_rb) def test_revision(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_revision.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) revision = metadata['revision-full'] self.assertEqual( self.storage.load(filename, revision=revision), content ) metadata2 = self.storage.metadata(filename, revision=revision) self.assertEqual( metadata, metadata2) # check broken revision revision3 = "xxx{}".format(revision) with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata(filename, revision=revision3) with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.load(filename, revision=revision3) def test_store_subdir(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test_subdir/test_subdir.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) # check if file exists self.assertTrue( self.storage.exists(filename) ) # check via file list files = self.storage.list_files() self.assertIn(filename, files) dn, fn = filename.split('/') files = self.storage.list_files(dn) self.assertIn(fn, files) def test_diff(self): author = ("Example Author", "mail@example.com") filename = "test_revert.md" content1 = "aaa" message1 = "added {}".format(content1) content2 = "bbb" message2 = "added {}".format(content2) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) # check content self.assertEqual( self.storage.load(filename), content1 ) # change content self.assertTrue( self.storage.store(filename, content=content2, author=author, message=message2) ) # check content self.assertEqual( self.storage.load(filename), content2 ) # get log log = self.storage.log() # get revisions rev_b, rev_a= log[0]['revision'], log[1]['revision'] # get diff diff = self.storage.diff(filename, rev_a, rev_b) # check -/+ strings self.assertIn("-aaa", diff) self.assertIn("+bbb", diff) def test_rename(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename1 = "test_rename1.md" filename2 = "test_rename2.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename1, content=content, author=author, message=message) ) # rename self.storage.rename( filename1, filename2, author=author ) # check if file exists self.assertFalse( self.storage.exists(filename1) ) self.assertTrue( self.storage.exists(filename2) ) # check if file exists via list_files files = self.storage.list_files() self.assertNotIn(filename1, files) self.assertIn(filename2, files) # check content self.assertEqual( self.storage.load(filename2), content ) # test rename fail with self.assertRaises(otterwiki.storage.StorageError): self.storage.rename( filename1, "", author=author ) def test_delete(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_revision.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) # check if file exists files = self.storage.list_files() self.assertIn(filename, files) # remove file self.storage.delete(filename, author=author) # check that file doesn't exist anymore files = self.storage.list_files() self.assertNotIn(filename, files) class TestEmptyStorage(unittest.TestCase): def test_log(self): with tempfile.TemporaryDirectory() as path: storage = otterwiki.storage.GitStorage(path=path, initialize=True) self.assertEqual(storage.log(), []) def test_load_fail(self): with tempfile.TemporaryDirectory() as path: storage = otterwiki.storage.GitStorage(path=path, initialize=True) with self.assertRaises(otterwiki.storage.StorageNotFound): storage.load("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): storage.log("non-existent.md") def test_init_fail(self): with tempfile.TemporaryDirectory() as path: with self.assertRaises(otterwiki.storage.StorageError): storage = otterwiki.storage.GitStorage(path=path) if __name__ == '__main__': unittest.main()
43.70412
115
0.646756
import tempfile import unittest from pprint import pprint import otterwiki.storage class TestStorage(unittest.TestCase): def setUp(self): self.tempdir = tempfile.TemporaryDirectory() self.path = '/tmp/xxx' self.path = self.tempdir.name self.storage = otterwiki.storage.GitStorage(path=self.path, initialize=True) def tearDown(self): self.tempdir.cleanup() def test_store_and_load(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) self.assertEqual( self.storage.load(filename), content ) metadata = self.storage.metadata(filename) self.assertEqual( metadata['author_name'], author[0] ) self.assertEqual( metadata['author_email'], author[1] ) self.assertEqual( metadata['message'], message ) self.assertIn( filename, self.storage.list_files() ) self.assertFalse( self.storage.store(filename, content=content, author=author) ) def test_load_fail(self): with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.load("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata("non-existent.md", revision="xxx") with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.log("non-existent.md") def test_broken_author(self): content = "This is test content.\n" message = "Test commit" filename = "test_broken_author.md" author = ("Example Author", "") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) self.assertEqual( metadata['author_name'], author[0] ) def test_broken_message(self): content = "This is test content.\n" message = None filename = "test_broken_message.md" author = ("Example Author", "mail@example.org") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) self.assertEqual( metadata['message'], '' ) def test_log(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test_log.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) log = self.storage.log(filename) msg = log[-1] self.assertEqual(msg['message'], message) log = self.storage.log() msg = log[-1] self.assertEqual(msg['message'], message) def test_revert(self): author = ("Example Author", "mail@example.com") filename = "test_revert.md" content1 = "aaa" message1 = "added {}".format(content1) content2 = "bbb" message2 = "added {}".format(content2) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) self.assertEqual( self.storage.load(filename), content1 ) self.assertTrue( self.storage.store(filename, content=content2, author=author, message=message2) ) self.assertEqual( self.storage.load(filename), content2 ) log = self.storage.log(filename) revision = log[0]['revision'] self.storage.revert(revision, message="reverted {}".format(revision), author=author) self.assertEqual( self.storage.load(filename), content1 ) def test_revert_fail(self): author = ("Example Author", "mail@example.com") filename = "test_revert_fail.md" content1 = "aaa" message1 = "added {}".format(content1) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) log = self.storage.log(filename) revision = log[0]['revision'] self.storage.revert(revision, message="reverted {}".format(revision), author=author) files = self.storage.list_files() self.assertNotIn(filename, files) def test_ascii_binary(self): content = u"kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_binary.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) content_rb = self.storage.load(filename, mode="rb") content_r = self.storage.load(filename, mode="r") self.assertIs(type(content_rb), bytes) self.assertIs(type(content_r), str) content_utf8 = content_rb.decode("utf-8") self.assertIs(type(content_utf8), str) self.assertEqual(content_utf8, content_r) def test_binary(self): content = b"GIF89a\x01\x00\x01\x00\x80\x01\x00\xff\xff"\ b"\xff\x00\x00\x00!\xf9\x04\x01\n\x00\x01\x00,"\ b"\x00\x00\x00\x00\x01\x00\x01\x00\x00\x02\x02L\x01\x00;" message = "Test commit" filename = "test_binary.gif" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message, mode='wb') ) content_rb = self.storage.load(filename, mode="rb") self.assertEqual(content, content_rb) log = self.storage.log() revision = log[0]['revision'] content_rb2 = self.storage.load(filename, mode="rb", revision=revision) self.assertEqual(content_rb2, content_rb) def test_revision(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_revision.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) metadata = self.storage.metadata(filename) revision = metadata['revision-full'] self.assertEqual( self.storage.load(filename, revision=revision), content ) metadata2 = self.storage.metadata(filename, revision=revision) self.assertEqual( metadata, metadata2) revision3 = "xxx{}".format(revision) with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.metadata(filename, revision=revision3) with self.assertRaises(otterwiki.storage.StorageNotFound): self.storage.load(filename, revision=revision3) def test_store_subdir(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename = "test_subdir/test_subdir.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) self.assertTrue( self.storage.exists(filename) ) files = self.storage.list_files() self.assertIn(filename, files) dn, fn = filename.split('/') files = self.storage.list_files(dn) self.assertIn(fn, files) def test_diff(self): author = ("Example Author", "mail@example.com") filename = "test_revert.md" content1 = "aaa" message1 = "added {}".format(content1) content2 = "bbb" message2 = "added {}".format(content2) self.assertTrue( self.storage.store(filename, content=content1, author=author, message=message1) ) self.assertEqual( self.storage.load(filename), content1 ) self.assertTrue( self.storage.store(filename, content=content2, author=author, message=message2) ) self.assertEqual( self.storage.load(filename), content2 ) log = self.storage.log() rev_b, rev_a= log[0]['revision'], log[1]['revision'] diff = self.storage.diff(filename, rev_a, rev_b) self.assertIn("-aaa", diff) self.assertIn("+bbb", diff) def test_rename(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad\n" message = "Test commit" filename1 = "test_rename1.md" filename2 = "test_rename2.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename1, content=content, author=author, message=message) ) self.storage.rename( filename1, filename2, author=author ) self.assertFalse( self.storage.exists(filename1) ) self.assertTrue( self.storage.exists(filename2) ) files = self.storage.list_files() self.assertNotIn(filename1, files) self.assertIn(filename2, files) self.assertEqual( self.storage.load(filename2), content ) with self.assertRaises(otterwiki.storage.StorageError): self.storage.rename( filename1, "", author=author ) def test_delete(self): content = "kdfjlhg gdklfjghdf gkl;djshfg dgf;lkjhs glkshjad" message = "Test commit" filename = "test_revision.md" author = ("Example Author", "mail@example.com") self.assertTrue( self.storage.store(filename, content=content, author=author, message=message) ) files = self.storage.list_files() self.assertIn(filename, files) self.storage.delete(filename, author=author) files = self.storage.list_files() self.assertNotIn(filename, files) class TestEmptyStorage(unittest.TestCase): def test_log(self): with tempfile.TemporaryDirectory() as path: storage = otterwiki.storage.GitStorage(path=path, initialize=True) self.assertEqual(storage.log(), []) def test_load_fail(self): with tempfile.TemporaryDirectory() as path: storage = otterwiki.storage.GitStorage(path=path, initialize=True) with self.assertRaises(otterwiki.storage.StorageNotFound): storage.load("non-existent.md") with self.assertRaises(otterwiki.storage.StorageNotFound): storage.log("non-existent.md") def test_init_fail(self): with tempfile.TemporaryDirectory() as path: with self.assertRaises(otterwiki.storage.StorageError): storage = otterwiki.storage.GitStorage(path=path) if __name__ == '__main__': unittest.main()
true
true
f717a54cab0f7211990d1d0fff0102426e318378
4,412
py
Python
TTY_Utils/StatusLine.py
paul-ollis/cleversheep3
86e6ca76ea4e8524f16e2348d38484dcfafb07d0
[ "Apache-2.0" ]
null
null
null
TTY_Utils/StatusLine.py
paul-ollis/cleversheep3
86e6ca76ea4e8524f16e2348d38484dcfafb07d0
[ "Apache-2.0" ]
null
null
null
TTY_Utils/StatusLine.py
paul-ollis/cleversheep3
86e6ca76ea4e8524f16e2348d38484dcfafb07d0
[ "Apache-2.0" ]
null
null
null
import sys import time from cleversheep3.TTY_Utils import RichTerm, registerForWinch class Status: """A fairly general purpose status line for a simple terminal. """ def __init__(self, startActive=True): self.setTerm(RichTerm.RichTerminal(sys.stdout)) self.spinner = None self.prevLine = None self.leftFields = [] self.leftFieldsByName = {} self.rightFields = [] self.rightFieldsByName = {} self.prevTime = time.time() self.updateInterval = 0.1 self.killed = True self.active = startActive registerForWinch(lambda: self.onWinch()) def onWinch(self): h, self.w = self.term.getDims(force=1) def setTerm(self, term): self.term = term h, self.w = self.term.getDims() def stop(self): self.kill() self.active = False def start(self): self.active = True def addLeftField(self, name, w): if name not in self.leftFieldsByName: self.leftFieldsByName[name] = len(self.leftFields) self.leftFields.append((w, "")) addField = addLeftField def addRightField(self, name, w): if name not in self.rightFieldsByName: self.rightFieldsByName[name] = len(self.rightFields) self.rightFields.append((w, "")) def setField(self, name, text): if text: text = text.splitlines()[0] try: idx = self.leftFieldsByName[name] w, _ = self.leftFields[idx] self.leftFields[idx] = w, text except KeyError: idx = self.rightFieldsByName[name] w, _ = self.rightFields[idx] self.rightFields[idx] = w, text self.update() def addSpinner(self, seq=r'\|/-'): def chars(): while True: for c in seq: yield c def s(): cc = chars() c = next(cc) t = time.time() while True: if time.time() - t > 0.1: t = time.time() c = next(cc) yield c self.spinner = s() def buildLine(self): ww = self.w - 2 rline = "" lline = "" for w, text in self.rightFields: bar = "" if rline: bar = "|" rline = "%-*.*s%s%s" % (w, w, text, bar, rline) lline = "" if self.spinner: lline += "%s " % next(self.spinner) for w, text in self.leftFields: if lline: lline += "|" if w is not None: lline += "%-*.*s" % (w, w, text) else: lline += "%s" % (text,) l = len(lline) + len(rline) if l > ww: rline = rline.rstrip() l = len(lline) + len(rline) if l > ww: lline = lline.rstrip() l = len(lline) + len(rline) while len(lline) > len(rline) and l > ww: lline = lline[:-1] l = len(lline) + len(rline) while len(rline) > len(lline) and l > ww: rline = rline[1:] l = len(lline) + len(rline) while l > ww: lline = lline[:-1] rline = rline[1:] l = len(lline) + len(rline) if lline and rline: pad = " " * (ww - len(lline) - len(rline) - 1) line = "%s%s|%s" % (lline, pad, rline) elif rline: pad = " " * (ww - len(lline) - len(rline)) line = "%s%s%s" % (lline, pad, rline) else: line = lline return line def update(self, force=False): if not (self.active and self.term.isatty()): return if not force and time.time() - self.prevTime < self.updateInterval: return self.prevTime = time.time() self.line = self.buildLine() if self.line != self.prevLine: self.term.write("\r%s" % self.line) self.term.flush() self.prevLine = self.line self.killed = False def kill(self): if not (self.active and self.term.isatty()): return if self.killed: return self.term.write("\r%s\r" % (" " * (self.w - 1))) self.term.flush() self.prevLine = None self.killed = True
29.026316
75
0.486174
import sys import time from cleversheep3.TTY_Utils import RichTerm, registerForWinch class Status: def __init__(self, startActive=True): self.setTerm(RichTerm.RichTerminal(sys.stdout)) self.spinner = None self.prevLine = None self.leftFields = [] self.leftFieldsByName = {} self.rightFields = [] self.rightFieldsByName = {} self.prevTime = time.time() self.updateInterval = 0.1 self.killed = True self.active = startActive registerForWinch(lambda: self.onWinch()) def onWinch(self): h, self.w = self.term.getDims(force=1) def setTerm(self, term): self.term = term h, self.w = self.term.getDims() def stop(self): self.kill() self.active = False def start(self): self.active = True def addLeftField(self, name, w): if name not in self.leftFieldsByName: self.leftFieldsByName[name] = len(self.leftFields) self.leftFields.append((w, "")) addField = addLeftField def addRightField(self, name, w): if name not in self.rightFieldsByName: self.rightFieldsByName[name] = len(self.rightFields) self.rightFields.append((w, "")) def setField(self, name, text): if text: text = text.splitlines()[0] try: idx = self.leftFieldsByName[name] w, _ = self.leftFields[idx] self.leftFields[idx] = w, text except KeyError: idx = self.rightFieldsByName[name] w, _ = self.rightFields[idx] self.rightFields[idx] = w, text self.update() def addSpinner(self, seq=r'\|/-'): def chars(): while True: for c in seq: yield c def s(): cc = chars() c = next(cc) t = time.time() while True: if time.time() - t > 0.1: t = time.time() c = next(cc) yield c self.spinner = s() def buildLine(self): ww = self.w - 2 rline = "" lline = "" for w, text in self.rightFields: bar = "" if rline: bar = "|" rline = "%-*.*s%s%s" % (w, w, text, bar, rline) lline = "" if self.spinner: lline += "%s " % next(self.spinner) for w, text in self.leftFields: if lline: lline += "|" if w is not None: lline += "%-*.*s" % (w, w, text) else: lline += "%s" % (text,) l = len(lline) + len(rline) if l > ww: rline = rline.rstrip() l = len(lline) + len(rline) if l > ww: lline = lline.rstrip() l = len(lline) + len(rline) while len(lline) > len(rline) and l > ww: lline = lline[:-1] l = len(lline) + len(rline) while len(rline) > len(lline) and l > ww: rline = rline[1:] l = len(lline) + len(rline) while l > ww: lline = lline[:-1] rline = rline[1:] l = len(lline) + len(rline) if lline and rline: pad = " " * (ww - len(lline) - len(rline) - 1) line = "%s%s|%s" % (lline, pad, rline) elif rline: pad = " " * (ww - len(lline) - len(rline)) line = "%s%s%s" % (lline, pad, rline) else: line = lline return line def update(self, force=False): if not (self.active and self.term.isatty()): return if not force and time.time() - self.prevTime < self.updateInterval: return self.prevTime = time.time() self.line = self.buildLine() if self.line != self.prevLine: self.term.write("\r%s" % self.line) self.term.flush() self.prevLine = self.line self.killed = False def kill(self): if not (self.active and self.term.isatty()): return if self.killed: return self.term.write("\r%s\r" % (" " * (self.w - 1))) self.term.flush() self.prevLine = None self.killed = True
true
true
f717a6a6f7df269ba5f932fb06f6150bdffcbc05
2,958
py
Python
etk/unit_tests/test_filter_results.py
linqyd/etk
dcf0cae4076619f5261573d47b4f5f26baaf15b7
[ "MIT" ]
null
null
null
etk/unit_tests/test_filter_results.py
linqyd/etk
dcf0cae4076619f5261573d47b4f5f26baaf15b7
[ "MIT" ]
null
null
null
etk/unit_tests/test_filter_results.py
linqyd/etk
dcf0cae4076619f5261573d47b4f5f26baaf15b7
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import unittest import sys, os sys.path.append('../../') from etk.core import Core import json import codecs class TestExtractionsFilterResults(unittest.TestCase): def test_filter_results(self): doc = { "url":"http:www.testurl.com", "doc_id": "19B0EAB211CD1D3C63063FAB0B2937043EA1F07B5341014A80E7473BA7318D9E", "knowledge_graph": { "name": [ { "provenance": [ { "extracted_value": "Very", "method": "extract_using_custom_spacy", "confidence": { "extraction": 1 }, "source": { "segment": "content_strict", "context": { "rule_id": 1, "input": "tokens", "identifier": "name_rule_02", "start": 18, "end": 21, "text": ". \n Well Guess What <etk 'attribute' = 'name'>i am Very</etk> Real \n I DON ' " }, "document_id": "19B0EAB211CD1D3C63063FAB0B2937043EA1F07B5341014A80E7473BA7318D9E" } } ], "confidence": 1.0, "value": "Very", "key": "very" } ] } } stop_words_path = os.path.join(os.path.dirname(__file__), "resources/stop_word_names.json") e_config = { "document_id":"doc_id", "resources": { "stop_word_dictionaries": { "name": stop_words_path } }, "kg_enhancement": { "fields": { "name": { "priority": 0, "extractors": { "filter_results": { "config": { "stop_word_dictionaries": "name" } } } } }, "input_path": "knowledge_graph.`parent`" }} c = Core(extraction_config=e_config) r = c.process(doc) self.assertTrue('knowledge_graph' in doc) self.assertTrue('name' in doc['knowledge_graph']) self.assertTrue(len(doc['knowledge_graph']['name']) == 1) self.assertTrue(doc['knowledge_graph']['name'][0]['confidence'] == 0.3) if __name__ == '__main__': unittest.main()
36.518519
129
0.373563
import unittest import sys, os sys.path.append('../../') from etk.core import Core import json import codecs class TestExtractionsFilterResults(unittest.TestCase): def test_filter_results(self): doc = { "url":"http:www.testurl.com", "doc_id": "19B0EAB211CD1D3C63063FAB0B2937043EA1F07B5341014A80E7473BA7318D9E", "knowledge_graph": { "name": [ { "provenance": [ { "extracted_value": "Very", "method": "extract_using_custom_spacy", "confidence": { "extraction": 1 }, "source": { "segment": "content_strict", "context": { "rule_id": 1, "input": "tokens", "identifier": "name_rule_02", "start": 18, "end": 21, "text": ". \n Well Guess What <etk 'attribute' = 'name'>i am Very</etk> Real \n I DON ' " }, "document_id": "19B0EAB211CD1D3C63063FAB0B2937043EA1F07B5341014A80E7473BA7318D9E" } } ], "confidence": 1.0, "value": "Very", "key": "very" } ] } } stop_words_path = os.path.join(os.path.dirname(__file__), "resources/stop_word_names.json") e_config = { "document_id":"doc_id", "resources": { "stop_word_dictionaries": { "name": stop_words_path } }, "kg_enhancement": { "fields": { "name": { "priority": 0, "extractors": { "filter_results": { "config": { "stop_word_dictionaries": "name" } } } } }, "input_path": "knowledge_graph.`parent`" }} c = Core(extraction_config=e_config) r = c.process(doc) self.assertTrue('knowledge_graph' in doc) self.assertTrue('name' in doc['knowledge_graph']) self.assertTrue(len(doc['knowledge_graph']['name']) == 1) self.assertTrue(doc['knowledge_graph']['name'][0]['confidence'] == 0.3) if __name__ == '__main__': unittest.main()
true
true
f717a6c9a05c6c18f54f5308390fe72e04c86105
10,727
py
Python
dataset/cs_labels.py
ChunGaoY/DSPnet
8fad61059d85ad0cd1f7790c37b5e0478dccb158
[ "MIT" ]
33
2018-04-18T06:52:16.000Z
2021-09-26T20:57:56.000Z
dataset/cs_labels.py
ChunGaoY/DSPnet
8fad61059d85ad0cd1f7790c37b5e0478dccb158
[ "MIT" ]
4
2018-07-09T07:09:15.000Z
2020-04-12T12:43:36.000Z
dataset/cs_labels.py
liangfu/dspnet
8fad61059d85ad0cd1f7790c37b5e0478dccb158
[ "MIT" ]
10
2018-04-19T08:17:01.000Z
2021-09-26T20:57:57.000Z
#!/usr/bin/python # # Cityscapes labels # from collections import namedtuple #-------------------------------------------------------------------------------- # Definitions #-------------------------------------------------------------------------------- # a label and all meta information Label = namedtuple( 'Label' , [ 'name' , # The identifier of this label, e.g. 'car', 'person', ... . # We use them to uniquely name a class 'id' , # An integer ID that is associated with this label. # The IDs are used to represent the label in ground truth images # An ID of -1 means that this label does not have an ID and thus # is ignored when creating ground truth images (e.g. license plate). # Do not modify these IDs, since exactly these IDs are expected by the # evaluation server. 'trainId' , # Feel free to modify these IDs as suitable for your method. Then create # ground truth images with train IDs, using the tools provided in the # 'preparation' folder. However, make sure to validate or submit results # to our evaluation server using the regular IDs above! # For trainIds, multiple labels might have the same ID. Then, these labels # are mapped to the same class in the ground truth images. For the inverse # mapping, we use the label that is defined first in the list below. # For example, mapping all void-type classes to the same ID in training, # might make sense for some approaches. # Max value is 255! 'category' , # The name of the category that this label belongs to 'categoryId' , # The ID of this category. Used to create ground truth images # on category level. 'hasInstances', # Whether this label distinguishes between single instances or not 'ignoreInEval', # Whether pixels having this class as ground truth label are ignored # during evaluations or not 'color' , # The color of this label ] ) #-------------------------------------------------------------------------------- # A list of all labels #-------------------------------------------------------------------------------- # Please adapt the train IDs as appropriate for you approach. # Note that you might want to ignore labels with ID 255 during training. # Further note that the current train IDs are only a suggestion. You can use whatever you like. # Make sure to provide your results using the original IDs and not the training IDs. # Note that many IDs are ignored in evaluation and thus you never need to predict these! labels = [ # name id trainId category catId hasInstances ignoreInEval color Label( 'unlabeled' , 0 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'ego vehicle' , 1 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'rectification border' , 2 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'out of roi' , 3 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'static' , 4 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'dynamic' , 5 , 255 , 'void' , 0 , False , True , (111, 74, 0) ), Label( 'ground' , 6 , 255 , 'void' , 0 , False , True , ( 81, 0, 81) ), Label( 'road' , 7 , 0 , 'flat' , 1 , False , False , (128, 64,128) ), Label( 'sidewalk' , 8 , 1 , 'flat' , 1 , False , False , (244, 35,232) ), Label( 'parking' , 9 , 255 , 'flat' , 1 , False , True , (250,170,160) ), Label( 'rail track' , 10 , 255 , 'flat' , 1 , False , True , (230,150,140) ), Label( 'building' , 11 , 2 , 'construction' , 2 , False , False , ( 70, 70, 70) ), Label( 'wall' , 12 , 3 , 'construction' , 2 , False , False , (102,102,156) ), Label( 'fence' , 13 , 4 , 'construction' , 2 , False , False , (190,153,153) ), Label( 'guard rail' , 14 , 255 , 'construction' , 2 , False , True , (180,165,180) ), Label( 'bridge' , 15 , 255 , 'construction' , 2 , False , True , (150,100,100) ), Label( 'tunnel' , 16 , 255 , 'construction' , 2 , False , True , (150,120, 90) ), Label( 'pole' , 17 , 5 , 'object' , 3 , False , False , (153,153,153) ), Label( 'polegroup' , 18 , 255 , 'object' , 3 , False , True , (153,153,153) ), Label( 'traffic light' , 19 , 6 , 'object' , 3 , False , False , (250,170, 30) ), Label( 'traffic sign' , 20 , 7 , 'object' , 3 , False , False , (220,220, 0) ), Label( 'vegetation' , 21 , 8 , 'nature' , 4 , False , False , (107,142, 35) ), Label( 'terrain' , 22 , 9 , 'nature' , 4 , False , False , (152,251,152) ), Label( 'sky' , 23 , 10 , 'sky' , 5 , False , False , ( 70,130,180) ), Label( 'person' , 24 , 11 , 'human' , 6 , True , False , (220, 20, 60) ), Label( 'rider' , 25 , 12 , 'human' , 6 , True , False , (255, 0, 0) ), Label( 'car' , 26 , 13 , 'vehicle' , 7 , True , False , ( 0, 0,142) ), Label( 'truck' , 27 , 14 , 'vehicle' , 7 , True , False , ( 0, 0, 70) ), Label( 'bus' , 28 , 15 , 'vehicle' , 7 , True , False , ( 0, 60,100) ), Label( 'caravan' , 29 , 255 , 'vehicle' , 7 , True , True , ( 0, 0, 90) ), Label( 'trailer' , 30 , 255 , 'vehicle' , 7 , True , True , ( 0, 0,110) ), Label( 'train' , 31 , 16 , 'vehicle' , 7 , True , False , ( 0, 80,100) ), Label( 'motorcycle' , 32 , 17 , 'vehicle' , 7 , True , False , ( 0, 0,230) ), Label( 'bicycle' , 33 , 18 , 'vehicle' , 7 , True , False , (119, 11, 32) ), Label( 'license plate' , -1 , -1 , 'vehicle' , 7 , False , True , ( 0, 0,142) ), Label( 'lane marking' , 34 , 19 , 'vehicle' , 1 , False , True , (192, 64,192) ), ] #-------------------------------------------------------------------------------- # Create dictionaries for a fast lookup #-------------------------------------------------------------------------------- # Please refer to the main method below for example usages! # name to label object name2label = { label.name : label for label in labels } # id to label object id2label = { label.id : label for label in labels } # trainId to label object trainId2label = { label.trainId : label for label in reversed(labels) } # category to list of label objects category2labels = {} for label in labels: category = label.category if category in category2labels: category2labels[category].append(label) else: category2labels[category] = [label] #-------------------------------------------------------------------------------- # Assure single instance name #-------------------------------------------------------------------------------- # returns the label name that describes a single instance (if possible) # e.g. input | output # ---------------------- # car | car # cargroup | car # foo | None # foogroup | None # skygroup | None def assureSingleInstanceName( name ): # if the name is known, it is not a group if name in name2label: return name # test if the name actually denotes a group if not name.endswith("group"): return None # remove group name = name[:-len("group")] # test if the new name exists if not name in name2label: return None # test if the new name denotes a label that actually has instances if not name2label[name].hasInstances: return None # all good then return name #-------------------------------------------------------------------------------- # Main for testing #-------------------------------------------------------------------------------- # just a dummy main if __name__ == "__main__": # Print all the labels print("List of cityscapes labels:") print("") print(" {:>21} | {:>3} | {:>7} | {:>14} | {:>10} | {:>12} | {:>12}".format( 'name', 'id', 'trainId', 'category', 'categoryId', 'hasInstances', 'ignoreInEval' )) print(" " + ('-' * 98)) for label in labels: print(" {:>21} | {:>3} | {:>7} | {:>14} | {:>10} | {:>12} | {:>12}".format( label.name, label.id, label.trainId, label.category, label.categoryId, label.hasInstances, label.ignoreInEval )) print("") print("Example usages:") # Map from name to label name = 'car' id = name2label[name].id print("ID of label '{name}': {id}".format( name=name, id=id )) # Map from ID to label category = id2label[id].category print("Category of label with ID '{id}': {category}".format( id=id, category=category )) # Map from trainID to label trainId = 0 name = trainId2label[trainId].name print("Name of label with trainID '{id}': {name}".format( id=trainId, name=name ))
58.617486
199
0.430223
from collections import namedtuple Label = namedtuple( 'Label' , [ 'name' , 'id' , 'trainId' , 'category' , 'categoryId' , 'hasInstances', 'ignoreInEval', 'color' , ] ) labels = [ Label( 'unlabeled' , 0 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'ego vehicle' , 1 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'rectification border' , 2 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'out of roi' , 3 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'static' , 4 , 255 , 'void' , 0 , False , True , ( 0, 0, 0) ), Label( 'dynamic' , 5 , 255 , 'void' , 0 , False , True , (111, 74, 0) ), Label( 'ground' , 6 , 255 , 'void' , 0 , False , True , ( 81, 0, 81) ), Label( 'road' , 7 , 0 , 'flat' , 1 , False , False , (128, 64,128) ), Label( 'sidewalk' , 8 , 1 , 'flat' , 1 , False , False , (244, 35,232) ), Label( 'parking' , 9 , 255 , 'flat' , 1 , False , True , (250,170,160) ), Label( 'rail track' , 10 , 255 , 'flat' , 1 , False , True , (230,150,140) ), Label( 'building' , 11 , 2 , 'construction' , 2 , False , False , ( 70, 70, 70) ), Label( 'wall' , 12 , 3 , 'construction' , 2 , False , False , (102,102,156) ), Label( 'fence' , 13 , 4 , 'construction' , 2 , False , False , (190,153,153) ), Label( 'guard rail' , 14 , 255 , 'construction' , 2 , False , True , (180,165,180) ), Label( 'bridge' , 15 , 255 , 'construction' , 2 , False , True , (150,100,100) ), Label( 'tunnel' , 16 , 255 , 'construction' , 2 , False , True , (150,120, 90) ), Label( 'pole' , 17 , 5 , 'object' , 3 , False , False , (153,153,153) ), Label( 'polegroup' , 18 , 255 , 'object' , 3 , False , True , (153,153,153) ), Label( 'traffic light' , 19 , 6 , 'object' , 3 , False , False , (250,170, 30) ), Label( 'traffic sign' , 20 , 7 , 'object' , 3 , False , False , (220,220, 0) ), Label( 'vegetation' , 21 , 8 , 'nature' , 4 , False , False , (107,142, 35) ), Label( 'terrain' , 22 , 9 , 'nature' , 4 , False , False , (152,251,152) ), Label( 'sky' , 23 , 10 , 'sky' , 5 , False , False , ( 70,130,180) ), Label( 'person' , 24 , 11 , 'human' , 6 , True , False , (220, 20, 60) ), Label( 'rider' , 25 , 12 , 'human' , 6 , True , False , (255, 0, 0) ), Label( 'car' , 26 , 13 , 'vehicle' , 7 , True , False , ( 0, 0,142) ), Label( 'truck' , 27 , 14 , 'vehicle' , 7 , True , False , ( 0, 0, 70) ), Label( 'bus' , 28 , 15 , 'vehicle' , 7 , True , False , ( 0, 60,100) ), Label( 'caravan' , 29 , 255 , 'vehicle' , 7 , True , True , ( 0, 0, 90) ), Label( 'trailer' , 30 , 255 , 'vehicle' , 7 , True , True , ( 0, 0,110) ), Label( 'train' , 31 , 16 , 'vehicle' , 7 , True , False , ( 0, 80,100) ), Label( 'motorcycle' , 32 , 17 , 'vehicle' , 7 , True , False , ( 0, 0,230) ), Label( 'bicycle' , 33 , 18 , 'vehicle' , 7 , True , False , (119, 11, 32) ), Label( 'license plate' , -1 , -1 , 'vehicle' , 7 , False , True , ( 0, 0,142) ), Label( 'lane marking' , 34 , 19 , 'vehicle' , 1 , False , True , (192, 64,192) ), ] name2label = { label.name : label for label in labels } id2label = { label.id : label for label in labels } trainId2label = { label.trainId : label for label in reversed(labels) } category2labels = {} for label in labels: category = label.category if category in category2labels: category2labels[category].append(label) else: category2labels[category] = [label] def assureSingleInstanceName( name ): if name in name2label: return name if not name.endswith("group"): return None name = name[:-len("group")] if not name in name2label: return None if not name2label[name].hasInstances: return None return name if __name__ == "__main__": print("List of cityscapes labels:") print("") print(" {:>21} | {:>3} | {:>7} | {:>14} | {:>10} | {:>12} | {:>12}".format( 'name', 'id', 'trainId', 'category', 'categoryId', 'hasInstances', 'ignoreInEval' )) print(" " + ('-' * 98)) for label in labels: print(" {:>21} | {:>3} | {:>7} | {:>14} | {:>10} | {:>12} | {:>12}".format( label.name, label.id, label.trainId, label.category, label.categoryId, label.hasInstances, label.ignoreInEval )) print("") print("Example usages:") name = 'car' id = name2label[name].id print("ID of label '{name}': {id}".format( name=name, id=id )) category = id2label[id].category print("Category of label with ID '{id}': {category}".format( id=id, category=category )) trainId = 0 name = trainId2label[trainId].name print("Name of label with trainID '{id}': {name}".format( id=trainId, name=name ))
true
true
f717a7d323d2826aa7fc0c90ce4ba34258907b10
1,050
py
Python
0301-0400/0391-Perfect Rectangle/0391-Perfect Rectangle.py
jiadaizhao/LeetCode
4ddea0a532fe7c5d053ffbd6870174ec99fc2d60
[ "MIT" ]
49
2018-05-05T02:53:10.000Z
2022-03-30T12:08:09.000Z
0301-0400/0391-Perfect Rectangle/0391-Perfect Rectangle.py
jolly-fellow/LeetCode
ab20b3ec137ed05fad1edda1c30db04ab355486f
[ "MIT" ]
11
2017-12-15T22:31:44.000Z
2020-10-02T12:42:49.000Z
0301-0400/0391-Perfect Rectangle/0391-Perfect Rectangle.py
jolly-fellow/LeetCode
ab20b3ec137ed05fad1edda1c30db04ab355486f
[ "MIT" ]
28
2017-12-05T10:56:51.000Z
2022-01-26T18:18:27.000Z
import math class Solution: def isRectangleCover(self, rectangles: 'List[List[int]]') -> 'bool': area = 0 x1 = y1 = math.inf x2 = y2 = -math.inf table = set() for rec in rectangles: x1 = min(x1, rec[0]) y1 = min(y1, rec[1]) x2 = max(x2, rec[2]) y2 = max(y2, rec[3]) area += (rec[2] - rec[0])*(rec[3] - rec[1]) points = ((rec[0], rec[1]), (rec[0], rec[3]), (rec[2], rec[1]), (rec[2], rec[3])) points = (str(p1) + '_' + str(p2) for p1, p2 in points) for point in points: if point in table: table.remove(point) else: table.add(point) return (len(table) == 4 and str(x1) + '_' + str(y1) in table and str(x1) + '_' + str(y2) in table and str(x2) + '_' + str(y1) in table and str(x2) + '_' + str(y2) in table and area == (x2 - x1)*(y2 - y1))
37.5
93
0.414286
import math class Solution: def isRectangleCover(self, rectangles: 'List[List[int]]') -> 'bool': area = 0 x1 = y1 = math.inf x2 = y2 = -math.inf table = set() for rec in rectangles: x1 = min(x1, rec[0]) y1 = min(y1, rec[1]) x2 = max(x2, rec[2]) y2 = max(y2, rec[3]) area += (rec[2] - rec[0])*(rec[3] - rec[1]) points = ((rec[0], rec[1]), (rec[0], rec[3]), (rec[2], rec[1]), (rec[2], rec[3])) points = (str(p1) + '_' + str(p2) for p1, p2 in points) for point in points: if point in table: table.remove(point) else: table.add(point) return (len(table) == 4 and str(x1) + '_' + str(y1) in table and str(x1) + '_' + str(y2) in table and str(x2) + '_' + str(y1) in table and str(x2) + '_' + str(y2) in table and area == (x2 - x1)*(y2 - y1))
true
true
f717a8e5702adf2484d162a384d9cd7e1c5d590d
2,944
py
Python
tests/test_sftp.py
lfbos/qftplib
2a1b79c7f36cc2dfd139a7a3d179fb23f7c6dc1f
[ "MIT" ]
null
null
null
tests/test_sftp.py
lfbos/qftplib
2a1b79c7f36cc2dfd139a7a3d179fb23f7c6dc1f
[ "MIT" ]
null
null
null
tests/test_sftp.py
lfbos/qftplib
2a1b79c7f36cc2dfd139a7a3d179fb23f7c6dc1f
[ "MIT" ]
null
null
null
import os import tempfile import time import unittest from qftplib.client import FTPClient class FTPTest(unittest.TestCase): host = os.environ.get('SFTP_HOST_TEST') user = os.environ.get('SFTP_USER_TEST') password = os.environ.get('SFTP_PASS_TEST') dir = os.environ.get('SFTP_DIR_TEST') port = 22 def test_name(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: assert str(sftp) == f'<FTPClient({self.host}, {self.user}, xxxx, port = {self.port}, protocol = sftp)>' def test_listdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: directory_files = sftp.listdir(self.dir) assert len(directory_files) > 0 def test_pwd_chdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) assert sftp.pwd() == f'/home/{self.user}/{self.dir}' def test_rmdir_mkdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) new_dir = str(int(time.time())) sftp.mkdir(new_dir) files = sftp.listdir('.') assert f'{new_dir}' in files sftp.rmdir(f'/home/{self.user}/{self.dir}/{new_dir}') files = sftp.listdir('.') assert f'{new_dir}' not in files def test_rename(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) new_dir = str(int(time.time())) sftp.mkdir(new_dir) sftp.rename(f'/home/{self.user}/{self.dir}/{new_dir}', f'/home/{self.user}/{self.dir}/{new_dir}_renamed') files = sftp.listdir('.') assert f'{new_dir}_renamed' in files sftp.rmdir(f'/home/{self.user}/{self.dir}/{new_dir}_renamed') def test_put_delete(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp, tempfile.NamedTemporaryFile() as f: remotepath = os.path.join(self.dir, os.path.basename(f.name)) sftp.put(f.name, remotepath) sftp.delete(remotepath) files = sftp.listdir(self.dir) assert remotepath not in files def test_get(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp, tempfile.NamedTemporaryFile(mode='w') as f: files = sftp.listdir(self.dir) filename = next((f for f in files if f.endswith('.csv') or f.endswith('.txt')), None) sftp.get(f'/home/{self.user}/{self.dir}/{filename}', f.name) assert os.path.exists(f.name) assert os.path.isfile(f.name)
35.047619
117
0.605978
import os import tempfile import time import unittest from qftplib.client import FTPClient class FTPTest(unittest.TestCase): host = os.environ.get('SFTP_HOST_TEST') user = os.environ.get('SFTP_USER_TEST') password = os.environ.get('SFTP_PASS_TEST') dir = os.environ.get('SFTP_DIR_TEST') port = 22 def test_name(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: assert str(sftp) == f'<FTPClient({self.host}, {self.user}, xxxx, port = {self.port}, protocol = sftp)>' def test_listdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: directory_files = sftp.listdir(self.dir) assert len(directory_files) > 0 def test_pwd_chdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) assert sftp.pwd() == f'/home/{self.user}/{self.dir}' def test_rmdir_mkdir(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) new_dir = str(int(time.time())) sftp.mkdir(new_dir) files = sftp.listdir('.') assert f'{new_dir}' in files sftp.rmdir(f'/home/{self.user}/{self.dir}/{new_dir}') files = sftp.listdir('.') assert f'{new_dir}' not in files def test_rename(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp: sftp.chdir(self.dir) new_dir = str(int(time.time())) sftp.mkdir(new_dir) sftp.rename(f'/home/{self.user}/{self.dir}/{new_dir}', f'/home/{self.user}/{self.dir}/{new_dir}_renamed') files = sftp.listdir('.') assert f'{new_dir}_renamed' in files sftp.rmdir(f'/home/{self.user}/{self.dir}/{new_dir}_renamed') def test_put_delete(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp, tempfile.NamedTemporaryFile() as f: remotepath = os.path.join(self.dir, os.path.basename(f.name)) sftp.put(f.name, remotepath) sftp.delete(remotepath) files = sftp.listdir(self.dir) assert remotepath not in files def test_get(self): with FTPClient(self.host, self.user, self.password, port=self.port, protocol='sftp')as sftp, tempfile.NamedTemporaryFile(mode='w') as f: files = sftp.listdir(self.dir) filename = next((f for f in files if f.endswith('.csv') or f.endswith('.txt')), None) sftp.get(f'/home/{self.user}/{self.dir}/{filename}', f.name) assert os.path.exists(f.name) assert os.path.isfile(f.name)
true
true
f717a96bf2f36b05dfb39bd231a309d8afe03392
723
py
Python
django-rest-framework-quick-start/tutorial/quickstart/views.py
madhikarma/python-django-demo
d18ca7e25d01f12813cd6e011ae253e792ac2d03
[ "MIT" ]
null
null
null
django-rest-framework-quick-start/tutorial/quickstart/views.py
madhikarma/python-django-demo
d18ca7e25d01f12813cd6e011ae253e792ac2d03
[ "MIT" ]
null
null
null
django-rest-framework-quick-start/tutorial/quickstart/views.py
madhikarma/python-django-demo
d18ca7e25d01f12813cd6e011ae253e792ac2d03
[ "MIT" ]
null
null
null
from django.contrib.auth.models import User, Group from rest_framework import viewsets from rest_framework import permissions from tutorial.quickstart.serializers import UserSerializer, GroupSerializer class UserViewSet(viewsets.ModelViewSet): """ API endpoint that allows users to be viewed or edited. """ queryset = User.objects.all().order_by('-date_joined') serializer_class = UserSerializer permission_classes = [permissions.IsAuthenticated] class GroupViewSet(viewsets.ModelViewSet): """ API endpoint that allows groups to be viewed or edited. """ queryset = Group.objects.all() serializer_class = GroupSerializer permission_classes = [permissions.IsAuthenticated]
31.434783
75
0.763485
from django.contrib.auth.models import User, Group from rest_framework import viewsets from rest_framework import permissions from tutorial.quickstart.serializers import UserSerializer, GroupSerializer class UserViewSet(viewsets.ModelViewSet): queryset = User.objects.all().order_by('-date_joined') serializer_class = UserSerializer permission_classes = [permissions.IsAuthenticated] class GroupViewSet(viewsets.ModelViewSet): queryset = Group.objects.all() serializer_class = GroupSerializer permission_classes = [permissions.IsAuthenticated]
true
true
f717a97ac0ae713afc793d70fe3a8c677d331876
767
py
Python
yatube/posts/migrations/0004_auto_20210318_1753.py
Rusich90/yatube
8593443fe64b0422b2ebbedc9966d347a668e0a3
[ "BSD-3-Clause" ]
1
2021-06-06T11:15:57.000Z
2021-06-06T11:15:57.000Z
yatube/posts/migrations/0004_auto_20210318_1753.py
Rusich90/yatube
8593443fe64b0422b2ebbedc9966d347a668e0a3
[ "BSD-3-Clause" ]
null
null
null
yatube/posts/migrations/0004_auto_20210318_1753.py
Rusich90/yatube
8593443fe64b0422b2ebbedc9966d347a668e0a3
[ "BSD-3-Clause" ]
1
2021-06-06T11:15:58.000Z
2021-06-06T11:15:58.000Z
# Generated by Django 2.2 on 2021-03-18 17:53 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('posts', '0003_auto_20210307_1711'), ] operations = [ migrations.AlterField( model_name='post', name='group', field=models.ForeignKey(blank=True, help_text='Выберите группу для поста (необязательно)', null=True, on_delete=django.db.models.deletion.CASCADE, related_name='posts', to='posts.Group', verbose_name='Группа'), ), migrations.AlterField( model_name='post', name='text', field=models.TextField(help_text='Введите текст', verbose_name='Текст'), ), ]
30.68
222
0.636245
from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('posts', '0003_auto_20210307_1711'), ] operations = [ migrations.AlterField( model_name='post', name='group', field=models.ForeignKey(blank=True, help_text='Выберите группу для поста (необязательно)', null=True, on_delete=django.db.models.deletion.CASCADE, related_name='posts', to='posts.Group', verbose_name='Группа'), ), migrations.AlterField( model_name='post', name='text', field=models.TextField(help_text='Введите текст', verbose_name='Текст'), ), ]
true
true
f717aa4a2dd6cc01429db2f797d20ef18a428275
8,080
py
Python
neutron/services/trunk/rpc/server.py
knodir/neutron
ac4e28478ac8a8a0c9f5c5785f6a6bcf532c66b8
[ "Apache-2.0" ]
null
null
null
neutron/services/trunk/rpc/server.py
knodir/neutron
ac4e28478ac8a8a0c9f5c5785f6a6bcf532c66b8
[ "Apache-2.0" ]
5
2019-08-14T06:46:03.000Z
2021-12-13T20:01:25.000Z
neutron/services/trunk/rpc/server.py
knodir/neutron
ac4e28478ac8a8a0c9f5c5785f6a6bcf532c66b8
[ "Apache-2.0" ]
2
2020-03-15T01:24:15.000Z
2020-07-22T20:34:26.000Z
# Copyright 2016 Hewlett Packard Enterprise Development LP # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import collections from neutron_lib.api.definitions import portbindings from neutron_lib.db import api as db_api from neutron_lib.plugins import directory from neutron_lib import rpc as n_rpc from oslo_log import helpers as log_helpers from oslo_log import log as logging import oslo_messaging from neutron.api.rpc.callbacks import events from neutron.api.rpc.callbacks.producer import registry from neutron.api.rpc.callbacks import resources from neutron.api.rpc.handlers import resources_rpc from neutron.objects import trunk as trunk_objects from neutron.services.trunk import constants as trunk_consts from neutron.services.trunk import exceptions as trunk_exc from neutron.services.trunk.rpc import constants LOG = logging.getLogger(__name__) # This module contains stub (client-side) and skeleton (server-side) # proxy code that executes in the Neutron server process space. This # is needed if any of the trunk service plugin drivers has a remote # component (e.g. agent), that needs to communicate with the Neutron # Server. # The Server side exposes the following remote methods: # # - lookup method to retrieve trunk details: used by the agent to learn # about the trunk. # - update methods for trunk and its subports: used by the agent to # inform the server about local trunk status changes. # # For agent-side stub and skeleton proxy code, please look at agent.py def trunk_by_port_provider(resource, port_id, context, **kwargs): """Provider callback to supply trunk information by parent port.""" return trunk_objects.Trunk.get_object(context, port_id=port_id) class TrunkSkeleton(object): """Skeleton proxy code for agent->server communication.""" # API version history: # 1.0 Initial version target = oslo_messaging.Target(version='1.0', namespace=constants.TRUNK_BASE_NAMESPACE) _core_plugin = None def __init__(self): # Used to provide trunk lookups for the agent. registry.provide(trunk_by_port_provider, resources.TRUNK) self._connection = n_rpc.Connection() self._connection.create_consumer( constants.TRUNK_BASE_TOPIC, [self], fanout=False) self._connection.consume_in_threads() @property def core_plugin(self): if not self._core_plugin: self._core_plugin = directory.get_plugin() return self._core_plugin @log_helpers.log_method_call def update_subport_bindings(self, context, subports): """Update subport bindings to match trunk host binding.""" el = context.elevated() ports_by_trunk_id = collections.defaultdict(list) updated_ports = collections.defaultdict(list) for s in subports: ports_by_trunk_id[s['trunk_id']].append(s['port_id']) for trunk_id, subport_ids in ports_by_trunk_id.items(): trunk = trunk_objects.Trunk.get_object(el, id=trunk_id) if not trunk: LOG.debug("Trunk not found. id: %s", trunk_id) continue trunk_updated_ports = self._process_trunk_subport_bindings( el, trunk, subport_ids) updated_ports[trunk.id].extend(trunk_updated_ports) return updated_ports def update_trunk_status(self, context, trunk_id, status): """Update the trunk status to reflect outcome of data plane wiring.""" with db_api.autonested_transaction(context.session): trunk = trunk_objects.Trunk.get_object(context, id=trunk_id) if trunk: trunk.update(status=status) def _process_trunk_subport_bindings(self, context, trunk, port_ids): """Process port bindings for subports on the given trunk.""" updated_ports = [] trunk_port_id = trunk.port_id trunk_port = self.core_plugin.get_port(context, trunk_port_id) trunk_host = trunk_port.get(portbindings.HOST_ID) # NOTE(status_police) Set the trunk in BUILD state before processing # subport bindings. The trunk will stay in BUILD state until an # attempt has been made to bind all subports passed here and the # agent acknowledges the operation was successful. trunk.update(status=trunk_consts.BUILD_STATUS) for port_id in port_ids: try: updated_port = self._handle_port_binding(context, port_id, trunk, trunk_host) # NOTE(fitoduarte): consider trimming down the content # of the port data structure. updated_ports.append(updated_port) except trunk_exc.SubPortBindingError as e: LOG.error("Failed to bind subport: %s", e) # NOTE(status_police) The subport binding has failed in a # manner in which we cannot proceed and the user must take # action to bring the trunk back to a sane state. trunk.update(status=trunk_consts.ERROR_STATUS) return [] except Exception as e: msg = ("Failed to bind subport port %(port)s on trunk " "%(trunk)s: %(exc)s") LOG.error(msg, {'port': port_id, 'trunk': trunk.id, 'exc': e}) if len(port_ids) != len(updated_ports): trunk.update(status=trunk_consts.DEGRADED_STATUS) return updated_ports def _handle_port_binding(self, context, port_id, trunk, trunk_host): """Bind the given port to the given host. :param context: The context to use for the operation :param port_id: The UUID of the port to be bound :param trunk: The trunk that the given port belongs to :param trunk_host: The host to bind the given port to """ port = self.core_plugin.update_port( context, port_id, {'port': {portbindings.HOST_ID: trunk_host, 'device_owner': trunk_consts.TRUNK_SUBPORT_OWNER}}) vif_type = port.get(portbindings.VIF_TYPE) if vif_type == portbindings.VIF_TYPE_BINDING_FAILED: raise trunk_exc.SubPortBindingError(port_id=port_id, trunk_id=trunk.id) return port class TrunkStub(object): """Stub proxy code for server->agent communication.""" def __init__(self): self._resource_rpc = resources_rpc.ResourcesPushRpcApi() @log_helpers.log_method_call def trunk_created(self, context, trunk): """Tell the agent about a trunk being created.""" self._resource_rpc.push(context, [trunk], events.CREATED) @log_helpers.log_method_call def trunk_deleted(self, context, trunk): """Tell the agent about a trunk being deleted.""" self._resource_rpc.push(context, [trunk], events.DELETED) @log_helpers.log_method_call def subports_added(self, context, subports): """Tell the agent about new subports to add.""" self._resource_rpc.push(context, subports, events.CREATED) @log_helpers.log_method_call def subports_deleted(self, context, subports): """Tell the agent about existing subports to remove.""" self._resource_rpc.push(context, subports, events.DELETED)
41.865285
78
0.664604
import collections from neutron_lib.api.definitions import portbindings from neutron_lib.db import api as db_api from neutron_lib.plugins import directory from neutron_lib import rpc as n_rpc from oslo_log import helpers as log_helpers from oslo_log import log as logging import oslo_messaging from neutron.api.rpc.callbacks import events from neutron.api.rpc.callbacks.producer import registry from neutron.api.rpc.callbacks import resources from neutron.api.rpc.handlers import resources_rpc from neutron.objects import trunk as trunk_objects from neutron.services.trunk import constants as trunk_consts from neutron.services.trunk import exceptions as trunk_exc from neutron.services.trunk.rpc import constants LOG = logging.getLogger(__name__) def trunk_by_port_provider(resource, port_id, context, **kwargs): return trunk_objects.Trunk.get_object(context, port_id=port_id) class TrunkSkeleton(object): target = oslo_messaging.Target(version='1.0', namespace=constants.TRUNK_BASE_NAMESPACE) _core_plugin = None def __init__(self): registry.provide(trunk_by_port_provider, resources.TRUNK) self._connection = n_rpc.Connection() self._connection.create_consumer( constants.TRUNK_BASE_TOPIC, [self], fanout=False) self._connection.consume_in_threads() @property def core_plugin(self): if not self._core_plugin: self._core_plugin = directory.get_plugin() return self._core_plugin @log_helpers.log_method_call def update_subport_bindings(self, context, subports): el = context.elevated() ports_by_trunk_id = collections.defaultdict(list) updated_ports = collections.defaultdict(list) for s in subports: ports_by_trunk_id[s['trunk_id']].append(s['port_id']) for trunk_id, subport_ids in ports_by_trunk_id.items(): trunk = trunk_objects.Trunk.get_object(el, id=trunk_id) if not trunk: LOG.debug("Trunk not found. id: %s", trunk_id) continue trunk_updated_ports = self._process_trunk_subport_bindings( el, trunk, subport_ids) updated_ports[trunk.id].extend(trunk_updated_ports) return updated_ports def update_trunk_status(self, context, trunk_id, status): with db_api.autonested_transaction(context.session): trunk = trunk_objects.Trunk.get_object(context, id=trunk_id) if trunk: trunk.update(status=status) def _process_trunk_subport_bindings(self, context, trunk, port_ids): updated_ports = [] trunk_port_id = trunk.port_id trunk_port = self.core_plugin.get_port(context, trunk_port_id) trunk_host = trunk_port.get(portbindings.HOST_ID) trunk.update(status=trunk_consts.BUILD_STATUS) for port_id in port_ids: try: updated_port = self._handle_port_binding(context, port_id, trunk, trunk_host) updated_ports.append(updated_port) except trunk_exc.SubPortBindingError as e: LOG.error("Failed to bind subport: %s", e) trunk.update(status=trunk_consts.ERROR_STATUS) return [] except Exception as e: msg = ("Failed to bind subport port %(port)s on trunk " "%(trunk)s: %(exc)s") LOG.error(msg, {'port': port_id, 'trunk': trunk.id, 'exc': e}) if len(port_ids) != len(updated_ports): trunk.update(status=trunk_consts.DEGRADED_STATUS) return updated_ports def _handle_port_binding(self, context, port_id, trunk, trunk_host): port = self.core_plugin.update_port( context, port_id, {'port': {portbindings.HOST_ID: trunk_host, 'device_owner': trunk_consts.TRUNK_SUBPORT_OWNER}}) vif_type = port.get(portbindings.VIF_TYPE) if vif_type == portbindings.VIF_TYPE_BINDING_FAILED: raise trunk_exc.SubPortBindingError(port_id=port_id, trunk_id=trunk.id) return port class TrunkStub(object): def __init__(self): self._resource_rpc = resources_rpc.ResourcesPushRpcApi() @log_helpers.log_method_call def trunk_created(self, context, trunk): self._resource_rpc.push(context, [trunk], events.CREATED) @log_helpers.log_method_call def trunk_deleted(self, context, trunk): self._resource_rpc.push(context, [trunk], events.DELETED) @log_helpers.log_method_call def subports_added(self, context, subports): self._resource_rpc.push(context, subports, events.CREATED) @log_helpers.log_method_call def subports_deleted(self, context, subports): self._resource_rpc.push(context, subports, events.DELETED)
true
true
f717aaa141f06ebfe41236718707dffb420687d5
11,262
py
Python
mars/worker/service.py
wdkwyf/mars
3f750e360e64380eab779301a5103994d4886b6a
[ "Apache-2.0" ]
null
null
null
mars/worker/service.py
wdkwyf/mars
3f750e360e64380eab779301a5103994d4886b6a
[ "Apache-2.0" ]
null
null
null
mars/worker/service.py
wdkwyf/mars
3f750e360e64380eab779301a5103994d4886b6a
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # Copyright 1999-2018 Alibaba Group Holding Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import logging try: from pyarrow import plasma except ImportError: # pragma: no cover plasma = None from ..config import options from .. import resource from ..utils import parse_readable_size, readable_size from ..compat import six from .status import StatusActor from .quota import QuotaActor, MemQuotaActor from .chunkholder import ChunkHolderActor from .dispatcher import DispatchActor from .events import EventsActor from .execution import ExecutionActor from .calc import CpuCalcActor from .transfer import ReceiverActor, SenderActor from .prochelper import ProcessHelperActor from .transfer import ResultSenderActor from .spill import SpillActor from .utils import WorkerClusterInfoActor logger = logging.getLogger(__name__) class WorkerService(object): def __init__(self, **kwargs): self._plasma_store = None self._chunk_holder_ref = None self._task_queue_ref = None self._mem_quota_ref = None self._dispatch_ref = None self._events_ref = None self._status_ref = None self._execution_ref = None self._daemon_ref = None self._cluster_info_ref = None self._cpu_calc_actors = [] self._sender_actors = [] self._receiver_actors = [] self._spill_actors = [] self._process_helper_actors = [] self._result_sender_ref = None self._advertise_addr = kwargs.pop('advertise_addr', None) self._n_cpu_process = int(kwargs.pop('n_cpu_process', None) or resource.cpu_count()) self._n_net_process = int(kwargs.pop('n_net_process', None) or '1') self._spill_dirs = kwargs.pop('spill_dirs', None) if self._spill_dirs: if isinstance(self._spill_dirs, six.string_types): from .spill import parse_spill_dirs self._spill_dirs = options.worker.spill_directory = parse_spill_dirs(self._spill_dirs) else: options.worker.spill_directory = self._spill_dirs else: self._spill_dirs = options.worker.spill_directory = [] options.worker.disk_compression = kwargs.pop('disk_compression', None) or \ options.worker.disk_compression options.worker.transfer_compression = kwargs.pop('transfer_compression', None) or \ options.worker.transfer_compression self._total_mem = kwargs.pop('total_mem', None) self._cache_mem_limit = kwargs.pop('cache_mem_limit', None) self._soft_mem_limit = kwargs.pop('soft_mem_limit', None) or '80%' self._hard_mem_limit = kwargs.pop('hard_mem_limit', None) or '90%' self._ignore_avail_mem = kwargs.pop('ignore_avail_mem', None) or False self._min_mem_size = kwargs.pop('min_mem_size', None) or 128 * 1024 ** 2 self._soft_quota_limit = self._soft_mem_limit self._calc_memory_limits() if kwargs: # pragma: no cover raise TypeError('Keyword arguments %r cannot be recognized.' % ', '.join(kwargs)) @property def n_process(self): return 1 + self._n_cpu_process + self._n_net_process + (1 if self._spill_dirs else 0) def _calc_memory_limits(self): def _calc_size_limit(limit_str, total_size): if limit_str is None: return None if isinstance(limit_str, int): return limit_str mem_limit, is_percent = parse_readable_size(limit_str) if is_percent: return int(total_size * mem_limit) else: return int(mem_limit) mem_stats = resource.virtual_memory() if self._total_mem: self._total_mem = _calc_size_limit(self._total_mem, mem_stats.total) else: self._total_mem = mem_stats.total self._min_mem_size = _calc_size_limit(self._min_mem_size, self._total_mem) self._hard_mem_limit = _calc_size_limit(self._hard_mem_limit, self._total_mem) self._cache_mem_limit = _calc_size_limit(self._cache_mem_limit, self._total_mem) if self._cache_mem_limit is None: self._cache_mem_limit = mem_stats.free // 2 self._soft_mem_limit = _calc_size_limit(self._soft_mem_limit, self._total_mem) actual_used = self._total_mem - mem_stats.available if self._ignore_avail_mem: self._soft_quota_limit = self._soft_mem_limit else: self._soft_quota_limit = self._soft_mem_limit - self._cache_mem_limit - actual_used if self._soft_quota_limit < self._min_mem_size: raise MemoryError('Memory not enough. soft_limit=%s, cache_limit=%s, used=%s' % tuple(readable_size(k) for k in ( self._soft_mem_limit, self._cache_mem_limit, actual_used))) logger.info('Setting soft limit to %s.', readable_size(self._soft_quota_limit)) def start_plasma(self): self._plasma_store = plasma.start_plasma_store(self._cache_mem_limit) options.worker.plasma_socket, _ = self._plasma_store.__enter__() def start(self, endpoint, pool, distributed=True, discoverer=None, process_start_index=0): # create plasma key mapper from .chunkstore import PlasmaKeyMapActor pool.create_actor(PlasmaKeyMapActor, uid=PlasmaKeyMapActor.default_uid()) # create WorkerClusterInfoActor self._cluster_info_ref = pool.create_actor( WorkerClusterInfoActor, discoverer, uid=WorkerClusterInfoActor.default_uid()) if distributed: # create process daemon from .daemon import WorkerDaemonActor actor_holder = self._daemon_ref = pool.create_actor( WorkerDaemonActor, uid=WorkerDaemonActor.default_uid()) # create StatusActor if ':' not in self._advertise_addr: self._advertise_addr += ':' + endpoint.rsplit(':', 1)[-1] self._status_ref = pool.create_actor( StatusActor, self._advertise_addr, uid=StatusActor.default_uid()) else: # create StatusActor self._status_ref = pool.create_actor( StatusActor, endpoint, uid=StatusActor.default_uid()) actor_holder = pool if self._ignore_avail_mem: # start a QuotaActor instead of MemQuotaActor to avoid memory size detection # for debug purpose only, DON'T USE IN PRODUCTION self._mem_quota_ref = pool.create_actor( QuotaActor, self._soft_mem_limit, uid=MemQuotaActor.default_uid()) else: self._mem_quota_ref = pool.create_actor( MemQuotaActor, self._soft_quota_limit, self._hard_mem_limit, uid=MemQuotaActor.default_uid()) # create ChunkHolderActor self._chunk_holder_ref = pool.create_actor( ChunkHolderActor, self._cache_mem_limit, uid=ChunkHolderActor.default_uid()) # create DispatchActor self._dispatch_ref = pool.create_actor(DispatchActor, uid=DispatchActor.default_uid()) # create EventsActor self._events_ref = pool.create_actor(EventsActor, uid=EventsActor.default_uid()) # create ExecutionActor self._execution_ref = pool.create_actor(ExecutionActor, uid=ExecutionActor.default_uid()) # create CpuCalcActor if not distributed: self._n_cpu_process = pool.cluster_info.n_process - 1 - process_start_index for cpu_id in range(self._n_cpu_process): uid = 'w:%d:mars-calc-%d-%d' % (cpu_id + 1, os.getpid(), cpu_id) actor = actor_holder.create_actor(CpuCalcActor, uid=uid) self._cpu_calc_actors.append(actor) start_pid = 1 + process_start_index + self._n_cpu_process if distributed: # create SenderActor and ReceiverActor for sender_id in range(self._n_net_process): uid = 'w:%d:mars-sender-%d-%d' % (start_pid + sender_id, os.getpid(), sender_id) actor = actor_holder.create_actor(SenderActor, uid=uid) self._sender_actors.append(actor) # Mutable requires ReceiverActor (with LocalClusterSession) for receiver_id in range(2 * self._n_net_process): uid = 'w:%d:mars-receiver-%d-%d' % (start_pid + receiver_id // 2, os.getpid(), receiver_id) actor = actor_holder.create_actor(ReceiverActor, uid=uid) self._receiver_actors.append(actor) # create ProcessHelperActor for proc_id in range(pool.cluster_info.n_process - process_start_index): uid = 'w:%d:mars-process-helper-%d-%d' % (proc_id, os.getpid(), proc_id) actor = actor_holder.create_actor(ProcessHelperActor, uid=uid) self._process_helper_actors.append(actor) # create ResultSenderActor self._result_sender_ref = pool.create_actor(ResultSenderActor, uid=ResultSenderActor.default_uid()) # create SpillActor start_pid = pool.cluster_info.n_process - 1 if options.worker.spill_directory: for spill_id in range(len(options.worker.spill_directory) * 2): uid = 'w:%d:mars-spill-%d-%d' % (start_pid, os.getpid(), spill_id) actor = actor_holder.create_actor(SpillActor, uid=uid) self._spill_actors.append(actor) # worker can be registered when everything is ready self._status_ref.enable_status_upload(_tell=True) def handle_process_down(self, pool, proc_indices): logger.debug('Process %r halt. Trying to recover.', proc_indices) for pid in proc_indices: pool.restart_process(pid) self._daemon_ref.handle_process_down(proc_indices, _tell=True) def stop(self): try: if self._result_sender_ref: self._result_sender_ref.destroy(wait=False) if self._status_ref: self._status_ref.destroy(wait=False) if self._chunk_holder_ref: self._chunk_holder_ref.destroy(wait=False) if self._events_ref: self._events_ref.destroy(wait=False) if self._dispatch_ref: self._dispatch_ref.destroy(wait=False) if self._execution_ref: self._execution_ref.destroy(wait=False) for actor in (self._cpu_calc_actors + self._sender_actors + self._receiver_actors + self._spill_actors + self._process_helper_actors): actor.destroy(wait=False) finally: self._plasma_store.__exit__(None, None, None)
42.659091
109
0.665335
import os import logging try: from pyarrow import plasma except ImportError: plasma = None from ..config import options from .. import resource from ..utils import parse_readable_size, readable_size from ..compat import six from .status import StatusActor from .quota import QuotaActor, MemQuotaActor from .chunkholder import ChunkHolderActor from .dispatcher import DispatchActor from .events import EventsActor from .execution import ExecutionActor from .calc import CpuCalcActor from .transfer import ReceiverActor, SenderActor from .prochelper import ProcessHelperActor from .transfer import ResultSenderActor from .spill import SpillActor from .utils import WorkerClusterInfoActor logger = logging.getLogger(__name__) class WorkerService(object): def __init__(self, **kwargs): self._plasma_store = None self._chunk_holder_ref = None self._task_queue_ref = None self._mem_quota_ref = None self._dispatch_ref = None self._events_ref = None self._status_ref = None self._execution_ref = None self._daemon_ref = None self._cluster_info_ref = None self._cpu_calc_actors = [] self._sender_actors = [] self._receiver_actors = [] self._spill_actors = [] self._process_helper_actors = [] self._result_sender_ref = None self._advertise_addr = kwargs.pop('advertise_addr', None) self._n_cpu_process = int(kwargs.pop('n_cpu_process', None) or resource.cpu_count()) self._n_net_process = int(kwargs.pop('n_net_process', None) or '1') self._spill_dirs = kwargs.pop('spill_dirs', None) if self._spill_dirs: if isinstance(self._spill_dirs, six.string_types): from .spill import parse_spill_dirs self._spill_dirs = options.worker.spill_directory = parse_spill_dirs(self._spill_dirs) else: options.worker.spill_directory = self._spill_dirs else: self._spill_dirs = options.worker.spill_directory = [] options.worker.disk_compression = kwargs.pop('disk_compression', None) or \ options.worker.disk_compression options.worker.transfer_compression = kwargs.pop('transfer_compression', None) or \ options.worker.transfer_compression self._total_mem = kwargs.pop('total_mem', None) self._cache_mem_limit = kwargs.pop('cache_mem_limit', None) self._soft_mem_limit = kwargs.pop('soft_mem_limit', None) or '80%' self._hard_mem_limit = kwargs.pop('hard_mem_limit', None) or '90%' self._ignore_avail_mem = kwargs.pop('ignore_avail_mem', None) or False self._min_mem_size = kwargs.pop('min_mem_size', None) or 128 * 1024 ** 2 self._soft_quota_limit = self._soft_mem_limit self._calc_memory_limits() if kwargs: raise TypeError('Keyword arguments %r cannot be recognized.' % ', '.join(kwargs)) @property def n_process(self): return 1 + self._n_cpu_process + self._n_net_process + (1 if self._spill_dirs else 0) def _calc_memory_limits(self): def _calc_size_limit(limit_str, total_size): if limit_str is None: return None if isinstance(limit_str, int): return limit_str mem_limit, is_percent = parse_readable_size(limit_str) if is_percent: return int(total_size * mem_limit) else: return int(mem_limit) mem_stats = resource.virtual_memory() if self._total_mem: self._total_mem = _calc_size_limit(self._total_mem, mem_stats.total) else: self._total_mem = mem_stats.total self._min_mem_size = _calc_size_limit(self._min_mem_size, self._total_mem) self._hard_mem_limit = _calc_size_limit(self._hard_mem_limit, self._total_mem) self._cache_mem_limit = _calc_size_limit(self._cache_mem_limit, self._total_mem) if self._cache_mem_limit is None: self._cache_mem_limit = mem_stats.free // 2 self._soft_mem_limit = _calc_size_limit(self._soft_mem_limit, self._total_mem) actual_used = self._total_mem - mem_stats.available if self._ignore_avail_mem: self._soft_quota_limit = self._soft_mem_limit else: self._soft_quota_limit = self._soft_mem_limit - self._cache_mem_limit - actual_used if self._soft_quota_limit < self._min_mem_size: raise MemoryError('Memory not enough. soft_limit=%s, cache_limit=%s, used=%s' % tuple(readable_size(k) for k in ( self._soft_mem_limit, self._cache_mem_limit, actual_used))) logger.info('Setting soft limit to %s.', readable_size(self._soft_quota_limit)) def start_plasma(self): self._plasma_store = plasma.start_plasma_store(self._cache_mem_limit) options.worker.plasma_socket, _ = self._plasma_store.__enter__() def start(self, endpoint, pool, distributed=True, discoverer=None, process_start_index=0): from .chunkstore import PlasmaKeyMapActor pool.create_actor(PlasmaKeyMapActor, uid=PlasmaKeyMapActor.default_uid()) self._cluster_info_ref = pool.create_actor( WorkerClusterInfoActor, discoverer, uid=WorkerClusterInfoActor.default_uid()) if distributed: from .daemon import WorkerDaemonActor actor_holder = self._daemon_ref = pool.create_actor( WorkerDaemonActor, uid=WorkerDaemonActor.default_uid()) if ':' not in self._advertise_addr: self._advertise_addr += ':' + endpoint.rsplit(':', 1)[-1] self._status_ref = pool.create_actor( StatusActor, self._advertise_addr, uid=StatusActor.default_uid()) else: self._status_ref = pool.create_actor( StatusActor, endpoint, uid=StatusActor.default_uid()) actor_holder = pool if self._ignore_avail_mem: self._mem_quota_ref = pool.create_actor( QuotaActor, self._soft_mem_limit, uid=MemQuotaActor.default_uid()) else: self._mem_quota_ref = pool.create_actor( MemQuotaActor, self._soft_quota_limit, self._hard_mem_limit, uid=MemQuotaActor.default_uid()) # create ChunkHolderActor self._chunk_holder_ref = pool.create_actor( ChunkHolderActor, self._cache_mem_limit, uid=ChunkHolderActor.default_uid()) # create DispatchActor self._dispatch_ref = pool.create_actor(DispatchActor, uid=DispatchActor.default_uid()) # create EventsActor self._events_ref = pool.create_actor(EventsActor, uid=EventsActor.default_uid()) # create ExecutionActor self._execution_ref = pool.create_actor(ExecutionActor, uid=ExecutionActor.default_uid()) # create CpuCalcActor if not distributed: self._n_cpu_process = pool.cluster_info.n_process - 1 - process_start_index for cpu_id in range(self._n_cpu_process): uid = 'w:%d:mars-calc-%d-%d' % (cpu_id + 1, os.getpid(), cpu_id) actor = actor_holder.create_actor(CpuCalcActor, uid=uid) self._cpu_calc_actors.append(actor) start_pid = 1 + process_start_index + self._n_cpu_process if distributed: # create SenderActor and ReceiverActor for sender_id in range(self._n_net_process): uid = 'w:%d:mars-sender-%d-%d' % (start_pid + sender_id, os.getpid(), sender_id) actor = actor_holder.create_actor(SenderActor, uid=uid) self._sender_actors.append(actor) # Mutable requires ReceiverActor (with LocalClusterSession) for receiver_id in range(2 * self._n_net_process): uid = 'w:%d:mars-receiver-%d-%d' % (start_pid + receiver_id // 2, os.getpid(), receiver_id) actor = actor_holder.create_actor(ReceiverActor, uid=uid) self._receiver_actors.append(actor) # create ProcessHelperActor for proc_id in range(pool.cluster_info.n_process - process_start_index): uid = 'w:%d:mars-process-helper-%d-%d' % (proc_id, os.getpid(), proc_id) actor = actor_holder.create_actor(ProcessHelperActor, uid=uid) self._process_helper_actors.append(actor) # create ResultSenderActor self._result_sender_ref = pool.create_actor(ResultSenderActor, uid=ResultSenderActor.default_uid()) # create SpillActor start_pid = pool.cluster_info.n_process - 1 if options.worker.spill_directory: for spill_id in range(len(options.worker.spill_directory) * 2): uid = 'w:%d:mars-spill-%d-%d' % (start_pid, os.getpid(), spill_id) actor = actor_holder.create_actor(SpillActor, uid=uid) self._spill_actors.append(actor) # worker can be registered when everything is ready self._status_ref.enable_status_upload(_tell=True) def handle_process_down(self, pool, proc_indices): logger.debug('Process %r halt. Trying to recover.', proc_indices) for pid in proc_indices: pool.restart_process(pid) self._daemon_ref.handle_process_down(proc_indices, _tell=True) def stop(self): try: if self._result_sender_ref: self._result_sender_ref.destroy(wait=False) if self._status_ref: self._status_ref.destroy(wait=False) if self._chunk_holder_ref: self._chunk_holder_ref.destroy(wait=False) if self._events_ref: self._events_ref.destroy(wait=False) if self._dispatch_ref: self._dispatch_ref.destroy(wait=False) if self._execution_ref: self._execution_ref.destroy(wait=False) for actor in (self._cpu_calc_actors + self._sender_actors + self._receiver_actors + self._spill_actors + self._process_helper_actors): actor.destroy(wait=False) finally: self._plasma_store.__exit__(None, None, None)
true
true
f717ac92ce24bf53f59cb78906005b06e6b861f3
483
py
Python
final_project/machinetranslation/tests/test_translator.py
BrunoYH/xzceb-flask_eng_fr
a86a32ef56fcae1a8a9e281f17eb73dca05aec10
[ "Apache-2.0" ]
null
null
null
final_project/machinetranslation/tests/test_translator.py
BrunoYH/xzceb-flask_eng_fr
a86a32ef56fcae1a8a9e281f17eb73dca05aec10
[ "Apache-2.0" ]
null
null
null
final_project/machinetranslation/tests/test_translator.py
BrunoYH/xzceb-flask_eng_fr
a86a32ef56fcae1a8a9e281f17eb73dca05aec10
[ "Apache-2.0" ]
null
null
null
import unittest from translator import english_to_french, french_to_english class TestEnglishToFrench(unittest.TestCase): def testE2F(self): self.assertEqual(english_to_french("Hello"), "Bonjour") self.assertNotEqual(english_to_french("Null"), "") class FrenchEnglishToEnglish(unittest.TestCase): def testF2E(self): self.assertEqual(french_to_english("Bonjour"), "Hello") self.assertNotEqual(french_to_english("Null"), "") unittest.main()
30.1875
63
0.73913
import unittest from translator import english_to_french, french_to_english class TestEnglishToFrench(unittest.TestCase): def testE2F(self): self.assertEqual(english_to_french("Hello"), "Bonjour") self.assertNotEqual(english_to_french("Null"), "") class FrenchEnglishToEnglish(unittest.TestCase): def testF2E(self): self.assertEqual(french_to_english("Bonjour"), "Hello") self.assertNotEqual(french_to_english("Null"), "") unittest.main()
true
true
f717ad7f8ee5170fed096dfbfb9360232cb0abc4
157,342
py
Python
tests/test_cadquery.py
RubenRubens/cadquery
94030ad5e4d5489b29cb021282ecc8db34172471
[ "Apache-2.0" ]
1
2021-09-01T20:16:29.000Z
2021-09-01T20:16:29.000Z
tests/test_cadquery.py
RubenRubens/cadquery
94030ad5e4d5489b29cb021282ecc8db34172471
[ "Apache-2.0" ]
null
null
null
tests/test_cadquery.py
RubenRubens/cadquery
94030ad5e4d5489b29cb021282ecc8db34172471
[ "Apache-2.0" ]
null
null
null
""" This module tests cadquery creation and manipulation functions """ # system modules import math, os.path, time, tempfile from random import choice from random import random from random import randrange from itertools import product from pytest import approx, raises # my modules from cadquery import * from cadquery import exporters from cadquery import occ_impl from tests import ( BaseTest, writeStringToFile, makeUnitCube, readFileAsString, makeUnitSquareWire, makeCube, ) # test data directory testdataDir = os.path.join(os.path.dirname(__file__), "testdata") # where unit test output will be saved OUTDIR = tempfile.gettempdir() SUMMARY_FILE = os.path.join(OUTDIR, "testSummary.html") SUMMARY_TEMPLATE = """<html> <head> <style type="text/css"> .testResult{ background: #eeeeee; margin: 50px; border: 1px solid black; } </style> </head> <body> <!--TEST_CONTENT--> </body> </html>""" TEST_RESULT_TEMPLATE = """ <div class="testResult"><h3>%(name)s</h3> %(svg)s </div> <!--TEST_CONTENT--> """ # clean up any summary file that is in the output directory. # i know, this sux, but there is no other way to do this in 2.6, as we cannot do class fixtures till 2.7 writeStringToFile(SUMMARY_TEMPLATE, SUMMARY_FILE) class TestCadQuery(BaseTest): def tearDown(self): """ Update summary with data from this test. This is a really hacky way of doing it-- we get a startup event from module load, but there is no way in unittest to get a single shutdown event-- except for stuff in 2.7 and above So what we do here is to read the existing file, stick in more content, and leave it """ svgFile = os.path.join(OUTDIR, self._testMethodName + ".svg") # all tests do not produce output if os.path.exists(svgFile): existingSummary = readFileAsString(SUMMARY_FILE) svgText = readFileAsString(svgFile) svgText = svgText.replace( '<?xml version="1.0" encoding="UTF-8" standalone="no"?>', "" ) # now write data into the file # the content we are replacing it with also includes the marker, so it can be replaced again existingSummary = existingSummary.replace( "<!--TEST_CONTENT-->", TEST_RESULT_TEMPLATE % (dict(svg=svgText, name=self._testMethodName)), ) writeStringToFile(existingSummary, SUMMARY_FILE) def saveModel(self, shape): """ shape must be a CQ object Save models in SVG and STEP format """ shape.exportSvg(os.path.join(OUTDIR, self._testMethodName + ".svg")) shape.val().exportStep(os.path.join(OUTDIR, self._testMethodName + ".step")) def testToOCC(self): """ Tests to make sure that a CadQuery object is converted correctly to a OCC object. """ r = Workplane("XY").rect(5, 5).extrude(5) r = r.toOCC() import OCP self.assertEqual(type(r), OCP.TopoDS.TopoDS_Compound) def testToSVG(self): """ Tests to make sure that a CadQuery object is converted correctly to SVG """ r = Workplane("XY").rect(5, 5).extrude(5) r_str = r.toSvg() # Make sure that a couple of sections from the SVG output make sense self.assertTrue(r_str.index('path d="M') > 0) self.assertTrue( r_str.index('line x1="30" y1="-30" x2="58" y2="-15" stroke-width="3"') > 0 ) def testCubePlugin(self): """ Tests a plugin that combines cubes together with a base :return: """ # make the plugin method def makeCubes(self, length): # self refers to the CQ or Workplane object # create the solid s = Solid.makeBox(length, length, length, Vector(0, 0, 0)) # use CQ utility method to iterate over the stack an position the cubes return self.eachpoint(lambda loc: s.located(loc), True) # link the plugin in Workplane.makeCubes = makeCubes # call it result = ( Workplane("XY") .box(6.0, 8.0, 0.5) .faces(">Z") .rect(4.0, 4.0, forConstruction=True) .vertices() ) result = result.makeCubes(1.0) result = result.combineSolids() self.saveModel(result) self.assertEqual(1, result.solids().size()) def testCylinderPlugin(self): """ Tests a cylinder plugin. The plugin creates cylinders of the specified radius and height for each item on the stack This is a very short plugin that illustrates just about the simplest possible plugin """ def cylinders(self, radius, height): # construct a cylinder at (0,0,0) c = Solid.makeCylinder(radius, height, Vector(0, 0, 0)) # combine all the cylinders into a single compound r = self.eachpoint(lambda loc: c.located(loc), True).combineSolids() return r Workplane.cyl = cylinders # now test. here we want weird workplane to see if the objects are transformed right s = ( Workplane(Plane(Vector((0, 0, 0)), Vector((1, -1, 0)), Vector((1, 1, 0)))) .rect(2.0, 3.0, forConstruction=True) .vertices() .cyl(0.25, 0.5) ) self.assertEqual(4, s.solids().size()) self.saveModel(s) def testPolygonPlugin(self): """ Tests a plugin to make regular polygons around points on the stack Demonstrations using eachpoint to allow working in local coordinates to create geometry """ def rPoly(self, nSides, diameter): def _makePolygon(loc): # pnt is a vector in local coordinates angle = 2.0 * math.pi / nSides pnts = [] for i in range(nSides + 1): pnts.append( Vector( (diameter / 2.0 * math.cos(angle * i)), (diameter / 2.0 * math.sin(angle * i)), 0, ) ) return Wire.makePolygon(pnts).located(loc) return self.eachpoint(_makePolygon, True) Workplane.rPoly = rPoly s = ( Workplane("XY") .box(4.0, 4.0, 0.25) .faces(">Z") .workplane() .rect(2.0, 2.0, forConstruction=True) .vertices() .rPoly(5, 0.5) .cutThruAll() ) # 6 base sides, 4 pentagons, 5 sides each = 26 self.assertEqual(26, s.faces().size()) self.saveModel(s) def testFluentMethodInheritance(self): """ Tests that a derived class inherits fluent methods which return instances of derived class when inherited. """ class ExtendedWorkplane(Workplane): def nonExistentInWorkplane(self): pass # Call an inherited fluent method: wp = ExtendedWorkplane("XY").moveTo(1, 2) # Verify that the inherited method returned an instance of the derived # class: self.assertEqual(type(wp), ExtendedWorkplane) # The following is redundant, but can make the use case clearer. # This must not raise an AttributeError: wp.nonExistentInWorkplane() def testPointList(self): """ Tests adding points and using them """ c = CQ(makeUnitCube()) s = c.faces(">Z").workplane().pushPoints([(-0.3, 0.3), (0.3, 0.3), (0, 0)]) self.assertEqual(3, s.size()) # TODO: is the ability to iterate over points with circle really worth it? # maybe we should just require using all() and a loop for this. the semantics and # possible combinations got too hard ( ie, .circle().circle() ) was really odd body = s.circle(0.05).cutThruAll() self.saveModel(body) self.assertEqual(9, body.faces().size()) # Test the case when using eachpoint with only a blank workplane def callback_fn(loc): self.assertEqual( Vector(0, 0, 0), Vector(loc.wrapped.Transformation().TranslationPart()) ) r = Workplane("XY") r.objects = [] r.eachpoint(callback_fn) def testWorkplaneFromFace(self): # make a workplane on the top face s = CQ(makeUnitCube()).faces(">Z").workplane() r = s.circle(0.125).cutBlind(-2.0) self.saveModel(r) # the result should have 7 faces self.assertEqual(7, r.faces().size()) self.assertEqual(type(r.val()), Compound) self.assertEqual(type(r.first().val()), Compound) def testFrontReference(self): # make a workplane on the top face s = CQ(makeUnitCube()).faces("front").workplane() r = s.circle(0.125).cutBlind(-2.0) self.saveModel(r) # the result should have 7 faces self.assertEqual(7, r.faces().size()) self.assertEqual(type(r.val()), Compound) self.assertEqual(type(r.first().val()), Compound) def testRotate(self): """Test solid rotation at the CQ object level.""" box = Workplane("XY").box(1, 1, 5) box.rotate((0, 0, 0), (1, 0, 0), 90) startPoint = box.faces("<Y").edges("<X").first().val().startPoint().toTuple() endPoint = box.faces("<Y").edges("<X").first().val().endPoint().toTuple() self.assertEqual(-0.5, startPoint[0]) self.assertEqual(-0.5, startPoint[1]) self.assertEqual(-2.5, startPoint[2]) self.assertEqual(-0.5, endPoint[0]) self.assertEqual(-0.5, endPoint[1]) self.assertEqual(2.5, endPoint[2]) def testPlaneRotateZNormal(self): """ Rotation of a plane in the Z direction should never alter its normal. This test creates random planes. The plane is rotated a random angle in the Z-direction to verify that the resulting plane maintains the same normal. The test also checks that the random origin is unaltered after rotation. """ for _ in range(100): angle = (random() - 0.5) * 720 xdir = Vector(random(), random(), random()).normalized() rdir = Vector(random(), random(), random()).normalized() zdir = xdir.cross(rdir).normalized() origin = (random(), random(), random()) plane = Plane(origin=origin, xDir=xdir, normal=zdir) rotated = plane.rotated((0, 0, angle)) assert rotated.zDir.toTuple() == approx(zdir.toTuple()) assert rotated.origin.toTuple() == approx(origin) def testPlaneRotateConcat(self): """ Test the result of a well-known concatenated rotation example. """ xdir = (1, 0, 0) normal = (0, 0, 1) k = 2.0 ** 0.5 / 2.0 origin = (2, -1, 1) plane = Plane(origin=origin, xDir=xdir, normal=normal) plane = plane.rotated((0, 0, 45)) assert plane.xDir.toTuple() == approx((k, k, 0)) assert plane.yDir.toTuple() == approx((-k, k, 0)) assert plane.zDir.toTuple() == approx((0, 0, 1)) plane = plane.rotated((0, 45, 0)) assert plane.xDir.toTuple() == approx((0.5, 0.5, -k)) assert plane.yDir.toTuple() == approx((-k, k, 0)) assert plane.zDir.toTuple() == approx((0.5, 0.5, k)) assert plane.origin.toTuple() == origin def testPlaneRotateConcatRandom(self): """ Rotation of a plane in a given direction should never alter that direction. This test creates a plane and rotates it a random angle in a given direction. After the rotation, the direction of the resulting plane in the rotation-direction should be constant. The test also checks that the origin is unaltered after all rotations. """ origin = (2, -1, 1) plane = Plane(origin=origin, xDir=(1, 0, 0), normal=(0, 0, 1)) for _ in range(100): before = { 0: plane.xDir.toTuple(), 1: plane.yDir.toTuple(), 2: plane.zDir.toTuple(), } angle = (random() - 0.5) * 720 direction = randrange(3) rotation = [0, 0, 0] rotation[direction] = angle plane = plane.rotated(rotation) after = { 0: plane.xDir.toTuple(), 1: plane.yDir.toTuple(), 2: plane.zDir.toTuple(), } assert before[direction] == approx(after[direction]) assert plane.origin.toTuple() == origin def testPlaneNoXDir(self): """ Plane should pick an arbitrary x direction if None is passed in. """ for z_dir in [(0, 0, 1), (1, 0, 0), (-1, 0, 0), Vector(-1, 0, 0)]: result = Plane(origin=(1, 2, 3), xDir=None, normal=z_dir) assert result.zDir == Vector(z_dir) assert result.xDir.Length == approx(1) assert result.origin == Vector(1, 2, 3) # unspecified xDir should be the same as xDir=None result2 = Plane(origin=(1, 2, 3), normal=z_dir) assert result2 == result def testPlaneToPln(self): plane = Plane(origin=(1, 2, 3), xDir=(-1, 0, 0), normal=(0, 1, 0)) gppln = plane.toPln() assert Vector(gppln.XAxis().Direction()) == Vector(-1, 0, 0) assert Vector(gppln.YAxis().Direction()) == plane.yDir assert Vector(gppln.Axis().Direction()) == plane.zDir def testRect(self): x = 10 y = 11 s = Workplane().rect(x, y) # a rectangle has 4 sides self.assertEqual(s.edges().size(), 4) # assert that the lower left corner is in the correct spot for all # possible values of centered for centered_x, xval in zip([True, False], [-x / 2, 0]): for centered_y, yval in zip([True, False], [-y / 2, 0]): s = ( Workplane() .rect(x, y, centered=(centered_x, centered_y)) .vertices("<X and <Y") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, 0), 3) # check that centered=True is the same as centered=(True, True) for option0 in [True, False]: v0 = ( Workplane() .rect(x, y, centered=option0) .vertices(">X and >Y") .val() .toTuple() ) v1 = ( Workplane() .rect(x, y, centered=(option0, option0)) .vertices(">X and >Y") .val() .toTuple() ) self.assertTupleAlmostEquals(v0, v1, 3) # test negative lengths r0 = Workplane().rect(-x, -y, centered=False) self.assertTupleAlmostEquals( (0, 0, 0), r0.vertices(">X and >Y").val().toTuple(), 3 ) self.assertTupleAlmostEquals( (-x, -y, 0), r0.vertices("<X and <Y").val().toTuple(), 3 ) # test move plus negative length r1 = Workplane().move(x, y).rect(-x, -y, centered=False) self.assertTupleAlmostEquals( (x, y, 0), r1.vertices(">X and >Y").val().toTuple(), 3 ) self.assertTupleAlmostEquals( (0, 0, 0), r1.vertices("<X and <Y").val().toTuple(), 3 ) # negative length should have no effect with centered=True v2 = Workplane().rect(x, y).vertices(">X and >Y").val().toTuple() v3 = Workplane().rect(-x, -y).vertices(">X and >Y").val().toTuple() self.assertTupleAlmostEquals(v2, v3, 3) def testLoft(self): """ Test making a lofted solid """ s = Workplane("XY").circle(4.0).workplane(5.0).rect(2.0, 2.0).loft() self.saveModel(s) # the result should have 7 faces self.assertEqual(1, s.solids().size()) # the resulting loft had a split on the side, not sure why really, i expected only 3 faces self.assertEqual(7, s.faces().size()) def testLoftRaisesValueError(self): s0 = Workplane().hLine(1) # no wires with raises(ValueError): s0.loft() s1 = Workplane("XY").circle(5) # one wire with self.assertRaises(ValueError) as cm: s1.loft() err = cm.exception self.assertEqual(str(err), "More than one wire is required") def testLoftCombine(self): """ test combining a lof with another feature :return: """ s = ( Workplane("front") .box(4.0, 4.0, 0.25) .faces(">Z") .circle(1.5) .workplane(offset=3.0) .rect(0.75, 0.5) .loft(combine=True) ) self.saveModel(s) # self.assertEqual(1,s.solids().size() ) # self.assertEqual(8,s.faces().size() ) def testRevolveCylinder(self): """ Test creating a solid using the revolve operation. :return: """ # The dimensions of the model. These can be modified rather than changing the # shape's code directly. rectangle_width = 10.0 rectangle_length = 10.0 angle_degrees = 360.0 # Test revolve without any options for making a cylinder result = ( Workplane("XY").rect(rectangle_width, rectangle_length, False).revolve() ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) # Test revolve when only setting the angle to revolve through result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, False) .revolve(angle_degrees) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, False) .revolve(270.0) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Test when passing revolve the angle and the axis of revolution's start point result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5)) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5)) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Test when passing revolve the angle and both the start and ends of the axis of revolution result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5), (-5, 5)) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5), (-5, 5)) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Testing all of the above without combine result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5), (-5, 5), False) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5), (-5, 5), False) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) def testRevolveDonut(self): """ Test creating a solid donut shape with square walls :return: """ # The dimensions of the model. These can be modified rather than changing the # shape's code directly. rectangle_width = 10.0 rectangle_length = 10.0 angle_degrees = 360.0 result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, True) .revolve(angle_degrees, (20, 0), (20, 10)) ) self.assertEqual(4, result.faces().size()) self.assertEqual(4, result.vertices().size()) self.assertEqual(6, result.edges().size()) def testRevolveCone(self): """ Test creating a solid from a revolved triangle :return: """ result = Workplane("XY").lineTo(0, 10).lineTo(5, 0).close().revolve() self.assertEqual(2, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(2, result.edges().size()) def testRevolveErrors(self): """ Test that revolve raises errors when used incorrectly. """ result = Workplane("XY").lineTo(0, 10).lineTo(5, 0) with raises(ValueError): result.revolve() def testSpline(self): """ Tests construction of splines """ pts = [(0, 0), (0, 1), (1, 2), (2, 4)] # Spline path - just a smoke test path = Workplane("XZ").spline(pts).val() # Closed spline path_closed = Workplane("XZ").spline(pts, periodic=True).val() self.assertTrue(path_closed.IsClosed()) # attempt to build a valid face w = Wire.assembleEdges([path_closed,]) f = Face.makeFromWires(w) self.assertTrue(f.isValid()) # attempt to build an invalid face w = Wire.assembleEdges([path,]) f = Face.makeFromWires(w) self.assertFalse(f.isValid()) # Spline with explicit tangents path_const = Workplane("XZ").spline(pts, tangents=((0, 1), (1, 0))).val() self.assertFalse(path.tangentAt(0) == path_const.tangentAt(0)) self.assertFalse(path.tangentAt(1) == path_const.tangentAt(1)) # test include current path1 = Workplane("XZ").spline(pts[1:], includeCurrent=True).val() self.assertAlmostEqual(path.Length(), path1.Length()) # test tangents and offset plane pts = [(0, 0), (-1, 1), (-2, 0), (-1, 0)] tangents = [(0, 1), (1, 0)] path2 = Workplane("XY", (0, 0, 10)).spline(pts, tangents=tangents) self.assertAlmostEqual(path2.val().tangentAt(0).z, 0) def testSplineWithMultipleTangents(self): """ Tests specifying B-spline tangents, besides the start point and end point tangents. """ points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters) .consolidateWires() ) test_point = spline.edges().val().positionAt(2.5, mode="parameter") expected_test_point = Vector(1.875, -0.625, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineWithSpecifiedAndUnspecifiedTangents(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), None, (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters) .consolidateWires() ) test_point = spline.edges().val().positionAt(1.5, mode="parameter") expected_test_point = Vector(1.6875, 0.875, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineSpecifyingParameters(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] spline1 = ( Workplane("XY") .spline(points, tangents=tangents, parameters=[0, 1, 2, 3]) .consolidateWires() ) # Multiply all parameter values by 10: spline2 = ( Workplane("XY") .spline(points, tangents=tangents, parameters=[0, 10, 20, 30]) .consolidateWires() ) # Test point equivalence for parameter, and parameter multiplied by 10: test_point1 = spline1.edges().val().positionAt(1.5, mode="parameter") test_point2 = spline2.edges().val().positionAt(15, mode="parameter") expected_test_point = Vector(1.625, 0.625, 0.0) self.assertAlmostEqual((test_point1 - test_point2).Length, 0) self.assertAlmostEqual((test_point1 - expected_test_point).Length, 0) # test periodic with parameters spline3 = Workplane().spline( points, periodic=True, parameters=[x for x in range(len(points) + 1)] ) self.assertTrue(spline3.val().IsClosed()) def testSplineWithScaleTrue(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters, scale=True) .consolidateWires() ) test_point = spline.edges().val().positionAt(0.5, mode="parameter") expected_test_point = Vector(0.375, 0.875, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineWithScaleFalse(self): """ Like testSplineWithScaleTrue, but verifies the tangent vector is different when scale=False. The interpolation points and tangent vectors are the same in `testSplineWithScaleTrue`, and `testSplineWithScaleFalse`. A test point is rendered at the same parameter value in both cases, but its coordinates are different in each case. """ points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters, scale=False) .consolidateWires() ) test_point = spline.edges().val().positionAt(0.5, mode="parameter") expected_test_point = Vector(0.375, 0.625, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineTangentMagnitudeBelowToleranceThrows(self): import OCP points = [(0, 0), (1, 1), (2, 0), (1, -1)] # Use a tangent vector with magnitude 0.5: tangents = [(0, 0.5), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) # Set tolerance above the 0.5 length of the tangent vector. This # should throw an exception: with raises( (OCP.Standard.Standard_ConstructionError, OCP.Standard.Standard_Failure) ): spline = ( Workplane("XY") .spline(points, tangents=tangents, tol=1) .consolidateWires() ) def testSplineInputValidation(self): points = [(0, 0), (1, 1), (2, 0)] tangents = [(0, 0.5), (1, 0), (0, -1), (-1, 0)] with raises(ValueError): spline = Workplane().spline(points, tangents=tangents) with raises(ValueError): Workplane().spline( points, periodic=False, parameters=[x for x in range(len(points) + 1)], ) with raises(ValueError): Workplane().spline( points, periodic=True, parameters=[x for x in range(len(points))], ) def testRotatedEllipse(self): def rotatePoint(x, y, alpha): # rotation matrix a = alpha * DEG2RAD r = ((math.cos(a), math.sin(a)), (-math.sin(a), math.cos(a))) return ((x * r[0][0] + y * r[1][0]), (x * r[0][1] + y * r[1][1])) def ellipsePoints(r1, r2, a): return (r1 * math.cos(a * DEG2RAD), r2 * math.sin(a * DEG2RAD)) DEG2RAD = math.pi / 180.0 p0 = (10, 20) a1, a2 = 30, -60 r1, r2 = 20, 10 ra = 25 sx_rot, sy_rot = rotatePoint(*ellipsePoints(r1, r2, a1), ra) ex_rot, ey_rot = rotatePoint(*ellipsePoints(r1, r2, a2), ra) # startAtCurrent=False, sense = 1 ellipseArc1 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) start = ellipseArc1.vertices().objects[0] end = ellipseArc1.vertices().objects[1] self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + sx_rot, p0[1] + sy_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + ex_rot, p0[1] + ey_rot), 3 ) # startAtCurrent=True, sense = 1 ellipseArc2 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=True, angle1=a1, angle2=a2, rotation_angle=ra ) ) start = ellipseArc2.vertices().objects[0] end = ellipseArc2.vertices().objects[1] self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + sx_rot - sx_rot, p0[1] + sy_rot - sy_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + ex_rot - sx_rot, p0[1] + ey_rot - sy_rot), 3 ) # startAtCurrent=False, sense = -1 ellipseArc3 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra, sense=-1, ) ) start = ellipseArc3.vertices().objects[0] end = ellipseArc3.vertices().objects[1] # swap start and end points for comparison due to different sense self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + ex_rot, p0[1] + ey_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + sx_rot, p0[1] + sy_rot), 3 ) # startAtCurrent=True, sense = -1 ellipseArc4 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=True, angle1=a1, angle2=a2, rotation_angle=ra, sense=-1, makeWire=True, ) ) self.assertEqual(len(ellipseArc4.ctx.pendingWires), 1) start = ellipseArc4.vertices().objects[0] end = ellipseArc4.vertices().objects[1] # swap start and end points for comparison due to different sense self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + ex_rot - ex_rot, p0[1] + ey_rot - ey_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + sx_rot - ex_rot, p0[1] + sy_rot - ey_rot), 3 ) def testEllipseArcsClockwise(self): ellipseArc = ( Workplane("XY") .moveTo(10, 15) .ellipseArc(5, 4, -10, 190, 45, sense=-1, startAtCurrent=False) ) sp = ellipseArc.val().startPoint() ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals( (sp.x, sp.y), (7.009330014275797, 11.027027582524015), 3 ) self.assertTupleAlmostEquals( (ep.x, ep.y), (13.972972417475985, 17.990669985724203), 3 ) ellipseArc = ( ellipseArc.ellipseArc(5, 4, -10, 190, 315, sense=-1) .ellipseArc(5, 4, -10, 190, 225, sense=-1) .ellipseArc(5, 4, -10, 190, 135, sense=-1) ) ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals((sp.x, sp.y), (ep.x, ep.y), 3) def testEllipseArcsCounterClockwise(self): ellipseArc = ( Workplane("XY") .moveTo(10, 15) .ellipseArc(5, 4, -10, 190, 45, startAtCurrent=False) ) sp = ellipseArc.val().startPoint() ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals( (sp.x, sp.y), (13.972972417475985, 17.990669985724203), 3 ) self.assertTupleAlmostEquals( (ep.x, ep.y), (7.009330014275797, 11.027027582524015), 3 ) ellipseArc = ( ellipseArc.ellipseArc(5, 4, -10, 190, 135) .ellipseArc(5, 4, -10, 190, 225) .ellipseArc(5, 4, -10, 190, 315) ) ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals((sp.x, sp.y), (ep.x, ep.y), 3) def testEllipseCenterAndMoveTo(self): # Whether we start from a center() call or a moveTo call, it should be the same ellipse Arc p0 = (10, 20) a1, a2 = 30, -60 r1, r2 = 20, 10 ra = 25 ellipseArc1 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) sp1 = ellipseArc1.val().startPoint() ep1 = ellipseArc1.val().endPoint() ellipseArc2 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) sp2 = ellipseArc2.val().startPoint() ep2 = ellipseArc2.val().endPoint() self.assertTupleAlmostEquals(sp1.toTuple(), sp2.toTuple(), 3) self.assertTupleAlmostEquals(ep1.toTuple(), ep2.toTuple(), 3) def testMakeEllipse(self): el = Wire.makeEllipse( 1, 2, Vector(0, 0, 0), Vector(0, 0, 1), Vector(1, 0, 0), 0, 90, 45, True, ) self.assertTrue(el.IsClosed()) self.assertTrue(el.isValid()) def testSweep(self): """ Tests the operation of sweeping a wire(s) along a path """ pts = [(0, 0), (0, 1), (1, 2), (2, 4)] # Spline path path = Workplane("XZ").spline(pts) # Test defaults result = Workplane("XY").circle(1.0).sweep(path) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with makeSolid False result = Workplane("XY").circle(1.0).sweep(path, makeSolid=False) self.assertEqual(1, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with isFrenet True result = Workplane("XY").circle(1.0).sweep(path, isFrenet=True) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with makeSolid False and isFrenet True result = Workplane("XY").circle(1.0).sweep(path, makeSolid=False, isFrenet=True) self.assertEqual(1, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test rectangle with defaults result = Workplane("XY").rect(1.0, 1.0).sweep(path) self.assertEqual(6, result.faces().size()) self.assertEqual(12, result.edges().size()) # Test fixed normal result = Workplane().circle(0.5).sweep(path, normal=Vector(0, 0, 1)) self.assertTupleAlmostEquals( result.faces(">Z").val().normalAt().toTuple(), (0, 0, 1), 6 ) # Polyline path path = Workplane("XZ").polyline(pts) # Test defaults result = Workplane("XY").circle(0.1).sweep(path, transition="transformed") self.assertEqual(5, result.faces().size()) self.assertEqual(7, result.edges().size()) # Polyline path and one inner profiles path = Workplane("XZ").polyline(pts) # Test defaults result = ( Workplane("XY") .circle(0.2) .circle(0.1) .sweep(path, transition="transformed") ) self.assertEqual(8, result.faces().size()) self.assertEqual(14, result.edges().size()) # Polyline path and different transition settings for t in ("transformed", "right", "round"): path = Workplane("XZ").polyline(pts) result = ( Workplane("XY") .circle(0.2) .rect(0.2, 0.1) .rect(0.1, 0.2) .sweep(path, transition=t) ) self.assertTrue(result.solids().val().isValid()) # Polyline path and multiple inner profiles path = Workplane("XZ").polyline(pts) # Test defaults result = ( Workplane("XY") .circle(0.2) .rect(0.2, 0.1) .rect(0.1, 0.2) .circle(0.1) .sweep(path) ) self.assertTrue(result.solids().val().isValid()) # Arc path path = Workplane("XZ").threePointArc((1.0, 1.5), (0.0, 1.0)) # Test defaults result = Workplane("XY").circle(0.1).sweep(path) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test aux spine pts1 = [(0, 0), (20, 100)] pts2 = [(0, 20, 0), (20, 0, 100)] path = Workplane("YZ").spline(pts1, tangents=[(0, 1, 0), (0, 1, 0)]) aux_path = Workplane("XY").spline(pts2, tangents=[(0, 0, 1), (0, 0, 1)]) result = Workplane("XY").rect(10, 20).sweep(path, auxSpine=aux_path) bottom = result.faces("<Z") top = result.faces(">Z") v1 = bottom.wires().val().tangentAt(0.0) v2 = top.wires().val().tangentAt(0.0) self.assertAlmostEqual(v1.getAngle(v2), math.pi / 4, 6) # test for ValueError if pending wires is empty w0 = Workplane().hLine(1).vLine(1) with raises(ValueError): w0.sweep(path) # Test aux spine invalid input handling with raises(ValueError): result = ( Workplane("XY") .rect(10, 20) .sweep(path, auxSpine=Workplane().box(1, 1, 1)) ) def testMultisectionSweep(self): """ Tests the operation of sweeping along a list of wire(s) along a path """ # X axis line length 20.0 path = Workplane("XZ").moveTo(-10, 0).lineTo(10, 0) # Sweep a circle from diameter 2.0 to diameter 1.0 to diameter 2.0 along X axis length 10.0 + 10.0 defaultSweep = ( Workplane("YZ") .workplane(offset=-10.0) .circle(2.0) .workplane(offset=10.0) .circle(1.0) .workplane(offset=10.0) .circle(2.0) .sweep(path, multisection=True) ) # We can sweep through different shapes recttocircleSweep = ( Workplane("YZ") .workplane(offset=-10.0) .rect(2.0, 2.0) .workplane(offset=8.0) .circle(1.0) .workplane(offset=4.0) .circle(1.0) .workplane(offset=8.0) .rect(2.0, 2.0) .sweep(path, multisection=True) ) circletorectSweep = ( Workplane("YZ") .workplane(offset=-10.0) .circle(1.0) .workplane(offset=7.0) .rect(2.0, 2.0) .workplane(offset=6.0) .rect(2.0, 2.0) .workplane(offset=7.0) .circle(1.0) .sweep(path, multisection=True) ) # Placement of the Shape is important otherwise could produce unexpected shape specialSweep = ( Workplane("YZ") .circle(1.0) .workplane(offset=10.0) .rect(2.0, 2.0) .sweep(path, multisection=True) ) # Switch to an arc for the path : line l=5.0 then half circle r=4.0 then line l=5.0 path = ( Workplane("XZ") .moveTo(-5, 4) .lineTo(0, 4) .threePointArc((4, 0), (0, -4)) .lineTo(-5, -4) ) # Placement of different shapes should follow the path # cylinder r=1.5 along first line # then sweep along arc from r=1.5 to r=1.0 # then cylinder r=1.0 along last line arcSweep = ( Workplane("YZ") .workplane(offset=-5) .moveTo(0, 4) .circle(1.5) .workplane(offset=5, centerOption="CenterOfMass") .circle(1.5) .moveTo(0, -8) .circle(1.0) .workplane(offset=-5, centerOption="CenterOfMass") .circle(1.0) .sweep(path, multisection=True) ) # Test multisection with normal pts = [(0, 0), (20, 100)] path = Workplane("YZ").spline(pts, tangents=[(0, 1, 0), (0, 1, 0)]) normalSweep = ( Workplane() .rect(10, 10) .workplane(offset=100) .rect(10, 20) .sweep(path, multisection=True, normal=(0, 1, 1)) ) self.assertTupleAlmostEquals( normalSweep.faces("<Z").val().normalAt().toTuple(), Vector(0, -1, -1).normalized().toTuple(), 6, ) self.assertTupleAlmostEquals( normalSweep.faces(">Z").val().normalAt().toTuple(), Vector(0, 1, 1).normalized().toTuple(), 6, ) # Test and saveModel self.assertEqual(1, defaultSweep.solids().size()) self.assertEqual(1, circletorectSweep.solids().size()) self.assertEqual(1, recttocircleSweep.solids().size()) self.assertEqual(1, specialSweep.solids().size()) self.assertEqual(1, arcSweep.solids().size()) self.saveModel(defaultSweep) def testTwistExtrude(self): """ Tests extrusion while twisting through an angle. """ profile = Workplane("XY").rect(10, 10) r = profile.twistExtrude(10, 45, False) self.assertEqual(6, r.faces().size()) def testTwistExtrudeCombine(self): """ Tests extrusion while twisting through an angle, combining with other solids. """ profile = Workplane("XY").rect(10, 10) r = profile.twistExtrude(10, 45) self.assertEqual(6, r.faces().size()) def testRectArray(self): x_num = 3 y_num = 3 x_spacing = 8.0 y_spacing = 8.0 s = ( Workplane("XY") .box(40, 40, 5, centered=(True, True, True)) .faces(">Z") .workplane() .rarray(x_spacing, y_spacing, x_num, y_num, True) .circle(2.0) .extrude(2.0) ) self.saveModel(s) # 6 faces for the box, 2 faces for each cylinder self.assertEqual(6 + x_num * y_num * 2, s.faces().size()) with raises(ValueError): Workplane().rarray(0, 0, x_num, y_num, True) # check lower and upper corner points are correct for all combinations of centering for x_opt, x_min, x_max in zip( [True, False], [-x_spacing, 0.0], [x_spacing, x_spacing * 2] ): for y_opt, y_min, y_max in zip( [True, False], [-y_spacing, 0.0], [y_spacing, y_spacing * 2] ): s = Workplane().rarray( x_spacing, y_spacing, x_num, y_num, center=(x_opt, y_opt) ) lower = Vector(x_min, y_min, 0) upper = Vector(x_max, y_max, 0) self.assertTrue(lower in s.objects) self.assertTrue(upper in s.objects) # check centered=True is equivalent to centered=(True, True) for val in [True, False]: s0 = Workplane().rarray(x_spacing, y_spacing, x_num, y_num, center=val) s1 = Workplane().rarray( x_spacing, y_spacing, x_num, y_num, center=(val, val) ) # check all the points in s0 are present in s1 self.assertTrue(all(pnt in s1.objects for pnt in s0.objects)) self.assertEqual(s0.size(), s1.size()) def testPolarArray(self): radius = 10 # Test for proper number of elements s = Workplane("XY").polarArray(radius, 0, 180, 1) self.assertEqual(1, s.size()) s = Workplane("XY").polarArray(radius, 0, 180, 6) self.assertEqual(6, s.size()) to_x = lambda l: l.wrapped.Transformation().TranslationPart().X() to_y = lambda l: l.wrapped.Transformation().TranslationPart().Y() to_angle = ( lambda l: l.wrapped.Transformation().GetRotation().GetRotationAngle() * 180.0 / math.pi ) # Test for proper placement when fill == True s = Workplane("XY").polarArray(radius, 0, 180, 3) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper placement when angle to fill is multiple of 360 deg s = Workplane("XY").polarArray(radius, 0, 360, 4) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper placement when fill == False s = Workplane("XY").polarArray(radius, 0, 90, 3, fill=False) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper operation of startAngle s = Workplane("XY").polarArray(radius, 90, 180, 3) self.assertAlmostEqual(radius, to_x(s.objects[0])) self.assertAlmostEqual(0, to_y(s.objects[0])) # Test for local rotation s = Workplane().polarArray(radius, 0, 180, 3) self.assertAlmostEqual(0, to_angle(s.objects[0])) self.assertAlmostEqual(90, to_angle(s.objects[1])) s = Workplane().polarArray(radius, 0, 180, 3, rotate=False) self.assertAlmostEqual(0, to_angle(s.objects[0])) self.assertAlmostEqual(0, to_angle(s.objects[1])) def testNestedCircle(self): s = ( Workplane("XY") .box(40, 40, 5) .pushPoints([(10, 0), (0, 10)]) .circle(4) .circle(2) .extrude(4) ) self.saveModel(s) self.assertEqual(14, s.faces().size()) def testConcentricEllipses(self): concentricEllipses = ( Workplane("XY").center(10, 20).ellipse(100, 10).center(0, 0).ellipse(50, 5) ) v = concentricEllipses.vertices().objects[0] self.assertTupleAlmostEquals((v.X, v.Y), (10 + 50, 20), 3) def testLegoBrick(self): # test making a simple lego brick # which of the below # inputs lbumps = 8 wbumps = 2 # lego brick constants P = 8.0 # nominal pitch c = 0.1 # clearance on each brick side H = 1.2 * P # nominal height of a brick bumpDiam = 4.8 # the standard bump diameter # the nominal thickness of the walls, normally 1.5 t = (P - (2 * c) - bumpDiam) / 2.0 postDiam = P - t # works out to 6.5 total_length = lbumps * P - 2.0 * c total_width = wbumps * P - 2.0 * c # build the brick s = Workplane("XY").box(total_length, total_width, H) # make the base s = s.faces("<Z").shell(-1.0 * t) # shell inwards not outwards s = ( s.faces(">Z") .workplane() .rarray(P, P, lbumps, wbumps, True) .circle(bumpDiam / 2.0) .extrude(1.8) ) # make the bumps on the top # add posts on the bottom. posts are different diameter depending on geometry # solid studs for 1 bump, tubes for multiple, none for 1x1 # this is cheating a little-- how to select the inner face from the shell? tmp = s.faces("<Z").workplane(invert=True) if lbumps > 1 and wbumps > 1: tmp = ( tmp.rarray(P, P, lbumps - 1, wbumps - 1, center=True) .circle(postDiam / 2.0) .circle(bumpDiam / 2.0) .extrude(H - t) ) elif lbumps > 1: tmp = tmp.rarray(P, P, lbumps - 1, 1, center=True).circle(t).extrude(H - t) elif wbumps > 1: tmp = tmp.rarray(P, P, 1, wbumps - 1, center=True).circle(t).extrude(H - t) self.saveModel(s) def testAngledHoles(self): s = ( Workplane("front") .box(4.0, 4.0, 0.25) .faces(">Z") .workplane() .transformed(offset=Vector(0, -1.5, 1.0), rotate=Vector(60, 0, 0)) .rect(1.5, 1.5, forConstruction=True) .vertices() .hole(0.25) ) self.saveModel(s) self.assertEqual(10, s.faces().size()) def testTranslateSolid(self): c = CQ(makeUnitCube()) self.assertAlmostEqual(0.0, c.faces("<Z").vertices().item(0).val().Z, 3) # TODO: it might be nice to provide a version of translate that modifies the existing geometry too d = c.translate(Vector(0, 0, 1.5)) self.assertAlmostEqual(1.5, d.faces("<Z").vertices().item(0).val().Z, 3) def testTranslateWire(self): c = CQ(makeUnitSquareWire()) self.assertAlmostEqual(0.0, c.edges().vertices().item(0).val().Z, 3) d = c.translate(Vector(0, 0, 1.5)) self.assertAlmostEqual(1.5, d.edges().vertices().item(0).val().Z, 3) def testSolidReferencesCombine(self): "test that solid references are preserved correctly" c = CQ(makeUnitCube()) # the cube is the context solid self.assertEqual(6, c.faces().size()) # cube has six faces r = ( c.faces(">Z").workplane().circle(0.125).extrude(0.5, True) ) # make a boss, not updating the original self.assertEqual(8, r.faces().size()) # just the boss faces self.assertEqual(6, c.faces().size()) # original is not modified def testSolidReferencesCombineTrue(self): s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).extrude(0.5) # the result of course has 6 faces self.assertEqual(6, r.faces().size()) # the original workplane does not, because it did not have a solid initially self.assertEqual(0, s.faces().size()) t = r.faces(">Z").workplane().rect(0.25, 0.25).extrude(0.5, True) # of course the result has 11 faces self.assertEqual(11, t.faces().size()) # r (being the parent) remains unmodified self.assertEqual(6, r.faces().size()) self.saveModel(r) def testSolidReferenceCombineFalse(self): s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).extrude(0.5) # the result of course has 6 faces self.assertEqual(6, r.faces().size()) # the original workplane does not, because it did not have a solid initially self.assertEqual(0, s.faces().size()) t = r.faces(">Z").workplane().rect(0.25, 0.25).extrude(0.5, False) # result has 6 faces, because it was not combined with the original self.assertEqual(6, t.faces().size()) self.assertEqual(6, r.faces().size()) # original is unmodified as well # subsequent operations use that context solid afterwards def testSimpleWorkplane(self): """ A simple square part with a hole in it """ s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .extrude(0.5) .faces(">Z") .workplane() .circle(0.25) .cutBlind(-1.0) ) self.saveModel(r) self.assertEqual(7, r.faces().size()) def testMultiFaceWorkplane(self): """ Test Creation of workplane from multiple co-planar face selection. """ s = Workplane("XY").box(1, 1, 1).faces(">Z").rect(1, 0.5).cutBlind(-0.2) w = s.faces(">Z").workplane() o = w.val() # origin of the workplane self.assertAlmostEqual(o.x, 0.0, 3) self.assertAlmostEqual(o.y, 0.0, 3) self.assertAlmostEqual(o.z, 0.5, 3) def testTriangularPrism(self): s = Workplane("XY").lineTo(1, 0).lineTo(1, 1).close().extrude(0.2) self.saveModel(s) def testMultiWireWorkplane(self): """ A simple square part with a hole in it-- but this time done as a single extrusion with two wires, as opposed to s cut """ s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).circle(0.25).extrude(0.5) self.saveModel(r) self.assertEqual(7, r.faces().size()) def testConstructionWire(self): """ Tests a wire with several holes, that are based on the vertices of a square also tests using a workplane plane other than XY """ s = Workplane(Plane.YZ()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) self.saveModel(r) # 10 faces-- 6 plus 4 holes, the vertices of the second rect. self.assertEqual(10, r.faces().size()) def testTwoWorkplanes(self): """ Tests a model that uses more than one workplane """ # base block s = Workplane(Plane.XY()) # TODO: this syntax is nice, but the iteration might not be worth # the complexity. # the simpler and slightly longer version would be: # r = s.rect(2.0,2.0).rect(1.3,1.3,forConstruction=True).vertices() # for c in r.all(): # c.circle(0.125).extrude(0.5,True) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # side hole, blind deep 1.9 t = r.faces(">Y").workplane().circle(0.125).cutBlind(-1.9) self.saveModel(t) self.assertEqual(12, t.faces().size()) def testCut(self): """ Tests the cut function by itself to catch the case where a Solid object is passed. """ s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toCut = s.rect(1.0, 1.0).extrude(0.5) resS = currentS.cut(toCut.val()) self.assertEqual(10, resS.faces().size()) with self.assertRaises(ValueError): currentS.cut(toCut.faces().val()) # Test syntactic sugar [__sub__ method] sugar = currentS - toCut.val() self.assertEqual(resS.faces().size(), sugar.faces().size()) # test ValueError on no solid found s0 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): s0.cut(toCut.val()) def testIntersect(self): """ Tests the intersect function. """ s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toIntersect = s.rect(1.0, 1.0).extrude(1) resS = currentS.intersect(toIntersect.val()) self.assertEqual(6, resS.faces().size()) self.assertAlmostEqual(resS.val().Volume(), 0.5) resS = currentS.intersect(toIntersect) self.assertEqual(6, resS.faces().size()) self.assertAlmostEqual(resS.val().Volume(), 0.5) b1 = Workplane("XY").box(1, 1, 1) b2 = Workplane("XY", origin=(0, 0, 0.5)).box(1, 1, 1) resS = b1.intersect(b2) self.assertAlmostEqual(resS.val().Volume(), 0.5) with self.assertRaises(ValueError): b1.intersect(b2.faces().val()) # Test syntactic sugar [__mul__ method] sugar = b1 & b2 self.assertEqual(resS.val().Volume(), sugar.val().Volume()) # raise ValueError when no solid found with raises(ValueError): Workplane().intersect(toIntersect) def testBoundingBox(self): """ Tests the boudingbox center of a model """ result0 = ( Workplane("XY") .moveTo(10, 0) .lineTo(5, 0) .threePointArc((3.9393, 0.4393), (3.5, 1.5)) .threePointArc((3.0607, 2.5607), (2, 3)) .lineTo(1.5, 3) .threePointArc((0.4393, 3.4393), (0, 4.5)) .lineTo(0, 13.5) .threePointArc((0.4393, 14.5607), (1.5, 15)) .lineTo(28, 15) .lineTo(28, 13.5) .lineTo(24, 13.5) .lineTo(24, 11.5) .lineTo(27, 11.5) .lineTo(27, 10) .lineTo(22, 10) .lineTo(22, 13.2) .lineTo(14.5, 13.2) .lineTo(14.5, 10) .lineTo(12.5, 10) .lineTo(12.5, 13.2) .lineTo(5.5, 13.2) .lineTo(5.5, 2) .threePointArc((5.793, 1.293), (6.5, 1)) .lineTo(10, 1) .close() ) result = result0.extrude(100) bb_center = result.val().BoundingBox().center self.saveModel(result) self.assertAlmostEqual(14.0, bb_center.x, 3) self.assertAlmostEqual(7.5, bb_center.y, 3) self.assertAlmostEqual(50.0, bb_center.z, 3) # The following will raise with the default tolerance of TOL 1e-2 bb = result.val().BoundingBox(tolerance=1e-3) self.assertAlmostEqual(0.0, bb.xmin, 2) self.assertAlmostEqual(28, bb.xmax, 2) self.assertAlmostEqual(0.0, bb.ymin, 2) self.assertAlmostEqual(15.0, bb.ymax, 2) self.assertAlmostEqual(0.0, bb.zmin, 2) self.assertAlmostEqual(100.0, bb.zmax, 2) def testCutThroughAll(self): """ Tests a model that uses more than one workplane """ # base block s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # thru all without explicit face selection t = r.circle(0.5).cutThruAll() self.assertEqual(11, t.faces().size()) # side hole, thru all t = ( t.faces(">Y") .workplane(centerOption="CenterOfMass") .circle(0.125) .cutThruAll() ) self.saveModel(t) self.assertEqual(13, t.faces().size()) # no planar faces sphere_r = 10.0 r = ( Workplane() .sphere(sphere_r) .workplane() .circle(sphere_r / 2.0) .cutThruAll() .workplane() .transformed(rotate=(90, 0, 0)) .circle(sphere_r / 2.0) .cutThruAll() .workplane() .transformed(rotate=(0, 90, 0)) .circle(sphere_r / 2.0) .cutThruAll() ) self.assertTrue(r.val().isValid()) self.assertEqual(r.faces().size(), 7) # test errors box0 = Workplane().box(1, 1, 1).faces(">Z").workplane().hLine(1) with raises(ValueError): box0.cutThruAll() no_box = Workplane().hLine(1).vLine(1).close() with raises(ValueError): no_box.cutThruAll() def testCutToFaceOffsetNOTIMPLEMENTEDYET(self): """ Tests cutting up to a given face, or an offset from a face """ # base block s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # side hole, up to 0.1 from the last face try: t = ( r.faces(">Y") .workplane() .circle(0.125) .cutToOffsetFromFace(r.faces().mminDist(Dir.Y), 0.1) ) # should end up being a blind hole self.assertEqual(10, t.faces().size()) t.first().val().exportStep("c:/temp/testCutToFace.STEP") except: pass # Not Implemented Yet def testWorkplaneOnExistingSolid(self): "Tests extruding on an existing solid" c = ( CQ(makeUnitCube()) .faces(">Z") .workplane() .circle(0.25) .circle(0.125) .extrude(0.25) ) self.saveModel(c) self.assertEqual(10, c.faces().size()) def testWorkplaneCenterMove(self): # this workplane is centered at x=0.5,y=0.5, the center of the upper face s = ( Workplane("XY").box(1, 1, 1).faces(">Z").workplane().center(-0.5, -0.5) ) # move the center to the corner t = s.circle(0.25).extrude(0.2) # make a boss self.assertEqual(9, t.faces().size()) self.saveModel(t) def testBasicLines(self): "Make a triangular boss" global OUTDIR s = Workplane(Plane.XY()) # TODO: extrude() should imply wire() if not done already # most users dont understand what a wire is, they are just drawing r = s.lineTo(1.0, 0).lineTo(0, 1.0).close().wire().extrude(0.25) r.val().exportStep(os.path.join(OUTDIR, "testBasicLinesStep1.STEP")) # no faces on the original workplane self.assertEqual(0, s.faces().size()) # 5 faces on newly created object self.assertEqual(5, r.faces().size()) # now add a circle through a side face r1 = ( r.faces("+XY") .workplane(centerOption="CenterOfMass") .circle(0.08) .cutThruAll() ) self.assertEqual(6, r1.faces().size()) r1.val().exportStep(os.path.join(OUTDIR, "testBasicLinesXY.STEP")) # now add a circle through a top r2 = ( r1.faces("+Z") .workplane(centerOption="CenterOfMass") .circle(0.08) .cutThruAll() ) self.assertEqual(9, r2.faces().size()) r2.val().exportStep(os.path.join(OUTDIR, "testBasicLinesZ.STEP")) self.saveModel(r2) def test2DDrawing(self): """ Draw things like 2D lines and arcs, should be expanded later to include all 2D constructs """ s = Workplane(Plane.XY()) r = ( s.lineTo(1.0, 0.0) .lineTo(1.0, 1.0) .threePointArc((1.0, 1.5), (0.0, 1.0)) .lineTo(0.0, 0.0) .moveTo(1.0, 0.0) .lineTo(2.0, 0.0) .lineTo(2.0, 2.0) .threePointArc((2.0, 2.5), (0.0, 2.0)) .lineTo(-2.0, 2.0) .lineTo(-2.0, 0.0) .close() ) self.assertEqual(1, r.wires().size()) # Test the *LineTo functions s = Workplane(Plane.XY()) r = s.hLineTo(1.0).vLineTo(1.0).hLineTo(0.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) # Test the *Line functions s = Workplane(Plane.XY()) r = s.hLine(1.0).vLine(1.0).hLine(-1.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) # Test the move function s = Workplane(Plane.XY()) r = s.move(1.0, 1.0).hLine(1.0).vLine(1.0).hLine(-1.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) self.assertEqual( (1.0, 1.0), ( r.vertices(selectors.NearestToPointSelector((0.0, 0.0, 0.0))) .first() .val() .X, r.vertices(selectors.NearestToPointSelector((0.0, 0.0, 0.0))) .first() .val() .Y, ), ) # Test the sagittaArc and radiusArc functions a1 = Workplane(Plane.YZ()).threePointArc((5, 1), (10, 0)) a2 = Workplane(Plane.YZ()).sagittaArc((10, 0), -1) a3 = Workplane(Plane.YZ()).threePointArc((6, 2), (12, 0)) a4 = Workplane(Plane.YZ()).radiusArc((12, 0), -10) assert a1.edges().first().val().geomType() == "CIRCLE" assert a2.edges().first().val().geomType() == "CIRCLE" assert a3.edges().first().val().geomType() == "CIRCLE" assert a4.edges().first().val().geomType() == "CIRCLE" assert a1.edges().first().val().Length() == a2.edges().first().val().Length() assert a3.edges().first().val().Length() == a4.edges().first().val().Length() def testPolarLines(self): """ Draw some polar lines and check expected results """ # Test the PolarLine* functions s = Workplane(Plane.XY()) r = ( s.polarLine(10, 45) .polarLineTo(10, -45) .polarLine(10, -180) .polarLine(-10, -90) .close() ) # a single wire, 5 edges self.assertEqual(1, r.wires().size()) self.assertEqual(5, r.wires().edges().size()) def testLargestDimension(self): """ Tests the largestDimension function when no solids are on the stack and when there are """ r = Workplane("XY").box(1, 1, 1) dim = r.largestDimension() self.assertAlmostEqual(1.76, dim, 1) r = Workplane("XY").rect(1, 1).extrude(1) dim = r.largestDimension() self.assertAlmostEqual(1.76, dim, 1) r = Workplane("XY") with raises(ValueError): r.largestDimension() def testOccBottle(self): """ Make the OCC bottle example. """ L = 20.0 w = 6.0 t = 3.0 s = Workplane(Plane.XY()) # draw half the profile of the bottle p = ( s.center(-L / 2.0, 0) .vLine(w / 2.0) .threePointArc((L / 2.0, w / 2.0 + t), (L, w / 2.0)) .vLine(-w / 2.0) .mirrorX() .extrude(30.0, True) ) # make the neck p.faces(">Z").workplane().circle(3.0).extrude( 2.0, True ) # .edges().fillet(0.05) # make a shell p.faces(">Z").shell(0.3) self.saveModel(p) def testSplineShape(self): """ Tests making a shape with an edge that is a spline """ s = Workplane(Plane.XY()) sPnts = [ (2.75, 1.5), (2.5, 1.75), (2.0, 1.5), (1.5, 1.0), (1.0, 1.25), (0.5, 1.0), (0, 1.0), ] r = s.lineTo(3.0, 0).lineTo(3.0, 1.0).spline(sPnts).close() r = r.extrude(0.5) self.saveModel(r) def testSimpleMirror(self): """ Tests a simple mirroring operation """ s = ( Workplane("XY") .lineTo(2, 2) .threePointArc((3, 1), (2, 0)) .mirrorX() .extrude(0.25) ) self.assertEqual(6, s.faces().size()) self.saveModel(s) def testUnorderedMirror(self): """ Tests whether or not a wire can be mirrored if its mirror won't connect to it """ r = 20 s = 7 t = 1.5 points = [ (0, 0), (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] r = Workplane("XY").polyline(points).mirrorX() self.assertEqual(1, r.wires().size()) self.assertEqual(18, r.edges().size()) # try the same with includeCurrent=True r = Workplane("XY").polyline(points[1:], includeCurrent=True).mirrorX() self.assertEqual(1, r.wires().size()) self.assertEqual(18, r.edges().size()) def testChainedMirror(self): """ Tests whether or not calling mirrorX().mirrorY() works correctly """ r = 20 s = 7 t = 1.5 points = [ (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] r = Workplane("XY").polyline(points).mirrorX().mirrorY().extrude(1).faces(">Z") self.assertEqual(1, r.wires().size()) self.assertEqual(32, r.edges().size()) # TODO: Re-work testIbeam test below now that chaining works # TODO: Add toLocalCoords and toWorldCoords tests def testIbeam(self): """ Make an ibeam. demonstrates fancy mirroring """ s = Workplane(Plane.XY()) L = 100.0 H = 20.0 W = 20.0 t = 1.0 # TODO: for some reason doing 1/4 of the profile and mirroring twice ( .mirrorX().mirrorY() ) # did not work, due to a bug in freecad-- it was losing edges when creating a composite wire. # i just side-stepped it for now pts = [ (0, 0), (0, H / 2.0), (W / 2.0, H / 2.0), (W / 2.0, (H / 2.0 - t)), (t / 2.0, (H / 2.0 - t)), (t / 2.0, (t - H / 2.0)), (W / 2.0, (t - H / 2.0)), (W / 2.0, H / -2.0), (0, H / -2.0), ] r = s.polyline(pts).mirrorY() # these other forms also work res = r.extrude(L) self.saveModel(res) def testCone(self): """ Tests that a simple cone works """ s = Solid.makeCone(0, 1.0, 2.0) t = CQ(s) self.saveModel(t) self.assertEqual(2, t.faces().size()) def testFillet(self): """ Tests filleting edges on a solid """ c = ( CQ(makeUnitCube()) .faces(">Z") .workplane() .circle(0.25) .extrude(0.25, True) .edges("|Z") .fillet(0.2) ) self.saveModel(c) self.assertEqual(12, c.faces().size()) # should raise an error if no solid c1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): c1.fillet(0.1) def testChamfer(self): """ Test chamfer API with a box shape """ cube = CQ(makeUnitCube()).faces(">Z").chamfer(0.1) self.saveModel(cube) self.assertEqual(10, cube.faces().size()) # should raise an error if no solid c1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): c1.chamfer(0.1) def testChamferAsymmetrical(self): """ Test chamfer API with a box shape for asymmetrical lengths """ cube = CQ(makeUnitCube()).faces(">Z").chamfer(0.1, 0.2) self.saveModel(cube) self.assertEqual(10, cube.faces().size()) # test if edge lengths are different edge = cube.edges(">Z").vals()[0] self.assertAlmostEqual(0.6, edge.Length(), 3) edge = cube.edges("|Z").vals()[0] self.assertAlmostEqual(0.9, edge.Length(), 3) def testChamferCylinder(self): """ Test chamfer API with a cylinder shape """ cylinder = Workplane("XY").circle(1).extrude(1).faces(">Z").chamfer(0.1) self.saveModel(cylinder) self.assertEqual(4, cylinder.faces().size()) def testCounterBores(self): """ Tests making a set of counterbored holes in a face """ c = CQ(makeCube(3.0)) pnts = [(-1.0, -1.0), (0.0, 0.0), (1.0, 1.0)] c = c.faces(">Z").workplane().pushPoints(pnts).cboreHole(0.1, 0.25, 0.25, 0.75) self.assertEqual(18, c.faces().size()) self.saveModel(c) # Tests the case where the depth of the cboreHole is not specified c2 = CQ(makeCube(3.0)) c2 = c2.faces(">Z").workplane().pushPoints(pnts).cboreHole(0.1, 0.25, 0.25) self.assertEqual(15, c2.faces().size()) def testCounterSinks(self): """ Tests countersinks """ s = Workplane(Plane.XY()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) def testSplitKeepingHalf(self): """ Tests splitting a solid """ # drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=True) self.saveModel(result) self.assertEqual(8, result.faces().size()) def testSplitKeepingBoth(self): """ Tests splitting a solid """ # drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=True, keepBottom=True) # stack will have both halves, original will be unchanged # two solids are on the stack, eac self.assertEqual(2, result.solids().size()) self.assertEqual(8, result.solids().item(0).faces().size()) self.assertEqual(8, result.solids().item(1).faces().size()) def testSplitKeepingBottom(self): """ Tests splitting a solid improperly """ # Drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # Now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=False, keepBottom=True) # stack will have both halves, original will be unchanged # one solid is on the stack self.assertEqual(1, result.solids().size()) self.assertEqual(8, result.solids().item(0).faces().size()) def testSplitError(self): # Test split produces the correct error when called with no solid to split. w = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w.split(keepTop=True) # Split should raise ValueError when called with no side kept with raises(ValueError): w.split(keepTop=False, keepBottom=False) def testBoxDefaults(self): """ Tests creating a single box """ s = Workplane("XY").box(2, 3, 4) self.assertEqual(1, s.solids().size()) self.saveModel(s) def testSimpleShell(self): """ Create s simple box """ s1 = Workplane("XY").box(2, 2, 2).faces("+Z").shell(0.05) self.saveModel(s1) self.assertEqual(23, s1.faces().size()) s2 = ( Workplane() .ellipse(4, 2) .extrude(4) .faces(">Z") .shell(+2, kind="intersection") ) self.assertEqual(5, s2.faces().size()) s3 = Workplane().ellipse(4, 2).extrude(4).faces(">Z").shell(+2, kind="arc") self.assertEqual(6, s3.faces().size()) # test error on no solid found s4 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): s4.shell(1) def testClosedShell(self): """ Create a hollow box """ s1 = Workplane("XY").box(2, 2, 2).shell(-0.1) self.assertEqual(12, s1.faces().size()) self.assertTrue(s1.val().isValid()) s2 = Workplane("XY").box(2, 2, 2).shell(0.1) self.assertEqual(32, s2.faces().size()) self.assertTrue(s2.val().isValid()) pts = [(1.0, 0.0), (0.3, 0.2), (0.0, 0.0), (0.3, -0.1), (1.0, -0.03)] s3 = Workplane().polyline(pts).close().extrude(1).shell(-0.05) self.assertTrue(s3.val().isValid()) s4_shape = Workplane("XY").box(2, 2, 2).val() # test that None and empty list both work and are equivalent s4_shell_1 = s4_shape.shell(faceList=None, thickness=-0.1) s4_shell_2 = s4_shape.shell(faceList=[], thickness=-0.1) # this should be the same as the first shape self.assertEqual(len(s4_shell_1.Faces()), s1.faces().size()) self.assertEqual(len(s4_shell_2.Faces()), s1.faces().size()) def testOpenCornerShell(self): s = Workplane("XY").box(1, 1, 1) s1 = s.faces("+Z") s1.add(s.faces("+Y")).add(s.faces("+X")) self.saveModel(s1.shell(0.2)) # Tests the list option variation of add s1 = s.faces("+Z") s1.add(s.faces("+Y")).add([s.faces("+X")]) # Tests the raw object option variation of add s1 = s.faces("+Z") s1.add(s.faces("+Y")).add(s.faces("+X").val().wrapped) def testTopFaceFillet(self): s = Workplane("XY").box(1, 1, 1).faces("+Z").edges().fillet(0.1) self.assertEqual(s.faces().size(), 10) self.saveModel(s) def testBoxPointList(self): """ Tests creating an array of boxes """ s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=True) ) # 1 object, 4 solids because the object is a compound self.assertEqual(4, s.solids().size()) self.assertEqual(1, s.size()) self.saveModel(s) s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=False) ) # 4 objects, 4 solids, because each is a separate solid self.assertEqual(4, s.size()) self.assertEqual(4, s.solids().size()) def testBoxCombine(self): s = ( Workplane("XY") .box(4, 4, 0.5) .faces(">Z") .workplane() .rect(3, 3, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=True) ) self.saveModel(s) self.assertEqual(1, s.solids().size()) # we should have one big solid # should have 26 faces. 6 for the box, and 4x5 for the smaller cubes self.assertEqual(26, s.faces().size()) def testBoxCentered(self): x, y, z = 10, 11, 12 # check that the bottom corner is where we expect it for all possible combinations of centered b = [True, False] expected_x = [-x / 2, 0] expected_y = [-y / 2, 0] expected_z = [-z / 2, 0] for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = ( Workplane() .box(x, y, z, centered=(xopt, yopt, zopt)) .vertices("<X and <Y and <Z") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, zval), 3) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = Workplane().box(x, y, z, centered=val).vertices(">X and >Y and >Z") self.assertEqual(s0.size(), 1) s1 = ( Workplane() .box(x, y, z, centered=(val, val, val)) .vertices(">X and >Y and >Z") ) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals(s0.val().toTuple(), s1.val().toTuple(), 3) def testSphereDefaults(self): s = Workplane("XY").sphere(10) self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(1, s.solids().size()) self.assertEqual(1, s.faces().size()) def testSphereCustom(self): s = Workplane("XY").sphere( 10, angle1=0, angle2=90, angle3=360, centered=(False, False, False) ) self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(2, s.faces().size()) # check that the bottom corner is where we expect it for all possible combinations of centered radius = 10 for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip((True, False), (0, radius)), repeat=3 ): s = Workplane().sphere(radius, centered=(xopt, yopt, zopt)) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals( s.val().Center().toTuple(), (xval, yval, zval), 3 ) # check centered=True produces the same result as centered=(True, True, True) for val in (True, False): s0 = Workplane().sphere(radius, centered=val) self.assertEqual(s0.size(), 1) s1 = Workplane().sphere(radius, centered=(val, val, val)) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals( s0.val().Center().toTuple(), s1.val().Center().toTuple(), 3 ) def testSpherePointList(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .sphere(0.25, combine=False) ) # self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(4, s.solids().size()) self.assertEqual(4, s.faces().size()) def testSphereCombine(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .sphere(2.25, combine=True) ) # self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(1, s.solids().size()) self.assertEqual(4, s.faces().size()) def testCylinderDefaults(self): s = Workplane("XY").cylinder(20, 10) self.assertEqual(1, s.size()) self.assertEqual(1, s.solids().size()) self.assertEqual(3, s.faces().size()) self.assertEqual(2, s.vertices().size()) self.assertTupleAlmostEquals(s.val().Center().toTuple(), (0, 0, 0), 3) def testCylinderCentering(self): radius = 10 height = 40 b = (True, False) expected_x = (0, radius) expected_y = (0, radius) expected_z = (0, height / 2) for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = Workplane("XY").cylinder(height, radius, centered=(xopt, yopt, zopt)) self.assertEqual(1, s.size()) self.assertTupleAlmostEquals( s.val().Center().toTuple(), (xval, yval, zval), 3 ) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = Workplane("XY").cylinder(height, radius, centered=val) self.assertEqual(s0.size(), 1) s1 = Workplane("XY").cylinder(height, radius, centered=(val, val, val)) self.assertEqual(s1.size(), 1) self.assertTupleAlmostEquals( s0.val().Center().toTuple(), s1.val().Center().toTuple(), 3 ) def testWedgeDefaults(self): s = Workplane("XY").wedge(10, 10, 10, 5, 5, 5, 5) self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(5, s.faces().size()) self.assertEqual(5, s.vertices().size()) def testWedgeCentering(self): s = Workplane("XY").wedge( 10, 10, 10, 5, 5, 5, 5, centered=(False, False, False) ) # self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(5, s.faces().size()) self.assertEqual(5, s.vertices().size()) # check that the bottom corner is where we expect it for all possible combinations of centered x, y, z = 10, 11, 12 b = [True, False] expected_x = [-x / 2, 0] expected_y = [-y / 2, 0] expected_z = [-z / 2, 0] for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=(xopt, yopt, zopt)) .vertices("<X and <Y and <Z") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, zval), 3) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=val) .vertices(">X and >Z") ) self.assertEqual(s0.size(), 1) s1 = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=(val, val, val)) .vertices(">X and >Z") ) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals(s0.val().toTuple(), s1.val().toTuple(), 3) def testWedgePointList(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .wedge(10, 10, 10, 5, 5, 5, 5, combine=False) ) # self.saveModel(s) self.assertEqual(4, s.solids().size()) self.assertEqual(20, s.faces().size()) self.assertEqual(20, s.vertices().size()) def testWedgeCombined(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .wedge(10, 10, 10, 5, 5, 5, 5, combine=True) ) # self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(12, s.faces().size()) self.assertEqual(16, s.vertices().size()) def testQuickStartXY(self): s = ( Workplane(Plane.XY()) .box(2, 4, 0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testQuickStartYZ(self): s = ( Workplane(Plane.YZ()) .box(2, 4, 0.5) .faces(">X") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testQuickStartXZ(self): s = ( Workplane(Plane.XZ()) .box(2, 4, 0.5) .faces(">Y") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testDoubleTwistedLoft(self): s = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) s2 = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=-4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) # self.assertEquals(10,s.faces().size()) # self.assertEquals(1,s.solids().size()) s3 = s.combineSolids(s2) self.saveModel(s3) def testTwistedLoft(self): s = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) self.assertEqual(10, s.faces().size()) self.assertEqual(1, s.solids().size()) self.saveModel(s) def testUnions(self): # duplicates a memory problem of some kind reported when combining lots of objects s = Workplane("XY").rect(0.5, 0.5).extrude(5.0) o = [] beginTime = time.time() for i in range(15): t = Workplane("XY").center(10.0 * i, 0).rect(0.5, 0.5).extrude(5.0) o.append(t) # union stuff for oo in o: s = s.union(oo) print("Total time %0.3f" % (time.time() - beginTime)) # Test unioning a Solid object s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toUnion = s.rect(1.0, 1.0).extrude(1.0) resS = currentS.union(toUnion) self.assertEqual(11, resS.faces().size()) with self.assertRaises(ValueError): resS.union(toUnion.faces().val()) # Test syntactic sugar [__add__ method] sugar1 = currentS | toUnion sugar2 = currentS + toUnion self.assertEqual(resS.faces().size(), sugar1.faces().size()) self.assertEqual(resS.faces().size(), sugar2.faces().size()) def testCombine(self): s = Workplane(Plane.XY()) objects1 = s.rect(2.0, 2.0).extrude(0.5).faces(">Z").rect(1.0, 1.0).extrude(0.5) objects1.combine() self.assertEqual(11, objects1.faces().size()) objects1 = s.rect(2.0, 2.0).extrude(0.5) objects2 = s.rect(1.0, 1.0).extrude(0.5).translate((0, 0, 0.5)) objects2 = objects1.add(objects2).combine(glue=True, tol=None) self.assertEqual(11, objects2.faces().size()) def testCombineSolidsInLoop(self): # duplicates a memory problem of some kind reported when combining lots of objects s = Workplane("XY").rect(0.5, 0.5).extrude(5.0) o = [] beginTime = time.time() for i in range(15): t = Workplane("XY").center(10.0 * i, 0).rect(0.5, 0.5).extrude(5.0) o.append(t) # append the 'good way' for oo in o: s.add(oo) s = s.combineSolids() print("Total time %0.3f" % (time.time() - beginTime)) self.saveModel(s) def testClean(self): """ Tests the `clean()` method which is called automatically. """ # make a cube with a splitter edge on one of the faces # autosimplify should remove the splitter s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10) ) self.assertEqual(6, s.faces().size()) # test removal of splitter caused by union operation s = Workplane("XY").box(10, 10, 10).union(Workplane("XY").box(20, 10, 10)) self.assertEqual(6, s.faces().size()) # test removal of splitter caused by extrude+combine operation s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(5, 10, True) .extrude(20) ) self.assertEqual(10, s.faces().size()) # test removal of splitter caused by double hole operation s = ( Workplane("XY") .box(10, 10, 10) .faces(">Z") .workplane() .hole(3, 5) .faces(">Z") .workplane() .hole(3, 10) ) self.assertEqual(7, s.faces().size()) # test removal of splitter caused by cutThruAll s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(10, 5) .cutBlind(-5) .faces(">Z") .workplane(centerOption="CenterOfMass") .center(0, 2.5) .rect(5, 5) .cutThruAll() ) self.assertEqual(18, s.faces().size()) # test removal of splitter with box s = Workplane("XY").box(5, 5, 5).box(10, 5, 2) self.assertEqual(14, s.faces().size()) def testNoClean(self): """ Test the case when clean is disabled. """ # test disabling autoSimplify s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10, clean=False) ) self.assertEqual(7, s.faces().size()) s = ( Workplane("XY") .box(10, 10, 10) .union(Workplane("XY").box(20, 10, 10), clean=False) ) self.assertEqual(14, s.faces().size()) s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(5, 10, True) .extrude(20, clean=False) ) self.assertEqual(12, s.faces().size()) def testExplicitClean(self): """ Test running of `clean()` method explicitly. """ s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10, clean=False) .clean() ) self.assertEqual(6, s.faces().size()) def testPlanes(self): """ Test other planes other than the normal ones (XY, YZ) """ # ZX plane s = Workplane(Plane.ZX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # YX plane s = Workplane(Plane.YX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # YX plane s = Workplane(Plane.YX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # ZY plane s = Workplane(Plane.ZY()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # front plane s = Workplane(Plane.front()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # back plane s = Workplane(Plane.back()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # left plane s = Workplane(Plane.left()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # right plane s = Workplane(Plane.right()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # top plane s = Workplane(Plane.top()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # bottom plane s = Workplane(Plane.bottom()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) def testIsInside(self): """ Testing if one box is inside of another. """ box1 = Workplane(Plane.XY()).box(10, 10, 10) box2 = Workplane(Plane.XY()).box(5, 5, 5) self.assertFalse(box2.val().BoundingBox().isInside(box1.val().BoundingBox())) self.assertTrue(box1.val().BoundingBox().isInside(box2.val().BoundingBox())) def testCup(self): """ UOM = "mm" # # PARAMETERS and PRESETS # These parameters can be manipulated by end users # bottomDiameter = FloatParam(min=10.0,presets={'default':50.0,'tumbler':50.0,'shot':35.0,'tea':50.0,'saucer':100.0},group="Basics", desc="Bottom diameter") topDiameter = FloatParam(min=10.0,presets={'default':85.0,'tumbler':85.0,'shot':50.0,'tea':51.0,'saucer':400.0 },group="Basics", desc="Top diameter") thickness = FloatParam(min=0.1,presets={'default':2.0,'tumbler':2.0,'shot':2.66,'tea':2.0,'saucer':2.0},group="Basics", desc="Thickness") height = FloatParam(min=1.0,presets={'default':80.0,'tumbler':80.0,'shot':59.0,'tea':125.0,'saucer':40.0},group="Basics", desc="Overall height") lipradius = FloatParam(min=1.0,presets={'default':1.0,'tumbler':1.0,'shot':0.8,'tea':1.0,'saucer':1.0},group="Basics", desc="Lip Radius") bottomThickness = FloatParam(min=1.0,presets={'default':5.0,'tumbler':5.0,'shot':10.0,'tea':10.0,'saucer':5.0},group="Basics", desc="BottomThickness") # # Your build method. It must return a solid object # def build(): br = bottomDiameter.value / 2.0 tr = topDiameter.value / 2.0 t = thickness.value s1 = Workplane("XY").circle(br).workplane(offset=height.value).circle(tr).loft() s2 = Workplane("XY").workplane(offset=bottomThickness.value).circle(br - t ).workplane(offset=height.value - t ).circle(tr - t).loft() cup = s1.cut(s2) cup.faces(">Z").edges().fillet(lipradius.value) return cup """ # for some reason shell doesn't work on this simple shape. how disappointing! td = 50.0 bd = 20.0 h = 10.0 t = 1.0 s1 = Workplane("XY").circle(bd).workplane(offset=h).circle(td).loft() s2 = ( Workplane("XY") .workplane(offset=t) .circle(bd - (2.0 * t)) .workplane(offset=(h - t)) .circle(td - (2.0 * t)) .loft() ) s3 = s1.cut(s2) self.saveModel(s3) def testEnclosure(self): """ Builds an electronics enclosure Original FreeCAD script: 81 source statements ,not including variables This script: 34 """ # parameter definitions p_outerWidth = 100.0 # Outer width of box enclosure p_outerLength = 150.0 # Outer length of box enclosure p_outerHeight = 50.0 # Outer height of box enclosure p_thickness = 3.0 # Thickness of the box walls p_sideRadius = 10.0 # Radius for the curves around the sides of the bo # Radius for the curves on the top and bottom edges of the box p_topAndBottomRadius = 2.0 # How far in from the edges the screwposts should be place. p_screwpostInset = 12.0 # Inner Diameter of the screwpost holes, should be roughly screw diameter not including threads p_screwpostID = 4.0 # Outer Diameter of the screwposts.\nDetermines overall thickness of the posts p_screwpostOD = 10.0 p_boreDiameter = 8.0 # Diameter of the counterbore hole, if any p_boreDepth = 1.0 # Depth of the counterbore hole, if # Outer diameter of countersink. Should roughly match the outer diameter of the screw head p_countersinkDiameter = 0.0 # Countersink angle (complete angle between opposite sides, not from center to one side) p_countersinkAngle = 90.0 # Whether to place the lid with the top facing down or not. p_flipLid = True # Height of lip on the underside of the lid.\nSits inside the box body for a snug fit. p_lipHeight = 1.0 # outer shell oshell = ( Workplane("XY") .rect(p_outerWidth, p_outerLength) .extrude(p_outerHeight + p_lipHeight) ) # weird geometry happens if we make the fillets in the wrong order if p_sideRadius > p_topAndBottomRadius: oshell = ( oshell.edges("|Z") .fillet(p_sideRadius) .edges("#Z") .fillet(p_topAndBottomRadius) ) else: oshell = ( oshell.edges("#Z") .fillet(p_topAndBottomRadius) .edges("|Z") .fillet(p_sideRadius) ) # inner shell ishell = ( oshell.faces("<Z") .workplane(p_thickness, True) .rect( (p_outerWidth - 2.0 * p_thickness), (p_outerLength - 2.0 * p_thickness) ) .extrude((p_outerHeight - 2.0 * p_thickness), False) ) # set combine false to produce just the new boss ishell = ishell.edges("|Z").fillet(p_sideRadius - p_thickness) # make the box outer box box = oshell.cut(ishell) # make the screwposts POSTWIDTH = p_outerWidth - 2.0 * p_screwpostInset POSTLENGTH = p_outerLength - 2.0 * p_screwpostInset box = ( box.faces(">Z") .workplane(-p_thickness) .rect(POSTWIDTH, POSTLENGTH, forConstruction=True) .vertices() .circle(p_screwpostOD / 2.0) .circle(p_screwpostID / 2.0) .extrude((-1.0) * (p_outerHeight + p_lipHeight - p_thickness), True) ) # split lid into top and bottom parts (lid, bottom) = ( box.faces(">Z") .workplane(-p_thickness - p_lipHeight) .split(keepTop=True, keepBottom=True) .all() ) # splits into two solids # translate the lid, and subtract the bottom from it to produce the lid inset lowerLid = lid.translate((0, 0, -p_lipHeight)) cutlip = lowerLid.cut(bottom).translate( (p_outerWidth + p_thickness, 0, p_thickness - p_outerHeight + p_lipHeight) ) # compute centers for counterbore/countersink or counterbore topOfLidCenters = ( cutlip.faces(">Z") .workplane() .rect(POSTWIDTH, POSTLENGTH, forConstruction=True) .vertices() ) # add holes of the desired type if p_boreDiameter > 0 and p_boreDepth > 0: topOfLid = topOfLidCenters.cboreHole( p_screwpostID, p_boreDiameter, p_boreDepth, (2.0) * p_thickness ) elif p_countersinkDiameter > 0 and p_countersinkAngle > 0: topOfLid = topOfLidCenters.cskHole( p_screwpostID, p_countersinkDiameter, p_countersinkAngle, (2.0) * p_thickness, ) else: topOfLid = topOfLidCenters.hole(p_screwpostID, (2.0) * p_thickness) # flip lid upside down if desired if p_flipLid: topOfLid.rotateAboutCenter((1, 0, 0), 180) # return the combined result result = topOfLid.union(bottom) self.saveModel(result) def testExtrude(self): """ Test extrude """ r = 1.0 h = 1.0 decimal_places = 9.0 # extrude in one direction s = Workplane("XY").circle(r).extrude(h, both=False) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # calculate the distance between the top and the bottom face delta = top_face.val().Center().sub(bottom_face.val().Center()) self.assertTupleAlmostEquals(delta.toTuple(), (0.0, 0.0, h), decimal_places) # extrude symmetrically s = Workplane("XY").circle(r).extrude(h, both=True) self.assertTrue(len(s.val().Solids()) == 1) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # calculate the distance between the top and the bottom face delta = top_face.val().Center().sub(bottom_face.val().Center()) self.assertTupleAlmostEquals( delta.toTuple(), (0.0, 0.0, 2.0 * h), decimal_places ) # check that non-conplanar extrusion raises with self.assertRaises(ValueError): Workplane().box(1, 1, 1).faces().circle(0.1).extrude(0.1) def testTaperedExtrudeCutBlind(self): h = 1.0 r = 1.0 t = 5 # extrude with a positive taper s = Workplane("XY").circle(r).extrude(h, taper=t) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # top and bottom face area delta = top_face.val().Area() - bottom_face.val().Area() self.assertTrue(delta < 0) # extrude with a negative taper s = Workplane("XY").circle(r).extrude(h, taper=-t) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # top and bottom face area delta = top_face.val().Area() - bottom_face.val().Area() self.assertTrue(delta > 0) # cut a tapered hole s = ( Workplane("XY") .rect(2 * r, 2 * r) .extrude(2 * h) .faces(">Z") .workplane() .rect(r, r) .cutBlind(-h, taper=t) ) middle_face = s.faces(">Z[-2]") self.assertTrue(middle_face.val().Area() < 1) def testClose(self): # Close without endPoint and startPoint coincide. # Create a half-circle a = Workplane(Plane.XY()).sagittaArc((10, 0), 2).close().extrude(2) # Close when endPoint and startPoint coincide. # Create a double half-circle b = ( Workplane(Plane.XY()) .sagittaArc((10, 0), 2) .sagittaArc((0, 0), 2) .close() .extrude(2) ) # The b shape shall have twice the volume of the a shape. self.assertAlmostEqual(a.val().Volume() * 2.0, b.val().Volume()) # Testcase 3 from issue #238 thickness = 3.0 length = 10.0 width = 5.0 obj1 = ( Workplane("XY", origin=(0, 0, -thickness / 2)) .moveTo(length / 2, 0) .threePointArc((0, width / 2), (-length / 2, 0)) .threePointArc((0, -width / 2), (length / 2, 0)) .close() .extrude(thickness) ) os_x = 8.0 # Offset in X os_y = -19.5 # Offset in Y obj2 = ( Workplane("YZ", origin=(os_x, os_y, -thickness / 2)) .moveTo(os_x + length / 2, os_y) .sagittaArc((os_x - length / 2, os_y), width / 2) .sagittaArc((os_x + length / 2, os_y), width / 2) .close() .extrude(thickness) ) # The obj1 shape shall have the same volume as the obj2 shape. self.assertAlmostEqual(obj1.val().Volume(), obj2.val().Volume()) def testText(self): global testdataDir box = Workplane("XY").box(4, 4, 0.5) obj1 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, -0.05, cut=True, halign="left", valign="bottom", font="Sans", ) ) # combined object should have smaller volume self.assertGreater(box.val().Volume(), obj1.val().Volume()) obj2 = ( box.faces(">Z") .workplane() .text("CQ 2.0", 0.5, 0.05, cut=False, combine=True, font="Sans") ) # combined object should have bigger volume self.assertLess(box.val().Volume(), obj2.val().Volume()) # verify that the number of top faces is correct (NB: this is font specific) self.assertEqual(len(obj2.faces(">Z").vals()), 5) obj3 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj3.solids().vals()), 5) obj4 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, fontPath=os.path.join(testdataDir, "OpenSans-Regular.ttf"), cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj4.solids().vals()), 5) # test to see if non-existent file causes segfault obj5 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, fontPath=os.path.join(testdataDir, "OpenSans-Irregular.ttf"), cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj5.solids().vals()), 5) # check it doesn't fall over with int sizes obj1 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 10, -1, cut=True, halign="left", valign="bottom", font="Sans", ) ) def testParametricCurve(self): from math import sin, cos, pi k = 4 r = 1 func = lambda t: ( r * (k + 1) * cos(t) - r * cos((k + 1) * t), r * (k + 1) * sin(t) - r * sin((k + 1) * t), ) res_open = Workplane("XY").parametricCurve(func).extrude(3) # open profile generates an invalid solid self.assertFalse(res_open.solids().val().isValid()) res_closed = ( Workplane("XY").parametricCurve(func, start=0, stop=2 * pi).extrude(3) ) # closed profile will generate a valid solid with 3 faces self.assertTrue(res_closed.solids().val().isValid()) self.assertEqual(len(res_closed.faces().vals()), 3) res_edge = Workplane("XY").parametricCurve(func, makeWire=False) self.assertEqual(len(res_edge.ctx.pendingEdges), 1) self.assertEqual(len(res_edge.ctx.pendingWires), 0) def testMakeShellSolid(self): c0 = math.sqrt(2) / 4 vertices = [[c0, -c0, c0], [c0, c0, -c0], [-c0, c0, c0], [-c0, -c0, -c0]] faces_ixs = [[0, 1, 2, 0], [1, 0, 3, 1], [2, 3, 0, 2], [3, 2, 1, 3]] faces = [] for ixs in faces_ixs: lines = [] for v1, v2 in zip(ixs, ixs[1:]): lines.append( Edge.makeLine(Vector(*vertices[v1]), Vector(*vertices[v2])) ) wire = Wire.combine(lines)[0] faces.append(Face.makeFromWires(wire)) shell = Shell.makeShell(faces) solid = Solid.makeSolid(shell) self.assertTrue(shell.isValid()) self.assertTrue(solid.isValid()) self.assertEqual(len(solid.Vertices()), 4) self.assertEqual(len(solid.Faces()), 4) def testIsInsideSolid(self): # test solid model = Workplane("XY").box(10, 10, 10) solid = model.val() # get first object on stack self.assertTrue(solid.isInside((0, 0, 0))) self.assertFalse(solid.isInside((10, 10, 10))) self.assertTrue(solid.isInside((Vector(3, 3, 3)))) self.assertFalse(solid.isInside((Vector(30.0, 30.0, 30.0)))) self.assertTrue(solid.isInside((0, 0, 4.99), tolerance=0.1)) self.assertTrue(solid.isInside((0, 0, 5))) # check point on surface self.assertTrue(solid.isInside((0, 0, 5.01), tolerance=0.1)) self.assertFalse(solid.isInside((0, 0, 5.1), tolerance=0.1)) # test compound solid model = Workplane("XY").box(10, 10, 10) model = model.moveTo(50, 50).box(10, 10, 10) solid = model.val() self.assertTrue(solid.isInside((0, 0, 0))) self.assertTrue(solid.isInside((50, 50, 0))) self.assertFalse(solid.isInside((50, 56, 0))) # make sure raises on non solid model = Workplane("XY").rect(10, 10) solid = model.val() with self.assertRaises(AttributeError): solid.isInside((0, 0, 0)) # test solid with an internal void void = Workplane("XY").box(10, 10, 10) model = Workplane("XY").box(100, 100, 100).cut(void) solid = model.val() self.assertFalse(solid.isInside((0, 0, 0))) self.assertTrue(solid.isInside((40, 40, 40))) self.assertFalse(solid.isInside((55, 55, 55))) def testWorkplaneCenterOptions(self): """ Test options for specifying origin of workplane """ decimal_places = 9 pts = [(0, 0), (90, 0), (90, 30), (30, 30), (30, 60), (0.0, 60)] r = Workplane("XY").polyline(pts).close().extrude(10.0) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, 0.0, 10.0), decimal_places) origin = ( r.faces(">Z").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (37.5, 22.5, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (45.0, 30.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=(30, 10, 20)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 10.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=Vector(30, 10, 20)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 10.0, 10.0), decimal_places) with self.assertRaises(ValueError): origin = r.faces(">Z").workplane(centerOption="undefined") # test case where plane origin is shifted with center call r = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .center(30, 0) .hole(90) ) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 0.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=(0, 0, 0)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, 0.0, 10.0), decimal_places) # make sure projection works in all directions r = Workplane("YZ").polyline(pts).close().extrude(10.0) origin = ( r.faces(">X") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 0.0, 0.0), decimal_places) origin = ( r.faces(">X").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 37.5, 22.5), decimal_places) origin = ( r.faces(">X") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 45.0, 30.0), decimal_places) r = Workplane("XZ").polyline(pts).close().extrude(10.0) origin = ( r.faces("<Y") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, -10.0, 0.0), decimal_places) origin = ( r.faces("<Y").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (37.5, -10.0, 22.5), decimal_places) origin = ( r.faces("<Y") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (45.0, -10.0, 30.0), decimal_places) def testFindSolid(self): r = Workplane("XY").pushPoints([(-2, 0), (2, 0)]).box(1, 1, 1, combine=False) # there should be two solids on the stack self.assertEqual(len(r.objects), 2) self.assertTrue(isinstance(r.val(), Solid)) # find solid should return a compound of two solids s = r.findSolid() self.assertEqual(len(s.Solids()), 2) self.assertTrue(isinstance(s, Compound)) # if no solids are found, should raise ValueError w = Workplane().hLine(1).close() with raises(ValueError): w.findSolid() def testSlot2D(self): decimal_places = 9 # Ensure it produces a solid with the correct volume result = Workplane("XY").slot2D(4, 1, 0).extrude(1) self.assertAlmostEqual(result.val().Volume(), 3.785398163, decimal_places) # Test for proper expected behaviour when cutting box = Workplane("XY").box(5, 5, 1) result = box.faces(">Z").workplane().slot2D(4, 1, 0).cutThruAll() self.assertAlmostEqual(result.val().Volume(), 21.214601837, decimal_places) result = box.faces(">Z").workplane().slot2D(4, 1, 0).cutBlind(-0.5) self.assertAlmostEqual(result.val().Volume(), 23.107300918, decimal_places) # Test to see if slot is rotated correctly result = Workplane("XY").slot2D(4, 1, 45).extrude(1) point = result.faces(">Z").edges(">X").first().val().startPoint().toTuple() self.assertTupleAlmostEquals( point, (0.707106781, 1.414213562, 1.0), decimal_places ) def test_assembleEdges(self): # Plate with 5 sides and 2 bumps, one side is not co-planar with the other sides # Passes an open wire to assembleEdges so that IsDone is true but Error returns 2 to test the warning functionality. edge_points = [ [-7.0, -7.0, 0.0], [-3.0, -10.0, 3.0], [7.0, -7.0, 0.0], [7.0, 7.0, 0.0], [-7.0, 7.0, 0.0], ] edge_wire = Workplane("XY").polyline( [(-7.0, -7.0), (7.0, -7.0), (7.0, 7.0), (-7.0, 7.0)] ) edge_wire = edge_wire.add( Workplane("YZ") .workplane() .transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)) .spline([(-7.0, 0.0), (3, -3), (7.0, 0.0)]) ) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Embossed star, need to change optional parameters to obtain nice looking result. r1 = 3.0 r2 = 10.0 fn = 6 edge_points = [ [r1 * math.cos(i * math.pi / fn), r1 * math.sin(i * math.pi / fn)] if i % 2 == 0 else [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn)] for i in range(2 * fn + 1) ] edge_wire = Workplane("XY").polyline(edge_points) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Points on hexagonal pattern coordinates, use of pushpoints. r1 = 1.0 fn = 6 edge_points = [ [r1 * math.cos(i * 2 * math.pi / fn), r1 * math.sin(i * 2 * math.pi / fn)] for i in range(fn + 1) ] surface_points = [ [0.25, 0, 0.75], [-0.25, 0, 0.75], [0, 0.25, 0.75], [0, -0.25, 0.75], [0, 0, 2], ] edge_wire = Workplane("XY").polyline(edge_points) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Gyroïd, all edges are splines on different workplanes. edge_points = [ [[3.54, 3.54], [1.77, 0.0], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [-1.77, 0.0], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], ] plane_list = ["XZ", "XY", "YZ", "XZ", "YZ", "XY"] offset_list = [-3.54, 3.54, 3.54, 3.54, -3.54, -3.54] edge_wire = ( Workplane(plane_list[0]) .workplane(offset=-offset_list[0]) .spline(edge_points[0]) ) for i in range(len(edge_points) - 1): edge_wire = edge_wire.add( Workplane(plane_list[i + 1]) .workplane(offset=-offset_list[i + 1]) .spline(edge_points[i + 1]) ) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) def testTag(self): # test tagging result = ( Workplane("XY") .pushPoints([(-2, 0), (2, 0)]) .box(1, 1, 1, combine=False) .tag("2 solids") .union(Workplane("XY").box(6, 1, 1)) ) self.assertEqual(len(result.objects), 1) result = result._getTagged("2 solids") self.assertEqual(len(result.objects), 2) with self.assertRaises(ValueError): result = result._getTagged("3 solids") def testCopyWorkplane(self): obj0 = Workplane("XY").box(1, 1, 10).faces(">Z").workplane() obj1 = Workplane("XY").copyWorkplane(obj0).box(1, 1, 1) self.assertTupleAlmostEquals((0, 0, 5), obj1.val().Center().toTuple(), 9) def testWorkplaneFromTagged(self): # create a flat, wide base. Extrude one object 4 units high, another # object on top of it 6 units high. Go back to base plane. Extrude an # object 11 units high. Assert that top face is 11 units high. result = ( Workplane("XY") .box(10, 10, 1, centered=(True, True, False)) .faces(">Z") .workplane() .tag("base") .center(3, 0) .rect(2, 2) .extrude(4) .faces(">Z") .workplane() .circle(1) .extrude(6) .workplaneFromTagged("base") .center(-3, 0) .circle(1) .extrude(11) ) self.assertTupleAlmostEquals( result.faces(">Z").val().Center().toTuple(), (-3, 0, 12), 9 ) def testWorkplaneOrientationOnVertex(self): # create a 10 unit sized cube on the XY plane parent = Workplane("XY").rect(10.0, 10.0).extrude(10) # assert that the direction tuples reflect accordingly assert parent.plane.xDir.toTuple() == approx((1.0, 0.0, 0.0)) assert parent.plane.zDir.toTuple() == approx((0.0, 0.0, 1.0)) # select the <XZ vertex on the <Y face and create a new workplane. child = parent.faces("<Y").vertices("<XZ").workplane() # assert that the direction tuples reflect the new workplane on the <Y face assert child.plane.xDir.toTuple() == approx((1.0, 0.0, -0.0)) assert child.plane.zDir.toTuple() == approx((0.0, -1.0, -0.0)) def testTagSelectors(self): result0 = Workplane("XY").box(1, 1, 1).tag("box").sphere(1) # result is currently a sphere self.assertEqual(1, result0.faces().size()) # a box has 8 vertices self.assertEqual(8, result0.vertices(tag="box").size()) # 6 faces self.assertEqual(6, result0.faces(tag="box").size()) # 12 edges self.assertEqual(12, result0.edges(tag="box").size()) # 6 wires self.assertEqual(6, result0.wires(tag="box").size()) # create two solids, tag them, join to one solid result1 = ( Workplane("XY") .pushPoints([(1, 0), (-1, 0)]) .box(1, 1, 1) .tag("boxes") .sphere(1) ) self.assertEqual(1, result1.solids().size()) self.assertEqual(2, result1.solids(tag="boxes").size()) self.assertEqual(1, result1.shells().size()) self.assertEqual(2, result1.shells(tag="boxes").size()) # create 4 individual objects, tag it, then combine to one compound result2 = ( Workplane("XY") .rect(4, 4) .vertices() .box(1, 1, 1, combine=False) .tag("4 objs") ) result2 = result2.newObject([Compound.makeCompound(result2.objects)]) self.assertEqual(1, result2.compounds().size()) self.assertEqual(0, result2.compounds(tag="4 objs").size()) def test_interpPlate(self): """ Tests the interpPlate() functionalities Numerical values of Areas and Volumes were obtained with the Area() and Volume() functions on a Linux machine under Debian 10 with python 3.7. """ # example from PythonOCC core_geometry_geomplate.py, use of thickness = 0 returns 2D surface. thickness = 0 edge_points = [ [0.0, 0.0, 0.0], [0.0, 10.0, 0.0], [0.0, 10.0, 10.0], [0.0, 0.0, 10.0], ] surface_points = [[5.0, 5.0, 5.0]] plate_0 = Workplane("XY").interpPlate(edge_points, surface_points, thickness) self.assertTrue(plate_0.val().isValid()) self.assertAlmostEqual(plate_0.val().Area(), 141.218823892, 1) # Plate with 5 sides and 2 bumps, one side is not co-planar with the other sides thickness = 0.1 edge_points = [ [-7.0, -7.0, 0.0], [-3.0, -10.0, 3.0], [7.0, -7.0, 0.0], [7.0, 7.0, 0.0], [-7.0, 7.0, 0.0], ] edge_wire = Workplane("XY").polyline( [(-7.0, -7.0), (7.0, -7.0), (7.0, 7.0), (-7.0, 7.0)] ) # edge_wire = edge_wire.add(Workplane('YZ').workplane().transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)).polyline([(-7.,0.), (3,-3), (7.,0.)])) # In CadQuery Sept-2019 it worked with rotate=Vector(0, 45, 0). In CadQuery Dec-2019 rotate=Vector(45, 0, 0) only closes the wire. edge_wire = edge_wire.add( Workplane("YZ") .workplane() .transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)) .spline([(-7.0, 0.0), (3, -3), (7.0, 0.0)]) ) surface_points = [[-3.0, -3.0, -3.0], [3.0, 3.0, 3.0]] plate_1 = Workplane("XY").interpPlate(edge_wire, surface_points, thickness) self.assertTrue(plate_1.val().isValid()) self.assertAlmostEqual(plate_1.val().Volume(), 26.124970206, 2) # Embossed star, need to change optional parameters to obtain nice looking result. r1 = 3.0 r2 = 10.0 fn = 6 thickness = 0.1 edge_points = [ [r1 * math.cos(i * math.pi / fn), r1 * math.sin(i * math.pi / fn)] if i % 2 == 0 else [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn)] for i in range(2 * fn + 1) ] edge_wire = Workplane("XY").polyline(edge_points) r2 = 4.5 surface_points = [ [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn), 1.0] for i in range(2 * fn) ] + [[0.0, 0.0, -2.0]] plate_2 = Workplane("XY").interpPlate( edge_wire, surface_points, thickness, combine=True, clean=True, degree=3, nbPtsOnCur=15, nbIter=2, anisotropy=False, tol2d=0.00001, tol3d=0.0001, tolAng=0.01, tolCurv=0.1, maxDeg=8, maxSegments=49, ) self.assertTrue(plate_2.val().isValid()) self.assertAlmostEqual(plate_2.val().Volume(), 10.956054314, 0) # Points on hexagonal pattern coordinates, use of pushpoints. r1 = 1.0 N = 3 ca = math.cos(30.0 * math.pi / 180.0) sa = math.sin(30.0 * math.pi / 180.0) # EVEN ROWS pts = [ (-3.0, -3.0), (-1.267949, -3.0), (0.464102, -3.0), (2.196152, -3.0), (-3.0, 0.0), (-1.267949, 0.0), (0.464102, 0.0), (2.196152, 0.0), (-2.133974, -1.5), (-0.401923, -1.5), (1.330127, -1.5), (3.062178, -1.5), (-2.133975, 1.5), (-0.401924, 1.5), (1.330127, 1.5), (3.062178, 1.5), ] # Spike surface thickness = 0.1 fn = 6 edge_points = [ [ r1 * math.cos(i * 2 * math.pi / fn + 30 * math.pi / 180), r1 * math.sin(i * 2 * math.pi / fn + 30 * math.pi / 180), ] for i in range(fn + 1) ] surface_points = [ [ r1 / 4 * math.cos(i * 2 * math.pi / fn + 30 * math.pi / 180), r1 / 4 * math.sin(i * 2 * math.pi / fn + 30 * math.pi / 180), 0.75, ] for i in range(fn + 1) ] + [[0, 0, 2]] edge_wire = Workplane("XY").polyline(edge_points) plate_3 = ( Workplane("XY") .pushPoints(pts) .interpPlate( edge_wire, surface_points, thickness, combine=False, clean=False, degree=2, nbPtsOnCur=20, nbIter=2, anisotropy=False, tol2d=0.00001, tol3d=0.0001, tolAng=0.01, tolCurv=0.1, maxDeg=8, maxSegments=9, ) ) self.assertTrue(plate_3.val().isValid()) self.assertAlmostEqual(plate_3.val().Volume(), 0.45893954685189414, 1) # Gyroïd, all edges are splines on different workplanes. thickness = 0.1 edge_points = [ [[3.54, 3.54], [1.77, 0.0], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [-1.77, 0.0], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], ] plane_list = ["XZ", "XY", "YZ", "XZ", "YZ", "XY"] offset_list = [-3.54, 3.54, 3.54, 3.54, -3.54, -3.54] edge_wire = ( Workplane(plane_list[0]) .workplane(offset=-offset_list[0]) .spline(edge_points[0]) ) for i in range(len(edge_points) - 1): edge_wire = edge_wire.add( Workplane(plane_list[i + 1]) .workplane(offset=-offset_list[i + 1]) .spline(edge_points[i + 1]) ) surface_points = [[0, 0, 0]] plate_4 = Workplane("XY").interpPlate(edge_wire, surface_points, thickness) self.assertTrue(plate_4.val().isValid()) self.assertAlmostEqual(plate_4.val().Volume(), 7.760559490, 2) def testTangentArcToPoint(self): # create a simple shape with tangents of straight edges and see if it has the correct area s0 = ( Workplane("XY") .hLine(1) .tangentArcPoint((1, 1), relative=False) .hLineTo(0) .tangentArcPoint((0, 0), relative=False) .close() .extrude(1) ) area0 = s0.faces(">Z").val().Area() self.assertAlmostEqual(area0, (1 + math.pi * 0.5 ** 2), 4) # test relative coords s1 = ( Workplane("XY") .hLine(1) .tangentArcPoint((0, 1), relative=True) .hLineTo(0) .tangentArcPoint((0, -1), relative=True) .close() .extrude(1) ) self.assertTupleAlmostEquals( s1.val().Center().toTuple(), s0.val().Center().toTuple(), 4 ) self.assertAlmostEqual(s1.val().Volume(), s0.val().Volume(), 4) # consecutive tangent arcs s1 = ( Workplane("XY") .vLine(2) .tangentArcPoint((1, 0)) .tangentArcPoint((1, 0)) .tangentArcPoint((1, 0)) .vLine(-2) .close() .extrude(1) ) self.assertAlmostEqual( s1.faces(">Z").val().Area(), 2 * 3 + 0.5 * math.pi * 0.5 ** 2, 4 ) # tangentArc on the end of a spline # spline will be a simple arc of a circle, then finished off with a # tangentArcPoint angles = [idx * 1.5 * math.pi / 10 for idx in range(10)] pts = [(math.sin(a), math.cos(a)) for a in angles] s2 = ( Workplane("XY") .spline(pts) .tangentArcPoint((0, 1), relative=False) .close() .extrude(1) ) # volume should almost be pi, but not accurately because we need to # start with a spline self.assertAlmostEqual(s2.val().Volume(), math.pi, 1) # assert local coords are mapped to global correctly arc0 = Workplane("XZ", origin=(1, 1, 1)).hLine(1).tangentArcPoint((1, 1)).val() self.assertTupleAlmostEquals(arc0.endPoint().toTuple(), (3, 1, 2), 4) # tangentArcPoint with 3-tuple argument w0 = Workplane("XY").lineTo(1, 1).tangentArcPoint((1, 1, 1)).wire() zmax = w0.val().BoundingBox().zmax self.assertAlmostEqual(zmax, 1, 1) def test_findFromEdge(self): part = Workplane("XY", origin=(1, 1, 1)).hLine(1) found_edge = part._findFromEdge(useLocalCoords=False) self.assertTupleAlmostEquals(found_edge.startPoint().toTuple(), (1, 1, 1), 3) self.assertTupleAlmostEquals(found_edge.Center().toTuple(), (1.5, 1, 1), 3) self.assertTupleAlmostEquals(found_edge.endPoint().toTuple(), (2, 1, 1), 3) found_edge = part._findFromEdge(useLocalCoords=True) self.assertTupleAlmostEquals(found_edge.endPoint().toTuple(), (1, 0, 0), 3) # check _findFromEdge can find a spline pts = [(0, 0), (0, 1), (1, 2), (2, 4)] spline0 = Workplane("XZ").spline(pts)._findFromEdge() self.assertTupleAlmostEquals((2, 0, 4), spline0.endPoint().toTuple(), 3) # check method fails if no edge is present part2 = Workplane("XY").box(1, 1, 1) with self.assertRaises(RuntimeError): part2._findFromEdge() with self.assertRaises(RuntimeError): part2._findFromEdge(useLocalCoords=True) def testMakeHelix(self): h = 10 pitch = 1.5 r = 1.2 obj = Wire.makeHelix(pitch, h, r) bb = obj.BoundingBox() self.assertAlmostEqual(bb.zlen, h, 1) def testUnionCompound(self): box1 = Workplane("XY").box(10, 20, 30) box2 = Workplane("YZ").box(10, 20, 30) shape_to_cut = Workplane("XY").box(15, 15, 15).translate((8, 8, 8)) list_of_shapes = [] for o in box1.all(): list_of_shapes.extend(o.vals()) for o in box2.all(): list_of_shapes.extend(o.vals()) obj = Workplane("XY").newObject(list_of_shapes).cut(shape_to_cut) assert obj.val().isValid() def testSection(self): box = Workplane("XY", origin=(1, 2, 3)).box(1, 1, 1) s1 = box.section() s2 = box.section(0.5) self.assertAlmostEqual(s1.faces().val().Area(), 1) self.assertAlmostEqual(s2.faces().val().Area(), 1) line = Workplane("XY").hLine(1) with self.assertRaises(ValueError): line.section() def testGlue(self): box1 = Workplane("XY").rect(1, 1).extrude(2) box2 = Workplane("XY", origin=(0, 1, 0)).rect(1, 1).extrude(1) res = box1.union(box2, glue=True) self.assertEqual(res.faces().size(), 8) obj = obj = ( Workplane("XY").rect(1, 1).extrude(2).moveTo(0, 2).rect(1, 1).extrude(2) ) res = obj.union(box2, glue=True) self.assertEqual(res.faces().size(), 10) def testFuzzyBoolOp(self): eps = 1e-3 box1 = Workplane("XY").box(1, 1, 1) box2 = Workplane("XY", origin=(1 + eps, 0.0)).box(1, 1, 1) box3 = Workplane("XY", origin=(2, 0, 0)).box(1, 1, 1) res = box1.union(box2) res_fuzzy = box1.union(box2, tol=eps) res_fuzzy2 = box1.union(box3).union(box2, tol=eps) self.assertEqual(res.solids().size(), 2) self.assertEqual(res_fuzzy.solids().size(), 1) self.assertEqual(res_fuzzy2.solids().size(), 1) def testLocatedMoved(self): box = Solid.makeBox(1, 1, 1, Vector(-0.5, -0.5, -0.5)) loc = Location(Vector(1, 1, 1)) box1 = box.located(loc) self.assertTupleAlmostEquals(box1.Center().toTuple(), (1, 1, 1), 6) self.assertTupleAlmostEquals(box.Center().toTuple(), (0, 0, 0), 6) box.locate(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (1, 1, 1), 6) box2 = box.moved(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (1, 1, 1), 6) self.assertTupleAlmostEquals(box2.Center().toTuple(), (2, 2, 2), 6) box.move(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (2, 2, 2), 6) def testNullShape(self): from OCP.TopoDS import TopoDS_Shape s = TopoDS_Shape() # make sure raises on non solid with self.assertRaises(ValueError): r = occ_impl.shapes.downcast(s) def testCenterOfBoundBox(self): obj = Workplane().pushPoints([(0, 0), (2, 2)]).box(1, 1, 1) c = obj.workplane(centerOption="CenterOfBoundBox").plane.origin self.assertTupleAlmostEquals(c.toTuple(), (1, 1, 0), 6) def testOffset2D(self): w1 = Workplane().rect(1, 1).offset2D(0.5, "arc") self.assertEqual(w1.edges().size(), 8) w2 = Workplane().rect(1, 1).offset2D(0.5, "tangent") self.assertEqual(w2.edges().size(), 4) w3 = Workplane().rect(1, 1).offset2D(0.5, "intersection") self.assertEqual(w3.edges().size(), 4) w4 = Workplane().pushPoints([(0, 0), (0, 5)]).rect(1, 1).offset2D(-0.5) self.assertEqual(w4.wires().size(), 0) w5 = Workplane().pushPoints([(0, 0), (0, 5)]).rect(1, 1).offset2D(-0.25) self.assertEqual(w5.wires().size(), 2) r = 20 s = 7 t = 1.5 points = [ (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] s = ( Workplane("XY") .polyline(points) .mirrorX() .mirrorY() .offset2D(-0.9) .extrude(1) ) self.assertEqual(s.solids().size(), 4) # test forConstruction # forConstruction=True should place results in objects, not ctx.pendingWires w6 = Workplane().hLine(1).vLine(1).close().offset2D(0.5, forConstruction=True) self.assertEqual(len(w6.ctx.pendingWires), 0) self.assertEqual(w6.size(), 1) self.assertEqual(type(w6.val()), Wire) # make sure the resulting wire has forConstruction set self.assertEqual(w6.val().forConstruction, True) def testConsolidateWires(self): w1 = Workplane().lineTo(0, 1).lineTo(1, 1).consolidateWires() self.assertEqual(w1.size(), 1) w1 = Workplane().consolidateWires() self.assertEqual(w1.size(), 0) def testLocationAt(self): r = 1 e = Wire.makeHelix(r, r, r).Edges()[0] locs_frenet = e.locations([0, 1], frame="frenet") T1 = locs_frenet[0].wrapped.Transformation() T2 = locs_frenet[1].wrapped.Transformation() self.assertAlmostEqual(T1.TranslationPart().X(), r, 6) self.assertAlmostEqual(T2.TranslationPart().X(), r, 6) self.assertAlmostEqual( T1.GetRotation().GetRotationAngle(), -T2.GetRotation().GetRotationAngle(), 6 ) ga = e._geomAdaptor() locs_corrected = e.locations( [ga.FirstParameter(), ga.LastParameter()], mode="parameter", frame="corrected", ) T3 = locs_corrected[0].wrapped.Transformation() T4 = locs_corrected[1].wrapped.Transformation() self.assertAlmostEqual(T3.TranslationPart().X(), r, 6) self.assertAlmostEqual(T4.TranslationPart().X(), r, 6) w = Wire.assembleEdges( [ Edge.makeLine(Vector(), Vector(0, 1)), Edge.makeLine(Vector(0, 1), Vector(1, 1)), ] ) locs_wire = e.locations([0, 1]) T5 = locs_wire[0].wrapped.Transformation() T6 = locs_wire[1].wrapped.Transformation() self.assertAlmostEqual(T5.TranslationPart().X(), r, 0) self.assertAlmostEqual(T6.TranslationPart().X(), r, 1) def testNormal(self): circ = Workplane().circle(1).edges().val() n = circ.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) ell = Workplane().ellipse(1, 2).edges().val() n = ell.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) r = Workplane().rect(1, 2).wires().val() n = r.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) with self.assertRaises(ValueError): edge = Workplane().rect(1, 2).edges().val() n = edge.normal() def testPositionAt(self): # test with an open wire w = Workplane().lineTo(0, 1).lineTo(1, 1).wire().val() p0 = w.positionAt(0.0) p1 = w.positionAt(0.5) p2 = w.positionAt(1.0) self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 1, 0), 6) self.assertTupleAlmostEquals(p2.toTuple(), (1, 1, 0), 6) p0 = w.positionAt(0.0, mode="param") self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) p0, p1, p2 = w.positions([0.0, 0.25, 0.5]) self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 0.5, 0), 6) self.assertTupleAlmostEquals(p2.toTuple(), (0, 1, 0), 6) # test with a closed wire w = Workplane().lineTo(0, 1).close().wire().val() p0 = w.positionAt(0.0) p1 = w.positionAt(0.5) p2 = w.positionAt(1.0) self.assertTupleAlmostEquals(p0.toTuple(), p2.toTuple(), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 1, 0), 6) def testTangengAt(self): pts = [(0, 0), (-1, 1), (-2, 0), (-1, 0)] path = Workplane("XZ").spline(pts, tangents=((0, 1), (1, 0))).val() self.assertTrue( path.tangentAt(0.0, mode="parameter") == path.tangentAt(0.0, mode="length") ) self.assertFalse( path.tangentAt(0.5, mode="parameter") == path.tangentAt(0.5, mode="length") ) arc = Workplane().radiusArc((2, 0), 1).val() self.assertTupleAlmostEquals( arc.tangentAt(math.pi / 2, "parameter").toTuple(), (1, 0, 0), 6 ) self.assertTupleAlmostEquals( arc.tangentAt(0.5, "length").toTuple(), (1, 0, 0), 6 ) def testEnd(self): with self.assertRaises(ValueError): Workplane().end() self.assertTrue(Workplane().objects == []) self.assertTrue(Workplane().box(1, 1, 1).end().objects == []) self.assertTrue(Workplane().box(1, 1, 1).box(2, 2, 1).end(2).objects == []) def testCutEach(self): # base shape: w = Workplane().box(3, 2, 2) # cutter: c = Workplane().box(2, 2, 2).val() # cut all the corners off w0 = w.vertices().cutEach(lambda loc: c.located(loc)) # we are left with a 1x2x2 box: self.assertAlmostEqual(w0.val().Volume(), 4, 3) # test error on no solid found w1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w1.cutEach(lambda loc: c.located(loc)) def testCutBlind(self): # cutBlind is already tested in several of the complicated tests, so this method is short. # test ValueError on no solid found w0 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w0.cutBlind(1) def testFindFace(self): # if there are no faces to find, should raise ValueError w0 = Workplane() with raises(ValueError): w0.findFace() w1 = Workplane().box(1, 1, 1).faces(">Z") self.assertTrue(isinstance(w1.findFace(), Face)) with raises(ValueError): w1.findFace(searchStack=False) w2 = w1.workplane().circle(0.1).extrude(0.1) self.assertTrue(isinstance(w2.findFace(searchParents=True), Face)) with raises(ValueError): w2.findFace(searchParents=False) def testPopPending(self): # test pending edges w0 = Workplane().hLine(1) self.assertEqual(len(w0.ctx.pendingEdges), 1) edges = w0.ctx.popPendingEdges() self.assertEqual(len(edges), 1) self.assertEqual(edges[0], w0.val()) # pending edges should now be cleared self.assertEqual(len(w0.ctx.pendingEdges), 0) # test pending wires w1 = Workplane().hLine(1).vLine(1).close() wire = w1.val() self.assertEqual(w1.ctx.pendingWires[0], wire) pop_pending_output = w1.ctx.popPendingWires() self.assertEqual(pop_pending_output[0], wire) # pending wires should now be cleared self.assertEqual(len(w1.ctx.pendingWires), 0) # test error when empty pending edges w2 = Workplane() # the following 2 should not raise an exception w2.ctx.popPendingEdges(errorOnEmpty=False) w2.ctx.popPendingWires(errorOnEmpty=False) # empty edges w3 = Workplane().hLine(1).vLine(1).close() with self.assertRaises(ValueError): w3.ctx.popPendingEdges() # empty wires w4 = Workplane().circle(1).extrude(1) with self.assertRaises(ValueError): w4.ctx.popPendingWires() # test via cutBlind w5 = Workplane().circle(1).extrude(1) with self.assertRaises(ValueError): w5.cutBlind(-1) def testCompSolid(self): from OCP.BRepPrimAPI import BRepPrimAPI_MakePrism tool = Solid.makeSphere(1, angleDegrees3=120) shell = tool.Shells()[0] v = Vector(0, 0, 1) builder = BRepPrimAPI_MakePrism(shell.wrapped, v.wrapped) result = Shape.cast(builder.Shape()) self.assertEqual(len(result.CompSolids()), 1) self.assertEqual(len(result.Solids()), 4) def test2Dfillet(self): r = Workplane().rect(1, 2).wires().val() f = Face.makeFromWires(r) verts = r.Vertices() self.assertEqual(len(f.fillet2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.fillet2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.fillet2D(0.25, verts).Vertices()), 8) # Test fillet2D with open wire and single vertex w0 = Workplane().hLine(1).vLine(1).wire() w0_verts = w0.vertices(">X and <Y").vals() unfilleted_wire0 = w0.val() filleted_wire0 = unfilleted_wire0.fillet2D(0.5, w0_verts) self.assertEqual(len(filleted_wire0.Vertices()), 4) # the filleted wire is shorter than the original self.assertGreater(unfilleted_wire0.Length() - filleted_wire0.Length(), 0.1) def test2Dchamfer(self): r = Workplane().rect(1, 2).wires().val() f = Face.makeFromWires(r) verts = r.Vertices() self.assertEqual(len(f.chamfer2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.chamfer2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.chamfer2D(0.25, verts).Vertices()), 8) r = Workplane().hLine(1).vLine(1).wire().val() vs = r.Vertices() self.assertEqual(len(r.chamfer2D(0.25, [vs[1]]).Vertices()), 4) with raises(ValueError): r.chamfer2D(0.25, [vs[0]]) def testSplineApprox(self): from .naca import naca5305 from math import pi, cos pts = [Vector(e[0], e[1], 0) for e in naca5305] # spline e1 = Edge.makeSplineApprox(pts, 1e-6, maxDeg=6, smoothing=(1, 1, 1)) e2 = Edge.makeSplineApprox(pts, 1e-6, minDeg=2, maxDeg=6) self.assertTrue(e1.isValid()) self.assertTrue(e2.isValid()) self.assertTrue(e1.Length() > e2.Length()) with raises(ValueError): e4 = Edge.makeSplineApprox(pts, 1e-6, maxDeg=3, smoothing=(1, 1, 1.0)) pts_closed = pts + [pts[0]] e3 = Edge.makeSplineApprox(pts_closed) w = Edge.makeSplineApprox(pts).close() self.assertTrue(e3.IsClosed()) self.assertTrue(w.IsClosed()) # Workplane method w1 = Workplane().splineApprox(pts) w2 = Workplane().splineApprox(pts, forConstruction=True) w3 = Workplane().splineApprox(pts, makeWire=True) w4 = Workplane().splineApprox(pts, makeWire=True, forConstruction=True) self.assertEqual(w1.edges().size(), 1) self.assertEqual(len(w1.ctx.pendingEdges), 1) self.assertEqual(w2.edges().size(), 1) self.assertEqual(len(w2.ctx.pendingEdges), 0) self.assertEqual(w3.wires().size(), 1) self.assertEqual(len(w3.ctx.pendingWires), 1) self.assertEqual(w4.wires().size(), 1) self.assertEqual(len(w4.ctx.pendingWires), 0) # spline surface N = 40 T = 20 A = 5 pts = [ [ Vector(i, j, A * cos(2 * pi * i / T) * cos(2 * pi * j / T)) for i in range(N + 1) ] for j in range(N + 1) ] f1 = Face.makeSplineApprox(pts, smoothing=(1, 1, 1), maxDeg=6) f2 = Face.makeSplineApprox(pts) self.assertTrue(f1.isValid()) self.assertTrue(f2.isValid()) with raises(ValueError): f3 = Face.makeSplineApprox(pts, smoothing=(1, 1, 1), maxDeg=3) def testParametricSurface(self): from math import pi, cos r1 = Workplane().parametricSurface( lambda u, v: (u, v, cos(pi * u) * cos(pi * v)), start=-1, stop=1 ) self.assertTrue(r1.faces().val().isValid()) r2 = Workplane().box(1, 1, 3).split(r1) self.assertTrue(r2.solids().val().isValid()) self.assertEqual(r2.solids().size(), 2) def testEdgeWireClose(self): # test with edge e0 = Edge.makeThreePointArc(Vector(0, 0, 0), Vector(1, 1, 0), Vector(0, 2, 0)) self.assertFalse(e0.IsClosed()) w0 = e0.close() self.assertTrue(w0.IsClosed()) # test with already closed edge e1 = Edge.makeCircle(1) self.assertTrue(e1.IsClosed()) e2 = e1.close() self.assertTrue(e2.IsClosed()) self.assertEqual(type(e1), type(e2)) # test with already closed WIRE w1 = Wire.makeCircle(1, Vector(), Vector(0, 0, 1)) self.assertTrue(w1.IsClosed()) w2 = w1.close() self.assertTrue(w1 is w2) def testSplitShape(self): """ Testing the Shape.split method. """ # split an edge with a vertex e0 = Edge.makeCircle(1, (0, 0, 0), (0, 0, 1)) v0 = Vertex.makeVertex(0, 1, 0) list_of_edges = e0.split(v0).Edges() self.assertEqual(len(list_of_edges), 2) self.assertTrue(Vector(0, 1, 0) in [e.endPoint() for e in list_of_edges]) # split a circle with multiple vertices angles = [2 * math.pi * idx / 10 for idx in range(10)] vecs = [Vector(math.sin(a), math.cos(a), 0) for a in angles] vertices = [Vertex.makeVertex(*v.toTuple()) for v in vecs] edges = e0.split(*vertices).Edges() self.assertEqual(len(edges), len(vertices) + 1) endpoints = [e.endPoint() for e in edges] self.assertTrue(all([v in endpoints for v in vecs])) def testBrepImportExport(self): # import/export to file s = Workplane().box(1, 1, 1).val() s.exportBrep("test.brep") si = Shape.importBrep("test.brep") self.assertTrue(si.isValid()) self.assertAlmostEqual(si.Volume(), 1) # import/export to BytesIO from io import BytesIO bio = BytesIO() s.exportBrep(bio) bio.seek(0) si = Shape.importBrep("test.brep") self.assertTrue(si.isValid()) self.assertAlmostEqual(si.Volume(), 1) def testFaceToPln(self): origin = (1, 2, 3) normal = (1, 1, 1) f0 = Face.makePlane(length=None, width=None, basePnt=origin, dir=normal) p0 = f0.toPln() self.assertTrue(Vector(p0.Location()) == Vector(origin)) self.assertTrue(Vector(p0.Axis().Direction()) == Vector(normal).normalized()) origin1 = (0, 0, -3) normal1 = (-1, 1, -1) f1 = Face.makePlane(length=0.1, width=100, basePnt=origin1, dir=normal1) p1 = f1.toPln() self.assertTrue(Vector(p1.Location()) == Vector(origin1)) self.assertTrue(Vector(p1.Axis().Direction()) == Vector(normal1).normalized()) f2 = Workplane().box(1, 1, 10, centered=False).faces(">Z").val() p2 = f2.toPln() self.assertTrue(p2.Contains(f2.Center().toPnt(), 0.1)) self.assertTrue(Vector(p2.Axis().Direction()) == f2.normalAt())
33.749893
164
0.528441
import math, os.path, time, tempfile from random import choice from random import random from random import randrange from itertools import product from pytest import approx, raises from cadquery import * from cadquery import exporters from cadquery import occ_impl from tests import ( BaseTest, writeStringToFile, makeUnitCube, readFileAsString, makeUnitSquareWire, makeCube, ) testdataDir = os.path.join(os.path.dirname(__file__), "testdata") OUTDIR = tempfile.gettempdir() SUMMARY_FILE = os.path.join(OUTDIR, "testSummary.html") SUMMARY_TEMPLATE = """<html> <head> <style type="text/css"> .testResult{ background: #eeeeee; margin: 50px; border: 1px solid black; } </style> </head> <body> <!--TEST_CONTENT--> </body> </html>""" TEST_RESULT_TEMPLATE = """ <div class="testResult"><h3>%(name)s</h3> %(svg)s </div> <!--TEST_CONTENT--> """ writeStringToFile(SUMMARY_TEMPLATE, SUMMARY_FILE) class TestCadQuery(BaseTest): def tearDown(self): svgFile = os.path.join(OUTDIR, self._testMethodName + ".svg") if os.path.exists(svgFile): existingSummary = readFileAsString(SUMMARY_FILE) svgText = readFileAsString(svgFile) svgText = svgText.replace( '<?xml version="1.0" encoding="UTF-8" standalone="no"?>', "" ) existingSummary = existingSummary.replace( "<!--TEST_CONTENT-->", TEST_RESULT_TEMPLATE % (dict(svg=svgText, name=self._testMethodName)), ) writeStringToFile(existingSummary, SUMMARY_FILE) def saveModel(self, shape): shape.exportSvg(os.path.join(OUTDIR, self._testMethodName + ".svg")) shape.val().exportStep(os.path.join(OUTDIR, self._testMethodName + ".step")) def testToOCC(self): r = Workplane("XY").rect(5, 5).extrude(5) r = r.toOCC() import OCP self.assertEqual(type(r), OCP.TopoDS.TopoDS_Compound) def testToSVG(self): r = Workplane("XY").rect(5, 5).extrude(5) r_str = r.toSvg() self.assertTrue(r_str.index('path d="M') > 0) self.assertTrue( r_str.index('line x1="30" y1="-30" x2="58" y2="-15" stroke-width="3"') > 0 ) def testCubePlugin(self): # make the plugin method def makeCubes(self, length): # self refers to the CQ or Workplane object # create the solid s = Solid.makeBox(length, length, length, Vector(0, 0, 0)) # use CQ utility method to iterate over the stack an position the cubes return self.eachpoint(lambda loc: s.located(loc), True) # link the plugin in Workplane.makeCubes = makeCubes # call it result = ( Workplane("XY") .box(6.0, 8.0, 0.5) .faces(">Z") .rect(4.0, 4.0, forConstruction=True) .vertices() ) result = result.makeCubes(1.0) result = result.combineSolids() self.saveModel(result) self.assertEqual(1, result.solids().size()) def testCylinderPlugin(self): def cylinders(self, radius, height): # construct a cylinder at (0,0,0) c = Solid.makeCylinder(radius, height, Vector(0, 0, 0)) # combine all the cylinders into a single compound r = self.eachpoint(lambda loc: c.located(loc), True).combineSolids() return r Workplane.cyl = cylinders # now test. here we want weird workplane to see if the objects are transformed right s = ( Workplane(Plane(Vector((0, 0, 0)), Vector((1, -1, 0)), Vector((1, 1, 0)))) .rect(2.0, 3.0, forConstruction=True) .vertices() .cyl(0.25, 0.5) ) self.assertEqual(4, s.solids().size()) self.saveModel(s) def testPolygonPlugin(self): def rPoly(self, nSides, diameter): def _makePolygon(loc): # pnt is a vector in local coordinates angle = 2.0 * math.pi / nSides pnts = [] for i in range(nSides + 1): pnts.append( Vector( (diameter / 2.0 * math.cos(angle * i)), (diameter / 2.0 * math.sin(angle * i)), 0, ) ) return Wire.makePolygon(pnts).located(loc) return self.eachpoint(_makePolygon, True) Workplane.rPoly = rPoly s = ( Workplane("XY") .box(4.0, 4.0, 0.25) .faces(">Z") .workplane() .rect(2.0, 2.0, forConstruction=True) .vertices() .rPoly(5, 0.5) .cutThruAll() ) # 6 base sides, 4 pentagons, 5 sides each = 26 self.assertEqual(26, s.faces().size()) self.saveModel(s) def testFluentMethodInheritance(self): class ExtendedWorkplane(Workplane): def nonExistentInWorkplane(self): pass # Call an inherited fluent method: wp = ExtendedWorkplane("XY").moveTo(1, 2) # Verify that the inherited method returned an instance of the derived # class: self.assertEqual(type(wp), ExtendedWorkplane) # The following is redundant, but can make the use case clearer. # This must not raise an AttributeError: wp.nonExistentInWorkplane() def testPointList(self): c = CQ(makeUnitCube()) s = c.faces(">Z").workplane().pushPoints([(-0.3, 0.3), (0.3, 0.3), (0, 0)]) self.assertEqual(3, s.size()) # TODO: is the ability to iterate over points with circle really worth it? # maybe we should just require using all() and a loop for this. the semantics and # possible combinations got too hard ( ie, .circle().circle() ) was really odd body = s.circle(0.05).cutThruAll() self.saveModel(body) self.assertEqual(9, body.faces().size()) # Test the case when using eachpoint with only a blank workplane def callback_fn(loc): self.assertEqual( Vector(0, 0, 0), Vector(loc.wrapped.Transformation().TranslationPart()) ) r = Workplane("XY") r.objects = [] r.eachpoint(callback_fn) def testWorkplaneFromFace(self): # make a workplane on the top face s = CQ(makeUnitCube()).faces(">Z").workplane() r = s.circle(0.125).cutBlind(-2.0) self.saveModel(r) # the result should have 7 faces self.assertEqual(7, r.faces().size()) self.assertEqual(type(r.val()), Compound) self.assertEqual(type(r.first().val()), Compound) def testFrontReference(self): # make a workplane on the top face s = CQ(makeUnitCube()).faces("front").workplane() r = s.circle(0.125).cutBlind(-2.0) self.saveModel(r) # the result should have 7 faces self.assertEqual(7, r.faces().size()) self.assertEqual(type(r.val()), Compound) self.assertEqual(type(r.first().val()), Compound) def testRotate(self): box = Workplane("XY").box(1, 1, 5) box.rotate((0, 0, 0), (1, 0, 0), 90) startPoint = box.faces("<Y").edges("<X").first().val().startPoint().toTuple() endPoint = box.faces("<Y").edges("<X").first().val().endPoint().toTuple() self.assertEqual(-0.5, startPoint[0]) self.assertEqual(-0.5, startPoint[1]) self.assertEqual(-2.5, startPoint[2]) self.assertEqual(-0.5, endPoint[0]) self.assertEqual(-0.5, endPoint[1]) self.assertEqual(2.5, endPoint[2]) def testPlaneRotateZNormal(self): for _ in range(100): angle = (random() - 0.5) * 720 xdir = Vector(random(), random(), random()).normalized() rdir = Vector(random(), random(), random()).normalized() zdir = xdir.cross(rdir).normalized() origin = (random(), random(), random()) plane = Plane(origin=origin, xDir=xdir, normal=zdir) rotated = plane.rotated((0, 0, angle)) assert rotated.zDir.toTuple() == approx(zdir.toTuple()) assert rotated.origin.toTuple() == approx(origin) def testPlaneRotateConcat(self): xdir = (1, 0, 0) normal = (0, 0, 1) k = 2.0 ** 0.5 / 2.0 origin = (2, -1, 1) plane = Plane(origin=origin, xDir=xdir, normal=normal) plane = plane.rotated((0, 0, 45)) assert plane.xDir.toTuple() == approx((k, k, 0)) assert plane.yDir.toTuple() == approx((-k, k, 0)) assert plane.zDir.toTuple() == approx((0, 0, 1)) plane = plane.rotated((0, 45, 0)) assert plane.xDir.toTuple() == approx((0.5, 0.5, -k)) assert plane.yDir.toTuple() == approx((-k, k, 0)) assert plane.zDir.toTuple() == approx((0.5, 0.5, k)) assert plane.origin.toTuple() == origin def testPlaneRotateConcatRandom(self): origin = (2, -1, 1) plane = Plane(origin=origin, xDir=(1, 0, 0), normal=(0, 0, 1)) for _ in range(100): before = { 0: plane.xDir.toTuple(), 1: plane.yDir.toTuple(), 2: plane.zDir.toTuple(), } angle = (random() - 0.5) * 720 direction = randrange(3) rotation = [0, 0, 0] rotation[direction] = angle plane = plane.rotated(rotation) after = { 0: plane.xDir.toTuple(), 1: plane.yDir.toTuple(), 2: plane.zDir.toTuple(), } assert before[direction] == approx(after[direction]) assert plane.origin.toTuple() == origin def testPlaneNoXDir(self): for z_dir in [(0, 0, 1), (1, 0, 0), (-1, 0, 0), Vector(-1, 0, 0)]: result = Plane(origin=(1, 2, 3), xDir=None, normal=z_dir) assert result.zDir == Vector(z_dir) assert result.xDir.Length == approx(1) assert result.origin == Vector(1, 2, 3) # unspecified xDir should be the same as xDir=None result2 = Plane(origin=(1, 2, 3), normal=z_dir) assert result2 == result def testPlaneToPln(self): plane = Plane(origin=(1, 2, 3), xDir=(-1, 0, 0), normal=(0, 1, 0)) gppln = plane.toPln() assert Vector(gppln.XAxis().Direction()) == Vector(-1, 0, 0) assert Vector(gppln.YAxis().Direction()) == plane.yDir assert Vector(gppln.Axis().Direction()) == plane.zDir def testRect(self): x = 10 y = 11 s = Workplane().rect(x, y) # a rectangle has 4 sides self.assertEqual(s.edges().size(), 4) # assert that the lower left corner is in the correct spot for all # possible values of centered for centered_x, xval in zip([True, False], [-x / 2, 0]): for centered_y, yval in zip([True, False], [-y / 2, 0]): s = ( Workplane() .rect(x, y, centered=(centered_x, centered_y)) .vertices("<X and <Y") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, 0), 3) # check that centered=True is the same as centered=(True, True) for option0 in [True, False]: v0 = ( Workplane() .rect(x, y, centered=option0) .vertices(">X and >Y") .val() .toTuple() ) v1 = ( Workplane() .rect(x, y, centered=(option0, option0)) .vertices(">X and >Y") .val() .toTuple() ) self.assertTupleAlmostEquals(v0, v1, 3) # test negative lengths r0 = Workplane().rect(-x, -y, centered=False) self.assertTupleAlmostEquals( (0, 0, 0), r0.vertices(">X and >Y").val().toTuple(), 3 ) self.assertTupleAlmostEquals( (-x, -y, 0), r0.vertices("<X and <Y").val().toTuple(), 3 ) # test move plus negative length r1 = Workplane().move(x, y).rect(-x, -y, centered=False) self.assertTupleAlmostEquals( (x, y, 0), r1.vertices(">X and >Y").val().toTuple(), 3 ) self.assertTupleAlmostEquals( (0, 0, 0), r1.vertices("<X and <Y").val().toTuple(), 3 ) # negative length should have no effect with centered=True v2 = Workplane().rect(x, y).vertices(">X and >Y").val().toTuple() v3 = Workplane().rect(-x, -y).vertices(">X and >Y").val().toTuple() self.assertTupleAlmostEquals(v2, v3, 3) def testLoft(self): s = Workplane("XY").circle(4.0).workplane(5.0).rect(2.0, 2.0).loft() self.saveModel(s) # the result should have 7 faces self.assertEqual(1, s.solids().size()) # the resulting loft had a split on the side, not sure why really, i expected only 3 faces self.assertEqual(7, s.faces().size()) def testLoftRaisesValueError(self): s0 = Workplane().hLine(1) # no wires with raises(ValueError): s0.loft() s1 = Workplane("XY").circle(5) # one wire with self.assertRaises(ValueError) as cm: s1.loft() err = cm.exception self.assertEqual(str(err), "More than one wire is required") def testLoftCombine(self): s = ( Workplane("front") .box(4.0, 4.0, 0.25) .faces(">Z") .circle(1.5) .workplane(offset=3.0) .rect(0.75, 0.5) .loft(combine=True) ) self.saveModel(s) # self.assertEqual(1,s.solids().size() ) # self.assertEqual(8,s.faces().size() ) def testRevolveCylinder(self): # The dimensions of the model. These can be modified rather than changing the # shape's code directly. rectangle_width = 10.0 rectangle_length = 10.0 angle_degrees = 360.0 # Test revolve without any options for making a cylinder result = ( Workplane("XY").rect(rectangle_width, rectangle_length, False).revolve() ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) # Test revolve when only setting the angle to revolve through result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, False) .revolve(angle_degrees) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, False) .revolve(270.0) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Test when passing revolve the angle and the axis of revolution's start point result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5)) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5)) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Test when passing revolve the angle and both the start and ends of the axis of revolution result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5), (-5, 5)) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5), (-5, 5)) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) # Testing all of the above without combine result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(angle_degrees, (-5, -5), (-5, 5), False) ) self.assertEqual(3, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(3, result.edges().size()) result = ( Workplane("XY") .rect(rectangle_width, rectangle_length) .revolve(270.0, (-5, -5), (-5, 5), False) ) self.assertEqual(5, result.faces().size()) self.assertEqual(6, result.vertices().size()) self.assertEqual(9, result.edges().size()) def testRevolveDonut(self): # The dimensions of the model. These can be modified rather than changing the # shape's code directly. rectangle_width = 10.0 rectangle_length = 10.0 angle_degrees = 360.0 result = ( Workplane("XY") .rect(rectangle_width, rectangle_length, True) .revolve(angle_degrees, (20, 0), (20, 10)) ) self.assertEqual(4, result.faces().size()) self.assertEqual(4, result.vertices().size()) self.assertEqual(6, result.edges().size()) def testRevolveCone(self): result = Workplane("XY").lineTo(0, 10).lineTo(5, 0).close().revolve() self.assertEqual(2, result.faces().size()) self.assertEqual(2, result.vertices().size()) self.assertEqual(2, result.edges().size()) def testRevolveErrors(self): result = Workplane("XY").lineTo(0, 10).lineTo(5, 0) with raises(ValueError): result.revolve() def testSpline(self): pts = [(0, 0), (0, 1), (1, 2), (2, 4)] # Spline path - just a smoke test path = Workplane("XZ").spline(pts).val() # Closed spline path_closed = Workplane("XZ").spline(pts, periodic=True).val() self.assertTrue(path_closed.IsClosed()) # attempt to build a valid face w = Wire.assembleEdges([path_closed,]) f = Face.makeFromWires(w) self.assertTrue(f.isValid()) # attempt to build an invalid face w = Wire.assembleEdges([path,]) f = Face.makeFromWires(w) self.assertFalse(f.isValid()) # Spline with explicit tangents path_const = Workplane("XZ").spline(pts, tangents=((0, 1), (1, 0))).val() self.assertFalse(path.tangentAt(0) == path_const.tangentAt(0)) self.assertFalse(path.tangentAt(1) == path_const.tangentAt(1)) # test include current path1 = Workplane("XZ").spline(pts[1:], includeCurrent=True).val() self.assertAlmostEqual(path.Length(), path1.Length()) # test tangents and offset plane pts = [(0, 0), (-1, 1), (-2, 0), (-1, 0)] tangents = [(0, 1), (1, 0)] path2 = Workplane("XY", (0, 0, 10)).spline(pts, tangents=tangents) self.assertAlmostEqual(path2.val().tangentAt(0).z, 0) def testSplineWithMultipleTangents(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters) .consolidateWires() ) test_point = spline.edges().val().positionAt(2.5, mode="parameter") expected_test_point = Vector(1.875, -0.625, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineWithSpecifiedAndUnspecifiedTangents(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), None, (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters) .consolidateWires() ) test_point = spline.edges().val().positionAt(1.5, mode="parameter") expected_test_point = Vector(1.6875, 0.875, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineSpecifyingParameters(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] spline1 = ( Workplane("XY") .spline(points, tangents=tangents, parameters=[0, 1, 2, 3]) .consolidateWires() ) # Multiply all parameter values by 10: spline2 = ( Workplane("XY") .spline(points, tangents=tangents, parameters=[0, 10, 20, 30]) .consolidateWires() ) # Test point equivalence for parameter, and parameter multiplied by 10: test_point1 = spline1.edges().val().positionAt(1.5, mode="parameter") test_point2 = spline2.edges().val().positionAt(15, mode="parameter") expected_test_point = Vector(1.625, 0.625, 0.0) self.assertAlmostEqual((test_point1 - test_point2).Length, 0) self.assertAlmostEqual((test_point1 - expected_test_point).Length, 0) # test periodic with parameters spline3 = Workplane().spline( points, periodic=True, parameters=[x for x in range(len(points) + 1)] ) self.assertTrue(spline3.val().IsClosed()) def testSplineWithScaleTrue(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters, scale=True) .consolidateWires() ) test_point = spline.edges().val().positionAt(0.5, mode="parameter") expected_test_point = Vector(0.375, 0.875, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineWithScaleFalse(self): points = [(0, 0), (1, 1), (2, 0), (1, -1)] tangents = [(0, 1), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) spline = ( Workplane("XY") .spline(points, tangents=tangents, parameters=parameters, scale=False) .consolidateWires() ) test_point = spline.edges().val().positionAt(0.5, mode="parameter") expected_test_point = Vector(0.375, 0.625, 0.0) self.assertAlmostEqual((test_point - expected_test_point).Length, 0) def testSplineTangentMagnitudeBelowToleranceThrows(self): import OCP points = [(0, 0), (1, 1), (2, 0), (1, -1)] # Use a tangent vector with magnitude 0.5: tangents = [(0, 0.5), (1, 0), (0, -1), (-1, 0)] parameters = range(len(points)) # Set tolerance above the 0.5 length of the tangent vector. This # should throw an exception: with raises( (OCP.Standard.Standard_ConstructionError, OCP.Standard.Standard_Failure) ): spline = ( Workplane("XY") .spline(points, tangents=tangents, tol=1) .consolidateWires() ) def testSplineInputValidation(self): points = [(0, 0), (1, 1), (2, 0)] tangents = [(0, 0.5), (1, 0), (0, -1), (-1, 0)] with raises(ValueError): spline = Workplane().spline(points, tangents=tangents) with raises(ValueError): Workplane().spline( points, periodic=False, parameters=[x for x in range(len(points) + 1)], ) with raises(ValueError): Workplane().spline( points, periodic=True, parameters=[x for x in range(len(points))], ) def testRotatedEllipse(self): def rotatePoint(x, y, alpha): # rotation matrix a = alpha * DEG2RAD r = ((math.cos(a), math.sin(a)), (-math.sin(a), math.cos(a))) return ((x * r[0][0] + y * r[1][0]), (x * r[0][1] + y * r[1][1])) def ellipsePoints(r1, r2, a): return (r1 * math.cos(a * DEG2RAD), r2 * math.sin(a * DEG2RAD)) DEG2RAD = math.pi / 180.0 p0 = (10, 20) a1, a2 = 30, -60 r1, r2 = 20, 10 ra = 25 sx_rot, sy_rot = rotatePoint(*ellipsePoints(r1, r2, a1), ra) ex_rot, ey_rot = rotatePoint(*ellipsePoints(r1, r2, a2), ra) # startAtCurrent=False, sense = 1 ellipseArc1 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) start = ellipseArc1.vertices().objects[0] end = ellipseArc1.vertices().objects[1] self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + sx_rot, p0[1] + sy_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + ex_rot, p0[1] + ey_rot), 3 ) # startAtCurrent=True, sense = 1 ellipseArc2 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=True, angle1=a1, angle2=a2, rotation_angle=ra ) ) start = ellipseArc2.vertices().objects[0] end = ellipseArc2.vertices().objects[1] self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + sx_rot - sx_rot, p0[1] + sy_rot - sy_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + ex_rot - sx_rot, p0[1] + ey_rot - sy_rot), 3 ) # startAtCurrent=False, sense = -1 ellipseArc3 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra, sense=-1, ) ) start = ellipseArc3.vertices().objects[0] end = ellipseArc3.vertices().objects[1] # swap start and end points for comparison due to different sense self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + ex_rot, p0[1] + ey_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + sx_rot, p0[1] + sy_rot), 3 ) # startAtCurrent=True, sense = -1 ellipseArc4 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=True, angle1=a1, angle2=a2, rotation_angle=ra, sense=-1, makeWire=True, ) ) self.assertEqual(len(ellipseArc4.ctx.pendingWires), 1) start = ellipseArc4.vertices().objects[0] end = ellipseArc4.vertices().objects[1] # swap start and end points for comparison due to different sense self.assertTupleAlmostEquals( (start.X, start.Y), (p0[0] + ex_rot - ex_rot, p0[1] + ey_rot - ey_rot), 3 ) self.assertTupleAlmostEquals( (end.X, end.Y), (p0[0] + sx_rot - ex_rot, p0[1] + sy_rot - ey_rot), 3 ) def testEllipseArcsClockwise(self): ellipseArc = ( Workplane("XY") .moveTo(10, 15) .ellipseArc(5, 4, -10, 190, 45, sense=-1, startAtCurrent=False) ) sp = ellipseArc.val().startPoint() ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals( (sp.x, sp.y), (7.009330014275797, 11.027027582524015), 3 ) self.assertTupleAlmostEquals( (ep.x, ep.y), (13.972972417475985, 17.990669985724203), 3 ) ellipseArc = ( ellipseArc.ellipseArc(5, 4, -10, 190, 315, sense=-1) .ellipseArc(5, 4, -10, 190, 225, sense=-1) .ellipseArc(5, 4, -10, 190, 135, sense=-1) ) ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals((sp.x, sp.y), (ep.x, ep.y), 3) def testEllipseArcsCounterClockwise(self): ellipseArc = ( Workplane("XY") .moveTo(10, 15) .ellipseArc(5, 4, -10, 190, 45, startAtCurrent=False) ) sp = ellipseArc.val().startPoint() ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals( (sp.x, sp.y), (13.972972417475985, 17.990669985724203), 3 ) self.assertTupleAlmostEquals( (ep.x, ep.y), (7.009330014275797, 11.027027582524015), 3 ) ellipseArc = ( ellipseArc.ellipseArc(5, 4, -10, 190, 135) .ellipseArc(5, 4, -10, 190, 225) .ellipseArc(5, 4, -10, 190, 315) ) ep = ellipseArc.val().endPoint() self.assertTupleAlmostEquals((sp.x, sp.y), (ep.x, ep.y), 3) def testEllipseCenterAndMoveTo(self): # Whether we start from a center() call or a moveTo call, it should be the same ellipse Arc p0 = (10, 20) a1, a2 = 30, -60 r1, r2 = 20, 10 ra = 25 ellipseArc1 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) sp1 = ellipseArc1.val().startPoint() ep1 = ellipseArc1.val().endPoint() ellipseArc2 = ( Workplane("XY") .moveTo(*p0) .ellipseArc( r1, r2, startAtCurrent=False, angle1=a1, angle2=a2, rotation_angle=ra ) ) sp2 = ellipseArc2.val().startPoint() ep2 = ellipseArc2.val().endPoint() self.assertTupleAlmostEquals(sp1.toTuple(), sp2.toTuple(), 3) self.assertTupleAlmostEquals(ep1.toTuple(), ep2.toTuple(), 3) def testMakeEllipse(self): el = Wire.makeEllipse( 1, 2, Vector(0, 0, 0), Vector(0, 0, 1), Vector(1, 0, 0), 0, 90, 45, True, ) self.assertTrue(el.IsClosed()) self.assertTrue(el.isValid()) def testSweep(self): pts = [(0, 0), (0, 1), (1, 2), (2, 4)] # Spline path path = Workplane("XZ").spline(pts) # Test defaults result = Workplane("XY").circle(1.0).sweep(path) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with makeSolid False result = Workplane("XY").circle(1.0).sweep(path, makeSolid=False) self.assertEqual(1, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with isFrenet True result = Workplane("XY").circle(1.0).sweep(path, isFrenet=True) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test with makeSolid False and isFrenet True result = Workplane("XY").circle(1.0).sweep(path, makeSolid=False, isFrenet=True) self.assertEqual(1, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test rectangle with defaults result = Workplane("XY").rect(1.0, 1.0).sweep(path) self.assertEqual(6, result.faces().size()) self.assertEqual(12, result.edges().size()) # Test fixed normal result = Workplane().circle(0.5).sweep(path, normal=Vector(0, 0, 1)) self.assertTupleAlmostEquals( result.faces(">Z").val().normalAt().toTuple(), (0, 0, 1), 6 ) # Polyline path path = Workplane("XZ").polyline(pts) # Test defaults result = Workplane("XY").circle(0.1).sweep(path, transition="transformed") self.assertEqual(5, result.faces().size()) self.assertEqual(7, result.edges().size()) # Polyline path and one inner profiles path = Workplane("XZ").polyline(pts) # Test defaults result = ( Workplane("XY") .circle(0.2) .circle(0.1) .sweep(path, transition="transformed") ) self.assertEqual(8, result.faces().size()) self.assertEqual(14, result.edges().size()) # Polyline path and different transition settings for t in ("transformed", "right", "round"): path = Workplane("XZ").polyline(pts) result = ( Workplane("XY") .circle(0.2) .rect(0.2, 0.1) .rect(0.1, 0.2) .sweep(path, transition=t) ) self.assertTrue(result.solids().val().isValid()) # Polyline path and multiple inner profiles path = Workplane("XZ").polyline(pts) # Test defaults result = ( Workplane("XY") .circle(0.2) .rect(0.2, 0.1) .rect(0.1, 0.2) .circle(0.1) .sweep(path) ) self.assertTrue(result.solids().val().isValid()) # Arc path path = Workplane("XZ").threePointArc((1.0, 1.5), (0.0, 1.0)) # Test defaults result = Workplane("XY").circle(0.1).sweep(path) self.assertEqual(3, result.faces().size()) self.assertEqual(3, result.edges().size()) # Test aux spine pts1 = [(0, 0), (20, 100)] pts2 = [(0, 20, 0), (20, 0, 100)] path = Workplane("YZ").spline(pts1, tangents=[(0, 1, 0), (0, 1, 0)]) aux_path = Workplane("XY").spline(pts2, tangents=[(0, 0, 1), (0, 0, 1)]) result = Workplane("XY").rect(10, 20).sweep(path, auxSpine=aux_path) bottom = result.faces("<Z") top = result.faces(">Z") v1 = bottom.wires().val().tangentAt(0.0) v2 = top.wires().val().tangentAt(0.0) self.assertAlmostEqual(v1.getAngle(v2), math.pi / 4, 6) # test for ValueError if pending wires is empty w0 = Workplane().hLine(1).vLine(1) with raises(ValueError): w0.sweep(path) # Test aux spine invalid input handling with raises(ValueError): result = ( Workplane("XY") .rect(10, 20) .sweep(path, auxSpine=Workplane().box(1, 1, 1)) ) def testMultisectionSweep(self): # X axis line length 20.0 path = Workplane("XZ").moveTo(-10, 0).lineTo(10, 0) # Sweep a circle from diameter 2.0 to diameter 1.0 to diameter 2.0 along X axis length 10.0 + 10.0 defaultSweep = ( Workplane("YZ") .workplane(offset=-10.0) .circle(2.0) .workplane(offset=10.0) .circle(1.0) .workplane(offset=10.0) .circle(2.0) .sweep(path, multisection=True) ) # We can sweep through different shapes recttocircleSweep = ( Workplane("YZ") .workplane(offset=-10.0) .rect(2.0, 2.0) .workplane(offset=8.0) .circle(1.0) .workplane(offset=4.0) .circle(1.0) .workplane(offset=8.0) .rect(2.0, 2.0) .sweep(path, multisection=True) ) circletorectSweep = ( Workplane("YZ") .workplane(offset=-10.0) .circle(1.0) .workplane(offset=7.0) .rect(2.0, 2.0) .workplane(offset=6.0) .rect(2.0, 2.0) .workplane(offset=7.0) .circle(1.0) .sweep(path, multisection=True) ) # Placement of the Shape is important otherwise could produce unexpected shape specialSweep = ( Workplane("YZ") .circle(1.0) .workplane(offset=10.0) .rect(2.0, 2.0) .sweep(path, multisection=True) ) # Switch to an arc for the path : line l=5.0 then half circle r=4.0 then line l=5.0 path = ( Workplane("XZ") .moveTo(-5, 4) .lineTo(0, 4) .threePointArc((4, 0), (0, -4)) .lineTo(-5, -4) ) # Placement of different shapes should follow the path # cylinder r=1.5 along first line # then sweep along arc from r=1.5 to r=1.0 # then cylinder r=1.0 along last line arcSweep = ( Workplane("YZ") .workplane(offset=-5) .moveTo(0, 4) .circle(1.5) .workplane(offset=5, centerOption="CenterOfMass") .circle(1.5) .moveTo(0, -8) .circle(1.0) .workplane(offset=-5, centerOption="CenterOfMass") .circle(1.0) .sweep(path, multisection=True) ) # Test multisection with normal pts = [(0, 0), (20, 100)] path = Workplane("YZ").spline(pts, tangents=[(0, 1, 0), (0, 1, 0)]) normalSweep = ( Workplane() .rect(10, 10) .workplane(offset=100) .rect(10, 20) .sweep(path, multisection=True, normal=(0, 1, 1)) ) self.assertTupleAlmostEquals( normalSweep.faces("<Z").val().normalAt().toTuple(), Vector(0, -1, -1).normalized().toTuple(), 6, ) self.assertTupleAlmostEquals( normalSweep.faces(">Z").val().normalAt().toTuple(), Vector(0, 1, 1).normalized().toTuple(), 6, ) # Test and saveModel self.assertEqual(1, defaultSweep.solids().size()) self.assertEqual(1, circletorectSweep.solids().size()) self.assertEqual(1, recttocircleSweep.solids().size()) self.assertEqual(1, specialSweep.solids().size()) self.assertEqual(1, arcSweep.solids().size()) self.saveModel(defaultSweep) def testTwistExtrude(self): profile = Workplane("XY").rect(10, 10) r = profile.twistExtrude(10, 45, False) self.assertEqual(6, r.faces().size()) def testTwistExtrudeCombine(self): profile = Workplane("XY").rect(10, 10) r = profile.twistExtrude(10, 45) self.assertEqual(6, r.faces().size()) def testRectArray(self): x_num = 3 y_num = 3 x_spacing = 8.0 y_spacing = 8.0 s = ( Workplane("XY") .box(40, 40, 5, centered=(True, True, True)) .faces(">Z") .workplane() .rarray(x_spacing, y_spacing, x_num, y_num, True) .circle(2.0) .extrude(2.0) ) self.saveModel(s) # 6 faces for the box, 2 faces for each cylinder self.assertEqual(6 + x_num * y_num * 2, s.faces().size()) with raises(ValueError): Workplane().rarray(0, 0, x_num, y_num, True) # check lower and upper corner points are correct for all combinations of centering for x_opt, x_min, x_max in zip( [True, False], [-x_spacing, 0.0], [x_spacing, x_spacing * 2] ): for y_opt, y_min, y_max in zip( [True, False], [-y_spacing, 0.0], [y_spacing, y_spacing * 2] ): s = Workplane().rarray( x_spacing, y_spacing, x_num, y_num, center=(x_opt, y_opt) ) lower = Vector(x_min, y_min, 0) upper = Vector(x_max, y_max, 0) self.assertTrue(lower in s.objects) self.assertTrue(upper in s.objects) # check centered=True is equivalent to centered=(True, True) for val in [True, False]: s0 = Workplane().rarray(x_spacing, y_spacing, x_num, y_num, center=val) s1 = Workplane().rarray( x_spacing, y_spacing, x_num, y_num, center=(val, val) ) # check all the points in s0 are present in s1 self.assertTrue(all(pnt in s1.objects for pnt in s0.objects)) self.assertEqual(s0.size(), s1.size()) def testPolarArray(self): radius = 10 # Test for proper number of elements s = Workplane("XY").polarArray(radius, 0, 180, 1) self.assertEqual(1, s.size()) s = Workplane("XY").polarArray(radius, 0, 180, 6) self.assertEqual(6, s.size()) to_x = lambda l: l.wrapped.Transformation().TranslationPart().X() to_y = lambda l: l.wrapped.Transformation().TranslationPart().Y() to_angle = ( lambda l: l.wrapped.Transformation().GetRotation().GetRotationAngle() * 180.0 / math.pi ) # Test for proper placement when fill == True s = Workplane("XY").polarArray(radius, 0, 180, 3) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper placement when angle to fill is multiple of 360 deg s = Workplane("XY").polarArray(radius, 0, 360, 4) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper placement when fill == False s = Workplane("XY").polarArray(radius, 0, 90, 3, fill=False) self.assertAlmostEqual(0, to_y(s.objects[1])) self.assertAlmostEqual(radius, to_x(s.objects[1])) # Test for proper operation of startAngle s = Workplane("XY").polarArray(radius, 90, 180, 3) self.assertAlmostEqual(radius, to_x(s.objects[0])) self.assertAlmostEqual(0, to_y(s.objects[0])) # Test for local rotation s = Workplane().polarArray(radius, 0, 180, 3) self.assertAlmostEqual(0, to_angle(s.objects[0])) self.assertAlmostEqual(90, to_angle(s.objects[1])) s = Workplane().polarArray(radius, 0, 180, 3, rotate=False) self.assertAlmostEqual(0, to_angle(s.objects[0])) self.assertAlmostEqual(0, to_angle(s.objects[1])) def testNestedCircle(self): s = ( Workplane("XY") .box(40, 40, 5) .pushPoints([(10, 0), (0, 10)]) .circle(4) .circle(2) .extrude(4) ) self.saveModel(s) self.assertEqual(14, s.faces().size()) def testConcentricEllipses(self): concentricEllipses = ( Workplane("XY").center(10, 20).ellipse(100, 10).center(0, 0).ellipse(50, 5) ) v = concentricEllipses.vertices().objects[0] self.assertTupleAlmostEquals((v.X, v.Y), (10 + 50, 20), 3) def testLegoBrick(self): # test making a simple lego brick # which of the below # inputs lbumps = 8 wbumps = 2 # lego brick constants P = 8.0 # nominal pitch c = 0.1 # clearance on each brick side H = 1.2 * P # nominal height of a brick bumpDiam = 4.8 # the standard bump diameter # the nominal thickness of the walls, normally 1.5 t = (P - (2 * c) - bumpDiam) / 2.0 postDiam = P - t # works out to 6.5 total_length = lbumps * P - 2.0 * c total_width = wbumps * P - 2.0 * c # build the brick s = Workplane("XY").box(total_length, total_width, H) # make the base s = s.faces("<Z").shell(-1.0 * t) # shell inwards not outwards s = ( s.faces(">Z") .workplane() .rarray(P, P, lbumps, wbumps, True) .circle(bumpDiam / 2.0) .extrude(1.8) ) # make the bumps on the top # add posts on the bottom. posts are different diameter depending on geometry # solid studs for 1 bump, tubes for multiple, none for 1x1 # this is cheating a little-- how to select the inner face from the shell? tmp = s.faces("<Z").workplane(invert=True) if lbumps > 1 and wbumps > 1: tmp = ( tmp.rarray(P, P, lbumps - 1, wbumps - 1, center=True) .circle(postDiam / 2.0) .circle(bumpDiam / 2.0) .extrude(H - t) ) elif lbumps > 1: tmp = tmp.rarray(P, P, lbumps - 1, 1, center=True).circle(t).extrude(H - t) elif wbumps > 1: tmp = tmp.rarray(P, P, 1, wbumps - 1, center=True).circle(t).extrude(H - t) self.saveModel(s) def testAngledHoles(self): s = ( Workplane("front") .box(4.0, 4.0, 0.25) .faces(">Z") .workplane() .transformed(offset=Vector(0, -1.5, 1.0), rotate=Vector(60, 0, 0)) .rect(1.5, 1.5, forConstruction=True) .vertices() .hole(0.25) ) self.saveModel(s) self.assertEqual(10, s.faces().size()) def testTranslateSolid(self): c = CQ(makeUnitCube()) self.assertAlmostEqual(0.0, c.faces("<Z").vertices().item(0).val().Z, 3) # TODO: it might be nice to provide a version of translate that modifies the existing geometry too d = c.translate(Vector(0, 0, 1.5)) self.assertAlmostEqual(1.5, d.faces("<Z").vertices().item(0).val().Z, 3) def testTranslateWire(self): c = CQ(makeUnitSquareWire()) self.assertAlmostEqual(0.0, c.edges().vertices().item(0).val().Z, 3) d = c.translate(Vector(0, 0, 1.5)) self.assertAlmostEqual(1.5, d.edges().vertices().item(0).val().Z, 3) def testSolidReferencesCombine(self): c = CQ(makeUnitCube()) # the cube is the context solid self.assertEqual(6, c.faces().size()) # cube has six faces r = ( c.faces(">Z").workplane().circle(0.125).extrude(0.5, True) ) # make a boss, not updating the original self.assertEqual(8, r.faces().size()) # just the boss faces self.assertEqual(6, c.faces().size()) # original is not modified def testSolidReferencesCombineTrue(self): s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).extrude(0.5) # the result of course has 6 faces self.assertEqual(6, r.faces().size()) # the original workplane does not, because it did not have a solid initially self.assertEqual(0, s.faces().size()) t = r.faces(">Z").workplane().rect(0.25, 0.25).extrude(0.5, True) # of course the result has 11 faces self.assertEqual(11, t.faces().size()) # r (being the parent) remains unmodified self.assertEqual(6, r.faces().size()) self.saveModel(r) def testSolidReferenceCombineFalse(self): s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).extrude(0.5) # the result of course has 6 faces self.assertEqual(6, r.faces().size()) # the original workplane does not, because it did not have a solid initially self.assertEqual(0, s.faces().size()) t = r.faces(">Z").workplane().rect(0.25, 0.25).extrude(0.5, False) # result has 6 faces, because it was not combined with the original self.assertEqual(6, t.faces().size()) self.assertEqual(6, r.faces().size()) # original is unmodified as well # subsequent operations use that context solid afterwards def testSimpleWorkplane(self): s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .extrude(0.5) .faces(">Z") .workplane() .circle(0.25) .cutBlind(-1.0) ) self.saveModel(r) self.assertEqual(7, r.faces().size()) def testMultiFaceWorkplane(self): s = Workplane("XY").box(1, 1, 1).faces(">Z").rect(1, 0.5).cutBlind(-0.2) w = s.faces(">Z").workplane() o = w.val() # origin of the workplane self.assertAlmostEqual(o.x, 0.0, 3) self.assertAlmostEqual(o.y, 0.0, 3) self.assertAlmostEqual(o.z, 0.5, 3) def testTriangularPrism(self): s = Workplane("XY").lineTo(1, 0).lineTo(1, 1).close().extrude(0.2) self.saveModel(s) def testMultiWireWorkplane(self): s = Workplane(Plane.XY()) r = s.rect(2.0, 2.0).circle(0.25).extrude(0.5) self.saveModel(r) self.assertEqual(7, r.faces().size()) def testConstructionWire(self): s = Workplane(Plane.YZ()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) self.saveModel(r) # 10 faces-- 6 plus 4 holes, the vertices of the second rect. self.assertEqual(10, r.faces().size()) def testTwoWorkplanes(self): # base block s = Workplane(Plane.XY()) # TODO: this syntax is nice, but the iteration might not be worth # the complexity. # the simpler and slightly longer version would be: # r = s.rect(2.0,2.0).rect(1.3,1.3,forConstruction=True).vertices() # for c in r.all(): # c.circle(0.125).extrude(0.5,True) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # side hole, blind deep 1.9 t = r.faces(">Y").workplane().circle(0.125).cutBlind(-1.9) self.saveModel(t) self.assertEqual(12, t.faces().size()) def testCut(self): s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toCut = s.rect(1.0, 1.0).extrude(0.5) resS = currentS.cut(toCut.val()) self.assertEqual(10, resS.faces().size()) with self.assertRaises(ValueError): currentS.cut(toCut.faces().val()) # Test syntactic sugar [__sub__ method] sugar = currentS - toCut.val() self.assertEqual(resS.faces().size(), sugar.faces().size()) # test ValueError on no solid found s0 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): s0.cut(toCut.val()) def testIntersect(self): s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toIntersect = s.rect(1.0, 1.0).extrude(1) resS = currentS.intersect(toIntersect.val()) self.assertEqual(6, resS.faces().size()) self.assertAlmostEqual(resS.val().Volume(), 0.5) resS = currentS.intersect(toIntersect) self.assertEqual(6, resS.faces().size()) self.assertAlmostEqual(resS.val().Volume(), 0.5) b1 = Workplane("XY").box(1, 1, 1) b2 = Workplane("XY", origin=(0, 0, 0.5)).box(1, 1, 1) resS = b1.intersect(b2) self.assertAlmostEqual(resS.val().Volume(), 0.5) with self.assertRaises(ValueError): b1.intersect(b2.faces().val()) # Test syntactic sugar [__mul__ method] sugar = b1 & b2 self.assertEqual(resS.val().Volume(), sugar.val().Volume()) # raise ValueError when no solid found with raises(ValueError): Workplane().intersect(toIntersect) def testBoundingBox(self): result0 = ( Workplane("XY") .moveTo(10, 0) .lineTo(5, 0) .threePointArc((3.9393, 0.4393), (3.5, 1.5)) .threePointArc((3.0607, 2.5607), (2, 3)) .lineTo(1.5, 3) .threePointArc((0.4393, 3.4393), (0, 4.5)) .lineTo(0, 13.5) .threePointArc((0.4393, 14.5607), (1.5, 15)) .lineTo(28, 15) .lineTo(28, 13.5) .lineTo(24, 13.5) .lineTo(24, 11.5) .lineTo(27, 11.5) .lineTo(27, 10) .lineTo(22, 10) .lineTo(22, 13.2) .lineTo(14.5, 13.2) .lineTo(14.5, 10) .lineTo(12.5, 10) .lineTo(12.5, 13.2) .lineTo(5.5, 13.2) .lineTo(5.5, 2) .threePointArc((5.793, 1.293), (6.5, 1)) .lineTo(10, 1) .close() ) result = result0.extrude(100) bb_center = result.val().BoundingBox().center self.saveModel(result) self.assertAlmostEqual(14.0, bb_center.x, 3) self.assertAlmostEqual(7.5, bb_center.y, 3) self.assertAlmostEqual(50.0, bb_center.z, 3) # The following will raise with the default tolerance of TOL 1e-2 bb = result.val().BoundingBox(tolerance=1e-3) self.assertAlmostEqual(0.0, bb.xmin, 2) self.assertAlmostEqual(28, bb.xmax, 2) self.assertAlmostEqual(0.0, bb.ymin, 2) self.assertAlmostEqual(15.0, bb.ymax, 2) self.assertAlmostEqual(0.0, bb.zmin, 2) self.assertAlmostEqual(100.0, bb.zmax, 2) def testCutThroughAll(self): # base block s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # thru all without explicit face selection t = r.circle(0.5).cutThruAll() self.assertEqual(11, t.faces().size()) # side hole, thru all t = ( t.faces(">Y") .workplane(centerOption="CenterOfMass") .circle(0.125) .cutThruAll() ) self.saveModel(t) self.assertEqual(13, t.faces().size()) # no planar faces sphere_r = 10.0 r = ( Workplane() .sphere(sphere_r) .workplane() .circle(sphere_r / 2.0) .cutThruAll() .workplane() .transformed(rotate=(90, 0, 0)) .circle(sphere_r / 2.0) .cutThruAll() .workplane() .transformed(rotate=(0, 90, 0)) .circle(sphere_r / 2.0) .cutThruAll() ) self.assertTrue(r.val().isValid()) self.assertEqual(r.faces().size(), 7) # test errors box0 = Workplane().box(1, 1, 1).faces(">Z").workplane().hLine(1) with raises(ValueError): box0.cutThruAll() no_box = Workplane().hLine(1).vLine(1).close() with raises(ValueError): no_box.cutThruAll() def testCutToFaceOffsetNOTIMPLEMENTEDYET(self): # base block s = Workplane(Plane.XY()) r = ( s.rect(2.0, 2.0) .rect(1.3, 1.3, forConstruction=True) .vertices() .circle(0.125) .extrude(0.5) ) # side hole, up to 0.1 from the last face try: t = ( r.faces(">Y") .workplane() .circle(0.125) .cutToOffsetFromFace(r.faces().mminDist(Dir.Y), 0.1) ) # should end up being a blind hole self.assertEqual(10, t.faces().size()) t.first().val().exportStep("c:/temp/testCutToFace.STEP") except: pass # Not Implemented Yet def testWorkplaneOnExistingSolid(self): c = ( CQ(makeUnitCube()) .faces(">Z") .workplane() .circle(0.25) .circle(0.125) .extrude(0.25) ) self.saveModel(c) self.assertEqual(10, c.faces().size()) def testWorkplaneCenterMove(self): # this workplane is centered at x=0.5,y=0.5, the center of the upper face s = ( Workplane("XY").box(1, 1, 1).faces(">Z").workplane().center(-0.5, -0.5) ) # move the center to the corner t = s.circle(0.25).extrude(0.2) # make a boss self.assertEqual(9, t.faces().size()) self.saveModel(t) def testBasicLines(self): global OUTDIR s = Workplane(Plane.XY()) # TODO: extrude() should imply wire() if not done already # most users dont understand what a wire is, they are just drawing r = s.lineTo(1.0, 0).lineTo(0, 1.0).close().wire().extrude(0.25) r.val().exportStep(os.path.join(OUTDIR, "testBasicLinesStep1.STEP")) # no faces on the original workplane self.assertEqual(0, s.faces().size()) # 5 faces on newly created object self.assertEqual(5, r.faces().size()) # now add a circle through a side face r1 = ( r.faces("+XY") .workplane(centerOption="CenterOfMass") .circle(0.08) .cutThruAll() ) self.assertEqual(6, r1.faces().size()) r1.val().exportStep(os.path.join(OUTDIR, "testBasicLinesXY.STEP")) # now add a circle through a top r2 = ( r1.faces("+Z") .workplane(centerOption="CenterOfMass") .circle(0.08) .cutThruAll() ) self.assertEqual(9, r2.faces().size()) r2.val().exportStep(os.path.join(OUTDIR, "testBasicLinesZ.STEP")) self.saveModel(r2) def test2DDrawing(self): s = Workplane(Plane.XY()) r = ( s.lineTo(1.0, 0.0) .lineTo(1.0, 1.0) .threePointArc((1.0, 1.5), (0.0, 1.0)) .lineTo(0.0, 0.0) .moveTo(1.0, 0.0) .lineTo(2.0, 0.0) .lineTo(2.0, 2.0) .threePointArc((2.0, 2.5), (0.0, 2.0)) .lineTo(-2.0, 2.0) .lineTo(-2.0, 0.0) .close() ) self.assertEqual(1, r.wires().size()) # Test the *LineTo functions s = Workplane(Plane.XY()) r = s.hLineTo(1.0).vLineTo(1.0).hLineTo(0.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) # Test the *Line functions s = Workplane(Plane.XY()) r = s.hLine(1.0).vLine(1.0).hLine(-1.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) # Test the move function s = Workplane(Plane.XY()) r = s.move(1.0, 1.0).hLine(1.0).vLine(1.0).hLine(-1.0).close() self.assertEqual(1, r.wire().size()) self.assertEqual(4, r.edges().size()) self.assertEqual( (1.0, 1.0), ( r.vertices(selectors.NearestToPointSelector((0.0, 0.0, 0.0))) .first() .val() .X, r.vertices(selectors.NearestToPointSelector((0.0, 0.0, 0.0))) .first() .val() .Y, ), ) # Test the sagittaArc and radiusArc functions a1 = Workplane(Plane.YZ()).threePointArc((5, 1), (10, 0)) a2 = Workplane(Plane.YZ()).sagittaArc((10, 0), -1) a3 = Workplane(Plane.YZ()).threePointArc((6, 2), (12, 0)) a4 = Workplane(Plane.YZ()).radiusArc((12, 0), -10) assert a1.edges().first().val().geomType() == "CIRCLE" assert a2.edges().first().val().geomType() == "CIRCLE" assert a3.edges().first().val().geomType() == "CIRCLE" assert a4.edges().first().val().geomType() == "CIRCLE" assert a1.edges().first().val().Length() == a2.edges().first().val().Length() assert a3.edges().first().val().Length() == a4.edges().first().val().Length() def testPolarLines(self): # Test the PolarLine* functions s = Workplane(Plane.XY()) r = ( s.polarLine(10, 45) .polarLineTo(10, -45) .polarLine(10, -180) .polarLine(-10, -90) .close() ) # a single wire, 5 edges self.assertEqual(1, r.wires().size()) self.assertEqual(5, r.wires().edges().size()) def testLargestDimension(self): r = Workplane("XY").box(1, 1, 1) dim = r.largestDimension() self.assertAlmostEqual(1.76, dim, 1) r = Workplane("XY").rect(1, 1).extrude(1) dim = r.largestDimension() self.assertAlmostEqual(1.76, dim, 1) r = Workplane("XY") with raises(ValueError): r.largestDimension() def testOccBottle(self): L = 20.0 w = 6.0 t = 3.0 s = Workplane(Plane.XY()) # draw half the profile of the bottle p = ( s.center(-L / 2.0, 0) .vLine(w / 2.0) .threePointArc((L / 2.0, w / 2.0 + t), (L, w / 2.0)) .vLine(-w / 2.0) .mirrorX() .extrude(30.0, True) ) # make the neck p.faces(">Z").workplane().circle(3.0).extrude( 2.0, True ) # .edges().fillet(0.05) # make a shell p.faces(">Z").shell(0.3) self.saveModel(p) def testSplineShape(self): s = Workplane(Plane.XY()) sPnts = [ (2.75, 1.5), (2.5, 1.75), (2.0, 1.5), (1.5, 1.0), (1.0, 1.25), (0.5, 1.0), (0, 1.0), ] r = s.lineTo(3.0, 0).lineTo(3.0, 1.0).spline(sPnts).close() r = r.extrude(0.5) self.saveModel(r) def testSimpleMirror(self): s = ( Workplane("XY") .lineTo(2, 2) .threePointArc((3, 1), (2, 0)) .mirrorX() .extrude(0.25) ) self.assertEqual(6, s.faces().size()) self.saveModel(s) def testUnorderedMirror(self): r = 20 s = 7 t = 1.5 points = [ (0, 0), (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] r = Workplane("XY").polyline(points).mirrorX() self.assertEqual(1, r.wires().size()) self.assertEqual(18, r.edges().size()) # try the same with includeCurrent=True r = Workplane("XY").polyline(points[1:], includeCurrent=True).mirrorX() self.assertEqual(1, r.wires().size()) self.assertEqual(18, r.edges().size()) def testChainedMirror(self): r = 20 s = 7 t = 1.5 points = [ (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] r = Workplane("XY").polyline(points).mirrorX().mirrorY().extrude(1).faces(">Z") self.assertEqual(1, r.wires().size()) self.assertEqual(32, r.edges().size()) # TODO: Re-work testIbeam test below now that chaining works # TODO: Add toLocalCoords and toWorldCoords tests def testIbeam(self): s = Workplane(Plane.XY()) L = 100.0 H = 20.0 W = 20.0 t = 1.0 # TODO: for some reason doing 1/4 of the profile and mirroring twice ( .mirrorX().mirrorY() ) # did not work, due to a bug in freecad-- it was losing edges when creating a composite wire. # i just side-stepped it for now pts = [ (0, 0), (0, H / 2.0), (W / 2.0, H / 2.0), (W / 2.0, (H / 2.0 - t)), (t / 2.0, (H / 2.0 - t)), (t / 2.0, (t - H / 2.0)), (W / 2.0, (t - H / 2.0)), (W / 2.0, H / -2.0), (0, H / -2.0), ] r = s.polyline(pts).mirrorY() # these other forms also work res = r.extrude(L) self.saveModel(res) def testCone(self): s = Solid.makeCone(0, 1.0, 2.0) t = CQ(s) self.saveModel(t) self.assertEqual(2, t.faces().size()) def testFillet(self): c = ( CQ(makeUnitCube()) .faces(">Z") .workplane() .circle(0.25) .extrude(0.25, True) .edges("|Z") .fillet(0.2) ) self.saveModel(c) self.assertEqual(12, c.faces().size()) # should raise an error if no solid c1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): c1.fillet(0.1) def testChamfer(self): cube = CQ(makeUnitCube()).faces(">Z").chamfer(0.1) self.saveModel(cube) self.assertEqual(10, cube.faces().size()) # should raise an error if no solid c1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): c1.chamfer(0.1) def testChamferAsymmetrical(self): cube = CQ(makeUnitCube()).faces(">Z").chamfer(0.1, 0.2) self.saveModel(cube) self.assertEqual(10, cube.faces().size()) # test if edge lengths are different edge = cube.edges(">Z").vals()[0] self.assertAlmostEqual(0.6, edge.Length(), 3) edge = cube.edges("|Z").vals()[0] self.assertAlmostEqual(0.9, edge.Length(), 3) def testChamferCylinder(self): cylinder = Workplane("XY").circle(1).extrude(1).faces(">Z").chamfer(0.1) self.saveModel(cylinder) self.assertEqual(4, cylinder.faces().size()) def testCounterBores(self): c = CQ(makeCube(3.0)) pnts = [(-1.0, -1.0), (0.0, 0.0), (1.0, 1.0)] c = c.faces(">Z").workplane().pushPoints(pnts).cboreHole(0.1, 0.25, 0.25, 0.75) self.assertEqual(18, c.faces().size()) self.saveModel(c) # Tests the case where the depth of the cboreHole is not specified c2 = CQ(makeCube(3.0)) c2 = c2.faces(">Z").workplane().pushPoints(pnts).cboreHole(0.1, 0.25, 0.25) self.assertEqual(15, c2.faces().size()) def testCounterSinks(self): s = Workplane(Plane.XY()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) def testSplitKeepingHalf(self): # drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=True) self.saveModel(result) self.assertEqual(8, result.faces().size()) def testSplitKeepingBoth(self): # drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=True, keepBottom=True) # stack will have both halves, original will be unchanged # two solids are on the stack, eac self.assertEqual(2, result.solids().size()) self.assertEqual(8, result.solids().item(0).faces().size()) self.assertEqual(8, result.solids().item(1).faces().size()) def testSplitKeepingBottom(self): # Drill a hole in the side c = CQ(makeUnitCube()).faces(">Z").workplane().circle(0.25).cutThruAll() self.assertEqual(7, c.faces().size()) # Now cut it in half sideways result = c.faces(">Y").workplane(-0.5).split(keepTop=False, keepBottom=True) # stack will have both halves, original will be unchanged # one solid is on the stack self.assertEqual(1, result.solids().size()) self.assertEqual(8, result.solids().item(0).faces().size()) def testSplitError(self): # Test split produces the correct error when called with no solid to split. w = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w.split(keepTop=True) # Split should raise ValueError when called with no side kept with raises(ValueError): w.split(keepTop=False, keepBottom=False) def testBoxDefaults(self): s = Workplane("XY").box(2, 3, 4) self.assertEqual(1, s.solids().size()) self.saveModel(s) def testSimpleShell(self): s1 = Workplane("XY").box(2, 2, 2).faces("+Z").shell(0.05) self.saveModel(s1) self.assertEqual(23, s1.faces().size()) s2 = ( Workplane() .ellipse(4, 2) .extrude(4) .faces(">Z") .shell(+2, kind="intersection") ) self.assertEqual(5, s2.faces().size()) s3 = Workplane().ellipse(4, 2).extrude(4).faces(">Z").shell(+2, kind="arc") self.assertEqual(6, s3.faces().size()) # test error on no solid found s4 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): s4.shell(1) def testClosedShell(self): s1 = Workplane("XY").box(2, 2, 2).shell(-0.1) self.assertEqual(12, s1.faces().size()) self.assertTrue(s1.val().isValid()) s2 = Workplane("XY").box(2, 2, 2).shell(0.1) self.assertEqual(32, s2.faces().size()) self.assertTrue(s2.val().isValid()) pts = [(1.0, 0.0), (0.3, 0.2), (0.0, 0.0), (0.3, -0.1), (1.0, -0.03)] s3 = Workplane().polyline(pts).close().extrude(1).shell(-0.05) self.assertTrue(s3.val().isValid()) s4_shape = Workplane("XY").box(2, 2, 2).val() # test that None and empty list both work and are equivalent s4_shell_1 = s4_shape.shell(faceList=None, thickness=-0.1) s4_shell_2 = s4_shape.shell(faceList=[], thickness=-0.1) # this should be the same as the first shape self.assertEqual(len(s4_shell_1.Faces()), s1.faces().size()) self.assertEqual(len(s4_shell_2.Faces()), s1.faces().size()) def testOpenCornerShell(self): s = Workplane("XY").box(1, 1, 1) s1 = s.faces("+Z") s1.add(s.faces("+Y")).add(s.faces("+X")) self.saveModel(s1.shell(0.2)) # Tests the list option variation of add s1 = s.faces("+Z") s1.add(s.faces("+Y")).add([s.faces("+X")]) # Tests the raw object option variation of add s1 = s.faces("+Z") s1.add(s.faces("+Y")).add(s.faces("+X").val().wrapped) def testTopFaceFillet(self): s = Workplane("XY").box(1, 1, 1).faces("+Z").edges().fillet(0.1) self.assertEqual(s.faces().size(), 10) self.saveModel(s) def testBoxPointList(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=True) ) # 1 object, 4 solids because the object is a compound self.assertEqual(4, s.solids().size()) self.assertEqual(1, s.size()) self.saveModel(s) s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=False) ) # 4 objects, 4 solids, because each is a separate solid self.assertEqual(4, s.size()) self.assertEqual(4, s.solids().size()) def testBoxCombine(self): s = ( Workplane("XY") .box(4, 4, 0.5) .faces(">Z") .workplane() .rect(3, 3, forConstruction=True) .vertices() .box(0.25, 0.25, 0.25, combine=True) ) self.saveModel(s) self.assertEqual(1, s.solids().size()) # we should have one big solid # should have 26 faces. 6 for the box, and 4x5 for the smaller cubes self.assertEqual(26, s.faces().size()) def testBoxCentered(self): x, y, z = 10, 11, 12 # check that the bottom corner is where we expect it for all possible combinations of centered b = [True, False] expected_x = [-x / 2, 0] expected_y = [-y / 2, 0] expected_z = [-z / 2, 0] for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = ( Workplane() .box(x, y, z, centered=(xopt, yopt, zopt)) .vertices("<X and <Y and <Z") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, zval), 3) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = Workplane().box(x, y, z, centered=val).vertices(">X and >Y and >Z") self.assertEqual(s0.size(), 1) s1 = ( Workplane() .box(x, y, z, centered=(val, val, val)) .vertices(">X and >Y and >Z") ) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals(s0.val().toTuple(), s1.val().toTuple(), 3) def testSphereDefaults(self): s = Workplane("XY").sphere(10) self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(1, s.solids().size()) self.assertEqual(1, s.faces().size()) def testSphereCustom(self): s = Workplane("XY").sphere( 10, angle1=0, angle2=90, angle3=360, centered=(False, False, False) ) self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(2, s.faces().size()) # check that the bottom corner is where we expect it for all possible combinations of centered radius = 10 for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip((True, False), (0, radius)), repeat=3 ): s = Workplane().sphere(radius, centered=(xopt, yopt, zopt)) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals( s.val().Center().toTuple(), (xval, yval, zval), 3 ) # check centered=True produces the same result as centered=(True, True, True) for val in (True, False): s0 = Workplane().sphere(radius, centered=val) self.assertEqual(s0.size(), 1) s1 = Workplane().sphere(radius, centered=(val, val, val)) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals( s0.val().Center().toTuple(), s1.val().Center().toTuple(), 3 ) def testSpherePointList(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .sphere(0.25, combine=False) ) # self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(4, s.solids().size()) self.assertEqual(4, s.faces().size()) def testSphereCombine(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .sphere(2.25, combine=True) ) # self.saveModel(s) # Until FreeCAD fixes their sphere operation self.assertEqual(1, s.solids().size()) self.assertEqual(4, s.faces().size()) def testCylinderDefaults(self): s = Workplane("XY").cylinder(20, 10) self.assertEqual(1, s.size()) self.assertEqual(1, s.solids().size()) self.assertEqual(3, s.faces().size()) self.assertEqual(2, s.vertices().size()) self.assertTupleAlmostEquals(s.val().Center().toTuple(), (0, 0, 0), 3) def testCylinderCentering(self): radius = 10 height = 40 b = (True, False) expected_x = (0, radius) expected_y = (0, radius) expected_z = (0, height / 2) for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = Workplane("XY").cylinder(height, radius, centered=(xopt, yopt, zopt)) self.assertEqual(1, s.size()) self.assertTupleAlmostEquals( s.val().Center().toTuple(), (xval, yval, zval), 3 ) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = Workplane("XY").cylinder(height, radius, centered=val) self.assertEqual(s0.size(), 1) s1 = Workplane("XY").cylinder(height, radius, centered=(val, val, val)) self.assertEqual(s1.size(), 1) self.assertTupleAlmostEquals( s0.val().Center().toTuple(), s1.val().Center().toTuple(), 3 ) def testWedgeDefaults(self): s = Workplane("XY").wedge(10, 10, 10, 5, 5, 5, 5) self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(5, s.faces().size()) self.assertEqual(5, s.vertices().size()) def testWedgeCentering(self): s = Workplane("XY").wedge( 10, 10, 10, 5, 5, 5, 5, centered=(False, False, False) ) # self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(5, s.faces().size()) self.assertEqual(5, s.vertices().size()) # check that the bottom corner is where we expect it for all possible combinations of centered x, y, z = 10, 11, 12 b = [True, False] expected_x = [-x / 2, 0] expected_y = [-y / 2, 0] expected_z = [-z / 2, 0] for (xopt, xval), (yopt, yval), (zopt, zval) in product( zip(b, expected_x), zip(b, expected_y), zip(b, expected_z) ): s = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=(xopt, yopt, zopt)) .vertices("<X and <Y and <Z") ) self.assertEqual(s.size(), 1) self.assertTupleAlmostEquals(s.val().toTuple(), (xval, yval, zval), 3) # check centered=True produces the same result as centered=(True, True, True) for val in b: s0 = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=val) .vertices(">X and >Z") ) self.assertEqual(s0.size(), 1) s1 = ( Workplane() .wedge(x, y, z, 2, 2, x - 2, z - 2, centered=(val, val, val)) .vertices(">X and >Z") ) self.assertEqual(s0.size(), 1) self.assertTupleAlmostEquals(s0.val().toTuple(), s1.val().toTuple(), 3) def testWedgePointList(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .wedge(10, 10, 10, 5, 5, 5, 5, combine=False) ) # self.saveModel(s) self.assertEqual(4, s.solids().size()) self.assertEqual(20, s.faces().size()) self.assertEqual(20, s.vertices().size()) def testWedgeCombined(self): s = ( Workplane("XY") .rect(4.0, 4.0, forConstruction=True) .vertices() .wedge(10, 10, 10, 5, 5, 5, 5, combine=True) ) # self.saveModel(s) self.assertEqual(1, s.solids().size()) self.assertEqual(12, s.faces().size()) self.assertEqual(16, s.vertices().size()) def testQuickStartXY(self): s = ( Workplane(Plane.XY()) .box(2, 4, 0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testQuickStartYZ(self): s = ( Workplane(Plane.YZ()) .box(2, 4, 0.5) .faces(">X") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testQuickStartXZ(self): s = ( Workplane(Plane.XZ()) .box(2, 4, 0.5) .faces(">Y") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.assertEqual(1, s.solids().size()) self.assertEqual(14, s.faces().size()) self.saveModel(s) def testDoubleTwistedLoft(self): s = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) s2 = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=-4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) # self.assertEquals(10,s.faces().size()) # self.assertEquals(1,s.solids().size()) s3 = s.combineSolids(s2) self.saveModel(s3) def testTwistedLoft(self): s = ( Workplane("XY") .polygon(8, 20.0) .workplane(offset=4.0) .transformed(rotate=Vector(0, 0, 15.0)) .polygon(8, 20) .loft() ) self.assertEqual(10, s.faces().size()) self.assertEqual(1, s.solids().size()) self.saveModel(s) def testUnions(self): # duplicates a memory problem of some kind reported when combining lots of objects s = Workplane("XY").rect(0.5, 0.5).extrude(5.0) o = [] beginTime = time.time() for i in range(15): t = Workplane("XY").center(10.0 * i, 0).rect(0.5, 0.5).extrude(5.0) o.append(t) # union stuff for oo in o: s = s.union(oo) print("Total time %0.3f" % (time.time() - beginTime)) # Test unioning a Solid object s = Workplane(Plane.XY()) currentS = s.rect(2.0, 2.0).extrude(0.5) toUnion = s.rect(1.0, 1.0).extrude(1.0) resS = currentS.union(toUnion) self.assertEqual(11, resS.faces().size()) with self.assertRaises(ValueError): resS.union(toUnion.faces().val()) # Test syntactic sugar [__add__ method] sugar1 = currentS | toUnion sugar2 = currentS + toUnion self.assertEqual(resS.faces().size(), sugar1.faces().size()) self.assertEqual(resS.faces().size(), sugar2.faces().size()) def testCombine(self): s = Workplane(Plane.XY()) objects1 = s.rect(2.0, 2.0).extrude(0.5).faces(">Z").rect(1.0, 1.0).extrude(0.5) objects1.combine() self.assertEqual(11, objects1.faces().size()) objects1 = s.rect(2.0, 2.0).extrude(0.5) objects2 = s.rect(1.0, 1.0).extrude(0.5).translate((0, 0, 0.5)) objects2 = objects1.add(objects2).combine(glue=True, tol=None) self.assertEqual(11, objects2.faces().size()) def testCombineSolidsInLoop(self): # duplicates a memory problem of some kind reported when combining lots of objects s = Workplane("XY").rect(0.5, 0.5).extrude(5.0) o = [] beginTime = time.time() for i in range(15): t = Workplane("XY").center(10.0 * i, 0).rect(0.5, 0.5).extrude(5.0) o.append(t) # append the 'good way' for oo in o: s.add(oo) s = s.combineSolids() print("Total time %0.3f" % (time.time() - beginTime)) self.saveModel(s) def testClean(self): # make a cube with a splitter edge on one of the faces # autosimplify should remove the splitter s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10) ) self.assertEqual(6, s.faces().size()) # test removal of splitter caused by union operation s = Workplane("XY").box(10, 10, 10).union(Workplane("XY").box(20, 10, 10)) self.assertEqual(6, s.faces().size()) # test removal of splitter caused by extrude+combine operation s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(5, 10, True) .extrude(20) ) self.assertEqual(10, s.faces().size()) # test removal of splitter caused by double hole operation s = ( Workplane("XY") .box(10, 10, 10) .faces(">Z") .workplane() .hole(3, 5) .faces(">Z") .workplane() .hole(3, 10) ) self.assertEqual(7, s.faces().size()) # test removal of splitter caused by cutThruAll s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(10, 5) .cutBlind(-5) .faces(">Z") .workplane(centerOption="CenterOfMass") .center(0, 2.5) .rect(5, 5) .cutThruAll() ) self.assertEqual(18, s.faces().size()) # test removal of splitter with box s = Workplane("XY").box(5, 5, 5).box(10, 5, 2) self.assertEqual(14, s.faces().size()) def testNoClean(self): # test disabling autoSimplify s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10, clean=False) ) self.assertEqual(7, s.faces().size()) s = ( Workplane("XY") .box(10, 10, 10) .union(Workplane("XY").box(20, 10, 10), clean=False) ) self.assertEqual(14, s.faces().size()) s = ( Workplane("XY") .box(10, 10, 10) .faces(">Y") .workplane() .rect(5, 10, True) .extrude(20, clean=False) ) self.assertEqual(12, s.faces().size()) def testExplicitClean(self): s = ( Workplane("XY") .moveTo(0, 0) .line(5, 0) .line(5, 0) .line(0, 10) .line(-10, 0) .close() .extrude(10, clean=False) .clean() ) self.assertEqual(6, s.faces().size()) def testPlanes(self): # ZX plane s = Workplane(Plane.ZX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # YX plane s = Workplane(Plane.YX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # YX plane s = Workplane(Plane.YX()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # ZY plane s = Workplane(Plane.ZY()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # front plane s = Workplane(Plane.front()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # back plane s = Workplane(Plane.back()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # left plane s = Workplane(Plane.left()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # right plane s = Workplane(Plane.right()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # top plane s = Workplane(Plane.top()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) # bottom plane s = Workplane(Plane.bottom()) result = ( s.rect(2.0, 4.0) .extrude(0.5) .faces(">Z") .workplane() .rect(1.5, 3.5, forConstruction=True) .vertices() .cskHole(0.125, 0.25, 82, depth=None) ) self.saveModel(result) def testIsInside(self): box1 = Workplane(Plane.XY()).box(10, 10, 10) box2 = Workplane(Plane.XY()).box(5, 5, 5) self.assertFalse(box2.val().BoundingBox().isInside(box1.val().BoundingBox())) self.assertTrue(box1.val().BoundingBox().isInside(box2.val().BoundingBox())) def testCup(self): # for some reason shell doesn't work on this simple shape. how disappointing! td = 50.0 bd = 20.0 h = 10.0 t = 1.0 s1 = Workplane("XY").circle(bd).workplane(offset=h).circle(td).loft() s2 = ( Workplane("XY") .workplane(offset=t) .circle(bd - (2.0 * t)) .workplane(offset=(h - t)) .circle(td - (2.0 * t)) .loft() ) s3 = s1.cut(s2) self.saveModel(s3) def testEnclosure(self): # parameter definitions p_outerWidth = 100.0 # Outer width of box enclosure p_outerLength = 150.0 # Outer length of box enclosure p_outerHeight = 50.0 # Outer height of box enclosure p_thickness = 3.0 # Thickness of the box walls p_sideRadius = 10.0 # Radius for the curves around the sides of the bo # Radius for the curves on the top and bottom edges of the box p_topAndBottomRadius = 2.0 # How far in from the edges the screwposts should be place. p_screwpostInset = 12.0 # Inner Diameter of the screwpost holes, should be roughly screw diameter not including threads p_screwpostID = 4.0 # Outer Diameter of the screwposts.\nDetermines overall thickness of the posts p_screwpostOD = 10.0 p_boreDiameter = 8.0 # Diameter of the counterbore hole, if any p_boreDepth = 1.0 # Depth of the counterbore hole, if # Outer diameter of countersink. Should roughly match the outer diameter of the screw head p_countersinkDiameter = 0.0 # Countersink angle (complete angle between opposite sides, not from center to one side) p_countersinkAngle = 90.0 # Whether to place the lid with the top facing down or not. p_flipLid = True # Height of lip on the underside of the lid.\nSits inside the box body for a snug fit. p_lipHeight = 1.0 # outer shell oshell = ( Workplane("XY") .rect(p_outerWidth, p_outerLength) .extrude(p_outerHeight + p_lipHeight) ) # weird geometry happens if we make the fillets in the wrong order if p_sideRadius > p_topAndBottomRadius: oshell = ( oshell.edges("|Z") .fillet(p_sideRadius) .edges(" .fillet(p_topAndBottomRadius) ) else: oshell = ( oshell.edges(" .fillet(p_topAndBottomRadius) .edges("|Z") .fillet(p_sideRadius) ) # inner shell ishell = ( oshell.faces("<Z") .workplane(p_thickness, True) .rect( (p_outerWidth - 2.0 * p_thickness), (p_outerLength - 2.0 * p_thickness) ) .extrude((p_outerHeight - 2.0 * p_thickness), False) ) # set combine false to produce just the new boss ishell = ishell.edges("|Z").fillet(p_sideRadius - p_thickness) # make the box outer box box = oshell.cut(ishell) # make the screwposts POSTWIDTH = p_outerWidth - 2.0 * p_screwpostInset POSTLENGTH = p_outerLength - 2.0 * p_screwpostInset box = ( box.faces(">Z") .workplane(-p_thickness) .rect(POSTWIDTH, POSTLENGTH, forConstruction=True) .vertices() .circle(p_screwpostOD / 2.0) .circle(p_screwpostID / 2.0) .extrude((-1.0) * (p_outerHeight + p_lipHeight - p_thickness), True) ) # split lid into top and bottom parts (lid, bottom) = ( box.faces(">Z") .workplane(-p_thickness - p_lipHeight) .split(keepTop=True, keepBottom=True) .all() ) # splits into two solids # translate the lid, and subtract the bottom from it to produce the lid inset lowerLid = lid.translate((0, 0, -p_lipHeight)) cutlip = lowerLid.cut(bottom).translate( (p_outerWidth + p_thickness, 0, p_thickness - p_outerHeight + p_lipHeight) ) # compute centers for counterbore/countersink or counterbore topOfLidCenters = ( cutlip.faces(">Z") .workplane() .rect(POSTWIDTH, POSTLENGTH, forConstruction=True) .vertices() ) # add holes of the desired type if p_boreDiameter > 0 and p_boreDepth > 0: topOfLid = topOfLidCenters.cboreHole( p_screwpostID, p_boreDiameter, p_boreDepth, (2.0) * p_thickness ) elif p_countersinkDiameter > 0 and p_countersinkAngle > 0: topOfLid = topOfLidCenters.cskHole( p_screwpostID, p_countersinkDiameter, p_countersinkAngle, (2.0) * p_thickness, ) else: topOfLid = topOfLidCenters.hole(p_screwpostID, (2.0) * p_thickness) # flip lid upside down if desired if p_flipLid: topOfLid.rotateAboutCenter((1, 0, 0), 180) # return the combined result result = topOfLid.union(bottom) self.saveModel(result) def testExtrude(self): r = 1.0 h = 1.0 decimal_places = 9.0 # extrude in one direction s = Workplane("XY").circle(r).extrude(h, both=False) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # calculate the distance between the top and the bottom face delta = top_face.val().Center().sub(bottom_face.val().Center()) self.assertTupleAlmostEquals(delta.toTuple(), (0.0, 0.0, h), decimal_places) # extrude symmetrically s = Workplane("XY").circle(r).extrude(h, both=True) self.assertTrue(len(s.val().Solids()) == 1) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # calculate the distance between the top and the bottom face delta = top_face.val().Center().sub(bottom_face.val().Center()) self.assertTupleAlmostEquals( delta.toTuple(), (0.0, 0.0, 2.0 * h), decimal_places ) # check that non-conplanar extrusion raises with self.assertRaises(ValueError): Workplane().box(1, 1, 1).faces().circle(0.1).extrude(0.1) def testTaperedExtrudeCutBlind(self): h = 1.0 r = 1.0 t = 5 # extrude with a positive taper s = Workplane("XY").circle(r).extrude(h, taper=t) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # top and bottom face area delta = top_face.val().Area() - bottom_face.val().Area() self.assertTrue(delta < 0) # extrude with a negative taper s = Workplane("XY").circle(r).extrude(h, taper=-t) top_face = s.faces(">Z") bottom_face = s.faces("<Z") # top and bottom face area delta = top_face.val().Area() - bottom_face.val().Area() self.assertTrue(delta > 0) # cut a tapered hole s = ( Workplane("XY") .rect(2 * r, 2 * r) .extrude(2 * h) .faces(">Z") .workplane() .rect(r, r) .cutBlind(-h, taper=t) ) middle_face = s.faces(">Z[-2]") self.assertTrue(middle_face.val().Area() < 1) def testClose(self): # Close without endPoint and startPoint coincide. # Create a half-circle a = Workplane(Plane.XY()).sagittaArc((10, 0), 2).close().extrude(2) # Close when endPoint and startPoint coincide. # Create a double half-circle b = ( Workplane(Plane.XY()) .sagittaArc((10, 0), 2) .sagittaArc((0, 0), 2) .close() .extrude(2) ) # The b shape shall have twice the volume of the a shape. self.assertAlmostEqual(a.val().Volume() * 2.0, b.val().Volume()) # Testcase 3 from issue #238 thickness = 3.0 length = 10.0 width = 5.0 obj1 = ( Workplane("XY", origin=(0, 0, -thickness / 2)) .moveTo(length / 2, 0) .threePointArc((0, width / 2), (-length / 2, 0)) .threePointArc((0, -width / 2), (length / 2, 0)) .close() .extrude(thickness) ) os_x = 8.0 # Offset in X os_y = -19.5 # Offset in Y obj2 = ( Workplane("YZ", origin=(os_x, os_y, -thickness / 2)) .moveTo(os_x + length / 2, os_y) .sagittaArc((os_x - length / 2, os_y), width / 2) .sagittaArc((os_x + length / 2, os_y), width / 2) .close() .extrude(thickness) ) # The obj1 shape shall have the same volume as the obj2 shape. self.assertAlmostEqual(obj1.val().Volume(), obj2.val().Volume()) def testText(self): global testdataDir box = Workplane("XY").box(4, 4, 0.5) obj1 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, -0.05, cut=True, halign="left", valign="bottom", font="Sans", ) ) # combined object should have smaller volume self.assertGreater(box.val().Volume(), obj1.val().Volume()) obj2 = ( box.faces(">Z") .workplane() .text("CQ 2.0", 0.5, 0.05, cut=False, combine=True, font="Sans") ) # combined object should have bigger volume self.assertLess(box.val().Volume(), obj2.val().Volume()) # verify that the number of top faces is correct (NB: this is font specific) self.assertEqual(len(obj2.faces(">Z").vals()), 5) obj3 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj3.solids().vals()), 5) obj4 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, fontPath=os.path.join(testdataDir, "OpenSans-Regular.ttf"), cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj4.solids().vals()), 5) # test to see if non-existent file causes segfault obj5 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 0.5, 0.05, fontPath=os.path.join(testdataDir, "OpenSans-Irregular.ttf"), cut=False, combine=False, halign="right", valign="top", font="Sans", ) ) # verify that the number of solids is correct self.assertEqual(len(obj5.solids().vals()), 5) # check it doesn't fall over with int sizes obj1 = ( box.faces(">Z") .workplane() .text( "CQ 2.0", 10, -1, cut=True, halign="left", valign="bottom", font="Sans", ) ) def testParametricCurve(self): from math import sin, cos, pi k = 4 r = 1 func = lambda t: ( r * (k + 1) * cos(t) - r * cos((k + 1) * t), r * (k + 1) * sin(t) - r * sin((k + 1) * t), ) res_open = Workplane("XY").parametricCurve(func).extrude(3) # open profile generates an invalid solid self.assertFalse(res_open.solids().val().isValid()) res_closed = ( Workplane("XY").parametricCurve(func, start=0, stop=2 * pi).extrude(3) ) # closed profile will generate a valid solid with 3 faces self.assertTrue(res_closed.solids().val().isValid()) self.assertEqual(len(res_closed.faces().vals()), 3) res_edge = Workplane("XY").parametricCurve(func, makeWire=False) self.assertEqual(len(res_edge.ctx.pendingEdges), 1) self.assertEqual(len(res_edge.ctx.pendingWires), 0) def testMakeShellSolid(self): c0 = math.sqrt(2) / 4 vertices = [[c0, -c0, c0], [c0, c0, -c0], [-c0, c0, c0], [-c0, -c0, -c0]] faces_ixs = [[0, 1, 2, 0], [1, 0, 3, 1], [2, 3, 0, 2], [3, 2, 1, 3]] faces = [] for ixs in faces_ixs: lines = [] for v1, v2 in zip(ixs, ixs[1:]): lines.append( Edge.makeLine(Vector(*vertices[v1]), Vector(*vertices[v2])) ) wire = Wire.combine(lines)[0] faces.append(Face.makeFromWires(wire)) shell = Shell.makeShell(faces) solid = Solid.makeSolid(shell) self.assertTrue(shell.isValid()) self.assertTrue(solid.isValid()) self.assertEqual(len(solid.Vertices()), 4) self.assertEqual(len(solid.Faces()), 4) def testIsInsideSolid(self): # test solid model = Workplane("XY").box(10, 10, 10) solid = model.val() # get first object on stack self.assertTrue(solid.isInside((0, 0, 0))) self.assertFalse(solid.isInside((10, 10, 10))) self.assertTrue(solid.isInside((Vector(3, 3, 3)))) self.assertFalse(solid.isInside((Vector(30.0, 30.0, 30.0)))) self.assertTrue(solid.isInside((0, 0, 4.99), tolerance=0.1)) self.assertTrue(solid.isInside((0, 0, 5))) # check point on surface self.assertTrue(solid.isInside((0, 0, 5.01), tolerance=0.1)) self.assertFalse(solid.isInside((0, 0, 5.1), tolerance=0.1)) # test compound solid model = Workplane("XY").box(10, 10, 10) model = model.moveTo(50, 50).box(10, 10, 10) solid = model.val() self.assertTrue(solid.isInside((0, 0, 0))) self.assertTrue(solid.isInside((50, 50, 0))) self.assertFalse(solid.isInside((50, 56, 0))) # make sure raises on non solid model = Workplane("XY").rect(10, 10) solid = model.val() with self.assertRaises(AttributeError): solid.isInside((0, 0, 0)) # test solid with an internal void void = Workplane("XY").box(10, 10, 10) model = Workplane("XY").box(100, 100, 100).cut(void) solid = model.val() self.assertFalse(solid.isInside((0, 0, 0))) self.assertTrue(solid.isInside((40, 40, 40))) self.assertFalse(solid.isInside((55, 55, 55))) def testWorkplaneCenterOptions(self): decimal_places = 9 pts = [(0, 0), (90, 0), (90, 30), (30, 30), (30, 60), (0.0, 60)] r = Workplane("XY").polyline(pts).close().extrude(10.0) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, 0.0, 10.0), decimal_places) origin = ( r.faces(">Z").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (37.5, 22.5, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (45.0, 30.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=(30, 10, 20)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 10.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=Vector(30, 10, 20)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 10.0, 10.0), decimal_places) with self.assertRaises(ValueError): origin = r.faces(">Z").workplane(centerOption="undefined") # test case where plane origin is shifted with center call r = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .center(30, 0) .hole(90) ) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (30.0, 0.0, 10.0), decimal_places) origin = ( r.faces(">Z") .workplane(centerOption="ProjectedOrigin", origin=(0, 0, 0)) .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, 0.0, 10.0), decimal_places) # make sure projection works in all directions r = Workplane("YZ").polyline(pts).close().extrude(10.0) origin = ( r.faces(">X") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 0.0, 0.0), decimal_places) origin = ( r.faces(">X").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 37.5, 22.5), decimal_places) origin = ( r.faces(">X") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (10.0, 45.0, 30.0), decimal_places) r = Workplane("XZ").polyline(pts).close().extrude(10.0) origin = ( r.faces("<Y") .workplane(centerOption="ProjectedOrigin") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (0.0, -10.0, 0.0), decimal_places) origin = ( r.faces("<Y").workplane(centerOption="CenterOfMass").plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (37.5, -10.0, 22.5), decimal_places) origin = ( r.faces("<Y") .workplane(centerOption="CenterOfBoundBox") .plane.origin.toTuple() ) self.assertTupleAlmostEquals(origin, (45.0, -10.0, 30.0), decimal_places) def testFindSolid(self): r = Workplane("XY").pushPoints([(-2, 0), (2, 0)]).box(1, 1, 1, combine=False) # there should be two solids on the stack self.assertEqual(len(r.objects), 2) self.assertTrue(isinstance(r.val(), Solid)) # find solid should return a compound of two solids s = r.findSolid() self.assertEqual(len(s.Solids()), 2) self.assertTrue(isinstance(s, Compound)) # if no solids are found, should raise ValueError w = Workplane().hLine(1).close() with raises(ValueError): w.findSolid() def testSlot2D(self): decimal_places = 9 # Ensure it produces a solid with the correct volume result = Workplane("XY").slot2D(4, 1, 0).extrude(1) self.assertAlmostEqual(result.val().Volume(), 3.785398163, decimal_places) # Test for proper expected behaviour when cutting box = Workplane("XY").box(5, 5, 1) result = box.faces(">Z").workplane().slot2D(4, 1, 0).cutThruAll() self.assertAlmostEqual(result.val().Volume(), 21.214601837, decimal_places) result = box.faces(">Z").workplane().slot2D(4, 1, 0).cutBlind(-0.5) self.assertAlmostEqual(result.val().Volume(), 23.107300918, decimal_places) # Test to see if slot is rotated correctly result = Workplane("XY").slot2D(4, 1, 45).extrude(1) point = result.faces(">Z").edges(">X").first().val().startPoint().toTuple() self.assertTupleAlmostEquals( point, (0.707106781, 1.414213562, 1.0), decimal_places ) def test_assembleEdges(self): # Plate with 5 sides and 2 bumps, one side is not co-planar with the other sides # Passes an open wire to assembleEdges so that IsDone is true but Error returns 2 to test the warning functionality. edge_points = [ [-7.0, -7.0, 0.0], [-3.0, -10.0, 3.0], [7.0, -7.0, 0.0], [7.0, 7.0, 0.0], [-7.0, 7.0, 0.0], ] edge_wire = Workplane("XY").polyline( [(-7.0, -7.0), (7.0, -7.0), (7.0, 7.0), (-7.0, 7.0)] ) edge_wire = edge_wire.add( Workplane("YZ") .workplane() .transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)) .spline([(-7.0, 0.0), (3, -3), (7.0, 0.0)]) ) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Embossed star, need to change optional parameters to obtain nice looking result. r1 = 3.0 r2 = 10.0 fn = 6 edge_points = [ [r1 * math.cos(i * math.pi / fn), r1 * math.sin(i * math.pi / fn)] if i % 2 == 0 else [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn)] for i in range(2 * fn + 1) ] edge_wire = Workplane("XY").polyline(edge_points) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Points on hexagonal pattern coordinates, use of pushpoints. r1 = 1.0 fn = 6 edge_points = [ [r1 * math.cos(i * 2 * math.pi / fn), r1 * math.sin(i * 2 * math.pi / fn)] for i in range(fn + 1) ] surface_points = [ [0.25, 0, 0.75], [-0.25, 0, 0.75], [0, 0.25, 0.75], [0, -0.25, 0.75], [0, 0, 2], ] edge_wire = Workplane("XY").polyline(edge_points) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) # Gyroïd, all edges are splines on different workplanes. edge_points = [ [[3.54, 3.54], [1.77, 0.0], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [-1.77, 0.0], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], ] plane_list = ["XZ", "XY", "YZ", "XZ", "YZ", "XY"] offset_list = [-3.54, 3.54, 3.54, 3.54, -3.54, -3.54] edge_wire = ( Workplane(plane_list[0]) .workplane(offset=-offset_list[0]) .spline(edge_points[0]) ) for i in range(len(edge_points) - 1): edge_wire = edge_wire.add( Workplane(plane_list[i + 1]) .workplane(offset=-offset_list[i + 1]) .spline(edge_points[i + 1]) ) edge_wire = [o.vals()[0] for o in edge_wire.all()] edge_wire = Wire.assembleEdges(edge_wire) def testTag(self): # test tagging result = ( Workplane("XY") .pushPoints([(-2, 0), (2, 0)]) .box(1, 1, 1, combine=False) .tag("2 solids") .union(Workplane("XY").box(6, 1, 1)) ) self.assertEqual(len(result.objects), 1) result = result._getTagged("2 solids") self.assertEqual(len(result.objects), 2) with self.assertRaises(ValueError): result = result._getTagged("3 solids") def testCopyWorkplane(self): obj0 = Workplane("XY").box(1, 1, 10).faces(">Z").workplane() obj1 = Workplane("XY").copyWorkplane(obj0).box(1, 1, 1) self.assertTupleAlmostEquals((0, 0, 5), obj1.val().Center().toTuple(), 9) def testWorkplaneFromTagged(self): # create a flat, wide base. Extrude one object 4 units high, another # object on top of it 6 units high. Go back to base plane. Extrude an # object 11 units high. Assert that top face is 11 units high. result = ( Workplane("XY") .box(10, 10, 1, centered=(True, True, False)) .faces(">Z") .workplane() .tag("base") .center(3, 0) .rect(2, 2) .extrude(4) .faces(">Z") .workplane() .circle(1) .extrude(6) .workplaneFromTagged("base") .center(-3, 0) .circle(1) .extrude(11) ) self.assertTupleAlmostEquals( result.faces(">Z").val().Center().toTuple(), (-3, 0, 12), 9 ) def testWorkplaneOrientationOnVertex(self): # create a 10 unit sized cube on the XY plane parent = Workplane("XY").rect(10.0, 10.0).extrude(10) # assert that the direction tuples reflect accordingly assert parent.plane.xDir.toTuple() == approx((1.0, 0.0, 0.0)) assert parent.plane.zDir.toTuple() == approx((0.0, 0.0, 1.0)) # select the <XZ vertex on the <Y face and create a new workplane. child = parent.faces("<Y").vertices("<XZ").workplane() # assert that the direction tuples reflect the new workplane on the <Y face assert child.plane.xDir.toTuple() == approx((1.0, 0.0, -0.0)) assert child.plane.zDir.toTuple() == approx((0.0, -1.0, -0.0)) def testTagSelectors(self): result0 = Workplane("XY").box(1, 1, 1).tag("box").sphere(1) # result is currently a sphere self.assertEqual(1, result0.faces().size()) # a box has 8 vertices self.assertEqual(8, result0.vertices(tag="box").size()) # 6 faces self.assertEqual(6, result0.faces(tag="box").size()) # 12 edges self.assertEqual(12, result0.edges(tag="box").size()) # 6 wires self.assertEqual(6, result0.wires(tag="box").size()) # create two solids, tag them, join to one solid result1 = ( Workplane("XY") .pushPoints([(1, 0), (-1, 0)]) .box(1, 1, 1) .tag("boxes") .sphere(1) ) self.assertEqual(1, result1.solids().size()) self.assertEqual(2, result1.solids(tag="boxes").size()) self.assertEqual(1, result1.shells().size()) self.assertEqual(2, result1.shells(tag="boxes").size()) # create 4 individual objects, tag it, then combine to one compound result2 = ( Workplane("XY") .rect(4, 4) .vertices() .box(1, 1, 1, combine=False) .tag("4 objs") ) result2 = result2.newObject([Compound.makeCompound(result2.objects)]) self.assertEqual(1, result2.compounds().size()) self.assertEqual(0, result2.compounds(tag="4 objs").size()) def test_interpPlate(self): # example from PythonOCC core_geometry_geomplate.py, use of thickness = 0 returns 2D surface. thickness = 0 edge_points = [ [0.0, 0.0, 0.0], [0.0, 10.0, 0.0], [0.0, 10.0, 10.0], [0.0, 0.0, 10.0], ] surface_points = [[5.0, 5.0, 5.0]] plate_0 = Workplane("XY").interpPlate(edge_points, surface_points, thickness) self.assertTrue(plate_0.val().isValid()) self.assertAlmostEqual(plate_0.val().Area(), 141.218823892, 1) # Plate with 5 sides and 2 bumps, one side is not co-planar with the other sides thickness = 0.1 edge_points = [ [-7.0, -7.0, 0.0], [-3.0, -10.0, 3.0], [7.0, -7.0, 0.0], [7.0, 7.0, 0.0], [-7.0, 7.0, 0.0], ] edge_wire = Workplane("XY").polyline( [(-7.0, -7.0), (7.0, -7.0), (7.0, 7.0), (-7.0, 7.0)] ) # edge_wire = edge_wire.add(Workplane('YZ').workplane().transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)).polyline([(-7.,0.), (3,-3), (7.,0.)])) # In CadQuery Sept-2019 it worked with rotate=Vector(0, 45, 0). In CadQuery Dec-2019 rotate=Vector(45, 0, 0) only closes the wire. edge_wire = edge_wire.add( Workplane("YZ") .workplane() .transformed(offset=Vector(0, 0, -7), rotate=Vector(45, 0, 0)) .spline([(-7.0, 0.0), (3, -3), (7.0, 0.0)]) ) surface_points = [[-3.0, -3.0, -3.0], [3.0, 3.0, 3.0]] plate_1 = Workplane("XY").interpPlate(edge_wire, surface_points, thickness) self.assertTrue(plate_1.val().isValid()) self.assertAlmostEqual(plate_1.val().Volume(), 26.124970206, 2) # Embossed star, need to change optional parameters to obtain nice looking result. r1 = 3.0 r2 = 10.0 fn = 6 thickness = 0.1 edge_points = [ [r1 * math.cos(i * math.pi / fn), r1 * math.sin(i * math.pi / fn)] if i % 2 == 0 else [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn)] for i in range(2 * fn + 1) ] edge_wire = Workplane("XY").polyline(edge_points) r2 = 4.5 surface_points = [ [r2 * math.cos(i * math.pi / fn), r2 * math.sin(i * math.pi / fn), 1.0] for i in range(2 * fn) ] + [[0.0, 0.0, -2.0]] plate_2 = Workplane("XY").interpPlate( edge_wire, surface_points, thickness, combine=True, clean=True, degree=3, nbPtsOnCur=15, nbIter=2, anisotropy=False, tol2d=0.00001, tol3d=0.0001, tolAng=0.01, tolCurv=0.1, maxDeg=8, maxSegments=49, ) self.assertTrue(plate_2.val().isValid()) self.assertAlmostEqual(plate_2.val().Volume(), 10.956054314, 0) # Points on hexagonal pattern coordinates, use of pushpoints. r1 = 1.0 N = 3 ca = math.cos(30.0 * math.pi / 180.0) sa = math.sin(30.0 * math.pi / 180.0) # EVEN ROWS pts = [ (-3.0, -3.0), (-1.267949, -3.0), (0.464102, -3.0), (2.196152, -3.0), (-3.0, 0.0), (-1.267949, 0.0), (0.464102, 0.0), (2.196152, 0.0), (-2.133974, -1.5), (-0.401923, -1.5), (1.330127, -1.5), (3.062178, -1.5), (-2.133975, 1.5), (-0.401924, 1.5), (1.330127, 1.5), (3.062178, 1.5), ] # Spike surface thickness = 0.1 fn = 6 edge_points = [ [ r1 * math.cos(i * 2 * math.pi / fn + 30 * math.pi / 180), r1 * math.sin(i * 2 * math.pi / fn + 30 * math.pi / 180), ] for i in range(fn + 1) ] surface_points = [ [ r1 / 4 * math.cos(i * 2 * math.pi / fn + 30 * math.pi / 180), r1 / 4 * math.sin(i * 2 * math.pi / fn + 30 * math.pi / 180), 0.75, ] for i in range(fn + 1) ] + [[0, 0, 2]] edge_wire = Workplane("XY").polyline(edge_points) plate_3 = ( Workplane("XY") .pushPoints(pts) .interpPlate( edge_wire, surface_points, thickness, combine=False, clean=False, degree=2, nbPtsOnCur=20, nbIter=2, anisotropy=False, tol2d=0.00001, tol3d=0.0001, tolAng=0.01, tolCurv=0.1, maxDeg=8, maxSegments=9, ) ) self.assertTrue(plate_3.val().isValid()) self.assertAlmostEqual(plate_3.val().Volume(), 0.45893954685189414, 1) # Gyroïd, all edges are splines on different workplanes. thickness = 0.1 edge_points = [ [[3.54, 3.54], [1.77, 0.0], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [0.0, -1.77], [3.54, -3.54]], [[-3.54, -3.54], [-1.77, 0.0], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], [[3.54, 3.54], [0.0, 1.77], [-3.54, 3.54]], ] plane_list = ["XZ", "XY", "YZ", "XZ", "YZ", "XY"] offset_list = [-3.54, 3.54, 3.54, 3.54, -3.54, -3.54] edge_wire = ( Workplane(plane_list[0]) .workplane(offset=-offset_list[0]) .spline(edge_points[0]) ) for i in range(len(edge_points) - 1): edge_wire = edge_wire.add( Workplane(plane_list[i + 1]) .workplane(offset=-offset_list[i + 1]) .spline(edge_points[i + 1]) ) surface_points = [[0, 0, 0]] plate_4 = Workplane("XY").interpPlate(edge_wire, surface_points, thickness) self.assertTrue(plate_4.val().isValid()) self.assertAlmostEqual(plate_4.val().Volume(), 7.760559490, 2) def testTangentArcToPoint(self): # create a simple shape with tangents of straight edges and see if it has the correct area s0 = ( Workplane("XY") .hLine(1) .tangentArcPoint((1, 1), relative=False) .hLineTo(0) .tangentArcPoint((0, 0), relative=False) .close() .extrude(1) ) area0 = s0.faces(">Z").val().Area() self.assertAlmostEqual(area0, (1 + math.pi * 0.5 ** 2), 4) # test relative coords s1 = ( Workplane("XY") .hLine(1) .tangentArcPoint((0, 1), relative=True) .hLineTo(0) .tangentArcPoint((0, -1), relative=True) .close() .extrude(1) ) self.assertTupleAlmostEquals( s1.val().Center().toTuple(), s0.val().Center().toTuple(), 4 ) self.assertAlmostEqual(s1.val().Volume(), s0.val().Volume(), 4) # consecutive tangent arcs s1 = ( Workplane("XY") .vLine(2) .tangentArcPoint((1, 0)) .tangentArcPoint((1, 0)) .tangentArcPoint((1, 0)) .vLine(-2) .close() .extrude(1) ) self.assertAlmostEqual( s1.faces(">Z").val().Area(), 2 * 3 + 0.5 * math.pi * 0.5 ** 2, 4 ) # tangentArc on the end of a spline # spline will be a simple arc of a circle, then finished off with a # tangentArcPoint angles = [idx * 1.5 * math.pi / 10 for idx in range(10)] pts = [(math.sin(a), math.cos(a)) for a in angles] s2 = ( Workplane("XY") .spline(pts) .tangentArcPoint((0, 1), relative=False) .close() .extrude(1) ) # volume should almost be pi, but not accurately because we need to # start with a spline self.assertAlmostEqual(s2.val().Volume(), math.pi, 1) # assert local coords are mapped to global correctly arc0 = Workplane("XZ", origin=(1, 1, 1)).hLine(1).tangentArcPoint((1, 1)).val() self.assertTupleAlmostEquals(arc0.endPoint().toTuple(), (3, 1, 2), 4) # tangentArcPoint with 3-tuple argument w0 = Workplane("XY").lineTo(1, 1).tangentArcPoint((1, 1, 1)).wire() zmax = w0.val().BoundingBox().zmax self.assertAlmostEqual(zmax, 1, 1) def test_findFromEdge(self): part = Workplane("XY", origin=(1, 1, 1)).hLine(1) found_edge = part._findFromEdge(useLocalCoords=False) self.assertTupleAlmostEquals(found_edge.startPoint().toTuple(), (1, 1, 1), 3) self.assertTupleAlmostEquals(found_edge.Center().toTuple(), (1.5, 1, 1), 3) self.assertTupleAlmostEquals(found_edge.endPoint().toTuple(), (2, 1, 1), 3) found_edge = part._findFromEdge(useLocalCoords=True) self.assertTupleAlmostEquals(found_edge.endPoint().toTuple(), (1, 0, 0), 3) # check _findFromEdge can find a spline pts = [(0, 0), (0, 1), (1, 2), (2, 4)] spline0 = Workplane("XZ").spline(pts)._findFromEdge() self.assertTupleAlmostEquals((2, 0, 4), spline0.endPoint().toTuple(), 3) # check method fails if no edge is present part2 = Workplane("XY").box(1, 1, 1) with self.assertRaises(RuntimeError): part2._findFromEdge() with self.assertRaises(RuntimeError): part2._findFromEdge(useLocalCoords=True) def testMakeHelix(self): h = 10 pitch = 1.5 r = 1.2 obj = Wire.makeHelix(pitch, h, r) bb = obj.BoundingBox() self.assertAlmostEqual(bb.zlen, h, 1) def testUnionCompound(self): box1 = Workplane("XY").box(10, 20, 30) box2 = Workplane("YZ").box(10, 20, 30) shape_to_cut = Workplane("XY").box(15, 15, 15).translate((8, 8, 8)) list_of_shapes = [] for o in box1.all(): list_of_shapes.extend(o.vals()) for o in box2.all(): list_of_shapes.extend(o.vals()) obj = Workplane("XY").newObject(list_of_shapes).cut(shape_to_cut) assert obj.val().isValid() def testSection(self): box = Workplane("XY", origin=(1, 2, 3)).box(1, 1, 1) s1 = box.section() s2 = box.section(0.5) self.assertAlmostEqual(s1.faces().val().Area(), 1) self.assertAlmostEqual(s2.faces().val().Area(), 1) line = Workplane("XY").hLine(1) with self.assertRaises(ValueError): line.section() def testGlue(self): box1 = Workplane("XY").rect(1, 1).extrude(2) box2 = Workplane("XY", origin=(0, 1, 0)).rect(1, 1).extrude(1) res = box1.union(box2, glue=True) self.assertEqual(res.faces().size(), 8) obj = obj = ( Workplane("XY").rect(1, 1).extrude(2).moveTo(0, 2).rect(1, 1).extrude(2) ) res = obj.union(box2, glue=True) self.assertEqual(res.faces().size(), 10) def testFuzzyBoolOp(self): eps = 1e-3 box1 = Workplane("XY").box(1, 1, 1) box2 = Workplane("XY", origin=(1 + eps, 0.0)).box(1, 1, 1) box3 = Workplane("XY", origin=(2, 0, 0)).box(1, 1, 1) res = box1.union(box2) res_fuzzy = box1.union(box2, tol=eps) res_fuzzy2 = box1.union(box3).union(box2, tol=eps) self.assertEqual(res.solids().size(), 2) self.assertEqual(res_fuzzy.solids().size(), 1) self.assertEqual(res_fuzzy2.solids().size(), 1) def testLocatedMoved(self): box = Solid.makeBox(1, 1, 1, Vector(-0.5, -0.5, -0.5)) loc = Location(Vector(1, 1, 1)) box1 = box.located(loc) self.assertTupleAlmostEquals(box1.Center().toTuple(), (1, 1, 1), 6) self.assertTupleAlmostEquals(box.Center().toTuple(), (0, 0, 0), 6) box.locate(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (1, 1, 1), 6) box2 = box.moved(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (1, 1, 1), 6) self.assertTupleAlmostEquals(box2.Center().toTuple(), (2, 2, 2), 6) box.move(loc) self.assertTupleAlmostEquals(box.Center().toTuple(), (2, 2, 2), 6) def testNullShape(self): from OCP.TopoDS import TopoDS_Shape s = TopoDS_Shape() # make sure raises on non solid with self.assertRaises(ValueError): r = occ_impl.shapes.downcast(s) def testCenterOfBoundBox(self): obj = Workplane().pushPoints([(0, 0), (2, 2)]).box(1, 1, 1) c = obj.workplane(centerOption="CenterOfBoundBox").plane.origin self.assertTupleAlmostEquals(c.toTuple(), (1, 1, 0), 6) def testOffset2D(self): w1 = Workplane().rect(1, 1).offset2D(0.5, "arc") self.assertEqual(w1.edges().size(), 8) w2 = Workplane().rect(1, 1).offset2D(0.5, "tangent") self.assertEqual(w2.edges().size(), 4) w3 = Workplane().rect(1, 1).offset2D(0.5, "intersection") self.assertEqual(w3.edges().size(), 4) w4 = Workplane().pushPoints([(0, 0), (0, 5)]).rect(1, 1).offset2D(-0.5) self.assertEqual(w4.wires().size(), 0) w5 = Workplane().pushPoints([(0, 0), (0, 5)]).rect(1, 1).offset2D(-0.25) self.assertEqual(w5.wires().size(), 2) r = 20 s = 7 t = 1.5 points = [ (0, t / 2), (r / 2 - 1.5 * t, r / 2 - t), (s / 2, r / 2 - t), (s / 2, r / 2), (r / 2, r / 2), (r / 2, s / 2), (r / 2 - t, s / 2), (r / 2 - t, r / 2 - 1.5 * t), (t / 2, 0), ] s = ( Workplane("XY") .polyline(points) .mirrorX() .mirrorY() .offset2D(-0.9) .extrude(1) ) self.assertEqual(s.solids().size(), 4) # test forConstruction # forConstruction=True should place results in objects, not ctx.pendingWires w6 = Workplane().hLine(1).vLine(1).close().offset2D(0.5, forConstruction=True) self.assertEqual(len(w6.ctx.pendingWires), 0) self.assertEqual(w6.size(), 1) self.assertEqual(type(w6.val()), Wire) # make sure the resulting wire has forConstruction set self.assertEqual(w6.val().forConstruction, True) def testConsolidateWires(self): w1 = Workplane().lineTo(0, 1).lineTo(1, 1).consolidateWires() self.assertEqual(w1.size(), 1) w1 = Workplane().consolidateWires() self.assertEqual(w1.size(), 0) def testLocationAt(self): r = 1 e = Wire.makeHelix(r, r, r).Edges()[0] locs_frenet = e.locations([0, 1], frame="frenet") T1 = locs_frenet[0].wrapped.Transformation() T2 = locs_frenet[1].wrapped.Transformation() self.assertAlmostEqual(T1.TranslationPart().X(), r, 6) self.assertAlmostEqual(T2.TranslationPart().X(), r, 6) self.assertAlmostEqual( T1.GetRotation().GetRotationAngle(), -T2.GetRotation().GetRotationAngle(), 6 ) ga = e._geomAdaptor() locs_corrected = e.locations( [ga.FirstParameter(), ga.LastParameter()], mode="parameter", frame="corrected", ) T3 = locs_corrected[0].wrapped.Transformation() T4 = locs_corrected[1].wrapped.Transformation() self.assertAlmostEqual(T3.TranslationPart().X(), r, 6) self.assertAlmostEqual(T4.TranslationPart().X(), r, 6) w = Wire.assembleEdges( [ Edge.makeLine(Vector(), Vector(0, 1)), Edge.makeLine(Vector(0, 1), Vector(1, 1)), ] ) locs_wire = e.locations([0, 1]) T5 = locs_wire[0].wrapped.Transformation() T6 = locs_wire[1].wrapped.Transformation() self.assertAlmostEqual(T5.TranslationPart().X(), r, 0) self.assertAlmostEqual(T6.TranslationPart().X(), r, 1) def testNormal(self): circ = Workplane().circle(1).edges().val() n = circ.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) ell = Workplane().ellipse(1, 2).edges().val() n = ell.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) r = Workplane().rect(1, 2).wires().val() n = r.normal() self.assertTupleAlmostEquals(n.toTuple(), (0, 0, 1), 6) with self.assertRaises(ValueError): edge = Workplane().rect(1, 2).edges().val() n = edge.normal() def testPositionAt(self): # test with an open wire w = Workplane().lineTo(0, 1).lineTo(1, 1).wire().val() p0 = w.positionAt(0.0) p1 = w.positionAt(0.5) p2 = w.positionAt(1.0) self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 1, 0), 6) self.assertTupleAlmostEquals(p2.toTuple(), (1, 1, 0), 6) p0 = w.positionAt(0.0, mode="param") self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) p0, p1, p2 = w.positions([0.0, 0.25, 0.5]) self.assertTupleAlmostEquals(p0.toTuple(), (0, 0, 0), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 0.5, 0), 6) self.assertTupleAlmostEquals(p2.toTuple(), (0, 1, 0), 6) # test with a closed wire w = Workplane().lineTo(0, 1).close().wire().val() p0 = w.positionAt(0.0) p1 = w.positionAt(0.5) p2 = w.positionAt(1.0) self.assertTupleAlmostEquals(p0.toTuple(), p2.toTuple(), 6) self.assertTupleAlmostEquals(p1.toTuple(), (0, 1, 0), 6) def testTangengAt(self): pts = [(0, 0), (-1, 1), (-2, 0), (-1, 0)] path = Workplane("XZ").spline(pts, tangents=((0, 1), (1, 0))).val() self.assertTrue( path.tangentAt(0.0, mode="parameter") == path.tangentAt(0.0, mode="length") ) self.assertFalse( path.tangentAt(0.5, mode="parameter") == path.tangentAt(0.5, mode="length") ) arc = Workplane().radiusArc((2, 0), 1).val() self.assertTupleAlmostEquals( arc.tangentAt(math.pi / 2, "parameter").toTuple(), (1, 0, 0), 6 ) self.assertTupleAlmostEquals( arc.tangentAt(0.5, "length").toTuple(), (1, 0, 0), 6 ) def testEnd(self): with self.assertRaises(ValueError): Workplane().end() self.assertTrue(Workplane().objects == []) self.assertTrue(Workplane().box(1, 1, 1).end().objects == []) self.assertTrue(Workplane().box(1, 1, 1).box(2, 2, 1).end(2).objects == []) def testCutEach(self): # base shape: w = Workplane().box(3, 2, 2) # cutter: c = Workplane().box(2, 2, 2).val() # cut all the corners off w0 = w.vertices().cutEach(lambda loc: c.located(loc)) # we are left with a 1x2x2 box: self.assertAlmostEqual(w0.val().Volume(), 4, 3) # test error on no solid found w1 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w1.cutEach(lambda loc: c.located(loc)) def testCutBlind(self): # cutBlind is already tested in several of the complicated tests, so this method is short. # test ValueError on no solid found w0 = Workplane().hLine(1).vLine(1).close() with raises(ValueError): w0.cutBlind(1) def testFindFace(self): # if there are no faces to find, should raise ValueError w0 = Workplane() with raises(ValueError): w0.findFace() w1 = Workplane().box(1, 1, 1).faces(">Z") self.assertTrue(isinstance(w1.findFace(), Face)) with raises(ValueError): w1.findFace(searchStack=False) w2 = w1.workplane().circle(0.1).extrude(0.1) self.assertTrue(isinstance(w2.findFace(searchParents=True), Face)) with raises(ValueError): w2.findFace(searchParents=False) def testPopPending(self): # test pending edges w0 = Workplane().hLine(1) self.assertEqual(len(w0.ctx.pendingEdges), 1) edges = w0.ctx.popPendingEdges() self.assertEqual(len(edges), 1) self.assertEqual(edges[0], w0.val()) # pending edges should now be cleared self.assertEqual(len(w0.ctx.pendingEdges), 0) # test pending wires w1 = Workplane().hLine(1).vLine(1).close() wire = w1.val() self.assertEqual(w1.ctx.pendingWires[0], wire) pop_pending_output = w1.ctx.popPendingWires() self.assertEqual(pop_pending_output[0], wire) # pending wires should now be cleared self.assertEqual(len(w1.ctx.pendingWires), 0) # test error when empty pending edges w2 = Workplane() # the following 2 should not raise an exception w2.ctx.popPendingEdges(errorOnEmpty=False) w2.ctx.popPendingWires(errorOnEmpty=False) # empty edges w3 = Workplane().hLine(1).vLine(1).close() with self.assertRaises(ValueError): w3.ctx.popPendingEdges() # empty wires w4 = Workplane().circle(1).extrude(1) with self.assertRaises(ValueError): w4.ctx.popPendingWires() # test via cutBlind w5 = Workplane().circle(1).extrude(1) with self.assertRaises(ValueError): w5.cutBlind(-1) def testCompSolid(self): from OCP.BRepPrimAPI import BRepPrimAPI_MakePrism tool = Solid.makeSphere(1, angleDegrees3=120) shell = tool.Shells()[0] v = Vector(0, 0, 1) builder = BRepPrimAPI_MakePrism(shell.wrapped, v.wrapped) result = Shape.cast(builder.Shape()) self.assertEqual(len(result.CompSolids()), 1) self.assertEqual(len(result.Solids()), 4) def test2Dfillet(self): r = Workplane().rect(1, 2).wires().val() f = Face.makeFromWires(r) verts = r.Vertices() self.assertEqual(len(f.fillet2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.fillet2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.fillet2D(0.25, verts).Vertices()), 8) # Test fillet2D with open wire and single vertex w0 = Workplane().hLine(1).vLine(1).wire() w0_verts = w0.vertices(">X and <Y").vals() unfilleted_wire0 = w0.val() filleted_wire0 = unfilleted_wire0.fillet2D(0.5, w0_verts) self.assertEqual(len(filleted_wire0.Vertices()), 4) # the filleted wire is shorter than the original self.assertGreater(unfilleted_wire0.Length() - filleted_wire0.Length(), 0.1) def test2Dchamfer(self): r = Workplane().rect(1, 2).wires().val() f = Face.makeFromWires(r) verts = r.Vertices() self.assertEqual(len(f.chamfer2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.chamfer2D(0.5, verts).Vertices()), 6) self.assertEqual(len(r.chamfer2D(0.25, verts).Vertices()), 8) r = Workplane().hLine(1).vLine(1).wire().val() vs = r.Vertices() self.assertEqual(len(r.chamfer2D(0.25, [vs[1]]).Vertices()), 4) with raises(ValueError): r.chamfer2D(0.25, [vs[0]]) def testSplineApprox(self): from .naca import naca5305 from math import pi, cos pts = [Vector(e[0], e[1], 0) for e in naca5305] # spline e1 = Edge.makeSplineApprox(pts, 1e-6, maxDeg=6, smoothing=(1, 1, 1)) e2 = Edge.makeSplineApprox(pts, 1e-6, minDeg=2, maxDeg=6) self.assertTrue(e1.isValid()) self.assertTrue(e2.isValid()) self.assertTrue(e1.Length() > e2.Length()) with raises(ValueError): e4 = Edge.makeSplineApprox(pts, 1e-6, maxDeg=3, smoothing=(1, 1, 1.0)) pts_closed = pts + [pts[0]] e3 = Edge.makeSplineApprox(pts_closed) w = Edge.makeSplineApprox(pts).close() self.assertTrue(e3.IsClosed()) self.assertTrue(w.IsClosed()) # Workplane method w1 = Workplane().splineApprox(pts) w2 = Workplane().splineApprox(pts, forConstruction=True) w3 = Workplane().splineApprox(pts, makeWire=True) w4 = Workplane().splineApprox(pts, makeWire=True, forConstruction=True) self.assertEqual(w1.edges().size(), 1) self.assertEqual(len(w1.ctx.pendingEdges), 1) self.assertEqual(w2.edges().size(), 1) self.assertEqual(len(w2.ctx.pendingEdges), 0) self.assertEqual(w3.wires().size(), 1) self.assertEqual(len(w3.ctx.pendingWires), 1) self.assertEqual(w4.wires().size(), 1) self.assertEqual(len(w4.ctx.pendingWires), 0) # spline surface N = 40 T = 20 A = 5 pts = [ [ Vector(i, j, A * cos(2 * pi * i / T) * cos(2 * pi * j / T)) for i in range(N + 1) ] for j in range(N + 1) ] f1 = Face.makeSplineApprox(pts, smoothing=(1, 1, 1), maxDeg=6) f2 = Face.makeSplineApprox(pts) self.assertTrue(f1.isValid()) self.assertTrue(f2.isValid()) with raises(ValueError): f3 = Face.makeSplineApprox(pts, smoothing=(1, 1, 1), maxDeg=3) def testParametricSurface(self): from math import pi, cos r1 = Workplane().parametricSurface( lambda u, v: (u, v, cos(pi * u) * cos(pi * v)), start=-1, stop=1 ) self.assertTrue(r1.faces().val().isValid()) r2 = Workplane().box(1, 1, 3).split(r1) self.assertTrue(r2.solids().val().isValid()) self.assertEqual(r2.solids().size(), 2) def testEdgeWireClose(self): # test with edge e0 = Edge.makeThreePointArc(Vector(0, 0, 0), Vector(1, 1, 0), Vector(0, 2, 0)) self.assertFalse(e0.IsClosed()) w0 = e0.close() self.assertTrue(w0.IsClosed()) # test with already closed edge e1 = Edge.makeCircle(1) self.assertTrue(e1.IsClosed()) e2 = e1.close() self.assertTrue(e2.IsClosed()) self.assertEqual(type(e1), type(e2)) # test with already closed WIRE w1 = Wire.makeCircle(1, Vector(), Vector(0, 0, 1)) self.assertTrue(w1.IsClosed()) w2 = w1.close() self.assertTrue(w1 is w2) def testSplitShape(self): # split an edge with a vertex e0 = Edge.makeCircle(1, (0, 0, 0), (0, 0, 1)) v0 = Vertex.makeVertex(0, 1, 0) list_of_edges = e0.split(v0).Edges() self.assertEqual(len(list_of_edges), 2) self.assertTrue(Vector(0, 1, 0) in [e.endPoint() for e in list_of_edges]) # split a circle with multiple vertices angles = [2 * math.pi * idx / 10 for idx in range(10)] vecs = [Vector(math.sin(a), math.cos(a), 0) for a in angles] vertices = [Vertex.makeVertex(*v.toTuple()) for v in vecs] edges = e0.split(*vertices).Edges() self.assertEqual(len(edges), len(vertices) + 1) endpoints = [e.endPoint() for e in edges] self.assertTrue(all([v in endpoints for v in vecs])) def testBrepImportExport(self): # import/export to file s = Workplane().box(1, 1, 1).val() s.exportBrep("test.brep") si = Shape.importBrep("test.brep") self.assertTrue(si.isValid()) self.assertAlmostEqual(si.Volume(), 1) # import/export to BytesIO from io import BytesIO bio = BytesIO() s.exportBrep(bio) bio.seek(0) si = Shape.importBrep("test.brep") self.assertTrue(si.isValid()) self.assertAlmostEqual(si.Volume(), 1) def testFaceToPln(self): origin = (1, 2, 3) normal = (1, 1, 1) f0 = Face.makePlane(length=None, width=None, basePnt=origin, dir=normal) p0 = f0.toPln() self.assertTrue(Vector(p0.Location()) == Vector(origin)) self.assertTrue(Vector(p0.Axis().Direction()) == Vector(normal).normalized()) origin1 = (0, 0, -3) normal1 = (-1, 1, -1) f1 = Face.makePlane(length=0.1, width=100, basePnt=origin1, dir=normal1) p1 = f1.toPln() self.assertTrue(Vector(p1.Location()) == Vector(origin1)) self.assertTrue(Vector(p1.Axis().Direction()) == Vector(normal1).normalized()) f2 = Workplane().box(1, 1, 10, centered=False).faces(">Z").val() p2 = f2.toPln() self.assertTrue(p2.Contains(f2.Center().toPnt(), 0.1)) self.assertTrue(Vector(p2.Axis().Direction()) == f2.normalAt())
true
true
f717ae14e4c0f23eb4e044fafbe02e361dc9be71
6,397
py
Python
src/datadog_api_client/v2/model/incident_service_update_data.py
DataDog/datadog-api-client-python
de2fc57dbde9acf4b8c8eef94ac29911227a62a2
[ "Apache-2.0" ]
32
2021-01-07T15:09:56.000Z
2022-01-30T05:49:23.000Z
src/datadog_api_client/v2/model/incident_service_update_data.py
DataDog/datadog-api-client-python
de2fc57dbde9acf4b8c8eef94ac29911227a62a2
[ "Apache-2.0" ]
228
2020-09-03T14:03:54.000Z
2022-03-31T20:16:12.000Z
src/datadog_api_client/v2/model/incident_service_update_data.py
DataDog/datadog-api-client-python
de2fc57dbde9acf4b8c8eef94ac29911227a62a2
[ "Apache-2.0" ]
12
2020-09-15T21:36:03.000Z
2022-03-31T17:13:17.000Z
# Unless explicitly stated otherwise all files in this repository are licensed under the Apache-2.0 License. # This product includes software developed at Datadog (https://www.datadoghq.com/). # Copyright 2019-Present Datadog, Inc. from datadog_api_client.v2.model_utils import ( # noqa: F401 ApiTypeError, ModelComposed, ModelNormal, ModelSimple, cached_property, change_keys_js_to_python, convert_js_args_to_python_args, date, datetime, file_type, none_type, ) def lazy_import(): from datadog_api_client.v2.model.incident_service_relationships import IncidentServiceRelationships from datadog_api_client.v2.model.incident_service_type import IncidentServiceType from datadog_api_client.v2.model.incident_service_update_attributes import IncidentServiceUpdateAttributes globals()["IncidentServiceRelationships"] = IncidentServiceRelationships globals()["IncidentServiceType"] = IncidentServiceType globals()["IncidentServiceUpdateAttributes"] = IncidentServiceUpdateAttributes class IncidentServiceUpdateData(ModelNormal): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. Attributes: allowed_values (dict): The key is the tuple path to the attribute and the for var_name this is (var_name,). The value is a dict with a capitalized key describing the allowed value and an allowed value. These dicts store the allowed enum values. attribute_map (dict): The key is attribute name and the value is json key in definition. discriminator_value_class_map (dict): A dict to go from the discriminator variable value to the discriminator class name. validations (dict): The key is the tuple path to the attribute and the for var_name this is (var_name,). The value is a dict that stores validations for max_length, min_length, max_items, min_items, exclusive_maximum, inclusive_maximum, exclusive_minimum, inclusive_minimum, and regex. additional_properties_type (tuple): A tuple of classes accepted as additional properties values. """ allowed_values = {} validations = {} additional_properties_type = None _nullable = False @cached_property def openapi_types(): """ This must be a method because a model may have properties that are of type self, this must run after the class is loaded Returns openapi_types (dict): The key is attribute name and the value is attribute type. """ lazy_import() return { "type": (IncidentServiceType,), # noqa: E501 "attributes": (IncidentServiceUpdateAttributes,), # noqa: E501 "id": (str,), # noqa: E501 "relationships": (IncidentServiceRelationships,), # noqa: E501 } discriminator = None attribute_map = { "type": "type", # noqa: E501 "attributes": "attributes", # noqa: E501 "id": "id", # noqa: E501 "relationships": "relationships", # noqa: E501 } read_only_vars = {} _composed_schemas = {} @convert_js_args_to_python_args def __init__(self, type, *args, **kwargs): # noqa: E501 """IncidentServiceUpdateData - a model defined in OpenAPI Args: type (IncidentServiceType): Keyword Args: _check_type (bool): if True, values for parameters in openapi_types will be type checked and a TypeError will be raised if the wrong type is input. Defaults to True _path_to_item (tuple/list): This is a list of keys or values to drill down to the model in received_data when deserializing a response _spec_property_naming (bool): True if the variable names in the input data are serialized names, as specified in the OpenAPI document. False if the variable names in the input data are pythonic names, e.g. snake case (default) _configuration (Configuration): the instance to use when deserializing a file_type parameter. If passed, type conversion is attempted If omitted no type conversion is done. _visited_composed_classes (tuple): This stores a tuple of classes that we have traveled through so that if we see that class again we will not use its discriminator again. When traveling through a discriminator, the composed schema that is is traveled through is added to this set. For example if Animal has a discriminator petType and we pass in "Dog", and the class Dog allOf includes Animal, we move through Animal once using the discriminator, and pick Dog. Then in Dog, we will make an instance of the Animal class but this time we won't travel through its discriminator because we passed in _visited_composed_classes = (Animal,) attributes (IncidentServiceUpdateAttributes): [optional] # noqa: E501 id (str): The incident service's ID.. [optional] # noqa: E501 relationships (IncidentServiceRelationships): [optional] # noqa: E501 """ super().__init__(kwargs) self._check_pos_args(args) self.type = type @classmethod @convert_js_args_to_python_args def _from_openapi_data(cls, type, *args, **kwargs): # noqa: E501 """Helper creating a new instance from a response.""" self = super(IncidentServiceUpdateData, cls)._from_openapi_data(kwargs) self._check_pos_args(args) self.type = type return self
41.810458
110
0.612318
from datadog_api_client.v2.model_utils import ( ApiTypeError, ModelComposed, ModelNormal, ModelSimple, cached_property, change_keys_js_to_python, convert_js_args_to_python_args, date, datetime, file_type, none_type, ) def lazy_import(): from datadog_api_client.v2.model.incident_service_relationships import IncidentServiceRelationships from datadog_api_client.v2.model.incident_service_type import IncidentServiceType from datadog_api_client.v2.model.incident_service_update_attributes import IncidentServiceUpdateAttributes globals()["IncidentServiceRelationships"] = IncidentServiceRelationships globals()["IncidentServiceType"] = IncidentServiceType globals()["IncidentServiceUpdateAttributes"] = IncidentServiceUpdateAttributes class IncidentServiceUpdateData(ModelNormal): allowed_values = {} validations = {} additional_properties_type = None _nullable = False @cached_property def openapi_types(): lazy_import() return { "type": (IncidentServiceType,), "attributes": (IncidentServiceUpdateAttributes,), "id": (str,), "relationships": (IncidentServiceRelationships,), } discriminator = None attribute_map = { "type": "type", "attributes": "attributes", "id": "id", "relationships": "relationships", } read_only_vars = {} _composed_schemas = {} @convert_js_args_to_python_args def __init__(self, type, *args, **kwargs): super().__init__(kwargs) self._check_pos_args(args) self.type = type @classmethod @convert_js_args_to_python_args def _from_openapi_data(cls, type, *args, **kwargs): self = super(IncidentServiceUpdateData, cls)._from_openapi_data(kwargs) self._check_pos_args(args) self.type = type return self
true
true
f717ae4b52146d6d410534878463dfe84c747e04
1,598
py
Python
configs/lesa/lesa_mask_rcnn_r50_dconv_c3-c5_8x1_20e.py
Chenglin-Yang/LESA_detection
3238bd9c6f3eb4a2a746837043cd9b3b56c77216
[ "Apache-2.0" ]
4
2021-07-18T08:09:09.000Z
2022-01-06T06:19:00.000Z
configs/lesa/lesa_mask_rcnn_r50_dconv_c3-c5_8x1_20e.py
Chenglin-Yang/LESA_detection
3238bd9c6f3eb4a2a746837043cd9b3b56c77216
[ "Apache-2.0" ]
null
null
null
configs/lesa/lesa_mask_rcnn_r50_dconv_c3-c5_8x1_20e.py
Chenglin-Yang/LESA_detection
3238bd9c6f3eb4a2a746837043cd9b3b56c77216
[ "Apache-2.0" ]
1
2022-01-06T06:18:53.000Z
2022-01-06T06:18:53.000Z
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance_1024.py', '../_base_/schedules/schedule_20e.py', '../_base_/default_runtime.py', ] optimizer = dict(lr=0.01) model = dict( pretrained=\ './checkpoints/lesa_pretrained_imagenet/'+\ 'lesa_resnet50_pretrained/'+\ 'lesa_resnet50/'+\ 'checkpoint.pth', backbone=dict( type='ResNet', strides=(1,2,2,2), wrn=False, dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True), stage_spatial_res=[256, 128, 64, 32], # 1024: [256, 128, 64, 32], 1280: [320, 160, 80, 40] stage_with_first_conv = [True, True, True, False], lesa=dict( type='LESA', with_cp_UB_terms_only=True, # cp used on the unary and binary terms only. pe_type='detection_qr', # ('classification', 'detection_qr') groups = 8, df_channel_shrink = [2], # df: dynamic fusion df_kernel_size = [1,1], df_group = [1,1], ), stage_with_lesa = (False, False, True, True), ), rpn_head=dict( anchor_generator=dict( type='AnchorGenerator', scales=[8], ratios=[0.5, 1.0, 2.0], strides=[4, 8, 16, 32, 64]), ), ) data_root = 'data/coco/' data = dict( samples_per_gpu=1, workers_per_gpu=1, # test=dict( # ann_file=data_root + 'annotations/image_info_test-dev2017.json', # img_prefix=data_root + 'test2017/'), )
28.535714
98
0.570088
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance_1024.py', '../_base_/schedules/schedule_20e.py', '../_base_/default_runtime.py', ] optimizer = dict(lr=0.01) model = dict( pretrained=\ './checkpoints/lesa_pretrained_imagenet/'+\ 'lesa_resnet50_pretrained/'+\ 'lesa_resnet50/'+\ 'checkpoint.pth', backbone=dict( type='ResNet', strides=(1,2,2,2), wrn=False, dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True), stage_spatial_res=[256, 128, 64, 32], stage_with_first_conv = [True, True, True, False], lesa=dict( type='LESA', with_cp_UB_terms_only=True, pe_type='detection_qr', groups = 8, df_channel_shrink = [2], df_kernel_size = [1,1], df_group = [1,1], ), stage_with_lesa = (False, False, True, True), ), rpn_head=dict( anchor_generator=dict( type='AnchorGenerator', scales=[8], ratios=[0.5, 1.0, 2.0], strides=[4, 8, 16, 32, 64]), ), ) data_root = 'data/coco/' data = dict( samples_per_gpu=1, workers_per_gpu=1, )
true
true
f717aeda2009266b9983c6e194696efc982f1189
1,312
py
Python
injectable/testing/clear_injectables_util.py
mt3o/injectable
0ffc5c758b63d9391134cd822158e1846999b404
[ "MIT" ]
71
2018-02-05T04:12:27.000Z
2022-02-15T23:08:16.000Z
injectable/testing/clear_injectables_util.py
Euraxluo/injectable
74e640f0911480fb06fa97c1a468c3863541c0fd
[ "MIT" ]
104
2018-02-06T23:37:36.000Z
2021-08-25T04:50:15.000Z
injectable/testing/clear_injectables_util.py
Euraxluo/injectable
74e640f0911480fb06fa97c1a468c3863541c0fd
[ "MIT" ]
13
2019-02-10T18:52:50.000Z
2022-01-26T17:12:35.000Z
from typing import Union, Set from injectable import InjectionContainer from injectable.common_utils import get_dependency_name from injectable.container.injectable import Injectable from injectable.constants import DEFAULT_NAMESPACE def clear_injectables( dependency: Union[type, str], namespace: str = None ) -> Set[Injectable]: """ Utility function to clear all injectables registered for the dependency in a given namespace. Returns a set containing all cleared injectables. :param dependency: class or qualifier of the dependency. :param namespace: (optional) namespace in which the injectable will be registered. Defaults to :const:`injectable.constants.DEFAULT_NAMESPACE`. Usage:: >>> from injectable.testing import clear_injectables >>> clear_injectables("foo") .. versionadded:: 3.3.0 """ namespace = InjectionContainer.NAMESPACES[namespace or DEFAULT_NAMESPACE] if isinstance(dependency, str): injectables = namespace.qualifier_registry[dependency] namespace.qualifier_registry[dependency] = set() else: dependency_name = get_dependency_name(dependency) injectables = namespace.class_registry[dependency_name] namespace.class_registry[dependency_name] = set() return injectables
36.444444
86
0.747713
from typing import Union, Set from injectable import InjectionContainer from injectable.common_utils import get_dependency_name from injectable.container.injectable import Injectable from injectable.constants import DEFAULT_NAMESPACE def clear_injectables( dependency: Union[type, str], namespace: str = None ) -> Set[Injectable]: namespace = InjectionContainer.NAMESPACES[namespace or DEFAULT_NAMESPACE] if isinstance(dependency, str): injectables = namespace.qualifier_registry[dependency] namespace.qualifier_registry[dependency] = set() else: dependency_name = get_dependency_name(dependency) injectables = namespace.class_registry[dependency_name] namespace.class_registry[dependency_name] = set() return injectables
true
true
f717afa83d677a0b51cae1b85b8a1bc081942f7a
5,143
py
Python
setup.py
lead-ratings/python-nvd3
2e677886d7b36f3aca641bbb2cc1117c3589b849
[ "MIT" ]
null
null
null
setup.py
lead-ratings/python-nvd3
2e677886d7b36f3aca641bbb2cc1117c3589b849
[ "MIT" ]
null
null
null
setup.py
lead-ratings/python-nvd3
2e677886d7b36f3aca641bbb2cc1117c3589b849
[ "MIT" ]
null
null
null
#! /usr/bin/env python # coding=utf-8 import os from setuptools import setup, find_packages from fnmatch import fnmatchcase from distutils.util import convert_path def read(*parts): return open(os.path.join(os.path.dirname(__file__), *parts)).read() # Provided as an attribute, so you can append to these instead # of replicating them: standard_exclude = ('*.py', '*.pyc', '*$py.class', '*~', '.*', '*.bak') standard_exclude_directories = ('.*', 'CVS', '_darcs', './build', './dist', 'EGG-INFO', '*.egg-info') # (c) 2005 Ian Bicking and contributors; written for Paste (http://pythonpaste.org) # Licensed under the MIT license: http://www.opensource.org/licenses/mit-license.php # Note: you may want to copy this into your setup.py file verbatim, as # you can't import this from another package, when you don't know if # that package is installed yet. def find_package_data(where='.', package='', exclude=standard_exclude, exclude_directories=standard_exclude_directories, only_in_packages=True, show_ignored=False): """ Return a dictionary suitable for use in ``package_data`` in a distutils ``setup.py`` file. The dictionary looks like:: {'package': [files]} Where ``files`` is a list of all the files in that package that don't match anything in ``exclude``. If ``only_in_packages`` is true, then top-level directories that are not packages won't be included (but directories under packages will). Directories matching any pattern in ``exclude_directories`` will be ignored; by default directories with leading ``.``, ``CVS``, and ``_darcs`` will be ignored. If ``show_ignored`` is true, then all the files that aren't included in package data are shown on stderr (for debugging purposes). Note patterns use wildcards, or can be exact paths (including leading ``./``), and all searching is case-insensitive. """ out = {} stack = [(convert_path(where), '', package, only_in_packages)] while stack: where, prefix, package, only_in_packages = stack.pop(0) for name in os.listdir(where): fn = os.path.join(where, name) if os.path.isdir(fn): bad_name = False for pattern in exclude_directories: if (fnmatchcase(name, pattern) or fn.lower() == pattern.lower()): bad_name = True if show_ignored: print ("Directory %s ignored by pattern %s" % (fn, pattern)) break if bad_name: continue if (os.path.isfile(os.path.join(fn, '__init__.py')) and not prefix): if not package: new_package = name else: new_package = package + '.' + name stack.append((fn, '', new_package, False)) else: stack.append((fn, prefix + name + '/', package, only_in_packages)) elif package or not only_in_packages: # is a file bad_name = False for pattern in exclude: if (fnmatchcase(name, pattern) or fn.lower() == pattern.lower()): bad_name = True if show_ignored: print ("File %s ignored by pattern %s" % (fn, pattern)) break if bad_name: continue out.setdefault(package, []).append(prefix + name) return out setup( name='python-nvd3', version='0.13.5', description="Python NVD3 - Chart Library for d3.js", long_description=read('README.rst'), keywords='plot, graph, nvd3, d3', author='Belaid Arezqui', author_email='areski@gmail.com', url='http://github.com/areski/python-nvd3', license='MIT license', py_modules=['nvd3'], namespace_packages=[], test_suite='tests', packages=find_packages(), package_data=find_package_data(), zip_safe=False, install_requires=[ 'setuptools', 'python-slugify==0.0.9', 'Jinja2>=2.7.2' # -*- Extra requirements: -*- ], entry_points={ 'console_scripts': [ 'nvd3 = nvd3:_main', ], }, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Topic :: Multimedia :: Graphics :: Presentation', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
36.735714
88
0.569512
import os from setuptools import setup, find_packages from fnmatch import fnmatchcase from distutils.util import convert_path def read(*parts): return open(os.path.join(os.path.dirname(__file__), *parts)).read() standard_exclude = ('*.py', '*.pyc', '*$py.class', '*~', '.*', '*.bak') standard_exclude_directories = ('.*', 'CVS', '_darcs', './build', './dist', 'EGG-INFO', '*.egg-info') def find_package_data(where='.', package='', exclude=standard_exclude, exclude_directories=standard_exclude_directories, only_in_packages=True, show_ignored=False): out = {} stack = [(convert_path(where), '', package, only_in_packages)] while stack: where, prefix, package, only_in_packages = stack.pop(0) for name in os.listdir(where): fn = os.path.join(where, name) if os.path.isdir(fn): bad_name = False for pattern in exclude_directories: if (fnmatchcase(name, pattern) or fn.lower() == pattern.lower()): bad_name = True if show_ignored: print ("Directory %s ignored by pattern %s" % (fn, pattern)) break if bad_name: continue if (os.path.isfile(os.path.join(fn, '__init__.py')) and not prefix): if not package: new_package = name else: new_package = package + '.' + name stack.append((fn, '', new_package, False)) else: stack.append((fn, prefix + name + '/', package, only_in_packages)) elif package or not only_in_packages: bad_name = False for pattern in exclude: if (fnmatchcase(name, pattern) or fn.lower() == pattern.lower()): bad_name = True if show_ignored: print ("File %s ignored by pattern %s" % (fn, pattern)) break if bad_name: continue out.setdefault(package, []).append(prefix + name) return out setup( name='python-nvd3', version='0.13.5', description="Python NVD3 - Chart Library for d3.js", long_description=read('README.rst'), keywords='plot, graph, nvd3, d3', author='Belaid Arezqui', author_email='areski@gmail.com', url='http://github.com/areski/python-nvd3', license='MIT license', py_modules=['nvd3'], namespace_packages=[], test_suite='tests', packages=find_packages(), package_data=find_package_data(), zip_safe=False, install_requires=[ 'setuptools', 'python-slugify==0.0.9', 'Jinja2>=2.7.2' ], entry_points={ 'console_scripts': [ 'nvd3 = nvd3:_main', ], }, classifiers=[ 'Development Status :: 5 - Production/Stable', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.3', 'Topic :: Multimedia :: Graphics :: Presentation', 'Topic :: Software Development :: Libraries :: Python Modules', ], )
true
true
f717afcac3389e0e78ff975905922f94ec7b01ce
55,864
py
Python
lib/arrow/arrow.py
matke-84/script.module.arrow
7aa0903bc846541e4d620f10aca52957ba1e0e5b
[ "Apache-2.0" ]
null
null
null
lib/arrow/arrow.py
matke-84/script.module.arrow
7aa0903bc846541e4d620f10aca52957ba1e0e5b
[ "Apache-2.0" ]
null
null
null
lib/arrow/arrow.py
matke-84/script.module.arrow
7aa0903bc846541e4d620f10aca52957ba1e0e5b
[ "Apache-2.0" ]
null
null
null
""" Provides the :class:`Arrow <arrow.arrow.Arrow>` class, an enhanced ``datetime`` replacement. """ import calendar import sys from datetime import date from datetime import datetime as dt_datetime from datetime import time as dt_time from datetime import timedelta from datetime import tzinfo as dt_tzinfo from math import trunc from time import struct_time from typing import ( Any, ClassVar, Generator, Iterable, List, Mapping, Optional, Tuple, Union, cast, overload, ) from dateutil import tz as dateutil_tz from dateutil.relativedelta import relativedelta from arrow import formatter, locales, parser, util from arrow.locales import TimeFrameLiteral if sys.version_info < (3, 8): # pragma: no cover from typing_extensions import Final, Literal else: from typing import Final, Literal # pragma: no cover TZ_EXPR = Union[dt_tzinfo, str] _T_FRAMES = Literal[ "year", "years", "month", "months", "day", "days", "hour", "hours", "minute", "minutes", "second", "seconds", "microsecond", "microseconds", "week", "weeks", "quarter", "quarters", ] _BOUNDS = Literal["[)", "()", "(]", "[]"] _GRANULARITY = Literal[ "auto", "second", "minute", "hour", "day", "week", "month", "year", ] class Arrow: """An :class:`Arrow <arrow.arrow.Arrow>` object. Implements the ``datetime`` interface, behaving as an aware ``datetime`` while implementing additional functionality. :param year: the calendar year. :param month: the calendar month. :param day: the calendar day. :param hour: (optional) the hour. Defaults to 0. :param minute: (optional) the minute, Defaults to 0. :param second: (optional) the second, Defaults to 0. :param microsecond: (optional) the microsecond. Defaults to 0. :param tzinfo: (optional) A timezone expression. Defaults to UTC. :param fold: (optional) 0 or 1, used to disambiguate repeated times. Defaults to 0. .. _tz-expr: Recognized timezone expressions: - A ``tzinfo`` object. - A ``str`` describing a timezone, similar to 'US/Pacific', or 'Europe/Berlin'. - A ``str`` in ISO 8601 style, as in '+07:00'. - A ``str``, one of the following: 'local', 'utc', 'UTC'. Usage:: >>> import arrow >>> arrow.Arrow(2013, 5, 5, 12, 30, 45) <Arrow [2013-05-05T12:30:45+00:00]> """ resolution: ClassVar[timedelta] = dt_datetime.resolution min: ClassVar["Arrow"] max: ClassVar["Arrow"] _ATTRS: Final[List[str]] = [ "year", "month", "day", "hour", "minute", "second", "microsecond", ] _ATTRS_PLURAL: Final[List[str]] = [f"{a}s" for a in _ATTRS] _MONTHS_PER_QUARTER: Final[int] = 3 _SECS_PER_MINUTE: Final[int] = 60 _SECS_PER_HOUR: Final[int] = 60 * 60 _SECS_PER_DAY: Final[int] = 60 * 60 * 24 _SECS_PER_WEEK: Final[int] = 60 * 60 * 24 * 7 _SECS_PER_MONTH: Final[float] = 60 * 60 * 24 * 30.5 _SECS_PER_YEAR: Final[int] = 60 * 60 * 24 * 365 _SECS_MAP: Final[Mapping[TimeFrameLiteral, float]] = { "second": 1.0, "minute": _SECS_PER_MINUTE, "hour": _SECS_PER_HOUR, "day": _SECS_PER_DAY, "week": _SECS_PER_WEEK, "month": _SECS_PER_MONTH, "year": _SECS_PER_YEAR, } _datetime: dt_datetime def __init__( self, year: int, month: int, day: int, hour: int = 0, minute: int = 0, second: int = 0, microsecond: int = 0, tzinfo: Optional[TZ_EXPR] = None, **kwargs: Any, ) -> None: if tzinfo is None: tzinfo = dateutil_tz.tzutc() # detect that tzinfo is a pytz object (issue #626) elif ( isinstance(tzinfo, dt_tzinfo) and hasattr(tzinfo, "localize") and hasattr(tzinfo, "zone") and tzinfo.zone # type: ignore[attr-defined] ): tzinfo = parser.TzinfoParser.parse(tzinfo.zone) # type: ignore[attr-defined] elif isinstance(tzinfo, str): tzinfo = parser.TzinfoParser.parse(tzinfo) fold = kwargs.get("fold", 0) self._datetime = dt_datetime( year, month, day, hour, minute, second, microsecond, tzinfo, fold=fold ) # factories: single object, both original and from datetime. @classmethod def now(cls, tzinfo: Optional[dt_tzinfo] = None) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object, representing "now" in the given timezone. :param tzinfo: (optional) a ``tzinfo`` object. Defaults to local time. Usage:: >>> arrow.now('Asia/Baku') <Arrow [2019-01-24T20:26:31.146412+04:00]> """ if tzinfo is None: tzinfo = dateutil_tz.tzlocal() dt = dt_datetime.now(tzinfo) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def utcnow(cls) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object, representing "now" in UTC time. Usage:: >>> arrow.utcnow() <Arrow [2019-01-24T16:31:40.651108+00:00]> """ dt = dt_datetime.now(dateutil_tz.tzutc()) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def fromtimestamp( cls, timestamp: Union[int, float, str], tzinfo: Optional[TZ_EXPR] = None, ) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object from a timestamp, converted to the given timezone. :param timestamp: an ``int`` or ``float`` timestamp, or a ``str`` that converts to either. :param tzinfo: (optional) a ``tzinfo`` object. Defaults to local time. """ if tzinfo is None: tzinfo = dateutil_tz.tzlocal() elif isinstance(tzinfo, str): tzinfo = parser.TzinfoParser.parse(tzinfo) if not util.is_timestamp(timestamp): raise ValueError(f"The provided timestamp {timestamp!r} is invalid.") timestamp = util.normalize_timestamp(float(timestamp)) dt = dt_datetime.fromtimestamp(timestamp, tzinfo) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def utcfromtimestamp(cls, timestamp: Union[int, float, str]) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object from a timestamp, in UTC time. :param timestamp: an ``int`` or ``float`` timestamp, or a ``str`` that converts to either. """ if not util.is_timestamp(timestamp): raise ValueError(f"The provided timestamp {timestamp!r} is invalid.") timestamp = util.normalize_timestamp(float(timestamp)) dt = dt_datetime.utcfromtimestamp(timestamp) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dateutil_tz.tzutc(), fold=getattr(dt, "fold", 0), ) @classmethod def fromdatetime(cls, dt: dt_datetime, tzinfo: Optional[TZ_EXPR] = None) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object from a ``datetime`` and optional replacement timezone. :param dt: the ``datetime`` :param tzinfo: (optional) A :ref:`timezone expression <tz-expr>`. Defaults to ``dt``'s timezone, or UTC if naive. If you only want to replace the timezone of naive datetimes:: >>> dt datetime.datetime(2013, 5, 5, 0, 0, tzinfo=tzutc()) >>> arrow.Arrow.fromdatetime(dt, dt.tzinfo or 'US/Pacific') <Arrow [2013-05-05T00:00:00+00:00]> """ try: if dt.tzinfo is None: tzinfo = dateutil_tz.tzutc() else: tzinfo = dt.tzinfo except: tzinfo = dateutil_tz.tzutc() try: return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, tzinfo, fold=getattr(dt, "fold", 0), ) except: return cls( 1980, 1, 1, 0, 0, 0, 0, tzinfo, fold=0, ) @classmethod def fromdate(cls, date: date, tzinfo: Optional[TZ_EXPR] = None) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object from a ``date`` and optional replacement timezone. Time values are set to 0. :param date: the ``date`` :param tzinfo: (optional) A :ref:`timezone expression <tz-expr>`. Defaults to UTC. """ if tzinfo is None: tzinfo = dateutil_tz.tzutc() return cls(date.year, date.month, date.day, tzinfo=tzinfo) @classmethod def strptime( cls, date_str: str, fmt: str, tzinfo: Optional[TZ_EXPR] = None ) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object from a date string and format, in the style of ``datetime.strptime``. Optionally replaces the parsed timezone. :param date_str: the date string. :param fmt: the format string. :param tzinfo: (optional) A :ref:`timezone expression <tz-expr>`. Defaults to the parsed timezone if ``fmt`` contains a timezone directive, otherwise UTC. Usage:: >>> arrow.Arrow.strptime('20-01-2019 15:49:10', '%d-%m-%Y %H:%M:%S') <Arrow [2019-01-20T15:49:10+00:00]> """ dt = dt_datetime.strptime(date_str, fmt) if tzinfo is None: tzinfo = dt.tzinfo return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def fromordinal(cls, ordinal: int) -> "Arrow": """Constructs an :class:`Arrow <arrow.arrow.Arrow>` object corresponding to the Gregorian Ordinal. :param ordinal: an ``int`` corresponding to a Gregorian Ordinal. Usage:: >>> arrow.fromordinal(737741) <Arrow [2020-11-12T00:00:00+00:00]> """ util.validate_ordinal(ordinal) dt = dt_datetime.fromordinal(ordinal) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) # factories: ranges and spans @classmethod def range( cls, frame: _T_FRAMES, start: Union["Arrow", dt_datetime], end: Union["Arrow", dt_datetime, None] = None, tz: Optional[TZ_EXPR] = None, limit: Optional[int] = None, ) -> Generator["Arrow", None, None]: """Returns an iterator of :class:`Arrow <arrow.arrow.Arrow>` objects, representing points in time between two inputs. :param frame: The timeframe. Can be any ``datetime`` property (day, hour, minute...). :param start: A datetime expression, the start of the range. :param end: (optional) A datetime expression, the end of the range. :param tz: (optional) A :ref:`timezone expression <tz-expr>`. Defaults to ``start``'s timezone, or UTC if ``start`` is naive. :param limit: (optional) A maximum number of tuples to return. **NOTE**: The ``end`` or ``limit`` must be provided. Call with ``end`` alone to return the entire range. Call with ``limit`` alone to return a maximum # of results from the start. Call with both to cap a range at a maximum # of results. **NOTE**: ``tz`` internally **replaces** the timezones of both ``start`` and ``end`` before iterating. As such, either call with naive objects and ``tz``, or aware objects from the same timezone and no ``tz``. Supported frame values: year, quarter, month, week, day, hour, minute, second. Recognized datetime expressions: - An :class:`Arrow <arrow.arrow.Arrow>` object. - A ``datetime`` object. Usage:: >>> start = datetime(2013, 5, 5, 12, 30) >>> end = datetime(2013, 5, 5, 17, 15) >>> for r in arrow.Arrow.range('hour', start, end): ... print(repr(r)) ... <Arrow [2013-05-05T12:30:00+00:00]> <Arrow [2013-05-05T13:30:00+00:00]> <Arrow [2013-05-05T14:30:00+00:00]> <Arrow [2013-05-05T15:30:00+00:00]> <Arrow [2013-05-05T16:30:00+00:00]> **NOTE**: Unlike Python's ``range``, ``end`` *may* be included in the returned iterator:: >>> start = datetime(2013, 5, 5, 12, 30) >>> end = datetime(2013, 5, 5, 13, 30) >>> for r in arrow.Arrow.range('hour', start, end): ... print(repr(r)) ... <Arrow [2013-05-05T12:30:00+00:00]> <Arrow [2013-05-05T13:30:00+00:00]> """ _, frame_relative, relative_steps = cls._get_frames(frame) tzinfo = cls._get_tzinfo(start.tzinfo if tz is None else tz) start = cls._get_datetime(start).replace(tzinfo=tzinfo) end, limit = cls._get_iteration_params(end, limit) end = cls._get_datetime(end).replace(tzinfo=tzinfo) current = cls.fromdatetime(start) original_day = start.day day_is_clipped = False i = 0 while current <= end and i < limit: i += 1 yield current values = [getattr(current, f) for f in cls._ATTRS] current = cls(*values, tzinfo=tzinfo).shift( # type: ignore **{frame_relative: relative_steps} ) if frame in ["month", "quarter", "year"] and current.day < original_day: day_is_clipped = True if day_is_clipped and not cls._is_last_day_of_month(current): current = current.replace(day=original_day) def span( self, frame: _T_FRAMES, count: int = 1, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Tuple["Arrow", "Arrow"]: """Returns two new :class:`Arrow <arrow.arrow.Arrow>` objects, representing the timespan of the :class:`Arrow <arrow.arrow.Arrow>` object in a given timeframe. :param frame: the timeframe. Can be any ``datetime`` property (day, hour, minute...). :param count: (optional) the number of frames to span. :param bounds: (optional) a ``str`` of either '()', '(]', '[)', or '[]' that specifies whether to include or exclude the start and end values in the span. '(' excludes the start, '[' includes the start, ')' excludes the end, and ']' includes the end. If the bounds are not specified, the default bound '[)' is used. :param exact: (optional) whether to have the start of the timespan begin exactly at the time specified by ``start`` and the end of the timespan truncated so as not to extend beyond ``end``. Supported frame values: year, quarter, month, week, day, hour, minute, second. Usage:: >>> arrow.utcnow() <Arrow [2013-05-09T03:32:36.186203+00:00]> >>> arrow.utcnow().span('hour') (<Arrow [2013-05-09T03:00:00+00:00]>, <Arrow [2013-05-09T03:59:59.999999+00:00]>) >>> arrow.utcnow().span('day') (<Arrow [2013-05-09T00:00:00+00:00]>, <Arrow [2013-05-09T23:59:59.999999+00:00]>) >>> arrow.utcnow().span('day', count=2) (<Arrow [2013-05-09T00:00:00+00:00]>, <Arrow [2013-05-10T23:59:59.999999+00:00]>) >>> arrow.utcnow().span('day', bounds='[]') (<Arrow [2013-05-09T00:00:00+00:00]>, <Arrow [2013-05-10T00:00:00+00:00]>) """ util.validate_bounds(bounds) frame_absolute, frame_relative, relative_steps = self._get_frames(frame) if frame_absolute == "week": attr = "day" elif frame_absolute == "quarter": attr = "month" else: attr = frame_absolute floor = self if not exact: index = self._ATTRS.index(attr) frames = self._ATTRS[: index + 1] values = [getattr(self, f) for f in frames] for _ in range(3 - len(values)): values.append(1) floor = self.__class__(*values, tzinfo=self.tzinfo) # type: ignore if frame_absolute == "week": floor = floor.shift(days=-(self.isoweekday() - 1)) elif frame_absolute == "quarter": floor = floor.shift(months=-((self.month - 1) % 3)) ceil = floor.shift(**{frame_relative: count * relative_steps}) if bounds[0] == "(": floor = floor.shift(microseconds=+1) if bounds[1] == ")": ceil = ceil.shift(microseconds=-1) return floor, ceil def floor(self, frame: _T_FRAMES) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object, representing the "floor" of the timespan of the :class:`Arrow <arrow.arrow.Arrow>` object in a given timeframe. Equivalent to the first element in the 2-tuple returned by :func:`span <arrow.arrow.Arrow.span>`. :param frame: the timeframe. Can be any ``datetime`` property (day, hour, minute...). Usage:: >>> arrow.utcnow().floor('hour') <Arrow [2013-05-09T03:00:00+00:00]> """ return self.span(frame)[0] def ceil(self, frame: _T_FRAMES) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object, representing the "ceiling" of the timespan of the :class:`Arrow <arrow.arrow.Arrow>` object in a given timeframe. Equivalent to the second element in the 2-tuple returned by :func:`span <arrow.arrow.Arrow.span>`. :param frame: the timeframe. Can be any ``datetime`` property (day, hour, minute...). Usage:: >>> arrow.utcnow().ceil('hour') <Arrow [2013-05-09T03:59:59.999999+00:00]> """ return self.span(frame)[1] @classmethod def span_range( cls, frame: _T_FRAMES, start: dt_datetime, end: dt_datetime, tz: Optional[TZ_EXPR] = None, limit: Optional[int] = None, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Iterable[Tuple["Arrow", "Arrow"]]: """Returns an iterator of tuples, each :class:`Arrow <arrow.arrow.Arrow>` objects, representing a series of timespans between two inputs. :param frame: The timeframe. Can be any ``datetime`` property (day, hour, minute...). :param start: A datetime expression, the start of the range. :param end: (optional) A datetime expression, the end of the range. :param tz: (optional) A :ref:`timezone expression <tz-expr>`. Defaults to ``start``'s timezone, or UTC if ``start`` is naive. :param limit: (optional) A maximum number of tuples to return. :param bounds: (optional) a ``str`` of either '()', '(]', '[)', or '[]' that specifies whether to include or exclude the start and end values in each span in the range. '(' excludes the start, '[' includes the start, ')' excludes the end, and ']' includes the end. If the bounds are not specified, the default bound '[)' is used. :param exact: (optional) whether to have the first timespan start exactly at the time specified by ``start`` and the final span truncated so as not to extend beyond ``end``. **NOTE**: The ``end`` or ``limit`` must be provided. Call with ``end`` alone to return the entire range. Call with ``limit`` alone to return a maximum # of results from the start. Call with both to cap a range at a maximum # of results. **NOTE**: ``tz`` internally **replaces** the timezones of both ``start`` and ``end`` before iterating. As such, either call with naive objects and ``tz``, or aware objects from the same timezone and no ``tz``. Supported frame values: year, quarter, month, week, day, hour, minute, second. Recognized datetime expressions: - An :class:`Arrow <arrow.arrow.Arrow>` object. - A ``datetime`` object. **NOTE**: Unlike Python's ``range``, ``end`` will *always* be included in the returned iterator of timespans. Usage: >>> start = datetime(2013, 5, 5, 12, 30) >>> end = datetime(2013, 5, 5, 17, 15) >>> for r in arrow.Arrow.span_range('hour', start, end): ... print(r) ... (<Arrow [2013-05-05T12:00:00+00:00]>, <Arrow [2013-05-05T12:59:59.999999+00:00]>) (<Arrow [2013-05-05T13:00:00+00:00]>, <Arrow [2013-05-05T13:59:59.999999+00:00]>) (<Arrow [2013-05-05T14:00:00+00:00]>, <Arrow [2013-05-05T14:59:59.999999+00:00]>) (<Arrow [2013-05-05T15:00:00+00:00]>, <Arrow [2013-05-05T15:59:59.999999+00:00]>) (<Arrow [2013-05-05T16:00:00+00:00]>, <Arrow [2013-05-05T16:59:59.999999+00:00]>) (<Arrow [2013-05-05T17:00:00+00:00]>, <Arrow [2013-05-05T17:59:59.999999+00:00]>) """ tzinfo = cls._get_tzinfo(start.tzinfo if tz is None else tz) start = cls.fromdatetime(start, tzinfo).span(frame, exact=exact)[0] end = cls.fromdatetime(end, tzinfo) _range = cls.range(frame, start, end, tz, limit) if not exact: for r in _range: yield r.span(frame, bounds=bounds, exact=exact) for r in _range: floor, ceil = r.span(frame, bounds=bounds, exact=exact) if ceil > end: ceil = end if bounds[1] == ")": ceil += relativedelta(microseconds=-1) if floor == end: break elif floor + relativedelta(microseconds=-1) == end: break yield floor, ceil @classmethod def interval( cls, frame: _T_FRAMES, start: dt_datetime, end: dt_datetime, interval: int = 1, tz: Optional[TZ_EXPR] = None, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Iterable[Tuple["Arrow", "Arrow"]]: """Returns an iterator of tuples, each :class:`Arrow <arrow.arrow.Arrow>` objects, representing a series of intervals between two inputs. :param frame: The timeframe. Can be any ``datetime`` property (day, hour, minute...). :param start: A datetime expression, the start of the range. :param end: (optional) A datetime expression, the end of the range. :param interval: (optional) Time interval for the given time frame. :param tz: (optional) A timezone expression. Defaults to UTC. :param bounds: (optional) a ``str`` of either '()', '(]', '[)', or '[]' that specifies whether to include or exclude the start and end values in the intervals. '(' excludes the start, '[' includes the start, ')' excludes the end, and ']' includes the end. If the bounds are not specified, the default bound '[)' is used. :param exact: (optional) whether to have the first timespan start exactly at the time specified by ``start`` and the final interval truncated so as not to extend beyond ``end``. Supported frame values: year, quarter, month, week, day, hour, minute, second Recognized datetime expressions: - An :class:`Arrow <arrow.arrow.Arrow>` object. - A ``datetime`` object. Recognized timezone expressions: - A ``tzinfo`` object. - A ``str`` describing a timezone, similar to 'US/Pacific', or 'Europe/Berlin'. - A ``str`` in ISO 8601 style, as in '+07:00'. - A ``str``, one of the following: 'local', 'utc', 'UTC'. Usage: >>> start = datetime(2013, 5, 5, 12, 30) >>> end = datetime(2013, 5, 5, 17, 15) >>> for r in arrow.Arrow.interval('hour', start, end, 2): ... print r ... (<Arrow [2013-05-05T12:00:00+00:00]>, <Arrow [2013-05-05T13:59:59.999999+00:00]>) (<Arrow [2013-05-05T14:00:00+00:00]>, <Arrow [2013-05-05T15:59:59.999999+00:00]>) (<Arrow [2013-05-05T16:00:00+00:00]>, <Arrow [2013-05-05T17:59:59.999999+00:0]>) """ if interval < 1: raise ValueError("interval has to be a positive integer") spanRange = iter( cls.span_range(frame, start, end, tz, bounds=bounds, exact=exact) ) while True: try: intvlStart, intvlEnd = next(spanRange) for _ in range(interval - 1): try: _, intvlEnd = next(spanRange) except StopIteration: continue yield intvlStart, intvlEnd except StopIteration: return # representations def __repr__(self) -> str: return f"<{self.__class__.__name__} [{self.__str__()}]>" def __str__(self) -> str: return self._datetime.isoformat() def __format__(self, formatstr: str) -> str: if len(formatstr) > 0: return self.format(formatstr) return str(self) def __hash__(self) -> int: return self._datetime.__hash__() # attributes and properties def __getattr__(self, name: str) -> int: if name == "week": return self.isocalendar()[1] if name == "quarter": return int((self.month - 1) / self._MONTHS_PER_QUARTER) + 1 if not name.startswith("_"): value: Optional[int] = getattr(self._datetime, name, None) if value is not None: return value return cast(int, object.__getattribute__(self, name)) @property def tzinfo(self) -> dt_tzinfo: """Gets the ``tzinfo`` of the :class:`Arrow <arrow.arrow.Arrow>` object. Usage:: >>> arw=arrow.utcnow() >>> arw.tzinfo tzutc() """ # In Arrow, `_datetime` cannot be naive. return cast(dt_tzinfo, self._datetime.tzinfo) @property def datetime(self) -> dt_datetime: """Returns a datetime representation of the :class:`Arrow <arrow.arrow.Arrow>` object. Usage:: >>> arw=arrow.utcnow() >>> arw.datetime datetime.datetime(2019, 1, 24, 16, 35, 27, 276649, tzinfo=tzutc()) """ return self._datetime @property def naive(self) -> dt_datetime: """Returns a naive datetime representation of the :class:`Arrow <arrow.arrow.Arrow>` object. Usage:: >>> nairobi = arrow.now('Africa/Nairobi') >>> nairobi <Arrow [2019-01-23T19:27:12.297999+03:00]> >>> nairobi.naive datetime.datetime(2019, 1, 23, 19, 27, 12, 297999) """ return self._datetime.replace(tzinfo=None) def timestamp(self) -> float: """Returns a timestamp representation of the :class:`Arrow <arrow.arrow.Arrow>` object, in UTC time. Usage:: >>> arrow.utcnow().timestamp 1548260567 """ return self._datetime.timestamp() @property def int_timestamp(self) -> int: """Returns a timestamp representation of the :class:`Arrow <arrow.arrow.Arrow>` object, in UTC time. Usage:: >>> arrow.utcnow().int_timestamp 1548260567 """ return int(self.timestamp()) @property def float_timestamp(self) -> float: """Returns a floating-point representation of the :class:`Arrow <arrow.arrow.Arrow>` object, in UTC time. Usage:: >>> arrow.utcnow().float_timestamp 1548260516.830896 """ return self.timestamp() @property def fold(self) -> int: """ Returns the ``fold`` value of the :class:`Arrow <arrow.arrow.Arrow>` object. """ return self._datetime.fold @property def ambiguous(self) -> bool: """ Returns a boolean indicating whether the :class:`Arrow <arrow.arrow.Arrow>` object is ambiguous.""" return dateutil_tz.datetime_ambiguous(self._datetime) @property def imaginary(self) -> bool: """Indicates whether the :class: `Arrow <arrow.arrow.Arrow>` object exists in the current timezone.""" return not dateutil_tz.datetime_exists(self._datetime) # mutation and duplication. def clone(self) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object, cloned from the current one. Usage: >>> arw = arrow.utcnow() >>> cloned = arw.clone() """ return self.fromdatetime(self._datetime) def replace(self, **kwargs: Any) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object with attributes updated according to inputs. Use property names to set their value absolutely:: >>> import arrow >>> arw = arrow.utcnow() >>> arw <Arrow [2013-05-11T22:27:34.787885+00:00]> >>> arw.replace(year=2014, month=6) <Arrow [2014-06-11T22:27:34.787885+00:00]> You can also replace the timezone without conversion, using a :ref:`timezone expression <tz-expr>`:: >>> arw.replace(tzinfo=tz.tzlocal()) <Arrow [2013-05-11T22:27:34.787885-07:00]> """ absolute_kwargs = {} for key, value in kwargs.items(): if key in self._ATTRS: absolute_kwargs[key] = value elif key in ["week", "quarter"]: raise ValueError(f"Setting absolute {key} is not supported.") elif key not in ["tzinfo", "fold"]: raise ValueError(f"Unknown attribute: {key!r}.") current = self._datetime.replace(**absolute_kwargs) tzinfo = kwargs.get("tzinfo") if tzinfo is not None: tzinfo = self._get_tzinfo(tzinfo) current = current.replace(tzinfo=tzinfo) fold = kwargs.get("fold") if fold is not None: current = current.replace(fold=fold) return self.fromdatetime(current) def shift(self, **kwargs: Any) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object with attributes updated according to inputs. Use pluralized property names to relatively shift their current value: >>> import arrow >>> arw = arrow.utcnow() >>> arw <Arrow [2013-05-11T22:27:34.787885+00:00]> >>> arw.shift(years=1, months=-1) <Arrow [2014-04-11T22:27:34.787885+00:00]> Day-of-the-week relative shifting can use either Python's weekday numbers (Monday = 0, Tuesday = 1 .. Sunday = 6) or using dateutil.relativedelta's day instances (MO, TU .. SU). When using weekday numbers, the returned date will always be greater than or equal to the starting date. Using the above code (which is a Saturday) and asking it to shift to Saturday: >>> arw.shift(weekday=5) <Arrow [2013-05-11T22:27:34.787885+00:00]> While asking for a Monday: >>> arw.shift(weekday=0) <Arrow [2013-05-13T22:27:34.787885+00:00]> """ relative_kwargs = {} additional_attrs = ["weeks", "quarters", "weekday"] for key, value in kwargs.items(): if key in self._ATTRS_PLURAL or key in additional_attrs: relative_kwargs[key] = value else: supported_attr = ", ".join(self._ATTRS_PLURAL + additional_attrs) raise ValueError( f"Invalid shift time frame. Please select one of the following: {supported_attr}." ) # core datetime does not support quarters, translate to months. relative_kwargs.setdefault("months", 0) relative_kwargs["months"] += ( relative_kwargs.pop("quarters", 0) * self._MONTHS_PER_QUARTER ) current = self._datetime + relativedelta(**relative_kwargs) if not dateutil_tz.datetime_exists(current): current = dateutil_tz.resolve_imaginary(current) return self.fromdatetime(current) def to(self, tz: TZ_EXPR) -> "Arrow": """Returns a new :class:`Arrow <arrow.arrow.Arrow>` object, converted to the target timezone. :param tz: A :ref:`timezone expression <tz-expr>`. Usage:: >>> utc = arrow.utcnow() >>> utc <Arrow [2013-05-09T03:49:12.311072+00:00]> >>> utc.to('US/Pacific') <Arrow [2013-05-08T20:49:12.311072-07:00]> >>> utc.to(tz.tzlocal()) <Arrow [2013-05-08T20:49:12.311072-07:00]> >>> utc.to('-07:00') <Arrow [2013-05-08T20:49:12.311072-07:00]> >>> utc.to('local') <Arrow [2013-05-08T20:49:12.311072-07:00]> >>> utc.to('local').to('utc') <Arrow [2013-05-09T03:49:12.311072+00:00]> """ if not isinstance(tz, dt_tzinfo): tz = parser.TzinfoParser.parse(tz) dt = self._datetime.astimezone(tz) return self.__class__( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) # string output and formatting def format(self, fmt: str = "YYYY-MM-DD HH:mm:ssZZ", locale: str = "en_us") -> str: """Returns a string representation of the :class:`Arrow <arrow.arrow.Arrow>` object, formatted according to a format string. :param fmt: the format string. :param locale: the locale to format. Usage:: >>> arrow.utcnow().format('YYYY-MM-DD HH:mm:ss ZZ') '2013-05-09 03:56:47 -00:00' >>> arrow.utcnow().format('X') '1368071882' >>> arrow.utcnow().format('MMMM DD, YYYY') 'May 09, 2013' >>> arrow.utcnow().format() '2013-05-09 03:56:47 -00:00' """ return formatter.DateTimeFormatter(locale).format(self._datetime, fmt) def humanize( self, other: Union["Arrow", dt_datetime, None] = None, locale: str = "en_us", only_distance: bool = False, granularity: Union[_GRANULARITY, List[_GRANULARITY]] = "auto", ) -> str: """Returns a localized, humanized representation of a relative difference in time. :param other: (optional) an :class:`Arrow <arrow.arrow.Arrow>` or ``datetime`` object. Defaults to now in the current :class:`Arrow <arrow.arrow.Arrow>` object's timezone. :param locale: (optional) a ``str`` specifying a locale. Defaults to 'en_us'. :param only_distance: (optional) returns only time difference eg: "11 seconds" without "in" or "ago" part. :param granularity: (optional) defines the precision of the output. Set it to strings 'second', 'minute', 'hour', 'day', 'week', 'month' or 'year' or a list of any combination of these strings Usage:: >>> earlier = arrow.utcnow().shift(hours=-2) >>> earlier.humanize() '2 hours ago' >>> later = earlier.shift(hours=4) >>> later.humanize(earlier) 'in 4 hours' """ locale_name = locale locale = locales.get_locale(locale) if other is None: utc = dt_datetime.utcnow().replace(tzinfo=dateutil_tz.tzutc()) dt = utc.astimezone(self._datetime.tzinfo) elif isinstance(other, Arrow): dt = other._datetime elif isinstance(other, dt_datetime): if other.tzinfo is None: dt = other.replace(tzinfo=self._datetime.tzinfo) else: dt = other.astimezone(self._datetime.tzinfo) else: raise TypeError( f"Invalid 'other' argument of type {type(other).__name__!r}. " "Argument must be of type None, Arrow, or datetime." ) if isinstance(granularity, list) and len(granularity) == 1: granularity = granularity[0] _delta = int(round((self._datetime - dt).total_seconds())) sign = -1 if _delta < 0 else 1 delta_second = diff = abs(_delta) try: if granularity == "auto": if diff < 10: return locale.describe("now", only_distance=only_distance) if diff < self._SECS_PER_MINUTE: seconds = sign * delta_second return locale.describe( "seconds", seconds, only_distance=only_distance ) elif diff < self._SECS_PER_MINUTE * 2: return locale.describe("minute", sign, only_distance=only_distance) elif diff < self._SECS_PER_HOUR: minutes = sign * max(delta_second // self._SECS_PER_MINUTE, 2) return locale.describe( "minutes", minutes, only_distance=only_distance ) elif diff < self._SECS_PER_HOUR * 2: return locale.describe("hour", sign, only_distance=only_distance) elif diff < self._SECS_PER_DAY: hours = sign * max(delta_second // self._SECS_PER_HOUR, 2) return locale.describe("hours", hours, only_distance=only_distance) elif diff < self._SECS_PER_DAY * 2: return locale.describe("day", sign, only_distance=only_distance) elif diff < self._SECS_PER_WEEK: days = sign * max(delta_second // self._SECS_PER_DAY, 2) return locale.describe("days", days, only_distance=only_distance) elif diff < self._SECS_PER_WEEK * 2: return locale.describe("week", sign, only_distance=only_distance) elif diff < self._SECS_PER_MONTH: weeks = sign * max(delta_second // self._SECS_PER_WEEK, 2) return locale.describe("weeks", weeks, only_distance=only_distance) elif diff < self._SECS_PER_MONTH * 2: return locale.describe("month", sign, only_distance=only_distance) elif diff < self._SECS_PER_YEAR: # TODO revisit for humanization during leap years self_months = self._datetime.year * 12 + self._datetime.month other_months = dt.year * 12 + dt.month months = sign * max(abs(other_months - self_months), 2) return locale.describe( "months", months, only_distance=only_distance ) elif diff < self._SECS_PER_YEAR * 2: return locale.describe("year", sign, only_distance=only_distance) else: years = sign * max(delta_second // self._SECS_PER_YEAR, 2) return locale.describe("years", years, only_distance=only_distance) elif isinstance(granularity, str): granularity = cast(TimeFrameLiteral, granularity) # type: ignore[assignment] if granularity == "second": delta = sign * float(delta_second) if abs(delta) < 2: return locale.describe("now", only_distance=only_distance) elif granularity == "minute": delta = sign * delta_second / self._SECS_PER_MINUTE elif granularity == "hour": delta = sign * delta_second / self._SECS_PER_HOUR elif granularity == "day": delta = sign * delta_second / self._SECS_PER_DAY elif granularity == "week": delta = sign * delta_second / self._SECS_PER_WEEK elif granularity == "month": delta = sign * delta_second / self._SECS_PER_MONTH elif granularity == "year": delta = sign * delta_second / self._SECS_PER_YEAR else: raise ValueError( "Invalid level of granularity. " "Please select between 'second', 'minute', 'hour', 'day', 'week', 'month' or 'year'." ) if trunc(abs(delta)) != 1: granularity += "s" # type: ignore return locale.describe(granularity, delta, only_distance=only_distance) else: timeframes: List[Tuple[TimeFrameLiteral, float]] = [] def gather_timeframes(_delta: float, _frame: TimeFrameLiteral) -> float: if _frame in granularity: value = sign * _delta / self._SECS_MAP[_frame] _delta %= self._SECS_MAP[_frame] if trunc(abs(value)) != 1: timeframes.append( (cast(TimeFrameLiteral, _frame + "s"), value) ) else: timeframes.append((_frame, value)) return _delta delta = float(delta_second) frames: Tuple[TimeFrameLiteral, ...] = ( "year", "month", "week", "day", "hour", "minute", "second", ) for frame in frames: delta = gather_timeframes(delta, frame) if len(timeframes) < len(granularity): raise ValueError( "Invalid level of granularity. " "Please select between 'second', 'minute', 'hour', 'day', 'week', 'month' or 'year'." ) return locale.describe_multi(timeframes, only_distance=only_distance) except KeyError as e: raise ValueError( f"Humanization of the {e} granularity is not currently translated in the {locale_name!r} locale. " "Please consider making a contribution to this locale." ) # query functions def is_between( self, start: "Arrow", end: "Arrow", bounds: _BOUNDS = "()", ) -> bool: """Returns a boolean denoting whether the specified date and time is between the start and end dates and times. :param start: an :class:`Arrow <arrow.arrow.Arrow>` object. :param end: an :class:`Arrow <arrow.arrow.Arrow>` object. :param bounds: (optional) a ``str`` of either '()', '(]', '[)', or '[]' that specifies whether to include or exclude the start and end values in the range. '(' excludes the start, '[' includes the start, ')' excludes the end, and ']' includes the end. If the bounds are not specified, the default bound '()' is used. Usage:: >>> start = arrow.get(datetime(2013, 5, 5, 12, 30, 10)) >>> end = arrow.get(datetime(2013, 5, 5, 12, 30, 36)) >>> arrow.get(datetime(2013, 5, 5, 12, 30, 27)).is_between(start, end) True >>> start = arrow.get(datetime(2013, 5, 5)) >>> end = arrow.get(datetime(2013, 5, 8)) >>> arrow.get(datetime(2013, 5, 8)).is_between(start, end, '[]') True >>> start = arrow.get(datetime(2013, 5, 5)) >>> end = arrow.get(datetime(2013, 5, 8)) >>> arrow.get(datetime(2013, 5, 8)).is_between(start, end, '[)') False """ util.validate_bounds(bounds) if not isinstance(start, Arrow): raise TypeError( f"Cannot parse start date argument type of {type(start)!r}." ) if not isinstance(end, Arrow): raise TypeError(f"Cannot parse end date argument type of {type(start)!r}.") include_start = bounds[0] == "[" include_end = bounds[1] == "]" target_ts = self.float_timestamp start_ts = start.float_timestamp end_ts = end.float_timestamp return ( (start_ts <= target_ts <= end_ts) and (include_start or start_ts < target_ts) and (include_end or target_ts < end_ts) ) # datetime methods def date(self) -> date: """Returns a ``date`` object with the same year, month and day. Usage:: >>> arrow.utcnow().date() datetime.date(2019, 1, 23) """ return self._datetime.date() def time(self) -> dt_time: """Returns a ``time`` object with the same hour, minute, second, microsecond. Usage:: >>> arrow.utcnow().time() datetime.time(12, 15, 34, 68352) """ return self._datetime.time() def timetz(self) -> dt_time: """Returns a ``time`` object with the same hour, minute, second, microsecond and tzinfo. Usage:: >>> arrow.utcnow().timetz() datetime.time(12, 5, 18, 298893, tzinfo=tzutc()) """ return self._datetime.timetz() def astimezone(self, tz: Optional[dt_tzinfo]) -> dt_datetime: """Returns a ``datetime`` object, converted to the specified timezone. :param tz: a ``tzinfo`` object. Usage:: >>> pacific=arrow.now('US/Pacific') >>> nyc=arrow.now('America/New_York').tzinfo >>> pacific.astimezone(nyc) datetime.datetime(2019, 1, 20, 10, 24, 22, 328172, tzinfo=tzfile('/usr/share/zoneinfo/America/New_York')) """ return self._datetime.astimezone(tz) def utcoffset(self) -> Optional[timedelta]: """Returns a ``timedelta`` object representing the whole number of minutes difference from UTC time. Usage:: >>> arrow.now('US/Pacific').utcoffset() datetime.timedelta(-1, 57600) """ return self._datetime.utcoffset() def dst(self) -> Optional[timedelta]: """Returns the daylight savings time adjustment. Usage:: >>> arrow.utcnow().dst() datetime.timedelta(0) """ return self._datetime.dst() def timetuple(self) -> struct_time: """Returns a ``time.struct_time``, in the current timezone. Usage:: >>> arrow.utcnow().timetuple() time.struct_time(tm_year=2019, tm_mon=1, tm_mday=20, tm_hour=15, tm_min=17, tm_sec=8, tm_wday=6, tm_yday=20, tm_isdst=0) """ return self._datetime.timetuple() def utctimetuple(self) -> struct_time: """Returns a ``time.struct_time``, in UTC time. Usage:: >>> arrow.utcnow().utctimetuple() time.struct_time(tm_year=2019, tm_mon=1, tm_mday=19, tm_hour=21, tm_min=41, tm_sec=7, tm_wday=5, tm_yday=19, tm_isdst=0) """ return self._datetime.utctimetuple() def toordinal(self) -> int: """Returns the proleptic Gregorian ordinal of the date. Usage:: >>> arrow.utcnow().toordinal() 737078 """ return self._datetime.toordinal() def weekday(self) -> int: """Returns the day of the week as an integer (0-6). Usage:: >>> arrow.utcnow().weekday() 5 """ return self._datetime.weekday() def isoweekday(self) -> int: """Returns the ISO day of the week as an integer (1-7). Usage:: >>> arrow.utcnow().isoweekday() 6 """ return self._datetime.isoweekday() def isocalendar(self) -> Tuple[int, int, int]: """Returns a 3-tuple, (ISO year, ISO week number, ISO weekday). Usage:: >>> arrow.utcnow().isocalendar() (2019, 3, 6) """ return self._datetime.isocalendar() def isoformat(self, sep: str = "T", timespec: str = "auto") -> str: """Returns an ISO 8601 formatted representation of the date and time. Usage:: >>> arrow.utcnow().isoformat() '2019-01-19T18:30:52.442118+00:00' """ return self._datetime.isoformat(sep, timespec) def ctime(self) -> str: """Returns a ctime formatted representation of the date and time. Usage:: >>> arrow.utcnow().ctime() 'Sat Jan 19 18:26:50 2019' """ return self._datetime.ctime() def strftime(self, format: str) -> str: """Formats in the style of ``datetime.strftime``. :param format: the format string. Usage:: >>> arrow.utcnow().strftime('%d-%m-%Y %H:%M:%S') '23-01-2019 12:28:17' """ return self._datetime.strftime(format) def for_json(self) -> str: """Serializes for the ``for_json`` protocol of simplejson. Usage:: >>> arrow.utcnow().for_json() '2019-01-19T18:25:36.760079+00:00' """ return self.isoformat() # math def __add__(self, other: Any) -> "Arrow": if isinstance(other, (timedelta, relativedelta)): return self.fromdatetime(self._datetime + other, self._datetime.tzinfo) return NotImplemented def __radd__(self, other: Union[timedelta, relativedelta]) -> "Arrow": return self.__add__(other) @overload def __sub__(self, other: Union[timedelta, relativedelta]) -> "Arrow": pass # pragma: no cover @overload def __sub__(self, other: Union[dt_datetime, "Arrow"]) -> timedelta: pass # pragma: no cover def __sub__(self, other: Any) -> Union[timedelta, "Arrow"]: if isinstance(other, (timedelta, relativedelta)): return self.fromdatetime(self._datetime - other, self._datetime.tzinfo) elif isinstance(other, dt_datetime): return self._datetime - other elif isinstance(other, Arrow): return self._datetime - other._datetime return NotImplemented def __rsub__(self, other: Any) -> timedelta: if isinstance(other, dt_datetime): return other - self._datetime return NotImplemented # comparisons def __eq__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return False return self._datetime == self._get_datetime(other) def __ne__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return True return not self.__eq__(other) def __gt__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime > self._get_datetime(other) def __ge__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime >= self._get_datetime(other) def __lt__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime < self._get_datetime(other) def __le__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime <= self._get_datetime(other) # internal methods @staticmethod def _get_tzinfo(tz_expr: Optional[TZ_EXPR]) -> dt_tzinfo: if tz_expr is None: return dateutil_tz.tzutc() if isinstance(tz_expr, dt_tzinfo): return tz_expr else: try: return parser.TzinfoParser.parse(tz_expr) except parser.ParserError: raise ValueError(f"{tz_expr!r} not recognized as a timezone.") @classmethod def _get_datetime( cls, expr: Union["Arrow", dt_datetime, int, float, str] ) -> dt_datetime: """Get datetime object for a specified expression.""" if isinstance(expr, Arrow): return expr.datetime elif isinstance(expr, dt_datetime): return expr elif util.is_timestamp(expr): timestamp = float(expr) return cls.utcfromtimestamp(timestamp).datetime else: raise ValueError(f"{expr!r} not recognized as a datetime or timestamp.") @classmethod def _get_frames(cls, name: _T_FRAMES) -> Tuple[str, str, int]: if name in cls._ATTRS: return name, f"{name}s", 1 elif name[-1] == "s" and name[:-1] in cls._ATTRS: return name[:-1], name, 1 elif name in ["week", "weeks"]: return "week", "weeks", 1 elif name in ["quarter", "quarters"]: return "quarter", "months", 3 else: supported = ", ".join( [ "year(s)", "month(s)", "day(s)", "hour(s)", "minute(s)", "second(s)", "microsecond(s)", "week(s)", "quarter(s)", ] ) raise ValueError( f"Range or span over frame {name} not supported. Supported frames: {supported}." ) @classmethod def _get_iteration_params(cls, end: Any, limit: Optional[int]) -> Tuple[Any, int]: if end is None: if limit is None: raise ValueError("One of 'end' or 'limit' is required.") return cls.max, limit else: if limit is None: return end, sys.maxsize return end, limit @staticmethod def _is_last_day_of_month(date: "Arrow") -> bool: return date.day == calendar.monthrange(date.year, date.month)[1] Arrow.min = Arrow.fromdatetime(dt_datetime.min) Arrow.max = Arrow.fromdatetime(dt_datetime.max)
32.47907
132
0.551751
import calendar import sys from datetime import date from datetime import datetime as dt_datetime from datetime import time as dt_time from datetime import timedelta from datetime import tzinfo as dt_tzinfo from math import trunc from time import struct_time from typing import ( Any, ClassVar, Generator, Iterable, List, Mapping, Optional, Tuple, Union, cast, overload, ) from dateutil import tz as dateutil_tz from dateutil.relativedelta import relativedelta from arrow import formatter, locales, parser, util from arrow.locales import TimeFrameLiteral if sys.version_info < (3, 8): from typing_extensions import Final, Literal else: from typing import Final, Literal TZ_EXPR = Union[dt_tzinfo, str] _T_FRAMES = Literal[ "year", "years", "month", "months", "day", "days", "hour", "hours", "minute", "minutes", "second", "seconds", "microsecond", "microseconds", "week", "weeks", "quarter", "quarters", ] _BOUNDS = Literal["[)", "()", "(]", "[]"] _GRANULARITY = Literal[ "auto", "second", "minute", "hour", "day", "week", "month", "year", ] class Arrow: resolution: ClassVar[timedelta] = dt_datetime.resolution min: ClassVar["Arrow"] max: ClassVar["Arrow"] _ATTRS: Final[List[str]] = [ "year", "month", "day", "hour", "minute", "second", "microsecond", ] _ATTRS_PLURAL: Final[List[str]] = [f"{a}s" for a in _ATTRS] _MONTHS_PER_QUARTER: Final[int] = 3 _SECS_PER_MINUTE: Final[int] = 60 _SECS_PER_HOUR: Final[int] = 60 * 60 _SECS_PER_DAY: Final[int] = 60 * 60 * 24 _SECS_PER_WEEK: Final[int] = 60 * 60 * 24 * 7 _SECS_PER_MONTH: Final[float] = 60 * 60 * 24 * 30.5 _SECS_PER_YEAR: Final[int] = 60 * 60 * 24 * 365 _SECS_MAP: Final[Mapping[TimeFrameLiteral, float]] = { "second": 1.0, "minute": _SECS_PER_MINUTE, "hour": _SECS_PER_HOUR, "day": _SECS_PER_DAY, "week": _SECS_PER_WEEK, "month": _SECS_PER_MONTH, "year": _SECS_PER_YEAR, } _datetime: dt_datetime def __init__( self, year: int, month: int, day: int, hour: int = 0, minute: int = 0, second: int = 0, microsecond: int = 0, tzinfo: Optional[TZ_EXPR] = None, **kwargs: Any, ) -> None: if tzinfo is None: tzinfo = dateutil_tz.tzutc() elif ( isinstance(tzinfo, dt_tzinfo) and hasattr(tzinfo, "localize") and hasattr(tzinfo, "zone") and tzinfo.zone ): tzinfo = parser.TzinfoParser.parse(tzinfo.zone) elif isinstance(tzinfo, str): tzinfo = parser.TzinfoParser.parse(tzinfo) fold = kwargs.get("fold", 0) self._datetime = dt_datetime( year, month, day, hour, minute, second, microsecond, tzinfo, fold=fold ) @classmethod def now(cls, tzinfo: Optional[dt_tzinfo] = None) -> "Arrow": if tzinfo is None: tzinfo = dateutil_tz.tzlocal() dt = dt_datetime.now(tzinfo) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def utcnow(cls) -> "Arrow": dt = dt_datetime.now(dateutil_tz.tzutc()) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def fromtimestamp( cls, timestamp: Union[int, float, str], tzinfo: Optional[TZ_EXPR] = None, ) -> "Arrow": if tzinfo is None: tzinfo = dateutil_tz.tzlocal() elif isinstance(tzinfo, str): tzinfo = parser.TzinfoParser.parse(tzinfo) if not util.is_timestamp(timestamp): raise ValueError(f"The provided timestamp {timestamp!r} is invalid.") timestamp = util.normalize_timestamp(float(timestamp)) dt = dt_datetime.fromtimestamp(timestamp, tzinfo) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def utcfromtimestamp(cls, timestamp: Union[int, float, str]) -> "Arrow": if not util.is_timestamp(timestamp): raise ValueError(f"The provided timestamp {timestamp!r} is invalid.") timestamp = util.normalize_timestamp(float(timestamp)) dt = dt_datetime.utcfromtimestamp(timestamp) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dateutil_tz.tzutc(), fold=getattr(dt, "fold", 0), ) @classmethod def fromdatetime(cls, dt: dt_datetime, tzinfo: Optional[TZ_EXPR] = None) -> "Arrow": try: if dt.tzinfo is None: tzinfo = dateutil_tz.tzutc() else: tzinfo = dt.tzinfo except: tzinfo = dateutil_tz.tzutc() try: return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, tzinfo, fold=getattr(dt, "fold", 0), ) except: return cls( 1980, 1, 1, 0, 0, 0, 0, tzinfo, fold=0, ) @classmethod def fromdate(cls, date: date, tzinfo: Optional[TZ_EXPR] = None) -> "Arrow": if tzinfo is None: tzinfo = dateutil_tz.tzutc() return cls(date.year, date.month, date.day, tzinfo=tzinfo) @classmethod def strptime( cls, date_str: str, fmt: str, tzinfo: Optional[TZ_EXPR] = None ) -> "Arrow": dt = dt_datetime.strptime(date_str, fmt) if tzinfo is None: tzinfo = dt.tzinfo return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def fromordinal(cls, ordinal: int) -> "Arrow": util.validate_ordinal(ordinal) dt = dt_datetime.fromordinal(ordinal) return cls( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) @classmethod def range( cls, frame: _T_FRAMES, start: Union["Arrow", dt_datetime], end: Union["Arrow", dt_datetime, None] = None, tz: Optional[TZ_EXPR] = None, limit: Optional[int] = None, ) -> Generator["Arrow", None, None]: _, frame_relative, relative_steps = cls._get_frames(frame) tzinfo = cls._get_tzinfo(start.tzinfo if tz is None else tz) start = cls._get_datetime(start).replace(tzinfo=tzinfo) end, limit = cls._get_iteration_params(end, limit) end = cls._get_datetime(end).replace(tzinfo=tzinfo) current = cls.fromdatetime(start) original_day = start.day day_is_clipped = False i = 0 while current <= end and i < limit: i += 1 yield current values = [getattr(current, f) for f in cls._ATTRS] current = cls(*values, tzinfo=tzinfo).shift( **{frame_relative: relative_steps} ) if frame in ["month", "quarter", "year"] and current.day < original_day: day_is_clipped = True if day_is_clipped and not cls._is_last_day_of_month(current): current = current.replace(day=original_day) def span( self, frame: _T_FRAMES, count: int = 1, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Tuple["Arrow", "Arrow"]: util.validate_bounds(bounds) frame_absolute, frame_relative, relative_steps = self._get_frames(frame) if frame_absolute == "week": attr = "day" elif frame_absolute == "quarter": attr = "month" else: attr = frame_absolute floor = self if not exact: index = self._ATTRS.index(attr) frames = self._ATTRS[: index + 1] values = [getattr(self, f) for f in frames] for _ in range(3 - len(values)): values.append(1) floor = self.__class__(*values, tzinfo=self.tzinfo) if frame_absolute == "week": floor = floor.shift(days=-(self.isoweekday() - 1)) elif frame_absolute == "quarter": floor = floor.shift(months=-((self.month - 1) % 3)) ceil = floor.shift(**{frame_relative: count * relative_steps}) if bounds[0] == "(": floor = floor.shift(microseconds=+1) if bounds[1] == ")": ceil = ceil.shift(microseconds=-1) return floor, ceil def floor(self, frame: _T_FRAMES) -> "Arrow": return self.span(frame)[0] def ceil(self, frame: _T_FRAMES) -> "Arrow": return self.span(frame)[1] @classmethod def span_range( cls, frame: _T_FRAMES, start: dt_datetime, end: dt_datetime, tz: Optional[TZ_EXPR] = None, limit: Optional[int] = None, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Iterable[Tuple["Arrow", "Arrow"]]: tzinfo = cls._get_tzinfo(start.tzinfo if tz is None else tz) start = cls.fromdatetime(start, tzinfo).span(frame, exact=exact)[0] end = cls.fromdatetime(end, tzinfo) _range = cls.range(frame, start, end, tz, limit) if not exact: for r in _range: yield r.span(frame, bounds=bounds, exact=exact) for r in _range: floor, ceil = r.span(frame, bounds=bounds, exact=exact) if ceil > end: ceil = end if bounds[1] == ")": ceil += relativedelta(microseconds=-1) if floor == end: break elif floor + relativedelta(microseconds=-1) == end: break yield floor, ceil @classmethod def interval( cls, frame: _T_FRAMES, start: dt_datetime, end: dt_datetime, interval: int = 1, tz: Optional[TZ_EXPR] = None, bounds: _BOUNDS = "[)", exact: bool = False, ) -> Iterable[Tuple["Arrow", "Arrow"]]: if interval < 1: raise ValueError("interval has to be a positive integer") spanRange = iter( cls.span_range(frame, start, end, tz, bounds=bounds, exact=exact) ) while True: try: intvlStart, intvlEnd = next(spanRange) for _ in range(interval - 1): try: _, intvlEnd = next(spanRange) except StopIteration: continue yield intvlStart, intvlEnd except StopIteration: return def __repr__(self) -> str: return f"<{self.__class__.__name__} [{self.__str__()}]>" def __str__(self) -> str: return self._datetime.isoformat() def __format__(self, formatstr: str) -> str: if len(formatstr) > 0: return self.format(formatstr) return str(self) def __hash__(self) -> int: return self._datetime.__hash__() def __getattr__(self, name: str) -> int: if name == "week": return self.isocalendar()[1] if name == "quarter": return int((self.month - 1) / self._MONTHS_PER_QUARTER) + 1 if not name.startswith("_"): value: Optional[int] = getattr(self._datetime, name, None) if value is not None: return value return cast(int, object.__getattribute__(self, name)) @property def tzinfo(self) -> dt_tzinfo: return cast(dt_tzinfo, self._datetime.tzinfo) @property def datetime(self) -> dt_datetime: return self._datetime @property def naive(self) -> dt_datetime: return self._datetime.replace(tzinfo=None) def timestamp(self) -> float: return self._datetime.timestamp() @property def int_timestamp(self) -> int: return int(self.timestamp()) @property def float_timestamp(self) -> float: return self.timestamp() @property def fold(self) -> int: return self._datetime.fold @property def ambiguous(self) -> bool: return dateutil_tz.datetime_ambiguous(self._datetime) @property def imaginary(self) -> bool: return not dateutil_tz.datetime_exists(self._datetime) def clone(self) -> "Arrow": return self.fromdatetime(self._datetime) def replace(self, **kwargs: Any) -> "Arrow": absolute_kwargs = {} for key, value in kwargs.items(): if key in self._ATTRS: absolute_kwargs[key] = value elif key in ["week", "quarter"]: raise ValueError(f"Setting absolute {key} is not supported.") elif key not in ["tzinfo", "fold"]: raise ValueError(f"Unknown attribute: {key!r}.") current = self._datetime.replace(**absolute_kwargs) tzinfo = kwargs.get("tzinfo") if tzinfo is not None: tzinfo = self._get_tzinfo(tzinfo) current = current.replace(tzinfo=tzinfo) fold = kwargs.get("fold") if fold is not None: current = current.replace(fold=fold) return self.fromdatetime(current) def shift(self, **kwargs: Any) -> "Arrow": relative_kwargs = {} additional_attrs = ["weeks", "quarters", "weekday"] for key, value in kwargs.items(): if key in self._ATTRS_PLURAL or key in additional_attrs: relative_kwargs[key] = value else: supported_attr = ", ".join(self._ATTRS_PLURAL + additional_attrs) raise ValueError( f"Invalid shift time frame. Please select one of the following: {supported_attr}." ) relative_kwargs.setdefault("months", 0) relative_kwargs["months"] += ( relative_kwargs.pop("quarters", 0) * self._MONTHS_PER_QUARTER ) current = self._datetime + relativedelta(**relative_kwargs) if not dateutil_tz.datetime_exists(current): current = dateutil_tz.resolve_imaginary(current) return self.fromdatetime(current) def to(self, tz: TZ_EXPR) -> "Arrow": if not isinstance(tz, dt_tzinfo): tz = parser.TzinfoParser.parse(tz) dt = self._datetime.astimezone(tz) return self.__class__( dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second, dt.microsecond, dt.tzinfo, fold=getattr(dt, "fold", 0), ) def format(self, fmt: str = "YYYY-MM-DD HH:mm:ssZZ", locale: str = "en_us") -> str: return formatter.DateTimeFormatter(locale).format(self._datetime, fmt) def humanize( self, other: Union["Arrow", dt_datetime, None] = None, locale: str = "en_us", only_distance: bool = False, granularity: Union[_GRANULARITY, List[_GRANULARITY]] = "auto", ) -> str: locale_name = locale locale = locales.get_locale(locale) if other is None: utc = dt_datetime.utcnow().replace(tzinfo=dateutil_tz.tzutc()) dt = utc.astimezone(self._datetime.tzinfo) elif isinstance(other, Arrow): dt = other._datetime elif isinstance(other, dt_datetime): if other.tzinfo is None: dt = other.replace(tzinfo=self._datetime.tzinfo) else: dt = other.astimezone(self._datetime.tzinfo) else: raise TypeError( f"Invalid 'other' argument of type {type(other).__name__!r}. " "Argument must be of type None, Arrow, or datetime." ) if isinstance(granularity, list) and len(granularity) == 1: granularity = granularity[0] _delta = int(round((self._datetime - dt).total_seconds())) sign = -1 if _delta < 0 else 1 delta_second = diff = abs(_delta) try: if granularity == "auto": if diff < 10: return locale.describe("now", only_distance=only_distance) if diff < self._SECS_PER_MINUTE: seconds = sign * delta_second return locale.describe( "seconds", seconds, only_distance=only_distance ) elif diff < self._SECS_PER_MINUTE * 2: return locale.describe("minute", sign, only_distance=only_distance) elif diff < self._SECS_PER_HOUR: minutes = sign * max(delta_second // self._SECS_PER_MINUTE, 2) return locale.describe( "minutes", minutes, only_distance=only_distance ) elif diff < self._SECS_PER_HOUR * 2: return locale.describe("hour", sign, only_distance=only_distance) elif diff < self._SECS_PER_DAY: hours = sign * max(delta_second // self._SECS_PER_HOUR, 2) return locale.describe("hours", hours, only_distance=only_distance) elif diff < self._SECS_PER_DAY * 2: return locale.describe("day", sign, only_distance=only_distance) elif diff < self._SECS_PER_WEEK: days = sign * max(delta_second // self._SECS_PER_DAY, 2) return locale.describe("days", days, only_distance=only_distance) elif diff < self._SECS_PER_WEEK * 2: return locale.describe("week", sign, only_distance=only_distance) elif diff < self._SECS_PER_MONTH: weeks = sign * max(delta_second // self._SECS_PER_WEEK, 2) return locale.describe("weeks", weeks, only_distance=only_distance) elif diff < self._SECS_PER_MONTH * 2: return locale.describe("month", sign, only_distance=only_distance) elif diff < self._SECS_PER_YEAR: self_months = self._datetime.year * 12 + self._datetime.month other_months = dt.year * 12 + dt.month months = sign * max(abs(other_months - self_months), 2) return locale.describe( "months", months, only_distance=only_distance ) elif diff < self._SECS_PER_YEAR * 2: return locale.describe("year", sign, only_distance=only_distance) else: years = sign * max(delta_second // self._SECS_PER_YEAR, 2) return locale.describe("years", years, only_distance=only_distance) elif isinstance(granularity, str): granularity = cast(TimeFrameLiteral, granularity) if granularity == "second": delta = sign * float(delta_second) if abs(delta) < 2: return locale.describe("now", only_distance=only_distance) elif granularity == "minute": delta = sign * delta_second / self._SECS_PER_MINUTE elif granularity == "hour": delta = sign * delta_second / self._SECS_PER_HOUR elif granularity == "day": delta = sign * delta_second / self._SECS_PER_DAY elif granularity == "week": delta = sign * delta_second / self._SECS_PER_WEEK elif granularity == "month": delta = sign * delta_second / self._SECS_PER_MONTH elif granularity == "year": delta = sign * delta_second / self._SECS_PER_YEAR else: raise ValueError( "Invalid level of granularity. " "Please select between 'second', 'minute', 'hour', 'day', 'week', 'month' or 'year'." ) if trunc(abs(delta)) != 1: granularity += "s" return locale.describe(granularity, delta, only_distance=only_distance) else: timeframes: List[Tuple[TimeFrameLiteral, float]] = [] def gather_timeframes(_delta: float, _frame: TimeFrameLiteral) -> float: if _frame in granularity: value = sign * _delta / self._SECS_MAP[_frame] _delta %= self._SECS_MAP[_frame] if trunc(abs(value)) != 1: timeframes.append( (cast(TimeFrameLiteral, _frame + "s"), value) ) else: timeframes.append((_frame, value)) return _delta delta = float(delta_second) frames: Tuple[TimeFrameLiteral, ...] = ( "year", "month", "week", "day", "hour", "minute", "second", ) for frame in frames: delta = gather_timeframes(delta, frame) if len(timeframes) < len(granularity): raise ValueError( "Invalid level of granularity. " "Please select between 'second', 'minute', 'hour', 'day', 'week', 'month' or 'year'." ) return locale.describe_multi(timeframes, only_distance=only_distance) except KeyError as e: raise ValueError( f"Humanization of the {e} granularity is not currently translated in the {locale_name!r} locale. " "Please consider making a contribution to this locale." ) def is_between( self, start: "Arrow", end: "Arrow", bounds: _BOUNDS = "()", ) -> bool: util.validate_bounds(bounds) if not isinstance(start, Arrow): raise TypeError( f"Cannot parse start date argument type of {type(start)!r}." ) if not isinstance(end, Arrow): raise TypeError(f"Cannot parse end date argument type of {type(start)!r}.") include_start = bounds[0] == "[" include_end = bounds[1] == "]" target_ts = self.float_timestamp start_ts = start.float_timestamp end_ts = end.float_timestamp return ( (start_ts <= target_ts <= end_ts) and (include_start or start_ts < target_ts) and (include_end or target_ts < end_ts) ) def date(self) -> date: return self._datetime.date() def time(self) -> dt_time: return self._datetime.time() def timetz(self) -> dt_time: return self._datetime.timetz() def astimezone(self, tz: Optional[dt_tzinfo]) -> dt_datetime: return self._datetime.astimezone(tz) def utcoffset(self) -> Optional[timedelta]: return self._datetime.utcoffset() def dst(self) -> Optional[timedelta]: return self._datetime.dst() def timetuple(self) -> struct_time: return self._datetime.timetuple() def utctimetuple(self) -> struct_time: return self._datetime.utctimetuple() def toordinal(self) -> int: return self._datetime.toordinal() def weekday(self) -> int: return self._datetime.weekday() def isoweekday(self) -> int: return self._datetime.isoweekday() def isocalendar(self) -> Tuple[int, int, int]: return self._datetime.isocalendar() def isoformat(self, sep: str = "T", timespec: str = "auto") -> str: return self._datetime.isoformat(sep, timespec) def ctime(self) -> str: return self._datetime.ctime() def strftime(self, format: str) -> str: return self._datetime.strftime(format) def for_json(self) -> str: return self.isoformat() def __add__(self, other: Any) -> "Arrow": if isinstance(other, (timedelta, relativedelta)): return self.fromdatetime(self._datetime + other, self._datetime.tzinfo) return NotImplemented def __radd__(self, other: Union[timedelta, relativedelta]) -> "Arrow": return self.__add__(other) @overload def __sub__(self, other: Union[timedelta, relativedelta]) -> "Arrow": pass @overload def __sub__(self, other: Union[dt_datetime, "Arrow"]) -> timedelta: pass def __sub__(self, other: Any) -> Union[timedelta, "Arrow"]: if isinstance(other, (timedelta, relativedelta)): return self.fromdatetime(self._datetime - other, self._datetime.tzinfo) elif isinstance(other, dt_datetime): return self._datetime - other elif isinstance(other, Arrow): return self._datetime - other._datetime return NotImplemented def __rsub__(self, other: Any) -> timedelta: if isinstance(other, dt_datetime): return other - self._datetime return NotImplemented def __eq__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return False return self._datetime == self._get_datetime(other) def __ne__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return True return not self.__eq__(other) def __gt__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime > self._get_datetime(other) def __ge__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime >= self._get_datetime(other) def __lt__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime < self._get_datetime(other) def __le__(self, other: Any) -> bool: if not isinstance(other, (Arrow, dt_datetime)): return NotImplemented return self._datetime <= self._get_datetime(other) @staticmethod def _get_tzinfo(tz_expr: Optional[TZ_EXPR]) -> dt_tzinfo: if tz_expr is None: return dateutil_tz.tzutc() if isinstance(tz_expr, dt_tzinfo): return tz_expr else: try: return parser.TzinfoParser.parse(tz_expr) except parser.ParserError: raise ValueError(f"{tz_expr!r} not recognized as a timezone.") @classmethod def _get_datetime( cls, expr: Union["Arrow", dt_datetime, int, float, str] ) -> dt_datetime: if isinstance(expr, Arrow): return expr.datetime elif isinstance(expr, dt_datetime): return expr elif util.is_timestamp(expr): timestamp = float(expr) return cls.utcfromtimestamp(timestamp).datetime else: raise ValueError(f"{expr!r} not recognized as a datetime or timestamp.") @classmethod def _get_frames(cls, name: _T_FRAMES) -> Tuple[str, str, int]: if name in cls._ATTRS: return name, f"{name}s", 1 elif name[-1] == "s" and name[:-1] in cls._ATTRS: return name[:-1], name, 1 elif name in ["week", "weeks"]: return "week", "weeks", 1 elif name in ["quarter", "quarters"]: return "quarter", "months", 3 else: supported = ", ".join( [ "year(s)", "month(s)", "day(s)", "hour(s)", "minute(s)", "second(s)", "microsecond(s)", "week(s)", "quarter(s)", ] ) raise ValueError( f"Range or span over frame {name} not supported. Supported frames: {supported}." ) @classmethod def _get_iteration_params(cls, end: Any, limit: Optional[int]) -> Tuple[Any, int]: if end is None: if limit is None: raise ValueError("One of 'end' or 'limit' is required.") return cls.max, limit else: if limit is None: return end, sys.maxsize return end, limit @staticmethod def _is_last_day_of_month(date: "Arrow") -> bool: return date.day == calendar.monthrange(date.year, date.month)[1] Arrow.min = Arrow.fromdatetime(dt_datetime.min) Arrow.max = Arrow.fromdatetime(dt_datetime.max)
true
true
f717aff2b542befa776d914dfd5cc573bbed5908
2,681
py
Python
make/photon/prepare/main.py
gaohuag/harbor
ebe2df06c84c3b6be33901e71d9bf14e80a4c003
[ "Apache-2.0" ]
1
2019-10-24T07:11:38.000Z
2019-10-24T07:11:38.000Z
make/photon/prepare/main.py
gaohuag/harbor
ebe2df06c84c3b6be33901e71d9bf14e80a4c003
[ "Apache-2.0" ]
null
null
null
make/photon/prepare/main.py
gaohuag/harbor
ebe2df06c84c3b6be33901e71d9bf14e80a4c003
[ "Apache-2.0" ]
null
null
null
# pylint: disable=no-value-for-parameter import click from utils.misc import delfile from utils.configs import validate, parse_yaml_config from utils.cert import prepare_ca, SSL_CERT_KEY_PATH, SSL_CERT_PATH, get_secret_key from utils.db import prepare_db from utils.jobservice import prepare_job_service from utils.registry import prepare_registry from utils.registry_ctl import prepare_registry_ctl from utils.core import prepare_core from utils.notary import prepare_notary from utils.log import prepare_log_configs from utils.clair import prepare_clair from utils.clair_adapter import prepare_clair_adapter from utils.chart import prepare_chartmuseum from utils.docker_compose import prepare_docker_compose from utils.nginx import prepare_nginx, nginx_confd_dir from utils.redis import prepare_redis from g import (config_dir, input_config_path, private_key_pem_path, root_crt_path, secret_key_dir, old_private_key_pem_path, old_crt_path) # Main function @click.command() @click.option('--conf', default=input_config_path, help="the path of Harbor configuration file") @click.option('--with-notary', is_flag=True, help="the Harbor instance is to be deployed with notary") @click.option('--with-clair', is_flag=True, help="the Harbor instance is to be deployed with clair") @click.option('--with-chartmuseum', is_flag=True, help="the Harbor instance is to be deployed with chart repository supporting") def main(conf, with_notary, with_clair, with_chartmuseum): delfile(config_dir) config_dict = parse_yaml_config(conf, with_notary=with_notary, with_clair=with_clair, with_chartmuseum=with_chartmuseum) validate(config_dict, notary_mode=with_notary) prepare_log_configs(config_dict) prepare_nginx(config_dict) prepare_core(config_dict, with_notary=with_notary, with_clair=with_clair, with_chartmuseum=with_chartmuseum) prepare_registry(config_dict) prepare_registry_ctl(config_dict) prepare_db(config_dict) prepare_job_service(config_dict) prepare_redis(config_dict) get_secret_key(secret_key_dir) # If Customized cert enabled prepare_ca( private_key_pem_path=private_key_pem_path, root_crt_path=root_crt_path, old_private_key_pem_path=old_private_key_pem_path, old_crt_path=old_crt_path) if with_notary: prepare_notary(config_dict, nginx_confd_dir, SSL_CERT_PATH, SSL_CERT_KEY_PATH) if with_clair: prepare_clair(config_dict) prepare_clair_adapter(config_dict) if with_chartmuseum: prepare_chartmuseum(config_dict) prepare_docker_compose(config_dict, with_clair, with_notary, with_chartmuseum) if __name__ == '__main__': main()
40.621212
128
0.806043
import click from utils.misc import delfile from utils.configs import validate, parse_yaml_config from utils.cert import prepare_ca, SSL_CERT_KEY_PATH, SSL_CERT_PATH, get_secret_key from utils.db import prepare_db from utils.jobservice import prepare_job_service from utils.registry import prepare_registry from utils.registry_ctl import prepare_registry_ctl from utils.core import prepare_core from utils.notary import prepare_notary from utils.log import prepare_log_configs from utils.clair import prepare_clair from utils.clair_adapter import prepare_clair_adapter from utils.chart import prepare_chartmuseum from utils.docker_compose import prepare_docker_compose from utils.nginx import prepare_nginx, nginx_confd_dir from utils.redis import prepare_redis from g import (config_dir, input_config_path, private_key_pem_path, root_crt_path, secret_key_dir, old_private_key_pem_path, old_crt_path) @click.command() @click.option('--conf', default=input_config_path, help="the path of Harbor configuration file") @click.option('--with-notary', is_flag=True, help="the Harbor instance is to be deployed with notary") @click.option('--with-clair', is_flag=True, help="the Harbor instance is to be deployed with clair") @click.option('--with-chartmuseum', is_flag=True, help="the Harbor instance is to be deployed with chart repository supporting") def main(conf, with_notary, with_clair, with_chartmuseum): delfile(config_dir) config_dict = parse_yaml_config(conf, with_notary=with_notary, with_clair=with_clair, with_chartmuseum=with_chartmuseum) validate(config_dict, notary_mode=with_notary) prepare_log_configs(config_dict) prepare_nginx(config_dict) prepare_core(config_dict, with_notary=with_notary, with_clair=with_clair, with_chartmuseum=with_chartmuseum) prepare_registry(config_dict) prepare_registry_ctl(config_dict) prepare_db(config_dict) prepare_job_service(config_dict) prepare_redis(config_dict) get_secret_key(secret_key_dir) prepare_ca( private_key_pem_path=private_key_pem_path, root_crt_path=root_crt_path, old_private_key_pem_path=old_private_key_pem_path, old_crt_path=old_crt_path) if with_notary: prepare_notary(config_dict, nginx_confd_dir, SSL_CERT_PATH, SSL_CERT_KEY_PATH) if with_clair: prepare_clair(config_dict) prepare_clair_adapter(config_dict) if with_chartmuseum: prepare_chartmuseum(config_dict) prepare_docker_compose(config_dict, with_clair, with_notary, with_chartmuseum) if __name__ == '__main__': main()
true
true
f717b0f7dc65a5798b28471eda50f0d297865349
3,637
py
Python
dendropy/dataio/nexmlyielder.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
dendropy/dataio/nexmlyielder.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
dendropy/dataio/nexmlyielder.py
EnjoyLifeFund/macHighSierra-py36-pkgs
5668b5785296b314ea1321057420bcd077dba9ea
[ "BSD-3-Clause", "BSD-2-Clause", "MIT" ]
null
null
null
#! /usr/bin/env python ############################################################################## ## DendroPy Phylogenetic Computing Library. ## ## Copyright 2010-2015 Jeet Sukumaran and Mark T. Holder. ## All rights reserved. ## ## See "LICENSE.rst" for terms and conditions of usage. ## ## If you use this work or any portion thereof in published work, ## please cite it as: ## ## Sukumaran, J. and M. T. Holder. 2010. DendroPy: a Python library ## for phylogenetic computing. Bioinformatics 26: 1569-1571. ## ############################################################################## """ Implementation of NEXML-schema tree iterator. """ import sys if not (sys.version_info.major >= 3 and sys.version_info.minor >= 4): from dendropy.utility.filesys import pre_py34_open as open from dendropy.dataio import ioservice from dendropy.dataio import nexmlreader from dendropy.dataio import xmlprocessing class NexmlTreeDataYielder( ioservice.TreeDataYielder, nexmlreader.NexmlReader): def __init__(self, files=None, taxon_namespace=None, tree_type=None, **kwargs): """ Parameters ---------- files : iterable of sources Iterable of sources, which can either be strings specifying file paths or file-like objects open for reading. If a source element is a string, then it is assumed to be a path to a file. Otherwise, the source is assumed to be a file-like object. taxon_namespace : |TaxonNamespace| instance The operational taxonomic unit concept namespace to use to manage taxon definitions. \*\*kwargs : keyword arguments These will be passed directly to the base `nexmlreader.NexusReader` class. See `nexmlreader.NexusReader` for details. """ ioservice.TreeDataYielder.__init__(self, files=files, taxon_namespace=taxon_namespace, tree_type=tree_type) nexmlreader.NexmlReader.__init__(self, **kwargs) self.attached_taxon_namespace = self.taxon_namespace ########################################################################### ## Implementation of DataYielder interface def _yield_items_from_stream(self, stream): xml_doc = xmlprocessing.XmlDocument(file_obj=stream, subelement_factory=self._subelement_factory) self._namespace_registry = xml_doc.namespace_registry xml_root = xml_doc.root self._parse_taxon_namespaces(xml_root) tree_parser = nexmlreader._NexmlTreeParser( id_taxon_map=self._id_taxon_map, annotations_processor_fn=self._parse_annotations, ) for trees_idx, trees_element in enumerate(xml_root.iter_trees()): trees_id = trees_element.get('id', "Trees" + str(trees_idx)) trees_label = trees_element.get('label', None) otus_id = trees_element.get('otus', None) if otus_id is None: raise Exception("Taxa block not specified for trees block '{}'".format(otus_id)) taxon_namespace = self._id_taxon_namespace_map.get(otus_id, None) if not taxon_namespace: raise Exception("Tree block '{}': Taxa block '{}' not found".format(trees_id, otus_id)) for tree_element in trees_element.findall_tree(): tree_obj = self.tree_factory() tree_parser.build_tree(tree_obj, tree_element, otus_id) yield tree_obj
40.865169
103
0.607369
true
true
f717b173935cbd0d7770fed38f1c470436bb75f4
970
py
Python
tests/testNextRule.py
andersonjwan/tltk-mtl-ext
a4f0314df6c55d532b7f9d169b6d08bff3811ef3
[ "BSD-3-Clause" ]
1
2020-12-11T21:32:46.000Z
2020-12-11T21:32:46.000Z
tests/testNextRule.py
andersonjwan/tltk-mtl-ext
a4f0314df6c55d532b7f9d169b6d08bff3811ef3
[ "BSD-3-Clause" ]
null
null
null
tests/testNextRule.py
andersonjwan/tltk-mtl-ext
a4f0314df6c55d532b7f9d169b6d08bff3811ef3
[ "BSD-3-Clause" ]
null
null
null
import tltk_mtl as MTL import tltk_mtl_ext as MTLE def testNextRule(): preds = {} preds['pred1'] = MTL.Predicate('pred1', 1, 2) preds['pred2'] = MTL.Predicate('pred2', 2, 4) preds['pred3'] = MTL.Predicate('pred3', 4, 8) assert isinstance(MTLE.parse_mtl('next pred1', preds), MTL.Next), printFail(1) printPass(1) assert isinstance(MTLE.parse_mtl('next X pred1', preds), MTL.Next), printFail(2) printPass(2) assert isinstance(MTLE.parse_mtl('next (pred1 && pred2)', preds), MTL.Next), printFail(3) printPass(3) assert isinstance(MTLE.parse_mtl('X (pred1 && X X X pred3)', preds), MTL.Next), printFail(4) printPass(4) assert isinstance(MTLE.parse_mtl('X X X X X X X X (pred1 & pred2)', preds), MTL.Next), printFail(5) printPass(5) def printPass(test_num): print("%-20s Assertion No. %02d PASSED." % ('[Next Rule Test]', test_num)) def printFail(test_num): return '%-20s Assertion No. %02d FAILED' % ('[Next Rule Test]', test_num) testNextRule()
34.642857
100
0.689691
import tltk_mtl as MTL import tltk_mtl_ext as MTLE def testNextRule(): preds = {} preds['pred1'] = MTL.Predicate('pred1', 1, 2) preds['pred2'] = MTL.Predicate('pred2', 2, 4) preds['pred3'] = MTL.Predicate('pred3', 4, 8) assert isinstance(MTLE.parse_mtl('next pred1', preds), MTL.Next), printFail(1) printPass(1) assert isinstance(MTLE.parse_mtl('next X pred1', preds), MTL.Next), printFail(2) printPass(2) assert isinstance(MTLE.parse_mtl('next (pred1 && pred2)', preds), MTL.Next), printFail(3) printPass(3) assert isinstance(MTLE.parse_mtl('X (pred1 && X X X pred3)', preds), MTL.Next), printFail(4) printPass(4) assert isinstance(MTLE.parse_mtl('X X X X X X X X (pred1 & pred2)', preds), MTL.Next), printFail(5) printPass(5) def printPass(test_num): print("%-20s Assertion No. %02d PASSED." % ('[Next Rule Test]', test_num)) def printFail(test_num): return '%-20s Assertion No. %02d FAILED' % ('[Next Rule Test]', test_num) testNextRule()
true
true
f717b362c8813b8acd724ce809248566680836c5
591
py
Python
data & result/history.py
robinzixuan/Dialog_Act_Bert_Classification
014cc8df0545e5bf85a22127e63e8490f3aa9012
[ "MIT" ]
null
null
null
data & result/history.py
robinzixuan/Dialog_Act_Bert_Classification
014cc8df0545e5bf85a22127e63e8490f3aa9012
[ "MIT" ]
null
null
null
data & result/history.py
robinzixuan/Dialog_Act_Bert_Classification
014cc8df0545e5bf85a22127e63e8490f3aa9012
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun Jun 30 20:49:32 2019 @author: rluo """ import keras import matplotlib.pyplot as plt from keras.models import load_model import pickle history = pickle.load(open('history.p','rb')) plt.plot(history['loss']) #plt.plot(history['val_loss']) plt.title('model loss') plt.ylabel('loss') plt.xlabel('Epoch') plt.legend(['train', 'test'], loc='upper left'); plt.plot(history['acc']) #plt.plot(history['val_acc']) plt.title('model accuracy') plt.ylabel('acc') plt.xlabel('Epoch') plt.legend(['train', 'test'], loc='upper left');
19.7
48
0.683587
import keras import matplotlib.pyplot as plt from keras.models import load_model import pickle history = pickle.load(open('history.p','rb')) plt.plot(history['loss']) plt.title('model loss') plt.ylabel('loss') plt.xlabel('Epoch') plt.legend(['train', 'test'], loc='upper left'); plt.plot(history['acc']) plt.title('model accuracy') plt.ylabel('acc') plt.xlabel('Epoch') plt.legend(['train', 'test'], loc='upper left');
true
true
f717b38081c7eaa8fa44370f116ae44f650ba3cb
689
py
Python
examples/seq2seq/paraphrasing/predict.py
liorshk/simpletransformers
226cf4d11edf5157c1beafcc44aaa78f65ccc985
[ "Apache-2.0" ]
3,151
2019-10-05T11:14:44.000Z
2022-03-31T17:02:54.000Z
examples/seq2seq/paraphrasing/predict.py
liorshk/simpletransformers
226cf4d11edf5157c1beafcc44aaa78f65ccc985
[ "Apache-2.0" ]
1,165
2019-10-05T14:48:55.000Z
2022-03-31T11:12:58.000Z
examples/seq2seq/paraphrasing/predict.py
liorshk/simpletransformers
226cf4d11edf5157c1beafcc44aaa78f65ccc985
[ "Apache-2.0" ]
739
2019-10-06T15:11:54.000Z
2022-03-28T11:07:36.000Z
import logging from simpletransformers.seq2seq import Seq2SeqModel logging.basicConfig(level=logging.INFO) transformers_logger = logging.getLogger("transformers") transformers_logger.setLevel(logging.ERROR) model = Seq2SeqModel(encoder_decoder_type="bart", encoder_decoder_name="outputs") while True: original = input("Enter text to paraphrase: ") to_predict = [original] preds = model.predict(to_predict) print("---------------------------------------------------------") print(original) print() print("Predictions >>>") for pred in preds[0]: print(pred) print("---------------------------------------------------------") print()
23.758621
81
0.593614
import logging from simpletransformers.seq2seq import Seq2SeqModel logging.basicConfig(level=logging.INFO) transformers_logger = logging.getLogger("transformers") transformers_logger.setLevel(logging.ERROR) model = Seq2SeqModel(encoder_decoder_type="bart", encoder_decoder_name="outputs") while True: original = input("Enter text to paraphrase: ") to_predict = [original] preds = model.predict(to_predict) print("---------------------------------------------------------") print(original) print() print("Predictions >>>") for pred in preds[0]: print(pred) print("---------------------------------------------------------") print()
true
true
f717b4377f6a9a74d53a66640675d4744b9f7a3a
2,171
py
Python
app.py
dennisfarmer/Ensemble-Video-Downloader
fa3cb334f9140b5fecdfcb1c2ee6020743a484a7
[ "MIT" ]
null
null
null
app.py
dennisfarmer/Ensemble-Video-Downloader
fa3cb334f9140b5fecdfcb1c2ee6020743a484a7
[ "MIT" ]
1
2021-08-02T19:08:13.000Z
2021-08-03T02:35:49.000Z
app.py
dennisfarmer/Ensemble-Video-Downloader
fa3cb334f9140b5fecdfcb1c2ee6020743a484a7
[ "MIT" ]
1
2021-08-03T04:44:13.000Z
2021-08-03T04:44:13.000Z
from PyQt5.QtWidgets import QApplication,QWidget,QTextEdit,QVBoxLayout,QPushButton from link_converter import convert_url import pyperclip import sys class LinkConverter(QWidget): def __init__(self,parent=None): super().__init__(parent) self.setWindowTitle("Ensemblevideo Link Converter - Dennis Farmer") self.resize(600,300) self.textEdit = QTextEdit() self.btnPress1 = QPushButton("Convert to Download Link") self.btnPress3 = QPushButton("Clear Field Contents") self.btnPress2 = QPushButton("Copy to Clipboard") layout = QVBoxLayout() layout.addWidget(self.textEdit) layout.addWidget(self.btnPress1) layout.addWidget(self.btnPress2) layout.addWidget(self.btnPress3) self.setLayout(layout) self.textEdit.setPlainText("https://cloud.ensemblevideo.com/hapi/v1/contents/e5ee9fb7-8a1e-48f2-9e4d-2ac8a77c2478/launch?embedAsThumbnail=false&displayTitle=True&displaySharing=False&autoplay=False&showCaptions=False&hideControls=True&audioPreviewImage=False&displayEmbedCode=False&displayAttachments=True&displayLinks=True&displayCredits=False&displayVideoDuration=False&displayNotes=True&displayCaptionSearch=True&displayMetaData=false&displayDownloadIcon=False&displayViewersReport=False&displayAxdxs=False&idn_content=e5ee9fb7-8a1e-48f2-9e4d-2ac8a77c2478&idn_init=False&idn_sig=62NO%2FKLkev7ZUVK696RTOVkjP1Q%3D") self.btnPress1.clicked.connect(self.btnPress1_Clicked) self.btnPress2.clicked.connect(self.btnPress2_Clicked) self.btnPress3.clicked.connect(self.btnPress3_Clicked) def btnPress1_Clicked(self): old_url = self.textEdit.toPlainText() new_url = convert_url(old_url) self.textEdit.setPlainText(new_url) def btnPress2_Clicked(self): url = self.textEdit.toPlainText() pyperclip.copy(url) def btnPress3_Clicked(self): self.textEdit.setPlainText("") def Start(): lc = LinkConverter() lc.show() return lc def main(): app = QApplication(sys.argv) win = Start() app.exec_() #sys.exit(app.exec_()) if __name__ == '__main__': main()
39.472727
624
0.738369
from PyQt5.QtWidgets import QApplication,QWidget,QTextEdit,QVBoxLayout,QPushButton from link_converter import convert_url import pyperclip import sys class LinkConverter(QWidget): def __init__(self,parent=None): super().__init__(parent) self.setWindowTitle("Ensemblevideo Link Converter - Dennis Farmer") self.resize(600,300) self.textEdit = QTextEdit() self.btnPress1 = QPushButton("Convert to Download Link") self.btnPress3 = QPushButton("Clear Field Contents") self.btnPress2 = QPushButton("Copy to Clipboard") layout = QVBoxLayout() layout.addWidget(self.textEdit) layout.addWidget(self.btnPress1) layout.addWidget(self.btnPress2) layout.addWidget(self.btnPress3) self.setLayout(layout) self.textEdit.setPlainText("https://cloud.ensemblevideo.com/hapi/v1/contents/e5ee9fb7-8a1e-48f2-9e4d-2ac8a77c2478/launch?embedAsThumbnail=false&displayTitle=True&displaySharing=False&autoplay=False&showCaptions=False&hideControls=True&audioPreviewImage=False&displayEmbedCode=False&displayAttachments=True&displayLinks=True&displayCredits=False&displayVideoDuration=False&displayNotes=True&displayCaptionSearch=True&displayMetaData=false&displayDownloadIcon=False&displayViewersReport=False&displayAxdxs=False&idn_content=e5ee9fb7-8a1e-48f2-9e4d-2ac8a77c2478&idn_init=False&idn_sig=62NO%2FKLkev7ZUVK696RTOVkjP1Q%3D") self.btnPress1.clicked.connect(self.btnPress1_Clicked) self.btnPress2.clicked.connect(self.btnPress2_Clicked) self.btnPress3.clicked.connect(self.btnPress3_Clicked) def btnPress1_Clicked(self): old_url = self.textEdit.toPlainText() new_url = convert_url(old_url) self.textEdit.setPlainText(new_url) def btnPress2_Clicked(self): url = self.textEdit.toPlainText() pyperclip.copy(url) def btnPress3_Clicked(self): self.textEdit.setPlainText("") def Start(): lc = LinkConverter() lc.show() return lc def main(): app = QApplication(sys.argv) win = Start() app.exec_() if __name__ == '__main__': main()
true
true
f717b45b3fbcc04aefd659d8153dad422e2f094b
3,280
py
Python
test/optimization/test_vertex_cover.py
kareem1925/qiskit-aqua
7056f9bdd9ece32c41e162faecdcd24cf483da6f
[ "Apache-2.0" ]
null
null
null
test/optimization/test_vertex_cover.py
kareem1925/qiskit-aqua
7056f9bdd9ece32c41e162faecdcd24cf483da6f
[ "Apache-2.0" ]
null
null
null
test/optimization/test_vertex_cover.py
kareem1925/qiskit-aqua
7056f9bdd9ece32c41e162faecdcd24cf483da6f
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- # This code is part of Qiskit. # # (C) Copyright IBM 2018, 2020. # # This code is licensed under the Apache License, Version 2.0. You may # obtain a copy of this license in the LICENSE.txt file in the root directory # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. # # Any modifications or derivative works of this code must retain this # copyright notice, and modified files need to carry a notice indicating # that they have been altered from the originals. """ Test Vertex Cover """ import unittest from test.optimization import QiskitOptimizationTestCase import numpy as np from qiskit import BasicAer from qiskit.aqua import aqua_globals, QuantumInstance from qiskit.optimization.applications.ising import vertex_cover from qiskit.optimization.applications.ising.common import random_graph, sample_most_likely from qiskit.aqua.algorithms import NumPyMinimumEigensolver, VQE from qiskit.aqua.components.variational_forms import RYRZ from qiskit.aqua.components.optimizers import SPSA class TestVertexCover(QiskitOptimizationTestCase): """Cplex Ising tests.""" def setUp(self): super().setUp() self.seed = 100 aqua_globals.random_seed = self.seed self.num_nodes = 3 self.w = random_graph(self.num_nodes, edge_prob=0.8, weight_range=10) self.qubit_op, self.offset = vertex_cover.get_operator(self.w) def _brute_force(self): # brute-force way def bitfield(n, length): result = np.binary_repr(n, length) return [int(digit) for digit in result] # [2:] to chop off the "0b" part nodes = self.num_nodes maximum = 2 ** nodes minimal_v = np.inf for i in range(maximum): cur = bitfield(i, nodes) cur_v = vertex_cover.check_full_edge_coverage(np.array(cur), self.w) if cur_v: nonzerocount = np.count_nonzero(cur) if nonzerocount < minimal_v: minimal_v = nonzerocount return minimal_v def test_vertex_cover(self): """ Vertex Cover test """ algo = NumPyMinimumEigensolver(self.qubit_op, aux_operators=[]) result = algo.run() x = sample_most_likely(result.eigenstate) sol = vertex_cover.get_graph_solution(x) np.testing.assert_array_equal(sol, [0, 1, 1]) oracle = self._brute_force() self.assertEqual(np.count_nonzero(sol), oracle) def test_vertex_cover_vqe(self): """ Vertex Cover VQE test """ aqua_globals.random_seed = self.seed result = VQE(self.qubit_op, RYRZ(self.qubit_op.num_qubits, depth=3), SPSA(max_trials=200), max_evals_grouped=2).run( QuantumInstance(BasicAer.get_backend('qasm_simulator'), seed_simulator=aqua_globals.random_seed, seed_transpiler=aqua_globals.random_seed)) x = sample_most_likely(result['eigvecs'][0]) sol = vertex_cover.get_graph_solution(x) oracle = self._brute_force() self.assertEqual(np.count_nonzero(sol), oracle) if __name__ == '__main__': unittest.main()
36.043956
90
0.655488
import unittest from test.optimization import QiskitOptimizationTestCase import numpy as np from qiskit import BasicAer from qiskit.aqua import aqua_globals, QuantumInstance from qiskit.optimization.applications.ising import vertex_cover from qiskit.optimization.applications.ising.common import random_graph, sample_most_likely from qiskit.aqua.algorithms import NumPyMinimumEigensolver, VQE from qiskit.aqua.components.variational_forms import RYRZ from qiskit.aqua.components.optimizers import SPSA class TestVertexCover(QiskitOptimizationTestCase): def setUp(self): super().setUp() self.seed = 100 aqua_globals.random_seed = self.seed self.num_nodes = 3 self.w = random_graph(self.num_nodes, edge_prob=0.8, weight_range=10) self.qubit_op, self.offset = vertex_cover.get_operator(self.w) def _brute_force(self): def bitfield(n, length): result = np.binary_repr(n, length) return [int(digit) for digit in result] nodes = self.num_nodes maximum = 2 ** nodes minimal_v = np.inf for i in range(maximum): cur = bitfield(i, nodes) cur_v = vertex_cover.check_full_edge_coverage(np.array(cur), self.w) if cur_v: nonzerocount = np.count_nonzero(cur) if nonzerocount < minimal_v: minimal_v = nonzerocount return minimal_v def test_vertex_cover(self): algo = NumPyMinimumEigensolver(self.qubit_op, aux_operators=[]) result = algo.run() x = sample_most_likely(result.eigenstate) sol = vertex_cover.get_graph_solution(x) np.testing.assert_array_equal(sol, [0, 1, 1]) oracle = self._brute_force() self.assertEqual(np.count_nonzero(sol), oracle) def test_vertex_cover_vqe(self): aqua_globals.random_seed = self.seed result = VQE(self.qubit_op, RYRZ(self.qubit_op.num_qubits, depth=3), SPSA(max_trials=200), max_evals_grouped=2).run( QuantumInstance(BasicAer.get_backend('qasm_simulator'), seed_simulator=aqua_globals.random_seed, seed_transpiler=aqua_globals.random_seed)) x = sample_most_likely(result['eigvecs'][0]) sol = vertex_cover.get_graph_solution(x) oracle = self._brute_force() self.assertEqual(np.count_nonzero(sol), oracle) if __name__ == '__main__': unittest.main()
true
true
f717b6f0791909056e625635779049e23c7cb669
1,651
py
Python
scripts/test.py
JIC-Image-Analysis/find-plasmodesmata-dockerised
940afd85687b8383203f65da586da58a234e2deb
[ "MIT" ]
null
null
null
scripts/test.py
JIC-Image-Analysis/find-plasmodesmata-dockerised
940afd85687b8383203f65da586da58a234e2deb
[ "MIT" ]
null
null
null
scripts/test.py
JIC-Image-Analysis/find-plasmodesmata-dockerised
940afd85687b8383203f65da586da58a234e2deb
[ "MIT" ]
null
null
null
"""Functional test.""" import os import os.path from jicbioimage.core.image import MicroscopyCollection from jicbioimage.core.io import ( AutoName, DataManager, FileBackend, _md5_hexdigest_from_file, ) from plasmodesmata_analysis import plasmodesmata_analysis def test_plasmodesmata_analysis(): output_dir = "/output/tmp" if not os.path.isdir(output_dir): os.mkdir(output_dir) try: AutoName.directory = output_dir backend_dir = "/backend" test_image = "/scripts/test_data/test.tif" file_backend = FileBackend(backend_dir) data_manager = DataManager(file_backend) data_manager.load(test_image) md5_hex = _md5_hexdigest_from_file(test_image) manifest_path = os.path.join(backend_dir, md5_hex, "manifest.json") microscopy_collection = MicroscopyCollection() microscopy_collection.parse_manifest(manifest_path) plasmodesmata_analysis(microscopy_collection, 0, 15000, 2, 50) expected_data_dir = "/scripts/test_data/expected" for fname in os.listdir(expected_data_dir): expected_fpath = os.path.join(expected_data_dir, fname) result_fpath = os.path.join(output_dir, fname) expected_md5 = _md5_hexdigest_from_file(expected_fpath) result_md5 = _md5_hexdigest_from_file(result_fpath) assert expected_md5 == result_md5 finally: for fname in os.listdir(output_dir): fpath = os.path.join(output_dir, fname) os.unlink(fpath) os.rmdir(output_dir) if __name__ == "__main__": test_plasmodesmata_analysis()
30.018182
75
0.69473
import os import os.path from jicbioimage.core.image import MicroscopyCollection from jicbioimage.core.io import ( AutoName, DataManager, FileBackend, _md5_hexdigest_from_file, ) from plasmodesmata_analysis import plasmodesmata_analysis def test_plasmodesmata_analysis(): output_dir = "/output/tmp" if not os.path.isdir(output_dir): os.mkdir(output_dir) try: AutoName.directory = output_dir backend_dir = "/backend" test_image = "/scripts/test_data/test.tif" file_backend = FileBackend(backend_dir) data_manager = DataManager(file_backend) data_manager.load(test_image) md5_hex = _md5_hexdigest_from_file(test_image) manifest_path = os.path.join(backend_dir, md5_hex, "manifest.json") microscopy_collection = MicroscopyCollection() microscopy_collection.parse_manifest(manifest_path) plasmodesmata_analysis(microscopy_collection, 0, 15000, 2, 50) expected_data_dir = "/scripts/test_data/expected" for fname in os.listdir(expected_data_dir): expected_fpath = os.path.join(expected_data_dir, fname) result_fpath = os.path.join(output_dir, fname) expected_md5 = _md5_hexdigest_from_file(expected_fpath) result_md5 = _md5_hexdigest_from_file(result_fpath) assert expected_md5 == result_md5 finally: for fname in os.listdir(output_dir): fpath = os.path.join(output_dir, fname) os.unlink(fpath) os.rmdir(output_dir) if __name__ == "__main__": test_plasmodesmata_analysis()
true
true
f717b71efc824a722853352eee8071fc691d8c22
2,118
py
Python
airflow/providers/amazon/aws/example_dags/example_s3_bucket.py
jpsalado92/airflow
6776586904ab2a15727dcfc1c8d03c7a32a78444
[ "Apache-2.0", "BSD-2-Clause", "MIT", "ECL-2.0", "BSD-3-Clause" ]
27
2019-02-25T14:20:36.000Z
2022-03-22T09:35:13.000Z
airflow/providers/amazon/aws/example_dags/example_s3_bucket.py
sdwivedi565/airflow
78e4d16d970dbcb599d9c7f8df2a1a4273649ae8
[ "Apache-2.0", "BSD-2-Clause", "MIT", "ECL-2.0", "BSD-3-Clause" ]
200
2019-01-09T15:33:06.000Z
2022-01-12T09:13:42.000Z
airflow/providers/amazon/aws/example_dags/example_s3_bucket.py
MatrixManAtYrService/airflow-challenge
cef498efe9f4818e5f8e85c179ca730e84b31f4a
[ "Apache-2.0", "BSD-2-Clause", "MIT", "ECL-2.0", "BSD-3-Clause" ]
14
2019-06-25T17:08:29.000Z
2022-03-29T13:25:53.000Z
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. import os from datetime import datetime from airflow.decorators import task from airflow.models.dag import DAG from airflow.providers.amazon.aws.hooks.s3 import S3Hook from airflow.providers.amazon.aws.operators.s3 import S3CreateBucketOperator, S3DeleteBucketOperator BUCKET_NAME = os.environ.get('BUCKET_NAME', 'test-airflow-12345') @task(task_id="s3_bucket_dag_add_keys_to_bucket") def upload_keys(): """This is a python callback to add keys into the s3 bucket""" # add keys to bucket s3_hook = S3Hook() for i in range(0, 3): s3_hook.load_string( string_data="input", key=f"path/data{i}", bucket_name=BUCKET_NAME, ) # [START howto_operator_s3_bucket] with DAG( dag_id='s3_bucket_dag', schedule_interval=None, start_date=datetime(2021, 1, 1), catchup=False, default_args={"bucket_name": BUCKET_NAME}, max_active_runs=1, tags=['example'], ) as dag: create_bucket = S3CreateBucketOperator(task_id='s3_bucket_dag_create', region_name='us-east-1') # Using a task-decorated function to add keys add_keys_to_bucket = upload_keys() delete_bucket = S3DeleteBucketOperator(task_id='s3_bucket_dag_delete', force_delete=True) create_bucket >> add_keys_to_bucket >> delete_bucket # [END howto_operator_s3_bucket]
34.721311
100
0.742682
import os from datetime import datetime from airflow.decorators import task from airflow.models.dag import DAG from airflow.providers.amazon.aws.hooks.s3 import S3Hook from airflow.providers.amazon.aws.operators.s3 import S3CreateBucketOperator, S3DeleteBucketOperator BUCKET_NAME = os.environ.get('BUCKET_NAME', 'test-airflow-12345') @task(task_id="s3_bucket_dag_add_keys_to_bucket") def upload_keys(): s3_hook = S3Hook() for i in range(0, 3): s3_hook.load_string( string_data="input", key=f"path/data{i}", bucket_name=BUCKET_NAME, ) with DAG( dag_id='s3_bucket_dag', schedule_interval=None, start_date=datetime(2021, 1, 1), catchup=False, default_args={"bucket_name": BUCKET_NAME}, max_active_runs=1, tags=['example'], ) as dag: create_bucket = S3CreateBucketOperator(task_id='s3_bucket_dag_create', region_name='us-east-1') add_keys_to_bucket = upload_keys() delete_bucket = S3DeleteBucketOperator(task_id='s3_bucket_dag_delete', force_delete=True) create_bucket >> add_keys_to_bucket >> delete_bucket
true
true
f717b89a3172f1c4b9bc6e5d62033e37e8a5b95c
753
py
Python
modulos/utils.py
oliverfaustino/NRPG-DataManager
71064cb79be304f712aabcceebd6647121d2cb6c
[ "MIT" ]
null
null
null
modulos/utils.py
oliverfaustino/NRPG-DataManager
71064cb79be304f712aabcceebd6647121d2cb6c
[ "MIT" ]
null
null
null
modulos/utils.py
oliverfaustino/NRPG-DataManager
71064cb79be304f712aabcceebd6647121d2cb6c
[ "MIT" ]
null
null
null
import pyperclip import pandas as pd from modulos.conecao import * def copiar(objeto): # função para copiar os objetos para área de transferência global copiar # para resolver o porblema UnboundLocalError: local variable 'copiar' referenced before assignment: opcao = int(input('Deseja copiar para área de transferência? "1" para sim e qualquer tecla para não\n\nR: ')) if opcao == 1: copiar = pyperclip.copy(objeto) print('\nCopiado com sucesso!') else: pass return copiar def select(sql): # função que decta qual tipo de ação eu desejo fazer try: df = pd.read_sql_query(sql, con=engine).to_string(index=False) finally: pass return df
25.965517
131
0.654714
import pyperclip import pandas as pd from modulos.conecao import * def copiar(objeto): global copiar opcao = int(input('Deseja copiar para área de transferência? "1" para sim e qualquer tecla para não\n\nR: ')) if opcao == 1: copiar = pyperclip.copy(objeto) print('\nCopiado com sucesso!') else: pass return copiar def select(sql): try: df = pd.read_sql_query(sql, con=engine).to_string(index=False) finally: pass return df
true
true
f717b8e1b7a02ea373928b5f35b6948527803150
5,822
py
Python
tests/admin_views/test_templatetags.py
downstreamimpact/django
6686238cdc5c826ca5aab39d771798ff98e90ae8
[ "CNRI-Python-GPL-Compatible", "BSD-3-Clause" ]
7
2020-01-13T18:26:41.000Z
2021-04-20T04:22:26.000Z
tests/admin_views/test_templatetags.py
downstreamimpact/django
6686238cdc5c826ca5aab39d771798ff98e90ae8
[ "CNRI-Python-GPL-Compatible", "BSD-3-Clause" ]
null
null
null
tests/admin_views/test_templatetags.py
downstreamimpact/django
6686238cdc5c826ca5aab39d771798ff98e90ae8
[ "CNRI-Python-GPL-Compatible", "BSD-3-Clause" ]
4
2019-11-07T01:22:16.000Z
2020-09-16T22:02:16.000Z
import datetime from django.contrib.admin import ModelAdmin from django.contrib.admin.templatetags.admin_list import date_hierarchy from django.contrib.admin.templatetags.admin_modify import submit_row from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User from django.test import RequestFactory, TestCase from django.urls import reverse from .admin import ArticleAdmin, site from .models import Article, Question from .tests import AdminViewBasicTestCase class AdminTemplateTagsTest(AdminViewBasicTestCase): request_factory = RequestFactory() def test_submit_row(self): """ submit_row template tag should pass whole context. """ request = self.request_factory.get(reverse('admin:auth_user_change', args=[self.superuser.pk])) request.user = self.superuser admin = UserAdmin(User, site) extra_context = {'extra': True} response = admin.change_view(request, str(self.superuser.pk), extra_context=extra_context) template_context = submit_row(response.context_data) self.assertIs(template_context['extra'], True) self.assertIs(template_context['show_save'], True) def test_override_show_save_and_add_another(self): request = self.request_factory.get( reverse('admin:auth_user_change', args=[self.superuser.pk]), ) request.user = self.superuser admin = UserAdmin(User, site) for extra_context, expected_flag in ( ({}, True), # Default. ({'show_save_and_add_another': False}, False), ): with self.subTest(show_save_and_add_another=expected_flag): response = admin.change_view( request, str(self.superuser.pk), extra_context=extra_context, ) template_context = submit_row(response.context_data) self.assertIs(template_context['show_save_and_add_another'], expected_flag) def test_override_change_form_template_tags(self): """ admin_modify template tags follow the standard search pattern admin/app_label/model/template.html. """ article = Article.objects.all()[0] request = self.request_factory.get(reverse('admin:admin_views_article_change', args=[article.pk])) request.user = self.superuser admin = ArticleAdmin(Article, site) extra_context = {'show_publish': True, 'extra': True} response = admin.change_view(request, str(article.pk), extra_context=extra_context) response.render() self.assertIs(response.context_data['show_publish'], True) self.assertIs(response.context_data['extra'], True) self.assertContains(response, 'name="_save"') self.assertContains(response, 'name="_publish"') self.assertContains(response, 'override-change_form_object_tools') self.assertContains(response, 'override-prepopulated_fields_js') def test_override_change_list_template_tags(self): """ admin_list template tags follow the standard search pattern admin/app_label/model/template.html. """ request = self.request_factory.get(reverse('admin:admin_views_article_changelist')) request.user = self.superuser admin = ArticleAdmin(Article, site) admin.date_hierarchy = 'date' admin.search_fields = ('title', 'content') response = admin.changelist_view(request) response.render() self.assertContains(response, 'override-actions') self.assertContains(response, 'override-change_list_object_tools') self.assertContains(response, 'override-change_list_results') self.assertContains(response, 'override-date_hierarchy') self.assertContains(response, 'override-pagination') self.assertContains(response, 'override-search_form') class DateHierarchyTests(TestCase): factory = RequestFactory() @classmethod def setUpTestData(cls): cls.superuser = User.objects.create_superuser(username='super', password='secret', email='super@example.com') def test_choice_links(self): modeladmin = ModelAdmin(Question, site) modeladmin.date_hierarchy = 'posted' posted_dates = ( datetime.date(2017, 10, 1), datetime.date(2017, 10, 1), datetime.date(2017, 12, 15), datetime.date(2017, 12, 15), datetime.date(2017, 12, 31), datetime.date(2018, 2, 1), ) Question.objects.bulk_create(Question(question='q', posted=posted) for posted in posted_dates) tests = ( ({}, [['year=2017'], ['year=2018']]), ({'year': 2016}, []), ({'year': 2017}, [['month=10', 'year=2017'], ['month=12', 'year=2017']]), ({'year': 2017, 'month': 9}, []), ({'year': 2017, 'month': 12}, [['day=15', 'month=12', 'year=2017'], ['day=31', 'month=12', 'year=2017']]), ) for query, expected_choices in tests: with self.subTest(query=query): query = {'posted__%s' % q: val for q, val in query.items()} request = self.factory.get('/', query) request.user = self.superuser changelist = modeladmin.get_changelist_instance(request) spec = date_hierarchy(changelist) choices = [choice['link'] for choice in spec['choices']] expected_choices = [ '&'.join('posted__%s' % c for c in choice) for choice in expected_choices ] expected_choices = [('?' + choice) if choice else '' for choice in expected_choices] self.assertEqual(choices, expected_choices)
44.442748
118
0.643422
import datetime from django.contrib.admin import ModelAdmin from django.contrib.admin.templatetags.admin_list import date_hierarchy from django.contrib.admin.templatetags.admin_modify import submit_row from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User from django.test import RequestFactory, TestCase from django.urls import reverse from .admin import ArticleAdmin, site from .models import Article, Question from .tests import AdminViewBasicTestCase class AdminTemplateTagsTest(AdminViewBasicTestCase): request_factory = RequestFactory() def test_submit_row(self): request = self.request_factory.get(reverse('admin:auth_user_change', args=[self.superuser.pk])) request.user = self.superuser admin = UserAdmin(User, site) extra_context = {'extra': True} response = admin.change_view(request, str(self.superuser.pk), extra_context=extra_context) template_context = submit_row(response.context_data) self.assertIs(template_context['extra'], True) self.assertIs(template_context['show_save'], True) def test_override_show_save_and_add_another(self): request = self.request_factory.get( reverse('admin:auth_user_change', args=[self.superuser.pk]), ) request.user = self.superuser admin = UserAdmin(User, site) for extra_context, expected_flag in ( ({}, True), ({'show_save_and_add_another': False}, False), ): with self.subTest(show_save_and_add_another=expected_flag): response = admin.change_view( request, str(self.superuser.pk), extra_context=extra_context, ) template_context = submit_row(response.context_data) self.assertIs(template_context['show_save_and_add_another'], expected_flag) def test_override_change_form_template_tags(self): article = Article.objects.all()[0] request = self.request_factory.get(reverse('admin:admin_views_article_change', args=[article.pk])) request.user = self.superuser admin = ArticleAdmin(Article, site) extra_context = {'show_publish': True, 'extra': True} response = admin.change_view(request, str(article.pk), extra_context=extra_context) response.render() self.assertIs(response.context_data['show_publish'], True) self.assertIs(response.context_data['extra'], True) self.assertContains(response, 'name="_save"') self.assertContains(response, 'name="_publish"') self.assertContains(response, 'override-change_form_object_tools') self.assertContains(response, 'override-prepopulated_fields_js') def test_override_change_list_template_tags(self): request = self.request_factory.get(reverse('admin:admin_views_article_changelist')) request.user = self.superuser admin = ArticleAdmin(Article, site) admin.date_hierarchy = 'date' admin.search_fields = ('title', 'content') response = admin.changelist_view(request) response.render() self.assertContains(response, 'override-actions') self.assertContains(response, 'override-change_list_object_tools') self.assertContains(response, 'override-change_list_results') self.assertContains(response, 'override-date_hierarchy') self.assertContains(response, 'override-pagination') self.assertContains(response, 'override-search_form') class DateHierarchyTests(TestCase): factory = RequestFactory() @classmethod def setUpTestData(cls): cls.superuser = User.objects.create_superuser(username='super', password='secret', email='super@example.com') def test_choice_links(self): modeladmin = ModelAdmin(Question, site) modeladmin.date_hierarchy = 'posted' posted_dates = ( datetime.date(2017, 10, 1), datetime.date(2017, 10, 1), datetime.date(2017, 12, 15), datetime.date(2017, 12, 15), datetime.date(2017, 12, 31), datetime.date(2018, 2, 1), ) Question.objects.bulk_create(Question(question='q', posted=posted) for posted in posted_dates) tests = ( ({}, [['year=2017'], ['year=2018']]), ({'year': 2016}, []), ({'year': 2017}, [['month=10', 'year=2017'], ['month=12', 'year=2017']]), ({'year': 2017, 'month': 9}, []), ({'year': 2017, 'month': 12}, [['day=15', 'month=12', 'year=2017'], ['day=31', 'month=12', 'year=2017']]), ) for query, expected_choices in tests: with self.subTest(query=query): query = {'posted__%s' % q: val for q, val in query.items()} request = self.factory.get('/', query) request.user = self.superuser changelist = modeladmin.get_changelist_instance(request) spec = date_hierarchy(changelist) choices = [choice['link'] for choice in spec['choices']] expected_choices = [ '&'.join('posted__%s' % c for c in choice) for choice in expected_choices ] expected_choices = [('?' + choice) if choice else '' for choice in expected_choices] self.assertEqual(choices, expected_choices)
true
true
f717b91705d7d8a081fb14e5a9579959e466aa1f
28,392
py
Python
allennlp/models/semantic_parsing/atis/atis_semantic_parser.py
ljch2018/allennlp
63ba3fb28897578d4798039d1713e2b7995eb753
[ "Apache-2.0" ]
null
null
null
allennlp/models/semantic_parsing/atis/atis_semantic_parser.py
ljch2018/allennlp
63ba3fb28897578d4798039d1713e2b7995eb753
[ "Apache-2.0" ]
null
null
null
allennlp/models/semantic_parsing/atis/atis_semantic_parser.py
ljch2018/allennlp
63ba3fb28897578d4798039d1713e2b7995eb753
[ "Apache-2.0" ]
null
null
null
import logging from typing import Any, Dict, List, Tuple import difflib import sqlparse from overrides import overrides import torch from allennlp.common.util import pad_sequence_to_length from allennlp.data import Vocabulary from allennlp.data.fields.production_rule_field import ProductionRuleArray from allennlp.semparse.executors import SqlExecutor from allennlp.models.model import Model from allennlp.modules import Attention, Seq2SeqEncoder, TextFieldEmbedder, \ Embedding from allennlp.nn import util from allennlp.semparse.worlds import AtisWorld from allennlp.semparse.contexts.sql_context_utils import action_sequence_to_sql from allennlp.state_machines.states import GrammarBasedState from allennlp.state_machines.transition_functions.linking_transition_function import LinkingTransitionFunction from allennlp.state_machines import BeamSearch from allennlp.state_machines.trainers import MaximumMarginalLikelihood from allennlp.state_machines.states import GrammarStatelet, RnnStatelet from allennlp.training.metrics import Average logger = logging.getLogger(__name__) # pylint: disable=invalid-name @Model.register("atis_parser") class AtisSemanticParser(Model): """ Parameters ---------- vocab : ``Vocabulary`` utterance_embedder : ``TextFieldEmbedder`` Embedder for utterances. action_embedding_dim : ``int`` Dimension to use for action embeddings. encoder : ``Seq2SeqEncoder`` The encoder to use for the input utterance. decoder_beam_search : ``BeamSearch`` Beam search used to retrieve best sequences after training. max_decoding_steps : ``int`` When we're decoding with a beam search, what's the maximum number of steps we should take? This only applies at evaluation time, not during training. input_attention: ``Attention`` We compute an attention over the input utterance at each step of the decoder, using the decoder hidden state as the query. Passed to the transition function. add_action_bias : ``bool``, optional (default=True) If ``True``, we will learn a bias weight for each action that gets used when predicting that action, in addition to its embedding. dropout : ``float``, optional (default=0) If greater than 0, we will apply dropout with this probability after all encoders (pytorch LSTMs do not apply dropout to their last layer). rule_namespace : ``str``, optional (default=rule_labels) The vocabulary namespace to use for production rules. The default corresponds to the default used in the dataset reader, so you likely don't need to modify this. database_file: ``str``, optional (default=/atis/atis.db) The path of the SQLite database when evaluating SQL queries. SQLite is disk based, so we need the file location to connect to it. """ def __init__(self, vocab: Vocabulary, utterance_embedder: TextFieldEmbedder, action_embedding_dim: int, encoder: Seq2SeqEncoder, decoder_beam_search: BeamSearch, max_decoding_steps: int, input_attention: Attention, add_action_bias: bool = True, training_beam_size: int = None, dropout: float = 0.0, rule_namespace: str = 'rule_labels', database_file='/atis/atis.db') -> None: # Atis semantic parser init super().__init__(vocab) self._utterance_embedder = utterance_embedder self._encoder = encoder self._max_decoding_steps = max_decoding_steps self._add_action_bias = add_action_bias if dropout > 0: self._dropout = torch.nn.Dropout(p=dropout) else: self._dropout = lambda x: x self._rule_namespace = rule_namespace self._exact_match = Average() self._valid_sql_query = Average() self._action_similarity = Average() self._denotation_accuracy = Average() self._executor = SqlExecutor(database_file) self._action_padding_index = -1 # the padding value used by IndexField num_actions = vocab.get_vocab_size(self._rule_namespace) if self._add_action_bias: input_action_dim = action_embedding_dim + 1 else: input_action_dim = action_embedding_dim self._action_embedder = Embedding(num_embeddings=num_actions, embedding_dim=input_action_dim) self._output_action_embedder = Embedding(num_embeddings=num_actions, embedding_dim=action_embedding_dim) # This is what we pass as input in the first step of decoding, when we don't have a # previous action, or a previous utterance attention. self._first_action_embedding = torch.nn.Parameter(torch.FloatTensor(action_embedding_dim)) self._first_attended_utterance = torch.nn.Parameter(torch.FloatTensor(encoder.get_output_dim())) torch.nn.init.normal_(self._first_action_embedding) torch.nn.init.normal_(self._first_attended_utterance) self._num_entity_types = 2 # TODO(kevin): get this in a more principled way somehow? self._entity_type_decoder_embedding = Embedding(self._num_entity_types, action_embedding_dim) self._beam_search = decoder_beam_search self._decoder_trainer = MaximumMarginalLikelihood(training_beam_size) self._transition_function = LinkingTransitionFunction(encoder_output_dim=self._encoder.get_output_dim(), action_embedding_dim=action_embedding_dim, input_attention=input_attention, predict_start_type_separately=False, add_action_bias=self._add_action_bias, dropout=dropout) @overrides def forward(self, # type: ignore utterance: Dict[str, torch.LongTensor], world: List[AtisWorld], actions: List[List[ProductionRuleArray]], linking_scores: torch.Tensor, target_action_sequence: torch.LongTensor = None, sql_queries: List[List[str]] = None) -> Dict[str, torch.Tensor]: # pylint: disable=arguments-differ """ We set up the initial state for the decoder, and pass that state off to either a DecoderTrainer, if we're training, or a BeamSearch for inference, if we're not. Parameters ---------- utterance : Dict[str, torch.LongTensor] The output of ``TextField.as_array()`` applied on the utterance ``TextField``. This will be passed through a ``TextFieldEmbedder`` and then through an encoder. world : ``List[AtisWorld]`` We use a ``MetadataField`` to get the ``World`` for each input instance. Because of how ``MetadataField`` works, this gets passed to us as a ``List[AtisWorld]``, actions : ``List[List[ProductionRuleArray]]`` A list of all possible actions for each ``World`` in the batch, indexed into a ``ProductionRuleArray`` using a ``ProductionRuleField``. We will embed all of these and use the embeddings to determine which action to take at each timestep in the decoder. linking_scores: ``torch.Tensor`` A matrix of the linking the utterance tokens and the entities. This is a binary matrix that is deterministically generated where each entry indicates whether a token generated an entity. This tensor has shape ``(batch_size, num_entities, num_utterance_tokens)``. target_action_sequence : torch.Tensor, optional (default=None) The action sequence for the correct action sequence, where each action is an index into the list of possible actions. This tensor has shape ``(batch_size, sequence_length, 1)``. We remove the trailing dimension. sql_queries : List[List[str]], optional (default=None) A list of the SQL queries that are given during training or validation. """ initial_state = self._get_initial_state(utterance, world, actions, linking_scores) batch_size = linking_scores.shape[0] if target_action_sequence is not None: # Remove the trailing dimension (from ListField[ListField[IndexField]]). target_action_sequence = target_action_sequence.squeeze(-1) target_mask = target_action_sequence != self._action_padding_index else: target_mask = None if self.training: # target_action_sequence is of shape (batch_size, 1, sequence_length) here after we unsqueeze it for # the MML trainer. return self._decoder_trainer.decode(initial_state, self._transition_function, (target_action_sequence.unsqueeze(1), target_mask.unsqueeze(1))) else: # TODO(kevin) Move some of this functionality to a separate method for computing validation outputs. action_mapping = {} for batch_index, batch_actions in enumerate(actions): for action_index, action in enumerate(batch_actions): action_mapping[(batch_index, action_index)] = action[0] outputs: Dict[str, Any] = {'action_mapping': action_mapping} outputs['linking_scores'] = linking_scores if target_action_sequence is not None: outputs['loss'] = self._decoder_trainer.decode(initial_state, self._transition_function, (target_action_sequence.unsqueeze(1), target_mask.unsqueeze(1)))['loss'] num_steps = self._max_decoding_steps # This tells the state to start keeping track of debug info, which we'll pass along in # our output dictionary. initial_state.debug_info = [[] for _ in range(batch_size)] best_final_states = self._beam_search.search(num_steps, initial_state, self._transition_function, keep_final_unfinished_states=False) outputs['best_action_sequence'] = [] outputs['debug_info'] = [] outputs['entities'] = [] outputs['predicted_sql_query'] = [] outputs['sql_queries'] = [] outputs['utterance'] = [] outputs['tokenized_utterance'] = [] for i in range(batch_size): # Decoding may not have terminated with any completed valid SQL queries, if `num_steps` # isn't long enough (or if the model is not trained enough and gets into an # infinite action loop). if i not in best_final_states: self._exact_match(0) self._denotation_accuracy(0) self._valid_sql_query(0) self._action_similarity(0) outputs['predicted_sql_query'].append('') continue best_action_indices = best_final_states[i][0].action_history[0] action_strings = [action_mapping[(i, action_index)] for action_index in best_action_indices] predicted_sql_query = action_sequence_to_sql(action_strings) if target_action_sequence is not None: # Use a Tensor, not a Variable, to avoid a memory leak. targets = target_action_sequence[i].data sequence_in_targets = 0 sequence_in_targets = self._action_history_match(best_action_indices, targets) self._exact_match(sequence_in_targets) similarity = difflib.SequenceMatcher(None, best_action_indices, targets) self._action_similarity(similarity.ratio()) if sql_queries and sql_queries[i]: denotation_correct = self._executor.evaluate_sql_query(predicted_sql_query, sql_queries[i]) self._denotation_accuracy(denotation_correct) outputs['sql_queries'].append(sql_queries[i]) outputs['utterance'].append(world[i].utterances[-1]) outputs['tokenized_utterance'].append([token.text for token in world[i].tokenized_utterances[-1]]) outputs['entities'].append(world[i].entities) outputs['best_action_sequence'].append(action_strings) outputs['predicted_sql_query'].append(sqlparse.format(predicted_sql_query, reindent=True)) outputs['debug_info'].append(best_final_states[i][0].debug_info[0]) # type: ignore return outputs def _get_initial_state(self, utterance: Dict[str, torch.LongTensor], worlds: List[AtisWorld], actions: List[List[ProductionRuleArray]], linking_scores: torch.Tensor) -> GrammarBasedState: embedded_utterance = self._utterance_embedder(utterance) utterance_mask = util.get_text_field_mask(utterance).float() batch_size = embedded_utterance.size(0) num_entities = max([len(world.entities) for world in worlds]) # entity_types: tensor with shape (batch_size, num_entities) entity_types, _ = self._get_type_vector(worlds, num_entities, embedded_utterance) # (batch_size, num_utterance_tokens, embedding_dim) encoder_input = embedded_utterance # (batch_size, utterance_length, encoder_output_dim) encoder_outputs = self._dropout(self._encoder(encoder_input, utterance_mask)) # This will be our initial hidden state and memory cell for the decoder LSTM. final_encoder_output = util.get_final_encoder_states(encoder_outputs, utterance_mask, self._encoder.is_bidirectional()) memory_cell = encoder_outputs.new_zeros(batch_size, self._encoder.get_output_dim()) initial_score = embedded_utterance.data.new_zeros(batch_size) # To make grouping states together in the decoder easier, we convert the batch dimension in # all of our tensors into an outer list. For instance, the encoder outputs have shape # `(batch_size, utterance_length, encoder_output_dim)`. We need to convert this into a list # of `batch_size` tensors, each of shape `(utterance_length, encoder_output_dim)`. Then we # won't have to do any index selects, or anything, we'll just do some `torch.cat()`s. initial_score_list = [initial_score[i] for i in range(batch_size)] encoder_output_list = [encoder_outputs[i] for i in range(batch_size)] utterance_mask_list = [utterance_mask[i] for i in range(batch_size)] initial_rnn_state = [] for i in range(batch_size): initial_rnn_state.append(RnnStatelet(final_encoder_output[i], memory_cell[i], self._first_action_embedding, self._first_attended_utterance, encoder_output_list, utterance_mask_list)) initial_grammar_state = [self._create_grammar_state(worlds[i], actions[i], linking_scores[i], entity_types[i]) for i in range(batch_size)] initial_state = GrammarBasedState(batch_indices=list(range(batch_size)), action_history=[[] for _ in range(batch_size)], score=initial_score_list, rnn_state=initial_rnn_state, grammar_state=initial_grammar_state, possible_actions=actions, debug_info=None) return initial_state @staticmethod def _get_type_vector(worlds: List[AtisWorld], num_entities: int, tensor: torch.Tensor = None) -> Tuple[torch.LongTensor, Dict[int, int]]: """ Produces the encoding for each entity's type. In addition, a map from a flattened entity index to type is returned to combine entity type operations into one method. Parameters ---------- worlds : ``List[AtisWorld]`` num_entities : ``int`` tensor : ``torch.Tensor`` Used for copying the constructed list onto the right device. Returns ------- A ``torch.LongTensor`` with shape ``(batch_size, num_entities, num_types)``. entity_types : ``Dict[int, int]`` This is a mapping from ((batch_index * num_entities) + entity_index) to entity type id. """ entity_types = {} batch_types = [] for batch_index, world in enumerate(worlds): types = [] entities = [('number', entity) if 'number' or 'time_range' in entity else ('string', entity) for entity in world.entities] for entity_index, entity in enumerate(entities): # We need numbers to be first, then strings, since our entities are going to be # sorted. We do a split by type and then a merge later, and it relies on this sorting. if entity[0] == 'number': entity_type = 1 else: entity_type = 0 types.append(entity_type) # For easier lookups later, we're actually using a _flattened_ version # of (batch_index, entity_index) for the key, because this is how the # linking scores are stored. flattened_entity_index = batch_index * num_entities + entity_index entity_types[flattened_entity_index] = entity_type padded = pad_sequence_to_length(types, num_entities, lambda: 0) batch_types.append(padded) return tensor.new_tensor(batch_types, dtype=torch.long), entity_types @staticmethod def _action_history_match(predicted: List[int], targets: torch.LongTensor) -> int: # TODO(mattg): this could probably be moved into a FullSequenceMatch metric, or something. # Check if target is big enough to cover prediction (including start/end symbols) if len(predicted) > targets.size(0): return 0 predicted_tensor = targets.new_tensor(predicted) targets_trimmed = targets[:len(predicted)] # Return 1 if the predicted sequence is anywhere in the list of targets. return predicted_tensor.equal(targets_trimmed) @staticmethod def is_nonterminal(token: str): if token[0] == '"' and token[-1] == '"': return False return True @overrides def get_metrics(self, reset: bool = False) -> Dict[str, float]: """ We track four metrics here: 1. exact_match, which is the percentage of the time that our best output action sequence matches the SQL query exactly. 2. denotation_acc, which is the percentage of examples where we get the correct denotation. This is the typical "accuracy" metric, and it is what you should usually report in an experimental result. You need to be careful, though, that you're computing this on the full data, and not just the subset that can be parsed. (make sure you pass "keep_if_unparseable=True" to the dataset reader, which we do for validation data, but not training data). 3. valid_sql_query, which is the percentage of time that decoding actually produces a valid SQL query. We might not produce a valid SQL query if the decoder gets into a repetitive loop, or we're trying to produce a super long SQL query and run out of time steps, or something. 4. action_similarity, which is how similar the action sequence predicted is to the actual action sequence. This is basically a soft measure of exact_match. """ return { 'exact_match': self._exact_match.get_metric(reset), 'denotation_acc': self._denotation_accuracy.get_metric(reset), 'valid_sql_query': self._valid_sql_query.get_metric(reset), 'action_similarity': self._action_similarity.get_metric(reset) } def _create_grammar_state(self, world: AtisWorld, possible_actions: List[ProductionRuleArray], linking_scores: torch.Tensor, entity_types: torch.Tensor) -> GrammarStatelet: """ This method creates the GrammarStatelet object that's used for decoding. Part of creating that is creating the `valid_actions` dictionary, which contains embedded representations of all of the valid actions. So, we create that here as well. The inputs to this method are for a `single instance in the batch`; none of the tensors we create here are batched. We grab the global action ids from the input ``ProductionRuleArrays``, and we use those to embed the valid actions for every non-terminal type. We use the input ``linking_scores`` for non-global actions. Parameters ---------- world : ``AtisWorld`` From the input to ``forward`` for a single batch instance. possible_actions : ``List[ProductionRuleArray]`` From the input to ``forward`` for a single batch instance. linking_scores : ``torch.Tensor`` Assumed to have shape ``(num_entities, num_utterance_tokens)`` (i.e., there is no batch dimension). entity_types : ``torch.Tensor`` Assumed to have shape ``(num_entities,)`` (i.e., there is no batch dimension). """ action_map = {} for action_index, action in enumerate(possible_actions): action_string = action[0] action_map[action_string] = action_index valid_actions = world.valid_actions entity_map = {} entities = world.entities for entity_index, entity in enumerate(entities): entity_map[entity] = entity_index translated_valid_actions: Dict[str, Dict[str, Tuple[torch.Tensor, torch.Tensor, List[int]]]] = {} for key, action_strings in valid_actions.items(): translated_valid_actions[key] = {} # `key` here is a non-terminal from the grammar, and `action_strings` are all the valid # productions of that non-terminal. We'll first split those productions by global vs. # linked action. action_indices = [action_map[action_string] for action_string in action_strings] production_rule_arrays = [(possible_actions[index], index) for index in action_indices] global_actions = [] linked_actions = [] for production_rule_array, action_index in production_rule_arrays: if production_rule_array[1]: global_actions.append((production_rule_array[2], action_index)) else: linked_actions.append((production_rule_array[0], action_index)) if global_actions: global_action_tensors, global_action_ids = zip(*global_actions) global_action_tensor = entity_types.new_tensor(torch.cat(global_action_tensors, dim=0), dtype=torch.long) global_input_embeddings = self._action_embedder(global_action_tensor) global_output_embeddings = self._output_action_embedder(global_action_tensor) translated_valid_actions[key]['global'] = (global_input_embeddings, global_output_embeddings, list(global_action_ids)) if linked_actions: linked_rules, linked_action_ids = zip(*linked_actions) entities = linked_rules entity_ids = [entity_map[entity] for entity in entities] entity_linking_scores = linking_scores[entity_ids] entity_type_tensor = entity_types[entity_ids] entity_type_embeddings = self._entity_type_decoder_embedding(entity_type_tensor) entity_type_embeddings = entity_types.new_tensor(entity_type_embeddings, dtype=torch.float) translated_valid_actions[key]['linked'] = (entity_linking_scores, entity_type_embeddings, list(linked_action_ids)) return GrammarStatelet(['statement'], {}, translated_valid_actions, {}, self.is_nonterminal, reverse_productions=False) @overrides def decode(self, output_dict: Dict[str, torch.Tensor]) -> Dict[str, torch.Tensor]: """ This method overrides ``Model.decode``, which gets called after ``Model.forward``, at test time, to finalize predictions. This is (confusingly) a separate notion from the "decoder" in "encoder/decoder", where that decoder logic lives in ``TransitionFunction``. This method trims the output predictions to the first end symbol, replaces indices with corresponding tokens, and adds a field called ``predicted_actions`` to the ``output_dict``. """ action_mapping = output_dict['action_mapping'] best_actions = output_dict["best_action_sequence"] debug_infos = output_dict['debug_info'] batch_action_info = [] for batch_index, (predicted_actions, debug_info) in enumerate(zip(best_actions, debug_infos)): instance_action_info = [] for predicted_action, action_debug_info in zip(predicted_actions, debug_info): action_info = {} action_info['predicted_action'] = predicted_action considered_actions = action_debug_info['considered_actions'] probabilities = action_debug_info['probabilities'] actions = [] for action, probability in zip(considered_actions, probabilities): if action != -1: actions.append((action_mapping[(batch_index, action)], probability)) actions.sort() considered_actions, probabilities = zip(*actions) action_info['considered_actions'] = considered_actions action_info['action_probabilities'] = probabilities action_info['utterance_attention'] = action_debug_info.get('question_attention', []) instance_action_info.append(action_info) batch_action_info.append(instance_action_info) output_dict["predicted_actions"] = batch_action_info return output_dict
54.705202
112
0.61151
import logging from typing import Any, Dict, List, Tuple import difflib import sqlparse from overrides import overrides import torch from allennlp.common.util import pad_sequence_to_length from allennlp.data import Vocabulary from allennlp.data.fields.production_rule_field import ProductionRuleArray from allennlp.semparse.executors import SqlExecutor from allennlp.models.model import Model from allennlp.modules import Attention, Seq2SeqEncoder, TextFieldEmbedder, \ Embedding from allennlp.nn import util from allennlp.semparse.worlds import AtisWorld from allennlp.semparse.contexts.sql_context_utils import action_sequence_to_sql from allennlp.state_machines.states import GrammarBasedState from allennlp.state_machines.transition_functions.linking_transition_function import LinkingTransitionFunction from allennlp.state_machines import BeamSearch from allennlp.state_machines.trainers import MaximumMarginalLikelihood from allennlp.state_machines.states import GrammarStatelet, RnnStatelet from allennlp.training.metrics import Average logger = logging.getLogger(__name__) @Model.register("atis_parser") class AtisSemanticParser(Model): def __init__(self, vocab: Vocabulary, utterance_embedder: TextFieldEmbedder, action_embedding_dim: int, encoder: Seq2SeqEncoder, decoder_beam_search: BeamSearch, max_decoding_steps: int, input_attention: Attention, add_action_bias: bool = True, training_beam_size: int = None, dropout: float = 0.0, rule_namespace: str = 'rule_labels', database_file='/atis/atis.db') -> None: super().__init__(vocab) self._utterance_embedder = utterance_embedder self._encoder = encoder self._max_decoding_steps = max_decoding_steps self._add_action_bias = add_action_bias if dropout > 0: self._dropout = torch.nn.Dropout(p=dropout) else: self._dropout = lambda x: x self._rule_namespace = rule_namespace self._exact_match = Average() self._valid_sql_query = Average() self._action_similarity = Average() self._denotation_accuracy = Average() self._executor = SqlExecutor(database_file) self._action_padding_index = -1 num_actions = vocab.get_vocab_size(self._rule_namespace) if self._add_action_bias: input_action_dim = action_embedding_dim + 1 else: input_action_dim = action_embedding_dim self._action_embedder = Embedding(num_embeddings=num_actions, embedding_dim=input_action_dim) self._output_action_embedder = Embedding(num_embeddings=num_actions, embedding_dim=action_embedding_dim) # previous action, or a previous utterance attention. self._first_action_embedding = torch.nn.Parameter(torch.FloatTensor(action_embedding_dim)) self._first_attended_utterance = torch.nn.Parameter(torch.FloatTensor(encoder.get_output_dim())) torch.nn.init.normal_(self._first_action_embedding) torch.nn.init.normal_(self._first_attended_utterance) self._num_entity_types = 2 # TODO(kevin): get this in a more principled way somehow? self._entity_type_decoder_embedding = Embedding(self._num_entity_types, action_embedding_dim) self._beam_search = decoder_beam_search self._decoder_trainer = MaximumMarginalLikelihood(training_beam_size) self._transition_function = LinkingTransitionFunction(encoder_output_dim=self._encoder.get_output_dim(), action_embedding_dim=action_embedding_dim, input_attention=input_attention, predict_start_type_separately=False, add_action_bias=self._add_action_bias, dropout=dropout) @overrides def forward(self, # type: ignore utterance: Dict[str, torch.LongTensor], world: List[AtisWorld], actions: List[List[ProductionRuleArray]], linking_scores: torch.Tensor, target_action_sequence: torch.LongTensor = None, sql_queries: List[List[str]] = None) -> Dict[str, torch.Tensor]: # pylint: disable=arguments-differ initial_state = self._get_initial_state(utterance, world, actions, linking_scores) batch_size = linking_scores.shape[0] if target_action_sequence is not None: # Remove the trailing dimension (from ListField[ListField[IndexField]]). target_action_sequence = target_action_sequence.squeeze(-1) target_mask = target_action_sequence != self._action_padding_index else: target_mask = None if self.training: # target_action_sequence is of shape (batch_size, 1, sequence_length) here after we unsqueeze it for # the MML trainer. return self._decoder_trainer.decode(initial_state, self._transition_function, (target_action_sequence.unsqueeze(1), target_mask.unsqueeze(1))) else: # TODO(kevin) Move some of this functionality to a separate method for computing validation outputs. action_mapping = {} for batch_index, batch_actions in enumerate(actions): for action_index, action in enumerate(batch_actions): action_mapping[(batch_index, action_index)] = action[0] outputs: Dict[str, Any] = {'action_mapping': action_mapping} outputs['linking_scores'] = linking_scores if target_action_sequence is not None: outputs['loss'] = self._decoder_trainer.decode(initial_state, self._transition_function, (target_action_sequence.unsqueeze(1), target_mask.unsqueeze(1)))['loss'] num_steps = self._max_decoding_steps # This tells the state to start keeping track of debug info, which we'll pass along in initial_state.debug_info = [[] for _ in range(batch_size)] best_final_states = self._beam_search.search(num_steps, initial_state, self._transition_function, keep_final_unfinished_states=False) outputs['best_action_sequence'] = [] outputs['debug_info'] = [] outputs['entities'] = [] outputs['predicted_sql_query'] = [] outputs['sql_queries'] = [] outputs['utterance'] = [] outputs['tokenized_utterance'] = [] for i in range(batch_size): # infinite action loop). if i not in best_final_states: self._exact_match(0) self._denotation_accuracy(0) self._valid_sql_query(0) self._action_similarity(0) outputs['predicted_sql_query'].append('') continue best_action_indices = best_final_states[i][0].action_history[0] action_strings = [action_mapping[(i, action_index)] for action_index in best_action_indices] predicted_sql_query = action_sequence_to_sql(action_strings) if target_action_sequence is not None: # Use a Tensor, not a Variable, to avoid a memory leak. targets = target_action_sequence[i].data sequence_in_targets = 0 sequence_in_targets = self._action_history_match(best_action_indices, targets) self._exact_match(sequence_in_targets) similarity = difflib.SequenceMatcher(None, best_action_indices, targets) self._action_similarity(similarity.ratio()) if sql_queries and sql_queries[i]: denotation_correct = self._executor.evaluate_sql_query(predicted_sql_query, sql_queries[i]) self._denotation_accuracy(denotation_correct) outputs['sql_queries'].append(sql_queries[i]) outputs['utterance'].append(world[i].utterances[-1]) outputs['tokenized_utterance'].append([token.text for token in world[i].tokenized_utterances[-1]]) outputs['entities'].append(world[i].entities) outputs['best_action_sequence'].append(action_strings) outputs['predicted_sql_query'].append(sqlparse.format(predicted_sql_query, reindent=True)) outputs['debug_info'].append(best_final_states[i][0].debug_info[0]) # type: ignore return outputs def _get_initial_state(self, utterance: Dict[str, torch.LongTensor], worlds: List[AtisWorld], actions: List[List[ProductionRuleArray]], linking_scores: torch.Tensor) -> GrammarBasedState: embedded_utterance = self._utterance_embedder(utterance) utterance_mask = util.get_text_field_mask(utterance).float() batch_size = embedded_utterance.size(0) num_entities = max([len(world.entities) for world in worlds]) # entity_types: tensor with shape (batch_size, num_entities) entity_types, _ = self._get_type_vector(worlds, num_entities, embedded_utterance) # (batch_size, num_utterance_tokens, embedding_dim) encoder_input = embedded_utterance # (batch_size, utterance_length, encoder_output_dim) encoder_outputs = self._dropout(self._encoder(encoder_input, utterance_mask)) # This will be our initial hidden state and memory cell for the decoder LSTM. final_encoder_output = util.get_final_encoder_states(encoder_outputs, utterance_mask, self._encoder.is_bidirectional()) memory_cell = encoder_outputs.new_zeros(batch_size, self._encoder.get_output_dim()) initial_score = embedded_utterance.data.new_zeros(batch_size) # To make grouping states together in the decoder easier, we convert the batch dimension in # all of our tensors into an outer list. For instance, the encoder outputs have shape # `(batch_size, utterance_length, encoder_output_dim)`. We need to convert this into a list # of `batch_size` tensors, each of shape `(utterance_length, encoder_output_dim)`. Then we # won't have to do any index selects, or anything, we'll just do some `torch.cat()`s. initial_score_list = [initial_score[i] for i in range(batch_size)] encoder_output_list = [encoder_outputs[i] for i in range(batch_size)] utterance_mask_list = [utterance_mask[i] for i in range(batch_size)] initial_rnn_state = [] for i in range(batch_size): initial_rnn_state.append(RnnStatelet(final_encoder_output[i], memory_cell[i], self._first_action_embedding, self._first_attended_utterance, encoder_output_list, utterance_mask_list)) initial_grammar_state = [self._create_grammar_state(worlds[i], actions[i], linking_scores[i], entity_types[i]) for i in range(batch_size)] initial_state = GrammarBasedState(batch_indices=list(range(batch_size)), action_history=[[] for _ in range(batch_size)], score=initial_score_list, rnn_state=initial_rnn_state, grammar_state=initial_grammar_state, possible_actions=actions, debug_info=None) return initial_state @staticmethod def _get_type_vector(worlds: List[AtisWorld], num_entities: int, tensor: torch.Tensor = None) -> Tuple[torch.LongTensor, Dict[int, int]]: entity_types = {} batch_types = [] for batch_index, world in enumerate(worlds): types = [] entities = [('number', entity) if 'number' or 'time_range' in entity else ('string', entity) for entity in world.entities] for entity_index, entity in enumerate(entities): # We need numbers to be first, then strings, since our entities are going to be # sorted. We do a split by type and then a merge later, and it relies on this sorting. if entity[0] == 'number': entity_type = 1 else: entity_type = 0 types.append(entity_type) # For easier lookups later, we're actually using a _flattened_ version flattened_entity_index = batch_index * num_entities + entity_index entity_types[flattened_entity_index] = entity_type padded = pad_sequence_to_length(types, num_entities, lambda: 0) batch_types.append(padded) return tensor.new_tensor(batch_types, dtype=torch.long), entity_types @staticmethod def _action_history_match(predicted: List[int], targets: torch.LongTensor) -> int: if len(predicted) > targets.size(0): return 0 predicted_tensor = targets.new_tensor(predicted) targets_trimmed = targets[:len(predicted)] return predicted_tensor.equal(targets_trimmed) @staticmethod def is_nonterminal(token: str): if token[0] == '"' and token[-1] == '"': return False return True @overrides def get_metrics(self, reset: bool = False) -> Dict[str, float]: return { 'exact_match': self._exact_match.get_metric(reset), 'denotation_acc': self._denotation_accuracy.get_metric(reset), 'valid_sql_query': self._valid_sql_query.get_metric(reset), 'action_similarity': self._action_similarity.get_metric(reset) } def _create_grammar_state(self, world: AtisWorld, possible_actions: List[ProductionRuleArray], linking_scores: torch.Tensor, entity_types: torch.Tensor) -> GrammarStatelet: action_map = {} for action_index, action in enumerate(possible_actions): action_string = action[0] action_map[action_string] = action_index valid_actions = world.valid_actions entity_map = {} entities = world.entities for entity_index, entity in enumerate(entities): entity_map[entity] = entity_index translated_valid_actions: Dict[str, Dict[str, Tuple[torch.Tensor, torch.Tensor, List[int]]]] = {} for key, action_strings in valid_actions.items(): translated_valid_actions[key] = {} # linked action. action_indices = [action_map[action_string] for action_string in action_strings] production_rule_arrays = [(possible_actions[index], index) for index in action_indices] global_actions = [] linked_actions = [] for production_rule_array, action_index in production_rule_arrays: if production_rule_array[1]: global_actions.append((production_rule_array[2], action_index)) else: linked_actions.append((production_rule_array[0], action_index)) if global_actions: global_action_tensors, global_action_ids = zip(*global_actions) global_action_tensor = entity_types.new_tensor(torch.cat(global_action_tensors, dim=0), dtype=torch.long) global_input_embeddings = self._action_embedder(global_action_tensor) global_output_embeddings = self._output_action_embedder(global_action_tensor) translated_valid_actions[key]['global'] = (global_input_embeddings, global_output_embeddings, list(global_action_ids)) if linked_actions: linked_rules, linked_action_ids = zip(*linked_actions) entities = linked_rules entity_ids = [entity_map[entity] for entity in entities] entity_linking_scores = linking_scores[entity_ids] entity_type_tensor = entity_types[entity_ids] entity_type_embeddings = self._entity_type_decoder_embedding(entity_type_tensor) entity_type_embeddings = entity_types.new_tensor(entity_type_embeddings, dtype=torch.float) translated_valid_actions[key]['linked'] = (entity_linking_scores, entity_type_embeddings, list(linked_action_ids)) return GrammarStatelet(['statement'], {}, translated_valid_actions, {}, self.is_nonterminal, reverse_productions=False) @overrides def decode(self, output_dict: Dict[str, torch.Tensor]) -> Dict[str, torch.Tensor]: action_mapping = output_dict['action_mapping'] best_actions = output_dict["best_action_sequence"] debug_infos = output_dict['debug_info'] batch_action_info = [] for batch_index, (predicted_actions, debug_info) in enumerate(zip(best_actions, debug_infos)): instance_action_info = [] for predicted_action, action_debug_info in zip(predicted_actions, debug_info): action_info = {} action_info['predicted_action'] = predicted_action considered_actions = action_debug_info['considered_actions'] probabilities = action_debug_info['probabilities'] actions = [] for action, probability in zip(considered_actions, probabilities): if action != -1: actions.append((action_mapping[(batch_index, action)], probability)) actions.sort() considered_actions, probabilities = zip(*actions) action_info['considered_actions'] = considered_actions action_info['action_probabilities'] = probabilities action_info['utterance_attention'] = action_debug_info.get('question_attention', []) instance_action_info.append(action_info) batch_action_info.append(instance_action_info) output_dict["predicted_actions"] = batch_action_info return output_dict
true
true
f717ba6db19fac8fe6c6bc3c5388211b577e2db4
871
py
Python
tests/unit/fake_data_root/kubernetes/var/lib/juju/agents/unit-containerd-2/charm/hooks/relations/untrusted-container-runtime/requires.py
KellenRenshaw/hotsos
e3fc51ab7f8af606a5846a3486a7fda23d761583
[ "Apache-2.0" ]
6
2021-10-01T19:46:14.000Z
2022-03-31T17:05:08.000Z
tests/unit/fake_data_root/kubernetes/var/lib/juju/agents/unit-containerd-2/charm/hooks/relations/untrusted-container-runtime/requires.py
KellenRenshaw/hotsos
e3fc51ab7f8af606a5846a3486a7fda23d761583
[ "Apache-2.0" ]
111
2021-10-01T18:18:17.000Z
2022-03-29T12:23:20.000Z
tests/unit/fake_data_root/kubernetes/var/lib/juju/agents/unit-containerd-2/charm/hooks/relations/untrusted-container-runtime/requires.py
KellenRenshaw/hotsos
e3fc51ab7f8af606a5846a3486a7fda23d761583
[ "Apache-2.0" ]
10
2021-09-29T14:47:54.000Z
2022-03-18T14:52:16.000Z
from charms.reactive import ( Endpoint, set_flag, clear_flag ) from charms.reactive import ( when, when_not ) class ContainerRuntimeRequires(Endpoint): @when('endpoint.{endpoint_name}.changed') def changed(self): set_flag(self.expand_name('endpoint.{endpoint_name}.available')) @when_not('endpoint.{endpoint_name}.joined') def broken(self): clear_flag(self.expand_name('endpoint.{endpoint_name}.available')) def set_config(self, name, binary_path): """ Set the configuration to be published. :param name: String name of runtime :param binary_path: String runtime executable :return: None """ for relation in self.relations: relation.to_publish.update({ 'name': name, 'binary_path': binary_path })
24.885714
74
0.626866
from charms.reactive import ( Endpoint, set_flag, clear_flag ) from charms.reactive import ( when, when_not ) class ContainerRuntimeRequires(Endpoint): @when('endpoint.{endpoint_name}.changed') def changed(self): set_flag(self.expand_name('endpoint.{endpoint_name}.available')) @when_not('endpoint.{endpoint_name}.joined') def broken(self): clear_flag(self.expand_name('endpoint.{endpoint_name}.available')) def set_config(self, name, binary_path): for relation in self.relations: relation.to_publish.update({ 'name': name, 'binary_path': binary_path })
true
true
f717ba7e0a6e1a58c6c5756909c027da10c010bb
878
py
Python
read_exif.py
kiyoon/camera-tools
2e269141597dd27ec7d41e49285e01ef566cb54c
[ "MIT" ]
1
2021-11-14T23:30:47.000Z
2021-11-14T23:30:47.000Z
read_exif.py
kiyoon/camera-tools
2e269141597dd27ec7d41e49285e01ef566cb54c
[ "MIT" ]
null
null
null
read_exif.py
kiyoon/camera-tools
2e269141597dd27ec7d41e49285e01ef566cb54c
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 import argparse class Formatter(argparse.ArgumentDefaultsHelpFormatter, argparse.RawDescriptionHelpFormatter): pass parser = argparse.ArgumentParser( description='''Read EXIF data Author: Kiyoon Kim (yoonkr33@gmail.com)''', formatter_class=Formatter) parser.add_argument('input_files', type=str, nargs='+', help='files to read metadata') args = parser.parse_args() import glob import os import exiftool import pprint if __name__ == "__main__": for origpath in args.input_files: for path in glob.glob(origpath): # glob: Windows wildcard support root, fname_ext = os.path.split(path) fname, fext = os.path.splitext(fname_ext) with exiftool.ExifTool() as et: metadata = et.get_metadata(path) print(path) pprint.pprint(metadata)
24.388889
94
0.671982
import argparse class Formatter(argparse.ArgumentDefaultsHelpFormatter, argparse.RawDescriptionHelpFormatter): pass parser = argparse.ArgumentParser( description='''Read EXIF data Author: Kiyoon Kim (yoonkr33@gmail.com)''', formatter_class=Formatter) parser.add_argument('input_files', type=str, nargs='+', help='files to read metadata') args = parser.parse_args() import glob import os import exiftool import pprint if __name__ == "__main__": for origpath in args.input_files: for path in glob.glob(origpath): root, fname_ext = os.path.split(path) fname, fext = os.path.splitext(fname_ext) with exiftool.ExifTool() as et: metadata = et.get_metadata(path) print(path) pprint.pprint(metadata)
true
true
f717bcdd89a1615aae5c49e9f6e2cbe27aa71630
458
py
Python
ProjectShop/ProductApp/urls.py
kolyasalubov/ProjectShop
41a62d27276542bf65aab51663220ffe88858024
[ "MIT" ]
3
2021-08-30T09:15:52.000Z
2022-03-24T14:40:11.000Z
ProjectShop/ProductApp/urls.py
kolyasalubov/ProjectShop
41a62d27276542bf65aab51663220ffe88858024
[ "MIT" ]
150
2021-08-30T12:37:17.000Z
2021-11-17T07:09:19.000Z
ProjectShop/ProductApp/urls.py
kolyasalubov/ProjectShop
41a62d27276542bf65aab51663220ffe88858024
[ "MIT" ]
null
null
null
from django.urls import path from ProductApp.views import ( HomePageView, CategoriesView, ProductDetailView, CategoryDetailView, ) urlpatterns = [ path("categories/<slug:slug>/", CategoryDetailView.as_view(), name="category-detail"), path("categories/", CategoriesView.as_view(), name="categories"), path("<slug:slug>/", ProductDetailView.as_view(), name="product-detail"), path("", HomePageView.as_view(), name="home"), ]
26.941176
90
0.696507
from django.urls import path from ProductApp.views import ( HomePageView, CategoriesView, ProductDetailView, CategoryDetailView, ) urlpatterns = [ path("categories/<slug:slug>/", CategoryDetailView.as_view(), name="category-detail"), path("categories/", CategoriesView.as_view(), name="categories"), path("<slug:slug>/", ProductDetailView.as_view(), name="product-detail"), path("", HomePageView.as_view(), name="home"), ]
true
true
f717bdb31e80489aaefa7fbd595c263379b27f37
15,376
py
Python
src/hub/dataload/sources/chembl/chembl_upload.py
ravila4/mychem.info
9b63b5f0957b5e7b252ca8122734a363905036b3
[ "Apache-2.0" ]
null
null
null
src/hub/dataload/sources/chembl/chembl_upload.py
ravila4/mychem.info
9b63b5f0957b5e7b252ca8122734a363905036b3
[ "Apache-2.0" ]
null
null
null
src/hub/dataload/sources/chembl/chembl_upload.py
ravila4/mychem.info
9b63b5f0957b5e7b252ca8122734a363905036b3
[ "Apache-2.0" ]
null
null
null
""" Chembl uploader """ # pylint: disable=E0401, E0611 import os import glob import pymongo import biothings.hub.dataload.storage as storage from biothings.hub.dataload.uploader import ParallelizedSourceUploader from hub.dataload.uploader import BaseDrugUploader from hub.datatransform.keylookup import MyChemKeyLookup from .chembl_parser import load_data SRC_META = { "url": 'https://www.ebi.ac.uk/chembl/', "license_url" : "https://www.ebi.ac.uk/about/terms-of-use", "license_url_short" : "http://bit.ly/2KAUCAm" } class ChemblUploader(BaseDrugUploader, ParallelizedSourceUploader): """ ChemblUploader - upload the Chembl data source """ name = "chembl" storage_class = storage.RootKeyMergerStorage __metadata__ = {"src_meta" : SRC_META} MOLECULE_PATTERN = "molecule.*.json" keylookup = MyChemKeyLookup( [("inchikey", "chembl.inchi_key"), ("inchi", "chembl.inchi"), ("chembl", "chembl.molecule_chembl_id"), ("chebi", "chembl.chebi_par_id"), ("drugcentral", "chembl.xrefs.drugcentral.id"), ("drugname", "chembl.pref_name")], # TODO: handle duplicate keys from pubchem # - we use RootKeyMergerStorage, but the num. duplicates # - is too high (>10000) # ("pubchem", "chembl.xrefs.pubchem.sid"), copy_from_doc=True) def jobs(self): """ this will generate arguments for self.load.data() method, allowing parallelization """ json_files = glob.glob(os.path.join(self.data_folder, self.__class__.MOLECULE_PATTERN)) return [(f,) for f in json_files] def load_data(self, data_folder): """load data from an input file""" self.logger.info("Load data from '%s'" % data_folder) return self.keylookup(load_data, debug=True)(data_folder) def post_update_data(self, *args, **kwargs): """create indexes following an update""" # pylint: disable=W0613 """ for idxname in ["chembl.chebi_par_id", "chembl.inchi", "chembl.molecule_chembl_id"]: self.logger.info("Indexing '%s'" % idxname) # background=true or it'll lock the whole database... self.collection.create_index(idxname, background=True) """ for idxname in ["chembl.chebi_par_id", "chembl.molecule_chembl_id"]: self.logger.info("Indexing '%s'" % idxname) # background=true or it'll lock the whole database... self.collection.create_index(idxname, background=True) @classmethod def get_mapping(cls): """return mapping data""" mapping = { "chembl": { "properties": { "biotherapeutic": { "properties": { "helm_notation": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "description": { "type": "text" }, "biocomponents": { "properties": { "organism": { "type": "text" }, "tax_id": { "type": "integer" }, "sequence": { "type": "text" }, "component_id": { "type": "integer" }, "description": { "type": "text" }, "component_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", 'copy_to': ['all'], } } }, "therapeutic_flag": { "type": "boolean" }, "usan_stem": { "type": "text" }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_properties": { "properties": { "heavy_atoms": { "type": "integer" }, "acd_most_bpka": { "type": "float" }, "mw_freebase": { "type": "float" }, "num_ro5_violations": { "type": "integer" }, "molecular_species": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "qed_weighted": { "type": "float" }, "ro3_pass": { "type": "boolean" }, "full_mwt": { "type": "float" }, "num_lipinski_ro5_violations": { "type": "integer" }, "rtb": { "type": "integer" }, "psa": { "type": "float" }, "alogp": { "type": "float" }, "hbd": { "type": "integer" }, "acd_most_apka": { "type": "float" }, "hbd_lipinski": { "type": "integer" }, "acd_logp": { "type": "float" }, "full_molformula": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "aromatic_rings": { "type": "integer" }, "hba_lipinski": { "type": "integer" }, "mw_monoisotopic": { "type": "float" }, "hba": { "type": "integer" }, "acd_logd": { "type": "float" } } }, "helm_notation": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "max_phase": { "type": "integer" }, "inorganic_flag": { "type": "integer" }, "usan_stem_definition": { "type": "text" }, "dosed_ingredient": { "type": "boolean" }, "chebi_par_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "withdrawn_reason": { "type": "text" }, "molecule_hierarchy": { "properties": { "parent_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "prodrug": { "type": "integer" }, "withdrawn_flag": { "type": "boolean" }, "usan_year": { "type": "integer" }, "parenteral": { "type": "boolean" }, "black_box_warning": { "type": "integer" }, "polymer_flag": { "type": "boolean" }, "molecule_synonyms": { "properties": { "molecule_synonym": { "type": "text" }, "synonyms": { "type": "text" }, "syn_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "atc_classifications": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_type": { "type": "text" }, "first_in_class": { "type": "integer" }, "inchi": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "structure_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "withdrawn_class": { "type": "text" }, "inchi_key": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "topical": { "type": "boolean" }, "oral": { "type": "boolean" }, "xrefs": { "properties": { "drugcentral": { "properties": { "id": { "type": "integer" }, "name": { "type": "text" } } }, "tg-gates": { "properties": { "id": { "type": "integer" }, "name": { "type": "text" } } }, "wikipedia": { "properties": { "url_stub": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "dailymed": { "properties": { "name": { "type": "text" } } }, "pubchem": { "properties": { "sid": { "type": "integer" } } } } }, "chirality": { "type": "integer" }, "usan_substem": { "type": "text" }, "indication_class": { "type": "text" }, "withdrawn_country": { "type": "text" }, "withdrawn_year": { "type": "integer" }, "availability_type": { "type": "integer" }, "smiles": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "natural_product": { "type": "integer" }, "pref_name": { "type": "text", "copy_to": ["all"] }, "first_approval": { "type": "integer" } } } } return mapping
40.569921
95
0.280437
import os import glob import pymongo import biothings.hub.dataload.storage as storage from biothings.hub.dataload.uploader import ParallelizedSourceUploader from hub.dataload.uploader import BaseDrugUploader from hub.datatransform.keylookup import MyChemKeyLookup from .chembl_parser import load_data SRC_META = { "url": 'https://www.ebi.ac.uk/chembl/', "license_url" : "https://www.ebi.ac.uk/about/terms-of-use", "license_url_short" : "http://bit.ly/2KAUCAm" } class ChemblUploader(BaseDrugUploader, ParallelizedSourceUploader): name = "chembl" storage_class = storage.RootKeyMergerStorage __metadata__ = {"src_meta" : SRC_META} MOLECULE_PATTERN = "molecule.*.json" keylookup = MyChemKeyLookup( [("inchikey", "chembl.inchi_key"), ("inchi", "chembl.inchi"), ("chembl", "chembl.molecule_chembl_id"), ("chebi", "chembl.chebi_par_id"), ("drugcentral", "chembl.xrefs.drugcentral.id"), ("drugname", "chembl.pref_name")], copy_from_doc=True) def jobs(self): json_files = glob.glob(os.path.join(self.data_folder, self.__class__.MOLECULE_PATTERN)) return [(f,) for f in json_files] def load_data(self, data_folder): self.logger.info("Load data from '%s'" % data_folder) return self.keylookup(load_data, debug=True)(data_folder) def post_update_data(self, *args, **kwargs): for idxname in ["chembl.chebi_par_id", "chembl.molecule_chembl_id"]: self.logger.info("Indexing '%s'" % idxname) self.collection.create_index(idxname, background=True) @classmethod def get_mapping(cls): mapping = { "chembl": { "properties": { "biotherapeutic": { "properties": { "helm_notation": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "description": { "type": "text" }, "biocomponents": { "properties": { "organism": { "type": "text" }, "tax_id": { "type": "integer" }, "sequence": { "type": "text" }, "component_id": { "type": "integer" }, "description": { "type": "text" }, "component_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", 'copy_to': ['all'], } } }, "therapeutic_flag": { "type": "boolean" }, "usan_stem": { "type": "text" }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_properties": { "properties": { "heavy_atoms": { "type": "integer" }, "acd_most_bpka": { "type": "float" }, "mw_freebase": { "type": "float" }, "num_ro5_violations": { "type": "integer" }, "molecular_species": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "qed_weighted": { "type": "float" }, "ro3_pass": { "type": "boolean" }, "full_mwt": { "type": "float" }, "num_lipinski_ro5_violations": { "type": "integer" }, "rtb": { "type": "integer" }, "psa": { "type": "float" }, "alogp": { "type": "float" }, "hbd": { "type": "integer" }, "acd_most_apka": { "type": "float" }, "hbd_lipinski": { "type": "integer" }, "acd_logp": { "type": "float" }, "full_molformula": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "aromatic_rings": { "type": "integer" }, "hba_lipinski": { "type": "integer" }, "mw_monoisotopic": { "type": "float" }, "hba": { "type": "integer" }, "acd_logd": { "type": "float" } } }, "helm_notation": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "max_phase": { "type": "integer" }, "inorganic_flag": { "type": "integer" }, "usan_stem_definition": { "type": "text" }, "dosed_ingredient": { "type": "boolean" }, "chebi_par_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "withdrawn_reason": { "type": "text" }, "molecule_hierarchy": { "properties": { "parent_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_chembl_id": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "prodrug": { "type": "integer" }, "withdrawn_flag": { "type": "boolean" }, "usan_year": { "type": "integer" }, "parenteral": { "type": "boolean" }, "black_box_warning": { "type": "integer" }, "polymer_flag": { "type": "boolean" }, "molecule_synonyms": { "properties": { "molecule_synonym": { "type": "text" }, "synonyms": { "type": "text" }, "syn_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "atc_classifications": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "molecule_type": { "type": "text" }, "first_in_class": { "type": "integer" }, "inchi": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "structure_type": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "withdrawn_class": { "type": "text" }, "inchi_key": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "topical": { "type": "boolean" }, "oral": { "type": "boolean" }, "xrefs": { "properties": { "drugcentral": { "properties": { "id": { "type": "integer" }, "name": { "type": "text" } } }, "tg-gates": { "properties": { "id": { "type": "integer" }, "name": { "type": "text" } } }, "wikipedia": { "properties": { "url_stub": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", } } }, "dailymed": { "properties": { "name": { "type": "text" } } }, "pubchem": { "properties": { "sid": { "type": "integer" } } } } }, "chirality": { "type": "integer" }, "usan_substem": { "type": "text" }, "indication_class": { "type": "text" }, "withdrawn_country": { "type": "text" }, "withdrawn_year": { "type": "integer" }, "availability_type": { "type": "integer" }, "smiles": { "normalizer": "keyword_lowercase_normalizer", "type": "keyword", }, "natural_product": { "type": "integer" }, "pref_name": { "type": "text", "copy_to": ["all"] }, "first_approval": { "type": "integer" } } } } return mapping
true
true
f717bdc2664f609c84cf83efbdea10b954fadfe4
2,307
py
Python
tools/test_cate_attr_predictor.py
engahmed1190/mmfashion
34ba2d8a9f2daadb4a04d24287664cebde4b14f9
[ "Apache-2.0" ]
6
2021-02-04T06:25:28.000Z
2021-07-16T05:49:13.000Z
tools/test_cate_attr_predictor.py
engahmed1190/mmfashion
34ba2d8a9f2daadb4a04d24287664cebde4b14f9
[ "Apache-2.0" ]
null
null
null
tools/test_cate_attr_predictor.py
engahmed1190/mmfashion
34ba2d8a9f2daadb4a04d24287664cebde4b14f9
[ "Apache-2.0" ]
2
2021-07-16T19:27:20.000Z
2021-12-15T09:52:54.000Z
from __future__ import division import argparse from mmcv import Config from mmcv.runner import load_checkpoint from mmfashion.apis import get_root_logger, init_dist, test_cate_attr_predictor from mmfashion.datasets.utils import get_dataset from mmfashion.models import build_predictor def parse_args(): parser = argparse.ArgumentParser( description='Test a Fashion Attribute Predictor') parser.add_argument( '--config', help='test config file path', default='configs/category_attribute_predict/roi_predictor_vgg.py') parser.add_argument('--work_dir', help='the dir to save logs and models') parser.add_argument( '--checkpoint', help='checkpoint file', default='checkpoint/CateAttrPredict/vgg/roi/epoch_40.pth') parser.add_argument( '--validate', action='store_true', help='whether to evaluate the checkpoint during training', default=True) parser.add_argument( '--launcher', choices=['none', 'pytorch', 'mpi', 'slurm'], default='none', help='job launcher') args = parser.parse_args() return args def main(): args = parse_args() cfg = Config.fromfile(args.config) if args.work_dir is not None: cfg.work_dir = args.work_dir if args.checkpoint is not None: cfg.load_from = args.checkpoint # init distributed env first if args.launcher == 'none': distributed = False else: distributed = True init_dist(args.launcher, **cfg.dist_params) # init logger logger = get_root_logger(cfg.log_level) logger.info('Distributed test: {}'.format(distributed)) # data loader dataset = get_dataset(cfg.data.test) print('dataset loaded') # build model and load checkpoint model = build_predictor(cfg.model) print('model built') load_checkpoint(model, cfg.load_from, map_location='cpu') print('load checkpoint from: {}'.format(cfg.load_from)) # test test_cate_attr_predictor(model, dataset, cfg, distributed=distributed, validate=args.validate, logger=logger) if __name__ == '__main__': main()
29.576923
79
0.638058
from __future__ import division import argparse from mmcv import Config from mmcv.runner import load_checkpoint from mmfashion.apis import get_root_logger, init_dist, test_cate_attr_predictor from mmfashion.datasets.utils import get_dataset from mmfashion.models import build_predictor def parse_args(): parser = argparse.ArgumentParser( description='Test a Fashion Attribute Predictor') parser.add_argument( '--config', help='test config file path', default='configs/category_attribute_predict/roi_predictor_vgg.py') parser.add_argument('--work_dir', help='the dir to save logs and models') parser.add_argument( '--checkpoint', help='checkpoint file', default='checkpoint/CateAttrPredict/vgg/roi/epoch_40.pth') parser.add_argument( '--validate', action='store_true', help='whether to evaluate the checkpoint during training', default=True) parser.add_argument( '--launcher', choices=['none', 'pytorch', 'mpi', 'slurm'], default='none', help='job launcher') args = parser.parse_args() return args def main(): args = parse_args() cfg = Config.fromfile(args.config) if args.work_dir is not None: cfg.work_dir = args.work_dir if args.checkpoint is not None: cfg.load_from = args.checkpoint if args.launcher == 'none': distributed = False else: distributed = True init_dist(args.launcher, **cfg.dist_params) logger = get_root_logger(cfg.log_level) logger.info('Distributed test: {}'.format(distributed)) dataset = get_dataset(cfg.data.test) print('dataset loaded') model = build_predictor(cfg.model) print('model built') load_checkpoint(model, cfg.load_from, map_location='cpu') print('load checkpoint from: {}'.format(cfg.load_from)) test_cate_attr_predictor(model, dataset, cfg, distributed=distributed, validate=args.validate, logger=logger) if __name__ == '__main__': main()
true
true
f717bedbafbc1c240de6e34e3435337fe00e3c60
4,260
py
Python
app/interfaces/Rubric.py
u8sand/FAIRshake
8f6f3dde42de29b88e9a43bdd43f848382e3bad7
[ "Apache-2.0" ]
null
null
null
app/interfaces/Rubric.py
u8sand/FAIRshake
8f6f3dde42de29b88e9a43bdd43f848382e3bad7
[ "Apache-2.0" ]
8
2018-06-05T17:01:43.000Z
2018-06-22T01:19:39.000Z
app/interfaces/Rubric.py
u8sand/FAIRshake
8f6f3dde42de29b88e9a43bdd43f848382e3bad7
[ "Apache-2.0" ]
1
2018-06-06T17:22:28.000Z
2018-06-06T17:22:28.000Z
from ..ioc import interface, model from ..types import HTTPResponse, UUID, Timestamp, Optional, List from ..util.generate_spec import generate_spec @model class CriterionModel: ''' type: object required: - user - name properties: id: type: string description: Criterion ID example: d290f1ee-6c54-4b01-90e6-d701748f0851 name: type: string description: Criterion being evaluated example: Coolness kind: type: string description: Value type of the criterion example: number ''' id: UUID name: str kind: str @model class RubricModel: ''' type: object description: An rubric for evaluating a digital object required: - user - name properties: id: type: string format: uuid description: ID of rubric example: d290f1ee-6c54-4b01-90e6-d701748f0851 user: type: string format: uuid description: ID of user who made the rubric example: d290f1ee-6c54-4b01-90e6-d701748f0851 name: type: string description: Name of the rubric example: Best Rubric criteria: type: array items: $ref: "#/definitions/Criterion" description: type: string description: Description of the rubric example: My favorite rubric tags: type: array items: type: string example: tool timestamp: type: string description: When the rubric was made format: dateTime example: '2018-05-20T15:59:59-08:00' ''' id: UUID user: UUID name: str criteria: List[CriterionModel] description: Optional[str] = None tags: Optional[List[str]] = None timestamp: Optional[Timestamp] = None @interface class RubricAPI: ''' swagger: '2.0' info: title: FAIRshakeRubric version: 1.0.0 description: A generic FAIRshake Rubric REST API for storing questions to be answered. contact: email: daniel.clarke@mssm.edu license: name: Apache 2.0 url: http://www.apache.org/licenses/LICENSE-2.0.html schemes: - https securityDefinitions: ApiKeyAuth: type: apiKey in: header name: X-API-Key paths: /: get: {RubricAPI__get} post: {RubricAPI__post} definitions: Criterion: {CriterionModel} Rubric: {RubricModel} ''' @staticmethod def get( id: Optional[UUID] = None, user: Optional[UUID] = None, object: Optional[UUID] = None, timestamp: Optional[Timestamp] = None, skip: Optional[int] = None, limit: Optional[int] = None, ) -> HTTPResponse[List[RubricModel]]: ''' summary: Query rubrics consumes: - application/json parameters: - name: id in: query description: Unique rubric identifier type: string format: uuid - name: user in: query description: Unique user identifier - name: object in: query description: Unique digital object identifier type: string format: uuid - name: timestamp in: query description: Updated at least since timestamp type: string format: dateTime - name: skip in: query description: Start index for limited results type: number - name: limit in: query description: Limit number of results type: number produces: - application/json responses: 200: description: Rubrics as json schema: type: array items: $ref: "#/definitions/Rubric" 404: description: Rubrics not found ''' raise NotImplemented @staticmethod def post( body: RubricModel ) -> HTTPResponse[None]: ''' summary: Submit rubric consumes: - application/json security: - ApiKeyAuth: [] parameters: - name: body in: body schema: $ref: "#/definitions/Rubric" produces: - application/json responses: 201: description: Rubric successfully made 400: description: Parameters invalid 500: description: Permission denied, not authenticated ''' raise NotImplemented RubricSpec = generate_spec(RubricAPI, [RubricModel, CriterionModel])
22.303665
90
0.621831
from ..ioc import interface, model from ..types import HTTPResponse, UUID, Timestamp, Optional, List from ..util.generate_spec import generate_spec @model class CriterionModel: id: UUID name: str kind: str @model class RubricModel: id: UUID user: UUID name: str criteria: List[CriterionModel] description: Optional[str] = None tags: Optional[List[str]] = None timestamp: Optional[Timestamp] = None @interface class RubricAPI: @staticmethod def get( id: Optional[UUID] = None, user: Optional[UUID] = None, object: Optional[UUID] = None, timestamp: Optional[Timestamp] = None, skip: Optional[int] = None, limit: Optional[int] = None, ) -> HTTPResponse[List[RubricModel]]: raise NotImplemented @staticmethod def post( body: RubricModel ) -> HTTPResponse[None]: raise NotImplemented RubricSpec = generate_spec(RubricAPI, [RubricModel, CriterionModel])
true
true
f717c0006f023d6f8d6d6a2811a17dfa5060ba8c
72,241
py
Python
mne/bem.py
jdammers/mne-python
1dc1502215a53385cda15c6c336fcc4341dc4d3b
[ "BSD-3-Clause" ]
null
null
null
mne/bem.py
jdammers/mne-python
1dc1502215a53385cda15c6c336fcc4341dc4d3b
[ "BSD-3-Clause" ]
null
null
null
mne/bem.py
jdammers/mne-python
1dc1502215a53385cda15c6c336fcc4341dc4d3b
[ "BSD-3-Clause" ]
1
2021-04-12T12:45:31.000Z
2021-04-12T12:45:31.000Z
# Authors: Alexandre Gramfort <alexandre.gramfort@telecom-paristech.fr> # Matti Hamalainen <msh@nmr.mgh.harvard.edu> # Eric Larson <larson.eric.d@gmail.com> # Lorenzo De Santis <lorenzo.de-santis@u-psud.fr> # # License: BSD (3-clause) from functools import partial import glob import os import os.path as op import shutil from copy import deepcopy import numpy as np from scipy import linalg from .transforms import _ensure_trans, apply_trans from .io import Info from .io.constants import FIFF from .io.write import (start_file, start_block, write_float, write_int, write_float_matrix, write_int_matrix, end_block, end_file) from .io.tag import find_tag from .io.tree import dir_tree_find from .io.open import fiff_open from .surface import (read_surface, write_surface, complete_surface_info, _compute_nearest, _get_ico_surface, read_tri, _fast_cross_nd_sum, _get_solids) from .utils import verbose, logger, run_subprocess, get_subjects_dir, warn, _pl from .fixes import einsum from .externals.six import string_types # ############################################################################ # Compute BEM solution # The following approach is based on: # # de Munck JC: "A linear discretization of the volume conductor boundary # integral equation using analytically integrated elements", # IEEE Trans Biomed Eng. 1992 39(9) : 986 - 990 # class ConductorModel(dict): """BEM or sphere model.""" def __repr__(self): # noqa: D105 if self['is_sphere']: center = ', '.join('%0.1f' % (x * 1000.) for x in self['r0']) rad = self.radius if rad is None: # no radius / MEG only extra = 'Sphere (no layers): r0=[%s] mm' % center else: extra = ('Sphere (%s layer%s): r0=[%s] R=%1.f mm' % (len(self['layers']) - 1, _pl(self['layers']), center, rad * 1000.)) else: extra = ('BEM (%s layer%s)' % (len(self['surfs']), _pl(self['surfs']))) return '<ConductorModel | %s>' % extra def copy(self): """Return copy of ConductorModel instance.""" return deepcopy(self) @property def radius(self): """Sphere radius if an EEG sphere model.""" if not self['is_sphere']: raise RuntimeError('radius undefined for BEM') return None if len(self['layers']) == 0 else self['layers'][-1]['rad'] def _calc_beta(rk, rk_norm, rk1, rk1_norm): """Compute coefficients for calculating the magic vector omega.""" rkk1 = rk1[0] - rk[0] size = np.linalg.norm(rkk1) rkk1 /= size num = rk_norm + np.dot(rk, rkk1) den = rk1_norm + np.dot(rk1, rkk1) res = np.log(num / den) / size return res def _lin_pot_coeff(fros, tri_rr, tri_nn, tri_area): """Compute the linear potential matrix element computations.""" omega = np.zeros((len(fros), 3)) # we replicate a little bit of the _get_solids code here for speed # (we need some of the intermediate values later) v1 = tri_rr[np.newaxis, 0, :] - fros v2 = tri_rr[np.newaxis, 1, :] - fros v3 = tri_rr[np.newaxis, 2, :] - fros triples = _fast_cross_nd_sum(v1, v2, v3) l1 = np.linalg.norm(v1, axis=1) l2 = np.linalg.norm(v2, axis=1) l3 = np.linalg.norm(v3, axis=1) ss = l1 * l2 * l3 ss += einsum('ij,ij,i->i', v1, v2, l3) ss += einsum('ij,ij,i->i', v1, v3, l2) ss += einsum('ij,ij,i->i', v2, v3, l1) solids = np.arctan2(triples, ss) # We *could* subselect the good points from v1, v2, v3, triples, solids, # l1, l2, and l3, but there are *very* few bad points. So instead we do # some unnecessary calculations, and then omit them from the final # solution. These three lines ensure we don't get invalid values in # _calc_beta. bad_mask = np.abs(solids) < np.pi / 1e6 l1[bad_mask] = 1. l2[bad_mask] = 1. l3[bad_mask] = 1. # Calculate the magic vector vec_omega beta = [_calc_beta(v1, l1, v2, l2)[:, np.newaxis], _calc_beta(v2, l2, v3, l3)[:, np.newaxis], _calc_beta(v3, l3, v1, l1)[:, np.newaxis]] vec_omega = (beta[2] - beta[0]) * v1 vec_omega += (beta[0] - beta[1]) * v2 vec_omega += (beta[1] - beta[2]) * v3 area2 = 2.0 * tri_area n2 = 1.0 / (area2 * area2) # leave omega = 0 otherwise # Put it all together... yys = [v1, v2, v3] idx = [0, 1, 2, 0, 2] for k in range(3): diff = yys[idx[k - 1]] - yys[idx[k + 1]] zdots = _fast_cross_nd_sum(yys[idx[k + 1]], yys[idx[k - 1]], tri_nn) omega[:, k] = -n2 * (area2 * zdots * 2. * solids - triples * (diff * vec_omega).sum(axis=-1)) # omit the bad points from the solution omega[bad_mask] = 0. return omega def _correct_auto_elements(surf, mat): """Improve auto-element approximation.""" pi2 = 2.0 * np.pi tris_flat = surf['tris'].ravel() misses = pi2 - mat.sum(axis=1) for j, miss in enumerate(misses): # How much is missing? n_memb = len(surf['neighbor_tri'][j]) # The node itself receives one half mat[j, j] = miss / 2.0 # The rest is divided evenly among the member nodes... miss /= (4.0 * n_memb) members = np.where(j == tris_flat)[0] mods = members % 3 offsets = np.array([[1, 2], [-1, 1], [-1, -2]]) tri_1 = members + offsets[mods, 0] tri_2 = members + offsets[mods, 1] for t1, t2 in zip(tri_1, tri_2): mat[j, tris_flat[t1]] += miss mat[j, tris_flat[t2]] += miss return def _fwd_bem_lin_pot_coeff(surfs): """Calculate the coefficients for linear collocation approach.""" # taken from fwd_bem_linear_collocation.c nps = [surf['np'] for surf in surfs] np_tot = sum(nps) coeff = np.zeros((np_tot, np_tot)) offsets = np.cumsum(np.concatenate(([0], nps))) for si_1, surf1 in enumerate(surfs): rr_ord = np.arange(nps[si_1]) for si_2, surf2 in enumerate(surfs): logger.info(" %s (%d) -> %s (%d) ..." % (_bem_explain_surface(surf1['id']), nps[si_1], _bem_explain_surface(surf2['id']), nps[si_2])) tri_rr = surf2['rr'][surf2['tris']] tri_nn = surf2['tri_nn'] tri_area = surf2['tri_area'] submat = coeff[offsets[si_1]:offsets[si_1 + 1], offsets[si_2]:offsets[si_2 + 1]] # view for k in range(surf2['ntri']): tri = surf2['tris'][k] if si_1 == si_2: skip_idx = ((rr_ord == tri[0]) | (rr_ord == tri[1]) | (rr_ord == tri[2])) else: skip_idx = list() # No contribution from a triangle that # this vertex belongs to # if sidx1 == sidx2 and (tri == j).any(): # continue # Otherwise do the hard job coeffs = _lin_pot_coeff(surf1['rr'], tri_rr[k], tri_nn[k], tri_area[k]) coeffs[skip_idx] = 0. submat[:, tri] -= coeffs if si_1 == si_2: _correct_auto_elements(surf1, submat) return coeff def _fwd_bem_multi_solution(solids, gamma, nps): """Do multi surface solution. * Invert I - solids/(2*M_PI) * Take deflation into account * The matrix is destroyed after inversion * This is the general multilayer case """ pi2 = 1.0 / (2 * np.pi) n_tot = np.sum(nps) assert solids.shape == (n_tot, n_tot) nsurf = len(nps) defl = 1.0 / n_tot # Modify the matrix offsets = np.cumsum(np.concatenate(([0], nps))) for si_1 in range(nsurf): for si_2 in range(nsurf): mult = pi2 if gamma is None else pi2 * gamma[si_1, si_2] slice_j = slice(offsets[si_1], offsets[si_1 + 1]) slice_k = slice(offsets[si_2], offsets[si_2 + 1]) solids[slice_j, slice_k] = defl - solids[slice_j, slice_k] * mult solids += np.eye(n_tot) return linalg.inv(solids, overwrite_a=True) def _fwd_bem_homog_solution(solids, nps): """Make a homogeneous solution.""" return _fwd_bem_multi_solution(solids, None, nps) def _fwd_bem_ip_modify_solution(solution, ip_solution, ip_mult, n_tri): """Modify the solution according to the IP approach.""" n_last = n_tri[-1] mult = (1.0 + ip_mult) / ip_mult logger.info(' Combining...') offsets = np.cumsum(np.concatenate(([0], n_tri))) for si in range(len(n_tri)): # Pick the correct submatrix (right column) and multiply sub = solution[offsets[si]:offsets[si + 1], np.sum(n_tri[:-1]):] # Multiply sub -= 2 * np.dot(sub, ip_solution) # The lower right corner is a special case sub[-n_last:, -n_last:] += mult * ip_solution # Final scaling logger.info(' Scaling...') solution *= ip_mult return def _fwd_bem_linear_collocation_solution(m): """Compute the linear collocation potential solution.""" # first, add surface geometries for surf in m['surfs']: complete_surface_info(surf, copy=False, verbose=False) logger.info('Computing the linear collocation solution...') logger.info(' Matrix coefficients...') coeff = _fwd_bem_lin_pot_coeff(m['surfs']) m['nsol'] = len(coeff) logger.info(" Inverting the coefficient matrix...") nps = [surf['np'] for surf in m['surfs']] m['solution'] = _fwd_bem_multi_solution(coeff, m['gamma'], nps) if len(m['surfs']) == 3: ip_mult = m['sigma'][1] / m['sigma'][2] if ip_mult <= FIFF.FWD_BEM_IP_APPROACH_LIMIT: logger.info('IP approach required...') logger.info(' Matrix coefficients (homog)...') coeff = _fwd_bem_lin_pot_coeff([m['surfs'][-1]]) logger.info(' Inverting the coefficient matrix (homog)...') ip_solution = _fwd_bem_homog_solution(coeff, [m['surfs'][-1]['np']]) logger.info(' Modify the original solution to incorporate ' 'IP approach...') _fwd_bem_ip_modify_solution(m['solution'], ip_solution, ip_mult, nps) m['bem_method'] = FIFF.FWD_BEM_LINEAR_COLL logger.info("Solution ready.") @verbose def make_bem_solution(surfs, verbose=None): """Create a BEM solution using the linear collocation approach. Parameters ---------- surfs : list of dict The BEM surfaces to use (`from make_bem_model`) verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- bem : instance of ConductorModel The BEM solution. Notes ----- .. versionadded:: 0.10.0 See Also -------- make_bem_model read_bem_surfaces write_bem_surfaces read_bem_solution write_bem_solution """ logger.info('Approximation method : Linear collocation\n') if isinstance(surfs, string_types): # Load the surfaces logger.info('Loading surfaces...') surfs = read_bem_surfaces(surfs) bem = ConductorModel(is_sphere=False, surfs=surfs) _add_gamma_multipliers(bem) if len(bem['surfs']) == 3: logger.info('Three-layer model surfaces loaded.') elif len(bem['surfs']) == 1: logger.info('Homogeneous model surface loaded.') else: raise RuntimeError('Only 1- or 3-layer BEM computations supported') _check_bem_size(bem['surfs']) _fwd_bem_linear_collocation_solution(bem) logger.info('BEM geometry computations complete.') return bem # ############################################################################ # Make BEM model def _ico_downsample(surf, dest_grade): """Downsample the surface if isomorphic to a subdivided icosahedron.""" n_tri = len(surf['tris']) found = -1 bad_msg = ("A surface with %d triangles cannot be isomorphic with a " "subdivided icosahedron." % n_tri) if n_tri % 20 != 0: raise RuntimeError(bad_msg) n_tri = n_tri // 20 found = int(round(np.log(n_tri) / np.log(4))) if n_tri != 4 ** found: raise RuntimeError(bad_msg) del n_tri if dest_grade > found: raise RuntimeError('For this surface, decimation grade should be %d ' 'or less, not %s.' % (found, dest_grade)) source = _get_ico_surface(found) dest = _get_ico_surface(dest_grade, patch_stats=True) del dest['tri_cent'] del dest['tri_nn'] del dest['neighbor_tri'] del dest['tri_area'] if not np.array_equal(source['tris'], surf['tris']): raise RuntimeError('The source surface has a matching number of ' 'triangles but ordering is wrong') logger.info('Going from %dth to %dth subdivision of an icosahedron ' '(n_tri: %d -> %d)' % (found, dest_grade, len(surf['tris']), len(dest['tris']))) # Find the mapping dest['rr'] = surf['rr'][_get_ico_map(source, dest)] return dest def _get_ico_map(fro, to): """Get a mapping between ico surfaces.""" nearest, dists = _compute_nearest(fro['rr'], to['rr'], return_dists=True) n_bads = (dists > 5e-3).sum() if n_bads > 0: raise RuntimeError('No matching vertex for %d destination vertices' % (n_bads)) return nearest def _order_surfaces(surfs): """Reorder the surfaces.""" if len(surfs) != 3: return surfs # we have three surfaces surf_order = [FIFF.FIFFV_BEM_SURF_ID_HEAD, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_BRAIN] ids = np.array([surf['id'] for surf in surfs]) if set(ids) != set(surf_order): raise RuntimeError('bad surface ids: %s' % ids) order = [np.where(ids == id_)[0][0] for id_ in surf_order] surfs = [surfs[idx] for idx in order] return surfs def _assert_complete_surface(surf, incomplete='raise'): """Check the sum of solid angles as seen from inside.""" # from surface_checks.c tot_angle = 0. # Center of mass.... cm = surf['rr'].mean(axis=0) logger.info('%s CM is %6.2f %6.2f %6.2f mm' % (_surf_name[surf['id']], 1000 * cm[0], 1000 * cm[1], 1000 * cm[2])) tot_angle = _get_solids(surf['rr'][surf['tris']], cm[np.newaxis, :])[0] prop = tot_angle / (2 * np.pi) if np.abs(prop - 1.0) > 1e-5: msg = ('Surface %s is not complete (sum of solid angles ' 'yielded %g, should be 1.)' % (_surf_name[surf['id']], prop)) if incomplete == 'raise': raise RuntimeError(msg) else: warn(msg) _surf_name = { FIFF.FIFFV_BEM_SURF_ID_HEAD: 'outer skin ', FIFF.FIFFV_BEM_SURF_ID_SKULL: 'outer skull', FIFF.FIFFV_BEM_SURF_ID_BRAIN: 'inner skull', FIFF.FIFFV_BEM_SURF_ID_UNKNOWN: 'unknown ', } def _assert_inside(fro, to): """Check one set of points is inside a surface.""" # this is "is_inside" in surface_checks.c tot_angle = _get_solids(to['rr'][to['tris']], fro['rr']) if (np.abs(tot_angle / (2 * np.pi) - 1.0) > 1e-5).any(): raise RuntimeError('Surface %s is not completely inside surface %s' % (_surf_name[fro['id']], _surf_name[to['id']])) def _check_surfaces(surfs, incomplete='raise'): """Check that the surfaces are complete and non-intersecting.""" for surf in surfs: _assert_complete_surface(surf, incomplete=incomplete) # Then check the topology for surf_1, surf_2 in zip(surfs[:-1], surfs[1:]): logger.info('Checking that %s surface is inside %s surface...' % (_surf_name[surf_2['id']], _surf_name[surf_1['id']])) _assert_inside(surf_2, surf_1) def _check_surface_size(surf): """Check that the coordinate limits are reasonable.""" sizes = surf['rr'].max(axis=0) - surf['rr'].min(axis=0) if (sizes < 0.05).any(): raise RuntimeError('Dimensions of the surface %s seem too small ' '(%9.5f mm). Maybe the the unit of measure is ' 'meters instead of mm' % (_surf_name[surf['id']], 1000 * sizes.min())) def _check_thicknesses(surfs): """Compute how close we are.""" for surf_1, surf_2 in zip(surfs[:-1], surfs[1:]): min_dist = _compute_nearest(surf_1['rr'], surf_2['rr'], return_dists=True)[0] min_dist = min_dist.min() logger.info('Checking distance between %s and %s surfaces...' % (_surf_name[surf_1['id']], _surf_name[surf_2['id']])) logger.info('Minimum distance between the %s and %s surfaces is ' 'approximately %6.1f mm' % (_surf_name[surf_1['id']], _surf_name[surf_2['id']], 1000 * min_dist)) def _surfaces_to_bem(surfs, ids, sigmas, ico=None, rescale=True, incomplete='raise'): """Convert surfaces to a BEM.""" # equivalent of mne_surf2bem # surfs can be strings (filenames) or surface dicts if len(surfs) not in (1, 3) or not (len(surfs) == len(ids) == len(sigmas)): raise ValueError('surfs, ids, and sigmas must all have the same ' 'number of elements (1 or 3)') surf = list(surfs) for si, surf in enumerate(surfs): if isinstance(surf, string_types): surfs[si] = read_surface(surf, return_dict=True)[-1] # Downsampling if the surface is isomorphic with a subdivided icosahedron if ico is not None: for si, surf in enumerate(surfs): surfs[si] = _ico_downsample(surf, ico) for surf, id_ in zip(surfs, ids): surf['id'] = id_ surf['coord_frame'] = surf.get('coord_frame', FIFF.FIFFV_COORD_MRI) surf.update(np=len(surf['rr']), ntri=len(surf['tris'])) if rescale: surf['rr'] /= 1000. # convert to meters # Shifting surfaces is not implemented here... # Order the surfaces for the benefit of the topology checks for surf, sigma in zip(surfs, sigmas): surf['sigma'] = sigma surfs = _order_surfaces(surfs) # Check topology as best we can _check_surfaces(surfs, incomplete=incomplete) for surf in surfs: _check_surface_size(surf) _check_thicknesses(surfs) logger.info('Surfaces passed the basic topology checks.') return surfs @verbose def make_bem_model(subject, ico=4, conductivity=(0.3, 0.006, 0.3), subjects_dir=None, verbose=None): """Create a BEM model for a subject. .. note:: To get a single layer bem corresponding to the --homog flag in the command line tool set the ``conductivity`` parameter to a list/tuple with a single value (e.g. [0.3]). Parameters ---------- subject : str The subject. ico : int | None The surface ico downsampling to use, e.g. 5=20484, 4=5120, 3=1280. If None, no subsampling is applied. conductivity : array of int, shape (3,) or (1,) The conductivities to use for each shell. Should be a single element for a one-layer model, or three elements for a three-layer model. Defaults to ``[0.3, 0.006, 0.3]``. The MNE-C default for a single-layer model would be ``[0.3]``. subjects_dir : string, or None Path to SUBJECTS_DIR if it is not set in the environment. verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- surfaces : list of dict The BEM surfaces. Use `make_bem_solution` to turn these into a `ConductorModel` suitable for forward calculation. Notes ----- .. versionadded:: 0.10.0 See Also -------- make_bem_solution make_sphere_model read_bem_surfaces write_bem_surfaces """ conductivity = np.array(conductivity, float) if conductivity.ndim != 1 or conductivity.size not in (1, 3): raise ValueError('conductivity must be 1D array-like with 1 or 3 ' 'elements') subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) subject_dir = op.join(subjects_dir, subject) bem_dir = op.join(subject_dir, 'bem') inner_skull = op.join(bem_dir, 'inner_skull.surf') outer_skull = op.join(bem_dir, 'outer_skull.surf') outer_skin = op.join(bem_dir, 'outer_skin.surf') surfaces = [inner_skull, outer_skull, outer_skin] ids = [FIFF.FIFFV_BEM_SURF_ID_BRAIN, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_HEAD] logger.info('Creating the BEM geometry...') if len(conductivity) == 1: surfaces = surfaces[:1] ids = ids[:1] surfaces = _surfaces_to_bem(surfaces, ids, conductivity, ico) _check_bem_size(surfaces) logger.info('Complete.\n') return surfaces # ############################################################################ # Compute EEG sphere model def _fwd_eeg_get_multi_sphere_model_coeffs(m, n_terms): """Get the model depended weighting factor for n.""" nlayer = len(m['layers']) if nlayer in (0, 1): return 1. # Initialize the arrays c1 = np.zeros(nlayer - 1) c2 = np.zeros(nlayer - 1) cr = np.zeros(nlayer - 1) cr_mult = np.zeros(nlayer - 1) for k in range(nlayer - 1): c1[k] = m['layers'][k]['sigma'] / m['layers'][k + 1]['sigma'] c2[k] = c1[k] - 1.0 cr_mult[k] = m['layers'][k]['rel_rad'] cr[k] = cr_mult[k] cr_mult[k] *= cr_mult[k] coeffs = np.zeros(n_terms - 1) for n in range(1, n_terms): # Increment the radius coefficients for k in range(nlayer - 1): cr[k] *= cr_mult[k] # Multiply the matrices M = np.eye(2) n1 = n + 1.0 for k in range(nlayer - 2, -1, -1): M = np.dot([[n + n1 * c1[k], n1 * c2[k] / cr[k]], [n * c2[k] * cr[k], n1 + n * c1[k]]], M) num = n * (2.0 * n + 1.0) ** (nlayer - 1) coeffs[n - 1] = num / (n * M[1, 1] + n1 * M[1, 0]) return coeffs def _compose_linear_fitting_data(mu, u): """Get the linear fitting data.""" # y is the data to be fitted (nterms-1 x 1) # M is the model matrix (nterms-1 x nfit-1) for k in range(u['nterms'] - 1): k1 = k + 1 mu1n = np.power(mu[0], k1) u['y'][k] = u['w'][k] * (u['fn'][k1] - mu1n * u['fn'][0]) for p in range(u['nfit'] - 1): u['M'][k][p] = u['w'][k] * (np.power(mu[p + 1], k1) - mu1n) def _compute_linear_parameters(mu, u): """Compute the best-fitting linear parameters.""" _compose_linear_fitting_data(mu, u) uu, sing, vv = linalg.svd(u['M'], full_matrices=False) # Compute the residuals u['resi'] = u['y'].copy() vec = np.empty(u['nfit'] - 1) for p in range(u['nfit'] - 1): vec[p] = np.dot(uu[:, p], u['y']) for k in range(u['nterms'] - 1): u['resi'][k] -= uu[k, p] * vec[p] vec[p] = vec[p] / sing[p] lambda_ = np.zeros(u['nfit']) for p in range(u['nfit'] - 1): sum_ = 0. for q in range(u['nfit'] - 1): sum_ += vv[q, p] * vec[q] lambda_[p + 1] = sum_ lambda_[0] = u['fn'][0] - np.sum(lambda_[1:]) rv = np.dot(u['resi'], u['resi']) / np.dot(u['y'], u['y']) return rv, lambda_ def _one_step(mu, u): """Evaluate the residual sum of squares fit for one set of mu values.""" if np.abs(mu).max() > 1.0: return 1.0 # Compose the data for the linear fitting, compute SVD, then residuals _compose_linear_fitting_data(mu, u) u['uu'], u['sing'], u['vv'] = linalg.svd(u['M']) u['resi'][:] = u['y'][:] for p in range(u['nfit'] - 1): dot = np.dot(u['uu'][p], u['y']) for k in range(u['nterms'] - 1): u['resi'][k] = u['resi'][k] - u['uu'][p, k] * dot # Return their sum of squares return np.dot(u['resi'], u['resi']) def _fwd_eeg_fit_berg_scherg(m, nterms, nfit): """Fit the Berg-Scherg equivalent spherical model dipole parameters.""" from scipy.optimize import fmin_cobyla assert nfit >= 2 u = dict(y=np.zeros(nterms - 1), resi=np.zeros(nterms - 1), nfit=nfit, nterms=nterms, M=np.zeros((nterms - 1, nfit - 1))) # (1) Calculate the coefficients of the true expansion u['fn'] = _fwd_eeg_get_multi_sphere_model_coeffs(m, nterms + 1) # (2) Calculate the weighting f = (min([layer['rad'] for layer in m['layers']]) / max([layer['rad'] for layer in m['layers']])) # correct weighting k = np.arange(1, nterms + 1) u['w'] = np.sqrt((2.0 * k + 1) * (3.0 * k + 1.0) / k) * np.power(f, (k - 1.0)) u['w'][-1] = 0 # Do the nonlinear minimization, constraining mu to the interval [-1, +1] mu_0 = np.random.RandomState(0).rand(nfit) * f fun = partial(_one_step, u=u) max_ = 1. - 2e-4 # adjust for fmin_cobyla "catol" that not all scipy have cons = [(lambda x: max_ - np.abs(x[ii])) for ii in range(nfit)] mu = fmin_cobyla(fun, mu_0, cons, rhobeg=0.5, rhoend=5e-3, disp=0) # (6) Do the final step: calculation of the linear parameters rv, lambda_ = _compute_linear_parameters(mu, u) order = np.argsort(mu)[::-1] mu, lambda_ = mu[order], lambda_[order] # sort: largest mu first m['mu'] = mu # This division takes into account the actual conductivities m['lambda'] = lambda_ / m['layers'][-1]['sigma'] m['nfit'] = nfit return rv @verbose def make_sphere_model(r0=(0., 0., 0.04), head_radius=0.09, info=None, relative_radii=(0.90, 0.92, 0.97, 1.0), sigmas=(0.33, 1.0, 0.004, 0.33), verbose=None): """Create a spherical model for forward solution calculation. Parameters ---------- r0 : array-like | str Head center to use (in head coordinates). If 'auto', the head center will be calculated from the digitization points in info. head_radius : float | str | None If float, compute spherical shells for EEG using the given radius. If 'auto', estimate an approriate radius from the dig points in Info, If None, exclude shells (single layer sphere model). info : instance of Info | None Measurement info. Only needed if ``r0`` or ``head_radius`` are ``'auto'``. relative_radii : array-like Relative radii for the spherical shells. sigmas : array-like Sigma values for the spherical shells. verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- sphere : instance of ConductorModel The resulting spherical conductor model. Notes ----- .. versionadded:: 0.9.0 See Also -------- make_bem_model make_bem_solution """ for name in ('r0', 'head_radius'): param = locals()[name] if isinstance(param, string_types): if param != 'auto': raise ValueError('%s, if str, must be "auto" not "%s"' % (name, param)) relative_radii = np.array(relative_radii, float).ravel() sigmas = np.array(sigmas, float).ravel() if len(relative_radii) != len(sigmas): raise ValueError('relative_radii length (%s) must match that of ' 'sigmas (%s)' % (len(relative_radii), len(sigmas))) if len(sigmas) <= 1 and head_radius is not None: raise ValueError('at least 2 sigmas must be supplied if ' 'head_radius is not None, got %s' % (len(sigmas),)) if (isinstance(r0, string_types) and r0 == 'auto') or \ (isinstance(head_radius, string_types) and head_radius == 'auto'): if info is None: raise ValueError('Info must not be None for auto mode') head_radius_fit, r0_fit = fit_sphere_to_headshape(info, units='m')[:2] if isinstance(r0, string_types): r0 = r0_fit if isinstance(head_radius, string_types): head_radius = head_radius_fit sphere = ConductorModel(is_sphere=True, r0=np.array(r0), coord_frame=FIFF.FIFFV_COORD_HEAD) sphere['layers'] = list() if head_radius is not None: # Eventually these could be configurable... relative_radii = np.array(relative_radii, float) sigmas = np.array(sigmas, float) order = np.argsort(relative_radii) relative_radii = relative_radii[order] sigmas = sigmas[order] for rel_rad, sig in zip(relative_radii, sigmas): # sort layers by (relative) radius, and scale radii layer = dict(rad=rel_rad, sigma=sig) layer['rel_rad'] = layer['rad'] = rel_rad sphere['layers'].append(layer) # scale the radii R = sphere['layers'][-1]['rad'] rR = sphere['layers'][-1]['rel_rad'] for layer in sphere['layers']: layer['rad'] /= R layer['rel_rad'] /= rR # # Setup the EEG sphere model calculations # # Scale the relative radii for k in range(len(relative_radii)): sphere['layers'][k]['rad'] = (head_radius * sphere['layers'][k]['rel_rad']) rv = _fwd_eeg_fit_berg_scherg(sphere, 200, 3) logger.info('\nEquiv. model fitting -> RV = %g %%' % (100 * rv)) for k in range(3): logger.info('mu%d = %g lambda%d = %g' % (k + 1, sphere['mu'][k], k + 1, sphere['layers'][-1]['sigma'] * sphere['lambda'][k])) logger.info('Set up EEG sphere model with scalp radius %7.1f mm\n' % (1000 * head_radius,)) return sphere # ############################################################################# # Sphere fitting _dig_kind_dict = { 'cardinal': FIFF.FIFFV_POINT_CARDINAL, 'hpi': FIFF.FIFFV_POINT_HPI, 'eeg': FIFF.FIFFV_POINT_EEG, 'extra': FIFF.FIFFV_POINT_EXTRA, } _dig_kind_rev = dict((val, key) for key, val in _dig_kind_dict.items()) _dig_kind_ints = tuple(_dig_kind_dict.values()) @verbose def fit_sphere_to_headshape(info, dig_kinds='auto', units='m', verbose=None): """Fit a sphere to the headshape points to determine head center. Parameters ---------- info : instance of Info Measurement info. dig_kinds : list of str | str Kind of digitization points to use in the fitting. These can be any combination of ('cardinal', 'hpi', 'eeg', 'extra'). Can also be 'auto' (default), which will use only the 'extra' points if enough (more than 10) are available, and if not, uses 'extra' and 'eeg' points. units : str Can be "m" (default) or "mm". .. versionadded:: 0.12 verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- radius : float Sphere radius. origin_head: ndarray, shape (3,) Head center in head coordinates. origin_device: ndarray, shape (3,) Head center in device coordinates. Notes ----- This function excludes any points that are low and frontal (``z < 0 and y > 0``) to improve the fit. """ if not isinstance(units, string_types) or units not in ('m', 'mm'): raise ValueError('units must be a "m" or "mm"') radius, origin_head, origin_device = _fit_sphere_to_headshape( info, dig_kinds) if units == 'mm': radius *= 1e3 origin_head *= 1e3 origin_device *= 1e3 return radius, origin_head, origin_device @verbose def get_fitting_dig(info, dig_kinds='auto', verbose=None): """Get digitization points suitable for sphere fitting. Parameters ---------- info : instance of Info The measurement info. dig_kinds : list of str | str Kind of digitization points to use in the fitting. These can be any combination of ('cardinal', 'hpi', 'eeg', 'extra'). Can also be 'auto' (default), which will use only the 'extra' points if enough (more than 10) are available, and if not, uses 'extra' and 'eeg' points. verbose : bool, str or None If not None, override default verbose level Returns ------- dig : array, shape (n_pts, 3) The digitization points (in head coordinates) to use for fitting. Notes ----- This will exclude digitization locations that have ``z < 0 and y > 0``, i.e. points on the nose and below the nose on the face. .. versionadded:: 0.14 """ if not isinstance(info, Info): raise TypeError('info must be an instance of Info not %s' % type(info)) if info['dig'] is None: raise RuntimeError('Cannot fit headshape without digitization ' ', info["dig"] is None') if isinstance(dig_kinds, string_types): if dig_kinds == 'auto': # try "extra" first try: return get_fitting_dig(info, 'extra') except ValueError: pass return get_fitting_dig(info, ('extra', 'eeg')) else: dig_kinds = (dig_kinds,) # convert string args to ints (first make dig_kinds mutable in case tuple) dig_kinds = list(dig_kinds) for di, d in enumerate(dig_kinds): dig_kinds[di] = _dig_kind_dict.get(d, d) if dig_kinds[di] not in _dig_kind_ints: raise ValueError('dig_kinds[#%d] (%s) must be one of %s' % (di, d, sorted(list(_dig_kind_dict.keys())))) # get head digization points of the specified kind(s) hsp = [p['r'] for p in info['dig'] if p['kind'] in dig_kinds] if any(p['coord_frame'] != FIFF.FIFFV_COORD_HEAD for p in info['dig']): raise RuntimeError('Digitization points not in head coordinates, ' 'contact mne-python developers') # exclude some frontal points (nose etc.) hsp = np.array([p for p in hsp if not (p[2] < -1e-6 and p[1] > 1e-6)]) if len(hsp) <= 10: kinds_str = ', '.join(['"%s"' % _dig_kind_rev[d] for d in sorted(dig_kinds)]) msg = ('Only %s head digitization points of the specified kind%s (%s,)' % (len(hsp), _pl(dig_kinds), kinds_str)) if len(hsp) < 4: raise ValueError(msg + ', at least 4 required') else: warn(msg + ', fitting may be inaccurate') return hsp @verbose def _fit_sphere_to_headshape(info, dig_kinds, verbose=None): """Fit a sphere to the given head shape.""" hsp = get_fitting_dig(info, dig_kinds) radius, origin_head = _fit_sphere(np.array(hsp), disp=False) # compute origin in device coordinates head_to_dev = _ensure_trans(info['dev_head_t'], 'head', 'meg') origin_device = apply_trans(head_to_dev, origin_head) logger.info('Fitted sphere radius:'.ljust(30) + '%0.1f mm' % (radius * 1e3,)) # 99th percentile on Wikipedia for Giabella to back of head is 21.7cm, # i.e. 108mm "radius", so let's go with 110mm # en.wikipedia.org/wiki/Human_head#/media/File:HeadAnthropometry.JPG if radius > 0.110: warn('Estimated head size (%0.1f mm) exceeded 99th ' 'percentile for adult head size' % (1e3 * radius,)) # > 2 cm away from head center in X or Y is strange if np.linalg.norm(origin_head[:2]) > 0.02: warn('(X, Y) fit (%0.1f, %0.1f) more than 20 mm from ' 'head frame origin' % tuple(1e3 * origin_head[:2])) logger.info('Origin head coordinates:'.ljust(30) + '%0.1f %0.1f %0.1f mm' % tuple(1e3 * origin_head)) logger.info('Origin device coordinates:'.ljust(30) + '%0.1f %0.1f %0.1f mm' % tuple(1e3 * origin_device)) return radius, origin_head, origin_device def _fit_sphere(points, disp='auto'): """Fit a sphere to an arbitrary set of points.""" from scipy.optimize import fmin_cobyla if isinstance(disp, string_types) and disp == 'auto': disp = True if logger.level <= 20 else False # initial guess for center and radius radii = (np.max(points, axis=1) - np.min(points, axis=1)) / 2. radius_init = radii.mean() center_init = np.median(points, axis=0) # optimization x0 = np.concatenate([center_init, [radius_init]]) def cost_fun(center_rad): d = np.linalg.norm(points - center_rad[:3], axis=1) - center_rad[3] d *= d return d.sum() def constraint(center_rad): return center_rad[3] # radius must be >= 0 x_opt = fmin_cobyla(cost_fun, x0, constraint, rhobeg=radius_init, rhoend=radius_init * 1e-6, disp=disp) origin = x_opt[:3] radius = x_opt[3] return radius, origin def _check_origin(origin, info, coord_frame='head', disp=False): """Check or auto-determine the origin.""" if isinstance(origin, string_types): if origin != 'auto': raise ValueError('origin must be a numerical array, or "auto", ' 'not %s' % (origin,)) if coord_frame == 'head': R, origin = fit_sphere_to_headshape(info, verbose=False, units='m')[:2] logger.info(' Automatic origin fit: head of radius %0.1f mm' % (R * 1000.,)) del R else: origin = (0., 0., 0.) origin = np.array(origin, float) if origin.shape != (3,): raise ValueError('origin must be a 3-element array') if disp: origin_str = ', '.join(['%0.1f' % (o * 1000) for o in origin]) msg = (' Using origin %s mm in the %s frame' % (origin_str, coord_frame)) if coord_frame == 'meg' and info['dev_head_t'] is not None: o_dev = apply_trans(info['dev_head_t'], origin) origin_str = ', '.join('%0.1f' % (o * 1000,) for o in o_dev) msg += ' (%s mm in the head frame)' % (origin_str,) logger.info(msg) return origin # ############################################################################ # Create BEM surfaces @verbose def make_watershed_bem(subject, subjects_dir=None, overwrite=False, volume='T1', atlas=False, gcaatlas=False, preflood=None, show=False, verbose=None): """Create BEM surfaces using the FreeSurfer watershed algorithm. Parameters ---------- subject : str Subject name (required) subjects_dir : str Directory containing subjects data. If None use the Freesurfer SUBJECTS_DIR environment variable. overwrite : bool Write over existing files volume : str Defaults to T1 atlas : bool Specify the --atlas option for mri_watershed gcaatlas : bool Use the subcortical atlas preflood : int Change the preflood height show : bool Show surfaces to visually inspect all three BEM surfaces (recommended). .. versionadded:: 0.12 verbose : bool, str or None If not None, override default verbose level Notes ----- .. versionadded:: 0.10 """ from .viz.misc import plot_bem env, mri_dir = _prepare_env(subject, subjects_dir, requires_freesurfer=True)[:2] subjects_dir = env['SUBJECTS_DIR'] subject_dir = op.join(subjects_dir, subject) mri_dir = op.join(subject_dir, 'mri') T1_dir = op.join(mri_dir, volume) T1_mgz = op.join(mri_dir, volume + '.mgz') bem_dir = op.join(subject_dir, 'bem') ws_dir = op.join(subject_dir, 'bem', 'watershed') if not op.isdir(bem_dir): os.makedirs(bem_dir) if not op.isdir(T1_dir) and not op.isfile(T1_mgz): raise RuntimeError('Could not find the MRI data') if op.isdir(ws_dir): if not overwrite: raise RuntimeError('%s already exists. Use the --overwrite option' ' to recreate it.' % ws_dir) else: shutil.rmtree(ws_dir) # put together the command cmd = ['mri_watershed'] if preflood: cmd += ["-h", "%s" % int(preflood)] if gcaatlas: cmd += ['-atlas', '-T1', '-brain_atlas', env['FREESURFER_HOME'] + '/average/RB_all_withskull_2007-08-08.gca', subject_dir + '/mri/transforms/talairach_with_skull.lta'] elif atlas: cmd += ['-atlas'] if op.exists(T1_mgz): cmd += ['-useSRAS', '-surf', op.join(ws_dir, subject), T1_mgz, op.join(ws_dir, 'ws')] else: cmd += ['-useSRAS', '-surf', op.join(ws_dir, subject), T1_dir, op.join(ws_dir, 'ws')] # report and run logger.info('\nRunning mri_watershed for BEM segmentation with the ' 'following parameters:\n\n' 'SUBJECTS_DIR = %s\n' 'SUBJECT = %s\n' 'Results dir = %s\n' % (subjects_dir, subject, ws_dir)) os.makedirs(op.join(ws_dir, 'ws')) run_subprocess(cmd, env=env) if op.isfile(T1_mgz): new_info = _extract_volume_info(T1_mgz) if new_info is None: warn('nibabel is required to replace the volume info. Volume info' 'not updated in the written surface.') new_info = dict() surfs = ['brain', 'inner_skull', 'outer_skull', 'outer_skin'] for s in surfs: surf_ws_out = op.join(ws_dir, '%s_%s_surface' % (subject, s)) rr, tris, volume_info = read_surface(surf_ws_out, read_metadata=True) volume_info.update(new_info) # replace volume info, 'head' stays write_surface(s, rr, tris, volume_info=volume_info) # Create symbolic links surf_out = op.join(bem_dir, '%s.surf' % s) if not overwrite and op.exists(surf_out): skip_symlink = True else: if op.exists(surf_out): os.remove(surf_out) _symlink(surf_ws_out, surf_out) skip_symlink = False if skip_symlink: logger.info("Unable to create all symbolic links to .surf files " "in bem folder. Use --overwrite option to recreate " "them.") dest = op.join(bem_dir, 'watershed') else: logger.info("Symbolic links to .surf files created in bem folder") dest = bem_dir logger.info("\nThank you for waiting.\nThe BEM triangulations for this " "subject are now available at:\n%s." % dest) # Write a head file for coregistration fname_head = op.join(bem_dir, subject + '-head.fif') if op.isfile(fname_head): os.remove(fname_head) surf = _surfaces_to_bem([op.join(ws_dir, subject + '_outer_skin_surface')], [FIFF.FIFFV_BEM_SURF_ID_HEAD], sigmas=[1]) write_bem_surfaces(fname_head, surf) # Show computed BEM surfaces if show: plot_bem(subject=subject, subjects_dir=subjects_dir, orientation='coronal', slices=None, show=True) logger.info('Created %s\n\nComplete.' % (fname_head,)) def _extract_volume_info(mgz, raise_error=True): """Extract volume info from a mgz file.""" try: import nibabel as nib except ImportError: return # warning raised elsewhere header = nib.load(mgz).header vol_info = dict() version = header['version'] if version == 1: version = '%s # volume info valid' % version else: raise ValueError('Volume info invalid.') vol_info['valid'] = version vol_info['filename'] = mgz vol_info['volume'] = header['dims'][:3] vol_info['voxelsize'] = header['delta'] vol_info['xras'], vol_info['yras'], vol_info['zras'] = header['Mdc'].T vol_info['cras'] = header['Pxyz_c'] return vol_info # ############################################################################ # Read @verbose def read_bem_surfaces(fname, patch_stats=False, s_id=None, verbose=None): """Read the BEM surfaces from a FIF file. Parameters ---------- fname : string The name of the file containing the surfaces. patch_stats : bool, optional (default False) Calculate and add cortical patch statistics to the surfaces. s_id : int | None If int, only read and return the surface with the given s_id. An error will be raised if it doesn't exist. If None, all surfaces are read and returned. verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- surf: list | dict A list of dictionaries that each contain a surface. If s_id is not None, only the requested surface will be returned. See Also -------- write_bem_surfaces, write_bem_solution, make_bem_model """ # Default coordinate frame coord_frame = FIFF.FIFFV_COORD_MRI # Open the file, create directory f, tree, _ = fiff_open(fname) with f as fid: # Find BEM bem = dir_tree_find(tree, FIFF.FIFFB_BEM) if bem is None or len(bem) == 0: raise ValueError('BEM data not found') bem = bem[0] # Locate all surfaces bemsurf = dir_tree_find(bem, FIFF.FIFFB_BEM_SURF) if bemsurf is None: raise ValueError('BEM surface data not found') logger.info(' %d BEM surfaces found' % len(bemsurf)) # Coordinate frame possibly at the top level tag = find_tag(fid, bem, FIFF.FIFF_BEM_COORD_FRAME) if tag is not None: coord_frame = tag.data # Read all surfaces if s_id is not None: surf = [_read_bem_surface(fid, bsurf, coord_frame, s_id) for bsurf in bemsurf] surf = [s for s in surf if s is not None] if not len(surf) == 1: raise ValueError('surface with id %d not found' % s_id) else: surf = list() for bsurf in bemsurf: logger.info(' Reading a surface...') this = _read_bem_surface(fid, bsurf, coord_frame) surf.append(this) logger.info('[done]') logger.info(' %d BEM surfaces read' % len(surf)) for this in surf: if patch_stats or this['nn'] is None: complete_surface_info(this, copy=False) return surf[0] if s_id is not None else surf def _read_bem_surface(fid, this, def_coord_frame, s_id=None): """Read one bem surface.""" # fid should be open as a context manager here res = dict() # Read all the interesting stuff tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_ID) if tag is None: res['id'] = FIFF.FIFFV_BEM_SURF_ID_UNKNOWN else: res['id'] = int(tag.data) if s_id is not None and res['id'] != s_id: return None tag = find_tag(fid, this, FIFF.FIFF_BEM_SIGMA) res['sigma'] = 1.0 if tag is None else float(tag.data) tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_NNODE) if tag is None: raise ValueError('Number of vertices not found') res['np'] = int(tag.data) tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_NTRI) if tag is None: raise ValueError('Number of triangles not found') res['ntri'] = int(tag.data) tag = find_tag(fid, this, FIFF.FIFF_MNE_COORD_FRAME) if tag is None: tag = find_tag(fid, this, FIFF.FIFF_BEM_COORD_FRAME) if tag is None: res['coord_frame'] = def_coord_frame else: res['coord_frame'] = tag.data else: res['coord_frame'] = tag.data # Vertices, normals, and triangles tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_NODES) if tag is None: raise ValueError('Vertex data not found') res['rr'] = tag.data.astype(np.float) # XXX : double because of mayavi bug if res['rr'].shape[0] != res['np']: raise ValueError('Vertex information is incorrect') tag = find_tag(fid, this, FIFF.FIFF_MNE_SOURCE_SPACE_NORMALS) if tag is None: tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_NORMALS) if tag is None: res['nn'] = None else: res['nn'] = tag.data.copy() if res['nn'].shape[0] != res['np']: raise ValueError('Vertex normal information is incorrect') tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_TRIANGLES) if tag is None: raise ValueError('Triangulation not found') res['tris'] = tag.data - 1 # index start at 0 in Python if res['tris'].shape[0] != res['ntri']: raise ValueError('Triangulation information is incorrect') return res @verbose def read_bem_solution(fname, verbose=None): """Read the BEM solution from a file. Parameters ---------- fname : string The file containing the BEM solution. verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Returns ------- bem : instance of ConductorModel The BEM solution. See Also -------- write_bem_solution, read_bem_surfaces, write_bem_surfaces, make_bem_solution """ # mirrors fwd_bem_load_surfaces from fwd_bem_model.c logger.info('Loading surfaces...') bem_surfs = read_bem_surfaces(fname, patch_stats=True, verbose=False) if len(bem_surfs) == 3: logger.info('Three-layer model surfaces loaded.') needed = np.array([FIFF.FIFFV_BEM_SURF_ID_HEAD, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_BRAIN]) if not all(x['id'] in needed for x in bem_surfs): raise RuntimeError('Could not find necessary BEM surfaces') # reorder surfaces as necessary (shouldn't need to?) reorder = [None] * 3 for x in bem_surfs: reorder[np.where(x['id'] == needed)[0][0]] = x bem_surfs = reorder elif len(bem_surfs) == 1: if not bem_surfs[0]['id'] == FIFF.FIFFV_BEM_SURF_ID_BRAIN: raise RuntimeError('BEM Surfaces not found') logger.info('Homogeneous model surface loaded.') # convert from surfaces to solution bem = ConductorModel(is_sphere=False, surfs=bem_surfs) logger.info('\nLoading the solution matrix...\n') f, tree, _ = fiff_open(fname) with f as fid: # Find the BEM data nodes = dir_tree_find(tree, FIFF.FIFFB_BEM) if len(nodes) == 0: raise RuntimeError('No BEM data in %s' % fname) bem_node = nodes[0] # Approximation method tag = find_tag(f, bem_node, FIFF.FIFF_BEM_APPROX) if tag is None: raise RuntimeError('No BEM solution found in %s' % fname) method = tag.data[0] if method not in (FIFF.FIFFV_BEM_APPROX_CONST, FIFF.FIFFV_BEM_APPROX_LINEAR): raise RuntimeError('Cannot handle BEM approximation method : %d' % method) tag = find_tag(fid, bem_node, FIFF.FIFF_BEM_POT_SOLUTION) dims = tag.data.shape if len(dims) != 2: raise RuntimeError('Expected a two-dimensional solution matrix ' 'instead of a %d dimensional one' % dims[0]) dim = 0 for surf in bem['surfs']: if method == FIFF.FIFFV_BEM_APPROX_LINEAR: dim += surf['np'] else: # method == FIFF.FIFFV_BEM_APPROX_CONST dim += surf['ntri'] if dims[0] != dim or dims[1] != dim: raise RuntimeError('Expected a %d x %d solution matrix instead of ' 'a %d x %d one' % (dim, dim, dims[1], dims[0])) sol = tag.data nsol = dims[0] bem['solution'] = sol bem['nsol'] = nsol bem['bem_method'] = method # Gamma factors and multipliers _add_gamma_multipliers(bem) kind = { FIFF.FIFFV_BEM_APPROX_CONST: 'constant collocation', FIFF.FIFFV_BEM_APPROX_LINEAR: 'linear_collocation', }[bem['bem_method']] logger.info('Loaded %s BEM solution from %s', kind, fname) return bem def _add_gamma_multipliers(bem): """Add gamma and multipliers in-place.""" bem['sigma'] = np.array([surf['sigma'] for surf in bem['surfs']]) # Dirty trick for the zero conductivity outside sigma = np.r_[0.0, bem['sigma']] bem['source_mult'] = 2.0 / (sigma[1:] + sigma[:-1]) bem['field_mult'] = sigma[1:] - sigma[:-1] # make sure subsequent "zip"s work correctly assert len(bem['surfs']) == len(bem['field_mult']) bem['gamma'] = ((sigma[1:] - sigma[:-1])[np.newaxis, :] / (sigma[1:] + sigma[:-1])[:, np.newaxis]) _surf_dict = {'inner_skull': FIFF.FIFFV_BEM_SURF_ID_BRAIN, 'outer_skull': FIFF.FIFFV_BEM_SURF_ID_SKULL, 'head': FIFF.FIFFV_BEM_SURF_ID_HEAD} def _bem_find_surface(bem, id_): """Find surface from already-loaded BEM.""" if isinstance(id_, string_types): name = id_ id_ = _surf_dict[id_] else: name = _bem_explain_surface(id_) idx = np.where(np.array([s['id'] for s in bem['surfs']]) == id_)[0] if len(idx) != 1: raise RuntimeError('BEM model does not have the %s triangulation' % name.replace('_', ' ')) return bem['surfs'][idx[0]] def _bem_explain_surface(id_): """Return a string corresponding to the given surface ID.""" _rev_dict = dict((val, key) for key, val in _surf_dict.items()) return _rev_dict[id_] # ############################################################################ # Write def write_bem_surfaces(fname, surfs): """Write BEM surfaces to a fiff file. Parameters ---------- fname : str Filename to write. surfs : dict | list of dict The surfaces, or a single surface. """ if isinstance(surfs, dict): surfs = [surfs] with start_file(fname) as fid: start_block(fid, FIFF.FIFFB_BEM) write_int(fid, FIFF.FIFF_BEM_COORD_FRAME, surfs[0]['coord_frame']) _write_bem_surfaces_block(fid, surfs) end_block(fid, FIFF.FIFFB_BEM) end_file(fid) def _write_bem_surfaces_block(fid, surfs): """Write bem surfaces to open file handle.""" for surf in surfs: start_block(fid, FIFF.FIFFB_BEM_SURF) write_float(fid, FIFF.FIFF_BEM_SIGMA, surf['sigma']) write_int(fid, FIFF.FIFF_BEM_SURF_ID, surf['id']) write_int(fid, FIFF.FIFF_MNE_COORD_FRAME, surf['coord_frame']) write_int(fid, FIFF.FIFF_BEM_SURF_NNODE, surf['np']) write_int(fid, FIFF.FIFF_BEM_SURF_NTRI, surf['ntri']) write_float_matrix(fid, FIFF.FIFF_BEM_SURF_NODES, surf['rr']) # index start at 0 in Python write_int_matrix(fid, FIFF.FIFF_BEM_SURF_TRIANGLES, surf['tris'] + 1) if 'nn' in surf and surf['nn'] is not None and len(surf['nn']) > 0: write_float_matrix(fid, FIFF.FIFF_BEM_SURF_NORMALS, surf['nn']) end_block(fid, FIFF.FIFFB_BEM_SURF) def write_bem_solution(fname, bem): """Write a BEM model with solution. Parameters ---------- fname : str The filename to use. bem : instance of ConductorModel The BEM model with solution to save. See Also -------- read_bem_solution """ _check_bem_size(bem['surfs']) with start_file(fname) as fid: start_block(fid, FIFF.FIFFB_BEM) # Coordinate frame (mainly for backward compatibility) write_int(fid, FIFF.FIFF_BEM_COORD_FRAME, bem['surfs'][0]['coord_frame']) # Surfaces _write_bem_surfaces_block(fid, bem['surfs']) # The potential solution if 'solution' in bem: if bem['bem_method'] != FIFF.FWD_BEM_LINEAR_COLL: raise RuntimeError('Only linear collocation supported') write_int(fid, FIFF.FIFF_BEM_APPROX, FIFF.FIFFV_BEM_APPROX_LINEAR) write_float_matrix(fid, FIFF.FIFF_BEM_POT_SOLUTION, bem['solution']) end_block(fid, FIFF.FIFFB_BEM) end_file(fid) # ############################################################################# # Create 3-Layers BEM model from Flash MRI images def _prepare_env(subject, subjects_dir, requires_freesurfer): """Prepare an env object for subprocess calls.""" env = os.environ.copy() if requires_freesurfer and not os.environ.get('FREESURFER_HOME'): raise RuntimeError('I cannot find freesurfer. The FREESURFER_HOME ' 'environment variable is not set.') if not isinstance(subject, string_types): raise TypeError('The subject argument must be set') subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) if not op.isdir(subjects_dir): raise RuntimeError('Could not find the MRI data directory "%s"' % subjects_dir) subject_dir = op.join(subjects_dir, subject) if not op.isdir(subject_dir): raise RuntimeError('Could not find the subject data directory "%s"' % (subject_dir,)) env['SUBJECT'] = subject env['SUBJECTS_DIR'] = subjects_dir mri_dir = op.join(subject_dir, 'mri') bem_dir = op.join(subject_dir, 'bem') return env, mri_dir, bem_dir @verbose def convert_flash_mris(subject, flash30=True, convert=True, unwarp=False, subjects_dir=None, verbose=None): """Convert DICOM files for use with make_flash_bem. Parameters ---------- subject : str Subject name. flash30 : bool Use 30-degree flip angle data. convert : bool Assume that the Flash MRI images have already been converted to mgz files. unwarp : bool Run grad_unwarp with -unwarp option on each of the converted data sets. It requires FreeSurfer's MATLAB toolbox to be properly installed. subjects_dir : string, or None Path to SUBJECTS_DIR if it is not set in the environment. verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Notes ----- Before running this script do the following: (unless convert=False is specified) 1. Copy all of your FLASH images in a single directory <source> and create a directory <dest> to hold the output of mne_organize_dicom 2. cd to <dest> and run $ mne_organize_dicom <source> to create an appropriate directory structure 3. Create symbolic links to make flash05 and flash30 point to the appropriate series: $ ln -s <FLASH 5 series dir> flash05 $ ln -s <FLASH 30 series dir> flash30 Some partition formats (e.g. FAT32) do not support symbolic links. In this case, copy the file to the appropriate series: $ cp <FLASH 5 series dir> flash05 $ cp <FLASH 30 series dir> flash30 4. cd to the directory where flash05 and flash30 links are 5. Set SUBJECTS_DIR and SUBJECT environment variables appropriately 6. Run this script This function assumes that the Freesurfer segmentation of the subject has been completed. In particular, the T1.mgz and brain.mgz MRI volumes should be, as usual, in the subject's mri directory. """ env, mri_dir = _prepare_env(subject, subjects_dir, requires_freesurfer=True)[:2] curdir = os.getcwd() # Step 1a : Data conversion to mgz format if not op.exists(op.join(mri_dir, 'flash', 'parameter_maps')): os.makedirs(op.join(mri_dir, 'flash', 'parameter_maps')) echos_done = 0 if convert: logger.info("\n---- Converting Flash images ----") echos = ['001', '002', '003', '004', '005', '006', '007', '008'] if flash30: flashes = ['05'] else: flashes = ['05', '30'] # missing = False for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): missing = True if missing: echos = ['002', '003', '004', '005', '006', '007', '008', '009'] for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): raise RuntimeError("Directory %s is missing." % op.join('flash' + flash, echo)) # for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): raise RuntimeError("Directory %s is missing." % op.join('flash' + flash, echo)) sample_file = glob.glob(op.join('flash' + flash, echo, '*'))[0] dest_file = op.join(mri_dir, 'flash', 'mef' + flash + '_' + echo + '.mgz') # do not redo if already present if op.isfile(dest_file): logger.info("The file %s is already there") else: cmd = ['mri_convert', sample_file, dest_file] run_subprocess(cmd, env=env) echos_done += 1 # Step 1b : Run grad_unwarp on converted files os.chdir(op.join(mri_dir, "flash")) files = glob.glob("mef*.mgz") if unwarp: logger.info("\n---- Unwarp mgz data sets ----") for infile in files: outfile = infile.replace(".mgz", "u.mgz") cmd = ['grad_unwarp', '-i', infile, '-o', outfile, '-unwarp', 'true'] run_subprocess(cmd, env=env) # Clear parameter maps if some of the data were reconverted if echos_done > 0 and op.exists("parameter_maps"): shutil.rmtree("parameter_maps") logger.info("\nParameter maps directory cleared") if not op.exists("parameter_maps"): os.makedirs("parameter_maps") # Step 2 : Create the parameter maps if flash30: logger.info("\n---- Creating the parameter maps ----") if unwarp: files = glob.glob("mef05*u.mgz") if len(os.listdir('parameter_maps')) == 0: cmd = ['mri_ms_fitparms'] + files + ['parameter_maps'] run_subprocess(cmd, env=env) else: logger.info("Parameter maps were already computed") # Step 3 : Synthesize the flash 5 images logger.info("\n---- Synthesizing flash 5 images ----") os.chdir('parameter_maps') if not op.exists('flash5.mgz'): cmd = ['mri_synthesize', '20 5 5', 'T1.mgz', 'PD.mgz', 'flash5.mgz'] run_subprocess(cmd, env=env) os.remove('flash5_reg.mgz') else: logger.info("Synthesized flash 5 volume is already there") else: logger.info("\n---- Averaging flash5 echoes ----") os.chdir('parameter_maps') if unwarp: files = glob.glob("mef05*u.mgz") else: files = glob.glob("mef05*.mgz") cmd = ['mri_average', '-noconform', files, 'flash5.mgz'] run_subprocess(cmd, env=env) if op.exists('flash5_reg.mgz'): os.remove('flash5_reg.mgz') # Go back to initial directory os.chdir(curdir) @verbose def make_flash_bem(subject, overwrite=False, show=True, subjects_dir=None, flash_path=None, verbose=None): """Create 3-Layer BEM model from prepared flash MRI images. Parameters ---------- subject : str Subject name. overwrite : bool Write over existing .surf files in bem folder. show : bool Show surfaces to visually inspect all three BEM surfaces (recommended). subjects_dir : string, or None Path to SUBJECTS_DIR if it is not set in the environment. flash_path : str | None Path to the flash images. If None (default), mri/flash/parameter_maps within the subject reconstruction is used. .. versionadded:: 0.13.0 verbose : bool, str, int, or None If not None, override default verbose level (see :func:`mne.verbose` and :ref:`Logging documentation <tut_logging>` for more). Notes ----- This program assumes that FreeSurfer is installed and sourced properly. This function extracts the BEM surfaces (outer skull, inner skull, and outer skin) from multiecho FLASH MRI data with spin angles of 5 and 30 degrees, in mgz format. See Also -------- convert_flash_mris """ from .viz.misc import plot_bem is_test = os.environ.get('MNE_SKIP_FS_FLASH_CALL', False) env, mri_dir, bem_dir = _prepare_env(subject, subjects_dir, requires_freesurfer=True) if flash_path is None: flash_path = op.join(mri_dir, 'flash', 'parameter_maps') else: flash_path = op.abspath(flash_path) curdir = os.getcwd() subjects_dir = env['SUBJECTS_DIR'] logger.info('\nProcessing the flash MRI data to produce BEM meshes with ' 'the following parameters:\n' 'SUBJECTS_DIR = %s\n' 'SUBJECT = %s\n' 'Result dir = %s\n' % (subjects_dir, subject, op.join(bem_dir, 'flash'))) # Step 4 : Register with MPRAGE logger.info("\n---- Registering flash 5 with MPRAGE ----") flash5 = op.join(flash_path, 'flash5.mgz') flash5_reg = op.join(flash_path, 'flash5_reg.mgz') if not op.exists(flash5_reg): if op.exists(op.join(mri_dir, 'T1.mgz')): ref_volume = op.join(mri_dir, 'T1.mgz') else: ref_volume = op.join(mri_dir, 'T1') cmd = ['fsl_rigid_register', '-r', ref_volume, '-i', flash5, '-o', flash5_reg] run_subprocess(cmd, env=env) else: logger.info("Registered flash 5 image is already there") # Step 5a : Convert flash5 into COR logger.info("\n---- Converting flash5 volume into COR format ----") shutil.rmtree(op.join(mri_dir, 'flash5'), ignore_errors=True) os.makedirs(op.join(mri_dir, 'flash5')) if not is_test: # CIs don't have freesurfer, skipped when testing. cmd = ['mri_convert', flash5_reg, op.join(mri_dir, 'flash5')] run_subprocess(cmd, env=env) # Step 5b and c : Convert the mgz volumes into COR os.chdir(mri_dir) convert_T1 = False if not op.isdir('T1') or len(glob.glob(op.join('T1', 'COR*'))) == 0: convert_T1 = True convert_brain = False if not op.isdir('brain') or len(glob.glob(op.join('brain', 'COR*'))) == 0: convert_brain = True logger.info("\n---- Converting T1 volume into COR format ----") if convert_T1: if not op.isfile('T1.mgz'): raise RuntimeError("Both T1 mgz and T1 COR volumes missing.") os.makedirs('T1') cmd = ['mri_convert', 'T1.mgz', 'T1'] run_subprocess(cmd, env=env) else: logger.info("T1 volume is already in COR format") logger.info("\n---- Converting brain volume into COR format ----") if convert_brain: if not op.isfile('brain.mgz'): raise RuntimeError("Both brain mgz and brain COR volumes missing.") os.makedirs('brain') cmd = ['mri_convert', 'brain.mgz', 'brain'] run_subprocess(cmd, env=env) else: logger.info("Brain volume is already in COR format") # Finally ready to go if not is_test: # CIs don't have freesurfer, skipped when testing. logger.info("\n---- Creating the BEM surfaces ----") cmd = ['mri_make_bem_surfaces', subject] run_subprocess(cmd, env=env) logger.info("\n---- Converting the tri files into surf files ----") os.chdir(bem_dir) if not op.exists('flash'): os.makedirs('flash') os.chdir('flash') surfs = ['inner_skull', 'outer_skull', 'outer_skin'] for surf in surfs: shutil.move(op.join(bem_dir, surf + '.tri'), surf + '.tri') nodes, tris = read_tri(surf + '.tri', swap=True) vol_info = _extract_volume_info(flash5_reg) if vol_info is None: warn('nibabel is required to update the volume info. Volume info ' 'omitted from the written surface.') else: vol_info['head'] = np.array([20]) write_surface(surf + '.surf', nodes, tris, volume_info=vol_info) # Cleanup section logger.info("\n---- Cleaning up ----") os.chdir(bem_dir) os.remove('inner_skull_tmp.tri') os.chdir(mri_dir) if convert_T1: shutil.rmtree('T1') logger.info("Deleted the T1 COR volume") if convert_brain: shutil.rmtree('brain') logger.info("Deleted the brain COR volume") shutil.rmtree('flash5') logger.info("Deleted the flash5 COR volume") # Create symbolic links to the .surf files in the bem folder logger.info("\n---- Creating symbolic links ----") os.chdir(bem_dir) for surf in surfs: surf = surf + '.surf' if not overwrite and op.exists(surf): skip_symlink = True else: if op.exists(surf): os.remove(surf) _symlink(op.join('flash', surf), op.join(surf)) skip_symlink = False if skip_symlink: logger.info("Unable to create all symbolic links to .surf files " "in bem folder. Use --overwrite option to recreate them.") dest = op.join(bem_dir, 'flash') else: logger.info("Symbolic links to .surf files created in bem folder") dest = bem_dir logger.info("\nThank you for waiting.\nThe BEM triangulations for this " "subject are now available at:\n%s.\nWe hope the BEM meshes " "created will facilitate your MEG and EEG data analyses." % dest) # Show computed BEM surfaces if show: plot_bem(subject=subject, subjects_dir=subjects_dir, orientation='coronal', slices=None, show=True) # Go back to initial directory os.chdir(curdir) def _check_bem_size(surfs): """Check bem surface sizes.""" if len(surfs) > 1 and surfs[0]['np'] > 10000: warn('The bem surfaces have %s data points. 5120 (ico grade=4) ' 'should be enough. Dense 3-layer bems may not save properly.' % surfs[0]['np']) def _symlink(src, dest): """Create a symlink.""" try: os.symlink(src, dest) except OSError: warn('Could not create symbolic link %s. Check that your partition ' 'handles symbolic links. The file will be copied instead.' % dest) shutil.copy(src, dest)
37.822513
79
0.584087
from functools import partial import glob import os import os.path as op import shutil from copy import deepcopy import numpy as np from scipy import linalg from .transforms import _ensure_trans, apply_trans from .io import Info from .io.constants import FIFF from .io.write import (start_file, start_block, write_float, write_int, write_float_matrix, write_int_matrix, end_block, end_file) from .io.tag import find_tag from .io.tree import dir_tree_find from .io.open import fiff_open from .surface import (read_surface, write_surface, complete_surface_info, _compute_nearest, _get_ico_surface, read_tri, _fast_cross_nd_sum, _get_solids) from .utils import verbose, logger, run_subprocess, get_subjects_dir, warn, _pl from .fixes import einsum from .externals.six import string_types rf['tris'].ravel() misses = pi2 - mat.sum(axis=1) for j, miss in enumerate(misses): # How much is missing? n_memb = len(surf['neighbor_tri'][j]) # The node itself receives one half mat[j, j] = miss / 2.0 # The rest is divided evenly among the member nodes... miss /= (4.0 * n_memb) members = np.where(j == tris_flat)[0] mods = members % 3 offsets = np.array([[1, 2], [-1, 1], [-1, -2]]) tri_1 = members + offsets[mods, 0] tri_2 = members + offsets[mods, 1] for t1, t2 in zip(tri_1, tri_2): mat[j, tris_flat[t1]] += miss mat[j, tris_flat[t2]] += miss return def _fwd_bem_lin_pot_coeff(surfs): # taken from fwd_bem_linear_collocation.c nps = [surf['np'] for surf in surfs] np_tot = sum(nps) coeff = np.zeros((np_tot, np_tot)) offsets = np.cumsum(np.concatenate(([0], nps))) for si_1, surf1 in enumerate(surfs): rr_ord = np.arange(nps[si_1]) for si_2, surf2 in enumerate(surfs): logger.info(" %s (%d) -> %s (%d) ..." % (_bem_explain_surface(surf1['id']), nps[si_1], _bem_explain_surface(surf2['id']), nps[si_2])) tri_rr = surf2['rr'][surf2['tris']] tri_nn = surf2['tri_nn'] tri_area = surf2['tri_area'] submat = coeff[offsets[si_1]:offsets[si_1 + 1], offsets[si_2]:offsets[si_2 + 1]] # view for k in range(surf2['ntri']): tri = surf2['tris'][k] if si_1 == si_2: skip_idx = ((rr_ord == tri[0]) | (rr_ord == tri[1]) | (rr_ord == tri[2])) else: skip_idx = list() # No contribution from a triangle that # this vertex belongs to # if sidx1 == sidx2 and (tri == j).any(): # continue # Otherwise do the hard job coeffs = _lin_pot_coeff(surf1['rr'], tri_rr[k], tri_nn[k], tri_area[k]) coeffs[skip_idx] = 0. submat[:, tri] -= coeffs if si_1 == si_2: _correct_auto_elements(surf1, submat) return coeff def _fwd_bem_multi_solution(solids, gamma, nps): pi2 = 1.0 / (2 * np.pi) n_tot = np.sum(nps) assert solids.shape == (n_tot, n_tot) nsurf = len(nps) defl = 1.0 / n_tot # Modify the matrix offsets = np.cumsum(np.concatenate(([0], nps))) for si_1 in range(nsurf): for si_2 in range(nsurf): mult = pi2 if gamma is None else pi2 * gamma[si_1, si_2] slice_j = slice(offsets[si_1], offsets[si_1 + 1]) slice_k = slice(offsets[si_2], offsets[si_2 + 1]) solids[slice_j, slice_k] = defl - solids[slice_j, slice_k] * mult solids += np.eye(n_tot) return linalg.inv(solids, overwrite_a=True) def _fwd_bem_homog_solution(solids, nps): return _fwd_bem_multi_solution(solids, None, nps) def _fwd_bem_ip_modify_solution(solution, ip_solution, ip_mult, n_tri): n_last = n_tri[-1] mult = (1.0 + ip_mult) / ip_mult logger.info(' Combining...') offsets = np.cumsum(np.concatenate(([0], n_tri))) for si in range(len(n_tri)): # Pick the correct submatrix (right column) and multiply sub = solution[offsets[si]:offsets[si + 1], np.sum(n_tri[:-1]):] # Multiply sub -= 2 * np.dot(sub, ip_solution) # The lower right corner is a special case sub[-n_last:, -n_last:] += mult * ip_solution # Final scaling logger.info(' Scaling...') solution *= ip_mult return def _fwd_bem_linear_collocation_solution(m): # first, add surface geometries for surf in m['surfs']: complete_surface_info(surf, copy=False, verbose=False) logger.info('Computing the linear collocation solution...') logger.info(' Matrix coefficients...') coeff = _fwd_bem_lin_pot_coeff(m['surfs']) m['nsol'] = len(coeff) logger.info(" Inverting the coefficient matrix...") nps = [surf['np'] for surf in m['surfs']] m['solution'] = _fwd_bem_multi_solution(coeff, m['gamma'], nps) if len(m['surfs']) == 3: ip_mult = m['sigma'][1] / m['sigma'][2] if ip_mult <= FIFF.FWD_BEM_IP_APPROACH_LIMIT: logger.info('IP approach required...') logger.info(' Matrix coefficients (homog)...') coeff = _fwd_bem_lin_pot_coeff([m['surfs'][-1]]) logger.info(' Inverting the coefficient matrix (homog)...') ip_solution = _fwd_bem_homog_solution(coeff, [m['surfs'][-1]['np']]) logger.info(' Modify the original solution to incorporate ' 'IP approach...') _fwd_bem_ip_modify_solution(m['solution'], ip_solution, ip_mult, nps) m['bem_method'] = FIFF.FWD_BEM_LINEAR_COLL logger.info("Solution ready.") @verbose def make_bem_solution(surfs, verbose=None): logger.info('Approximation method : Linear collocation\n') if isinstance(surfs, string_types): # Load the surfaces logger.info('Loading surfaces...') surfs = read_bem_surfaces(surfs) bem = ConductorModel(is_sphere=False, surfs=surfs) _add_gamma_multipliers(bem) if len(bem['surfs']) == 3: logger.info('Three-layer model surfaces loaded.') elif len(bem['surfs']) == 1: logger.info('Homogeneous model surface loaded.') else: raise RuntimeError('Only 1- or 3-layer BEM computations supported') _check_bem_size(bem['surfs']) _fwd_bem_linear_collocation_solution(bem) logger.info('BEM geometry computations complete.') return bem # ############################################################################ # Make BEM model def _ico_downsample(surf, dest_grade): n_tri = len(surf['tris']) found = -1 bad_msg = ("A surface with %d triangles cannot be isomorphic with a " "subdivided icosahedron." % n_tri) if n_tri % 20 != 0: raise RuntimeError(bad_msg) n_tri = n_tri // 20 found = int(round(np.log(n_tri) / np.log(4))) if n_tri != 4 ** found: raise RuntimeError(bad_msg) del n_tri if dest_grade > found: raise RuntimeError('For this surface, decimation grade should be %d ' 'or less, not %s.' % (found, dest_grade)) source = _get_ico_surface(found) dest = _get_ico_surface(dest_grade, patch_stats=True) del dest['tri_cent'] del dest['tri_nn'] del dest['neighbor_tri'] del dest['tri_area'] if not np.array_equal(source['tris'], surf['tris']): raise RuntimeError('The source surface has a matching number of ' 'triangles but ordering is wrong') logger.info('Going from %dth to %dth subdivision of an icosahedron ' '(n_tri: %d -> %d)' % (found, dest_grade, len(surf['tris']), len(dest['tris']))) # Find the mapping dest['rr'] = surf['rr'][_get_ico_map(source, dest)] return dest def _get_ico_map(fro, to): nearest, dists = _compute_nearest(fro['rr'], to['rr'], return_dists=True) n_bads = (dists > 5e-3).sum() if n_bads > 0: raise RuntimeError('No matching vertex for %d destination vertices' % (n_bads)) return nearest def _order_surfaces(surfs): if len(surfs) != 3: return surfs # we have three surfaces surf_order = [FIFF.FIFFV_BEM_SURF_ID_HEAD, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_BRAIN] ids = np.array([surf['id'] for surf in surfs]) if set(ids) != set(surf_order): raise RuntimeError('bad surface ids: %s' % ids) order = [np.where(ids == id_)[0][0] for id_ in surf_order] surfs = [surfs[idx] for idx in order] return surfs def _assert_complete_surface(surf, incomplete='raise'): # from surface_checks.c tot_angle = 0. # Center of mass.... cm = surf['rr'].mean(axis=0) logger.info('%s CM is %6.2f %6.2f %6.2f mm' % (_surf_name[surf['id']], 1000 * cm[0], 1000 * cm[1], 1000 * cm[2])) tot_angle = _get_solids(surf['rr'][surf['tris']], cm[np.newaxis, :])[0] prop = tot_angle / (2 * np.pi) if np.abs(prop - 1.0) > 1e-5: msg = ('Surface %s is not complete (sum of solid angles ' 'yielded %g, should be 1.)' % (_surf_name[surf['id']], prop)) if incomplete == 'raise': raise RuntimeError(msg) else: warn(msg) _surf_name = { FIFF.FIFFV_BEM_SURF_ID_HEAD: 'outer skin ', FIFF.FIFFV_BEM_SURF_ID_SKULL: 'outer skull', FIFF.FIFFV_BEM_SURF_ID_BRAIN: 'inner skull', FIFF.FIFFV_BEM_SURF_ID_UNKNOWN: 'unknown ', } def _assert_inside(fro, to): # this is "is_inside" in surface_checks.c tot_angle = _get_solids(to['rr'][to['tris']], fro['rr']) if (np.abs(tot_angle / (2 * np.pi) - 1.0) > 1e-5).any(): raise RuntimeError('Surface %s is not completely inside surface %s' % (_surf_name[fro['id']], _surf_name[to['id']])) def _check_surfaces(surfs, incomplete='raise'): for surf in surfs: _assert_complete_surface(surf, incomplete=incomplete) # Then check the topology for surf_1, surf_2 in zip(surfs[:-1], surfs[1:]): logger.info('Checking that %s surface is inside %s surface...' % (_surf_name[surf_2['id']], _surf_name[surf_1['id']])) _assert_inside(surf_2, surf_1) def _check_surface_size(surf): sizes = surf['rr'].max(axis=0) - surf['rr'].min(axis=0) if (sizes < 0.05).any(): raise RuntimeError('Dimensions of the surface %s seem too small ' '(%9.5f mm). Maybe the the unit of measure is ' 'meters instead of mm' % (_surf_name[surf['id']], 1000 * sizes.min())) def _check_thicknesses(surfs): for surf_1, surf_2 in zip(surfs[:-1], surfs[1:]): min_dist = _compute_nearest(surf_1['rr'], surf_2['rr'], return_dists=True)[0] min_dist = min_dist.min() logger.info('Checking distance between %s and %s surfaces...' % (_surf_name[surf_1['id']], _surf_name[surf_2['id']])) logger.info('Minimum distance between the %s and %s surfaces is ' 'approximately %6.1f mm' % (_surf_name[surf_1['id']], _surf_name[surf_2['id']], 1000 * min_dist)) def _surfaces_to_bem(surfs, ids, sigmas, ico=None, rescale=True, incomplete='raise'): # equivalent of mne_surf2bem # surfs can be strings (filenames) or surface dicts if len(surfs) not in (1, 3) or not (len(surfs) == len(ids) == len(sigmas)): raise ValueError('surfs, ids, and sigmas must all have the same ' 'number of elements (1 or 3)') surf = list(surfs) for si, surf in enumerate(surfs): if isinstance(surf, string_types): surfs[si] = read_surface(surf, return_dict=True)[-1] # Downsampling if the surface is isomorphic with a subdivided icosahedron if ico is not None: for si, surf in enumerate(surfs): surfs[si] = _ico_downsample(surf, ico) for surf, id_ in zip(surfs, ids): surf['id'] = id_ surf['coord_frame'] = surf.get('coord_frame', FIFF.FIFFV_COORD_MRI) surf.update(np=len(surf['rr']), ntri=len(surf['tris'])) if rescale: surf['rr'] /= 1000. # convert to meters # Shifting surfaces is not implemented here... # Order the surfaces for the benefit of the topology checks for surf, sigma in zip(surfs, sigmas): surf['sigma'] = sigma surfs = _order_surfaces(surfs) # Check topology as best we can _check_surfaces(surfs, incomplete=incomplete) for surf in surfs: _check_surface_size(surf) _check_thicknesses(surfs) logger.info('Surfaces passed the basic topology checks.') return surfs @verbose def make_bem_model(subject, ico=4, conductivity=(0.3, 0.006, 0.3), subjects_dir=None, verbose=None): conductivity = np.array(conductivity, float) if conductivity.ndim != 1 or conductivity.size not in (1, 3): raise ValueError('conductivity must be 1D array-like with 1 or 3 ' 'elements') subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) subject_dir = op.join(subjects_dir, subject) bem_dir = op.join(subject_dir, 'bem') inner_skull = op.join(bem_dir, 'inner_skull.surf') outer_skull = op.join(bem_dir, 'outer_skull.surf') outer_skin = op.join(bem_dir, 'outer_skin.surf') surfaces = [inner_skull, outer_skull, outer_skin] ids = [FIFF.FIFFV_BEM_SURF_ID_BRAIN, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_HEAD] logger.info('Creating the BEM geometry...') if len(conductivity) == 1: surfaces = surfaces[:1] ids = ids[:1] surfaces = _surfaces_to_bem(surfaces, ids, conductivity, ico) _check_bem_size(surfaces) logger.info('Complete.\n') return surfaces # ############################################################################ # Compute EEG sphere model def _fwd_eeg_get_multi_sphere_model_coeffs(m, n_terms): nlayer = len(m['layers']) if nlayer in (0, 1): return 1. # Initialize the arrays c1 = np.zeros(nlayer - 1) c2 = np.zeros(nlayer - 1) cr = np.zeros(nlayer - 1) cr_mult = np.zeros(nlayer - 1) for k in range(nlayer - 1): c1[k] = m['layers'][k]['sigma'] / m['layers'][k + 1]['sigma'] c2[k] = c1[k] - 1.0 cr_mult[k] = m['layers'][k]['rel_rad'] cr[k] = cr_mult[k] cr_mult[k] *= cr_mult[k] coeffs = np.zeros(n_terms - 1) for n in range(1, n_terms): # Increment the radius coefficients for k in range(nlayer - 1): cr[k] *= cr_mult[k] # Multiply the matrices M = np.eye(2) n1 = n + 1.0 for k in range(nlayer - 2, -1, -1): M = np.dot([[n + n1 * c1[k], n1 * c2[k] / cr[k]], [n * c2[k] * cr[k], n1 + n * c1[k]]], M) num = n * (2.0 * n + 1.0) ** (nlayer - 1) coeffs[n - 1] = num / (n * M[1, 1] + n1 * M[1, 0]) return coeffs def _compose_linear_fitting_data(mu, u): # y is the data to be fitted (nterms-1 x 1) # M is the model matrix (nterms-1 x nfit-1) for k in range(u['nterms'] - 1): k1 = k + 1 mu1n = np.power(mu[0], k1) u['y'][k] = u['w'][k] * (u['fn'][k1] - mu1n * u['fn'][0]) for p in range(u['nfit'] - 1): u['M'][k][p] = u['w'][k] * (np.power(mu[p + 1], k1) - mu1n) def _compute_linear_parameters(mu, u): _compose_linear_fitting_data(mu, u) uu, sing, vv = linalg.svd(u['M'], full_matrices=False) # Compute the residuals u['resi'] = u['y'].copy() vec = np.empty(u['nfit'] - 1) for p in range(u['nfit'] - 1): vec[p] = np.dot(uu[:, p], u['y']) for k in range(u['nterms'] - 1): u['resi'][k] -= uu[k, p] * vec[p] vec[p] = vec[p] / sing[p] lambda_ = np.zeros(u['nfit']) for p in range(u['nfit'] - 1): sum_ = 0. for q in range(u['nfit'] - 1): sum_ += vv[q, p] * vec[q] lambda_[p + 1] = sum_ lambda_[0] = u['fn'][0] - np.sum(lambda_[1:]) rv = np.dot(u['resi'], u['resi']) / np.dot(u['y'], u['y']) return rv, lambda_ def _one_step(mu, u): if np.abs(mu).max() > 1.0: return 1.0 # Compose the data for the linear fitting, compute SVD, then residuals _compose_linear_fitting_data(mu, u) u['uu'], u['sing'], u['vv'] = linalg.svd(u['M']) u['resi'][:] = u['y'][:] for p in range(u['nfit'] - 1): dot = np.dot(u['uu'][p], u['y']) for k in range(u['nterms'] - 1): u['resi'][k] = u['resi'][k] - u['uu'][p, k] * dot # Return their sum of squares return np.dot(u['resi'], u['resi']) def _fwd_eeg_fit_berg_scherg(m, nterms, nfit): from scipy.optimize import fmin_cobyla assert nfit >= 2 u = dict(y=np.zeros(nterms - 1), resi=np.zeros(nterms - 1), nfit=nfit, nterms=nterms, M=np.zeros((nterms - 1, nfit - 1))) # (1) Calculate the coefficients of the true expansion u['fn'] = _fwd_eeg_get_multi_sphere_model_coeffs(m, nterms + 1) # (2) Calculate the weighting f = (min([layer['rad'] for layer in m['layers']]) / max([layer['rad'] for layer in m['layers']])) # correct weighting k = np.arange(1, nterms + 1) u['w'] = np.sqrt((2.0 * k + 1) * (3.0 * k + 1.0) / k) * np.power(f, (k - 1.0)) u['w'][-1] = 0 # Do the nonlinear minimization, constraining mu to the interval [-1, +1] mu_0 = np.random.RandomState(0).rand(nfit) * f fun = partial(_one_step, u=u) max_ = 1. - 2e-4 # adjust for fmin_cobyla "catol" that not all scipy have cons = [(lambda x: max_ - np.abs(x[ii])) for ii in range(nfit)] mu = fmin_cobyla(fun, mu_0, cons, rhobeg=0.5, rhoend=5e-3, disp=0) # (6) Do the final step: calculation of the linear parameters rv, lambda_ = _compute_linear_parameters(mu, u) order = np.argsort(mu)[::-1] mu, lambda_ = mu[order], lambda_[order] # sort: largest mu first m['mu'] = mu # This division takes into account the actual conductivities m['lambda'] = lambda_ / m['layers'][-1]['sigma'] m['nfit'] = nfit return rv @verbose def make_sphere_model(r0=(0., 0., 0.04), head_radius=0.09, info=None, relative_radii=(0.90, 0.92, 0.97, 1.0), sigmas=(0.33, 1.0, 0.004, 0.33), verbose=None): for name in ('r0', 'head_radius'): param = locals()[name] if isinstance(param, string_types): if param != 'auto': raise ValueError('%s, if str, must be "auto" not "%s"' % (name, param)) relative_radii = np.array(relative_radii, float).ravel() sigmas = np.array(sigmas, float).ravel() if len(relative_radii) != len(sigmas): raise ValueError('relative_radii length (%s) must match that of ' 'sigmas (%s)' % (len(relative_radii), len(sigmas))) if len(sigmas) <= 1 and head_radius is not None: raise ValueError('at least 2 sigmas must be supplied if ' 'head_radius is not None, got %s' % (len(sigmas),)) if (isinstance(r0, string_types) and r0 == 'auto') or \ (isinstance(head_radius, string_types) and head_radius == 'auto'): if info is None: raise ValueError('Info must not be None for auto mode') head_radius_fit, r0_fit = fit_sphere_to_headshape(info, units='m')[:2] if isinstance(r0, string_types): r0 = r0_fit if isinstance(head_radius, string_types): head_radius = head_radius_fit sphere = ConductorModel(is_sphere=True, r0=np.array(r0), coord_frame=FIFF.FIFFV_COORD_HEAD) sphere['layers'] = list() if head_radius is not None: # Eventually these could be configurable... relative_radii = np.array(relative_radii, float) sigmas = np.array(sigmas, float) order = np.argsort(relative_radii) relative_radii = relative_radii[order] sigmas = sigmas[order] for rel_rad, sig in zip(relative_radii, sigmas): # sort layers by (relative) radius, and scale radii layer = dict(rad=rel_rad, sigma=sig) layer['rel_rad'] = layer['rad'] = rel_rad sphere['layers'].append(layer) # scale the radii R = sphere['layers'][-1]['rad'] rR = sphere['layers'][-1]['rel_rad'] for layer in sphere['layers']: layer['rad'] /= R layer['rel_rad'] /= rR # # Setup the EEG sphere model calculations # # Scale the relative radii for k in range(len(relative_radii)): sphere['layers'][k]['rad'] = (head_radius * sphere['layers'][k]['rel_rad']) rv = _fwd_eeg_fit_berg_scherg(sphere, 200, 3) logger.info('\nEquiv. model fitting -> RV = %g %%' % (100 * rv)) for k in range(3): logger.info('mu%d = %g lambda%d = %g' % (k + 1, sphere['mu'][k], k + 1, sphere['layers'][-1]['sigma'] * sphere['lambda'][k])) logger.info('Set up EEG sphere model with scalp radius %7.1f mm\n' % (1000 * head_radius,)) return sphere # ############################################################################# # Sphere fitting _dig_kind_dict = { 'cardinal': FIFF.FIFFV_POINT_CARDINAL, 'hpi': FIFF.FIFFV_POINT_HPI, 'eeg': FIFF.FIFFV_POINT_EEG, 'extra': FIFF.FIFFV_POINT_EXTRA, } _dig_kind_rev = dict((val, key) for key, val in _dig_kind_dict.items()) _dig_kind_ints = tuple(_dig_kind_dict.values()) @verbose def fit_sphere_to_headshape(info, dig_kinds='auto', units='m', verbose=None): if not isinstance(units, string_types) or units not in ('m', 'mm'): raise ValueError('units must be a "m" or "mm"') radius, origin_head, origin_device = _fit_sphere_to_headshape( info, dig_kinds) if units == 'mm': radius *= 1e3 origin_head *= 1e3 origin_device *= 1e3 return radius, origin_head, origin_device @verbose def get_fitting_dig(info, dig_kinds='auto', verbose=None): if not isinstance(info, Info): raise TypeError('info must be an instance of Info not %s' % type(info)) if info['dig'] is None: raise RuntimeError('Cannot fit headshape without digitization ' ', info["dig"] is None') if isinstance(dig_kinds, string_types): if dig_kinds == 'auto': # try "extra" first try: return get_fitting_dig(info, 'extra') except ValueError: pass return get_fitting_dig(info, ('extra', 'eeg')) else: dig_kinds = (dig_kinds,) # convert string args to ints (first make dig_kinds mutable in case tuple) dig_kinds = list(dig_kinds) for di, d in enumerate(dig_kinds): dig_kinds[di] = _dig_kind_dict.get(d, d) if dig_kinds[di] not in _dig_kind_ints: raise ValueError('dig_kinds[ % (di, d, sorted(list(_dig_kind_dict.keys())))) # get head digization points of the specified kind(s) hsp = [p['r'] for p in info['dig'] if p['kind'] in dig_kinds] if any(p['coord_frame'] != FIFF.FIFFV_COORD_HEAD for p in info['dig']): raise RuntimeError('Digitization points not in head coordinates, ' 'contact mne-python developers') # exclude some frontal points (nose etc.) hsp = np.array([p for p in hsp if not (p[2] < -1e-6 and p[1] > 1e-6)]) if len(hsp) <= 10: kinds_str = ', '.join(['"%s"' % _dig_kind_rev[d] for d in sorted(dig_kinds)]) msg = ('Only %s head digitization points of the specified kind%s (%s,)' % (len(hsp), _pl(dig_kinds), kinds_str)) if len(hsp) < 4: raise ValueError(msg + ', at least 4 required') else: warn(msg + ', fitting may be inaccurate') return hsp @verbose def _fit_sphere_to_headshape(info, dig_kinds, verbose=None): hsp = get_fitting_dig(info, dig_kinds) radius, origin_head = _fit_sphere(np.array(hsp), disp=False) # compute origin in device coordinates head_to_dev = _ensure_trans(info['dev_head_t'], 'head', 'meg') origin_device = apply_trans(head_to_dev, origin_head) logger.info('Fitted sphere radius:'.ljust(30) + '%0.1f mm' % (radius * 1e3,)) # 99th percentile on Wikipedia for Giabella to back of head is 21.7cm, # i.e. 108mm "radius", so let's go with 110mm rn('Estimated head size (%0.1f mm) exceeded 99th ' 'percentile for adult head size' % (1e3 * radius,)) if np.linalg.norm(origin_head[:2]) > 0.02: warn('(X, Y) fit (%0.1f, %0.1f) more than 20 mm from ' 'head frame origin' % tuple(1e3 * origin_head[:2])) logger.info('Origin head coordinates:'.ljust(30) + '%0.1f %0.1f %0.1f mm' % tuple(1e3 * origin_head)) logger.info('Origin device coordinates:'.ljust(30) + '%0.1f %0.1f %0.1f mm' % tuple(1e3 * origin_device)) return radius, origin_head, origin_device def _fit_sphere(points, disp='auto'): from scipy.optimize import fmin_cobyla if isinstance(disp, string_types) and disp == 'auto': disp = True if logger.level <= 20 else False radii = (np.max(points, axis=1) - np.min(points, axis=1)) / 2. radius_init = radii.mean() center_init = np.median(points, axis=0) x0 = np.concatenate([center_init, [radius_init]]) def cost_fun(center_rad): d = np.linalg.norm(points - center_rad[:3], axis=1) - center_rad[3] d *= d return d.sum() def constraint(center_rad): return center_rad[3] x_opt = fmin_cobyla(cost_fun, x0, constraint, rhobeg=radius_init, rhoend=radius_init * 1e-6, disp=disp) origin = x_opt[:3] radius = x_opt[3] return radius, origin def _check_origin(origin, info, coord_frame='head', disp=False): if isinstance(origin, string_types): if origin != 'auto': raise ValueError('origin must be a numerical array, or "auto", ' 'not %s' % (origin,)) if coord_frame == 'head': R, origin = fit_sphere_to_headshape(info, verbose=False, units='m')[:2] logger.info(' Automatic origin fit: head of radius %0.1f mm' % (R * 1000.,)) del R else: origin = (0., 0., 0.) origin = np.array(origin, float) if origin.shape != (3,): raise ValueError('origin must be a 3-element array') if disp: origin_str = ', '.join(['%0.1f' % (o * 1000) for o in origin]) msg = (' Using origin %s mm in the %s frame' % (origin_str, coord_frame)) if coord_frame == 'meg' and info['dev_head_t'] is not None: o_dev = apply_trans(info['dev_head_t'], origin) origin_str = ', '.join('%0.1f' % (o * 1000,) for o in o_dev) msg += ' (%s mm in the head frame)' % (origin_str,) logger.info(msg) return origin out): os.remove(surf_out) _symlink(surf_ws_out, surf_out) skip_symlink = False if skip_symlink: logger.info("Unable to create all symbolic links to .surf files " "in bem folder. Use --overwrite option to recreate " "them.") dest = op.join(bem_dir, 'watershed') else: logger.info("Symbolic links to .surf files created in bem folder") dest = bem_dir logger.info("\nThank you for waiting.\nThe BEM triangulations for this " "subject are now available at:\n%s." % dest) fname_head = op.join(bem_dir, subject + '-head.fif') if op.isfile(fname_head): os.remove(fname_head) surf = _surfaces_to_bem([op.join(ws_dir, subject + '_outer_skin_surface')], [FIFF.FIFFV_BEM_SURF_ID_HEAD], sigmas=[1]) write_bem_surfaces(fname_head, surf) if show: plot_bem(subject=subject, subjects_dir=subjects_dir, orientation='coronal', slices=None, show=True) logger.info('Created %s\n\nComplete.' % (fname_head,)) def _extract_volume_info(mgz, raise_error=True): try: import nibabel as nib except ImportError: return header = nib.load(mgz).header vol_info = dict() version = header['version'] if version == 1: version = '%s # volume info valid' % version else: raise ValueError('Volume info invalid.') vol_info['valid'] = version vol_info['filename'] = mgz vol_info['volume'] = header['dims'][:3] vol_info['voxelsize'] = header['delta'] vol_info['xras'], vol_info['yras'], vol_info['zras'] = header['Mdc'].T vol_info['cras'] = header['Pxyz_c'] return vol_info s, FIFF.FIFF_MNE_SOURCE_SPACE_NORMALS) if tag is None: tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_NORMALS) if tag is None: res['nn'] = None else: res['nn'] = tag.data.copy() if res['nn'].shape[0] != res['np']: raise ValueError('Vertex normal information is incorrect') tag = find_tag(fid, this, FIFF.FIFF_BEM_SURF_TRIANGLES) if tag is None: raise ValueError('Triangulation not found') res['tris'] = tag.data - 1 if res['tris'].shape[0] != res['ntri']: raise ValueError('Triangulation information is incorrect') return res @verbose def read_bem_solution(fname, verbose=None): logger.info('Loading surfaces...') bem_surfs = read_bem_surfaces(fname, patch_stats=True, verbose=False) if len(bem_surfs) == 3: logger.info('Three-layer model surfaces loaded.') needed = np.array([FIFF.FIFFV_BEM_SURF_ID_HEAD, FIFF.FIFFV_BEM_SURF_ID_SKULL, FIFF.FIFFV_BEM_SURF_ID_BRAIN]) if not all(x['id'] in needed for x in bem_surfs): raise RuntimeError('Could not find necessary BEM surfaces') reorder = [None] * 3 for x in bem_surfs: reorder[np.where(x['id'] == needed)[0][0]] = x bem_surfs = reorder elif len(bem_surfs) == 1: if not bem_surfs[0]['id'] == FIFF.FIFFV_BEM_SURF_ID_BRAIN: raise RuntimeError('BEM Surfaces not found') logger.info('Homogeneous model surface loaded.') # convert from surfaces to solution bem = ConductorModel(is_sphere=False, surfs=bem_surfs) logger.info('\nLoading the solution matrix...\n') f, tree, _ = fiff_open(fname) with f as fid: # Find the BEM data nodes = dir_tree_find(tree, FIFF.FIFFB_BEM) if len(nodes) == 0: raise RuntimeError('No BEM data in %s' % fname) bem_node = nodes[0] # Approximation method tag = find_tag(f, bem_node, FIFF.FIFF_BEM_APPROX) if tag is None: raise RuntimeError('No BEM solution found in %s' % fname) method = tag.data[0] if method not in (FIFF.FIFFV_BEM_APPROX_CONST, FIFF.FIFFV_BEM_APPROX_LINEAR): raise RuntimeError('Cannot handle BEM approximation method : %d' % method) tag = find_tag(fid, bem_node, FIFF.FIFF_BEM_POT_SOLUTION) dims = tag.data.shape if len(dims) != 2: raise RuntimeError('Expected a two-dimensional solution matrix ' 'instead of a %d dimensional one' % dims[0]) dim = 0 for surf in bem['surfs']: if method == FIFF.FIFFV_BEM_APPROX_LINEAR: dim += surf['np'] else: # method == FIFF.FIFFV_BEM_APPROX_CONST dim += surf['ntri'] if dims[0] != dim or dims[1] != dim: raise RuntimeError('Expected a %d x %d solution matrix instead of ' 'a %d x %d one' % (dim, dim, dims[1], dims[0])) sol = tag.data nsol = dims[0] bem['solution'] = sol bem['nsol'] = nsol bem['bem_method'] = method # Gamma factors and multipliers _add_gamma_multipliers(bem) kind = { FIFF.FIFFV_BEM_APPROX_CONST: 'constant collocation', FIFF.FIFFV_BEM_APPROX_LINEAR: 'linear_collocation', }[bem['bem_method']] logger.info('Loaded %s BEM solution from %s', kind, fname) return bem def _add_gamma_multipliers(bem): bem['sigma'] = np.array([surf['sigma'] for surf in bem['surfs']]) # Dirty trick for the zero conductivity outside sigma = np.r_[0.0, bem['sigma']] bem['source_mult'] = 2.0 / (sigma[1:] + sigma[:-1]) bem['field_mult'] = sigma[1:] - sigma[:-1] # make sure subsequent "zip"s work correctly assert len(bem['surfs']) == len(bem['field_mult']) bem['gamma'] = ((sigma[1:] - sigma[:-1])[np.newaxis, :] / (sigma[1:] + sigma[:-1])[:, np.newaxis]) _surf_dict = {'inner_skull': FIFF.FIFFV_BEM_SURF_ID_BRAIN, 'outer_skull': FIFF.FIFFV_BEM_SURF_ID_SKULL, 'head': FIFF.FIFFV_BEM_SURF_ID_HEAD} def _bem_find_surface(bem, id_): if isinstance(id_, string_types): name = id_ id_ = _surf_dict[id_] else: name = _bem_explain_surface(id_) idx = np.where(np.array([s['id'] for s in bem['surfs']]) == id_)[0] if len(idx) != 1: raise RuntimeError('BEM model does not have the %s triangulation' % name.replace('_', ' ')) return bem['surfs'][idx[0]] def _bem_explain_surface(id_): _rev_dict = dict((val, key) for key, val in _surf_dict.items()) return _rev_dict[id_] # ############################################################################ # Write def write_bem_surfaces(fname, surfs): if isinstance(surfs, dict): surfs = [surfs] with start_file(fname) as fid: start_block(fid, FIFF.FIFFB_BEM) write_int(fid, FIFF.FIFF_BEM_COORD_FRAME, surfs[0]['coord_frame']) _write_bem_surfaces_block(fid, surfs) end_block(fid, FIFF.FIFFB_BEM) end_file(fid) def _write_bem_surfaces_block(fid, surfs): for surf in surfs: start_block(fid, FIFF.FIFFB_BEM_SURF) write_float(fid, FIFF.FIFF_BEM_SIGMA, surf['sigma']) write_int(fid, FIFF.FIFF_BEM_SURF_ID, surf['id']) write_int(fid, FIFF.FIFF_MNE_COORD_FRAME, surf['coord_frame']) write_int(fid, FIFF.FIFF_BEM_SURF_NNODE, surf['np']) write_int(fid, FIFF.FIFF_BEM_SURF_NTRI, surf['ntri']) write_float_matrix(fid, FIFF.FIFF_BEM_SURF_NODES, surf['rr']) # index start at 0 in Python write_int_matrix(fid, FIFF.FIFF_BEM_SURF_TRIANGLES, surf['tris'] + 1) if 'nn' in surf and surf['nn'] is not None and len(surf['nn']) > 0: write_float_matrix(fid, FIFF.FIFF_BEM_SURF_NORMALS, surf['nn']) end_block(fid, FIFF.FIFFB_BEM_SURF) def write_bem_solution(fname, bem): _check_bem_size(bem['surfs']) with start_file(fname) as fid: start_block(fid, FIFF.FIFFB_BEM) # Coordinate frame (mainly for backward compatibility) write_int(fid, FIFF.FIFF_BEM_COORD_FRAME, bem['surfs'][0]['coord_frame']) # Surfaces _write_bem_surfaces_block(fid, bem['surfs']) # The potential solution if 'solution' in bem: if bem['bem_method'] != FIFF.FWD_BEM_LINEAR_COLL: raise RuntimeError('Only linear collocation supported') write_int(fid, FIFF.FIFF_BEM_APPROX, FIFF.FIFFV_BEM_APPROX_LINEAR) write_float_matrix(fid, FIFF.FIFF_BEM_POT_SOLUTION, bem['solution']) end_block(fid, FIFF.FIFFB_BEM) end_file(fid) # ############################################################################# # Create 3-Layers BEM model from Flash MRI images def _prepare_env(subject, subjects_dir, requires_freesurfer): env = os.environ.copy() if requires_freesurfer and not os.environ.get('FREESURFER_HOME'): raise RuntimeError('I cannot find freesurfer. The FREESURFER_HOME ' 'environment variable is not set.') if not isinstance(subject, string_types): raise TypeError('The subject argument must be set') subjects_dir = get_subjects_dir(subjects_dir, raise_error=True) if not op.isdir(subjects_dir): raise RuntimeError('Could not find the MRI data directory "%s"' % subjects_dir) subject_dir = op.join(subjects_dir, subject) if not op.isdir(subject_dir): raise RuntimeError('Could not find the subject data directory "%s"' % (subject_dir,)) env['SUBJECT'] = subject env['SUBJECTS_DIR'] = subjects_dir mri_dir = op.join(subject_dir, 'mri') bem_dir = op.join(subject_dir, 'bem') return env, mri_dir, bem_dir @verbose def convert_flash_mris(subject, flash30=True, convert=True, unwarp=False, subjects_dir=None, verbose=None): env, mri_dir = _prepare_env(subject, subjects_dir, requires_freesurfer=True)[:2] curdir = os.getcwd() # Step 1a : Data conversion to mgz format if not op.exists(op.join(mri_dir, 'flash', 'parameter_maps')): os.makedirs(op.join(mri_dir, 'flash', 'parameter_maps')) echos_done = 0 if convert: logger.info("\n---- Converting Flash images ----") echos = ['001', '002', '003', '004', '005', '006', '007', '008'] if flash30: flashes = ['05'] else: flashes = ['05', '30'] # missing = False for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): missing = True if missing: echos = ['002', '003', '004', '005', '006', '007', '008', '009'] for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): raise RuntimeError("Directory %s is missing." % op.join('flash' + flash, echo)) # for flash in flashes: for echo in echos: if not op.isdir(op.join('flash' + flash, echo)): raise RuntimeError("Directory %s is missing." % op.join('flash' + flash, echo)) sample_file = glob.glob(op.join('flash' + flash, echo, '*'))[0] dest_file = op.join(mri_dir, 'flash', 'mef' + flash + '_' + echo + '.mgz') # do not redo if already present if op.isfile(dest_file): logger.info("The file %s is already there") else: cmd = ['mri_convert', sample_file, dest_file] run_subprocess(cmd, env=env) echos_done += 1 # Step 1b : Run grad_unwarp on converted files os.chdir(op.join(mri_dir, "flash")) files = glob.glob("mef*.mgz") if unwarp: logger.info("\n---- Unwarp mgz data sets ----") for infile in files: outfile = infile.replace(".mgz", "u.mgz") cmd = ['grad_unwarp', '-i', infile, '-o', outfile, '-unwarp', 'true'] run_subprocess(cmd, env=env) # Clear parameter maps if some of the data were reconverted if echos_done > 0 and op.exists("parameter_maps"): shutil.rmtree("parameter_maps") logger.info("\nParameter maps directory cleared") if not op.exists("parameter_maps"): os.makedirs("parameter_maps") # Step 2 : Create the parameter maps if flash30: logger.info("\n---- Creating the parameter maps ----") if unwarp: files = glob.glob("mef05*u.mgz") if len(os.listdir('parameter_maps')) == 0: cmd = ['mri_ms_fitparms'] + files + ['parameter_maps'] run_subprocess(cmd, env=env) else: logger.info("Parameter maps were already computed") # Step 3 : Synthesize the flash 5 images logger.info("\n---- Synthesizing flash 5 images ----") os.chdir('parameter_maps') if not op.exists('flash5.mgz'): cmd = ['mri_synthesize', '20 5 5', 'T1.mgz', 'PD.mgz', 'flash5.mgz'] run_subprocess(cmd, env=env) os.remove('flash5_reg.mgz') else: logger.info("Synthesized flash 5 volume is already there") else: logger.info("\n---- Averaging flash5 echoes ----") os.chdir('parameter_maps') if unwarp: files = glob.glob("mef05*u.mgz") else: files = glob.glob("mef05*.mgz") cmd = ['mri_average', '-noconform', files, 'flash5.mgz'] run_subprocess(cmd, env=env) if op.exists('flash5_reg.mgz'): os.remove('flash5_reg.mgz') # Go back to initial directory os.chdir(curdir) @verbose def make_flash_bem(subject, overwrite=False, show=True, subjects_dir=None, flash_path=None, verbose=None): from .viz.misc import plot_bem is_test = os.environ.get('MNE_SKIP_FS_FLASH_CALL', False) env, mri_dir, bem_dir = _prepare_env(subject, subjects_dir, requires_freesurfer=True) if flash_path is None: flash_path = op.join(mri_dir, 'flash', 'parameter_maps') else: flash_path = op.abspath(flash_path) curdir = os.getcwd() subjects_dir = env['SUBJECTS_DIR'] logger.info('\nProcessing the flash MRI data to produce BEM meshes with ' 'the following parameters:\n' 'SUBJECTS_DIR = %s\n' 'SUBJECT = %s\n' 'Result dir = %s\n' % (subjects_dir, subject, op.join(bem_dir, 'flash'))) # Step 4 : Register with MPRAGE logger.info("\n---- Registering flash 5 with MPRAGE ----") flash5 = op.join(flash_path, 'flash5.mgz') flash5_reg = op.join(flash_path, 'flash5_reg.mgz') if not op.exists(flash5_reg): if op.exists(op.join(mri_dir, 'T1.mgz')): ref_volume = op.join(mri_dir, 'T1.mgz') else: ref_volume = op.join(mri_dir, 'T1') cmd = ['fsl_rigid_register', '-r', ref_volume, '-i', flash5, '-o', flash5_reg] run_subprocess(cmd, env=env) else: logger.info("Registered flash 5 image is already there") # Step 5a : Convert flash5 into COR logger.info("\n---- Converting flash5 volume into COR format ----") shutil.rmtree(op.join(mri_dir, 'flash5'), ignore_errors=True) os.makedirs(op.join(mri_dir, 'flash5')) if not is_test: # CIs don't have freesurfer, skipped when testing. cmd = ['mri_convert', flash5_reg, op.join(mri_dir, 'flash5')] run_subprocess(cmd, env=env) os.chdir(mri_dir) convert_T1 = False if not op.isdir('T1') or len(glob.glob(op.join('T1', 'COR*'))) == 0: convert_T1 = True convert_brain = False if not op.isdir('brain') or len(glob.glob(op.join('brain', 'COR*'))) == 0: convert_brain = True logger.info("\n---- Converting T1 volume into COR format ----") if convert_T1: if not op.isfile('T1.mgz'): raise RuntimeError("Both T1 mgz and T1 COR volumes missing.") os.makedirs('T1') cmd = ['mri_convert', 'T1.mgz', 'T1'] run_subprocess(cmd, env=env) else: logger.info("T1 volume is already in COR format") logger.info("\n---- Converting brain volume into COR format ----") if convert_brain: if not op.isfile('brain.mgz'): raise RuntimeError("Both brain mgz and brain COR volumes missing.") os.makedirs('brain') cmd = ['mri_convert', 'brain.mgz', 'brain'] run_subprocess(cmd, env=env) else: logger.info("Brain volume is already in COR format") if not is_test: logger.info("\n---- Creating the BEM surfaces ----") cmd = ['mri_make_bem_surfaces', subject] run_subprocess(cmd, env=env) logger.info("\n---- Converting the tri files into surf files ----") os.chdir(bem_dir) if not op.exists('flash'): os.makedirs('flash') os.chdir('flash') surfs = ['inner_skull', 'outer_skull', 'outer_skin'] for surf in surfs: shutil.move(op.join(bem_dir, surf + '.tri'), surf + '.tri') nodes, tris = read_tri(surf + '.tri', swap=True) vol_info = _extract_volume_info(flash5_reg) if vol_info is None: warn('nibabel is required to update the volume info. Volume info ' 'omitted from the written surface.') else: vol_info['head'] = np.array([20]) write_surface(surf + '.surf', nodes, tris, volume_info=vol_info) # Cleanup section logger.info("\n---- Cleaning up ----") os.chdir(bem_dir) os.remove('inner_skull_tmp.tri') os.chdir(mri_dir) if convert_T1: shutil.rmtree('T1') logger.info("Deleted the T1 COR volume") if convert_brain: shutil.rmtree('brain') logger.info("Deleted the brain COR volume") shutil.rmtree('flash5') logger.info("Deleted the flash5 COR volume") # Create symbolic links to the .surf files in the bem folder logger.info("\n---- Creating symbolic links ----") os.chdir(bem_dir) for surf in surfs: surf = surf + '.surf' if not overwrite and op.exists(surf): skip_symlink = True else: if op.exists(surf): os.remove(surf) _symlink(op.join('flash', surf), op.join(surf)) skip_symlink = False if skip_symlink: logger.info("Unable to create all symbolic links to .surf files " "in bem folder. Use --overwrite option to recreate them.") dest = op.join(bem_dir, 'flash') else: logger.info("Symbolic links to .surf files created in bem folder") dest = bem_dir logger.info("\nThank you for waiting.\nThe BEM triangulations for this " "subject are now available at:\n%s.\nWe hope the BEM meshes " "created will facilitate your MEG and EEG data analyses." % dest) # Show computed BEM surfaces if show: plot_bem(subject=subject, subjects_dir=subjects_dir, orientation='coronal', slices=None, show=True) # Go back to initial directory os.chdir(curdir) def _check_bem_size(surfs): if len(surfs) > 1 and surfs[0]['np'] > 10000: warn('The bem surfaces have %s data points. 5120 (ico grade=4) ' 'should be enough. Dense 3-layer bems may not save properly.' % surfs[0]['np']) def _symlink(src, dest): try: os.symlink(src, dest) except OSError: warn('Could not create symbolic link %s. Check that your partition ' 'handles symbolic links. The file will be copied instead.' % dest) shutil.copy(src, dest)
true
true
f717c05e56131daf3f0bdbca7cbed9fe45f2de19
930
py
Python
src/augment/basic_augmentation/noise.py
TencentYoutuResearch/SelfSupervisedLearning-DSM
655a0a23a47bf2559f3d435384ae59a8871a5ff5
[ "Apache-2.0" ]
27
2021-01-07T11:09:33.000Z
2021-08-31T02:46:23.000Z
src/augment/basic_augmentation/noise.py
TencentYoutuResearch/SelfSupervisedLearning-DSM
655a0a23a47bf2559f3d435384ae59a8871a5ff5
[ "Apache-2.0" ]
null
null
null
src/augment/basic_augmentation/noise.py
TencentYoutuResearch/SelfSupervisedLearning-DSM
655a0a23a47bf2559f3d435384ae59a8871a5ff5
[ "Apache-2.0" ]
3
2021-01-08T08:31:06.000Z
2021-11-26T04:10:23.000Z
import torch import torch.nn as nn """ usage z_rand = generate_noise([1,nzx,nzy], device=opt.device) z_rand = z_rand.expand(1,3,Z_opt.shape[2],Z_opt.shape[3]) z_prev1 = 0.95*Z_opt +0.05*z_rand """ def upsampling(im, sx, sy): m = nn.Upsample(size=[round(sx), round(sy)], mode='bilinear', align_corners=True) return m(im) def generate_noise(size, num_samp=1, device='cuda', type='gaussian', scale=1): if type == 'gaussian': noise = torch.randn(num_samp, size[0], round(size[1]/scale), round(size[2]/scale)) noise = upsampling(noise, size[1], size[2]) if type == 'gaussian_mixture': noise1 = torch.randn(num_samp, size[0], size[1], size[2]) + 5 noise2 = torch.randn(num_samp, size[0], size[1], size[2]) noise = noise1 + noise2 if type == 'uniform': noise = torch.randn(num_samp, size[0], size[1], size[2]) return noise
32.068966
90
0.609677
import torch import torch.nn as nn def upsampling(im, sx, sy): m = nn.Upsample(size=[round(sx), round(sy)], mode='bilinear', align_corners=True) return m(im) def generate_noise(size, num_samp=1, device='cuda', type='gaussian', scale=1): if type == 'gaussian': noise = torch.randn(num_samp, size[0], round(size[1]/scale), round(size[2]/scale)) noise = upsampling(noise, size[1], size[2]) if type == 'gaussian_mixture': noise1 = torch.randn(num_samp, size[0], size[1], size[2]) + 5 noise2 = torch.randn(num_samp, size[0], size[1], size[2]) noise = noise1 + noise2 if type == 'uniform': noise = torch.randn(num_samp, size[0], size[1], size[2]) return noise
true
true
f717c11ea68b58ce09246f3ab62dcc7714c8c8ca
1,921
py
Python
recipes/pneumothorax/binary_classification/train_example_deit.py
lbolanos/HugsVision
c9b1708928916bc413a795590d731308ea8c582b
[ "MIT" ]
154
2021-08-14T00:09:15.000Z
2022-03-16T02:42:01.000Z
recipes/pneumothorax/binary_classification/train_example_deit.py
lbolanos/HugsVision
c9b1708928916bc413a795590d731308ea8c582b
[ "MIT" ]
36
2021-08-13T23:42:15.000Z
2022-03-31T03:57:23.000Z
recipes/pneumothorax/binary_classification/train_example_deit.py
lbolanos/HugsVision
c9b1708928916bc413a795590d731308ea8c582b
[ "MIT" ]
12
2021-09-04T00:49:35.000Z
2022-02-24T13:20:54.000Z
import argparse from hugsvision.nnet.VisionClassifierTrainer import VisionClassifierTrainer from hugsvision.dataio.VisionDataset import VisionDataset from torchvision.datasets import ImageFolder from transformers import DeiTFeatureExtractor, DeiTForImageClassification parser = argparse.ArgumentParser(description='Image classifier') parser.add_argument('--name', type=str, default="MyVitModel", help='The name of the model') parser.add_argument('--imgs', type=str, default="./images/", help='The directory of the input images') parser.add_argument('--output', type=str, default="./out/", help='The output directory of the model') parser.add_argument('--epochs', type=int, default=1, help='Number of Epochs') args = parser.parse_args() # Load the dataset train, test, id2label, label2id = VisionDataset.fromImageFolder( args.imgs, test_ratio = 0.15, balanced = True, augmentation = True, ) # # Load the dataset # train, test, id2label, label2id = VisionDataset.fromImageFolders( # "/<PATH>/train/", # "/<PATH>/test/", # ) huggingface_model = "facebook/deit-base-distilled-patch16-224" # Train the model trainer = VisionClassifierTrainer( model_name = args.name, train = train, test = test, output_dir = args.output, max_epochs = args.epochs, cores = 4, batch_size = 32, model = DeiTForImageClassification.from_pretrained( huggingface_model, num_labels = len(label2id), label2id = label2id, id2label = id2label ), feature_extractor = DeiTFeatureExtractor.from_pretrained( huggingface_model ), ) # Evaluate on the test sub-dataset ref, hyp = trainer.evaluate_f1_score() # Test on a single image trainer.testing(img='./data/demo/42.png',expected=2) trainer.testing(img='./data/demo/3.jpg',expected=0) trainer.testing(img='./data/demo/5.jpg',expected=2) trainer.testing(img='./data/demo/4.jpg',expected=1)
32.559322
103
0.724623
import argparse from hugsvision.nnet.VisionClassifierTrainer import VisionClassifierTrainer from hugsvision.dataio.VisionDataset import VisionDataset from torchvision.datasets import ImageFolder from transformers import DeiTFeatureExtractor, DeiTForImageClassification parser = argparse.ArgumentParser(description='Image classifier') parser.add_argument('--name', type=str, default="MyVitModel", help='The name of the model') parser.add_argument('--imgs', type=str, default="./images/", help='The directory of the input images') parser.add_argument('--output', type=str, default="./out/", help='The output directory of the model') parser.add_argument('--epochs', type=int, default=1, help='Number of Epochs') args = parser.parse_args() train, test, id2label, label2id = VisionDataset.fromImageFolder( args.imgs, test_ratio = 0.15, balanced = True, augmentation = True, ) model = "facebook/deit-base-distilled-patch16-224" trainer = VisionClassifierTrainer( model_name = args.name, train = train, test = test, output_dir = args.output, max_epochs = args.epochs, cores = 4, batch_size = 32, model = DeiTForImageClassification.from_pretrained( huggingface_model, num_labels = len(label2id), label2id = label2id, id2label = id2label ), feature_extractor = DeiTFeatureExtractor.from_pretrained( huggingface_model ), ) ref, hyp = trainer.evaluate_f1_score() trainer.testing(img='./data/demo/42.png',expected=2) trainer.testing(img='./data/demo/3.jpg',expected=0) trainer.testing(img='./data/demo/5.jpg',expected=2) trainer.testing(img='./data/demo/4.jpg',expected=1)
true
true
f717c1795cee9e960f89d88b04da89c0fcbb68f9
1,593
py
Python
algorithms/LinkedList/reverseLinkedList.py
gadodia/Algorithms
714d43ce89b684d8e5c4bb77a654938957957f33
[ "MIT" ]
null
null
null
algorithms/LinkedList/reverseLinkedList.py
gadodia/Algorithms
714d43ce89b684d8e5c4bb77a654938957957f33
[ "MIT" ]
null
null
null
algorithms/LinkedList/reverseLinkedList.py
gadodia/Algorithms
714d43ce89b684d8e5c4bb77a654938957957f33
[ "MIT" ]
null
null
null
''' This problem was recently asked by Google: Given a singly-linked list, reverse the list. This can be done iteratively or recursively. Can you get both solutions? Example: Input: 4 -> 3 -> 2 -> 1 -> 0 -> NULL Output: 0 -> 1 -> 2 -> 3 -> 4 -> NULL ''' class ListNode(object): def __init__(self, x): self.val = x self.next = None # Function to print the list def printList(self): node = self output = '' while node != None: output += str(node.val) output += " " node = node.next print(output) # Iterative Solution def reverseIteratively(self, head): if not head or not head.next: return head prev = None cur = head while cur: temp = cur.next cur.next = prev prev = cur cur = temp return prev # Recursive Solution def reverseRecursively(self, head): if not head or not head.next: return head p = self.reverseRecursively(head.next) head.next.next = head head.next = None return p # Test Program # Initialize the test list: testHead = ListNode(4) node1 = ListNode(3) testHead.next = node1 node2 = ListNode(2) node1.next = node2 node3 = ListNode(1) node2.next = node3 testTail = ListNode(0) node3.next = testTail print("Initial list: ") testHead.printList() # 4 3 2 1 0 # testHead.reverseIteratively(testHead) testHead.reverseRecursively(testHead) print("List after reversal: ") testTail.printList() # 0 1 2 3 4
23.426471
118
0.591337
class ListNode(object): def __init__(self, x): self.val = x self.next = None def printList(self): node = self output = '' while node != None: output += str(node.val) output += " " node = node.next print(output) def reverseIteratively(self, head): if not head or not head.next: return head prev = None cur = head while cur: temp = cur.next cur.next = prev prev = cur cur = temp return prev def reverseRecursively(self, head): if not head or not head.next: return head p = self.reverseRecursively(head.next) head.next.next = head head.next = None return p testHead = ListNode(4) node1 = ListNode(3) testHead.next = node1 node2 = ListNode(2) node1.next = node2 node3 = ListNode(1) node2.next = node3 testTail = ListNode(0) node3.next = testTail print("Initial list: ") testHead.printList() testHead.reverseRecursively(testHead) print("List after reversal: ") testTail.printList()
true
true
f717c2255fe90f0a652e062275efe541f399cc62
22,811
py
Python
utils/exp_utils.py
NoeSamaille/medical-detection-toolkit
232d3d1444ccaac04e15a00d8030390560236871
[ "Apache-2.0" ]
1
2021-04-01T02:15:49.000Z
2021-04-01T02:15:49.000Z
utils/exp_utils.py
NoeSamaille/medical-detection-toolkit
232d3d1444ccaac04e15a00d8030390560236871
[ "Apache-2.0" ]
2
2021-06-02T02:14:01.000Z
2021-09-06T06:03:25.000Z
utils/exp_utils.py
NoeSamaille/medical-detection-toolkit
232d3d1444ccaac04e15a00d8030390560236871
[ "Apache-2.0" ]
3
2020-06-28T14:57:00.000Z
2021-09-06T08:26:26.000Z
#!/usr/bin/env python # Copyright 2018 Division of Medical Image Computing, German Cancer Research Center (DKFZ). # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== from typing import Iterable, Tuple, Any, Union import os, sys import subprocess from multiprocessing import Process import importlib.util import pickle import logging from torch.utils.tensorboard import SummaryWriter from collections import OrderedDict import numpy as np import torch import pandas as pd def split_off_process(target, *args, daemon: bool=False, **kwargs): """Start a process that won't block parent script. No join(), no return value. If daemon=False: before parent exits, it waits for this to finish. :param target: the target function of the process. :params *args: args to pass to target. :param daemon: if False: before parent exits, it waits for this process to finish. :params **kwargs: kwargs to pass to target. """ p = Process(target=target, args=tuple(args), kwargs=kwargs, daemon=daemon) p.start() return p def get_formatted_duration(seconds: float, format: str="hms") -> str: """Format a time in seconds. :param format: "hms" for hours mins secs or "ms" for min secs. """ mins, secs = divmod(seconds, 60) if format == "ms": t = "{:d}m:{:02d}s".format(int(mins), int(secs)) elif format == "hms": h, mins = divmod(mins, 60) t = "{:d}h:{:02d}m:{:02d}s".format(int(h), int(mins), int(secs)) else: raise Exception("Format {} not available, only 'hms' or 'ms'".format(format)) return t class CombinedLogger(object): """Combine console and tensorboard logger and record system metrics. """ def __init__(self, name: str, log_dir: str, server_env: bool=True, fold: Union[int, str]="all"): self.pylogger = logging.getLogger(name) self.tboard = SummaryWriter(log_dir=os.path.join(log_dir, "tboard")) self.log_dir = log_dir self.fold = str(fold) self.server_env = server_env self.pylogger.setLevel(logging.DEBUG) self.log_file = os.path.join(log_dir, "fold_"+self.fold, 'exec.log') os.makedirs(os.path.dirname(self.log_file), exist_ok=True) self.pylogger.addHandler(logging.FileHandler(self.log_file)) if not server_env: self.pylogger.addHandler(ColorHandler()) else: self.pylogger.addHandler(logging.StreamHandler()) self.pylogger.propagate = False def __getattr__(self, attr): """delegate all undefined method requests to objects of this class in order pylogger, tboard (first find first serve). E.g., combinedlogger.add_scalars(...) should trigger self.tboard.add_scalars(...) """ for obj in [self.pylogger, self.tboard]: if attr in dir(obj): return getattr(obj, attr) print("logger attr not found") def set_logfile(self, fold: Union[int, str, None]=None, log_file: Union[str, None]=None): if fold is not None: self.fold = str(fold) if log_file is None: self.log_file = os.path.join(self.log_dir, "fold_"+self.fold, 'exec.log') else: self.log_file = log_file os.makedirs(os.path.dirname(self.log_file), exist_ok=True) for hdlr in self.pylogger.handlers: hdlr.close() self.pylogger.handlers = [] self.pylogger.addHandler(logging.FileHandler(self.log_file)) if not self.server_env: self.pylogger.addHandler(ColorHandler()) else: self.pylogger.addHandler(logging.StreamHandler()) def metrics2tboard(self, metrics, global_step=None, suptitle=None): """ :param metrics: {'train': dataframe, 'val':df}, df as produced in evaluator.py.evaluate_predictions """ # print("metrics", metrics) if global_step is None: global_step = len(metrics['train'][list(metrics['train'].keys())[0]]) - 1 if suptitle is not None: suptitle = str(suptitle) else: suptitle = "Fold_" + str(self.fold) for key in ['train', 'val']: # series = {k:np.array(v[-1]) for (k,v) in metrics[key].items() if not np.isnan(v[-1]) and not 'Bin_Stats' in k} loss_series = {} mon_met_series = {} for tag, val in metrics[key].items(): val = val[-1] # maybe remove list wrapping, recording in evaluator? if 'loss' in tag.lower() and not np.isnan(val): loss_series["{}".format(tag)] = val elif not np.isnan(val): mon_met_series["{}".format(tag)] = val self.tboard.add_scalars(suptitle + "/Losses/{}".format(key), loss_series, global_step) self.tboard.add_scalars(suptitle + "/Monitor_Metrics/{}".format(key), mon_met_series, global_step) self.tboard.add_scalars(suptitle + "/Learning_Rate", metrics["lr"], global_step) return def __del__(self): # otherwise might produce multiple prints e.g. in ipython console for hdlr in self.pylogger.handlers: hdlr.close() self.pylogger.handlers = [] del self.pylogger self.tboard.flush() # close somehow prevents main script from exiting # maybe revise this issue in a later pytorch version #self.tboard.close() def get_logger(exp_dir: str, server_env: bool=False) -> CombinedLogger: """ creates logger instance. writing out info to file, to terminal and to tensorboard. :param exp_dir: experiment directory, where exec.log file is stored. :param server_env: True if operating in server environment (e.g., gpu cluster) :return: custom CombinedLogger instance. """ log_dir = os.path.join(exp_dir, "logs") logger = CombinedLogger('medicaldetectiontoolkit', log_dir, server_env=server_env) print("Logging to {}".format(logger.log_file)) return logger def prep_exp(dataset_path, exp_path, server_env, use_stored_settings=True, is_training=True): """ I/O handling, creating of experiment folder structure. Also creates a snapshot of configs/model scripts and copies them to the exp_dir. This way the exp_dir contains all info needed to conduct an experiment, independent to changes in actual source code. Thus, training/inference of this experiment can be started at anytime. Therefore, the model script is copied back to the source code dir as tmp_model (tmp_backbone). Provides robust structure for cloud deployment. :param dataset_path: path to source code for specific data set. (e.g. medicaldetectiontoolkit/lidc_exp) :param exp_path: path to experiment directory. :param server_env: boolean flag. pass to configs script for cloud deployment. :param use_stored_settings: boolean flag. When starting training: If True, starts training from snapshot in existing experiment directory, else creates experiment directory on the fly using configs/model scripts from source code. :param is_training: boolean flag. distinguishes train vs. inference mode. :return: """ if is_training: if use_stored_settings: cf_file = import_module('cf_file', os.path.join(exp_path, 'configs.py')) cf = cf_file.configs(server_env) # in this mode, previously saved model and backbone need to be found in exp dir. if not os.path.isfile(os.path.join(exp_path, 'mdt_model.py')) or \ not os.path.isfile(os.path.join(exp_path, 'backbone.py')): raise Exception( "Selected use_stored_settings option but no model and/or backbone source files exist in exp dir.") cf.model_path = os.path.join(exp_path, 'mdt_model.py') cf.backbone_path = os.path.join(exp_path, 'backbone.py') else: # this case overwrites settings files in exp dir, i.e., default_configs, configs, backbone, model os.makedirs(exp_path, exist_ok=True) # run training with source code info and copy snapshot of model to exp_dir for later testing (overwrite scripts if exp_dir already exists.) subprocess.call('cp {} {}'.format('default_configs.py', os.path.join(exp_path, 'default_configs.py')), shell=True) subprocess.call( 'cp {} {}'.format(os.path.join(dataset_path, 'configs.py'), os.path.join(exp_path, 'configs.py')), shell=True) cf_file = import_module('cf_file', os.path.join(dataset_path, 'configs.py')) cf = cf_file.configs(server_env) subprocess.call('cp {} {}'.format(cf.model_path, os.path.join(exp_path, 'mdt_model.py')), shell=True) subprocess.call('cp {} {}'.format(cf.backbone_path, os.path.join(exp_path, 'backbone.py')), shell=True) if os.path.isfile(os.path.join(exp_path, "fold_ids.pickle")): subprocess.call('rm {}'.format(os.path.join(exp_path, "fold_ids.pickle")), shell=True) else: # testing, use model and backbone stored in exp dir. cf_file = import_module('cf_file', os.path.join(exp_path, 'configs.py')) cf = cf_file.configs(server_env) cf.model_path = os.path.join(exp_path, 'mdt_model.py') cf.backbone_path = os.path.join(exp_path, 'backbone.py') cf.exp_dir = exp_path cf.test_dir = os.path.join(cf.exp_dir, 'test') cf.plot_dir = os.path.join(cf.exp_dir, 'plots') if not os.path.exists(cf.test_dir): os.mkdir(cf.test_dir) if not os.path.exists(cf.plot_dir): os.mkdir(cf.plot_dir) cf.experiment_name = exp_path.split("/")[-1] cf.created_fold_id_pickle = False return cf def import_module(name: str, path: str): """ correct way of importing a module dynamically in python 3. :param name: name given to module instance. :param path: path to module. :return: module: returned module instance. """ spec = importlib.util.spec_from_file_location(name, path) module = importlib.util.module_from_spec(spec) spec.loader.exec_module(module) return module def set_params_flag(module: torch.nn.Module, flag: Tuple[str, Any], check_overwrite: bool = True) -> torch.nn.Module: """Set an attribute for all passed module parameters. :param flag: tuple (str attribute name : attr value) :param check_overwrite: if True, assert that attribute not already exists. """ for param in module.parameters(): if check_overwrite: assert not hasattr(param, flag[0]), \ "param {} already has attr {} (w/ val {})".format(param, flag[0], getattr(param, flag[0])) setattr(param, flag[0], flag[1]) return module def parse_params_for_optim(net: torch.nn.Module, weight_decay: float = 0., exclude_from_wd: Iterable = ("norm",)) -> list: """Split network parameters into weight-decay dependent groups for the optimizer. :param net: network. :param weight_decay: weight decay value for the parameters that it is applied to. excluded parameters will have weight decay 0. :param exclude_from_wd: List of strings of parameter-group names to exclude from weight decay. Options: "norm", "bias". :return: """ if weight_decay is None: weight_decay = 0. # pytorch implements parameter groups as dicts {'params': ...} and # weight decay as p.data.mul_(1 - group['lr'] * group['weight_decay']) norm_types = [torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d, torch.nn.InstanceNorm1d, torch.nn.InstanceNorm2d, torch.nn.InstanceNorm3d, torch.nn.LayerNorm, torch.nn.GroupNorm, torch.nn.SyncBatchNorm, torch.nn.LocalResponseNorm] level_map = {"bias": "weight", "norm": "module"} type_map = {"norm": norm_types} exclude_from_wd = [str(name).lower() for name in exclude_from_wd] exclude_weight_names = [k for k, v in level_map.items() if k in exclude_from_wd and v == "weight"] exclude_module_types = tuple([type_ for k, v in level_map.items() if (k in exclude_from_wd and v == "module") for type_ in type_map[k]]) if exclude_from_wd: print("excluding {} from weight decay.".format(exclude_from_wd)) for module in net.modules(): if isinstance(module, exclude_module_types): set_params_flag(module, ("no_wd", True)) for param_name, param in net.named_parameters(): if np.any([ename in param_name for ename in exclude_weight_names]): setattr(param, "no_wd", True) with_dec, no_dec = [], [] for param in net.parameters(): if hasattr(param, "no_wd") and param.no_wd == True: no_dec.append(param) else: with_dec.append(param) orig_ps = sum(p.numel() for p in net.parameters()) with_ps = sum(p.numel() for p in with_dec) wo_ps = sum(p.numel() for p in no_dec) assert orig_ps == with_ps + wo_ps, "orig n parameters {} unequals sum of with wd {} and w/o wd {}."\ .format(orig_ps, with_ps, wo_ps) groups = [{'params': gr, 'weight_decay': wd} for (gr, wd) in [(no_dec, 0.), (with_dec, weight_decay)] if len(gr)>0] return groups class ModelSelector: ''' saves a checkpoint after each epoch as 'last_state' (can be loaded to continue interrupted training). saves the top-k (k=cf.save_n_models) ranked epochs. In inference, predictions of multiple epochs can be ensembled to improve performance. ''' def __init__(self, cf, logger): self.cf = cf self.saved_epochs = [-1] * cf.save_n_models self.logger = logger def run_model_selection(self, net: torch.nn.Module, optimizer: torch.optim.Optimizer, monitor_metrics: dict, epoch: int): # take the mean over all selection criteria in each epoch non_nan_scores = np.mean(np.array([[0 if (ii is None or np.isnan(ii)) else ii for ii in monitor_metrics['val'][sc]] for sc in self.cf.model_selection_criteria]), 0) epochs_scores = [ii for ii in non_nan_scores[1:]] # ranking of epochs according to model_selection_criterion epoch_ranking = np.argsort(epochs_scores, kind="stable")[::-1] + 1 #epochs start at 1 # if set in configs, epochs < min_save_thresh are discarded from saving process. epoch_ranking = epoch_ranking[epoch_ranking >= self.cf.min_save_thresh] # check if current epoch is among the top-k epochs. if epoch in epoch_ranking[:self.cf.save_n_models]: save_dir = os.path.join(self.cf.fold_dir, '{}_best_checkpoint'.format(epoch)) if not os.path.exists(save_dir): os.mkdir(save_dir) torch.save(net.state_dict(), os.path.join(save_dir, 'params.pth')) with open(os.path.join(save_dir, 'monitor_metrics.pickle'), 'wb') as handle: pickle.dump(monitor_metrics, handle) # save epoch_ranking to keep info for inference. np.save(os.path.join(self.cf.fold_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) np.save(os.path.join(save_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) self.logger.info( "saving current epoch {} at rank {}".format(epoch, np.argwhere(epoch_ranking == epoch))) # delete params of the epoch that just fell out of the top-k epochs. for se in [int(ii.split('_')[0]) for ii in os.listdir(self.cf.fold_dir) if 'best_checkpoint' in ii]: if se in epoch_ranking[self.cf.save_n_models:]: subprocess.call('rm -rf {}'.format(os.path.join(self.cf.fold_dir, '{}_best_checkpoint'.format(se))), shell=True) self.logger.info('deleting epoch {} at rank {}'.format(se, np.argwhere(epoch_ranking == se))) state = { 'epoch': epoch, 'state_dict': net.state_dict(), 'optimizer': optimizer.state_dict(), } # save checkpoint of current epoch. save_dir = os.path.join(self.cf.fold_dir, 'last_checkpoint'.format(epoch)) if not os.path.exists(save_dir): os.mkdir(save_dir) torch.save(state, os.path.join(save_dir, 'params.pth')) np.save(os.path.join(save_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) with open(os.path.join(save_dir, 'monitor_metrics.pickle'), 'wb') as handle: pickle.dump(monitor_metrics, handle) return os.path.join(os.path.join(self.cf.fold_dir, f'{epoch_ranking[0]}_best_checkpoint')) def load_checkpoint(checkpoint_path: str, net: torch.nn.Module, optimizer: torch.optim.Optimizer) -> Tuple: checkpoint = torch.load(os.path.join(checkpoint_path, 'params.pth')) net.load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) with open(os.path.join(checkpoint_path, 'monitor_metrics.pickle'), 'rb') as handle: monitor_metrics = pickle.load(handle) starting_epoch = checkpoint['epoch'] + 1 return starting_epoch, net, optimizer, monitor_metrics def prepare_monitoring(cf): """ creates dictionaries, where train/val metrics are stored. """ metrics = {} # first entry for loss dict accounts for epoch starting at 1. metrics['train'] = OrderedDict() metrics['val'] = OrderedDict() metric_classes = [] if 'rois' in cf.report_score_level: metric_classes.extend([v for k, v in cf.class_dict.items()]) if 'patient' in cf.report_score_level: metric_classes.extend(['patient']) for cl in metric_classes: metrics['train'][cl + '_ap'] = [np.nan] metrics['val'][cl + '_ap'] = [np.nan] if cl == 'patient': metrics['train'][cl + '_auc'] = [np.nan] metrics['val'][cl + '_auc'] = [np.nan] return metrics def create_csv_output(results_list, cf, logger): """ Write out test set predictions to .csv file. output format is one line per prediction: PatientID | PredictionID | [y1 x1 y2 x2 (z1) (z2)] | score | pred_classID Note, that prediction coordinates correspond to images as loaded for training/testing and need to be adapted when plotted over raw data (before preprocessing/resampling). :param results_list: [[patient_results, patient_id], [patient_results, patient_id], ...] """ logger.info('creating csv output file at {}'.format(os.path.join(cf.test_dir, 'results.csv'))) predictions_df = pd.DataFrame(columns = ['patientID', 'predictionID', 'coords', 'score', 'pred_classID']) for r in results_list: pid = r[1] #optionally load resampling info from preprocessing to match output predictions with raw data. #with open(os.path.join(cf.exp_dir, 'test_resampling_info', pid), 'rb') as handle: # resampling_info = pickle.load(handle) for bix, box in enumerate(r[0][0]): if box["box_type"] == "gt": continue assert box['box_type'] == 'det', box['box_type'] coords = box['box_coords'] score = box['box_score'] pred_class_id = box['box_pred_class_id'] out_coords = [] if score >= cf.min_det_thresh: out_coords.append(coords[0]) #* resampling_info['scale'][0]) out_coords.append(coords[1]) #* resampling_info['scale'][1]) out_coords.append(coords[2]) #* resampling_info['scale'][0]) out_coords.append(coords[3]) #* resampling_info['scale'][1]) if len(coords) > 4: out_coords.append(coords[4]) #* resampling_info['scale'][2] + resampling_info['z_crop']) out_coords.append(coords[5]) #* resampling_info['scale'][2] + resampling_info['z_crop']) predictions_df.loc[len(predictions_df)] = [pid, bix, out_coords, score, pred_class_id] try: fold = cf.fold except: fold = 'hold_out' predictions_df.to_csv(os.path.join(cf.exp_dir, 'results_{}.csv'.format(fold)), index=False) class _AnsiColorizer(object): """ A colorizer is an object that loosely wraps around a stream, allowing callers to write text to the stream in a particular color. Colorizer classes must implement C{supported()} and C{write(text, color)}. """ _colors = dict(black=30, red=31, green=32, yellow=33, blue=34, magenta=35, cyan=36, white=37, default=39) def __init__(self, stream): self.stream = stream @classmethod def supported(cls, stream=sys.stdout): """ A class method that returns True if the current platform supports coloring terminal output using this method. Returns False otherwise. """ if not stream.isatty(): return False # auto color only on TTYs try: import curses except ImportError: return False else: try: try: return curses.tigetnum("colors") > 2 except curses.error: curses.setupterm() return curses.tigetnum("colors") > 2 except: raise # guess false in case of error return False def write(self, text, color): """ Write the given text to the stream in the given color. @param text: Text to be written to the stream. @param color: A string label for a color. e.g. 'red', 'white'. """ color = self._colors[color] self.stream.write('\x1b[%sm%s\x1b[0m' % (color, text)) class ColorHandler(logging.StreamHandler): def __init__(self, stream=sys.stdout): super(ColorHandler, self).__init__(_AnsiColorizer(stream)) def emit(self, record): msg_colors = { logging.DEBUG: "green", logging.INFO: "default", logging.WARNING: "red", logging.ERROR: "red" } color = msg_colors.get(record.levelno, "blue") self.stream.write(record.msg + "\n", color)
44.903543
287
0.639691
from typing import Iterable, Tuple, Any, Union import os, sys import subprocess from multiprocessing import Process import importlib.util import pickle import logging from torch.utils.tensorboard import SummaryWriter from collections import OrderedDict import numpy as np import torch import pandas as pd def split_off_process(target, *args, daemon: bool=False, **kwargs): p = Process(target=target, args=tuple(args), kwargs=kwargs, daemon=daemon) p.start() return p def get_formatted_duration(seconds: float, format: str="hms") -> str: mins, secs = divmod(seconds, 60) if format == "ms": t = "{:d}m:{:02d}s".format(int(mins), int(secs)) elif format == "hms": h, mins = divmod(mins, 60) t = "{:d}h:{:02d}m:{:02d}s".format(int(h), int(mins), int(secs)) else: raise Exception("Format {} not available, only 'hms' or 'ms'".format(format)) return t class CombinedLogger(object): def __init__(self, name: str, log_dir: str, server_env: bool=True, fold: Union[int, str]="all"): self.pylogger = logging.getLogger(name) self.tboard = SummaryWriter(log_dir=os.path.join(log_dir, "tboard")) self.log_dir = log_dir self.fold = str(fold) self.server_env = server_env self.pylogger.setLevel(logging.DEBUG) self.log_file = os.path.join(log_dir, "fold_"+self.fold, 'exec.log') os.makedirs(os.path.dirname(self.log_file), exist_ok=True) self.pylogger.addHandler(logging.FileHandler(self.log_file)) if not server_env: self.pylogger.addHandler(ColorHandler()) else: self.pylogger.addHandler(logging.StreamHandler()) self.pylogger.propagate = False def __getattr__(self, attr): for obj in [self.pylogger, self.tboard]: if attr in dir(obj): return getattr(obj, attr) print("logger attr not found") def set_logfile(self, fold: Union[int, str, None]=None, log_file: Union[str, None]=None): if fold is not None: self.fold = str(fold) if log_file is None: self.log_file = os.path.join(self.log_dir, "fold_"+self.fold, 'exec.log') else: self.log_file = log_file os.makedirs(os.path.dirname(self.log_file), exist_ok=True) for hdlr in self.pylogger.handlers: hdlr.close() self.pylogger.handlers = [] self.pylogger.addHandler(logging.FileHandler(self.log_file)) if not self.server_env: self.pylogger.addHandler(ColorHandler()) else: self.pylogger.addHandler(logging.StreamHandler()) def metrics2tboard(self, metrics, global_step=None, suptitle=None): if global_step is None: global_step = len(metrics['train'][list(metrics['train'].keys())[0]]) - 1 if suptitle is not None: suptitle = str(suptitle) else: suptitle = "Fold_" + str(self.fold) for key in ['train', 'val']: loss_series = {} mon_met_series = {} for tag, val in metrics[key].items(): val = val[-1] if 'loss' in tag.lower() and not np.isnan(val): loss_series["{}".format(tag)] = val elif not np.isnan(val): mon_met_series["{}".format(tag)] = val self.tboard.add_scalars(suptitle + "/Losses/{}".format(key), loss_series, global_step) self.tboard.add_scalars(suptitle + "/Monitor_Metrics/{}".format(key), mon_met_series, global_step) self.tboard.add_scalars(suptitle + "/Learning_Rate", metrics["lr"], global_step) return def __del__(self): for hdlr in self.pylogger.handlers: hdlr.close() self.pylogger.handlers = [] del self.pylogger self.tboard.flush() def get_logger(exp_dir: str, server_env: bool=False) -> CombinedLogger: log_dir = os.path.join(exp_dir, "logs") logger = CombinedLogger('medicaldetectiontoolkit', log_dir, server_env=server_env) print("Logging to {}".format(logger.log_file)) return logger def prep_exp(dataset_path, exp_path, server_env, use_stored_settings=True, is_training=True): if is_training: if use_stored_settings: cf_file = import_module('cf_file', os.path.join(exp_path, 'configs.py')) cf = cf_file.configs(server_env) if not os.path.isfile(os.path.join(exp_path, 'mdt_model.py')) or \ not os.path.isfile(os.path.join(exp_path, 'backbone.py')): raise Exception( "Selected use_stored_settings option but no model and/or backbone source files exist in exp dir.") cf.model_path = os.path.join(exp_path, 'mdt_model.py') cf.backbone_path = os.path.join(exp_path, 'backbone.py') else: os.makedirs(exp_path, exist_ok=True) subprocess.call('cp {} {}'.format('default_configs.py', os.path.join(exp_path, 'default_configs.py')), shell=True) subprocess.call( 'cp {} {}'.format(os.path.join(dataset_path, 'configs.py'), os.path.join(exp_path, 'configs.py')), shell=True) cf_file = import_module('cf_file', os.path.join(dataset_path, 'configs.py')) cf = cf_file.configs(server_env) subprocess.call('cp {} {}'.format(cf.model_path, os.path.join(exp_path, 'mdt_model.py')), shell=True) subprocess.call('cp {} {}'.format(cf.backbone_path, os.path.join(exp_path, 'backbone.py')), shell=True) if os.path.isfile(os.path.join(exp_path, "fold_ids.pickle")): subprocess.call('rm {}'.format(os.path.join(exp_path, "fold_ids.pickle")), shell=True) else: cf_file = import_module('cf_file', os.path.join(exp_path, 'configs.py')) cf = cf_file.configs(server_env) cf.model_path = os.path.join(exp_path, 'mdt_model.py') cf.backbone_path = os.path.join(exp_path, 'backbone.py') cf.exp_dir = exp_path cf.test_dir = os.path.join(cf.exp_dir, 'test') cf.plot_dir = os.path.join(cf.exp_dir, 'plots') if not os.path.exists(cf.test_dir): os.mkdir(cf.test_dir) if not os.path.exists(cf.plot_dir): os.mkdir(cf.plot_dir) cf.experiment_name = exp_path.split("/")[-1] cf.created_fold_id_pickle = False return cf def import_module(name: str, path: str): spec = importlib.util.spec_from_file_location(name, path) module = importlib.util.module_from_spec(spec) spec.loader.exec_module(module) return module def set_params_flag(module: torch.nn.Module, flag: Tuple[str, Any], check_overwrite: bool = True) -> torch.nn.Module: for param in module.parameters(): if check_overwrite: assert not hasattr(param, flag[0]), \ "param {} already has attr {} (w/ val {})".format(param, flag[0], getattr(param, flag[0])) setattr(param, flag[0], flag[1]) return module def parse_params_for_optim(net: torch.nn.Module, weight_decay: float = 0., exclude_from_wd: Iterable = ("norm",)) -> list: if weight_decay is None: weight_decay = 0. norm_types = [torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d, torch.nn.InstanceNorm1d, torch.nn.InstanceNorm2d, torch.nn.InstanceNorm3d, torch.nn.LayerNorm, torch.nn.GroupNorm, torch.nn.SyncBatchNorm, torch.nn.LocalResponseNorm] level_map = {"bias": "weight", "norm": "module"} type_map = {"norm": norm_types} exclude_from_wd = [str(name).lower() for name in exclude_from_wd] exclude_weight_names = [k for k, v in level_map.items() if k in exclude_from_wd and v == "weight"] exclude_module_types = tuple([type_ for k, v in level_map.items() if (k in exclude_from_wd and v == "module") for type_ in type_map[k]]) if exclude_from_wd: print("excluding {} from weight decay.".format(exclude_from_wd)) for module in net.modules(): if isinstance(module, exclude_module_types): set_params_flag(module, ("no_wd", True)) for param_name, param in net.named_parameters(): if np.any([ename in param_name for ename in exclude_weight_names]): setattr(param, "no_wd", True) with_dec, no_dec = [], [] for param in net.parameters(): if hasattr(param, "no_wd") and param.no_wd == True: no_dec.append(param) else: with_dec.append(param) orig_ps = sum(p.numel() for p in net.parameters()) with_ps = sum(p.numel() for p in with_dec) wo_ps = sum(p.numel() for p in no_dec) assert orig_ps == with_ps + wo_ps, "orig n parameters {} unequals sum of with wd {} and w/o wd {}."\ .format(orig_ps, with_ps, wo_ps) groups = [{'params': gr, 'weight_decay': wd} for (gr, wd) in [(no_dec, 0.), (with_dec, weight_decay)] if len(gr)>0] return groups class ModelSelector: def __init__(self, cf, logger): self.cf = cf self.saved_epochs = [-1] * cf.save_n_models self.logger = logger def run_model_selection(self, net: torch.nn.Module, optimizer: torch.optim.Optimizer, monitor_metrics: dict, epoch: int): non_nan_scores = np.mean(np.array([[0 if (ii is None or np.isnan(ii)) else ii for ii in monitor_metrics['val'][sc]] for sc in self.cf.model_selection_criteria]), 0) epochs_scores = [ii for ii in non_nan_scores[1:]] epoch_ranking = np.argsort(epochs_scores, kind="stable")[::-1] + 1 epoch_ranking = epoch_ranking[epoch_ranking >= self.cf.min_save_thresh] if epoch in epoch_ranking[:self.cf.save_n_models]: save_dir = os.path.join(self.cf.fold_dir, '{}_best_checkpoint'.format(epoch)) if not os.path.exists(save_dir): os.mkdir(save_dir) torch.save(net.state_dict(), os.path.join(save_dir, 'params.pth')) with open(os.path.join(save_dir, 'monitor_metrics.pickle'), 'wb') as handle: pickle.dump(monitor_metrics, handle) np.save(os.path.join(self.cf.fold_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) np.save(os.path.join(save_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) self.logger.info( "saving current epoch {} at rank {}".format(epoch, np.argwhere(epoch_ranking == epoch))) for se in [int(ii.split('_')[0]) for ii in os.listdir(self.cf.fold_dir) if 'best_checkpoint' in ii]: if se in epoch_ranking[self.cf.save_n_models:]: subprocess.call('rm -rf {}'.format(os.path.join(self.cf.fold_dir, '{}_best_checkpoint'.format(se))), shell=True) self.logger.info('deleting epoch {} at rank {}'.format(se, np.argwhere(epoch_ranking == se))) state = { 'epoch': epoch, 'state_dict': net.state_dict(), 'optimizer': optimizer.state_dict(), } save_dir = os.path.join(self.cf.fold_dir, 'last_checkpoint'.format(epoch)) if not os.path.exists(save_dir): os.mkdir(save_dir) torch.save(state, os.path.join(save_dir, 'params.pth')) np.save(os.path.join(save_dir, 'epoch_ranking'), epoch_ranking[:self.cf.save_n_models]) with open(os.path.join(save_dir, 'monitor_metrics.pickle'), 'wb') as handle: pickle.dump(monitor_metrics, handle) return os.path.join(os.path.join(self.cf.fold_dir, f'{epoch_ranking[0]}_best_checkpoint')) def load_checkpoint(checkpoint_path: str, net: torch.nn.Module, optimizer: torch.optim.Optimizer) -> Tuple: checkpoint = torch.load(os.path.join(checkpoint_path, 'params.pth')) net.load_state_dict(checkpoint['state_dict']) optimizer.load_state_dict(checkpoint['optimizer']) with open(os.path.join(checkpoint_path, 'monitor_metrics.pickle'), 'rb') as handle: monitor_metrics = pickle.load(handle) starting_epoch = checkpoint['epoch'] + 1 return starting_epoch, net, optimizer, monitor_metrics def prepare_monitoring(cf): metrics = {} metrics['train'] = OrderedDict() metrics['val'] = OrderedDict() metric_classes = [] if 'rois' in cf.report_score_level: metric_classes.extend([v for k, v in cf.class_dict.items()]) if 'patient' in cf.report_score_level: metric_classes.extend(['patient']) for cl in metric_classes: metrics['train'][cl + '_ap'] = [np.nan] metrics['val'][cl + '_ap'] = [np.nan] if cl == 'patient': metrics['train'][cl + '_auc'] = [np.nan] metrics['val'][cl + '_auc'] = [np.nan] return metrics def create_csv_output(results_list, cf, logger): logger.info('creating csv output file at {}'.format(os.path.join(cf.test_dir, 'results.csv'))) predictions_df = pd.DataFrame(columns = ['patientID', 'predictionID', 'coords', 'score', 'pred_classID']) for r in results_list: pid = r[1] for bix, box in enumerate(r[0][0]): if box["box_type"] == "gt": continue assert box['box_type'] == 'det', box['box_type'] coords = box['box_coords'] score = box['box_score'] pred_class_id = box['box_pred_class_id'] out_coords = [] if score >= cf.min_det_thresh: out_coords.append(coords[0]) out_coords.append(coords[1]) out_coords.append(coords[2]) out_coords.append(coords[3]) if len(coords) > 4: out_coords.append(coords[4]) out_coords.append(coords[5]) predictions_df.loc[len(predictions_df)] = [pid, bix, out_coords, score, pred_class_id] try: fold = cf.fold except: fold = 'hold_out' predictions_df.to_csv(os.path.join(cf.exp_dir, 'results_{}.csv'.format(fold)), index=False) class _AnsiColorizer(object): _colors = dict(black=30, red=31, green=32, yellow=33, blue=34, magenta=35, cyan=36, white=37, default=39) def __init__(self, stream): self.stream = stream @classmethod def supported(cls, stream=sys.stdout): if not stream.isatty(): return False try: import curses except ImportError: return False else: try: try: return curses.tigetnum("colors") > 2 except curses.error: curses.setupterm() return curses.tigetnum("colors") > 2 except: raise return False def write(self, text, color): color = self._colors[color] self.stream.write('\x1b[%sm%s\x1b[0m' % (color, text)) class ColorHandler(logging.StreamHandler): def __init__(self, stream=sys.stdout): super(ColorHandler, self).__init__(_AnsiColorizer(stream)) def emit(self, record): msg_colors = { logging.DEBUG: "green", logging.INFO: "default", logging.WARNING: "red", logging.ERROR: "red" } color = msg_colors.get(record.levelno, "blue") self.stream.write(record.msg + "\n", color)
true
true
f717c2aceaf306d0f7eae56f9df7f70d7fb7e56b
2,488
py
Python
usort/util.py
thatch/usort
2ca1ff63d6cfc79e76ea95b69d162f4579c3fa3c
[ "MIT" ]
null
null
null
usort/util.py
thatch/usort
2ca1ff63d6cfc79e76ea95b69d162f4579c3fa3c
[ "MIT" ]
null
null
null
usort/util.py
thatch/usort
2ca1ff63d6cfc79e76ea95b69d162f4579c3fa3c
[ "MIT" ]
null
null
null
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from contextlib import contextmanager from contextvars import ContextVar from pathlib import Path from time import monotonic from typing import Callable, Generator, List, Optional, Tuple import libcst as cst TIMINGS: ContextVar[List[Tuple[str, float]]] = ContextVar("TIMINGS") @contextmanager def timed(msg: str) -> Generator[None, None, None]: """ Records the monotonic duration of the contained context, with a given description. Timings are stored for later use/printing with `print_timings()`. """ before = monotonic() yield after = monotonic() try: TIMINGS.get().append((msg, after - before)) except LookupError: pass @contextmanager def save_timings(to: List[Tuple[str, float]]) -> Generator[None, None, None]: token = TIMINGS.set([]) yield to.extend(TIMINGS.get()) TIMINGS.reset(token) def merge_timings(more: List[Tuple[str, float]]) -> None: TIMINGS.get().extend(more) def print_timings(fn: Callable[[str], None] = print) -> None: """ Print all stored timing values in microseconds. """ for msg, duration in TIMINGS.get(): fn(f"{msg + ':':50} {int(duration*1000000):7} µs") def try_parse(path: Path, data: Optional[bytes] = None) -> cst.Module: """ Attempts to parse the file with all syntax versions known by LibCST. If parsing fails on all supported grammar versions, then raises the parser error from the first/newest version attempted. """ if data is None: data = path.read_bytes() with timed(f"parsing {path}"): parse_error: Optional[cst.ParserSyntaxError] = None for version in cst.KNOWN_PYTHON_VERSION_STRINGS[::-1]: try: mod = cst.parse_module( data, cst.PartialParserConfig(python_version=version) ) return mod except cst.ParserSyntaxError as e: # keep the first error we see in case parsing fails on all versions if parse_error is None: parse_error = e # not caring about existing traceback here because it's not useful for parse # errors, and usort_path is already going to wrap it in a custom class raise parse_error or Exception("unknown parse failure")
30.716049
86
0.658762
from contextlib import contextmanager from contextvars import ContextVar from pathlib import Path from time import monotonic from typing import Callable, Generator, List, Optional, Tuple import libcst as cst TIMINGS: ContextVar[List[Tuple[str, float]]] = ContextVar("TIMINGS") @contextmanager def timed(msg: str) -> Generator[None, None, None]: before = monotonic() yield after = monotonic() try: TIMINGS.get().append((msg, after - before)) except LookupError: pass @contextmanager def save_timings(to: List[Tuple[str, float]]) -> Generator[None, None, None]: token = TIMINGS.set([]) yield to.extend(TIMINGS.get()) TIMINGS.reset(token) def merge_timings(more: List[Tuple[str, float]]) -> None: TIMINGS.get().extend(more) def print_timings(fn: Callable[[str], None] = print) -> None: for msg, duration in TIMINGS.get(): fn(f"{msg + ':':50} {int(duration*1000000):7} µs") def try_parse(path: Path, data: Optional[bytes] = None) -> cst.Module: if data is None: data = path.read_bytes() with timed(f"parsing {path}"): parse_error: Optional[cst.ParserSyntaxError] = None for version in cst.KNOWN_PYTHON_VERSION_STRINGS[::-1]: try: mod = cst.parse_module( data, cst.PartialParserConfig(python_version=version) ) return mod except cst.ParserSyntaxError as e: if parse_error is None: parse_error = e # errors, and usort_path is already going to wrap it in a custom class raise parse_error or Exception("unknown parse failure")
true
true
f717c2b8fff53fa6c6659e99b4b2a067e7f9823a
2,440
py
Python
model/experiment/gaussian_naive_bayes_model.py
theBraindonor/chicago-crime-arrests
64cdb82fbe828d1316cf945b67ddc205ef190293
[ "CC-BY-4.0" ]
1
2019-04-29T10:32:26.000Z
2019-04-29T10:32:26.000Z
model/experiment/gaussian_naive_bayes_model.py
theBraindonor/chicago-crime-arrests
64cdb82fbe828d1316cf945b67ddc205ef190293
[ "CC-BY-4.0" ]
4
2020-03-24T16:54:17.000Z
2021-06-01T23:33:48.000Z
model/experiment/gaussian_naive_bayes_model.py
theBraindonor/chicago-crime-arrests
64cdb82fbe828d1316cf945b67ddc205ef190293
[ "CC-BY-4.0" ]
null
null
null
#!/usr/bin/env python # -*- coding: utf-8 -*- """ Experiment with a gaussian naive bayes model with a variety of balancing techniques on the cleaned data set """ __author__ = "John Hoff" __email__ = "john.hoff@braindonor.net" __copyright__ = "Copyright 2019, John Hoff" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License" __version__ = "1.0" from imblearn.combine import SMOTEENN from imblearn.over_sampling import SMOTE from imblearn.under_sampling import RandomUnderSampler from sklearn.naive_bayes import GaussianNB from utility import Runner from model import load_clean_sample_data_frame, binned_geo_one_hot_data_mapper sample = None fit_increment = 10000 def test_gaussian_naive_bayes(): runner = Runner( 'model/experiment/output/gaussian_naive_bayes_basic', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1 ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_under_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=RandomUnderSampler() ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_over_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=SMOTE() ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_combine_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=SMOTEENN() ) if __name__ == '__main__': test_gaussian_naive_bayes()
26.813187
111
0.697951
__author__ = "John Hoff" __email__ = "john.hoff@braindonor.net" __copyright__ = "Copyright 2019, John Hoff" __license__ = "Creative Commons Attribution-ShareAlike 4.0 International License" __version__ = "1.0" from imblearn.combine import SMOTEENN from imblearn.over_sampling import SMOTE from imblearn.under_sampling import RandomUnderSampler from sklearn.naive_bayes import GaussianNB from utility import Runner from model import load_clean_sample_data_frame, binned_geo_one_hot_data_mapper sample = None fit_increment = 10000 def test_gaussian_naive_bayes(): runner = Runner( 'model/experiment/output/gaussian_naive_bayes_basic', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1 ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_under_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=RandomUnderSampler() ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_over_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=SMOTE() ) runner = Runner( 'model/experiment/output/gaussian_naive_bayes_combine_sampled', load_clean_sample_data_frame(), 'arrest', GaussianNB() ) runner.run_classification_experiment( sample=sample, record_predict_proba=True, transformer=binned_geo_one_hot_data_mapper, fit_increment=fit_increment, n_jobs=1, sampling=SMOTEENN() ) if __name__ == '__main__': test_gaussian_naive_bayes()
true
true
f717c597ce99c986760a12c223c014567cf34f38
96,886
py
Python
ThirdParty/ZopeInterface/zope/interface/tests/test_registry.py
OpenGeoscience/VTK
a373e975b9284a022f43a062ebf5042bb17b4e44
[ "BSD-3-Clause" ]
3
2016-02-01T02:29:51.000Z
2020-09-04T17:19:24.000Z
ThirdParty/ZopeInterface/zope/interface/tests/test_registry.py
OpenGeoscience/VTK
a373e975b9284a022f43a062ebf5042bb17b4e44
[ "BSD-3-Clause" ]
7
2021-02-08T20:22:15.000Z
2022-03-11T23:19:41.000Z
ThirdParty/ZopeInterface/zope/interface/tests/test_registry.py
OpenGeoscience/VTK
a373e975b9284a022f43a062ebf5042bb17b4e44
[ "BSD-3-Clause" ]
6
2017-02-13T09:11:02.000Z
2021-06-29T11:22:18.000Z
############################################################################## # # Copyright (c) 2001, 2002, 2009 Zope Foundation and Contributors. # All Rights Reserved. # # This software is subject to the provisions of the Zope Public License, # Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution. # THIS SOFTWARE IS PROVIDED "AS IS" AND ANY AND ALL EXPRESS OR IMPLIED # WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS # FOR A PARTICULAR PURPOSE. # ############################################################################## """Component Registry Tests""" import unittest class _SilencePy3Deprecations(unittest.TestCase): # silence deprecation warnings under py3 def failUnless(self, expr): # St00pid speling. return self.assertTrue(expr) def failIf(self, expr): # St00pid speling. return self.assertFalse(expr) class ComponentsTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import Components return Components def _makeOne(self, name='test', *args, **kw): return self._getTargetClass()(name, *args, **kw) def _wrapEvents(self): from zope.interface import registry _events = [] def _notify(*args, **kw): _events.append((args, kw)) _monkey = _Monkey(registry, notify=_notify) return _monkey, _events def test_ctor_no_bases(self): from zope.interface.adapter import AdapterRegistry comp = self._makeOne('testing') self.assertEqual(comp.__name__, 'testing') self.assertEqual(comp.__bases__, ()) self.failUnless(isinstance(comp.adapters, AdapterRegistry)) self.failUnless(isinstance(comp.utilities, AdapterRegistry)) self.assertEqual(comp.adapters.__bases__, ()) self.assertEqual(comp.utilities.__bases__, ()) self.assertEqual(comp._utility_registrations, {}) self.assertEqual(comp._adapter_registrations, {}) self.assertEqual(comp._subscription_registrations, []) self.assertEqual(comp._handler_registrations, []) def test_ctor_w_base(self): base = self._makeOne('base') comp = self._makeOne('testing', (base,)) self.assertEqual(comp.__name__, 'testing') self.assertEqual(comp.__bases__, (base,)) self.assertEqual(comp.adapters.__bases__, (base.adapters,)) self.assertEqual(comp.utilities.__bases__, (base.utilities,)) def test___repr__(self): comp = self._makeOne('testing') self.assertEqual(repr(comp), '<Components testing>') # test _init_registries / _init_registrations via only caller, __init__. def test_assign_to___bases__(self): base1 = self._makeOne('base1') base2 = self._makeOne('base2') comp = self._makeOne() comp.__bases__ = (base1, base2) self.assertEqual(comp.__bases__, (base1, base2)) self.assertEqual(comp.adapters.__bases__, (base1.adapters, base2.adapters)) self.assertEqual(comp.utilities.__bases__, (base1.utilities, base2.utilities)) def test_registerUtility_both_factory_and_component(self): def _factory(): pass _to_reg = object() comp = self._makeOne() self.assertRaises(TypeError, comp.registerUtility, component=_to_reg, factory=_factory) def test_registerUtility_w_component(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name, _info) self.failUnless(comp.utilities._adapters[0][ifoo][_name] is _to_reg) self.assertEqual(comp._utility_registrations[ifoo, _name], (_to_reg, _info, None)) self.assertEqual(comp.utilities._subscribers[0][ifoo][''], (_to_reg,)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_registerUtility_w_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(None, ifoo, _name, _info, factory=_factory) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerUtility_no_provided_available(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() comp = self._makeOne() self.assertRaises(TypeError, comp.registerUtility, _to_reg, None, _name, _info) def test_registerUtility_wo_provided(self): from zope.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, None, _name, _info) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_registerUtility_duplicates_existing_reg(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name, _info) self.assertEqual(len(_events), 0) def test_registerUtility_replaces_existing_reg(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _before, _after = object(), object() comp = self._makeOne() comp.registerUtility(_before, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_after, ifoo, _name, _info) self.assertEqual(len(_events), 2) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _before) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) args, kw = _events[1] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _after) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_registerUtility_w_existing_subscr(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name2, _info) self.assertEqual(comp.utilities._subscribers[0][ifoo][''], (_to_reg,)) def test_registerUtility_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name, _info, False) self.assertEqual(len(_events), 0) def test_unregisterUtility_neither_factory_nor_component_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterUtility, component=None, provided=None, factory=None) def test_unregisterUtility_both_factory_and_component(self): def _factory(): pass _to_reg = object() comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterUtility, component=_to_reg, factory=_factory) def test_unregisterUtility_w_component_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name = _u('name') _to_reg = object() comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, ifoo, _name) self.failIf(unreg) self.failIf(_events) def test_unregisterUtility_w_component(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name = _u('name') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, ifoo, _name) self.failUnless(unreg) self.failIf(comp.utilities._adapters) # all erased self.failIf((ifoo, _name) in comp._utility_registrations) self.failIf(comp.utilities._subscribers) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.factory is None) def test_unregisterUtility_w_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(): return _to_reg comp = self._makeOne() comp.registerUtility(None, ifoo, _name, _info, factory=_factory) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(None, ifoo, _name, factory=_factory) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.factory is _factory) def test_unregisterUtility_wo_explicit_provided(self): from zope.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, None, _name) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_unregisterUtility_wo_component_or_factory(self): from zope.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: # Just pass the interface / name unreg = comp.unregisterUtility(provided=ifoo, name=_name) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_unregisterUtility_w_existing_subscr(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) comp.registerUtility(_to_reg, ifoo, _name2, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.unregisterUtility(_to_reg, ifoo, _name2) self.assertEqual(comp.utilities._subscribers[0][ifoo][''], (_to_reg,)) def test_registeredUtilities_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredUtilities()), []) def test_registeredUtilities_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import UtilityRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) comp.registerUtility(_to_reg, ifoo, _name2, _info) reg = sorted(comp.registeredUtilities(), key=lambda r: r.name) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], UtilityRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.failUnless(reg[0].name is _name1) self.failUnless(reg[0].component is _to_reg) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is None) self.failUnless(isinstance(reg[1], UtilityRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.failUnless(reg[1].name is _name2) self.failUnless(reg[1].component is _to_reg) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is None) def test_queryUtility_miss_no_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.failUnless(comp.queryUtility(ifoo) is None) def test_queryUtility_miss_w_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _default = object() self.failUnless(comp.queryUtility(ifoo, default=_default) is _default) def test_queryUtility_hit(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo) self.failUnless(comp.queryUtility(ifoo) is _to_reg) def test_getUtility_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getUtility, ifoo) def test_getUtility_hit(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo) self.failUnless(comp.getUtility(ifoo) is _to_reg) def test_getUtilitiesFor_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertEqual(list(comp.getUtilitiesFor(ifoo)), []) def test_getUtilitiesFor_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, name=_name1) comp.registerUtility(_to_reg, ifoo, name=_name2) self.assertEqual(sorted(comp.getUtilitiesFor(ifoo)), [(_name1, _to_reg), (_name2, _to_reg)]) def test_getAllUtilitiesRegisteredFor_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertEqual(list(comp.getAllUtilitiesRegisteredFor(ifoo)), []) def test_getAllUtilitiesRegisteredFor_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, name=_name1) comp.registerUtility(_to_reg, ifoo, name=_name2) self.assertEqual(list(comp.getAllUtilitiesRegisteredFor(ifoo)), [_to_reg]) def test_registerAdapter_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_factory, (ibar,), ifoo, _name, _info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerAdapter_no_provided_available(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, (ibar,), name=_name, info=_info) def test_registerAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, (ibar,), name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_no_required_available(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, provided=ifoo, name=_name, info=_info) def test_registerAdapter_w_invalid_required(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, ibar, provided=ifoo, name=_name, info=_info) def test_registerAdapter_w_required_containing_None(self): from zope.interface.declarations import InterfaceClass from zope.interface.interface import Interface from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, [None], provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][Interface][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(Interface,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (Interface,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_w_required_containing_class(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.declarations import implementedBy from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context @implementer(ibar) class _Context(object): pass _ctx_impl = implementedBy(_Context) comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, [_Context], provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][_ctx_impl][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(_ctx_impl,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (_ctx_impl,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_w_required_containing_junk(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, [object()], provided=ifoo, name=_name, info=_info) def test_registerAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_factory, (ibar,), ifoo, _name, _info, event=False) self.assertEqual(len(_events), 0) def test_unregisterAdapter_neither_factory_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterAdapter, factory=None, provided=None) def test_unregisterAdapter_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterAdapter, factory=None, provided=ifoo, required=None) def test_unregisterAdapter_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,), ifoo) self.failIf(unreg) def test_unregisterAdapter_hit_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._adapters) self.failIf(comp._adapter_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, provided=ifoo) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_registeredAdapters_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredAdapters()), []) def test_registeredAdapters_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') class _Factory(object): def __init__(self, context): pass comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo, _name1, _info) comp.registerAdapter(_Factory, (ibar,), ifoo, _name2, _info) reg = sorted(comp.registeredAdapters(), key=lambda r: r.name) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], AdapterRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.assertEqual(reg[0].required, (ibar,)) self.failUnless(reg[0].name is _name1) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is _Factory) self.failUnless(isinstance(reg[1], AdapterRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.assertEqual(reg[1].required, (ibar,)) self.failUnless(reg[1].name is _name2) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is _Factory) def test_queryAdapter_miss_no_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _context = object() self.failUnless(comp.queryAdapter(_context, ifoo) is None) def test_queryAdapter_miss_w_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _context = object() _default = object() self.failUnless( comp.queryAdapter(_context, ifoo, default=_default) is _default) def test_queryAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self.context = context @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) adapter = comp.queryAdapter(_context, ifoo) self.failUnless(isinstance(adapter, _Factory)) self.failUnless(adapter.context is _context) def test_getAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getAdapter, _context, ifoo) def test_getAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self.context = context @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) adapter = comp.getAdapter(_context, ifoo) self.failUnless(isinstance(adapter, _Factory)) self.failUnless(adapter.context is _context) def test_queryMultiAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertEqual(comp.queryMultiAdapter((_context1, _context2), ifoo), None) def test_queryMultiAdapter_miss_w_default(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() _default = object() comp = self._makeOne() self.failUnless( comp.queryMultiAdapter((_context1, _context2), ifoo, default=_default) is _default) def test_queryMultiAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory(object): def __init__(self, context1, context2): self.context = context1, context2 comp = self._makeOne() comp.registerAdapter(_Factory, (ibar, ibaz), ifoo) adapter = comp.queryMultiAdapter((_context1, _context2), ifoo) self.failUnless(isinstance(adapter, _Factory)) self.assertEqual(adapter.context, (_context1, _context2)) def test_getMultiAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getMultiAdapter, (_context1, _context2), ifoo) def test_getMultiAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory(object): def __init__(self, context1, context2): self.context = context1, context2 comp = self._makeOne() comp.registerAdapter(_Factory, (ibar, ibaz), ifoo) adapter = comp.getMultiAdapter((_context1, _context2), ifoo) self.failUnless(isinstance(adapter, _Factory)) self.assertEqual(adapter.context, (_context1, _context2)) def test_getAdapters_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertEqual( list(comp.getAdapters((_context1, _context2), ifoo)), []) def test_getAdapters_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory1(object): def __init__(self, context1, context2): self.context = context1, context2 class _Factory2(object): def __init__(self, context1, context2): self.context = context1, context2 _name1 = _u('name1') _name2 = _u('name2') comp = self._makeOne() comp.registerAdapter(_Factory1, (ibar, ibaz), ifoo, name=_name1) comp.registerAdapter(_Factory2, (ibar, ibaz), ifoo, name=_name2) found = sorted(comp.getAdapters((_context1, _context2), ifoo)) self.assertEqual(len(found), 2) self.assertEqual(found[0][0], _name1) self.failUnless(isinstance(found[0][1], _Factory1)) self.assertEqual(found[1][0], _name2) self.failUnless(isinstance(found[1][1], _Factory2)) def test_registerSubscriptionAdapter_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _name = _u('name') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() self.assertRaises(TypeError, comp.registerSubscriptionAdapter, _factory, (ibar,), ifoo, _name, _info) def test_registerSubscriptionAdapter_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_factory, (ibar,), ifoo, info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerSubscriptionAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _blank = _u('') _to_reg = object() @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_Factory, (ibar,), info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerSubscriptionAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _blank = _u('') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter( _Factory, provided=ifoo, info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerSubscriptionAdapter_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_factory, (ibar,), ifoo, info=_info, event=False) self.assertEqual(len(_events), 0) def test_registeredSubscriptionAdapters_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredSubscriptionAdapters()), []) def test_registeredSubscriptionAdapters_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IFoo') _info = _u('info') _blank = _u('') class _Factory(object): def __init__(self, context): pass comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo, info=_info) comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo, info=_info) reg = list(comp.registeredSubscriptionAdapters()) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], SubscriptionRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.assertEqual(reg[0].required, (ibar,)) self.assertEqual(reg[0].name, _blank) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is _Factory) self.failUnless(isinstance(reg[1], SubscriptionRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.assertEqual(reg[1].required, (ibar,)) self.assertEqual(reg[1].name, _blank) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is _Factory) def test_unregisterSubscriptionAdapter_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _nonblank = _u('nonblank') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, required=ifoo, provided=ibar, name=_nonblank) def test_unregisterSubscriptionAdapter_neither_factory_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, factory=None, provided=None) def test_unregisterSubscriptionAdapter_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, factory=None, provided=ifoo, required=None) def test_unregisterSubscriptionAdapter_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,), ifoo) self.failIf(unreg) self.failIf(_events) def test_unregisterSubscriptionAdapter_hit_wo_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(None, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._subscribers) self.failIf(comp._subscription_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is None) def test_unregisterSubscriptionAdapter_hit_w_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._subscribers) self.failIf(comp._subscription_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterSubscriptionAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterSubscriptionAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, provided=ifoo) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_subscribers_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') comp = self._makeOne() @implementer(ibar) class Bar(object): pass bar = Bar() self.assertEqual(list(comp.subscribers((bar,), ifoo)), []) def test_subscribers_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context class _Derived(_Factory): pass comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) comp.registerSubscriptionAdapter(_Derived, (ibar,), ifoo) @implementer(ibar) class Bar(object): pass bar = Bar() subscribers = comp.subscribers((bar,), ifoo) def _klassname(x): return x.__class__.__name__ subscribers = sorted(subscribers, key=_klassname) self.assertEqual(len(subscribers), 2) self.failUnless(isinstance(subscribers[0], _Derived)) self.failUnless(isinstance(subscribers[1], _Factory)) def test_registerHandler_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _nonblank = _u('nonblank') comp = self._makeOne() def _factory(context): pass self.assertRaises(TypeError, comp.registerHandler, _factory, required=ifoo, name=_nonblank) def test_registerHandler_w_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import HandlerRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerHandler(_factory, (ifoo,), info=_info) reg = comp.adapters._subscribers[1][ifoo][None][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _factory) self.assertEqual(comp._handler_registrations, [((ifoo,), _blank, _factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerHandler_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import HandlerRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _blank = _u('') class _Factory(object): __component_adapts__ = (ifoo,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerHandler(_Factory, info=_info) reg = comp.adapters._subscribers[1][ifoo][None][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._handler_registrations, [((ifoo,), _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registeredHandlers_empty(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass comp = self._makeOne() self.failIf(list(comp.registeredHandlers())) def test_registeredHandlers_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') def _factory1(context): pass def _factory2(context): pass comp = self._makeOne() comp.registerHandler(_factory1, (ifoo,)) comp.registerHandler(_factory2, (ifoo,)) def _factory_name(x): return x.factory.__code__.co_name subscribers = sorted(comp.registeredHandlers(), key=_factory_name) self.assertEqual(len(subscribers), 2) self.failUnless(isinstance(subscribers[0], HandlerRegistration)) self.assertEqual(subscribers[0].required, (ifoo,)) self.assertEqual(subscribers[0].name, '') self.assertEqual(subscribers[0].factory, _factory1) self.assertEqual(subscribers[0].info, '') self.failUnless(isinstance(subscribers[1], HandlerRegistration)) self.assertEqual(subscribers[1].required, (ifoo,)) self.assertEqual(subscribers[1].name, '') self.assertEqual(subscribers[1].factory, _factory2) self.assertEqual(subscribers[1].info, '') def test_unregisterHandler_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _nonblank = _u('nonblank') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterHandler, required=(ifoo,), name=_nonblank) def test_unregisterHandler_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterHandler) def test_unregisterHandler_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() unreg = comp.unregisterHandler(required=(ifoo,)) self.failIf(unreg) def test_unregisterHandler_hit_w_factory_and_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() comp.registerHandler(_factory, (ifoo,)) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(_factory, (ifoo,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.failUnless(event.object.factory is _factory) def test_unregisterHandler_hit_w_only_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() comp.registerHandler(_factory, (ifoo,)) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(required=(ifoo,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.failUnless(event.object.factory is None) def test_unregisterHandler_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Factory(object): __component_adapts__ = (ifoo,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerHandler(_Factory) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(_Factory) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_handle_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() @implementer(ifoo) class Bar(object): pass bar = Bar() comp.handle((bar,)) # doesn't raise def test_handle_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _called_1 = [] def _factory_1(context): _called_1.append(context) _called_2 = [] def _factory_2(context): _called_2.append(context) comp = self._makeOne() comp.registerHandler(_factory_1, (ifoo,)) comp.registerHandler(_factory_2, (ifoo,)) @implementer(ifoo) class Bar(object): pass bar = Bar() comp.handle(bar) self.assertEqual(_called_1, [bar]) self.assertEqual(_called_2, [bar]) # Test _getUtilityProvided, _getAdapterProvided, _getAdapterRequired via their # callers (Component.registerUtility, Component.registerAdapter). class UtilityRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import UtilityRegistration return UtilityRegistration def _makeOne(self, component=None, factory=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, ifoo, name, component, doc, factory), registry, name, ) def test_class_conforms_to_IUtilityRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IUtilityRegistration verifyClass(IUtilityRegistration, self._getTargetClass()) def test_instance_conforms_to_IUtilityRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IUtilityRegistration ur, _, _ = self._makeOne() verifyObject(IUtilityRegistration, ur) def test___repr__(self): class _Component(object): __name__ = 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, IFoo, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___repr___provided_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) ur.provided = object() self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, None, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___repr___component_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) ur.provided = object() self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, None, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___hash__(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.assertEqual(ur.__hash__(), id(ur)) def test___eq___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur == ur) def test___eq___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failUnless(ur == ur2) def test___eq___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) self.failIf(ur == ur2) def test___ne___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur != ur) def test___ne___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur != ur2) def test___ne___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) self.failUnless(ur != ur2) def test___lt___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur < ur) def test___lt___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur < ur2) def test___lt___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur < ur2) def test___le___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur <= ur) def test___le___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failUnless(ur <= ur2) def test___le___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur <= ur2) def test___gt___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur > ur) def test___gt___hit(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur2 > ur) def test___gt___miss(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur2 > ur) def test___ge___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur >= ur) def test___ge___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failIf(ur >= ur2) def test___ge___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) ur2.name = _name + '2' self.failUnless(ur2 >= ur) class AdapterRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import AdapterRegistration return AdapterRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ibar,), ifoo, name, component, doc), registry, name, ) def test_class_conforms_to_IAdapterRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IAdapterRegistration verifyClass(IAdapterRegistration, self._getTargetClass()) def test_instance_conforms_to_IAdapterRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IAdapterRegistration ar, _, _ = self._makeOne() verifyObject(IAdapterRegistration, ar) def test___repr__(self): class _Component(object): __name__ = 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], IFoo, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___provided_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) ar.provided = object() self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], None, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___component_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) ar.provided = object() self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], None, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___hash__(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.assertEqual(ar.__hash__(), id(ar)) def test___eq___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar == ar) def test___eq___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failUnless(ar == ar2) def test___eq___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) self.failIf(ar == ar2) def test___ne___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar != ar) def test___ne___miss(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar != ar2) def test___ne___hit_component(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) self.failUnless(ar != ar2) def test___ne___hit_provided(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ibaz = IFoo('IBaz') _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) ar2.provided = ibaz self.failUnless(ar != ar2) def test___ne___hit_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ibaz = IFoo('IBaz') _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.required = (ibaz,) self.failUnless(ar != ar2) def test___lt___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar < ar) def test___lt___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar < ar2) def test___lt___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar < ar2) def test___le___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar <= ar) def test___le___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failUnless(ar <= ar2) def test___le___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar <= ar2) def test___gt___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar > ar) def test___gt___hit(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar2 > ar) def test___gt___miss(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar2 > ar) def test___ge___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar >= ar) def test___ge___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failIf(ar >= ar2) def test___ge___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) ar2.name = _name + '2' self.failUnless(ar2 >= ar) class SubscriptionRegistrationTests(unittest.TestCase): def _getTargetClass(self): from zope.interface.registry import SubscriptionRegistration return SubscriptionRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ibar,), ifoo, name, component, doc), registry, name, ) def test_class_conforms_to_ISubscriptionAdapterRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import ISubscriptionAdapterRegistration verifyClass(ISubscriptionAdapterRegistration, self._getTargetClass()) def test_instance_conforms_to_ISubscriptionAdapterRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import ISubscriptionAdapterRegistration sar, _, _ = self._makeOne() verifyObject(ISubscriptionAdapterRegistration, sar) class HandlerRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import HandlerRegistration return HandlerRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ifoo,), name, component, doc), registry, name, ) def test_class_conforms_to_IHandlerRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IHandlerRegistration verifyClass(IHandlerRegistration, self._getTargetClass()) def test_instance_conforms_to_IHandlerRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IHandlerRegistration hr, _, _ = self._makeOne() verifyObject(IHandlerRegistration, hr) def test_properties(self): def _factory(context): pass hr, _, _ = self._makeOne(_factory) self.failUnless(hr.handler is _factory) self.failUnless(hr.factory is hr.handler) self.failUnless(hr.provided is None) def test___repr___factory_w_name(self): class _Factory(object): __name__ = 'TEST' hr, _registry, _name = self._makeOne(_Factory()) self.assertEqual(repr(hr), ("HandlerRegistration(_REGISTRY, [IFoo], %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___factory_wo_name(self): class _Factory(object): def __repr__(self): return 'TEST' hr, _registry, _name = self._makeOne(_Factory()) self.assertEqual(repr(hr), ("HandlerRegistration(_REGISTRY, [IFoo], %r, TEST, " + "'DOCSTRING')") % (_name)) class _Monkey(object): # context-manager for replacing module names in the scope of a test. def __init__(self, module, **kw): self.module = module self.to_restore = dict([(key, getattr(module, key)) for key in kw]) for key, value in kw.items(): setattr(module, key, value) def __enter__(self): return self def __exit__(self, exc_type, exc_val, exc_tb): for key, value in self.to_restore.items(): setattr(self.module, key, value) def test_suite(): return unittest.TestSuite(( unittest.makeSuite(ComponentsTests), unittest.makeSuite(UtilityRegistrationTests), unittest.makeSuite(AdapterRegistrationTests), unittest.makeSuite(SubscriptionRegistrationTests), unittest.makeSuite(AdapterRegistrationTests), ))
39.464766
80
0.630783
e.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, None, _name, _info) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_registerUtility_duplicates_existing_reg(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name, _info) self.assertEqual(len(_events), 0) def test_registerUtility_replaces_existing_reg(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.interfaces import Registered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _before, _after = object(), object() comp = self._makeOne() comp.registerUtility(_before, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_after, ifoo, _name, _info) self.assertEqual(len(_events), 2) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _before) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) args, kw = _events[1] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _after) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_registerUtility_w_existing_subscr(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name2, _info) self.assertEqual(comp.utilities._subscribers[0][ifoo][''], (_to_reg,)) def test_registerUtility_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerUtility(_to_reg, ifoo, _name, _info, False) self.assertEqual(len(_events), 0) def test_unregisterUtility_neither_factory_nor_component_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterUtility, component=None, provided=None, factory=None) def test_unregisterUtility_both_factory_and_component(self): def _factory(): pass _to_reg = object() comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterUtility, component=_to_reg, factory=_factory) def test_unregisterUtility_w_component_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name = _u('name') _to_reg = object() comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, ifoo, _name) self.failIf(unreg) self.failIf(_events) def test_unregisterUtility_w_component(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name = _u('name') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, ifoo, _name) self.failUnless(unreg) self.failIf(comp.utilities._adapters) self.failIf((ifoo, _name) in comp._utility_registrations) self.failIf(comp.utilities._subscribers) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.factory is None) def test_unregisterUtility_w_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(): return _to_reg comp = self._makeOne() comp.registerUtility(None, ifoo, _name, _info, factory=_factory) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(None, ifoo, _name, factory=_factory) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.factory is _factory) def test_unregisterUtility_wo_explicit_provided(self): from zope.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(_to_reg, None, _name) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_unregisterUtility_wo_component_or_factory(self): from zope.interface.declarations import directlyProvides from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import UtilityRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass class Foo(object): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') _to_reg = Foo() directlyProvides(_to_reg, ifoo) comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name, _info) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterUtility(provided=ifoo, name=_name) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, UtilityRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.failUnless(event.object.name is _name) self.failUnless(event.object.component is _to_reg) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is None) def test_unregisterUtility_w_existing_subscr(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) comp.registerUtility(_to_reg, ifoo, _name2, _info) _monkey, _events = self._wrapEvents() with _monkey: comp.unregisterUtility(_to_reg, ifoo, _name2) self.assertEqual(comp.utilities._subscribers[0][ifoo][''], (_to_reg,)) def test_registeredUtilities_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredUtilities()), []) def test_registeredUtilities_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import UtilityRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, _name1, _info) comp.registerUtility(_to_reg, ifoo, _name2, _info) reg = sorted(comp.registeredUtilities(), key=lambda r: r.name) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], UtilityRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.failUnless(reg[0].name is _name1) self.failUnless(reg[0].component is _to_reg) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is None) self.failUnless(isinstance(reg[1], UtilityRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.failUnless(reg[1].name is _name2) self.failUnless(reg[1].component is _to_reg) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is None) def test_queryUtility_miss_no_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.failUnless(comp.queryUtility(ifoo) is None) def test_queryUtility_miss_w_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _default = object() self.failUnless(comp.queryUtility(ifoo, default=_default) is _default) def test_queryUtility_hit(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo) self.failUnless(comp.queryUtility(ifoo) is _to_reg) def test_getUtility_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getUtility, ifoo) def test_getUtility_hit(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo) self.failUnless(comp.getUtility(ifoo) is _to_reg) def test_getUtilitiesFor_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertEqual(list(comp.getUtilitiesFor(ifoo)), []) def test_getUtilitiesFor_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, name=_name1) comp.registerUtility(_to_reg, ifoo, name=_name2) self.assertEqual(sorted(comp.getUtilitiesFor(ifoo)), [(_name1, _to_reg), (_name2, _to_reg)]) def test_getAllUtilitiesRegisteredFor_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertEqual(list(comp.getAllUtilitiesRegisteredFor(ifoo)), []) def test_getAllUtilitiesRegisteredFor_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _name1 = _u('name1') _name2 = _u('name2') _to_reg = object() comp = self._makeOne() comp.registerUtility(_to_reg, ifoo, name=_name1) comp.registerUtility(_to_reg, ifoo, name=_name2) self.assertEqual(list(comp.getAllUtilitiesRegisteredFor(ifoo)), [_to_reg]) def test_registerAdapter_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_factory, (ibar,), ifoo, _name, _info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerAdapter_no_provided_available(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, (ibar,), name=_name, info=_info) def test_registerAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, (ibar,), name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_no_required_available(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, provided=ifoo, name=_name, info=_info) def test_registerAdapter_w_invalid_required(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, ibar, provided=ifoo, name=_name, info=_info) def test_registerAdapter_w_required_containing_None(self): from zope.interface.declarations import InterfaceClass from zope.interface.interface import Interface from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, [None], provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][Interface][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(Interface,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (Interface,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_w_required_containing_class(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.declarations import implementedBy from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context @implementer(ibar) class _Context(object): pass _ctx_impl = implementedBy(_Context) comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, [_Context], provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][_ctx_impl][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(_ctx_impl,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (_ctx_impl,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_w_required_containing_junk(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() self.assertRaises(TypeError, comp.registerAdapter, _Factory, [object()], provided=ifoo, name=_name, info=_info) def test_registerAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import AdapterRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_Factory, provided=ifoo, name=_name, info=_info) self.failUnless(comp.adapters._adapters[1][ibar][ifoo][_name] is _Factory) self.assertEqual(comp._adapter_registrations[(ibar,), ifoo, _name], (_Factory, _info)) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.failUnless(event.object.name is _name) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerAdapter_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _name = _u('name') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerAdapter(_factory, (ibar,), ifoo, _name, _info, event=False) self.assertEqual(len(_events), 0) def test_unregisterAdapter_neither_factory_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterAdapter, factory=None, provided=None) def test_unregisterAdapter_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterAdapter, factory=None, provided=ifoo, required=None) def test_unregisterAdapter_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,), ifoo) self.failIf(unreg) def test_unregisterAdapter_hit_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._adapters) self.failIf(comp._adapter_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, (ibar,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterAdapter(_Factory, provided=ifoo) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, AdapterRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_registeredAdapters_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredAdapters()), []) def test_registeredAdapters_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import AdapterRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IFoo') _info = _u('info') _name1 = _u('name1') _name2 = _u('name2') class _Factory(object): def __init__(self, context): pass comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo, _name1, _info) comp.registerAdapter(_Factory, (ibar,), ifoo, _name2, _info) reg = sorted(comp.registeredAdapters(), key=lambda r: r.name) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], AdapterRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.assertEqual(reg[0].required, (ibar,)) self.failUnless(reg[0].name is _name1) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is _Factory) self.failUnless(isinstance(reg[1], AdapterRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.assertEqual(reg[1].required, (ibar,)) self.failUnless(reg[1].name is _name2) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is _Factory) def test_queryAdapter_miss_no_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _context = object() self.failUnless(comp.queryAdapter(_context, ifoo) is None) def test_queryAdapter_miss_w_default(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _context = object() _default = object() self.failUnless( comp.queryAdapter(_context, ifoo, default=_default) is _default) def test_queryAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self.context = context @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) adapter = comp.queryAdapter(_context, ifoo) self.failUnless(isinstance(adapter, _Factory)) self.failUnless(adapter.context is _context) def test_getAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getAdapter, _context, ifoo) def test_getAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self.context = context @implementer(ibar) class _Context(object): pass _context = _Context() comp = self._makeOne() comp.registerAdapter(_Factory, (ibar,), ifoo) adapter = comp.getAdapter(_context, ifoo) self.failUnless(isinstance(adapter, _Factory)) self.failUnless(adapter.context is _context) def test_queryMultiAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertEqual(comp.queryMultiAdapter((_context1, _context2), ifoo), None) def test_queryMultiAdapter_miss_w_default(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() _default = object() comp = self._makeOne() self.failUnless( comp.queryMultiAdapter((_context1, _context2), ifoo, default=_default) is _default) def test_queryMultiAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory(object): def __init__(self, context1, context2): self.context = context1, context2 comp = self._makeOne() comp.registerAdapter(_Factory, (ibar, ibaz), ifoo) adapter = comp.queryMultiAdapter((_context1, _context2), ifoo) self.failUnless(isinstance(adapter, _Factory)) self.assertEqual(adapter.context, (_context1, _context2)) def test_getMultiAdapter_miss(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import ComponentLookupError class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertRaises(ComponentLookupError, comp.getMultiAdapter, (_context1, _context2), ifoo) def test_getMultiAdapter_hit(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory(object): def __init__(self, context1, context2): self.context = context1, context2 comp = self._makeOne() comp.registerAdapter(_Factory, (ibar, ibaz), ifoo) adapter = comp.getMultiAdapter((_context1, _context2), ifoo) self.failUnless(isinstance(adapter, _Factory)) self.assertEqual(adapter.context, (_context1, _context2)) def test_getAdapters_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() comp = self._makeOne() self.assertEqual( list(comp.getAdapters((_context1, _context2), ifoo)), []) def test_getAdapters_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') ibaz = IFoo('IBaz') @implementer(ibar) class _Context1(object): pass @implementer(ibaz) class _Context2(object): pass _context1 = _Context1() _context2 = _Context2() class _Factory1(object): def __init__(self, context1, context2): self.context = context1, context2 class _Factory2(object): def __init__(self, context1, context2): self.context = context1, context2 _name1 = _u('name1') _name2 = _u('name2') comp = self._makeOne() comp.registerAdapter(_Factory1, (ibar, ibaz), ifoo, name=_name1) comp.registerAdapter(_Factory2, (ibar, ibaz), ifoo, name=_name2) found = sorted(comp.getAdapters((_context1, _context2), ifoo)) self.assertEqual(len(found), 2) self.assertEqual(found[0][0], _name1) self.failUnless(isinstance(found[0][1], _Factory1)) self.assertEqual(found[1][0], _name2) self.failUnless(isinstance(found[1][1], _Factory2)) def test_registerSubscriptionAdapter_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _name = _u('name') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() self.assertRaises(TypeError, comp.registerSubscriptionAdapter, _factory, (ibar,), ifoo, _name, _info) def test_registerSubscriptionAdapter_w_explicit_provided_and_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_factory, (ibar,), ifoo, info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerSubscriptionAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _blank = _u('') _to_reg = object() @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_Factory, (ibar,), info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerSubscriptionAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import SubscriptionRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _info = _u('info') _blank = _u('') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter( _Factory, provided=ifoo, info=_info) reg = comp.adapters._subscribers[1][ibar][ifoo][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._subscription_registrations, [((ibar,), ifoo, _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registerSubscriptionAdapter_wo_event(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerSubscriptionAdapter(_factory, (ibar,), ifoo, info=_info, event=False) self.assertEqual(len(_events), 0) def test_registeredSubscriptionAdapters_empty(self): comp = self._makeOne() self.assertEqual(list(comp.registeredSubscriptionAdapters()), []) def test_registeredSubscriptionAdapters_notempty(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IFoo') _info = _u('info') _blank = _u('') class _Factory(object): def __init__(self, context): pass comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo, info=_info) comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo, info=_info) reg = list(comp.registeredSubscriptionAdapters()) self.assertEqual(len(reg), 2) self.failUnless(isinstance(reg[0], SubscriptionRegistration)) self.failUnless(reg[0].registry is comp) self.failUnless(reg[0].provided is ifoo) self.assertEqual(reg[0].required, (ibar,)) self.assertEqual(reg[0].name, _blank) self.failUnless(reg[0].info is _info) self.failUnless(reg[0].factory is _Factory) self.failUnless(isinstance(reg[1], SubscriptionRegistration)) self.failUnless(reg[1].registry is comp) self.failUnless(reg[1].provided is ifoo) self.assertEqual(reg[1].required, (ibar,)) self.assertEqual(reg[1].name, _blank) self.failUnless(reg[1].info is _info) self.failUnless(reg[1].factory is _Factory) def test_unregisterSubscriptionAdapter_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') _nonblank = _u('nonblank') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, required=ifoo, provided=ibar, name=_nonblank) def test_unregisterSubscriptionAdapter_neither_factory_nor_provided(self): comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, factory=None, provided=None) def test_unregisterSubscriptionAdapter_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterSubscriptionAdapter, factory=None, provided=ifoo, required=None) def test_unregisterSubscriptionAdapter_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,), ifoo) self.failIf(unreg) self.failIf(_events) def test_unregisterSubscriptionAdapter_hit_wo_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(None, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._subscribers) self.failIf(comp._subscription_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is None) def test_unregisterSubscriptionAdapter_hit_w_factory(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,), ifoo) self.failUnless(unreg) self.failIf(comp.adapters._subscribers) self.failIf(comp._subscription_registrations) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterSubscriptionAdapter_wo_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') @implementer(ifoo) class _Factory(object): def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, (ibar,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_unregisterSubscriptionAdapter_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import SubscriptionRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterSubscriptionAdapter(_Factory, provided=ifoo) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, SubscriptionRegistration)) self.failUnless(event.object.registry is comp) self.failUnless(event.object.provided is ifoo) self.assertEqual(event.object.required, (ibar,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_subscribers_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') comp = self._makeOne() @implementer(ibar) class Bar(object): pass bar = Bar() self.assertEqual(list(comp.subscribers((bar,), ifoo)), []) def test_subscribers_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Factory(object): __component_adapts__ = (ibar,) def __init__(self, context): self._context = context class _Derived(_Factory): pass comp = self._makeOne() comp.registerSubscriptionAdapter(_Factory, (ibar,), ifoo) comp.registerSubscriptionAdapter(_Derived, (ibar,), ifoo) @implementer(ibar) class Bar(object): pass bar = Bar() subscribers = comp.subscribers((bar,), ifoo) def _klassname(x): return x.__class__.__name__ subscribers = sorted(subscribers, key=_klassname) self.assertEqual(len(subscribers), 2) self.failUnless(isinstance(subscribers[0], _Derived)) self.failUnless(isinstance(subscribers[1], _Factory)) def test_registerHandler_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _nonblank = _u('nonblank') comp = self._makeOne() def _factory(context): pass self.assertRaises(TypeError, comp.registerHandler, _factory, required=ifoo, name=_nonblank) def test_registerHandler_w_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import HandlerRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _blank = _u('') _info = _u('info') _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerHandler(_factory, (ifoo,), info=_info) reg = comp.adapters._subscribers[1][ifoo][None][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _factory) self.assertEqual(comp._handler_registrations, [((ifoo,), _blank, _factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _factory) def test_registerHandler_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Registered from zope.interface.registry import HandlerRegistration from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _info = _u('info') _blank = _u('') class _Factory(object): __component_adapts__ = (ifoo,) def __init__(self, context): self._context = context comp = self._makeOne() _monkey, _events = self._wrapEvents() with _monkey: comp.registerHandler(_Factory, info=_info) reg = comp.adapters._subscribers[1][ifoo][None][_blank] self.assertEqual(len(reg), 1) self.failUnless(reg[0] is _Factory) self.assertEqual(comp._handler_registrations, [((ifoo,), _blank, _Factory, _info)]) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Registered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, _blank) self.failUnless(event.object.info is _info) self.failUnless(event.object.factory is _Factory) def test_registeredHandlers_empty(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass comp = self._makeOne() self.failIf(list(comp.registeredHandlers())) def test_registeredHandlers_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') def _factory1(context): pass def _factory2(context): pass comp = self._makeOne() comp.registerHandler(_factory1, (ifoo,)) comp.registerHandler(_factory2, (ifoo,)) def _factory_name(x): return x.factory.__code__.co_name subscribers = sorted(comp.registeredHandlers(), key=_factory_name) self.assertEqual(len(subscribers), 2) self.failUnless(isinstance(subscribers[0], HandlerRegistration)) self.assertEqual(subscribers[0].required, (ifoo,)) self.assertEqual(subscribers[0].name, '') self.assertEqual(subscribers[0].factory, _factory1) self.assertEqual(subscribers[0].info, '') self.failUnless(isinstance(subscribers[1], HandlerRegistration)) self.assertEqual(subscribers[1].required, (ifoo,)) self.assertEqual(subscribers[1].name, '') self.assertEqual(subscribers[1].factory, _factory2) self.assertEqual(subscribers[1].info, '') def test_unregisterHandler_w_nonblank_name(self): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _nonblank = _u('nonblank') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterHandler, required=(ifoo,), name=_nonblank) def test_unregisterHandler_neither_factory_nor_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() self.assertRaises(TypeError, comp.unregisterHandler) def test_unregisterHandler_miss(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() unreg = comp.unregisterHandler(required=(ifoo,)) self.failIf(unreg) def test_unregisterHandler_hit_w_factory_and_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() comp.registerHandler(_factory, (ifoo,)) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(_factory, (ifoo,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.failUnless(event.object.factory is _factory) def test_unregisterHandler_hit_w_only_explicit_provided(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() _to_reg = object() def _factory(context): return _to_reg comp = self._makeOne() comp.registerHandler(_factory, (ifoo,)) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(required=(ifoo,)) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.failUnless(event.object.factory is None) def test_unregisterHandler_wo_explicit_required(self): from zope.interface.declarations import InterfaceClass from zope.interface.interfaces import Unregistered from zope.interface.registry import HandlerRegistration class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Factory(object): __component_adapts__ = (ifoo,) def __init__(self, context): self._context = context comp = self._makeOne() comp.registerHandler(_Factory) _monkey, _events = self._wrapEvents() with _monkey: unreg = comp.unregisterHandler(_Factory) self.failUnless(unreg) self.assertEqual(len(_events), 1) args, kw = _events[0] event, = args self.assertEqual(kw, {}) self.failUnless(isinstance(event, Unregistered)) self.failUnless(isinstance(event.object, HandlerRegistration)) self.failUnless(event.object.registry is comp) self.assertEqual(event.object.required, (ifoo,)) self.assertEqual(event.object.name, '') self.assertEqual(event.object.info, '') self.failUnless(event.object.factory is _Factory) def test_handle_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') comp = self._makeOne() @implementer(ifoo) class Bar(object): pass bar = Bar() comp.handle((bar,)) def test_handle_non_empty(self): from zope.interface.declarations import InterfaceClass from zope.interface.declarations import implementer class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') _called_1 = [] def _factory_1(context): _called_1.append(context) _called_2 = [] def _factory_2(context): _called_2.append(context) comp = self._makeOne() comp.registerHandler(_factory_1, (ifoo,)) comp.registerHandler(_factory_2, (ifoo,)) @implementer(ifoo) class Bar(object): pass bar = Bar() comp.handle(bar) self.assertEqual(_called_1, [bar]) self.assertEqual(_called_2, [bar]) # Test _getUtilityProvided, _getAdapterProvided, _getAdapterRequired via their # callers (Component.registerUtility, Component.registerAdapter). class UtilityRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import UtilityRegistration return UtilityRegistration def _makeOne(self, component=None, factory=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, ifoo, name, component, doc, factory), registry, name, ) def test_class_conforms_to_IUtilityRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IUtilityRegistration verifyClass(IUtilityRegistration, self._getTargetClass()) def test_instance_conforms_to_IUtilityRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IUtilityRegistration ur, _, _ = self._makeOne() verifyObject(IUtilityRegistration, ur) def test___repr__(self): class _Component(object): __name__ = 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, IFoo, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___repr___provided_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) ur.provided = object() self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, None, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___repr___component_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ur, _registry, _name = self._makeOne(_component) ur.provided = object() self.assertEqual(repr(ur), "UtilityRegistration(_REGISTRY, None, %r, TEST, None, 'DOCSTRING')" % (_name)) def test___hash__(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.assertEqual(ur.__hash__(), id(ur)) def test___eq___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur == ur) def test___eq___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failUnless(ur == ur2) def test___eq___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) self.failIf(ur == ur2) def test___ne___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur != ur) def test___ne___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur != ur2) def test___ne___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) self.failUnless(ur != ur2) def test___lt___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur < ur) def test___lt___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur < ur2) def test___lt___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur < ur2) def test___le___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur <= ur) def test___le___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failUnless(ur <= ur2) def test___le___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur <= ur2) def test___gt___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failIf(ur > ur) def test___gt___hit(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failUnless(ur2 > ur) def test___gt___miss(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) self.failIf(ur2 > ur) def test___ge___identity(self): _component = object() ur, _registry, _name = self._makeOne(_component) self.failUnless(ur >= ur) def test___ge___miss(self): _component = object() _component2 = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component2) ur2.name = _name + '2' self.failIf(ur >= ur2) def test___ge___hit(self): _component = object() ur, _registry, _name = self._makeOne(_component) ur2, _, _ = self._makeOne(_component) ur2.name = _name + '2' self.failUnless(ur2 >= ur) class AdapterRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import AdapterRegistration return AdapterRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ibar,), ifoo, name, component, doc), registry, name, ) def test_class_conforms_to_IAdapterRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IAdapterRegistration verifyClass(IAdapterRegistration, self._getTargetClass()) def test_instance_conforms_to_IAdapterRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IAdapterRegistration ar, _, _ = self._makeOne() verifyObject(IAdapterRegistration, ar) def test___repr__(self): class _Component(object): __name__ = 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], IFoo, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___provided_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) ar.provided = object() self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], None, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___component_wo_name(self): class _Component(object): def __repr__(self): return 'TEST' _component = _Component() ar, _registry, _name = self._makeOne(_component) ar.provided = object() self.assertEqual(repr(ar), ("AdapterRegistration(_REGISTRY, [IBar], None, %r, TEST, " + "'DOCSTRING')") % (_name)) def test___hash__(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.assertEqual(ar.__hash__(), id(ar)) def test___eq___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar == ar) def test___eq___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failUnless(ar == ar2) def test___eq___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) self.failIf(ar == ar2) def test___ne___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar != ar) def test___ne___miss(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar != ar2) def test___ne___hit_component(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) self.failUnless(ar != ar2) def test___ne___hit_provided(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ibaz = IFoo('IBaz') _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) ar2.provided = ibaz self.failUnless(ar != ar2) def test___ne___hit_required(self): from zope.interface.declarations import InterfaceClass class IFoo(InterfaceClass): pass ibaz = IFoo('IBaz') _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.required = (ibaz,) self.failUnless(ar != ar2) def test___lt___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar < ar) def test___lt___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar < ar2) def test___lt___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar < ar2) def test___le___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar <= ar) def test___le___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failUnless(ar <= ar2) def test___le___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar <= ar2) def test___gt___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failIf(ar > ar) def test___gt___hit(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failUnless(ar2 > ar) def test___gt___miss(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) self.failIf(ar2 > ar) def test___ge___identity(self): _component = object() ar, _registry, _name = self._makeOne(_component) self.failUnless(ar >= ar) def test___ge___miss(self): _component = object() _component2 = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component2) ar2.name = _name + '2' self.failIf(ar >= ar2) def test___ge___hit(self): _component = object() ar, _registry, _name = self._makeOne(_component) ar2, _, _ = self._makeOne(_component) ar2.name = _name + '2' self.failUnless(ar2 >= ar) class SubscriptionRegistrationTests(unittest.TestCase): def _getTargetClass(self): from zope.interface.registry import SubscriptionRegistration return SubscriptionRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') ibar = IFoo('IBar') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ibar,), ifoo, name, component, doc), registry, name, ) def test_class_conforms_to_ISubscriptionAdapterRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import ISubscriptionAdapterRegistration verifyClass(ISubscriptionAdapterRegistration, self._getTargetClass()) def test_instance_conforms_to_ISubscriptionAdapterRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import ISubscriptionAdapterRegistration sar, _, _ = self._makeOne() verifyObject(ISubscriptionAdapterRegistration, sar) class HandlerRegistrationTests(_SilencePy3Deprecations): def _getTargetClass(self): from zope.interface.registry import HandlerRegistration return HandlerRegistration def _makeOne(self, component=None): from zope.interface.declarations import InterfaceClass from zope.interface._compat import _u class IFoo(InterfaceClass): pass ifoo = IFoo('IFoo') class _Registry(object): def __repr__(self): return '_REGISTRY' registry = _Registry() name = _u('name') doc = 'DOCSTRING' klass = self._getTargetClass() return (klass(registry, (ifoo,), name, component, doc), registry, name, ) def test_class_conforms_to_IHandlerRegistration(self): from zope.interface.verify import verifyClass from zope.interface.interfaces import IHandlerRegistration verifyClass(IHandlerRegistration, self._getTargetClass()) def test_instance_conforms_to_IHandlerRegistration(self): from zope.interface.verify import verifyObject from zope.interface.interfaces import IHandlerRegistration hr, _, _ = self._makeOne() verifyObject(IHandlerRegistration, hr) def test_properties(self): def _factory(context): pass hr, _, _ = self._makeOne(_factory) self.failUnless(hr.handler is _factory) self.failUnless(hr.factory is hr.handler) self.failUnless(hr.provided is None) def test___repr___factory_w_name(self): class _Factory(object): __name__ = 'TEST' hr, _registry, _name = self._makeOne(_Factory()) self.assertEqual(repr(hr), ("HandlerRegistration(_REGISTRY, [IFoo], %r, TEST, " + "'DOCSTRING')") % (_name)) def test___repr___factory_wo_name(self): class _Factory(object): def __repr__(self): return 'TEST' hr, _registry, _name = self._makeOne(_Factory()) self.assertEqual(repr(hr), ("HandlerRegistration(_REGISTRY, [IFoo], %r, TEST, " + "'DOCSTRING')") % (_name)) class _Monkey(object): # context-manager for replacing module names in the scope of a test. def __init__(self, module, **kw): self.module = module self.to_restore = dict([(key, getattr(module, key)) for key in kw]) for key, value in kw.items(): setattr(module, key, value) def __enter__(self): return self def __exit__(self, exc_type, exc_val, exc_tb): for key, value in self.to_restore.items(): setattr(self.module, key, value) def test_suite(): return unittest.TestSuite(( unittest.makeSuite(ComponentsTests), unittest.makeSuite(UtilityRegistrationTests), unittest.makeSuite(AdapterRegistrationTests), unittest.makeSuite(SubscriptionRegistrationTests), unittest.makeSuite(AdapterRegistrationTests), ))
true
true
f717c69030d602486ed1591a2382be46b01ff559
3,373
py
Python
hw/ip/otbn/dv/otbnsim/sim/csr.py
Jacob-Levy/opentitan
13ed3e95572168f7db0e3178d06a72e72ee5a745
[ "Apache-2.0" ]
2
2019-11-21T14:05:14.000Z
2020-07-10T12:40:54.000Z
hw/ip/otbn/dv/otbnsim/sim/csr.py
Jacob-Levy/opentitan
13ed3e95572168f7db0e3178d06a72e72ee5a745
[ "Apache-2.0" ]
null
null
null
hw/ip/otbn/dv/otbnsim/sim/csr.py
Jacob-Levy/opentitan
13ed3e95572168f7db0e3178d06a72e72ee5a745
[ "Apache-2.0" ]
null
null
null
# Copyright lowRISC contributors. # Licensed under the Apache License, Version 2.0, see LICENSE for details. # SPDX-License-Identifier: Apache-2.0 from .flags import FlagGroups from .wsr import WSRFile class CSRFile: '''A model of the CSR file''' def __init__(self) -> None: self.flags = FlagGroups() self._known_indices = set() self._known_indices.add(0x7c0) # FG0 self._known_indices.add(0x7c1) # FG1 self._known_indices.add(0x7c8) # FLAGS for idx in range(0x7d0, 0x7d8): self._known_indices.add(idx) # MODi self._known_indices.add(0x7d8) # RND_PREFETCH self._known_indices.add(0xfc0) # RND self._known_indices.add(0xfc1) # URND @staticmethod def _get_field(field_idx: int, field_size: int, val: int) -> int: mask = (1 << field_size) - 1 return (val >> (field_size * field_idx)) & mask @staticmethod def _set_field(field_idx: int, field_size: int, field_val: int, old_val: int) -> int: assert 0 <= field_val < (1 << field_size) mask = (1 << field_size) - 1 shift = field_size * field_idx return (old_val & ~(mask << shift)) | (field_val << shift) def check_idx(self, idx: int) -> bool: '''Return True if idx points to a valid CSR; False otherwise.''' return idx in self._known_indices def read_unsigned(self, wsrs: WSRFile, idx: int) -> int: if 0x7c0 <= idx <= 0x7c1: # FG0/FG1 fg = idx - 0x7c0 return self._get_field(fg, 4, self.flags.read_unsigned()) if idx == 0x7c8: # FLAGS register return self.flags.read_unsigned() if 0x7d0 <= idx <= 0x7d7: # MOD0 .. MOD7. MODi is bits [32*(i+1)-1..32*i] mod_n = idx - 0x7d0 return self._get_field(mod_n, 32, wsrs.MOD.read_unsigned()) if idx == 0x7d8: # RND_PREFETCH register return 0 if idx == 0xfc0: # RND register return wsrs.RND.read_u32() if idx == 0xfc1: # URND register return wsrs.URND.read_u32() raise RuntimeError('Unknown CSR index: {:#x}'.format(idx)) def write_unsigned(self, wsrs: WSRFile, idx: int, value: int) -> None: assert 0 <= value < (1 << 32) if 0x7c0 <= idx <= 0x7c1: # FG0/FG1 fg = idx - 0x7c0 old = self.flags.read_unsigned() self.flags.write_unsigned(self._set_field(fg, 4, value & 0xf, old)) return if idx == 0x7c8: # FLAGS register self.flags.write_unsigned(value) return if 0x7d0 <= idx <= 0x7d7: # MOD0 .. MOD7. MODi is bits [32*(i+1)-1..32*i]. read,modify,write. mod_n = idx - 0x7d0 old = wsrs.MOD.read_unsigned() wsrs.MOD.write_unsigned(self._set_field(mod_n, 32, value, old)) return if idx == 0x7d8: # RND_PREFETCH wsrs.RND.pending_request = True return if idx == 0xfc0: # RND register (which ignores writes) return if idx == 0xfc1: # URND register (which ignores writes) return raise RuntimeError('Unknown CSR index: {:#x}'.format(idx))
31.820755
79
0.555885
from .flags import FlagGroups from .wsr import WSRFile class CSRFile: def __init__(self) -> None: self.flags = FlagGroups() self._known_indices = set() self._known_indices.add(0x7c0) self._known_indices.add(0x7c1) self._known_indices.add(0x7c8) for idx in range(0x7d0, 0x7d8): self._known_indices.add(idx) self._known_indices.add(0x7d8) self._known_indices.add(0xfc0) self._known_indices.add(0xfc1) @staticmethod def _get_field(field_idx: int, field_size: int, val: int) -> int: mask = (1 << field_size) - 1 return (val >> (field_size * field_idx)) & mask @staticmethod def _set_field(field_idx: int, field_size: int, field_val: int, old_val: int) -> int: assert 0 <= field_val < (1 << field_size) mask = (1 << field_size) - 1 shift = field_size * field_idx return (old_val & ~(mask << shift)) | (field_val << shift) def check_idx(self, idx: int) -> bool: return idx in self._known_indices def read_unsigned(self, wsrs: WSRFile, idx: int) -> int: if 0x7c0 <= idx <= 0x7c1: fg = idx - 0x7c0 return self._get_field(fg, 4, self.flags.read_unsigned()) if idx == 0x7c8: return self.flags.read_unsigned() if 0x7d0 <= idx <= 0x7d7: mod_n = idx - 0x7d0 return self._get_field(mod_n, 32, wsrs.MOD.read_unsigned()) if idx == 0x7d8: return 0 if idx == 0xfc0: return wsrs.RND.read_u32() if idx == 0xfc1: return wsrs.URND.read_u32() raise RuntimeError('Unknown CSR index: {:#x}'.format(idx)) def write_unsigned(self, wsrs: WSRFile, idx: int, value: int) -> None: assert 0 <= value < (1 << 32) if 0x7c0 <= idx <= 0x7c1: fg = idx - 0x7c0 old = self.flags.read_unsigned() self.flags.write_unsigned(self._set_field(fg, 4, value & 0xf, old)) return if idx == 0x7c8: self.flags.write_unsigned(value) return if 0x7d0 <= idx <= 0x7d7: mod_n = idx - 0x7d0 old = wsrs.MOD.read_unsigned() wsrs.MOD.write_unsigned(self._set_field(mod_n, 32, value, old)) return if idx == 0x7d8: wsrs.RND.pending_request = True return if idx == 0xfc0: return if idx == 0xfc1: return raise RuntimeError('Unknown CSR index: {:#x}'.format(idx))
true
true
f717c7a5b3ca5f9a8a20ce7a7c9f7a50b5ae1c68
14,021
py
Python
tests/ci/sync_requirements.py
lolwww/rally
fcb1fb6c608e29dd62549cf6b3cec2e90529932f
[ "Apache-2.0" ]
null
null
null
tests/ci/sync_requirements.py
lolwww/rally
fcb1fb6c608e29dd62549cf6b3cec2e90529932f
[ "Apache-2.0" ]
null
null
null
tests/ci/sync_requirements.py
lolwww/rally
fcb1fb6c608e29dd62549cf6b3cec2e90529932f
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """ Synchronizes, formats and prepares requirements to release(obtains and adds maximum allowed version). """ import collections import logging import re import sys import textwrap import requests LOG = logging.getLogger(__name__) if not LOG.handlers: LOG.addHandler(logging.StreamHandler()) LOG.setLevel(logging.INFO) GLOBAL_REQUIREMENTS_LOCATIONS = ( "https://raw.githubusercontent.com/openstack/requirements/master/", "http://opendev.org/openstack/requirements/raw/branch/master/" ) RALLY_REQUIREMENTS_FILES = ( "requirements.txt", "test-requirements.txt" ) DO_NOT_TOUCH_TAG = "[do-not-touch]" class Comment(object): def __init__(self, s=None, finished=False): self._comments = [] self.is_finished = finished if s: self.append(s) def finish_him(self): self.is_finished = True def append(self, s): self._comments.append(s[1:].strip()) def __str__(self): return textwrap.fill("\n".join(self._comments), width=80, initial_indent="# ", subsequent_indent="# ") _PYPI_CACHE = {} class PYPIPackage(object): # NOTE(andreykurilin): one license can have different labels. Let's use # unified variant. LICENSE_MAP = {"MIT license": "MIT", "MIT License": "MIT", "BSD License": "BSD", "Apache 2.0": "Apache License, Version 2.0"} def __init__(self, package_name): self.package_name = package_name self._pypi_info = None self._pypi_license = None @property def pypi_info(self): if self._pypi_info is None: if self.package_name in _PYPI_CACHE: self._pypi_info = _PYPI_CACHE[self.package_name] else: resp = requests.get("https://pypi.org/pypi/%s/json" % self.package_name) if resp.status_code != 200: print("An error occurred while checking '%s' package at " "pypi." % self.package_name) raise Exception(resp.text) self._pypi_info = resp.json() # let's cache it for the case when we need to sync requirements # and update upper constrains _PYPI_CACHE[self.package_name] = self._pypi_info return self._pypi_info @property def pypi_version(self): return self.pypi_info["info"]["version"] @property def pypi_license(self): if self._pypi_license is None: if self.pypi_info["info"]["license"]: self._pypi_license = self.pypi_info["info"]["license"] else: # try to parse classifiers prefix = "License :: OSI Approved :: " classifiers = [c[len(prefix):] for c in self.pypi_info["info"]["classifiers"] if c.startswith(prefix)] self._pypi_license = "/".join(classifiers) self._license = self.LICENSE_MAP.get(self._pypi_license, self._pypi_license) if self._pypi_license == "UNKNOWN": self._pypi_license = None return self._license def __eq__(self, other): return (isinstance(other, PYPIPackage) and self.package_name == other.package_name) class Requirement(PYPIPackage): RE_NAME = re.compile(r"[a-zA-Z0-9-._]+") RE_CONST_VERSION = re.compile(r"==[a-zA-Z0-9.]+") RE_MIN_VERSION = re.compile(r">=?[a-zA-Z0-9.]+") RE_MAX_VERSION = re.compile(r"<=?[a-zA-Z0-9.]+") RE_NE_VERSIONS = re.compile(r"!=[a-zA-Z0-9.]+") def __init__(self, package_name, version): super(Requirement, self).__init__(package_name) self.version = version self.do_not_touch = False def sync_max_version_with_pypy(self): if isinstance(self.version, dict) and not self.do_not_touch: self.version["max"] = "<=%s" % self.pypi_version @classmethod def parse_line(cls, line): match = cls.RE_NAME.match(line) if match: name = match.group() # remove name versions = line.replace(name, "") # remove comments versions = versions.split("#")[0] # remove python classifiers versions = versions.split(";")[0].strip() if not cls.RE_CONST_VERSION.match(versions): versions = versions.strip().split(",") min_version = None max_version = None ne_versions = [] for version in versions: if cls.RE_MIN_VERSION.match(version): if min_version: raise Exception("Found several min versions for " "%s package." % name) min_version = version elif cls.RE_MAX_VERSION.match(version): if max_version: raise Exception("Found several max versions for " "%s package." % name) max_version = version elif cls.RE_NE_VERSIONS.match(version): ne_versions.append(version) versions = {"min": min_version, "max": max_version, "ne": ne_versions} return cls(name, versions) def __str__(self): if isinstance(self.version, dict): version = [] min_equal_to_max = False if self.version["min"] and self.version["max"]: if ( self.version["min"].startswith(">=") and self.version["max"].startswith("<=") and self.version["min"][2:] == self.version["max"][2:] ): # min and max versions are equal there is no need to write # both of them min_equal_to_max = True version.append("==%s" % self.version["min"][2:]) if not min_equal_to_max and self.version["min"]: version.append(self.version["min"]) if not min_equal_to_max and self.version["ne"]: version.extend(self.version["ne"]) if not min_equal_to_max and self.version["max"]: version.append(self.version["max"]) version = ",".join(version) else: if self.do_not_touch: version = self.version else: # remove const version version = ">=%s" % self.version[2:] string = "%s%s" % (self.package_name, version) if self.pypi_license: # NOTE(andreykurilin): When I start implementation of this script, # python-keystoneclient dependency string took around ~45-55 # chars, so let's use this length as indent. Feel free to modify # it to lower or greater value. magic_number = 55 if len(string) < magic_number: indent = magic_number - len(string) else: indent = 2 string += " " * indent + "# " + self.pypi_license return string def __eq__(self, other): return (isinstance(other, self.__class__) and self.package_name == other.package_name) def __ne__(self, other): return not self.__eq__(other) class UpperConstraint(PYPIPackage): RE_LINE = re.compile( r"(?P<package_name>[a-zA-Z0-9-._]+)===(?P<version>[a-zA-Z0-9.]+)") def __init__(self, package_name, version=None): super(UpperConstraint, self).__init__(package_name) self._version = version def __str__(self): return "%s===%s" % (self.package_name, self.version) @property def version(self): if self._version is None: self._version = self.pypi_version return self._version @classmethod def parse_line(cls, line): match = cls.RE_LINE.match(line) if match: return cls(**match.groupdict()) def update(self, version): self._version = version def parse_data(raw_data, include_comments=True, dependency_cls=Requirement): # first elem is None to simplify checks of last elem in requirements requirements = [None] for line in raw_data.split("\n"): if line.startswith("#"): if not include_comments: continue if getattr(requirements[-1], "is_finished", True): requirements.append(Comment()) requirements[-1].append(line) elif line == "": # just empty line if isinstance(requirements[-1], Comment): requirements[-1].finish_him() requirements.append(Comment(finished=True)) else: if (isinstance(requirements[-1], Comment) and not requirements[-1].is_finished): requirements[-1].finish_him() # parse_line dep = dependency_cls.parse_line(line) if dep: if (isinstance(requirements[-1], Comment) and DO_NOT_TOUCH_TAG in str(requirements[-1])): dep.do_not_touch = True requirements.append(dep) for i in range(len(requirements) - 1, 0, -1): # remove empty lines at the end of file if isinstance(requirements[i], Comment): if str(requirements[i]) == "": requirements.pop(i) else: break return collections.OrderedDict( (v if isinstance(v, Comment) else v.package_name, v) for v in requirements if v) def _fetch_from_gr(filename): """Try to fetch data from OpenStack global-requirements repo""" for i in range(0, len(GLOBAL_REQUIREMENTS_LOCATIONS)): url = GLOBAL_REQUIREMENTS_LOCATIONS[i] + filename LOG.debug("Try to obtain %s from %s" % (filename, url)) try: return requests.get(url).text except requests.ConnectionError as e: LOG.exception(e) raise Exception("Unable to obtain %s" % filename) def _write_requirements(filename, requirements): """Saves requirements to file.""" if isinstance(requirements, dict): requirements = requirements.values() LOG.info("Saving requirements to %s." % filename) with open(filename, "w") as f: for entity in requirements: f.write(str(entity)) f.write("\n") def sync_requirements(): """Synchronizes Rally requirements with OpenStack global-requirements.""" LOG.info("Obtaining global-requirements of OpenStack...") raw_gr = _fetch_from_gr("global-requirements.txt") # NOTE(andreykurilin): global-requirements includes comments which can be # unrelated to Rally project, so let's just ignore them gr = parse_data(raw_gr, include_comments=False) for file_name in RALLY_REQUIREMENTS_FILES: LOG.debug("Processing '%s'." % file_name) with open(file_name) as f: requirements = parse_data(f.read()) for name, req in requirements.items(): if isinstance(req, Requirement) and not req.do_not_touch: if name in gr: req.version = gr[req.package_name].version else: # it not g-r requirements if isinstance(req.version, dict): req.version["max"] = None _write_requirements(file_name, requirements) def update_upper_constraints(): """Obtains latest version of packages and put them to upper-constraints.""" LOG.info("Obtaining upper-constrains from OpenStack...") raw_g_uc = _fetch_from_gr("upper-constraints.txt") # NOTE(andreykurilin): global OpenStack upper-constraints file includes # comments which can be unrelated to Rally project, so let's just ignore # them. global_uc = parse_data(raw_g_uc, include_comments=False, dependency_cls=UpperConstraint) with open("upper-constraints.txt") as f: our_uc = parse_data(f.read(), dependency_cls=UpperConstraint) with open("requirements.txt") as f: our_requirements = parse_data(f.read(), include_comments=False) for name, req in our_requirements.items(): if isinstance(req, Comment): continue if name not in our_uc: our_uc[name] = UpperConstraint(name) if name in global_uc: # we cannot use whatever we want versions in CI. OpenStack CI # ignores versions listed in requirements of # particular project and use versions from global u-c file. # It means that we need to suggest to use the same versions our_uc[name].update(global_uc[name].version) our_uc = sorted(our_uc.values(), key=lambda o: o.package_name.upper()) _write_requirements("upper-constraints.txt", our_uc) def main(): sync_requirements() update_upper_constraints() if __name__ == "__main__": sys.exit(main())
36.136598
79
0.57799
import collections import logging import re import sys import textwrap import requests LOG = logging.getLogger(__name__) if not LOG.handlers: LOG.addHandler(logging.StreamHandler()) LOG.setLevel(logging.INFO) GLOBAL_REQUIREMENTS_LOCATIONS = ( "https://raw.githubusercontent.com/openstack/requirements/master/", "http://opendev.org/openstack/requirements/raw/branch/master/" ) RALLY_REQUIREMENTS_FILES = ( "requirements.txt", "test-requirements.txt" ) DO_NOT_TOUCH_TAG = "[do-not-touch]" class Comment(object): def __init__(self, s=None, finished=False): self._comments = [] self.is_finished = finished if s: self.append(s) def finish_him(self): self.is_finished = True def append(self, s): self._comments.append(s[1:].strip()) def __str__(self): return textwrap.fill("\n".join(self._comments), width=80, initial_indent="# ", subsequent_indent="# ") _PYPI_CACHE = {} class PYPIPackage(object): # unified variant. LICENSE_MAP = {"MIT license": "MIT", "MIT License": "MIT", "BSD License": "BSD", "Apache 2.0": "Apache License, Version 2.0"} def __init__(self, package_name): self.package_name = package_name self._pypi_info = None self._pypi_license = None @property def pypi_info(self): if self._pypi_info is None: if self.package_name in _PYPI_CACHE: self._pypi_info = _PYPI_CACHE[self.package_name] else: resp = requests.get("https://pypi.org/pypi/%s/json" % self.package_name) if resp.status_code != 200: print("An error occurred while checking '%s' package at " "pypi." % self.package_name) raise Exception(resp.text) self._pypi_info = resp.json() # let's cache it for the case when we need to sync requirements _PYPI_CACHE[self.package_name] = self._pypi_info return self._pypi_info @property def pypi_version(self): return self.pypi_info["info"]["version"] @property def pypi_license(self): if self._pypi_license is None: if self.pypi_info["info"]["license"]: self._pypi_license = self.pypi_info["info"]["license"] else: prefix = "License :: OSI Approved :: " classifiers = [c[len(prefix):] for c in self.pypi_info["info"]["classifiers"] if c.startswith(prefix)] self._pypi_license = "/".join(classifiers) self._license = self.LICENSE_MAP.get(self._pypi_license, self._pypi_license) if self._pypi_license == "UNKNOWN": self._pypi_license = None return self._license def __eq__(self, other): return (isinstance(other, PYPIPackage) and self.package_name == other.package_name) class Requirement(PYPIPackage): RE_NAME = re.compile(r"[a-zA-Z0-9-._]+") RE_CONST_VERSION = re.compile(r"==[a-zA-Z0-9.]+") RE_MIN_VERSION = re.compile(r">=?[a-zA-Z0-9.]+") RE_MAX_VERSION = re.compile(r"<=?[a-zA-Z0-9.]+") RE_NE_VERSIONS = re.compile(r"!=[a-zA-Z0-9.]+") def __init__(self, package_name, version): super(Requirement, self).__init__(package_name) self.version = version self.do_not_touch = False def sync_max_version_with_pypy(self): if isinstance(self.version, dict) and not self.do_not_touch: self.version["max"] = "<=%s" % self.pypi_version @classmethod def parse_line(cls, line): match = cls.RE_NAME.match(line) if match: name = match.group() versions = line.replace(name, "") versions = versions.split("#")[0] versions = versions.split(";")[0].strip() if not cls.RE_CONST_VERSION.match(versions): versions = versions.strip().split(",") min_version = None max_version = None ne_versions = [] for version in versions: if cls.RE_MIN_VERSION.match(version): if min_version: raise Exception("Found several min versions for " "%s package." % name) min_version = version elif cls.RE_MAX_VERSION.match(version): if max_version: raise Exception("Found several max versions for " "%s package." % name) max_version = version elif cls.RE_NE_VERSIONS.match(version): ne_versions.append(version) versions = {"min": min_version, "max": max_version, "ne": ne_versions} return cls(name, versions) def __str__(self): if isinstance(self.version, dict): version = [] min_equal_to_max = False if self.version["min"] and self.version["max"]: if ( self.version["min"].startswith(">=") and self.version["max"].startswith("<=") and self.version["min"][2:] == self.version["max"][2:] ): min_equal_to_max = True version.append("==%s" % self.version["min"][2:]) if not min_equal_to_max and self.version["min"]: version.append(self.version["min"]) if not min_equal_to_max and self.version["ne"]: version.extend(self.version["ne"]) if not min_equal_to_max and self.version["max"]: version.append(self.version["max"]) version = ",".join(version) else: if self.do_not_touch: version = self.version else: version = ">=%s" % self.version[2:] string = "%s%s" % (self.package_name, version) if self.pypi_license: # it to lower or greater value. magic_number = 55 if len(string) < magic_number: indent = magic_number - len(string) else: indent = 2 string += " " * indent + "# " + self.pypi_license return string def __eq__(self, other): return (isinstance(other, self.__class__) and self.package_name == other.package_name) def __ne__(self, other): return not self.__eq__(other) class UpperConstraint(PYPIPackage): RE_LINE = re.compile( r"(?P<package_name>[a-zA-Z0-9-._]+)===(?P<version>[a-zA-Z0-9.]+)") def __init__(self, package_name, version=None): super(UpperConstraint, self).__init__(package_name) self._version = version def __str__(self): return "%s===%s" % (self.package_name, self.version) @property def version(self): if self._version is None: self._version = self.pypi_version return self._version @classmethod def parse_line(cls, line): match = cls.RE_LINE.match(line) if match: return cls(**match.groupdict()) def update(self, version): self._version = version def parse_data(raw_data, include_comments=True, dependency_cls=Requirement): # first elem is None to simplify checks of last elem in requirements requirements = [None] for line in raw_data.split("\n"): if line.startswith("#"): if not include_comments: continue if getattr(requirements[-1], "is_finished", True): requirements.append(Comment()) requirements[-1].append(line) elif line == "": # just empty line if isinstance(requirements[-1], Comment): requirements[-1].finish_him() requirements.append(Comment(finished=True)) else: if (isinstance(requirements[-1], Comment) and not requirements[-1].is_finished): requirements[-1].finish_him() # parse_line dep = dependency_cls.parse_line(line) if dep: if (isinstance(requirements[-1], Comment) and DO_NOT_TOUCH_TAG in str(requirements[-1])): dep.do_not_touch = True requirements.append(dep) for i in range(len(requirements) - 1, 0, -1): # remove empty lines at the end of file if isinstance(requirements[i], Comment): if str(requirements[i]) == "": requirements.pop(i) else: break return collections.OrderedDict( (v if isinstance(v, Comment) else v.package_name, v) for v in requirements if v) def _fetch_from_gr(filename): for i in range(0, len(GLOBAL_REQUIREMENTS_LOCATIONS)): url = GLOBAL_REQUIREMENTS_LOCATIONS[i] + filename LOG.debug("Try to obtain %s from %s" % (filename, url)) try: return requests.get(url).text except requests.ConnectionError as e: LOG.exception(e) raise Exception("Unable to obtain %s" % filename) def _write_requirements(filename, requirements): if isinstance(requirements, dict): requirements = requirements.values() LOG.info("Saving requirements to %s." % filename) with open(filename, "w") as f: for entity in requirements: f.write(str(entity)) f.write("\n") def sync_requirements(): LOG.info("Obtaining global-requirements of OpenStack...") raw_gr = _fetch_from_gr("global-requirements.txt") # NOTE(andreykurilin): global-requirements includes comments which can be # unrelated to Rally project, so let's just ignore them gr = parse_data(raw_gr, include_comments=False) for file_name in RALLY_REQUIREMENTS_FILES: LOG.debug("Processing '%s'." % file_name) with open(file_name) as f: requirements = parse_data(f.read()) for name, req in requirements.items(): if isinstance(req, Requirement) and not req.do_not_touch: if name in gr: req.version = gr[req.package_name].version else: if isinstance(req.version, dict): req.version["max"] = None _write_requirements(file_name, requirements) def update_upper_constraints(): LOG.info("Obtaining upper-constrains from OpenStack...") raw_g_uc = _fetch_from_gr("upper-constraints.txt") # them. global_uc = parse_data(raw_g_uc, include_comments=False, dependency_cls=UpperConstraint) with open("upper-constraints.txt") as f: our_uc = parse_data(f.read(), dependency_cls=UpperConstraint) with open("requirements.txt") as f: our_requirements = parse_data(f.read(), include_comments=False) for name, req in our_requirements.items(): if isinstance(req, Comment): continue if name not in our_uc: our_uc[name] = UpperConstraint(name) if name in global_uc: # we cannot use whatever we want versions in CI. OpenStack CI # ignores versions listed in requirements of # particular project and use versions from global u-c file. # It means that we need to suggest to use the same versions our_uc[name].update(global_uc[name].version) our_uc = sorted(our_uc.values(), key=lambda o: o.package_name.upper()) _write_requirements("upper-constraints.txt", our_uc) def main(): sync_requirements() update_upper_constraints() if __name__ == "__main__": sys.exit(main())
true
true
f717c8b4305af7242c263a306f933192f7a5b976
491
py
Python
route/api_raw.py
susemeee/openNAMU
0f3d8acb5f0fe179cc8dbbfea1846da3ca0cd4e1
[ "BSD-3-Clause" ]
null
null
null
route/api_raw.py
susemeee/openNAMU
0f3d8acb5f0fe179cc8dbbfea1846da3ca0cd4e1
[ "BSD-3-Clause" ]
null
null
null
route/api_raw.py
susemeee/openNAMU
0f3d8acb5f0fe179cc8dbbfea1846da3ca0cd4e1
[ "BSD-3-Clause" ]
1
2020-01-04T09:43:35.000Z
2020-01-04T09:43:35.000Z
from .tool.func import * def api_raw_2(conn, name): curs = conn.cursor() if acl_check(name, 'render') != 1: curs.execute("select data from data where title = ?", [name]) data = curs.fetchall() if data: json_data = { "title" : name, "data" : render_set(title = name, data = data[0][0], s_data = 1) } return flask.jsonify(json_data) else: return flask.jsonify({}) else: return flask.jsonify({})
30.6875
108
0.545825
from .tool.func import * def api_raw_2(conn, name): curs = conn.cursor() if acl_check(name, 'render') != 1: curs.execute("select data from data where title = ?", [name]) data = curs.fetchall() if data: json_data = { "title" : name, "data" : render_set(title = name, data = data[0][0], s_data = 1) } return flask.jsonify(json_data) else: return flask.jsonify({}) else: return flask.jsonify({})
true
true
f717ca0d2a58ad1a206f65fb4972fcfb0f68141c
634
py
Python
rates/migrations/0003_rating.py
Gichimu/irater
2da90230368b2aa5b97b9ab84455703e510e51ef
[ "MIT" ]
null
null
null
rates/migrations/0003_rating.py
Gichimu/irater
2da90230368b2aa5b97b9ab84455703e510e51ef
[ "MIT" ]
16
2020-01-11T08:20:30.000Z
2022-02-10T09:18:48.000Z
rates/migrations/0003_rating.py
Gichimu/irater
2da90230368b2aa5b97b9ab84455703e510e51ef
[ "MIT" ]
null
null
null
# Generated by Django 2.2 on 2020-01-15 06:05 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('rates', '0002_auto_20200113_0757'), ] operations = [ migrations.CreateModel( name='Rating', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('usability', models.IntegerField(default=0)), ('design', models.IntegerField(default=0)), ('content', models.IntegerField(default=0)), ], ), ]
27.565217
114
0.572555
from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('rates', '0002_auto_20200113_0757'), ] operations = [ migrations.CreateModel( name='Rating', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('usability', models.IntegerField(default=0)), ('design', models.IntegerField(default=0)), ('content', models.IntegerField(default=0)), ], ), ]
true
true
f717cc90445df76a8ce5da7e005a9d1637cd561b
857
py
Python
setup.py
treamology/sphinx-variations
1f02e3bf32610ef32cf24748d4f91bab0263c173
[ "MIT" ]
null
null
null
setup.py
treamology/sphinx-variations
1f02e3bf32610ef32cf24748d4f91bab0263c173
[ "MIT" ]
2
2019-10-14T12:17:54.000Z
2019-11-29T09:06:54.000Z
setup.py
treamology/sphinx-variations
1f02e3bf32610ef32cf24748d4f91bab0263c173
[ "MIT" ]
1
2019-10-16T10:45:39.000Z
2019-10-16T10:45:39.000Z
import setuptools from variations import __version__ with open('README.md', 'r') as fh: long_description = fh.read() setuptools.setup( name='sphinx-variations', version=__version__, author='D. Lawrence', author_email='trea@treamous.com', description='Create multiple variations of your Sphinx HTML documentation.', long_description=long_description, long_description_content_type='text/markdown', url='https://github.com/treamology/sphinx-variations', packages=['variations'], classifiers=[ 'Programming Language :: Python :: 3', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Framework :: Sphinx', 'Topic :: Documentation', 'Topic :: Software Development :: Documentation', ], install_requires=[ 'sphinx' ], )
30.607143
80
0.662777
import setuptools from variations import __version__ with open('README.md', 'r') as fh: long_description = fh.read() setuptools.setup( name='sphinx-variations', version=__version__, author='D. Lawrence', author_email='trea@treamous.com', description='Create multiple variations of your Sphinx HTML documentation.', long_description=long_description, long_description_content_type='text/markdown', url='https://github.com/treamology/sphinx-variations', packages=['variations'], classifiers=[ 'Programming Language :: Python :: 3', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Framework :: Sphinx', 'Topic :: Documentation', 'Topic :: Software Development :: Documentation', ], install_requires=[ 'sphinx' ], )
true
true
f717cd31d8d75e9ba8cc375f6e83f9427f97fd2e
1,947
py
Python
apps/sumo/tests/test_json_decorator.py
storagebot/kitsune
613ba2ca09104f330ab77088b452391169096249
[ "BSD-3-Clause" ]
2
2019-08-19T17:08:47.000Z
2019-10-05T11:37:02.000Z
apps/sumo/tests/test_json_decorator.py
taliasman/kitsune
f8085205eef143011adb4c52d1f183da06c1c58e
[ "BSD-3-Clause" ]
null
null
null
apps/sumo/tests/test_json_decorator.py
taliasman/kitsune
f8085205eef143011adb4c52d1f183da06c1c58e
[ "BSD-3-Clause" ]
null
null
null
import json from django import http from django.core.exceptions import PermissionDenied from django.test import RequestFactory, TestCase from nose.tools import eq_ from sumo.decorators import json_view rf = RequestFactory() JSON = 'application/json' class JsonViewTests(TestCase): def test_object(self): data = { 'foo': 'bar', 'baz': 'qux', 'quz': [{'foo': 'bar'},], } expect = json.dumps(data) @json_view def temp(req): return data res = temp(rf.get('/')) eq_(200, res.status_code) eq_(expect, res.content) eq_(JSON, res['content-type']) def test_list(self): data = ['foo', 'bar', 'baz'] expect = json.dumps(data) @json_view def temp(req): return data res = temp(rf.get('/')) eq_(200, res.status_code) eq_(expect, res.content) eq_(JSON, res['content-type']) def test_404(self): @json_view def temp(req): raise http.Http404, 'foo' res = temp(rf.get('/')) eq_(404, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(404, data['error']) eq_('foo', data['message']) def test_permission(self): @json_view def temp(req): raise PermissionDenied, 'bar' res = temp(rf.get('/')) eq_(403, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(403, data['error']) eq_('bar', data['message']) def test_server_error(self): @json_view def temp(req): raise TypeError, 'fail' res = temp(rf.get('/')) eq_(500, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(500, data['error']) eq_('fail', data['message'])
24.037037
51
0.539291
import json from django import http from django.core.exceptions import PermissionDenied from django.test import RequestFactory, TestCase from nose.tools import eq_ from sumo.decorators import json_view rf = RequestFactory() JSON = 'application/json' class JsonViewTests(TestCase): def test_object(self): data = { 'foo': 'bar', 'baz': 'qux', 'quz': [{'foo': 'bar'},], } expect = json.dumps(data) @json_view def temp(req): return data res = temp(rf.get('/')) eq_(200, res.status_code) eq_(expect, res.content) eq_(JSON, res['content-type']) def test_list(self): data = ['foo', 'bar', 'baz'] expect = json.dumps(data) @json_view def temp(req): return data res = temp(rf.get('/')) eq_(200, res.status_code) eq_(expect, res.content) eq_(JSON, res['content-type']) def test_404(self): @json_view def temp(req): raise http.Http404, 'foo' res = temp(rf.get('/')) eq_(404, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(404, data['error']) eq_('foo', data['message']) def test_permission(self): @json_view def temp(req): raise PermissionDenied, 'bar' res = temp(rf.get('/')) eq_(403, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(403, data['error']) eq_('bar', data['message']) def test_server_error(self): @json_view def temp(req): raise TypeError, 'fail' res = temp(rf.get('/')) eq_(500, res.status_code) eq_(JSON, res['content-type']) data = json.loads(res.content) eq_(500, data['error']) eq_('fail', data['message'])
false
true
f717cd7177718b946f4674840996c1bb04bd9abb
7,798
py
Python
frame_finder.py
myddrin/frame_finder
88f8002157ed7d37e3c1d8f24e98be4f51198578
[ "MIT" ]
null
null
null
frame_finder.py
myddrin/frame_finder
88f8002157ed7d37e3c1d8f24e98be4f51198578
[ "MIT" ]
2
2016-04-04T19:26:59.000Z
2016-05-25T12:40:50.000Z
frame_finder.py
myddrin/frame_finder
88f8002157ed7d37e3c1d8f24e98be4f51198578
[ "MIT" ]
null
null
null
#!/usr/bin/python """ This python script was created to find missing files in given folders. Copyright (c) 2016 Thomas Richard Following MIT license (see copying.txt) The software is provided "as is", without warranty of any kind, express or implied, including but not limited to the warranties of merchantability, fitness for a particular purpose and noninfringement. """ import shutil import os import glob import re import argparse def touch(path): """ Create a file """ open(path, 'w').close() class FrameFinder: """ Object to check that files following a sequence of numbers exist """ class Operations: NONE, DUPLICATE, TOUCH = range(3) # NONE: do nothin besides printing when not finding a file # DUPLICATE: When not finding a file duplicate the last found one # TOUCH: When not finding a file create a new empty file def __init__(self, prefix="image", extension="exr", zfill=4): # local variable as not used after creation base_regexp = "([0-9]*)\." """base_regexp: to allow update in case prefix or extension changed""" self.prefix = prefix """filename prefix""" self.extension = extension """filename extension""" self.reg_exp = self.prefix + base_regexp + self.extension self.verbose = False self.zfill = 4 """number of digits to use to create a name with leading 0s""" def get_filename(self, number=0): """ Return filename based on prefix, number zfill and extension """ return (self.prefix + str(number).zfill(self.zfill) + "." + self.extension) def find_missing(self, basedir="."): """ Return a list of missing files in basedir by checking the range from existing files """ found_numbers = list() missing_files = list() print "searching in " + basedir for filename in glob.glob(basedir + "/*." + self.extension): found = re.findall(self.reg_exp, filename) if self.verbose: print "found " + filename + " found: " + str(found) if len(found) > 0: num = found[0] else: num = "" if num.isdigit(): found_numbers.append(int(num)) else: if self.verbose: print ("Could not find a digit in filename '" + filename + "' using '" + reg_exp + "'") continue # next frame if len(found_numbers) == 0: print "Did not find any files!" return missing_files found_numbers = sorted(found_numbers) prev = found_numbers[0] for num in found_numbers: while prev+1 < num: missing = prev+1 filename = self.get_filename(missing) missing_files.append(filename) prev = missing prev = num return missing_files def apply(self, basedir, missing_files, ope): """ Given a list of missing file will apply the given option using duplicate and createEmpty options Default operation is NONE """ for filename in missing_files: if ope == FrameFinder.Operations.DUPLICATE: found = re.findall(self.reg_exp, filename) if len(found) > 0: num = found[0] else: raise ValueError("regexp '" + reg_exp + "' failed to find 1 group in '" + str(filename) + "'") if num.isdigit(): prev = int(num) - 1 else: raise ValueError("could not find number in filename '" + str(filename) + "' given as input") prev_filename = self.get_filename(prev) print "copying " + prev_filename + " to " + filename shutil.copyfile(os.path.join(basedir, prev_filename), os.path.join(basedir, filename)) elif ope == FrameFinder.Operations.TOUCH: print "touch " + filename touch(os.path.join(basedir, filename)) elif ope == FrameFinder.Operations.NONE: # operation has to be NONE print "missing file " + filename else: raise ValueError("Invalid operation '" + ope + "' given") def verify(self, basedir, ope=Operations.NONE): """ Utility method to verify a directory """ missing_files = self.find_missing(basedir) self.apply(basedir, missing_files, ope) if __name__ == "__main__": """ Verify that a folder contains all images """ parser = argparse.ArgumentParser("frame_finder.py") finder = FrameFinder() # to have default values # long options parser.add_argument("--prefix", type=str, dest="prefix", default=finder.prefix, help="filename prefix - default: " + finder.prefix) parser.add_argument("--extension", type=str, dest="extension", default=finder.extension, help=("file extension without the dot - default: " + finder.extension)) parser.add_argument("--regexp", type=str, dest="reg_exp", default="", help=("regular expression to find the file. The " "regular expression is updated if filename " "prefix or extension is changed or can be " "specified here directly - default " + finder.reg_exp)) # short options parser.add_argument("-v", "--verbose", action="store_true", dest="verbose", help="verbose printing") parser.add_argument("-d", "--duplicate", action="store_true", dest="duplicate", help=("Duplicate the previous frame when not found " "(has priority over other options)")) parser.add_argument("-t", "--touch", action="store_true", dest="touch", help=("Create empty file when not found using prefix " "and extension")) parser.add_argument("-z", "--zerofill", type=int, dest="zfill", default=finder.zfill, help=("Number of digits to use with zero fill (with " "z=3 converts 40 to 040) - default " + str(finder.zfill))) # remaining parameters are directories to search parser.add_argument("dir", nargs="+", type=str, help="Verify files in a given directory") args = parser.parse_args() finder = FrameFinder(args.prefix, args.extension, args.zfill) if args.reg_exp != "": finder.reg_exp = args.reg_exp finder.verbose = args.verbose operation = FrameFinder.Operations.NONE if args.duplicate: operation = FrameFinder.Operations.DUPLICATE elif args.touch: operation = FrameFinder.Operations.TOUCH print "prefix: " + finder.prefix print "extension: " + finder.extension print "no digits: " + str(finder.zfill) print "" for directory in args.dir: if os.path.isdir(directory): finder.verify(directory, operation) print "" # empty line between dirs else: print directory + " is not a directory"
36.439252
78
0.543473
""" This python script was created to find missing files in given folders. Copyright (c) 2016 Thomas Richard Following MIT license (see copying.txt) The software is provided "as is", without warranty of any kind, express or implied, including but not limited to the warranties of merchantability, fitness for a particular purpose and noninfringement. """ import shutil import os import glob import re import argparse def touch(path): """ Create a file """ open(path, 'w').close() class FrameFinder: """ Object to check that files following a sequence of numbers exist """ class Operations: NONE, DUPLICATE, TOUCH = range(3) def __init__(self, prefix="image", extension="exr", zfill=4): base_regexp = "([0-9]*)\." """base_regexp: to allow update in case prefix or extension changed""" self.prefix = prefix """filename prefix""" self.extension = extension """filename extension""" self.reg_exp = self.prefix + base_regexp + self.extension self.verbose = False self.zfill = 4 """number of digits to use to create a name with leading 0s""" def get_filename(self, number=0): """ Return filename based on prefix, number zfill and extension """ return (self.prefix + str(number).zfill(self.zfill) + "." + self.extension) def find_missing(self, basedir="."): """ Return a list of missing files in basedir by checking the range from existing files """ found_numbers = list() missing_files = list() print "searching in " + basedir for filename in glob.glob(basedir + "/*." + self.extension): found = re.findall(self.reg_exp, filename) if self.verbose: print "found " + filename + " found: " + str(found) if len(found) > 0: num = found[0] else: num = "" if num.isdigit(): found_numbers.append(int(num)) else: if self.verbose: print ("Could not find a digit in filename '" + filename + "' using '" + reg_exp + "'") continue if len(found_numbers) == 0: print "Did not find any files!" return missing_files found_numbers = sorted(found_numbers) prev = found_numbers[0] for num in found_numbers: while prev+1 < num: missing = prev+1 filename = self.get_filename(missing) missing_files.append(filename) prev = missing prev = num return missing_files def apply(self, basedir, missing_files, ope): """ Given a list of missing file will apply the given option using duplicate and createEmpty options Default operation is NONE """ for filename in missing_files: if ope == FrameFinder.Operations.DUPLICATE: found = re.findall(self.reg_exp, filename) if len(found) > 0: num = found[0] else: raise ValueError("regexp '" + reg_exp + "' failed to find 1 group in '" + str(filename) + "'") if num.isdigit(): prev = int(num) - 1 else: raise ValueError("could not find number in filename '" + str(filename) + "' given as input") prev_filename = self.get_filename(prev) print "copying " + prev_filename + " to " + filename shutil.copyfile(os.path.join(basedir, prev_filename), os.path.join(basedir, filename)) elif ope == FrameFinder.Operations.TOUCH: print "touch " + filename touch(os.path.join(basedir, filename)) elif ope == FrameFinder.Operations.NONE: print "missing file " + filename else: raise ValueError("Invalid operation '" + ope + "' given") def verify(self, basedir, ope=Operations.NONE): """ Utility method to verify a directory """ missing_files = self.find_missing(basedir) self.apply(basedir, missing_files, ope) if __name__ == "__main__": """ Verify that a folder contains all images """ parser = argparse.ArgumentParser("frame_finder.py") finder = FrameFinder() parser.add_argument("--prefix", type=str, dest="prefix", default=finder.prefix, help="filename prefix - default: " + finder.prefix) parser.add_argument("--extension", type=str, dest="extension", default=finder.extension, help=("file extension without the dot - default: " + finder.extension)) parser.add_argument("--regexp", type=str, dest="reg_exp", default="", help=("regular expression to find the file. The " "regular expression is updated if filename " "prefix or extension is changed or can be " "specified here directly - default " + finder.reg_exp)) parser.add_argument("-v", "--verbose", action="store_true", dest="verbose", help="verbose printing") parser.add_argument("-d", "--duplicate", action="store_true", dest="duplicate", help=("Duplicate the previous frame when not found " "(has priority over other options)")) parser.add_argument("-t", "--touch", action="store_true", dest="touch", help=("Create empty file when not found using prefix " "and extension")) parser.add_argument("-z", "--zerofill", type=int, dest="zfill", default=finder.zfill, help=("Number of digits to use with zero fill (with " "z=3 converts 40 to 040) - default " + str(finder.zfill))) parser.add_argument("dir", nargs="+", type=str, help="Verify files in a given directory") args = parser.parse_args() finder = FrameFinder(args.prefix, args.extension, args.zfill) if args.reg_exp != "": finder.reg_exp = args.reg_exp finder.verbose = args.verbose operation = FrameFinder.Operations.NONE if args.duplicate: operation = FrameFinder.Operations.DUPLICATE elif args.touch: operation = FrameFinder.Operations.TOUCH print "prefix: " + finder.prefix print "extension: " + finder.extension print "no digits: " + str(finder.zfill) print "" for directory in args.dir: if os.path.isdir(directory): finder.verify(directory, operation) print "" else: print directory + " is not a directory"
false
true
f717cdb73df096844f350cf4219a8fcdfd000791
1,017
py
Python
lib/misc.py
ngreenberg/twitch-plays-wikipedia
ee72ce2605eed77fb13a89678bd4ba647d83ff9f
[ "MIT" ]
null
null
null
lib/misc.py
ngreenberg/twitch-plays-wikipedia
ee72ce2605eed77fb13a89678bd4ba647d83ff9f
[ "MIT" ]
null
null
null
lib/misc.py
ngreenberg/twitch-plays-wikipedia
ee72ce2605eed77fb13a89678bd4ba647d83ff9f
[ "MIT" ]
null
null
null
import time import os def pp(message, mtype='INFO'): mtype = mtype.upper() print '[%s] [%s] %s' % (time.strftime('%H:%M:%S', time.gmtime()), mtype, message) def ppi(channel, message, username): print '[%s %s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), channel, username.lower(), message) def pbot(message, channel=''): if channel: msg = '[%s %s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), channel, 'BOT', message) else: msg = '[%s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), 'BOT', message) print msg def pbutton(message_buffer): os.system('cls' if os.name == 'nt' else 'clear') print '\n\n' print '\n'.join([' {0:<12s} {1:>6s}' .format(message['username'][:12].title(), message['button'].lower()) for message in message_buffer])
33.9
76
0.462144
import time import os def pp(message, mtype='INFO'): mtype = mtype.upper() print '[%s] [%s] %s' % (time.strftime('%H:%M:%S', time.gmtime()), mtype, message) def ppi(channel, message, username): print '[%s %s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), channel, username.lower(), message) def pbot(message, channel=''): if channel: msg = '[%s %s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), channel, 'BOT', message) else: msg = '[%s] <%s> %s' % (time.strftime('%H:%M:%S', time.gmtime()), 'BOT', message) print msg def pbutton(message_buffer): os.system('cls' if os.name == 'nt' else 'clear') print '\n\n' print '\n'.join([' {0:<12s} {1:>6s}' .format(message['username'][:12].title(), message['button'].lower()) for message in message_buffer])
false
true
f717ce0bca4aa3ca6519610187210bdcf31b9d6a
2,934
py
Python
musicplayer/buttons.py
Gabe-Corral/Music_Player
9352266aa86e5927e4a603b6238cea4745cc6b26
[ "MIT" ]
null
null
null
musicplayer/buttons.py
Gabe-Corral/Music_Player
9352266aa86e5927e4a603b6238cea4745cc6b26
[ "MIT" ]
null
null
null
musicplayer/buttons.py
Gabe-Corral/Music_Player
9352266aa86e5927e4a603b6238cea4745cc6b26
[ "MIT" ]
null
null
null
import tkinter as tk from PIL import ImageTk, Image from file_import import FileImport class Buttons: def __init__(self, parent, player): self.player = player self.parent = parent #clean these up unskip_img = ImageTk.PhotoImage(Image.open("assets/unskip.png").resize((25,25))) unskip_btn = tk.Button(parent, image=unskip_img, command=self.player.play_previous) unskip_btn.image = unskip_img unskip_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) unskip_btn.grid(column=0, row=0) self.pause_img = ImageTk.PhotoImage(Image.open("assets/pause.png").resize((25,25))) self.play_img = ImageTk.PhotoImage(Image.open("assets/play.png").resize((25,25))) self.play_btn = tk.Button(parent, image=self.pause_img, borderwidth=0, bg='#282828', command=self.pause_play) self.play_btn.image = self.pause_img self.play_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) self.play_btn.grid(column=1, row=0) skip_img = ImageTk.PhotoImage(Image.open("assets/skip.png").resize((25,25))) skip_btn = tk.Button(parent, image=skip_img, command=self.player.play_next) skip_btn.image = skip_img skip_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) skip_btn.grid(column=2, row=0) volume_down_img = ImageTk.PhotoImage(Image.open("assets/volume_down.png").resize((25,25))) volume_up_btn = tk.Button(parent, image=volume_down_img, command=self.player.volume_control) volume_up_btn.image = volume_down_img volume_up_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) volume_up_btn.grid(column=3, row=0) volume_up_img = ImageTk.PhotoImage(Image.open("assets/volume_up.png").resize((25,25))) volume_down_btn = tk.Button(parent, image=volume_up_img, command=lambda: self.player.volume_control(True)) volume_down_btn.image = volume_up_img volume_down_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) volume_down_btn.grid(column=4, row=0) import_btn = tk.Button(parent, text="Import", command=self.import_new_files) import_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) import_btn.place(x=970, y=0) def import_new_files(self): importer = FileImport(self.parent, tk.Toplevel(self.parent)) def pause_play(self): if self.player.is_playing: self.play_btn.config(image=self.play_img) else: self.play_btn.config(image=self.pause_img) self.player.pause_resume()
44.454545
98
0.673142
import tkinter as tk from PIL import ImageTk, Image from file_import import FileImport class Buttons: def __init__(self, parent, player): self.player = player self.parent = parent unskip_img = ImageTk.PhotoImage(Image.open("assets/unskip.png").resize((25,25))) unskip_btn = tk.Button(parent, image=unskip_img, command=self.player.play_previous) unskip_btn.image = unskip_img unskip_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) unskip_btn.grid(column=0, row=0) self.pause_img = ImageTk.PhotoImage(Image.open("assets/pause.png").resize((25,25))) self.play_img = ImageTk.PhotoImage(Image.open("assets/play.png").resize((25,25))) self.play_btn = tk.Button(parent, image=self.pause_img, borderwidth=0, bg='#282828', command=self.pause_play) self.play_btn.image = self.pause_img self.play_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) self.play_btn.grid(column=1, row=0) skip_img = ImageTk.PhotoImage(Image.open("assets/skip.png").resize((25,25))) skip_btn = tk.Button(parent, image=skip_img, command=self.player.play_next) skip_btn.image = skip_img skip_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) skip_btn.grid(column=2, row=0) volume_down_img = ImageTk.PhotoImage(Image.open("assets/volume_down.png").resize((25,25))) volume_up_btn = tk.Button(parent, image=volume_down_img, command=self.player.volume_control) volume_up_btn.image = volume_down_img volume_up_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) volume_up_btn.grid(column=3, row=0) volume_up_img = ImageTk.PhotoImage(Image.open("assets/volume_up.png").resize((25,25))) volume_down_btn = tk.Button(parent, image=volume_up_img, command=lambda: self.player.volume_control(True)) volume_down_btn.image = volume_up_img volume_down_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) volume_down_btn.grid(column=4, row=0) import_btn = tk.Button(parent, text="Import", command=self.import_new_files) import_btn.config(highlightbackground="#282828", highlightcolor="#282828", bg='#282828', borderwidth=0) import_btn.place(x=970, y=0) def import_new_files(self): importer = FileImport(self.parent, tk.Toplevel(self.parent)) def pause_play(self): if self.player.is_playing: self.play_btn.config(image=self.play_img) else: self.play_btn.config(image=self.pause_img) self.player.pause_resume()
true
true
f717ce9240361f4597c503f589f1095a203484a7
20,067
py
Python
tests/wallet/did_wallet/test_did.py
Starcoder0x/chia-blockchain
5b2f5772780d0370f76b0134db6a7fdc7af42862
[ "Apache-2.0" ]
1
2022-03-04T13:12:39.000Z
2022-03-04T13:12:39.000Z
tests/wallet/did_wallet/test_did.py
zcomputerwiz/experiments-blockchain
841754b44494451a9e3e537575eeec431fe533d1
[ "Apache-2.0" ]
null
null
null
tests/wallet/did_wallet/test_did.py
zcomputerwiz/experiments-blockchain
841754b44494451a9e3e537575eeec431fe533d1
[ "Apache-2.0" ]
null
null
null
import asyncio import pytest import pytest_asyncio from chia.simulator.simulator_protocol import FarmNewBlockProtocol from chia.types.peer_info import PeerInfo from chia.util.ints import uint16, uint32, uint64 from tests.setup_nodes import self_hostname, setup_simulators_and_wallets from chia.wallet.did_wallet.did_wallet import DIDWallet from chia.types.blockchain_format.program import Program from blspy import AugSchemeMPL from chia.types.spend_bundle import SpendBundle from chia.consensus.block_rewards import calculate_pool_reward, calculate_base_farmer_reward from tests.time_out_assert import time_out_assert, time_out_assert_not_none pytestmark = pytest.mark.skip("TODO: Fix tests") @pytest.fixture(scope="module") def event_loop(): loop = asyncio.get_event_loop() yield loop class TestDIDWallet: @pytest_asyncio.fixture(scope="function") async def wallet_node(self): async for _ in setup_simulators_and_wallets(1, 1, {}): yield _ @pytest_asyncio.fixture(scope="function") async def two_wallet_nodes(self): async for _ in setup_simulators_and_wallets(1, 2, {}): yield _ @pytest_asyncio.fixture(scope="function") async def three_wallet_nodes(self): async for _ in setup_simulators_and_wallets(1, 3, {}): yield _ @pytest_asyncio.fixture(scope="function") async def two_wallet_nodes_five_freeze(self): async for _ in setup_simulators_and_wallets(1, 2, {}): yield _ @pytest_asyncio.fixture(scope="function") async def three_sim_two_wallets(self): async for _ in setup_simulators_and_wallets(3, 2, {}): yield _ @pytest.mark.asyncio async def test_creation_from_backup_file(self, three_wallet_nodes): num_blocks = 5 full_nodes, wallets = three_wallet_nodes full_node_api = full_nodes[0] full_node_server = full_node_api.server wallet_node_0, server_0 = wallets[0] wallet_node_1, server_1 = wallets[1] wallet_node_2, server_2 = wallets[2] wallet_0 = wallet_node_0.wallet_state_manager.main_wallet wallet_1 = wallet_node_1.wallet_state_manager.main_wallet wallet_2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet_0.get_new_puzzlehash() ph1 = await wallet_1.get_new_puzzlehash() ph2 = await wallet_2.get_new_puzzlehash() await server_0.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) await server_1.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) await server_2.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(10, wallet_0.get_unconfirmed_balance, funds) await time_out_assert(10, wallet_0.get_confirmed_balance, funds) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph1)) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) # Wallet1 sets up DIDWallet1 without any backup set async with wallet_node_0.wallet_state_manager.lock: did_wallet_0: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_0.wallet_state_manager, wallet_0, uint64(101) ) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_0.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_0.get_unconfirmed_balance, 101) await time_out_assert(15, did_wallet_0.get_pending_change_balance, 0) # Wallet1 sets up DIDWallet_1 with DIDWallet_0 as backup backup_ids = [bytes.fromhex(did_wallet_0.get_my_DID())] async with wallet_node_1.wallet_state_manager.lock: did_wallet_1: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_1.wallet_state_manager, wallet_1, uint64(201), backup_ids ) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_1.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_1.get_unconfirmed_balance, 201) await time_out_assert(15, did_wallet_1.get_pending_change_balance, 0) filename = "test.backup" did_wallet_1.create_backup(filename) # Wallet2 recovers DIDWallet2 to a new set of keys async with wallet_node_2.wallet_state_manager.lock: did_wallet_2 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node_2.wallet_state_manager, wallet_2, filename ) coins = await did_wallet_1.select_coins(1) coin = coins.copy().pop() assert did_wallet_2.did_info.temp_coin == coin newpuzhash = await did_wallet_2.get_new_inner_hash() pubkey = bytes( (await did_wallet_2.wallet_state_manager.get_unused_derivation_record(did_wallet_2.wallet_info.id)).pubkey ) message_spend_bundle = await did_wallet_0.create_attestment( did_wallet_2.did_info.temp_coin.name(), newpuzhash, pubkey, "test.attest" ) print(f"pubkey: {pubkey}") for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_2.load_attest_files_for_recovery_spend(["test.attest"]) assert message_spend_bundle == test_message_spend_bundle await did_wallet_2.recovery_spend( did_wallet_2.did_info.temp_coin, newpuzhash, test_info_list, pubkey, test_message_spend_bundle, ) print(f"pubkey: {did_wallet_2}") for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(45, did_wallet_2.get_confirmed_balance, 201) await time_out_assert(45, did_wallet_2.get_unconfirmed_balance, 201) some_ph = 32 * b"\2" await did_wallet_2.create_exit_spend(some_ph) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) async def get_coins_with_ph(): coins = await full_node_api.full_node.coin_store.get_coin_records_by_puzzle_hash(True, some_ph) if len(coins) == 1: return True return False await time_out_assert(15, get_coins_with_ph, True) await time_out_assert(45, did_wallet_2.get_confirmed_balance, 0) await time_out_assert(45, did_wallet_2.get_unconfirmed_balance, 0) @pytest.mark.asyncio async def test_did_recovery_with_multiple_backup_dids(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet wallet2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) ph = await wallet2.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) recovery_list = [bytes.fromhex(did_wallet.get_my_DID())] async with wallet_node_2.wallet_state_manager.lock: did_wallet_2: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(101), recovery_list ) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_2.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_2.get_unconfirmed_balance, 101) assert did_wallet_2.did_info.backup_ids == recovery_list recovery_list.append(bytes.fromhex(did_wallet_2.get_my_DID())) async with wallet_node_2.wallet_state_manager.lock: did_wallet_3: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(201), recovery_list ) ph2 = await wallet.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) assert did_wallet_3.did_info.backup_ids == recovery_list await time_out_assert(15, did_wallet_3.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 201) coins = await did_wallet_3.select_coins(1) coin = coins.pop() filename = "test.backup" did_wallet_3.create_backup(filename) async with wallet_node.wallet_state_manager.lock: did_wallet_4 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node.wallet_state_manager, wallet, filename, ) pubkey = ( await did_wallet_4.wallet_state_manager.get_unused_derivation_record(did_wallet_2.wallet_info.id) ).pubkey new_ph = await did_wallet_4.get_new_inner_hash() message_spend_bundle = await did_wallet.create_attestment(coin.name(), new_ph, pubkey, "test1.attest") message_spend_bundle2 = await did_wallet_2.create_attestment(coin.name(), new_ph, pubkey, "test2.attest") message_spend_bundle = message_spend_bundle.aggregate([message_spend_bundle, message_spend_bundle2]) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_4.load_attest_files_for_recovery_spend(["test1.attest", "test2.attest"]) assert message_spend_bundle == test_message_spend_bundle for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await did_wallet_4.recovery_spend(coin, new_ph, test_info_list, pubkey, message_spend_bundle) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet_4.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_4.get_unconfirmed_balance, 201) await time_out_assert(15, did_wallet_3.get_confirmed_balance, 0) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 0) @pytest.mark.asyncio async def test_did_recovery_with_empty_set(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) coins = await did_wallet.select_coins(1) coin = coins.pop() info = Program.to([]) pubkey = (await did_wallet.wallet_state_manager.get_unused_derivation_record(did_wallet.wallet_info.id)).pubkey spend_bundle = await did_wallet.recovery_spend( coin, ph, info, pubkey, SpendBundle([], AugSchemeMPL.aggregate([])) ) additions = spend_bundle.additions() assert additions == [] @pytest.mark.asyncio async def test_did_attest_after_recovery(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet wallet2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) ph2 = await wallet2.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) recovery_list = [bytes.fromhex(did_wallet.get_my_DID())] async with wallet_node_2.wallet_state_manager.lock: did_wallet_2: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(101), recovery_list ) ph = await wallet.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_2.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_2.get_unconfirmed_balance, 101) assert did_wallet_2.did_info.backup_ids == recovery_list # Update coin with new ID info recovery_list = [bytes.fromhex(did_wallet_2.get_my_DID())] await did_wallet.update_recovery_list(recovery_list, uint64(1)) assert did_wallet.did_info.backup_ids == recovery_list await did_wallet.create_update_spend() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) # DID Wallet 2 recovers into DID Wallet 3 with new innerpuz filename = "test.backup" did_wallet_2.create_backup(filename) async with wallet_node.wallet_state_manager.lock: did_wallet_3 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node.wallet_state_manager, wallet, filename, ) new_ph = await did_wallet_3.get_new_inner_hash() coins = await did_wallet_2.select_coins(1) coin = coins.pop() pubkey = ( await did_wallet_3.wallet_state_manager.get_unused_derivation_record(did_wallet_3.wallet_info.id) ).pubkey message_spend_bundle = await did_wallet.create_attestment(coin.name(), new_ph, pubkey, "test.attest") for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) ( info, message_spend_bundle, ) = await did_wallet_3.load_attest_files_for_recovery_spend(["test.attest"]) await did_wallet_3.recovery_spend(coin, new_ph, info, pubkey, message_spend_bundle) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_3.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 101) # DID Wallet 1 recovery spends into DID Wallet 4 filename = "test.backup" did_wallet.create_backup(filename) async with wallet_node_2.wallet_state_manager.lock: did_wallet_4 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node_2.wallet_state_manager, wallet2, filename, ) coins = await did_wallet.select_coins(1) coin = coins.pop() new_ph = await did_wallet_4.get_new_inner_hash() pubkey = ( await did_wallet_4.wallet_state_manager.get_unused_derivation_record(did_wallet_4.wallet_info.id) ).pubkey await did_wallet_3.create_attestment(coin.name(), new_ph, pubkey, "test.attest") for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_4.load_attest_files_for_recovery_spend(["test.attest"]) spend_bundle = await did_wallet_4.recovery_spend( coin, new_ph, test_info_list, pubkey, test_message_spend_bundle ) await time_out_assert_not_none(15, full_node_1.full_node.mempool_manager.get_spendbundle, spend_bundle.name()) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_4.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_4.get_unconfirmed_balance, 101) await time_out_assert(15, did_wallet.get_confirmed_balance, 0) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 0)
43.623913
119
0.695819
import asyncio import pytest import pytest_asyncio from chia.simulator.simulator_protocol import FarmNewBlockProtocol from chia.types.peer_info import PeerInfo from chia.util.ints import uint16, uint32, uint64 from tests.setup_nodes import self_hostname, setup_simulators_and_wallets from chia.wallet.did_wallet.did_wallet import DIDWallet from chia.types.blockchain_format.program import Program from blspy import AugSchemeMPL from chia.types.spend_bundle import SpendBundle from chia.consensus.block_rewards import calculate_pool_reward, calculate_base_farmer_reward from tests.time_out_assert import time_out_assert, time_out_assert_not_none pytestmark = pytest.mark.skip("TODO: Fix tests") @pytest.fixture(scope="module") def event_loop(): loop = asyncio.get_event_loop() yield loop class TestDIDWallet: @pytest_asyncio.fixture(scope="function") async def wallet_node(self): async for _ in setup_simulators_and_wallets(1, 1, {}): yield _ @pytest_asyncio.fixture(scope="function") async def two_wallet_nodes(self): async for _ in setup_simulators_and_wallets(1, 2, {}): yield _ @pytest_asyncio.fixture(scope="function") async def three_wallet_nodes(self): async for _ in setup_simulators_and_wallets(1, 3, {}): yield _ @pytest_asyncio.fixture(scope="function") async def two_wallet_nodes_five_freeze(self): async for _ in setup_simulators_and_wallets(1, 2, {}): yield _ @pytest_asyncio.fixture(scope="function") async def three_sim_two_wallets(self): async for _ in setup_simulators_and_wallets(3, 2, {}): yield _ @pytest.mark.asyncio async def test_creation_from_backup_file(self, three_wallet_nodes): num_blocks = 5 full_nodes, wallets = three_wallet_nodes full_node_api = full_nodes[0] full_node_server = full_node_api.server wallet_node_0, server_0 = wallets[0] wallet_node_1, server_1 = wallets[1] wallet_node_2, server_2 = wallets[2] wallet_0 = wallet_node_0.wallet_state_manager.main_wallet wallet_1 = wallet_node_1.wallet_state_manager.main_wallet wallet_2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet_0.get_new_puzzlehash() ph1 = await wallet_1.get_new_puzzlehash() ph2 = await wallet_2.get_new_puzzlehash() await server_0.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) await server_1.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) await server_2.start_client(PeerInfo(self_hostname, uint16(full_node_server._port)), None) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(10, wallet_0.get_unconfirmed_balance, funds) await time_out_assert(10, wallet_0.get_confirmed_balance, funds) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph1)) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) async with wallet_node_0.wallet_state_manager.lock: did_wallet_0: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_0.wallet_state_manager, wallet_0, uint64(101) ) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_0.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_0.get_unconfirmed_balance, 101) await time_out_assert(15, did_wallet_0.get_pending_change_balance, 0) backup_ids = [bytes.fromhex(did_wallet_0.get_my_DID())] async with wallet_node_1.wallet_state_manager.lock: did_wallet_1: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_1.wallet_state_manager, wallet_1, uint64(201), backup_ids ) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_1.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_1.get_unconfirmed_balance, 201) await time_out_assert(15, did_wallet_1.get_pending_change_balance, 0) filename = "test.backup" did_wallet_1.create_backup(filename) async with wallet_node_2.wallet_state_manager.lock: did_wallet_2 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node_2.wallet_state_manager, wallet_2, filename ) coins = await did_wallet_1.select_coins(1) coin = coins.copy().pop() assert did_wallet_2.did_info.temp_coin == coin newpuzhash = await did_wallet_2.get_new_inner_hash() pubkey = bytes( (await did_wallet_2.wallet_state_manager.get_unused_derivation_record(did_wallet_2.wallet_info.id)).pubkey ) message_spend_bundle = await did_wallet_0.create_attestment( did_wallet_2.did_info.temp_coin.name(), newpuzhash, pubkey, "test.attest" ) print(f"pubkey: {pubkey}") for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_2.load_attest_files_for_recovery_spend(["test.attest"]) assert message_spend_bundle == test_message_spend_bundle await did_wallet_2.recovery_spend( did_wallet_2.did_info.temp_coin, newpuzhash, test_info_list, pubkey, test_message_spend_bundle, ) print(f"pubkey: {did_wallet_2}") for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(45, did_wallet_2.get_confirmed_balance, 201) await time_out_assert(45, did_wallet_2.get_unconfirmed_balance, 201) some_ph = 32 * b"\2" await did_wallet_2.create_exit_spend(some_ph) for i in range(1, num_blocks): await full_node_api.farm_new_transaction_block(FarmNewBlockProtocol(ph)) async def get_coins_with_ph(): coins = await full_node_api.full_node.coin_store.get_coin_records_by_puzzle_hash(True, some_ph) if len(coins) == 1: return True return False await time_out_assert(15, get_coins_with_ph, True) await time_out_assert(45, did_wallet_2.get_confirmed_balance, 0) await time_out_assert(45, did_wallet_2.get_unconfirmed_balance, 0) @pytest.mark.asyncio async def test_did_recovery_with_multiple_backup_dids(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet wallet2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) ph = await wallet2.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) recovery_list = [bytes.fromhex(did_wallet.get_my_DID())] async with wallet_node_2.wallet_state_manager.lock: did_wallet_2: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(101), recovery_list ) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_2.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_2.get_unconfirmed_balance, 101) assert did_wallet_2.did_info.backup_ids == recovery_list recovery_list.append(bytes.fromhex(did_wallet_2.get_my_DID())) async with wallet_node_2.wallet_state_manager.lock: did_wallet_3: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(201), recovery_list ) ph2 = await wallet.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) assert did_wallet_3.did_info.backup_ids == recovery_list await time_out_assert(15, did_wallet_3.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 201) coins = await did_wallet_3.select_coins(1) coin = coins.pop() filename = "test.backup" did_wallet_3.create_backup(filename) async with wallet_node.wallet_state_manager.lock: did_wallet_4 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node.wallet_state_manager, wallet, filename, ) pubkey = ( await did_wallet_4.wallet_state_manager.get_unused_derivation_record(did_wallet_2.wallet_info.id) ).pubkey new_ph = await did_wallet_4.get_new_inner_hash() message_spend_bundle = await did_wallet.create_attestment(coin.name(), new_ph, pubkey, "test1.attest") message_spend_bundle2 = await did_wallet_2.create_attestment(coin.name(), new_ph, pubkey, "test2.attest") message_spend_bundle = message_spend_bundle.aggregate([message_spend_bundle, message_spend_bundle2]) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_4.load_attest_files_for_recovery_spend(["test1.attest", "test2.attest"]) assert message_spend_bundle == test_message_spend_bundle for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await did_wallet_4.recovery_spend(coin, new_ph, test_info_list, pubkey, message_spend_bundle) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet_4.get_confirmed_balance, 201) await time_out_assert(15, did_wallet_4.get_unconfirmed_balance, 201) await time_out_assert(15, did_wallet_3.get_confirmed_balance, 0) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 0) @pytest.mark.asyncio async def test_did_recovery_with_empty_set(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) coins = await did_wallet.select_coins(1) coin = coins.pop() info = Program.to([]) pubkey = (await did_wallet.wallet_state_manager.get_unused_derivation_record(did_wallet.wallet_info.id)).pubkey spend_bundle = await did_wallet.recovery_spend( coin, ph, info, pubkey, SpendBundle([], AugSchemeMPL.aggregate([])) ) additions = spend_bundle.additions() assert additions == [] @pytest.mark.asyncio async def test_did_attest_after_recovery(self, two_wallet_nodes): num_blocks = 5 full_nodes, wallets = two_wallet_nodes full_node_1 = full_nodes[0] server_1 = full_node_1.server wallet_node, server_2 = wallets[0] wallet_node_2, server_3 = wallets[1] wallet = wallet_node.wallet_state_manager.main_wallet wallet2 = wallet_node_2.wallet_state_manager.main_wallet ph = await wallet.get_new_puzzlehash() await server_2.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) await server_3.start_client(PeerInfo(self_hostname, uint16(server_1._port)), None) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) funds = sum( [ calculate_pool_reward(uint32(i)) + calculate_base_farmer_reward(uint32(i)) for i in range(1, num_blocks - 1) ] ) await time_out_assert(15, wallet.get_confirmed_balance, funds) async with wallet_node.wallet_state_manager.lock: did_wallet: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node.wallet_state_manager, wallet, uint64(101) ) ph2 = await wallet2.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) recovery_list = [bytes.fromhex(did_wallet.get_my_DID())] async with wallet_node_2.wallet_state_manager.lock: did_wallet_2: DIDWallet = await DIDWallet.create_new_did_wallet( wallet_node_2.wallet_state_manager, wallet2, uint64(101), recovery_list ) ph = await wallet.get_new_puzzlehash() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_2.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_2.get_unconfirmed_balance, 101) assert did_wallet_2.did_info.backup_ids == recovery_list recovery_list = [bytes.fromhex(did_wallet_2.get_my_DID())] await did_wallet.update_recovery_list(recovery_list, uint64(1)) assert did_wallet.did_info.backup_ids == recovery_list await did_wallet.create_update_spend() for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) await time_out_assert(15, did_wallet.get_confirmed_balance, 101) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 101) filename = "test.backup" did_wallet_2.create_backup(filename) async with wallet_node.wallet_state_manager.lock: did_wallet_3 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node.wallet_state_manager, wallet, filename, ) new_ph = await did_wallet_3.get_new_inner_hash() coins = await did_wallet_2.select_coins(1) coin = coins.pop() pubkey = ( await did_wallet_3.wallet_state_manager.get_unused_derivation_record(did_wallet_3.wallet_info.id) ).pubkey message_spend_bundle = await did_wallet.create_attestment(coin.name(), new_ph, pubkey, "test.attest") for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) ( info, message_spend_bundle, ) = await did_wallet_3.load_attest_files_for_recovery_spend(["test.attest"]) await did_wallet_3.recovery_spend(coin, new_ph, info, pubkey, message_spend_bundle) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_3.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_3.get_unconfirmed_balance, 101) filename = "test.backup" did_wallet.create_backup(filename) async with wallet_node_2.wallet_state_manager.lock: did_wallet_4 = await DIDWallet.create_new_did_wallet_from_recovery( wallet_node_2.wallet_state_manager, wallet2, filename, ) coins = await did_wallet.select_coins(1) coin = coins.pop() new_ph = await did_wallet_4.get_new_inner_hash() pubkey = ( await did_wallet_4.wallet_state_manager.get_unused_derivation_record(did_wallet_4.wallet_info.id) ).pubkey await did_wallet_3.create_attestment(coin.name(), new_ph, pubkey, "test.attest") for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph2)) ( test_info_list, test_message_spend_bundle, ) = await did_wallet_4.load_attest_files_for_recovery_spend(["test.attest"]) spend_bundle = await did_wallet_4.recovery_spend( coin, new_ph, test_info_list, pubkey, test_message_spend_bundle ) await time_out_assert_not_none(15, full_node_1.full_node.mempool_manager.get_spendbundle, spend_bundle.name()) for i in range(1, num_blocks): await full_node_1.farm_new_transaction_block(FarmNewBlockProtocol(ph)) await time_out_assert(15, did_wallet_4.get_confirmed_balance, 101) await time_out_assert(15, did_wallet_4.get_unconfirmed_balance, 101) await time_out_assert(15, did_wallet.get_confirmed_balance, 0) await time_out_assert(15, did_wallet.get_unconfirmed_balance, 0)
true
true
f717cec2bb576f4d7ab0e2542c02f2f63a1c8e21
582
py
Python
src/models/gender.py
evdhout/learn_students
4fe10df13065ae7c46930cb69fa46851c1dca9e6
[ "MIT" ]
null
null
null
src/models/gender.py
evdhout/learn_students
4fe10df13065ae7c46930cb69fa46851c1dca9e6
[ "MIT" ]
null
null
null
src/models/gender.py
evdhout/learn_students
4fe10df13065ae7c46930cb69fa46851c1dca9e6
[ "MIT" ]
null
null
null
class Gender: NEUTRAL = 1 FEMALE = 2 MALE = 3 GENDER_STRINGS = {NEUTRAL: "neutral", FEMALE: "female", MALE: "male" } def __init__(self, gender: int = 1): self.gender: int = gender def __str__(self): return self.GENDER_STRINGS[_(self.gender)] def __eq__(self, other): if isinstance(other, Gender): return self.gender == other.gender elif isinstance(other, int): return self.gender == other else: return False
24.25
50
0.512027
class Gender: NEUTRAL = 1 FEMALE = 2 MALE = 3 GENDER_STRINGS = {NEUTRAL: "neutral", FEMALE: "female", MALE: "male" } def __init__(self, gender: int = 1): self.gender: int = gender def __str__(self): return self.GENDER_STRINGS[_(self.gender)] def __eq__(self, other): if isinstance(other, Gender): return self.gender == other.gender elif isinstance(other, int): return self.gender == other else: return False
true
true
f717cfc7cab0f458b15996bf3bf047fcffdef11f
3,950
py
Python
test/functional/feature_minchainwork.py
bontecoin/bontecoin-core
a7c368cf25339b9495e94651ee2a036e4f3068ae
[ "MIT" ]
null
null
null
test/functional/feature_minchainwork.py
bontecoin/bontecoin-core
a7c368cf25339b9495e94651ee2a036e4f3068ae
[ "MIT" ]
null
null
null
test/functional/feature_minchainwork.py
bontecoin/bontecoin-core
a7c368cf25339b9495e94651ee2a036e4f3068ae
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # Copyright (c) 2017 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. """Test logic for setting nMinimumChainWork on command line. Nodes don't consider themselves out of "initial block download" until their active chain has more work than nMinimumChainWork. Nodes don't download blocks from a peer unless the peer's best known block has more work than nMinimumChainWork. While in initial block download, nodes won't relay blocks to their peers, so test that this parameter functions as intended by verifying that block relay only succeeds past a given node once its nMinimumChainWork has been exceeded. """ import time from test_framework.test_framework import BontecoinTestFramework from test_framework.util import connect_nodes, assert_equal # 2 hashes required per regtest block (with no difficulty adjustment) REGTEST_WORK_PER_BLOCK = 2 class MinimumChainWorkTest(BontecoinTestFramework): def set_test_params(self): self.setup_clean_chain = True self.num_nodes = 3 self.extra_args = [[], ["-minimumchainwork=0x65"], ["-minimumchainwork=0x65"]] self.node_min_work = [0, 101, 101] def setup_network(self): # This test relies on the chain setup being: # node0 <- node1 <- node2 # Before leaving IBD, nodes prefer to download blocks from outbound # peers, so ensure that we're mining on an outbound peer and testing # block relay to inbound peers. self.setup_nodes() for i in range(self.num_nodes-1): connect_nodes(self.nodes[i+1], i) def run_test(self): # Start building a chain on node0. node2 shouldn't be able to sync until node1's # minchainwork is exceeded starting_chain_work = REGTEST_WORK_PER_BLOCK # Genesis block's work self.log.info("Testing relay across node %d (minChainWork = %d)", 1, self.node_min_work[1]) starting_blockcount = self.nodes[2].getblockcount() num_blocks_to_generate = int((self.node_min_work[1] - starting_chain_work) / REGTEST_WORK_PER_BLOCK) self.log.info("Generating %d blocks on node0", num_blocks_to_generate) hashes = self.nodes[0].generate(num_blocks_to_generate) self.log.info("Node0 current chain work: %s", self.nodes[0].getblockheader(hashes[-1])['chainwork']) # Sleep a few seconds and verify that node2 didn't get any new blocks # or headers. We sleep, rather than sync_blocks(node0, node1) because # it's reasonable either way for node1 to get the blocks, or not get # them (since they're below node1's minchainwork). time.sleep(3) self.log.info("Verifying node 2 has no more blocks than before") self.log.info("Blockcounts: %s", [n.getblockcount() for n in self.nodes]) # Node2 shouldn't have any new headers yet, because node1 should not # have relayed anything. assert_equal(len(self.nodes[2].getchaintips()), 1) assert_equal(self.nodes[2].getchaintips()[0]['height'], 0) assert self.nodes[1].getbestblockhash() != self.nodes[0].getbestblockhash() assert_equal(self.nodes[2].getblockcount(), starting_blockcount) self.log.info("Generating one more block") self.nodes[0].generate(1) self.log.info("Verifying nodes are all synced") # Because nodes in regtest are all manual connections (eg using # addnode), node1 should not have disconnected node0. If not for that, # we'd expect node1 to have disconnected node0 for serving an # insufficient work chain, in which case we'd need to reconnect them to # continue the test. self.sync_all() self.log.info("Blockcounts: %s", [n.getblockcount() for n in self.nodes]) if __name__ == '__main__': MinimumChainWorkTest().main()
43.888889
108
0.702278
import time from test_framework.test_framework import BontecoinTestFramework from test_framework.util import connect_nodes, assert_equal REGTEST_WORK_PER_BLOCK = 2 class MinimumChainWorkTest(BontecoinTestFramework): def set_test_params(self): self.setup_clean_chain = True self.num_nodes = 3 self.extra_args = [[], ["-minimumchainwork=0x65"], ["-minimumchainwork=0x65"]] self.node_min_work = [0, 101, 101] def setup_network(self): # block relay to inbound peers. self.setup_nodes() for i in range(self.num_nodes-1): connect_nodes(self.nodes[i+1], i) def run_test(self): # Start building a chain on node0. node2 shouldn't be able to sync until node1's # minchainwork is exceeded starting_chain_work = REGTEST_WORK_PER_BLOCK # Genesis block's work self.log.info("Testing relay across node %d (minChainWork = %d)", 1, self.node_min_work[1]) starting_blockcount = self.nodes[2].getblockcount() num_blocks_to_generate = int((self.node_min_work[1] - starting_chain_work) / REGTEST_WORK_PER_BLOCK) self.log.info("Generating %d blocks on node0", num_blocks_to_generate) hashes = self.nodes[0].generate(num_blocks_to_generate) self.log.info("Node0 current chain work: %s", self.nodes[0].getblockheader(hashes[-1])['chainwork']) # or headers. We sleep, rather than sync_blocks(node0, node1) because # it's reasonable either way for node1 to get the blocks, or not get time.sleep(3) self.log.info("Verifying node 2 has no more blocks than before") self.log.info("Blockcounts: %s", [n.getblockcount() for n in self.nodes]) # have relayed anything. assert_equal(len(self.nodes[2].getchaintips()), 1) assert_equal(self.nodes[2].getchaintips()[0]['height'], 0) assert self.nodes[1].getbestblockhash() != self.nodes[0].getbestblockhash() assert_equal(self.nodes[2].getblockcount(), starting_blockcount) self.log.info("Generating one more block") self.nodes[0].generate(1) self.log.info("Verifying nodes are all synced") # Because nodes in regtest are all manual connections (eg using # addnode), node1 should not have disconnected node0. If not for that, # we'd expect node1 to have disconnected node0 for serving an # continue the test. self.sync_all() self.log.info("Blockcounts: %s", [n.getblockcount() for n in self.nodes]) if __name__ == '__main__': MinimumChainWorkTest().main()
true
true
f717cfc9fd68c709b7614798b5f73cd888527fbd
2,681
py
Python
data/LSSVM.py
tejas-9er/SVM-vs-LSSVM
e44f63458680c39df370ddfcdf22e8c450d23128
[ "MIT" ]
5
2019-12-16T03:29:24.000Z
2022-02-14T01:26:28.000Z
data/LSSVM.py
tejas-9er/SVM-vs-LSSVM
e44f63458680c39df370ddfcdf22e8c450d23128
[ "MIT" ]
null
null
null
data/LSSVM.py
tejas-9er/SVM-vs-LSSVM
e44f63458680c39df370ddfcdf22e8c450d23128
[ "MIT" ]
5
2019-06-18T06:59:35.000Z
2020-12-19T11:10:56.000Z
import numpy as np import scipy from scipy.sparse import linalg from sklearn.metrics import accuracy_score class LSSVM: def __init__(self, kernel = 'linear', C = 1.0,gamma = 1.0, d = 2.0): kernels = { 'rbf':self.rbf, 'poly':self.polynomial, 'linear':self.linear } self.kernel = kernels[kernel] self.C = C self.gamma = 1.0 self.d = d #Build the gram matrix def build_kernel_matrix(self, X, y): instances, dimensions = X.shape gram_matrix = np.zeros((instances,instances)) #computing the gram matrix, involves going over the dataset and computing pairwise kernel function for i in range(0, instances): for j in range(0, instances): gram_matrix[i, j] = self.kernel(X[i], X[j]) return gram_matrix def fit(self, X, y): self.kernel_matrix = self.build_kernel_matrix(X,y) identity_matrix = np.identity(X.shape[0]) #We wish to solve Ax = B, so we begin by defining the matrices A, B A = np.zeros((X.shape[0]+1, X.shape[0]+1)) B = np.ones(((X.shape[0]+1,1))) A[0][0] = 0 A[0,1:X.shape[0]+1] = np.hstack((np.ones(X.shape[0]))) A[1:X.shape[0]+1,0] = np.ones(X.shape[0]) A[1:X.shape[0]+1,1:X.shape[0]+1] = self.kernel_matrix + identity_matrix / self.C #B is a column vector. B[0][0] = 0 B[1:X.shape[0]+1,0] = y solution = scipy.sparse.linalg.cg(A,B) self.bias = solution[:-1] solution = solution[:-1] self.support_vector_alphas = [] self.support_vector_labels = [] self.support_vectors = [] for index,alpha in enumerate(solution[0]): if(alpha > 1e-3): self.support_vector_alphas.append(alpha) self.support_vector_labels.append(y[index]) self.support_vectors.append(X[index]) #define kernels def linear(self, x1, x2): return np.dot(x1, x2.T) def polynomial(self, x1, x2): return (np.dot(x1, x2.T) ** self.d) def rbf(self,xi,xj): return np.exp(-self.gamma * np.linalg.norm(xi-xj)**2) def predict(self,X_test): predictions = [] for instance in X_test: for index, sv in enumerate(self.support_vectors): prediction = np.sum(self.support_vector_alphas[index] * self.support_vector_labels[index] * self.kernel(sv,instance) + self.bias) predictions.append(np.sign(prediction).astype(int)) return np.array(predictions)
33.098765
145
0.564342
import numpy as np import scipy from scipy.sparse import linalg from sklearn.metrics import accuracy_score class LSSVM: def __init__(self, kernel = 'linear', C = 1.0,gamma = 1.0, d = 2.0): kernels = { 'rbf':self.rbf, 'poly':self.polynomial, 'linear':self.linear } self.kernel = kernels[kernel] self.C = C self.gamma = 1.0 self.d = d def build_kernel_matrix(self, X, y): instances, dimensions = X.shape gram_matrix = np.zeros((instances,instances)) for i in range(0, instances): for j in range(0, instances): gram_matrix[i, j] = self.kernel(X[i], X[j]) return gram_matrix def fit(self, X, y): self.kernel_matrix = self.build_kernel_matrix(X,y) identity_matrix = np.identity(X.shape[0]) A = np.zeros((X.shape[0]+1, X.shape[0]+1)) B = np.ones(((X.shape[0]+1,1))) A[0][0] = 0 A[0,1:X.shape[0]+1] = np.hstack((np.ones(X.shape[0]))) A[1:X.shape[0]+1,0] = np.ones(X.shape[0]) A[1:X.shape[0]+1,1:X.shape[0]+1] = self.kernel_matrix + identity_matrix / self.C B[0][0] = 0 B[1:X.shape[0]+1,0] = y solution = scipy.sparse.linalg.cg(A,B) self.bias = solution[:-1] solution = solution[:-1] self.support_vector_alphas = [] self.support_vector_labels = [] self.support_vectors = [] for index,alpha in enumerate(solution[0]): if(alpha > 1e-3): self.support_vector_alphas.append(alpha) self.support_vector_labels.append(y[index]) self.support_vectors.append(X[index]) def linear(self, x1, x2): return np.dot(x1, x2.T) def polynomial(self, x1, x2): return (np.dot(x1, x2.T) ** self.d) def rbf(self,xi,xj): return np.exp(-self.gamma * np.linalg.norm(xi-xj)**2) def predict(self,X_test): predictions = [] for instance in X_test: for index, sv in enumerate(self.support_vectors): prediction = np.sum(self.support_vector_alphas[index] * self.support_vector_labels[index] * self.kernel(sv,instance) + self.bias) predictions.append(np.sign(prediction).astype(int)) return np.array(predictions)
true
true
f717cfcf1da0e79443c3060c50aa0097551bd7be
1,034
py
Python
p56_Merge_Intervals.py
bzhou26/leetcode_sol
82506521e2cc412f96cd1dfc3c8c3ab635f67f73
[ "MIT" ]
null
null
null
p56_Merge_Intervals.py
bzhou26/leetcode_sol
82506521e2cc412f96cd1dfc3c8c3ab635f67f73
[ "MIT" ]
null
null
null
p56_Merge_Intervals.py
bzhou26/leetcode_sol
82506521e2cc412f96cd1dfc3c8c3ab635f67f73
[ "MIT" ]
null
null
null
''' - Leetcode problem: 56 - Difficulty: Medium - Brief problem description: Given a collection of intervals, merge all overlapping intervals. Example 1: Input: [[1,3],[2,6],[8,10],[15,18]] Output: [[1,6],[8,10],[15,18]] Explanation: Since intervals [1,3] and [2,6] overlaps, merge them into [1,6]. Example 2: Input: [[1,4],[4,5]] Output: [[1,5]] Explanation: Intervals [1,4] and [4,5] are considered overlapping. NOTE: input types have been changed on April 15, 2019. Please reset to default code definition to get new method signature. - Solution Summary: - Used Resources: --- Bo Zhou ''' class Solution: def merge(self, intervals: List[List[int]]) -> List[List[int]]: hp = [] for i in intervals: heapq.heappush(hp, i) result = [] while hp: inter = heapq.heappop(hp) if result and inter[0] <= result[-1][1]: result[-1][1] = max(inter[1], result[-1][1]) else: result.append(inter) return result
22.478261
112
0.597679
class Solution: def merge(self, intervals: List[List[int]]) -> List[List[int]]: hp = [] for i in intervals: heapq.heappush(hp, i) result = [] while hp: inter = heapq.heappop(hp) if result and inter[0] <= result[-1][1]: result[-1][1] = max(inter[1], result[-1][1]) else: result.append(inter) return result
true
true
f717cff9e873f6d4ee6ab179e6d452349b4cd40a
1,032
py
Python
env/Lib/site-packages/OpenGL/raw/EGL/EXT/device_query.py
5gconnectedbike/Navio2
8c3f2b5d8bbbcea1fc08739945183c12b206712c
[ "BSD-3-Clause" ]
210
2016-04-09T14:26:00.000Z
2022-03-25T18:36:19.000Z
env/Lib/site-packages/OpenGL/raw/EGL/EXT/device_query.py
5gconnectedbike/Navio2
8c3f2b5d8bbbcea1fc08739945183c12b206712c
[ "BSD-3-Clause" ]
72
2016-09-04T09:30:19.000Z
2022-03-27T17:06:53.000Z
env/Lib/site-packages/OpenGL/raw/EGL/EXT/device_query.py
5gconnectedbike/Navio2
8c3f2b5d8bbbcea1fc08739945183c12b206712c
[ "BSD-3-Clause" ]
64
2016-04-09T14:26:49.000Z
2022-03-21T11:19:47.000Z
'''Autogenerated by xml_generate script, do not edit!''' from OpenGL import platform as _p, arrays # Code generation uses this from OpenGL.raw.EGL import _types as _cs # End users want this... from OpenGL.raw.EGL._types import * from OpenGL.raw.EGL import _errors from OpenGL.constant import Constant as _C import ctypes _EXTENSION_NAME = 'EGL_EXT_device_query' def _f( function ): return _p.createFunction( function,_p.PLATFORM.EGL,'EGL_EXT_device_query',error_checker=_errors._error_checker) EGL_BAD_DEVICE_EXT=_C('EGL_BAD_DEVICE_EXT',0x322B) EGL_DEVICE_EXT=_C('EGL_DEVICE_EXT',0x322C) # EGL_NO_DEVICE_EXT=_C('EGL_NO_DEVICE_EXT',((EGLDeviceEXT)(0))) @_f @_p.types(_cs.EGLBoolean,_cs.EGLDeviceEXT,_cs.EGLint,arrays.EGLAttribArray) def eglQueryDeviceAttribEXT(device,attribute,value):pass @_f @_p.types(ctypes.c_char_p,_cs.EGLDeviceEXT,_cs.EGLint) def eglQueryDeviceStringEXT(device,name):pass @_f @_p.types(_cs.EGLBoolean,_cs.EGLDisplay,_cs.EGLint,arrays.EGLAttribArray) def eglQueryDisplayAttribEXT(dpy,attribute,value):pass
39.692308
115
0.814922
from OpenGL import platform as _p, arrays from OpenGL.raw.EGL import _types as _cs from OpenGL.raw.EGL._types import * from OpenGL.raw.EGL import _errors from OpenGL.constant import Constant as _C import ctypes _EXTENSION_NAME = 'EGL_EXT_device_query' def _f( function ): return _p.createFunction( function,_p.PLATFORM.EGL,'EGL_EXT_device_query',error_checker=_errors._error_checker) EGL_BAD_DEVICE_EXT=_C('EGL_BAD_DEVICE_EXT',0x322B) EGL_DEVICE_EXT=_C('EGL_DEVICE_EXT',0x322C) @_f @_p.types(_cs.EGLBoolean,_cs.EGLDeviceEXT,_cs.EGLint,arrays.EGLAttribArray) def eglQueryDeviceAttribEXT(device,attribute,value):pass @_f @_p.types(ctypes.c_char_p,_cs.EGLDeviceEXT,_cs.EGLint) def eglQueryDeviceStringEXT(device,name):pass @_f @_p.types(_cs.EGLBoolean,_cs.EGLDisplay,_cs.EGLint,arrays.EGLAttribArray) def eglQueryDisplayAttribEXT(dpy,attribute,value):pass
true
true
f717d0d34b30202c5d6aa299b38730b144f81368
238
py
Python
chapter1-cam.py
gcruchon/test-opencv
fdf7cb7a86f5606ca6df6170107a0264fbc43e9c
[ "CC0-1.0" ]
null
null
null
chapter1-cam.py
gcruchon/test-opencv
fdf7cb7a86f5606ca6df6170107a0264fbc43e9c
[ "CC0-1.0" ]
null
null
null
chapter1-cam.py
gcruchon/test-opencv
fdf7cb7a86f5606ca6df6170107a0264fbc43e9c
[ "CC0-1.0" ]
null
null
null
import cv2 import numpy as np cap = cv2.VideoCapture(0) kernel = np.ones((5, 5), np.uint8) while True: success, img = cap.read() cv2.imshow("Cam", cv2.Canny(img, 100, 100)) if cv2.waitKey(1) & 0xFF == ord('q'): break
21.636364
47
0.609244
import cv2 import numpy as np cap = cv2.VideoCapture(0) kernel = np.ones((5, 5), np.uint8) while True: success, img = cap.read() cv2.imshow("Cam", cv2.Canny(img, 100, 100)) if cv2.waitKey(1) & 0xFF == ord('q'): break
true
true
f717d2356d90cde15cbb3d0d6e7bc7ef270af89b
320
py
Python
Beecrowd/1187.py
Gugarauj07/Algorithms
21c0fe96c1c870a01644d398c446182ced37eb19
[ "MIT" ]
null
null
null
Beecrowd/1187.py
Gugarauj07/Algorithms
21c0fe96c1c870a01644d398c446182ced37eb19
[ "MIT" ]
null
null
null
Beecrowd/1187.py
Gugarauj07/Algorithms
21c0fe96c1c870a01644d398c446182ced37eb19
[ "MIT" ]
null
null
null
O = input() soma = count = 0 for linha in range(12): for coluna in range(12): num = float(input()) if linha <= 4 and (10 - linha) >= coluna > linha: soma += num count += 1 if O == "S": print(f"{soma:.1f}") elif O == "M": media = soma / count print(f"{media:.1f}")
22.857143
57
0.484375
O = input() soma = count = 0 for linha in range(12): for coluna in range(12): num = float(input()) if linha <= 4 and (10 - linha) >= coluna > linha: soma += num count += 1 if O == "S": print(f"{soma:.1f}") elif O == "M": media = soma / count print(f"{media:.1f}")
true
true
f717d265ddfccca7c0bcef14a21dd104634baabd
189
py
Python
Lesson/lesson-oo-415.py
v-v-tarasov/python-my-lesson
57f8985cdfe9d9439452d6d3d0f89cf4bfbf2d14
[ "MIT" ]
null
null
null
Lesson/lesson-oo-415.py
v-v-tarasov/python-my-lesson
57f8985cdfe9d9439452d6d3d0f89cf4bfbf2d14
[ "MIT" ]
null
null
null
Lesson/lesson-oo-415.py
v-v-tarasov/python-my-lesson
57f8985cdfe9d9439452d6d3d0f89cf4bfbf2d14
[ "MIT" ]
null
null
null
def isPointInSquare(x, y): return 1.0 >= x >= -1.0 and 1.0 >= y >= -1.0 x = float(input()) y = float(input()) if isPointInSquare(x, y) == True: print('YES') else: print('NO')
17.181818
48
0.544974
def isPointInSquare(x, y): return 1.0 >= x >= -1.0 and 1.0 >= y >= -1.0 x = float(input()) y = float(input()) if isPointInSquare(x, y) == True: print('YES') else: print('NO')
true
true
f717d32ab2f0db9bf60e9f86574ade09c2ab6e47
208
py
Python
3_Django/PROJECT_NAME/APP_NAME/urls.py
kevinbeirne1/CS50-course
b6247cdea50890369a5e0fd60017980643bf6e37
[ "MIT" ]
1
2021-08-31T20:52:41.000Z
2021-08-31T20:52:41.000Z
3_Django/PROJECT_NAME/APP_NAME/urls.py
kevinbeirne1/CS50-course
b6247cdea50890369a5e0fd60017980643bf6e37
[ "MIT" ]
null
null
null
3_Django/PROJECT_NAME/APP_NAME/urls.py
kevinbeirne1/CS50-course
b6247cdea50890369a5e0fd60017980643bf6e37
[ "MIT" ]
null
null
null
from django.urls import path from . import views urlpatterns = [ path('', views.index, name="index"), path('<str:name>', views.greet, name="greet"), # path('david/', views.david, name="david"), ]
26
50
0.629808
from django.urls import path from . import views urlpatterns = [ path('', views.index, name="index"), path('<str:name>', views.greet, name="greet"), ]
true
true
f717d4456b785afc6a7d4ab258496baf58e1cea9
2,004
py
Python
diary/forms.py
j3ygh/ctdb
3c4b90c5265125fc37aa9ac569b7f39ffd5b13c9
[ "MIT" ]
1
2021-12-27T02:18:19.000Z
2021-12-27T02:18:19.000Z
diary/forms.py
j3ygh/ctdb
3c4b90c5265125fc37aa9ac569b7f39ffd5b13c9
[ "MIT" ]
null
null
null
diary/forms.py
j3ygh/ctdb
3c4b90c5265125fc37aa9ac569b7f39ffd5b13c9
[ "MIT" ]
1
2022-03-14T06:52:48.000Z
2022-03-14T06:52:48.000Z
from django import forms from django.utils.translation import gettext_lazy as _ from .models import Diary class DiaryModelForm(forms.ModelForm): class Meta: widgets = { 'date': forms.DateInput(attrs={'type': 'date'}), 'daily_record': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'todo': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'remark': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'comment': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), # TODO: ckeditor4 or not? } model = Diary exclude = ['created_by'] def full_clean(self): super().full_clean() try: self.instance.validate_unique() except forms.ValidationError: self.add_error(field='date', error=_('The diary with this date has already existed.')) class DiaryCommentModelForm(forms.ModelForm): class Meta: widgets = { 'date': forms.DateInput(attrs={'type': 'date', 'readonly': ''}), # Attrs readonly & style="pointer-events: none" make the <select> tag work like a readonly field. 'daily_check': forms.Select(attrs={'readonly': '', 'style': 'pointer-events: none'}), 'daily_record': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'todo': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'remark': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'comment': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), # TODO: ckeditor4 or not? } model = Diary exclude = ['created_by'] def full_clean(self): super().full_clean() try: self.instance.validate_unique() except forms.ValidationError: self.add_error(field='date', error=_('The diary with this date has already existed.'))
42.638298
122
0.586826
from django import forms from django.utils.translation import gettext_lazy as _ from .models import Diary class DiaryModelForm(forms.ModelForm): class Meta: widgets = { 'date': forms.DateInput(attrs={'type': 'date'}), 'daily_record': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'todo': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'remark': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), 'comment': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), } model = Diary exclude = ['created_by'] def full_clean(self): super().full_clean() try: self.instance.validate_unique() except forms.ValidationError: self.add_error(field='date', error=_('The diary with this date has already existed.')) class DiaryCommentModelForm(forms.ModelForm): class Meta: widgets = { 'date': forms.DateInput(attrs={'type': 'date', 'readonly': ''}), 'daily_check': forms.Select(attrs={'readonly': '', 'style': 'pointer-events: none'}), 'daily_record': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'todo': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'remark': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4', 'readonly': ''}), 'comment': forms.Textarea(attrs={'rows': 4, 'class': 'ckeditor4'}), } model = Diary exclude = ['created_by'] def full_clean(self): super().full_clean() try: self.instance.validate_unique() except forms.ValidationError: self.add_error(field='date', error=_('The diary with this date has already existed.'))
true
true
f717d534a0b8a9554184e8b7505e68a755c8fbed
2,086
py
Python
app/models/Config.py
samousli/ikiru
4a4a002db398dd7ba1b112ea406c92b0a8cb6c37
[ "MIT" ]
null
null
null
app/models/Config.py
samousli/ikiru
4a4a002db398dd7ba1b112ea406c92b0a8cb6c37
[ "MIT" ]
null
null
null
app/models/Config.py
samousli/ikiru
4a4a002db398dd7ba1b112ea406c92b0a8cb6c37
[ "MIT" ]
null
null
null
from ast import literal_eval import enum import logging from . import db from .Base import Base LOG = logging.getLogger(__name__) class ValueType(enum.Enum): Int = (1, int, int) Bool = (2, bool, lambda b: b == 'True') Float = (3, float, float) Text = (4, str, lambda s: s) Tuple = (5, tuple, literal_eval) List = (6, list, literal_eval) Set = (7, set, literal_eval) Dict = (8, dict, literal_eval) ACCEPTED_TYPES = tuple(k.value[1] for k in ValueType) PYTHON_TYPE_TO_ENUM_TYPE = {k.value[1]: k for k in ValueType} VALUETYPE_TO_CONVERTER_FUNC = {k: k.value[2] for k in ValueType} IGNORE_LIST = {'SQLALCHEMY_DATABASE_URI', 'JWT_SECRET_KEY', 'SECRET_KEY'} class Config(Base): # ToDo: Check for SQLAlchemy record size optimizations # Allowing for large key sizes to allow tiered keys e.g. <parent>.<child>.<key_str> key = db.Column(db.String(192), unique=True) _type = db.Column(db.Enum(ValueType)) _value = db.Column(db.String(192)) @property def value(self): if self._type in VALUETYPE_TO_CONVERTER_FUNC: return VALUETYPE_TO_CONVERTER_FUNC[self._type](self._value) raise TypeError('Invalid config value type.') @value.setter def value(self, val): if not isinstance(val, ACCEPTED_TYPES): raise ValueError(f'Invalid config value type ([{type(val)}] {val}).') self._type = PYTHON_TYPE_TO_ENUM_TYPE[type(val)] self._value = str(val) @staticmethod def populate_from_conf_object(conf, name=None): val = name or conf.__name__ db.session.add(Config(key='IKIRU_ENV', value=val)) for k, v in conf.as_dict().items(): if k in IGNORE_LIST: continue db.session.add(Config(key=k, value=v)) db.session.commit() @staticmethod def load_from_db(app): with app.app_context(): for conf in Config.query: app.config[conf.key] = conf.value def __repr__(self): return f'{self.__class__.__name__}(key={self.key}, value={self.value})'
31.606061
87
0.645733
from ast import literal_eval import enum import logging from . import db from .Base import Base LOG = logging.getLogger(__name__) class ValueType(enum.Enum): Int = (1, int, int) Bool = (2, bool, lambda b: b == 'True') Float = (3, float, float) Text = (4, str, lambda s: s) Tuple = (5, tuple, literal_eval) List = (6, list, literal_eval) Set = (7, set, literal_eval) Dict = (8, dict, literal_eval) ACCEPTED_TYPES = tuple(k.value[1] for k in ValueType) PYTHON_TYPE_TO_ENUM_TYPE = {k.value[1]: k for k in ValueType} VALUETYPE_TO_CONVERTER_FUNC = {k: k.value[2] for k in ValueType} IGNORE_LIST = {'SQLALCHEMY_DATABASE_URI', 'JWT_SECRET_KEY', 'SECRET_KEY'} class Config(Base): key = db.Column(db.String(192), unique=True) _type = db.Column(db.Enum(ValueType)) _value = db.Column(db.String(192)) @property def value(self): if self._type in VALUETYPE_TO_CONVERTER_FUNC: return VALUETYPE_TO_CONVERTER_FUNC[self._type](self._value) raise TypeError('Invalid config value type.') @value.setter def value(self, val): if not isinstance(val, ACCEPTED_TYPES): raise ValueError(f'Invalid config value type ([{type(val)}] {val}).') self._type = PYTHON_TYPE_TO_ENUM_TYPE[type(val)] self._value = str(val) @staticmethod def populate_from_conf_object(conf, name=None): val = name or conf.__name__ db.session.add(Config(key='IKIRU_ENV', value=val)) for k, v in conf.as_dict().items(): if k in IGNORE_LIST: continue db.session.add(Config(key=k, value=v)) db.session.commit() @staticmethod def load_from_db(app): with app.app_context(): for conf in Config.query: app.config[conf.key] = conf.value def __repr__(self): return f'{self.__class__.__name__}(key={self.key}, value={self.value})'
true
true
f717d61c108d4a173265538f31a31ca0754a90ee
2,281
py
Python
Lectures/lec_05/genSymbolImg.py
diable201/ComputerVision
5ee153363fa6757d3cd8b1add3e5d48b01a499e2
[ "MIT" ]
1
2021-02-23T08:44:02.000Z
2021-02-23T08:44:02.000Z
Lectures/lec_05/genSymbolImg.py
diable201/ComputerVision
5ee153363fa6757d3cd8b1add3e5d48b01a499e2
[ "MIT" ]
1
2021-02-23T09:12:44.000Z
2021-02-27T17:05:58.000Z
Lectures/lec_05/genSymbolImg.py
diable201/ComputerVision
5ee153363fa6757d3cd8b1add3e5d48b01a499e2
[ "MIT" ]
1
2021-02-28T14:15:57.000Z
2021-02-28T14:15:57.000Z
import cv2 import numpy as np from random import randint, uniform import string, random def addNoise(image): row,col = image.shape s_vs_p = 0.4 amount = 0.01 out = np.copy(image) # Salt mode num_salt = np.ceil(amount * image.size * s_vs_p) coords = [np.random.randint(0, i - 1, int(num_salt)) for i in image.shape] out[tuple(coords)] = 1 # Pepper mode num_pepper = np.ceil(amount* image.size * (1. - s_vs_p)) coords = [np.random.randint(0, i - 1, int(num_pepper)) for i in image.shape] out[tuple(coords)] = 0 return out # def addLines(img): # for i in range(randint(0,2)): # y1 = randint(0, img.shape[0]) # y2 = randint(0, img.shape[0]) # cv2.line(img, (0, y1), (img.shape[1], y2), 0, 1) def addBlur(img, kw, kh): return cv2.blur(img, (kw, kh)) def text_generator(chars, size = 8): return ''.join(random.choice(chars) for _ in range(size)) def addText(img, chars, font, size, line_size): text = text_generator(chars, 1) cv2.putText(img, text, (0, img.shape[0]-4), font, size, (0, 0, 255), line_size, cv2.LINE_AA) return text sizes = [(70,58),(40,35),(75,70),(70,70),(70,70),(50,50)] def genSymbolImg(chars = string.ascii_uppercase + string.digits, font = None, line_size = None, blur = None, kw = None, kh = None): if font is None: font = randint(0, 5) # if size is None: # size = uniform(2.5, 3.5) if line_size is None: line_size = randint(1, 3) if blur is None: blur = randint(0, 1) if kw is None: kw = randint(3, 9) if kh is None: kh = randint(3, 9) genImg = np.full(sizes[font], 255, dtype= np.uint8) text = addText(genImg, chars, font, 3, line_size) if randint(0, 1): genImg = addNoise(genImg) # if lines: # addLines(genImg) if blur: genImg = addBlur(genImg, kw, kh) return genImg, text if __name__ == '__main__': for i in xrange(10000): img, text = genSymbolImg(kw = 5, kh = 5, blur = 1) print(text) cv2.imshow("W", img) k = cv2.waitKey(0) if k == 27: break
22.145631
96
0.551074
import cv2 import numpy as np from random import randint, uniform import string, random def addNoise(image): row,col = image.shape s_vs_p = 0.4 amount = 0.01 out = np.copy(image) num_salt = np.ceil(amount * image.size * s_vs_p) coords = [np.random.randint(0, i - 1, int(num_salt)) for i in image.shape] out[tuple(coords)] = 1 num_pepper = np.ceil(amount* image.size * (1. - s_vs_p)) coords = [np.random.randint(0, i - 1, int(num_pepper)) for i in image.shape] out[tuple(coords)] = 0 return out def addBlur(img, kw, kh): return cv2.blur(img, (kw, kh)) def text_generator(chars, size = 8): return ''.join(random.choice(chars) for _ in range(size)) def addText(img, chars, font, size, line_size): text = text_generator(chars, 1) cv2.putText(img, text, (0, img.shape[0]-4), font, size, (0, 0, 255), line_size, cv2.LINE_AA) return text sizes = [(70,58),(40,35),(75,70),(70,70),(70,70),(50,50)] def genSymbolImg(chars = string.ascii_uppercase + string.digits, font = None, line_size = None, blur = None, kw = None, kh = None): if font is None: font = randint(0, 5) if line_size is None: line_size = randint(1, 3) if blur is None: blur = randint(0, 1) if kw is None: kw = randint(3, 9) if kh is None: kh = randint(3, 9) genImg = np.full(sizes[font], 255, dtype= np.uint8) text = addText(genImg, chars, font, 3, line_size) if randint(0, 1): genImg = addNoise(genImg) if blur: genImg = addBlur(genImg, kw, kh) return genImg, text if __name__ == '__main__': for i in xrange(10000): img, text = genSymbolImg(kw = 5, kh = 5, blur = 1) print(text) cv2.imshow("W", img) k = cv2.waitKey(0) if k == 27: break
true
true
f717d65f584bd26a05f750dc31f00fd352a9f051
7,242
py
Python
filer/admin/clipboardadmin.py
haricot/django-filer
f3b90fbbb90a3c99ade104b1c3190621773fa7e1
[ "BSD-3-Clause" ]
null
null
null
filer/admin/clipboardadmin.py
haricot/django-filer
f3b90fbbb90a3c99ade104b1c3190621773fa7e1
[ "BSD-3-Clause" ]
11
2019-11-02T20:57:52.000Z
2020-09-27T09:08:33.000Z
filer/admin/clipboardadmin.py
haricot/django-filer
f3b90fbbb90a3c99ade104b1c3190621773fa7e1
[ "BSD-3-Clause" ]
null
null
null
# -*- coding: utf-8 -*- from __future__ import absolute_import from django.conf.urls import url from django.contrib import admin from django.forms.models import modelform_factory from django.http import JsonResponse from django.views.decorators.csrf import csrf_exempt from .. import settings as filer_settings from ..models import Clipboard, ClipboardItem, Folder from ..utils.files import ( UploadException, handle_request_files_upload, handle_upload, ) from ..utils.loader import load_model from . import views NO_FOLDER_ERROR = "Can't find folder to upload. Please refresh and try again" NO_PERMISSIONS_FOR_FOLDER = ( "Can't use this folder, Permission Denied. Please select another folder." ) Image = load_model(filer_settings.FILER_IMAGE_MODEL) # ModelAdmins class ClipboardItemInline(admin.TabularInline): model = ClipboardItem class ClipboardAdmin(admin.ModelAdmin): model = Clipboard inlines = [ClipboardItemInline] filter_horizontal = ('files',) raw_id_fields = ('user',) verbose_name = "DEBUG Clipboard" verbose_name_plural = "DEBUG Clipboards" def get_urls(self): return [ url(r'^operations/paste_clipboard_to_folder/$', self.admin_site.admin_view(views.paste_clipboard_to_folder), name='filer-paste_clipboard_to_folder'), url(r'^operations/discard_clipboard/$', self.admin_site.admin_view(views.discard_clipboard), name='filer-discard_clipboard'), url(r'^operations/delete_clipboard/$', self.admin_site.admin_view(views.delete_clipboard), name='filer-delete_clipboard'), url(r'^operations/upload/(?P<folder_id>[0-9]+)/$', ajax_upload, name='filer-ajax_upload'), url(r'^operations/upload/no_folder/$', ajax_upload, name='filer-ajax_upload'), ] + super(ClipboardAdmin, self).get_urls() def get_model_perms(self, *args, **kwargs): """ It seems this is only used for the list view. NICE :-) """ return { 'add': False, 'change': False, 'delete': False, } @csrf_exempt def ajax_upload(request, folder_id=None): """ Receives an upload from the uploader. Receives only one file at a time. """ folder = None if folder_id: try: # Get folder folder = Folder.objects.get(pk=folder_id) except Folder.DoesNotExist: return JsonResponse({'error': NO_FOLDER_ERROR}) # check permissions if folder and not folder.has_add_children_permission(request): return JsonResponse({'error': NO_PERMISSIONS_FOR_FOLDER}) try: if len(request.FILES) == 1: # dont check if request is ajax or not, just grab the file upload, filename, is_raw = handle_request_files_upload(request) else: # else process the request as usual upload, filename, is_raw = handle_upload(request) # TODO: Deprecated/refactor # Get clipboad # clipboard = Clipboard.objects.get_or_create(user=request.user)[0] # find the file type for filer_class in filer_settings.FILER_FILE_MODELS: FileSubClass = load_model(filer_class) # TODO: What if there are more than one that qualify? if FileSubClass.matches_file_type(filename, upload, request): FileForm = modelform_factory( model=FileSubClass, fields=('original_filename', 'owner', 'file') ) break uploadform = FileForm({'original_filename': filename, 'owner': request.user.pk}, {'file': upload}) if uploadform.is_valid(): file_obj = uploadform.save(commit=False) # Enforce the FILER_IS_PUBLIC_DEFAULT file_obj.is_public = filer_settings.FILER_IS_PUBLIC_DEFAULT file_obj.folder = folder file_with_thumbs = None data = {} file_obj.save() # TODO: Deprecated/refactor # clipboard_item = ClipboardItem( # clipboard=clipboard, file=file_obj) # clipboard_item.save() # Try to generate thumbnails. if not file_obj.icons: if file_obj.extension not in filer_settings.FILER_FILE_EXTENSION_NOTHUMBS: # There is no point to continue, as we can't generate # thumbnails for this file. Usual reasons: bad format or # filename. file_obj.delete() # This would be logged in BaseImage._generate_thumbnails() # if FILER_ENABLE_LOGGING is on. file_with_thumbs = True return JsonResponse( {'error': 'failed to generate icons for file'}, status=500, ) else: file_with_thumbs = True if file_with_thumbs: # Backwards compatibility: try to get specific icon size (32px) # first. Then try medium icon size (they are already sorted), # fallback to the first (smallest) configured icon. thumbnail = None for size in (['32'] + filer_settings.FILER_ADMIN_ICON_SIZES[1::-1]): try: thumbnail = file_obj.icons[size] break except KeyError: continue # prepare preview thumbnail if type(file_obj) == Image: thumbnail_180_options = { 'size': (180, 180), 'crop': True, 'upscale': True, } thumbnail_180 = file_obj.file.get_thumbnail( thumbnail_180_options) data_thumbs = { 'thumbnail': thumbnail, 'thumbnail_180': thumbnail_180.url } data.update(data_thumbs) data_common = { 'alt_text': '', 'label': str(file_obj), 'file_id': file_obj.pk, 'original_image': file_obj.url } data.update(data_common) return JsonResponse(data) else: form_errors = '; '.join(['%s: %s' % ( field, ', '.join(errors)) for field, errors in list( uploadform.errors.items()) ]) raise UploadException( "AJAX request not valid: form invalid '%s'" % ( form_errors,)) except UploadException as e: return JsonResponse({'error': str(e)}, status=500)
38.521277
90
0.543496
from __future__ import absolute_import from django.conf.urls import url from django.contrib import admin from django.forms.models import modelform_factory from django.http import JsonResponse from django.views.decorators.csrf import csrf_exempt from .. import settings as filer_settings from ..models import Clipboard, ClipboardItem, Folder from ..utils.files import ( UploadException, handle_request_files_upload, handle_upload, ) from ..utils.loader import load_model from . import views NO_FOLDER_ERROR = "Can't find folder to upload. Please refresh and try again" NO_PERMISSIONS_FOR_FOLDER = ( "Can't use this folder, Permission Denied. Please select another folder." ) Image = load_model(filer_settings.FILER_IMAGE_MODEL) class ClipboardItemInline(admin.TabularInline): model = ClipboardItem class ClipboardAdmin(admin.ModelAdmin): model = Clipboard inlines = [ClipboardItemInline] filter_horizontal = ('files',) raw_id_fields = ('user',) verbose_name = "DEBUG Clipboard" verbose_name_plural = "DEBUG Clipboards" def get_urls(self): return [ url(r'^operations/paste_clipboard_to_folder/$', self.admin_site.admin_view(views.paste_clipboard_to_folder), name='filer-paste_clipboard_to_folder'), url(r'^operations/discard_clipboard/$', self.admin_site.admin_view(views.discard_clipboard), name='filer-discard_clipboard'), url(r'^operations/delete_clipboard/$', self.admin_site.admin_view(views.delete_clipboard), name='filer-delete_clipboard'), url(r'^operations/upload/(?P<folder_id>[0-9]+)/$', ajax_upload, name='filer-ajax_upload'), url(r'^operations/upload/no_folder/$', ajax_upload, name='filer-ajax_upload'), ] + super(ClipboardAdmin, self).get_urls() def get_model_perms(self, *args, **kwargs): return { 'add': False, 'change': False, 'delete': False, } @csrf_exempt def ajax_upload(request, folder_id=None): folder = None if folder_id: try: folder = Folder.objects.get(pk=folder_id) except Folder.DoesNotExist: return JsonResponse({'error': NO_FOLDER_ERROR}) if folder and not folder.has_add_children_permission(request): return JsonResponse({'error': NO_PERMISSIONS_FOR_FOLDER}) try: if len(request.FILES) == 1: upload, filename, is_raw = handle_request_files_upload(request) else: upload, filename, is_raw = handle_upload(request) for filer_class in filer_settings.FILER_FILE_MODELS: FileSubClass = load_model(filer_class) if FileSubClass.matches_file_type(filename, upload, request): FileForm = modelform_factory( model=FileSubClass, fields=('original_filename', 'owner', 'file') ) break uploadform = FileForm({'original_filename': filename, 'owner': request.user.pk}, {'file': upload}) if uploadform.is_valid(): file_obj = uploadform.save(commit=False) file_obj.is_public = filer_settings.FILER_IS_PUBLIC_DEFAULT file_obj.folder = folder file_with_thumbs = None data = {} file_obj.save() if not file_obj.icons: if file_obj.extension not in filer_settings.FILER_FILE_EXTENSION_NOTHUMBS: # thumbnails for this file. Usual reasons: bad format or # filename. file_obj.delete() # This would be logged in BaseImage._generate_thumbnails() # if FILER_ENABLE_LOGGING is on. file_with_thumbs = True return JsonResponse( {'error': 'failed to generate icons for file'}, status=500, ) else: file_with_thumbs = True if file_with_thumbs: # Backwards compatibility: try to get specific icon size (32px) # first. Then try medium icon size (they are already sorted), # fallback to the first (smallest) configured icon. thumbnail = None for size in (['32'] + filer_settings.FILER_ADMIN_ICON_SIZES[1::-1]): try: thumbnail = file_obj.icons[size] break except KeyError: continue # prepare preview thumbnail if type(file_obj) == Image: thumbnail_180_options = { 'size': (180, 180), 'crop': True, 'upscale': True, } thumbnail_180 = file_obj.file.get_thumbnail( thumbnail_180_options) data_thumbs = { 'thumbnail': thumbnail, 'thumbnail_180': thumbnail_180.url } data.update(data_thumbs) data_common = { 'alt_text': '', 'label': str(file_obj), 'file_id': file_obj.pk, 'original_image': file_obj.url } data.update(data_common) return JsonResponse(data) else: form_errors = '; '.join(['%s: %s' % ( field, ', '.join(errors)) for field, errors in list( uploadform.errors.items()) ]) raise UploadException( "AJAX request not valid: form invalid '%s'" % ( form_errors,)) except UploadException as e: return JsonResponse({'error': str(e)}, status=500)
true
true
f717d6a1554caa5ee66a91c8ac8b847a9b74aadc
1,083
py
Python
indico/modules/events/reminders/blueprint.py
uxmaster/indico
ecd19f17ef6fdc9f5584f59c87ec647319ce5d31
[ "MIT" ]
1
2019-11-03T11:34:16.000Z
2019-11-03T11:34:16.000Z
indico/modules/events/reminders/blueprint.py
NP-compete/indico
80db7ca0ef9d1f3240a16b9ff2d84bf0bf26c549
[ "MIT" ]
null
null
null
indico/modules/events/reminders/blueprint.py
NP-compete/indico
80db7ca0ef9d1f3240a16b9ff2d84bf0bf26c549
[ "MIT" ]
null
null
null
# This file is part of Indico. # Copyright (C) 2002 - 2019 CERN # # Indico is free software; you can redistribute it and/or # modify it under the terms of the MIT License; see the # LICENSE file for more details. from __future__ import unicode_literals from indico.modules.events.reminders.controllers import (RHAddReminder, RHDeleteReminder, RHEditReminder, RHListReminders, RHPreviewReminder) from indico.web.flask.wrappers import IndicoBlueprint _bp = IndicoBlueprint('event_reminders', __name__, template_folder='templates', virtual_template_folder='events/reminders', url_prefix='/event/<confId>/manage/reminders') _bp.add_url_rule('/', 'list', RHListReminders) _bp.add_url_rule('/add', 'add', RHAddReminder, methods=('GET', 'POST')) _bp.add_url_rule('/preview', 'preview', RHPreviewReminder, methods=('POST',)) _bp.add_url_rule('/<int:reminder_id>/', 'edit', RHEditReminder, methods=('GET', 'POST')) _bp.add_url_rule('/<int:reminder_id>/delete', 'delete', RHDeleteReminder, methods=('POST',))
47.086957
112
0.710988
from __future__ import unicode_literals from indico.modules.events.reminders.controllers import (RHAddReminder, RHDeleteReminder, RHEditReminder, RHListReminders, RHPreviewReminder) from indico.web.flask.wrappers import IndicoBlueprint _bp = IndicoBlueprint('event_reminders', __name__, template_folder='templates', virtual_template_folder='events/reminders', url_prefix='/event/<confId>/manage/reminders') _bp.add_url_rule('/', 'list', RHListReminders) _bp.add_url_rule('/add', 'add', RHAddReminder, methods=('GET', 'POST')) _bp.add_url_rule('/preview', 'preview', RHPreviewReminder, methods=('POST',)) _bp.add_url_rule('/<int:reminder_id>/', 'edit', RHEditReminder, methods=('GET', 'POST')) _bp.add_url_rule('/<int:reminder_id>/delete', 'delete', RHDeleteReminder, methods=('POST',))
true
true
f717d74e6dbd251d84df1d67ed85b3b52ba68270
2,105
py
Python
labellab-flask/api/models/User.py
darkshredder/LabelLab
fc762e6eea52b9023e38ba5f32bbcaa7cbc17dbe
[ "Apache-2.0" ]
70
2019-01-25T19:16:00.000Z
2022-03-23T14:37:28.000Z
labellab-flask/api/models/User.py
darkshredder/LabelLab
fc762e6eea52b9023e38ba5f32bbcaa7cbc17dbe
[ "Apache-2.0" ]
350
2019-01-30T10:50:34.000Z
2022-03-31T19:58:44.000Z
labellab-flask/api/models/User.py
darkshredder/LabelLab
fc762e6eea52b9023e38ba5f32bbcaa7cbc17dbe
[ "Apache-2.0" ]
140
2019-01-30T08:53:35.000Z
2022-03-25T15:37:12.000Z
from datetime import datetime from flask import current_app, jsonify from flask_bcrypt import Bcrypt import json from api.extensions import db, Base, ma class User(db.Model): """ This model holds information about a user registered """ __tablename__ = "user" id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(80), nullable=False) username = db.Column(db.String(80), unique=True, nullable=False,) password = db.Column(db.String(128)) email = db.Column(db.String(100), nullable=False, unique=True) date = db.Column(db.DateTime, default=datetime.now()) thumbnail = db.Column(db.String(1500), default='https://react.semantic-ui.com/images/avatar/large/elliot.jpg') projects = db.relationship('Project', backref='user', lazy=True, cascade="all, save-update, delete", passive_deletes=True) project_members = db.relationship('ProjectMember', backref='user', lazy=True, cascade="all, save-update, delete", passive_deletes=True) def __init__(self, name, username, email, password=None): """ Initializes the user instance """ self.name = name self.username = username self.email = email if password: self.password = User.generate_password_hash(password) def __repr__(self): """ Returns the object reprensentation of user """ return "<User %r>" % self.name @staticmethod def generate_password_hash(password): """ Returns hash of password """ return Bcrypt().generate_password_hash(password,10).decode() def verify_password(self, password): """ Verify the password """ return Bcrypt().check_password_hash(self.password, password)
33.951613
94
0.560095
from datetime import datetime from flask import current_app, jsonify from flask_bcrypt import Bcrypt import json from api.extensions import db, Base, ma class User(db.Model): __tablename__ = "user" id = db.Column(db.Integer, primary_key=True) name = db.Column(db.String(80), nullable=False) username = db.Column(db.String(80), unique=True, nullable=False,) password = db.Column(db.String(128)) email = db.Column(db.String(100), nullable=False, unique=True) date = db.Column(db.DateTime, default=datetime.now()) thumbnail = db.Column(db.String(1500), default='https://react.semantic-ui.com/images/avatar/large/elliot.jpg') projects = db.relationship('Project', backref='user', lazy=True, cascade="all, save-update, delete", passive_deletes=True) project_members = db.relationship('ProjectMember', backref='user', lazy=True, cascade="all, save-update, delete", passive_deletes=True) def __init__(self, name, username, email, password=None): self.name = name self.username = username self.email = email if password: self.password = User.generate_password_hash(password) def __repr__(self): return "<User %r>" % self.name @staticmethod def generate_password_hash(password): return Bcrypt().generate_password_hash(password,10).decode() def verify_password(self, password): return Bcrypt().check_password_hash(self.password, password)
true
true
f717d76a526f8cc95d2243ac714a7311f53a0737
4,588
py
Python
Project2/program1/clean.py
Sandeep-AnilKumar/Data-Science-Projects
b7d890f855ffc6edd0544ff3bd115fa85d19fd4f
[ "MIT" ]
null
null
null
Project2/program1/clean.py
Sandeep-AnilKumar/Data-Science-Projects
b7d890f855ffc6edd0544ff3bd115fa85d19fd4f
[ "MIT" ]
null
null
null
Project2/program1/clean.py
Sandeep-AnilKumar/Data-Science-Projects
b7d890f855ffc6edd0544ff3bd115fa85d19fd4f
[ "MIT" ]
null
null
null
import sys import re import string # dictionary to store clean data. cleaned_data = {} # list of professors professors = [] # list of courses courses = [] def createdict(profname, course_list): new_course_list = [] is_course_match = False profname = profname.title() prof_courses = course_list.split('|') prof_courses = [course.strip() for course in prof_courses] if profname not in cleaned_data: cleaned_data.setdefault(profname, []) for c in prof_courses: # replace & with and if '&' in c: c = c.replace('&', 'and ') # replace intro. or intro with introduction matcher = re.match("intro\.?", c) if matcher: c = re.sub("intro\.?", "introduction ", c) # replace or make all roman numerals capitals. matcher = re.match(r"\bi+\b", c.lower()) if matcher: c = c.lower() c = re.sub(r"\bi\b", "I", c) c = re.sub(r"\bii\b", "II", c) c = re.sub(r"\biii\b", "III", c) # remove all punctuation marks. punctuation_regex = re.compile('[%s]' % re.escape(string.punctuation)) c = punctuation_regex.sub('', c) c_split = c.split() c_word_array = [] # make only non roman numeral words as "title", roman numerals as "uppercase". for c_split_constituent in c_split: matcher = re.match(r"\bi+\b", c_split_constituent.lower()) if matcher: c_split_constituent = c_split_constituent.upper() else: c_split_constituent = c_split_constituent.title() c_word_array.append(c_split_constituent) c = (" ".join(c_word for c_word in c_word_array)) if not courses: courses.append(c) is_course_match = False if c in courses: is_course_match = True else: # calculate courses similarity using edit distance using DP. for c2 in courses: c_length = len(c) c2_length = len(c2) table = [[0 for x in range(c2_length + 1)] for x in range(c_length + 1)] for i in range(c_length + 1): table[i][0] = i for j in range(c2_length + 1): table[0][j] = j for i in range(1, c_length + 1): for j in range(1, c2_length + 1): if c[i - 1] == c2[j - 1]: table[i][j] = table[i - 1][j - 1] else: table[i][j] = 1 + min(table[i][j - 1], table[i - 1][j], table[i - 1][j - 1]) distance = table[i][j] if distance <= 2: is_course_match = True c = c2 break if not is_course_match: courses.append(c) new_course_list.append(c) cleaned_data[profname] = cleaned_data[profname] + new_course_list return # output file where the cleaned data is stored. out = open("cleaned.txt", "w") # input file to read the data. inFile = sys.argv[1] file_buffer = open(inFile, "r").read().splitlines() for line in file_buffer: if not line.strip(): continue # separate the prof names and course lists. separator = line.split('-', 1) # if the professor name has comma, since we only need last name, we take only that. prof = separator[0].strip() if ',' in prof: prof = (prof.split(',')[0]).strip() # if professor name has a space in the last name, take only the last part from it. if ' ' in prof: prof = (prof.split()[-1]).strip() # if professor name has a '.' in the last name, take only the last part form it. elif '.' in prof: prof = (prof.split('.')[-1]).strip() # if professor name is in firstName.lastName format, take only lastName. elif '.' in prof: prof = ((prof.split('.')[-1]).split()[-1]).strip() # if professor name is in firstName lastName format, take only lastName. elif ' ' in prof: prof = (prof.split()[-1]).strip() else: prof = prof.strip() # create a dictionary of professor to their courses. createdict(prof, separator[1].strip()) for key, value in cleaned_data.items(): professors.append(key) # sort the courses list value = list(set(value)) value.sort() cleaned_data[key] = value professors.sort() for name in professors: out.write(name + " - " + ("|".join(cleaned_data[name]))+"\n")
35.022901
104
0.549259
import sys import re import string cleaned_data = {} professors = [] courses = [] def createdict(profname, course_list): new_course_list = [] is_course_match = False profname = profname.title() prof_courses = course_list.split('|') prof_courses = [course.strip() for course in prof_courses] if profname not in cleaned_data: cleaned_data.setdefault(profname, []) for c in prof_courses: if '&' in c: c = c.replace('&', 'and ') matcher = re.match("intro\.?", c) if matcher: c = re.sub("intro\.?", "introduction ", c) matcher = re.match(r"\bi+\b", c.lower()) if matcher: c = c.lower() c = re.sub(r"\bi\b", "I", c) c = re.sub(r"\bii\b", "II", c) c = re.sub(r"\biii\b", "III", c) punctuation_regex = re.compile('[%s]' % re.escape(string.punctuation)) c = punctuation_regex.sub('', c) c_split = c.split() c_word_array = [] for c_split_constituent in c_split: matcher = re.match(r"\bi+\b", c_split_constituent.lower()) if matcher: c_split_constituent = c_split_constituent.upper() else: c_split_constituent = c_split_constituent.title() c_word_array.append(c_split_constituent) c = (" ".join(c_word for c_word in c_word_array)) if not courses: courses.append(c) is_course_match = False if c in courses: is_course_match = True else: for c2 in courses: c_length = len(c) c2_length = len(c2) table = [[0 for x in range(c2_length + 1)] for x in range(c_length + 1)] for i in range(c_length + 1): table[i][0] = i for j in range(c2_length + 1): table[0][j] = j for i in range(1, c_length + 1): for j in range(1, c2_length + 1): if c[i - 1] == c2[j - 1]: table[i][j] = table[i - 1][j - 1] else: table[i][j] = 1 + min(table[i][j - 1], table[i - 1][j], table[i - 1][j - 1]) distance = table[i][j] if distance <= 2: is_course_match = True c = c2 break if not is_course_match: courses.append(c) new_course_list.append(c) cleaned_data[profname] = cleaned_data[profname] + new_course_list return out = open("cleaned.txt", "w") inFile = sys.argv[1] file_buffer = open(inFile, "r").read().splitlines() for line in file_buffer: if not line.strip(): continue separator = line.split('-', 1) prof = separator[0].strip() if ',' in prof: prof = (prof.split(',')[0]).strip() if ' ' in prof: prof = (prof.split()[-1]).strip() elif '.' in prof: prof = (prof.split('.')[-1]).strip() elif '.' in prof: prof = ((prof.split('.')[-1]).split()[-1]).strip() elif ' ' in prof: prof = (prof.split()[-1]).strip() else: prof = prof.strip() createdict(prof, separator[1].strip()) for key, value in cleaned_data.items(): professors.append(key) value = list(set(value)) value.sort() cleaned_data[key] = value professors.sort() for name in professors: out.write(name + " - " + ("|".join(cleaned_data[name]))+"\n")
true
true
f717d7ca70131a15c0bee7dca10a57ff4d0cb3db
9,967
py
Python
spacy_pytorch_transformers/pipeline/tok2vec.py
tamuhey/spacy-pytorch-transformers
1b4a58505ee3618a6288a47d4b5716981e39e581
[ "MIT" ]
1
2021-01-11T19:35:46.000Z
2021-01-11T19:35:46.000Z
spacy_pytorch_transformers/pipeline/tok2vec.py
tamuhey/spacy-pytorch-transformers
1b4a58505ee3618a6288a47d4b5716981e39e581
[ "MIT" ]
null
null
null
spacy_pytorch_transformers/pipeline/tok2vec.py
tamuhey/spacy-pytorch-transformers
1b4a58505ee3618a6288a47d4b5716981e39e581
[ "MIT" ]
null
null
null
from typing import Any, List from thinc.neural.ops import get_array_module from spacy.pipeline import Pipe from spacy.tokens import Doc from spacy.vocab import Vocab from spacy.util import minibatch from ..wrapper import PyTT_Wrapper from ..model_registry import get_model_function from ..activations import Activations, RaggedArray from ..util import get_pytt_config, get_pytt_model, get_sents class PyTT_TokenVectorEncoder(Pipe): """spaCy pipeline component to use PyTorch-Transformers models. The component assigns the output of the transformer to the `doc._.pytt_outputs` extension attribute. We also calculate an alignment between the word-piece tokens and the spaCy tokenization, so that we can use the last hidden states to set the doc.tensor attribute. When multiple word-piece tokens align to the same spaCy token, the spaCy token receives the sum of their values. """ name = "pytt_tok2vec" @classmethod def from_nlp(cls, nlp, **cfg): """Factory to add to Language.factories via entry point.""" return cls(nlp.vocab, **cfg) @classmethod def from_pretrained(cls, vocab: Vocab, name: str, **cfg): """Create a PyTT_TokenVectorEncoder instance using pre-trained weights from a PyTorch Transformer model, even if it's not installed as a spaCy package. vocab (spacy.vocab.Vocab): The spaCy vocab to use. name (unicode): Name of pre-trained model, e.g. 'bert-base-uncased'. RETURNS (PyTT_TokenVectorEncoder): The token vector encoder. """ cfg["pytt_name"] = name model = cls.Model(from_pretrained=True, **cfg) cfg["pytt_config"] = dict(model._model.pytt_model.config.to_dict()) self = cls(vocab, model=model, **cfg) return self @classmethod def Model(cls, **cfg) -> Any: """Create an instance of `PyTT_Wrapper`, which holds the PyTorch-Transformers model. **cfg: Optional config parameters. RETURNS (thinc.neural.Model): The wrapped model. """ name = cfg.get("pytt_name") if not name: raise ValueError("Need pytt_name argument, e.g. 'bert-base-uncased'") if cfg.get("from_pretrained"): pytt_wrap = PyTT_Wrapper.from_pretrained(name) else: pytt_config = cfg["pytt_config"] # Work around floating point limitation in ujson: # If we have the setting cfg["pytt_config"]["layer_norm_eps"] as 0, # that's because of misprecision in serializing. Fix that. pytt_config["layer_norm_eps"] = 1e-12 config_cls = get_pytt_config(name) model_cls = get_pytt_model(name) # Need to match the name their constructor expects. if "vocab_size" in cfg["pytt_config"]: vocab_size = cfg["pytt_config"]["vocab_size"] cfg["pytt_config"]["vocab_size_or_config_json_file"] = vocab_size pytt_wrap = PyTT_Wrapper( name, pytt_config, model_cls(config_cls(**pytt_config)) ) make_model = get_model_function(cfg.get("architecture", "tok2vec_per_sentence")) model = make_model(pytt_wrap, cfg) setattr(model, "nO", pytt_wrap.nO) setattr(model, "_model", pytt_wrap) return model def __init__(self, vocab, model=True, **cfg): """Initialize the component. model (thinc.neural.Model / True): The component's model or `True` if not initialized yet. **cfg: Optional config parameters. """ self.vocab = vocab self.model = model self.cfg = cfg @property def token_vector_width(self): return self.model._model.nO @property def pytt_model(self): return self.model._model.pytt_model def __call__(self, doc): """Process a Doc and assign the extracted features. doc (spacy.tokens.Doc): The Doc to process. RETURNS (spacy.tokens.Doc): The processed Doc. """ self.require_model() outputs = self.predict([doc]) self.set_annotations([doc], outputs) return doc def pipe(self, stream, batch_size=128): """Process Doc objects as a stream and assign the extracted features. stream (iterable): A stream of Doc objects. batch_size (int): The number of texts to buffer. YIELDS (spacy.tokens.Doc): Processed Docs in order. """ for docs in minibatch(stream, size=batch_size): docs = list(docs) outputs = self.predict(docs) self.set_annotations(docs, outputs) for doc in docs: yield doc def begin_update(self, docs, drop=None, **cfg): """Get the predictions and a callback to complete the gradient update. This is only used internally within PyTT_Language.update. """ outputs, backprop = self.model.begin_update(docs, drop=drop) def finish_update(docs, sgd=None): assert len(docs) d_lh = [] d_po = [] lh_lengths = [] po_lengths = [] for doc in docs: d_lh.append(doc._.pytt_d_last_hidden_state) d_po.append(doc._.pytt_d_pooler_output) lh_lengths.append(doc._.pytt_d_last_hidden_state.shape[0]) po_lengths.append(doc._.pytt_d_pooler_output.shape[0]) xp = self.model.ops.xp gradients = Activations( RaggedArray(xp.vstack(d_lh), lh_lengths), RaggedArray(xp.vstack(d_po), po_lengths), ) backprop(gradients, sgd=sgd) for doc in docs: doc._.pytt_d_last_hidden_state.fill(0) doc._.pytt_d_pooler_output.fill(0) return None return outputs, finish_update def predict(self, docs): """Run the transformer model on a batch of docs and return the extracted features. docs (iterable): A batch of Docs to process. RETURNS (list): A list of Activations objects, one per doc. """ return self.model.predict(docs) def set_annotations(self, docs: List[Doc], activations: Activations): """Assign the extracted features to the Doc objects and overwrite the vector and similarity hooks. docs (iterable): A batch of `Doc` objects. activations (iterable): A batch of activations. """ xp = activations.xp for i, doc in enumerate(docs): # Make it 2d -- acts are always 3d, to represent batch size. wp_tensor = activations.lh.get(i) doc.tensor = self.model.ops.allocate((len(doc), self.model.nO)) doc._.pytt_last_hidden_state = wp_tensor if activations.has_po: pooler_output = activations.po.get(i) doc._.pytt_pooler_output = pooler_output doc._.pytt_d_last_hidden_state = xp.zeros((0, 0), dtype=wp_tensor.dtype) doc._.pytt_d_pooler_output = xp.zeros((0, 0), dtype=wp_tensor.dtype) doc._.pytt_d_all_hidden_states = [] doc._.pytt_d_all_attentions = [] if wp_tensor.shape != (len(doc._.pytt_word_pieces), self.model.nO): raise ValueError( "Mismatch between tensor shape and word pieces. This usually " "means we did something wrong in the sentence reshaping, " "or possibly finding the separator tokens." ) # Count how often each word-piece token is represented. This allows # a weighted sum, so that we can make sure doc.tensor.sum() # equals wp_tensor.sum(). Do this with sensitivity to boundary tokens wp_rows, align_sizes = _get_boundary_sensitive_alignment(doc) wp_weighted = wp_tensor / xp.array(align_sizes, dtype="f").reshape((-1, 1)) # TODO: Obviously incrementing the rows individually is bad. How # to do in one shot without blowing up the memory? for i, word_piece_slice in enumerate(wp_rows): for j in word_piece_slice: doc.tensor[i] += wp_weighted[j] doc.user_hooks["vector"] = get_doc_vector_via_tensor doc.user_span_hooks["vector"] = get_span_vector_via_tensor doc.user_token_hooks["vector"] = get_token_vector_via_tensor doc.user_hooks["similarity"] = get_similarity_via_tensor doc.user_span_hooks["similarity"] = get_similarity_via_tensor doc.user_token_hooks["similarity"] = get_similarity_via_tensor def _get_boundary_sensitive_alignment(doc): align_sizes = [0 for _ in range(len(doc._.pytt_word_pieces))] wp_rows = [] for word_piece_slice in doc._.pytt_alignment: wp_rows.append(list(word_piece_slice)) for i in word_piece_slice: align_sizes[i] += 1 # To make this weighting work, we "align" the boundary tokens against # every token in their sentence. The boundary tokens are otherwise # unaligned, which is how we identify them. for sent in get_sents(doc): offset = sent._.pytt_start for i in range(len(sent._.pytt_word_pieces)): if align_sizes[offset + i] == 0: align_sizes[offset + i] = len(sent) for tok in sent: wp_rows[tok.i].append(offset + i) return wp_rows, align_sizes def get_doc_vector_via_tensor(doc): return doc.tensor.sum(axis=0) def get_span_vector_via_tensor(span): return span.doc.tensor[span.start : span.end].sum(axis=0) def get_token_vector_via_tensor(token): return token.doc.tensor[token.i] def get_similarity_via_tensor(doc1, doc2): v1 = doc1.vector v2 = doc2.vector xp = get_array_module(v1) return xp.dot(v1, v2) / (doc1.vector_norm * doc2.vector_norm)
40.681633
88
0.631685
from typing import Any, List from thinc.neural.ops import get_array_module from spacy.pipeline import Pipe from spacy.tokens import Doc from spacy.vocab import Vocab from spacy.util import minibatch from ..wrapper import PyTT_Wrapper from ..model_registry import get_model_function from ..activations import Activations, RaggedArray from ..util import get_pytt_config, get_pytt_model, get_sents class PyTT_TokenVectorEncoder(Pipe): name = "pytt_tok2vec" @classmethod def from_nlp(cls, nlp, **cfg): return cls(nlp.vocab, **cfg) @classmethod def from_pretrained(cls, vocab: Vocab, name: str, **cfg): cfg["pytt_name"] = name model = cls.Model(from_pretrained=True, **cfg) cfg["pytt_config"] = dict(model._model.pytt_model.config.to_dict()) self = cls(vocab, model=model, **cfg) return self @classmethod def Model(cls, **cfg) -> Any: name = cfg.get("pytt_name") if not name: raise ValueError("Need pytt_name argument, e.g. 'bert-base-uncased'") if cfg.get("from_pretrained"): pytt_wrap = PyTT_Wrapper.from_pretrained(name) else: pytt_config = cfg["pytt_config"] pytt_config["layer_norm_eps"] = 1e-12 config_cls = get_pytt_config(name) model_cls = get_pytt_model(name) # Need to match the name their constructor expects. if "vocab_size" in cfg["pytt_config"]: vocab_size = cfg["pytt_config"]["vocab_size"] cfg["pytt_config"]["vocab_size_or_config_json_file"] = vocab_size pytt_wrap = PyTT_Wrapper( name, pytt_config, model_cls(config_cls(**pytt_config)) ) make_model = get_model_function(cfg.get("architecture", "tok2vec_per_sentence")) model = make_model(pytt_wrap, cfg) setattr(model, "nO", pytt_wrap.nO) setattr(model, "_model", pytt_wrap) return model def __init__(self, vocab, model=True, **cfg): self.vocab = vocab self.model = model self.cfg = cfg @property def token_vector_width(self): return self.model._model.nO @property def pytt_model(self): return self.model._model.pytt_model def __call__(self, doc): self.require_model() outputs = self.predict([doc]) self.set_annotations([doc], outputs) return doc def pipe(self, stream, batch_size=128): for docs in minibatch(stream, size=batch_size): docs = list(docs) outputs = self.predict(docs) self.set_annotations(docs, outputs) for doc in docs: yield doc def begin_update(self, docs, drop=None, **cfg): outputs, backprop = self.model.begin_update(docs, drop=drop) def finish_update(docs, sgd=None): assert len(docs) d_lh = [] d_po = [] lh_lengths = [] po_lengths = [] for doc in docs: d_lh.append(doc._.pytt_d_last_hidden_state) d_po.append(doc._.pytt_d_pooler_output) lh_lengths.append(doc._.pytt_d_last_hidden_state.shape[0]) po_lengths.append(doc._.pytt_d_pooler_output.shape[0]) xp = self.model.ops.xp gradients = Activations( RaggedArray(xp.vstack(d_lh), lh_lengths), RaggedArray(xp.vstack(d_po), po_lengths), ) backprop(gradients, sgd=sgd) for doc in docs: doc._.pytt_d_last_hidden_state.fill(0) doc._.pytt_d_pooler_output.fill(0) return None return outputs, finish_update def predict(self, docs): return self.model.predict(docs) def set_annotations(self, docs: List[Doc], activations: Activations): xp = activations.xp for i, doc in enumerate(docs): # Make it 2d -- acts are always 3d, to represent batch size. wp_tensor = activations.lh.get(i) doc.tensor = self.model.ops.allocate((len(doc), self.model.nO)) doc._.pytt_last_hidden_state = wp_tensor if activations.has_po: pooler_output = activations.po.get(i) doc._.pytt_pooler_output = pooler_output doc._.pytt_d_last_hidden_state = xp.zeros((0, 0), dtype=wp_tensor.dtype) doc._.pytt_d_pooler_output = xp.zeros((0, 0), dtype=wp_tensor.dtype) doc._.pytt_d_all_hidden_states = [] doc._.pytt_d_all_attentions = [] if wp_tensor.shape != (len(doc._.pytt_word_pieces), self.model.nO): raise ValueError( "Mismatch between tensor shape and word pieces. This usually " "means we did something wrong in the sentence reshaping, " "or possibly finding the separator tokens." ) # Count how often each word-piece token is represented. This allows # a weighted sum, so that we can make sure doc.tensor.sum() # equals wp_tensor.sum(). Do this with sensitivity to boundary tokens wp_rows, align_sizes = _get_boundary_sensitive_alignment(doc) wp_weighted = wp_tensor / xp.array(align_sizes, dtype="f").reshape((-1, 1)) # TODO: Obviously incrementing the rows individually is bad. How # to do in one shot without blowing up the memory? for i, word_piece_slice in enumerate(wp_rows): for j in word_piece_slice: doc.tensor[i] += wp_weighted[j] doc.user_hooks["vector"] = get_doc_vector_via_tensor doc.user_span_hooks["vector"] = get_span_vector_via_tensor doc.user_token_hooks["vector"] = get_token_vector_via_tensor doc.user_hooks["similarity"] = get_similarity_via_tensor doc.user_span_hooks["similarity"] = get_similarity_via_tensor doc.user_token_hooks["similarity"] = get_similarity_via_tensor def _get_boundary_sensitive_alignment(doc): align_sizes = [0 for _ in range(len(doc._.pytt_word_pieces))] wp_rows = [] for word_piece_slice in doc._.pytt_alignment: wp_rows.append(list(word_piece_slice)) for i in word_piece_slice: align_sizes[i] += 1 # To make this weighting work, we "align" the boundary tokens against # every token in their sentence. The boundary tokens are otherwise # unaligned, which is how we identify them. for sent in get_sents(doc): offset = sent._.pytt_start for i in range(len(sent._.pytt_word_pieces)): if align_sizes[offset + i] == 0: align_sizes[offset + i] = len(sent) for tok in sent: wp_rows[tok.i].append(offset + i) return wp_rows, align_sizes def get_doc_vector_via_tensor(doc): return doc.tensor.sum(axis=0) def get_span_vector_via_tensor(span): return span.doc.tensor[span.start : span.end].sum(axis=0) def get_token_vector_via_tensor(token): return token.doc.tensor[token.i] def get_similarity_via_tensor(doc1, doc2): v1 = doc1.vector v2 = doc2.vector xp = get_array_module(v1) return xp.dot(v1, v2) / (doc1.vector_norm * doc2.vector_norm)
true
true
f717d7dbab9408394ce847244ea76b8cf45150f2
10,895
py
Python
tau.py
cradesto/pystella
f6f44ed12d9648585a52a09e15d494daa4c70c59
[ "MIT" ]
1
2019-08-08T13:11:57.000Z
2019-08-08T13:11:57.000Z
tau.py
cradesto/pystella
f6f44ed12d9648585a52a09e15d494daa4c70c59
[ "MIT" ]
null
null
null
tau.py
cradesto/pystella
f6f44ed12d9648585a52a09e15d494daa4c70c59
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 import argparse import logging import numpy as np import pystella as ps from pystella.model.sn_tau import StellaTauDetail mpl_logger = logging.getLogger('matplotlib') mpl_logger.setLevel(logging.WARNING) __author__ = 'bakl' # todo Show filters # todo show valuse for filters # todo compute SED = 4 pi R^2 sig T^4 def plot_tau_moments(tau, moments=None, xlim=None): import matplotlib.pyplot as plt moments = moments or np.exp(np.linspace(np.log(0.5), np.log(400.), 40)) fig, (axV, axT) = plt.subplots(2, figsize=(12, 12), sharex=True, gridspec_kw={'hspace': 0}) axV.set_title(tau.Name) axV.set_xlabel('') axV.set_ylabel('Velocity [1000 km/s]') axT.set_xlabel('Radius [cm]') axT.set_ylabel('Temperature [K]') for i, time in enumerate(moments): b = tau.block_nearest(time) n = int(2 - np.log10(max(1e-03, abs(b.Time)))) # if b.Time >= 10. else 4 # label format p = axV.semilogx(b.R, b.V8, label="t= {:.{}f}".format(b.Time, n)) color = p[0].get_color() axT.loglog(b.R, b.T, label="t={:.2f}".format(time), color=color) axV.legend(frameon=False) if xlim is not None: axT.set_xlim(xlim) axV.set_xlim(xlim) fig.tight_layout() return fig def plot_bands(ax, bnames, amp=30, alpha=0.5): """Plot the filter responses""" color_dic = ps.band.colors() res = {} for bname in bnames: b = ps.band.band_by_name(bname) wl = b.wl * ps.phys.cm_to_angs ax.plot(wl, b.resp_wl*amp, color_dic[bname], alpha=alpha) wl_eff = b.wl_eff_angs ax.axvline(x=wl_eff, ymin=0., ymax=0.99, linestyle='--', color=color_dic[bname], alpha=alpha) ax.text(wl_eff, 10, bname, fontsize=12) ax.text(wl_eff*.95, 3, "{:.0f}".format(wl_eff), fontsize=6) res[bname] = (wl_eff, color_dic[bname]) return res def plot_tau_phot(tau_data, pars, tau_ph, xlim=None, title='', bnames=None): """ Plot photosphere as Func(nu). Maybe: R, V, V8, T :param pars: the parameters of photosphere :param tau_data: the data at the optical depth tau_ph :param tau_ph: the photosphere location :param xlim: wave length interval [A] :param title: the plot title :param bnames: array of filter names to show the filter responses :return: figure """ import matplotlib.pyplot as plt def fr2wv(nu): return ps.phys.c / nu * ps.phys.cm_to_angs fig, axs = plt.subplots(len(pars)+1, figsize=(12, 12), sharex=True, gridspec_kw={'hspace': 0}) # Setup ax = axs[0] ax.set_ylabel(r'Zone ($\tau_{{ph}}= {:.2f}$)'.format(tau_ph)) ax.set_title(title) ax.xaxis.set_ticks_position('top') # ax.xaxis.tick_top() # ax.tick_params(axis="x", direction="in", pad=-22) # ax.tick_params(direction='in') for i, p in enumerate(pars, 1): ax = axs[i] ax.set_ylabel(r'{}$_{{ph}}$'.format(p)) if i < len(axs)-1: ax.set_xlabel('') ax.tick_params(which='both', top=False, bottom=False) else: ax.set_xlabel('Wavelength [A]') # Plot Zone_ph colors = [] for j, (t, freq, y) in enumerate(tau_data[StellaTauDetail.col_zon]): axzon = axs[0] n = int(3 - np.log10(max(1e-03, abs(t)))) # label format lbl = "t= {:.{}f} d".format(t, n) ll = axzon.semilogx(fr2wv(freq), y, label=lbl) color = ll[0].get_color() colors.append(color) bnames_waves = None if bnames is not None: ylim = axzon.get_ylim() bnames_waves = plot_bands(axzon, bnames, amp=ylim[1]*0.25, alpha=0.5) # Plot other params for i, p in enumerate(pars, 1): is_log = p.startswith('log') p_data = p.replace('log', '') if is_log else p ax = axs[i] for j, (t, freq, y) in enumerate(tau_data[p_data]): x = fr2wv(freq) if is_log: ax.loglog(x, y, color=colors[j]) else: ax.semilogx(x, y, color=colors[j]) if bnames_waves is not None: for bn, (wl, col) in bnames_waves.items(): ax.axvline(x=wl, ymin=0., ymax=0.99, linestyle='--', color=col, alpha=0.5) # Post-plotting for i, ax in enumerate(axs): ax.tick_params(which='both', left=True, right=True, direction="in") # ax.grid(axis="x", color="grey", alpha=.5, linewidth=1, linestyle=":") if xlim is not None: ax.set_xlim(xlim) axs[0].legend(frameon=False) fig.tight_layout() return fig def get_parser(times='0.1:1:10:25:65', bnames='U:B:V:R'): parser = argparse.ArgumentParser(description='Standard Candle Method.') print(" Plot the tau-wave diagram for STELLA models") parser.add_argument('-b', '--band', nargs='?', required=False, # default=bnames, const=bnames, type=str, dest="bnames", help="-b <bands>: string. If set only -b BNAMES is {}".format(bnames)) parser.add_argument('-i', '--input', required=True, dest="input", help="Model name, example: cat_R450_M15_Ni007") parser.add_argument('-p', '--path', required=False, type=str, default=False, dest="path", help="Model directory") parser.add_argument('-ph', '--phot', required=False, type=str, default=False, dest="phot", help='Plot photosphere parameter. Maybe: R, V, V8, T. Example: -ph R:V8:T ' 'You may use prefix log, e.g. logT or logV8') parser.add_argument('-s', '--save', action='store_const', const=True, dest="is_save", help="To save the result plot to pdf-file. Format: tau_[name]_t[times].pdf.") parser.add_argument('-t', '--time', required=False, type=str, default=times, dest="times", help="Plot tau snap for selected time moments. Default: {0}".format(times)) parser.add_argument('--tau_ph', required=False, type=float, default=2./3., dest="tau_ph", help="The optical depth at the photosphere. Default: 2/3") parser.add_argument('-x', '--xlim', required=False, type=str, default=None, dest="xlim", help="wave length interval [A]. Example: 1.:25e3. Default: all waves in the tau-file") parser.add_argument('-w', '--write', required=False, type=str, default=None, dest="write_prefix", help="The prefix of file + -ParamName.dat") return parser def str2float(s): return list(map(float, s.split(':'))) def main(): import os import sys try: import matplotlib.pyplot as plt except ImportError: plt = None ps.Band.load_settings() model_ext = '.tau' parser = get_parser() args, unknownargs = parser.parse_known_args() path = os.getcwd() if args.path: path = os.path.expanduser(args.path) # Set model names fname = None if args.input: fname = args.input.strip() fname = fname.replace(model_ext, '') if fname is None: parser.print_help() sys.exit(2) model = ps.Stella(fname, path=path) if not model.is_tau: print("No tau-data for: " + str(model)) return None fig = None xlim = None fplot = None print('\n Arguments') times = str2float(args.times) print(' The time moments: ', args.times) print(' The optical depth ', args.tau_ph) if args.phot: print(' The photospheric parameters ', args.phot) if args.xlim is not None: xlim = str2float(args.xlim) print(" xlim: ", xlim) # Set band names bnames = ('B',) ps.Band.load_settings() if args.bnames: bnames = [] for bname in args.bnames.split('-'): if not ps.band.is_exist(bname): print('No such band: ' + bname) parser.print_help() sys.exit(2) bnames.append(bname) tau = model.get_tau().load(is_info=False) print('\n Loaded data from {}'.format(tau.FName)) print('Model has Nzone= {} Ntimes= {}'.format(tau.Nzon, tau.Ntimes)) print("The model time interval: {:.3e} - {:3e} days".format(min(tau.Times), max(tau.Times))) print("The bnames are {}".format(', '.join(bnames))) # print(tau.Wl2angs) # tau = b.Tau # print(tau.shape) ### # Plot if args.phot: pars = args.phot.split(':') if isinstance(pars, str): pars = [pars] pars_data = [p.replace('log', '') for p in pars] tau_data = tau.params_ph(pars=pars_data, moments=times, tau_ph=args.tau_ph) if args.write_prefix: fwrite = os.path.expanduser(args.write_prefix) tau.data_save(fwrite, tau_data, pars_data) else: # Print parameters print('\nPhotospheric parameters:') for ii, p in enumerate(pars_data): print('{:9s} {}'.format('t_real', ' '.join([f'{p}_{b:10s}' for b in bnames]))) for i, (t, freq, y) in enumerate(tau_data[p]): s = '{:9.4f} '.format(t) for bname in bnames: b = ps.band.band_by_name(bname) fr_eff = b.freq_eff idx = (np.abs(freq - fr_eff)).argmin() s += ' {:10e}'.format( y[idx]) print(s) # Plot fig = plot_tau_phot(tau_data, pars, tau_ph=args.tau_ph, xlim=xlim, title=tau.Name, bnames=bnames) fplot = os.path.expanduser("~/tau_{}_{}.pdf".format(fname, str.replace(args.phot, ':', '-'))) else: fig = plot_tau_moments(tau, moments=times, xlim=xlim) if args.is_save: if fplot is None: fplot = os.path.expanduser("~/tau_{0}_t{1}.pdf".format(fname, str.replace(args.times, ':', '-'))) print("Save plot to {0}".format(fplot)) fig.savefig(fplot, bbox_inches='tight') else: plt.show() if __name__ == '__main__': main()
33.626543
110
0.534832
import argparse import logging import numpy as np import pystella as ps from pystella.model.sn_tau import StellaTauDetail mpl_logger = logging.getLogger('matplotlib') mpl_logger.setLevel(logging.WARNING) __author__ = 'bakl' def plot_tau_moments(tau, moments=None, xlim=None): import matplotlib.pyplot as plt moments = moments or np.exp(np.linspace(np.log(0.5), np.log(400.), 40)) fig, (axV, axT) = plt.subplots(2, figsize=(12, 12), sharex=True, gridspec_kw={'hspace': 0}) axV.set_title(tau.Name) axV.set_xlabel('') axV.set_ylabel('Velocity [1000 km/s]') axT.set_xlabel('Radius [cm]') axT.set_ylabel('Temperature [K]') for i, time in enumerate(moments): b = tau.block_nearest(time) n = int(2 - np.log10(max(1e-03, abs(b.Time)))) xV.semilogx(b.R, b.V8, label="t= {:.{}f}".format(b.Time, n)) color = p[0].get_color() axT.loglog(b.R, b.T, label="t={:.2f}".format(time), color=color) axV.legend(frameon=False) if xlim is not None: axT.set_xlim(xlim) axV.set_xlim(xlim) fig.tight_layout() return fig def plot_bands(ax, bnames, amp=30, alpha=0.5): color_dic = ps.band.colors() res = {} for bname in bnames: b = ps.band.band_by_name(bname) wl = b.wl * ps.phys.cm_to_angs ax.plot(wl, b.resp_wl*amp, color_dic[bname], alpha=alpha) wl_eff = b.wl_eff_angs ax.axvline(x=wl_eff, ymin=0., ymax=0.99, linestyle='--', color=color_dic[bname], alpha=alpha) ax.text(wl_eff, 10, bname, fontsize=12) ax.text(wl_eff*.95, 3, "{:.0f}".format(wl_eff), fontsize=6) res[bname] = (wl_eff, color_dic[bname]) return res def plot_tau_phot(tau_data, pars, tau_ph, xlim=None, title='', bnames=None): import matplotlib.pyplot as plt def fr2wv(nu): return ps.phys.c / nu * ps.phys.cm_to_angs fig, axs = plt.subplots(len(pars)+1, figsize=(12, 12), sharex=True, gridspec_kw={'hspace': 0}) ax = axs[0] ax.set_ylabel(r'Zone ($\tau_{{ph}}= {:.2f}$)'.format(tau_ph)) ax.set_title(title) ax.xaxis.set_ticks_position('top') for i, p in enumerate(pars, 1): ax = axs[i] ax.set_ylabel(r'{}$_{{ph}}$'.format(p)) if i < len(axs)-1: ax.set_xlabel('') ax.tick_params(which='both', top=False, bottom=False) else: ax.set_xlabel('Wavelength [A]') colors = [] for j, (t, freq, y) in enumerate(tau_data[StellaTauDetail.col_zon]): axzon = axs[0] n = int(3 - np.log10(max(1e-03, abs(t)))) lbl = "t= {:.{}f} d".format(t, n) ll = axzon.semilogx(fr2wv(freq), y, label=lbl) color = ll[0].get_color() colors.append(color) bnames_waves = None if bnames is not None: ylim = axzon.get_ylim() bnames_waves = plot_bands(axzon, bnames, amp=ylim[1]*0.25, alpha=0.5) for i, p in enumerate(pars, 1): is_log = p.startswith('log') p_data = p.replace('log', '') if is_log else p ax = axs[i] for j, (t, freq, y) in enumerate(tau_data[p_data]): x = fr2wv(freq) if is_log: ax.loglog(x, y, color=colors[j]) else: ax.semilogx(x, y, color=colors[j]) if bnames_waves is not None: for bn, (wl, col) in bnames_waves.items(): ax.axvline(x=wl, ymin=0., ymax=0.99, linestyle='--', color=col, alpha=0.5) for i, ax in enumerate(axs): ax.tick_params(which='both', left=True, right=True, direction="in") if xlim is not None: ax.set_xlim(xlim) axs[0].legend(frameon=False) fig.tight_layout() return fig def get_parser(times='0.1:1:10:25:65', bnames='U:B:V:R'): parser = argparse.ArgumentParser(description='Standard Candle Method.') print(" Plot the tau-wave diagram for STELLA models") parser.add_argument('-b', '--band', nargs='?', required=False, const=bnames, type=str, dest="bnames", help="-b <bands>: string. If set only -b BNAMES is {}".format(bnames)) parser.add_argument('-i', '--input', required=True, dest="input", help="Model name, example: cat_R450_M15_Ni007") parser.add_argument('-p', '--path', required=False, type=str, default=False, dest="path", help="Model directory") parser.add_argument('-ph', '--phot', required=False, type=str, default=False, dest="phot", help='Plot photosphere parameter. Maybe: R, V, V8, T. Example: -ph R:V8:T ' 'You may use prefix log, e.g. logT or logV8') parser.add_argument('-s', '--save', action='store_const', const=True, dest="is_save", help="To save the result plot to pdf-file. Format: tau_[name]_t[times].pdf.") parser.add_argument('-t', '--time', required=False, type=str, default=times, dest="times", help="Plot tau snap for selected time moments. Default: {0}".format(times)) parser.add_argument('--tau_ph', required=False, type=float, default=2./3., dest="tau_ph", help="The optical depth at the photosphere. Default: 2/3") parser.add_argument('-x', '--xlim', required=False, type=str, default=None, dest="xlim", help="wave length interval [A]. Example: 1.:25e3. Default: all waves in the tau-file") parser.add_argument('-w', '--write', required=False, type=str, default=None, dest="write_prefix", help="The prefix of file + -ParamName.dat") return parser def str2float(s): return list(map(float, s.split(':'))) def main(): import os import sys try: import matplotlib.pyplot as plt except ImportError: plt = None ps.Band.load_settings() model_ext = '.tau' parser = get_parser() args, unknownargs = parser.parse_known_args() path = os.getcwd() if args.path: path = os.path.expanduser(args.path) fname = None if args.input: fname = args.input.strip() fname = fname.replace(model_ext, '') if fname is None: parser.print_help() sys.exit(2) model = ps.Stella(fname, path=path) if not model.is_tau: print("No tau-data for: " + str(model)) return None fig = None xlim = None fplot = None print('\n Arguments') times = str2float(args.times) print(' The time moments: ', args.times) print(' The optical depth ', args.tau_ph) if args.phot: print(' The photospheric parameters ', args.phot) if args.xlim is not None: xlim = str2float(args.xlim) print(" xlim: ", xlim) bnames = ('B',) ps.Band.load_settings() if args.bnames: bnames = [] for bname in args.bnames.split('-'): if not ps.band.is_exist(bname): print('No such band: ' + bname) parser.print_help() sys.exit(2) bnames.append(bname) tau = model.get_tau().load(is_info=False) print('\n Loaded data from {}'.format(tau.FName)) print('Model has Nzone= {} Ntimes= {}'.format(tau.Nzon, tau.Ntimes)) print("The model time interval: {:.3e} - {:3e} days".format(min(tau.Times), max(tau.Times))) print("The bnames are {}".format(', '.join(bnames))) if args.phot: pars = args.phot.split(':') if isinstance(pars, str): pars = [pars] pars_data = [p.replace('log', '') for p in pars] tau_data = tau.params_ph(pars=pars_data, moments=times, tau_ph=args.tau_ph) if args.write_prefix: fwrite = os.path.expanduser(args.write_prefix) tau.data_save(fwrite, tau_data, pars_data) else: print('\nPhotospheric parameters:') for ii, p in enumerate(pars_data): print('{:9s} {}'.format('t_real', ' '.join([f'{p}_{b:10s}' for b in bnames]))) for i, (t, freq, y) in enumerate(tau_data[p]): s = '{:9.4f} '.format(t) for bname in bnames: b = ps.band.band_by_name(bname) fr_eff = b.freq_eff idx = (np.abs(freq - fr_eff)).argmin() s += ' {:10e}'.format( y[idx]) print(s) fig = plot_tau_phot(tau_data, pars, tau_ph=args.tau_ph, xlim=xlim, title=tau.Name, bnames=bnames) fplot = os.path.expanduser("~/tau_{}_{}.pdf".format(fname, str.replace(args.phot, ':', '-'))) else: fig = plot_tau_moments(tau, moments=times, xlim=xlim) if args.is_save: if fplot is None: fplot = os.path.expanduser("~/tau_{0}_t{1}.pdf".format(fname, str.replace(args.times, ':', '-'))) print("Save plot to {0}".format(fplot)) fig.savefig(fplot, bbox_inches='tight') else: plt.show() if __name__ == '__main__': main()
true
true
f717d82ecf5183dae516e756dfbcb6f492d9702a
1,019
py
Python
tests/__init__.py
dolfinus/pexpect
3453ea9b8b326179cf720351001e64c7ea6b07bc
[ "0BSD" ]
2,132
2015-01-02T12:48:45.000Z
2022-03-28T05:32:54.000Z
tests/__init__.py
dolfinus/pexpect
3453ea9b8b326179cf720351001e64c7ea6b07bc
[ "0BSD" ]
1,274
2015-09-22T20:06:16.000Z
2018-08-31T22:14:00.000Z
tests/__init__.py
dolfinus/pexpect
3453ea9b8b326179cf720351001e64c7ea6b07bc
[ "0BSD" ]
517
2015-01-07T02:09:44.000Z
2022-03-26T14:18:23.000Z
''' PEXPECT LICENSE This license is approved by the OSI and FSF as GPL-compatible. http://opensource.org/licenses/isc-license.txt Copyright (c) 2012, Noah Spurrier <noah@noah.org> PERMISSION TO USE, COPY, MODIFY, AND/OR DISTRIBUTE THIS SOFTWARE FOR ANY PURPOSE WITH OR WITHOUT FEE IS HEREBY GRANTED, PROVIDED THAT THE ABOVE COPYRIGHT NOTICE AND THIS PERMISSION NOTICE APPEAR IN ALL COPIES. THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. ''' # __init__.py # The mere presence of this file makes the dir a package. pass
39.192308
76
0.750736
pass
true
true
f717da674668d80ac70707def8bcf6d62e7209d0
2,410
py
Python
git_nit/tests/test_parse_review_id.py
dhellmann/git-nit
9fce9eb8806d6997182107eb5d755a1220fa5e88
[ "Apache-2.0" ]
8
2018-04-27T07:03:50.000Z
2018-10-02T08:05:40.000Z
git_nit/tests/test_parse_review_id.py
dhellmann/git-nit
9fce9eb8806d6997182107eb5d755a1220fa5e88
[ "Apache-2.0" ]
null
null
null
git_nit/tests/test_parse_review_id.py
dhellmann/git-nit
9fce9eb8806d6997182107eb5d755a1220fa5e88
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from git_nit import cmd import testscenarios.testcase import testtools class ParseReviewIDTest(testscenarios.testcase.WithScenarios, testtools.TestCase): scenarios = [ ('fragment with patchset', { 'url': 'https://review.openstack.org/#/c/564559/5/', 'review': '564559', 'patchset': '5', }), ('fragment with patchset, no trailing slash', { 'url': 'https://review.openstack.org/#/c/564559/5', 'review': '564559', 'patchset': '5', }), ('fragment without patchset', { 'url': 'https://review.openstack.org/#/c/564559/', 'review': '564559', 'patchset': None, }), ('fragment without patchset, no trailing slash', { 'url': 'https://review.openstack.org/#/c/564559', 'review': '564559', 'patchset': None, }), ('path with patchset', { 'url': 'https://review.openstack.org/564559/5/', 'review': '564559', 'patchset': '5', }), ('path with patchset, no trailing slash', { 'url': 'https://review.openstack.org/564559/5', 'review': '564559', 'patchset': '5', }), ('path without patchset', { 'url': 'https://review.openstack.org/564559/', 'review': '564559', 'patchset': None, }), ('path without patchset, no trailing slash', { 'url': 'https://review.openstack.org/564559', 'review': '564559', 'patchset': None, }), ] def test(self): review, patchset = cmd.parse_review_id(self.url) self.assertEqual( (self.review, self.patchset), (review, patchset), )
33.013699
74
0.554357
from git_nit import cmd import testscenarios.testcase import testtools class ParseReviewIDTest(testscenarios.testcase.WithScenarios, testtools.TestCase): scenarios = [ ('fragment with patchset', { 'url': 'https://review.openstack.org/#/c/564559/5/', 'review': '564559', 'patchset': '5', }), ('fragment with patchset, no trailing slash', { 'url': 'https://review.openstack.org/#/c/564559/5', 'review': '564559', 'patchset': '5', }), ('fragment without patchset', { 'url': 'https://review.openstack.org/#/c/564559/', 'review': '564559', 'patchset': None, }), ('fragment without patchset, no trailing slash', { 'url': 'https://review.openstack.org/#/c/564559', 'review': '564559', 'patchset': None, }), ('path with patchset', { 'url': 'https://review.openstack.org/564559/5/', 'review': '564559', 'patchset': '5', }), ('path with patchset, no trailing slash', { 'url': 'https://review.openstack.org/564559/5', 'review': '564559', 'patchset': '5', }), ('path without patchset', { 'url': 'https://review.openstack.org/564559/', 'review': '564559', 'patchset': None, }), ('path without patchset, no trailing slash', { 'url': 'https://review.openstack.org/564559', 'review': '564559', 'patchset': None, }), ] def test(self): review, patchset = cmd.parse_review_id(self.url) self.assertEqual( (self.review, self.patchset), (review, patchset), )
true
true
f717da9d442484cb8be34d4f36e19f573b0849be
10,841
py
Python
happybase_mock/table.py
evgenia-ch/happybase-mock
6fbf4a4f9685829b32ad8dc3de3e01b2a9fba964
[ "MIT" ]
null
null
null
happybase_mock/table.py
evgenia-ch/happybase-mock
6fbf4a4f9685829b32ad8dc3de3e01b2a9fba964
[ "MIT" ]
null
null
null
happybase_mock/table.py
evgenia-ch/happybase-mock
6fbf4a4f9685829b32ad8dc3de3e01b2a9fba964
[ "MIT" ]
3
2018-05-15T14:10:23.000Z
2020-08-12T13:45:28.000Z
import struct import time import six from six import iteritems from .batch import Batch def _check_table_existence(method): def wrap(table, *args, **kwargs): if not table._exists(): raise IOError('TableNotFoundException: %s' % table.name) return method(table, *args, **kwargs) return wrap # Copied from happybase.util def bytes_increment(b): """Increment and truncate a byte string (for sorting purposes) This functions returns the shortest string that sorts after the given string when compared using regular string comparison semantics. This function increments the last byte that is smaller than ``0xFF``, and drops everything after it. If the string only contains ``0xFF`` bytes, `None` is returned. """ assert isinstance(b, six.binary_type) b = bytearray(b) # Used subset of its API is the same on Python 2 and 3. for i in range(len(b) - 1, -1, -1): if b[i] != 0xff: b[i] += 1 return bytes(b[:i+1]) return None class Table(object): def __init__(self, name, connection): self.name = name self.connection = connection self._enabled = True # A multi-dimentional map, _data[rowkey][colname][timestamp] = value self._data = {} def __repr__(self): return '<%s.%s name=%r>' % ( __name__, self.__class__.__name__, self.name, ) @_check_table_existence def families(self): return self._families def regions(self): if not self._exists(): return [] # Table.regions() is meaningless for in-memory mocking, so just return # some fake data return [{ 'end_key': '', 'id': 1, 'name': '%s,,1.1234' % self.name, 'port': 60000, 'server_name': 'localhost', 'start_key': '', 'version': 1 }] @_check_table_existence def row(self, row, columns=None, timestamp=None, include_timestamp=False): if not isinstance(row, bytes): row = row.encode('utf-8') data = self._data.get(row, {}) result = {} if not columns: columns = data.keys() for colname in columns: if colname in data: cell = data[colname] timestamps = sorted(cell.keys(), reverse=True) if timestamp is None: # Use latest version if timestamp isn't specified ts = timestamps[0] if include_timestamp: result[colname] = cell[ts], ts else: result[colname] = cell[ts] else: # Find the first ts < timestamp for ts in timestamps: if ts < timestamp: if include_timestamp: result[colname] = cell[ts], ts else: result[colname] = cell[ts] break return result @_check_table_existence def rows(self, rows, columns=None, timestamp=None, include_timestamp=False): result = [] for row in rows: data = self.row(row, columns, timestamp, include_timestamp) result.append((row, data)) return result @_check_table_existence def cells(self, row, column, versions=None, timestamp=None, include_timestamp=False): if not isinstance(row, bytes): row = row.encode('utf-8') if not isinstance(column, bytes): column = column.encode('utf-8') result = [] timestamps = sorted(self._data.get(row, {}).get(column, {}).keys(), reverse=True) for ts in timestamps: value = self._data[row][column][ts] if timestamp is None or ts < timestamp: if include_timestamp: result.append((value, ts)) else: result.append(value) return result @_check_table_existence def scan(self, row_start=None, row_stop=None, row_prefix=None, columns=None, timestamp=None, include_timestamp=False, batch_size=1000, scan_batching=None, limit=None, reverse=False, sorted_columns=False, **kwargs): # encode columns key and data (for python3 compatibility) if reverse: old_row_start = row_start row_start = row_stop row_stop = old_row_start if columns: for i, col in enumerate(columns): if not isinstance(col, bytes): columns[i] = col.encode('utf-8') if row_prefix is not None: if not isinstance(row_prefix, bytes): row_prefix = row_prefix.encode('utf-8') if row_start is not None or row_stop is not None: raise TypeError( "'row_prefix' cannot be combined with 'row_start' " "or 'row_stop'") row_start = row_prefix row_stop = bytes_increment(row_prefix) if row_start is None: row_start = b'' else: if not isinstance(row_start, bytes): row_start = row_start.encode('utf-8') if columns: rows = (k for k, v in self._data.items() if set(columns).intersection(v)) else: rows = self._data.keys() if not reverse: rows = filter(lambda k: k >= row_start, rows) else: rows = filter(lambda k: k > row_start, rows) if row_stop is not None: if not isinstance(row_stop, bytes): row_stop = row_stop.encode('utf-8') if not reverse: rows = filter(lambda k: k < row_stop, rows) else: rows = filter(lambda k: k <= row_stop, rows) result = sorted([ (row, self.row(row, columns, timestamp, include_timestamp)) for row in rows ], reverse=reverse) if limit: if len(result) > limit: result = result[:limit] return iter(result) @_check_table_existence def put(self, row, data, timestamp=None, wal=True): # encode row key and data before put (for python3 compatibility) if not isinstance(row, bytes): row = row.encode('utf-8') data = { (k if isinstance(k, bytes) else k.encode('utf-8')): (v if isinstance(v, bytes) else v.encode('utf-8')) for k, v in iteritems(data) } # Check data against column families for colname in data: cf = colname.decode('utf-8').split(':')[0] if cf not in self._families: raise IOError('NoSuchColumnFamilyException: %s' % cf) if timestamp is None: timestamp = int(time.time() * 1000) columns = self._data.get(row) if columns is None: columns = {} self._data[row] = columns for colname, value in iteritems(data): column = columns.get(colname) if column is None: column = {} columns[colname] = column column[timestamp] = value # Check if it exceeds max_versions cf = colname.decode('utf-8').split(':')[0] max_versions = self._max_versions(cf) if len(column) > max_versions: # Delete cell with minimum timestamp del column[min(column.keys())] @_check_table_existence def delete(self, row, columns=None, timestamp=None, wal=True): if not isinstance(row, bytes): row = row.encode('utf-8') if columns: columns = [ column if isinstance(column, bytes) else column.encode('utf-8') for column in columns ] if not columns and timestamp is None: # Delete whole row self._data.pop(row, None) elif row in self._data: data = self._data[row] if not columns: # Delete all columns if not specified columns = data.keys() else: columns = list(set(columns) & data.keys()) if timestamp is None: timestamp = int(time.time() * 1000) to_be_deleted = [] for colname in columns: for ts in data[colname]: if ts <= timestamp: to_be_deleted.append((colname, ts)) for colname, ts in to_be_deleted: del data[colname][ts] if not data[colname]: # Delete a column if it doesn't have any timestamps del data[colname] def batch(self, timestamp=None, batch_size=None, transaction=False, wal=True): return Batch(self, timestamp, batch_size, transaction, wal) def counter_get(self, row, column): # Decode as long integer, big endian value = self.row(row, (column,)).get(column) if not value: return 0 return struct.unpack('>q', value)[0] @_check_table_existence def counter_set(self, row, column, value=0): # Encode as long integer, big endian value = struct.pack('>q', value) self.delete(row, (column,)) self.put(row, {column: value}) @_check_table_existence def counter_inc(self, row, column, value=1): orig_value = self.counter_get(row, column) self.counter_set(row, column, orig_value + value) @_check_table_existence def counter_dec(self, row, column, value=1): orig_value = self.counter_get(row, column) self.counter_set(row, column, orig_value - value) def _exists(self): return self.name in self.connection._tables def _max_versions(self, cf): return self._families[cf]['max_versions'] def _set_families(self, families): # Default family options defaults = { 'block_cache_enabled': False, 'bloom_filter_nb_hashes': 0, 'bloom_filter_type': 'NONE', 'bloom_filter_vector_size': 0, 'compression': 'NONE', 'in_memory': False, 'max_versions': 3, 'time_to_live': -1 } self._families = {} for name, opts in iteritems(families): family_options = defaults.copy() family_options['name'] = name family_options.update(opts) self._families[name] = family_options
33.984326
85
0.54091
import struct import time import six from six import iteritems from .batch import Batch def _check_table_existence(method): def wrap(table, *args, **kwargs): if not table._exists(): raise IOError('TableNotFoundException: %s' % table.name) return method(table, *args, **kwargs) return wrap def bytes_increment(b): assert isinstance(b, six.binary_type) b = bytearray(b) for i in range(len(b) - 1, -1, -1): if b[i] != 0xff: b[i] += 1 return bytes(b[:i+1]) return None class Table(object): def __init__(self, name, connection): self.name = name self.connection = connection self._enabled = True self._data = {} def __repr__(self): return '<%s.%s name=%r>' % ( __name__, self.__class__.__name__, self.name, ) @_check_table_existence def families(self): return self._families def regions(self): if not self._exists(): return [] return [{ 'end_key': '', 'id': 1, 'name': '%s,,1.1234' % self.name, 'port': 60000, 'server_name': 'localhost', 'start_key': '', 'version': 1 }] @_check_table_existence def row(self, row, columns=None, timestamp=None, include_timestamp=False): if not isinstance(row, bytes): row = row.encode('utf-8') data = self._data.get(row, {}) result = {} if not columns: columns = data.keys() for colname in columns: if colname in data: cell = data[colname] timestamps = sorted(cell.keys(), reverse=True) if timestamp is None: ts = timestamps[0] if include_timestamp: result[colname] = cell[ts], ts else: result[colname] = cell[ts] else: # Find the first ts < timestamp for ts in timestamps: if ts < timestamp: if include_timestamp: result[colname] = cell[ts], ts else: result[colname] = cell[ts] break return result @_check_table_existence def rows(self, rows, columns=None, timestamp=None, include_timestamp=False): result = [] for row in rows: data = self.row(row, columns, timestamp, include_timestamp) result.append((row, data)) return result @_check_table_existence def cells(self, row, column, versions=None, timestamp=None, include_timestamp=False): if not isinstance(row, bytes): row = row.encode('utf-8') if not isinstance(column, bytes): column = column.encode('utf-8') result = [] timestamps = sorted(self._data.get(row, {}).get(column, {}).keys(), reverse=True) for ts in timestamps: value = self._data[row][column][ts] if timestamp is None or ts < timestamp: if include_timestamp: result.append((value, ts)) else: result.append(value) return result @_check_table_existence def scan(self, row_start=None, row_stop=None, row_prefix=None, columns=None, timestamp=None, include_timestamp=False, batch_size=1000, scan_batching=None, limit=None, reverse=False, sorted_columns=False, **kwargs): # encode columns key and data (for python3 compatibility) if reverse: old_row_start = row_start row_start = row_stop row_stop = old_row_start if columns: for i, col in enumerate(columns): if not isinstance(col, bytes): columns[i] = col.encode('utf-8') if row_prefix is not None: if not isinstance(row_prefix, bytes): row_prefix = row_prefix.encode('utf-8') if row_start is not None or row_stop is not None: raise TypeError( "'row_prefix' cannot be combined with 'row_start' " "or 'row_stop'") row_start = row_prefix row_stop = bytes_increment(row_prefix) if row_start is None: row_start = b'' else: if not isinstance(row_start, bytes): row_start = row_start.encode('utf-8') if columns: rows = (k for k, v in self._data.items() if set(columns).intersection(v)) else: rows = self._data.keys() if not reverse: rows = filter(lambda k: k >= row_start, rows) else: rows = filter(lambda k: k > row_start, rows) if row_stop is not None: if not isinstance(row_stop, bytes): row_stop = row_stop.encode('utf-8') if not reverse: rows = filter(lambda k: k < row_stop, rows) else: rows = filter(lambda k: k <= row_stop, rows) result = sorted([ (row, self.row(row, columns, timestamp, include_timestamp)) for row in rows ], reverse=reverse) if limit: if len(result) > limit: result = result[:limit] return iter(result) @_check_table_existence def put(self, row, data, timestamp=None, wal=True): # encode row key and data before put (for python3 compatibility) if not isinstance(row, bytes): row = row.encode('utf-8') data = { (k if isinstance(k, bytes) else k.encode('utf-8')): (v if isinstance(v, bytes) else v.encode('utf-8')) for k, v in iteritems(data) } # Check data against column families for colname in data: cf = colname.decode('utf-8').split(':')[0] if cf not in self._families: raise IOError('NoSuchColumnFamilyException: %s' % cf) if timestamp is None: timestamp = int(time.time() * 1000) columns = self._data.get(row) if columns is None: columns = {} self._data[row] = columns for colname, value in iteritems(data): column = columns.get(colname) if column is None: column = {} columns[colname] = column column[timestamp] = value # Check if it exceeds max_versions cf = colname.decode('utf-8').split(':')[0] max_versions = self._max_versions(cf) if len(column) > max_versions: # Delete cell with minimum timestamp del column[min(column.keys())] @_check_table_existence def delete(self, row, columns=None, timestamp=None, wal=True): if not isinstance(row, bytes): row = row.encode('utf-8') if columns: columns = [ column if isinstance(column, bytes) else column.encode('utf-8') for column in columns ] if not columns and timestamp is None: # Delete whole row self._data.pop(row, None) elif row in self._data: data = self._data[row] if not columns: # Delete all columns if not specified columns = data.keys() else: columns = list(set(columns) & data.keys()) if timestamp is None: timestamp = int(time.time() * 1000) to_be_deleted = [] for colname in columns: for ts in data[colname]: if ts <= timestamp: to_be_deleted.append((colname, ts)) for colname, ts in to_be_deleted: del data[colname][ts] if not data[colname]: # Delete a column if it doesn't have any timestamps del data[colname] def batch(self, timestamp=None, batch_size=None, transaction=False, wal=True): return Batch(self, timestamp, batch_size, transaction, wal) def counter_get(self, row, column): value = self.row(row, (column,)).get(column) if not value: return 0 return struct.unpack('>q', value)[0] @_check_table_existence def counter_set(self, row, column, value=0): value = struct.pack('>q', value) self.delete(row, (column,)) self.put(row, {column: value}) @_check_table_existence def counter_inc(self, row, column, value=1): orig_value = self.counter_get(row, column) self.counter_set(row, column, orig_value + value) @_check_table_existence def counter_dec(self, row, column, value=1): orig_value = self.counter_get(row, column) self.counter_set(row, column, orig_value - value) def _exists(self): return self.name in self.connection._tables def _max_versions(self, cf): return self._families[cf]['max_versions'] def _set_families(self, families): defaults = { 'block_cache_enabled': False, 'bloom_filter_nb_hashes': 0, 'bloom_filter_type': 'NONE', 'bloom_filter_vector_size': 0, 'compression': 'NONE', 'in_memory': False, 'max_versions': 3, 'time_to_live': -1 } self._families = {} for name, opts in iteritems(families): family_options = defaults.copy() family_options['name'] = name family_options.update(opts) self._families[name] = family_options
true
true
f717dab203ae844f0e1a238ee942846c22823c19
38,297
py
Python
test/terra/backends/qasm_simulator/qasm_snapshot.py
ares201005/qiskit-aer
fb3bab00ab810e73ad333b0f538fa6c3c53f054e
[ "Apache-2.0" ]
null
null
null
test/terra/backends/qasm_simulator/qasm_snapshot.py
ares201005/qiskit-aer
fb3bab00ab810e73ad333b0f538fa6c3c53f054e
[ "Apache-2.0" ]
null
null
null
test/terra/backends/qasm_simulator/qasm_snapshot.py
ares201005/qiskit-aer
fb3bab00ab810e73ad333b0f538fa6c3c53f054e
[ "Apache-2.0" ]
null
null
null
# This code is part of Qiskit. # # (C) Copyright IBM 2018, 2019. # # This code is licensed under the Apache License, Version 2.0. You may # obtain a copy of this license in the LICENSE.txt file in the root directory # of this source tree or at http://www.apache.org/licenses/LICENSE-2.0. # # Any modifications or derivative works of this code must retain this # copyright notice, and modified files need to carry a notice indicating # that they have been altered from the originals. """ QasmSimulator Integration Tests for Snapshot instructions """ import logging import itertools as it import numpy as np from qiskit import QuantumCircuit from qiskit.compiler import assemble from qiskit.quantum_info import DensityMatrix, Pauli, Operator from qiskit.providers.aer import QasmSimulator from qiskit.providers.aer import AerError from test.terra.reference.ref_snapshot_state import ( snapshot_state_circuits_deterministic, snapshot_state_counts_deterministic, snapshot_state_pre_measure_statevector_deterministic, snapshot_state_post_measure_statevector_deterministic, snapshot_state_circuits_nondeterministic, snapshot_state_counts_nondeterministic, snapshot_state_pre_measure_statevector_nondeterministic, snapshot_state_post_measure_statevector_nondeterministic) from test.terra.reference.ref_snapshot_probabilities import ( snapshot_probabilities_circuits, snapshot_probabilities_counts, snapshot_probabilities_labels_qubits, snapshot_probabilities_post_meas_probs, snapshot_probabilities_pre_meas_probs) from test.terra.reference.ref_snapshot_expval import ( snapshot_expval_circuits, snapshot_expval_counts, snapshot_expval_labels, snapshot_expval_post_meas_values, snapshot_expval_pre_meas_values) class QasmSnapshotStatevectorTests: """QasmSimulator snapshot statevector tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'matrix_product_state' ] BACKEND_OPTS = {} def statevector_snapshots(self, data, label): """Format snapshots as list of Numpy arrays""" snaps = data.get("snapshots", {}).get("statevector", {}).get(label, []) statevecs = [] for snap in snaps: self.assertIsInstance(snap, np.ndarray) statevecs.append(snap) return statevecs def test_snapshot_statevector_pre_measure_det(self): """Test snapshot statevector before deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'statevector', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) self.assertTrue(len(snaps), 1) target = statevec_targets[j] value = snaps[0] self.assertTrue(np.allclose(value, target)) def test_snapshot_statevector_pre_measure_nondet(self): """Test snapshot statevector before non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'statevector', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) self.assertTrue(len(snaps), 1) target = statevec_targets[j] value = snaps[0] self.assertTrue(np.allclose(value, target)) def test_snapshot_statevector_post_measure_det(self): """Test snapshot statevector after deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'statevector', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: logging.getLogger().setLevel(logging.CRITICAL) self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) self.assertTrue(np.allclose(snaps[j], target)) def test_snapshot_statevector_post_measure_nondet(self): """Test snapshot statevector after non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'statevector', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) self.assertTrue(np.allclose(snaps[j], target)) class QasmSnapshotStabilizerTests: """QasmSimulator method snapshot stabilizer tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = ['automatic', 'stabilizer'] BACKEND_OPTS = {} @staticmethod def stabilizer_snapshots(data, label): """Get stabilizer snapshots""" return data.get("snapshots", {}).get("stabilizer", {}).get(label, []) @staticmethod def stabilizes_statevector(stabilizer, statevector): """Return True if two stabilizer states are equal.""" # Get stabilizer and destabilizers and convert to sets for stab in stabilizer: if stab[0] == '-': pauli_mat = -1 * Pauli.from_label(stab[1:]).to_matrix() else: pauli_mat = Pauli.from_label(stab).to_matrix() val = statevector.conj().dot(pauli_mat.dot(statevector)) if not np.isclose(val, 1): return False return True def test_snapshot_stabilizer_pre_measure_det(self): """Test snapshot stabilizer before deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'stabilizer', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) self.assertEqual(len(snaps), 1) statevec = statevec_targets[j] stabilizer = snaps[0] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_pre_measure_nondet(self): """Test snapshot stabilizer before non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'stabilizer', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) self.assertEqual(len(snaps), 1) statevec = statevec_targets[j] stabilizer = snaps[0] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_post_measure_det(self): """Test snapshot stabilizer after deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'stabilizer', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) for j, mem in enumerate(data['memory']): statevec = statevec_targets[i].get(mem) stabilizer = snaps[j] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_post_measure_nondet(self): """Test snapshot stabilizer after non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'stabilizer', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) for j, mem in enumerate(data['memory']): statevec = statevec_targets[i].get(mem) stabilizer = snaps[j] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) class QasmSnapshotDensityMatrixTests: """QasmSimulator snapshot density matrix tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust' ] BACKEND_OPTS = {} def density_snapshots(self, data, label): """Format snapshots as list of Numpy arrays""" # Check snapshot entry exists in data snaps = data.get("snapshots", {}).get("density_matrix", {}).get(label, []) # Convert nested lists to numpy arrays output = {} for snap_dict in snaps: memory = snap_dict['memory'] self.assertIsInstance(snap_dict['value'], np.ndarray) output[memory] = snap_dict['value'] return output def test_snapshot_density_matrix_pre_measure_det(self): """Test snapshot density matrix before deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'density_matrix', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) self.assertTrue(len(snaps), 1) target = np.outer(statevec_targets[j], statevec_targets[j].conj()) # Pre-measurement all memory bits should be 0 value = snaps.get('0x0') self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_pre_measure_nondet(self): """Test snapshot density matrix before non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'density_matrix', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) self.assertTrue(len(snaps), 1) target = np.outer(statevec_targets[j], statevec_targets[j].conj()) value = snaps.get('0x0') self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_post_measure_det(self): """Test snapshot density matrix after deterministic final measurement""" shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'density_matrix', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) target = np.outer(target, target.conj()) value = snaps.get(mem) self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_post_measure_nondet(self): """Test snapshot density matrix after non-deterministic final measurement""" shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'density_matrix', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) # Check snapshots for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) target = np.outer(target, target.conj()) value = snaps.get(mem) self.assertTrue(np.allclose(value, target)) class QasmSnapshotProbabilitiesTests: """QasmSimulator snapshot probabilities tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'stabilizer', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', ] BACKEND_OPTS = {} @staticmethod def probability_snapshots(data, labels): """Format snapshots as nested dicts""" # Check snapshot entry exists in data output = {} for label in labels: snaps = data.get("snapshots", {}).get("probabilities", {}).get(label, []) output[label] = { snap_dict['memory']: snap_dict['value'] for snap_dict in snaps } return output def test_snapshot_probabilities_pre_measure(self): """Test snapshot probabilities before final measurement""" shots = 1000 labels = list(snapshot_probabilities_labels_qubits().keys()) counts_targets = snapshot_probabilities_counts(shots) prob_targets = snapshot_probabilities_pre_meas_probs() circuits = snapshot_probabilities_circuits(post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotProbabilitiesTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.probability_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = prob_targets[j].get(label, {}).get(memory, {}) self.assertDictAlmostEqual(value, target, delta=1e-7) def test_snapshot_probabilities_post_measure(self): """Test snapshot probabilities after final measurement""" shots = 1000 labels = list(snapshot_probabilities_labels_qubits().keys()) counts_targets = snapshot_probabilities_counts(shots) prob_targets = snapshot_probabilities_post_meas_probs() circuits = snapshot_probabilities_circuits(post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotProbabilitiesTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.probability_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) for memory, value in snaps.items(): target = prob_targets[j].get(label, {}).get(memory, {}) self.assertDictAlmostEqual(value, target, delta=1e-7) class QasmSnapshotExpValPauliTests: """QasmSimulator snapshot pauli expectation value tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', 'stabilizer' ] BACKEND_OPTS = {} @staticmethod def expval_snapshots(data, labels): """Format snapshots as nested dicts""" # Check snapshot entry exists in data output = {} for label in labels: snaps = data.get("snapshots", {}).get("expectation_value", {}).get(label, []) # Convert list into dict inner = {} for snap_dict in snaps: val = snap_dict['value'] inner[snap_dict['memory']] = val output[label] = inner return output def test_snapshot_expval_pauli_pre_measure(self): """Test snapshot expectation value (pauli) before final measurement""" shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_pre_meas_values() circuits = snapshot_expval_circuits(pauli=True, post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValPauliTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) def test_snapshot_expval_pauli_post_measure(self): """Test snapshot expectation value (pauli) after final measurement""" shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_post_meas_values() circuits = snapshot_expval_circuits(pauli=True, post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValPauliTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) class QasmSnapshotExpvalPauliNCTests: """QasmSimulator snapshot pauli expectation value tests on random states.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', ] BACKEND_OPTS = {} def general_test(self, pauli, num_qubits=None, seed=None): """General test case""" pauli_qubits = list(range(len(pauli))) if num_qubits is None: num_qubits = len(pauli_qubits) # Prepare random N-qubit product input state # from seed rng = np.random.default_rng(seed) params = rng.uniform(-1, 1, size=(num_qubits, 3)) init_circ = QuantumCircuit(num_qubits) for i, par in enumerate(params): init_circ.u3(*par, i) # Compute the target expectation value rho = DensityMatrix.from_instruction(init_circ) op = Operator.from_label(pauli) target = np.trace(Operator(rho).compose(op, pauli_qubits).data) # Simulate expectation value qc = init_circ.copy() qc.snapshot_expectation_value('final', [(1, pauli)], pauli_qubits) qobj = assemble(qc) result = self.SIMULATOR.run( qobj, backend_options=self.BACKEND_OPTS).result() self.assertTrue(getattr(result, 'success', False)) snapshots = result.data(0).get('snapshots', {}) self.assertIn('expectation_value', snapshots) self.assertIn('final', snapshots['expectation_value']) expval = snapshots.get('expectation_value', {})['final'][0]['value'] self.assertAlmostEqual(expval, target) def test_pauli1(self): """Test all 1-qubit Pauli snapshots.""" seed = 100 for tup in ['I', 'X', 'Y', 'Z']: pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) def test_pauli2(self): """Test all 2-qubit Pauli snapshots.""" seed = 100 for tup in it.product(['I', 'X', 'Y', 'Z'], repeat=2): pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) def test_pauli3(self): """Test all 3-qubit Pauli snapshots.""" seed = 100 for tup in it.product(['I', 'X', 'Y', 'Z'], repeat=3): pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) class QasmSnapshotExpValMatrixTests: """QasmSimulator snapshot pauli expectation value tests.""" SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'matrix_product_state' ] BACKEND_OPTS = {} @staticmethod def expval_snapshots(data, labels): """Format snapshots as nested dicts""" # Check snapshot entry exists in data output = {} for label in labels: snaps = data.get("snapshots", {}).get("expectation_value", {}).get(label, []) # Convert list into dict inner = {} for snap_dict in snaps: inner[snap_dict['memory']] = snap_dict['value'] output[label] = inner return output def test_snapshot_expval_matrix_pre_measure(self): """Test snapshot expectation value (matrix) before final measurement""" shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_pre_meas_values() circuits = snapshot_expval_circuits(pauli=False, post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) def test_snapshot_expval_matrix_post_measure(self): """Test snapshot expectation value (matrix) after final measurement""" shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_post_meas_values() circuits = snapshot_expval_circuits(pauli=False, post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) # Check snapshots for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7)
44.844262
85
0.589733
import logging import itertools as it import numpy as np from qiskit import QuantumCircuit from qiskit.compiler import assemble from qiskit.quantum_info import DensityMatrix, Pauli, Operator from qiskit.providers.aer import QasmSimulator from qiskit.providers.aer import AerError from test.terra.reference.ref_snapshot_state import ( snapshot_state_circuits_deterministic, snapshot_state_counts_deterministic, snapshot_state_pre_measure_statevector_deterministic, snapshot_state_post_measure_statevector_deterministic, snapshot_state_circuits_nondeterministic, snapshot_state_counts_nondeterministic, snapshot_state_pre_measure_statevector_nondeterministic, snapshot_state_post_measure_statevector_nondeterministic) from test.terra.reference.ref_snapshot_probabilities import ( snapshot_probabilities_circuits, snapshot_probabilities_counts, snapshot_probabilities_labels_qubits, snapshot_probabilities_post_meas_probs, snapshot_probabilities_pre_meas_probs) from test.terra.reference.ref_snapshot_expval import ( snapshot_expval_circuits, snapshot_expval_counts, snapshot_expval_labels, snapshot_expval_post_meas_values, snapshot_expval_pre_meas_values) class QasmSnapshotStatevectorTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'matrix_product_state' ] BACKEND_OPTS = {} def statevector_snapshots(self, data, label): snaps = data.get("snapshots", {}).get("statevector", {}).get(label, []) statevecs = [] for snap in snaps: self.assertIsInstance(snap, np.ndarray) statevecs.append(snap) return statevecs def test_snapshot_statevector_pre_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'statevector', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) self.assertTrue(len(snaps), 1) target = statevec_targets[j] value = snaps[0] self.assertTrue(np.allclose(value, target)) def test_snapshot_statevector_pre_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'statevector', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) self.assertTrue(len(snaps), 1) target = statevec_targets[j] value = snaps[0] self.assertTrue(np.allclose(value, target)) def test_snapshot_statevector_post_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'statevector', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: logging.getLogger().setLevel(logging.CRITICAL) self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) self.assertTrue(np.allclose(snaps[j], target)) def test_snapshot_statevector_post_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'statevector', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStatevectorTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.statevector_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) self.assertTrue(np.allclose(snaps[j], target)) class QasmSnapshotStabilizerTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = ['automatic', 'stabilizer'] BACKEND_OPTS = {} @staticmethod def stabilizer_snapshots(data, label): return data.get("snapshots", {}).get("stabilizer", {}).get(label, []) @staticmethod def stabilizes_statevector(stabilizer, statevector): for stab in stabilizer: if stab[0] == '-': pauli_mat = -1 * Pauli.from_label(stab[1:]).to_matrix() else: pauli_mat = Pauli.from_label(stab).to_matrix() val = statevector.conj().dot(pauli_mat.dot(statevector)) if not np.isclose(val, 1): return False return True def test_snapshot_stabilizer_pre_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'stabilizer', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) self.assertEqual(len(snaps), 1) statevec = statevec_targets[j] stabilizer = snaps[0] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_pre_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'stabilizer', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) self.assertEqual(len(snaps), 1) statevec = statevec_targets[j] stabilizer = snaps[0] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_post_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'stabilizer', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) for j, mem in enumerate(data['memory']): statevec = statevec_targets[i].get(mem) stabilizer = snaps[j] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) def test_snapshot_stabilizer_post_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'stabilizer', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotStabilizerTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.stabilizer_snapshots(data, label) for j, mem in enumerate(data['memory']): statevec = statevec_targets[i].get(mem) stabilizer = snaps[j] self.assertTrue( self.stabilizes_statevector(stabilizer, statevec)) class QasmSnapshotDensityMatrixTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust' ] BACKEND_OPTS = {} def density_snapshots(self, data, label): snaps = data.get("snapshots", {}).get("density_matrix", {}).get(label, []) output = {} for snap_dict in snaps: memory = snap_dict['memory'] self.assertIsInstance(snap_dict['value'], np.ndarray) output[memory] = snap_dict['value'] return output def test_snapshot_density_matrix_pre_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'density_matrix', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) self.assertTrue(len(snaps), 1) target = np.outer(statevec_targets[j], statevec_targets[j].conj()) value = snaps.get('0x0') self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_pre_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_pre_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'density_matrix', post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) self.assertTrue(len(snaps), 1) target = np.outer(statevec_targets[j], statevec_targets[j].conj()) value = snaps.get('0x0') self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_post_measure_det(self): shots = 10 label = "snap" counts_targets = snapshot_state_counts_deterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_deterministic( ) circuits = snapshot_state_circuits_deterministic(label, 'density_matrix', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) target = np.outer(target, target.conj()) value = snaps.get(mem) self.assertTrue(np.allclose(value, target)) def test_snapshot_density_matrix_post_measure_nondet(self): shots = 100 label = "snap" counts_targets = snapshot_state_counts_nondeterministic(shots) statevec_targets = snapshot_state_post_measure_statevector_nondeterministic( ) circuits = snapshot_state_circuits_nondeterministic(label, 'density_matrix', post_measure=True) qobj = assemble(circuits, self.SIMULATOR, memory=True, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotDensityMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.2 * shots) for i, circuit in enumerate(circuits): data = result.data(circuit) snaps = self.density_snapshots(data, label) for j, mem in enumerate(data['memory']): target = statevec_targets[i].get(mem) target = np.outer(target, target.conj()) value = snaps.get(mem) self.assertTrue(np.allclose(value, target)) class QasmSnapshotProbabilitiesTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'stabilizer', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', ] BACKEND_OPTS = {} @staticmethod def probability_snapshots(data, labels): output = {} for label in labels: snaps = data.get("snapshots", {}).get("probabilities", {}).get(label, []) output[label] = { snap_dict['memory']: snap_dict['value'] for snap_dict in snaps } return output def test_snapshot_probabilities_pre_measure(self): shots = 1000 labels = list(snapshot_probabilities_labels_qubits().keys()) counts_targets = snapshot_probabilities_counts(shots) prob_targets = snapshot_probabilities_pre_meas_probs() circuits = snapshot_probabilities_circuits(post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotProbabilitiesTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.probability_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = prob_targets[j].get(label, {}).get(memory, {}) self.assertDictAlmostEqual(value, target, delta=1e-7) def test_snapshot_probabilities_post_measure(self): shots = 1000 labels = list(snapshot_probabilities_labels_qubits().keys()) counts_targets = snapshot_probabilities_counts(shots) prob_targets = snapshot_probabilities_post_meas_probs() circuits = snapshot_probabilities_circuits(post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotProbabilitiesTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.probability_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) for memory, value in snaps.items(): target = prob_targets[j].get(label, {}).get(memory, {}) self.assertDictAlmostEqual(value, target, delta=1e-7) class QasmSnapshotExpValPauliTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', 'stabilizer' ] BACKEND_OPTS = {} @staticmethod def expval_snapshots(data, labels): output = {} for label in labels: snaps = data.get("snapshots", {}).get("expectation_value", {}).get(label, []) inner = {} for snap_dict in snaps: val = snap_dict['value'] inner[snap_dict['memory']] = val output[label] = inner return output def test_snapshot_expval_pauli_pre_measure(self): shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_pre_meas_values() circuits = snapshot_expval_circuits(pauli=True, post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValPauliTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) def test_snapshot_expval_pauli_post_measure(self): shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_post_meas_values() circuits = snapshot_expval_circuits(pauli=True, post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValPauliTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) class QasmSnapshotExpvalPauliNCTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'density_matrix', 'density_matrix_gpu', 'density_matrix_thrust', 'matrix_product_state', ] BACKEND_OPTS = {} def general_test(self, pauli, num_qubits=None, seed=None): pauli_qubits = list(range(len(pauli))) if num_qubits is None: num_qubits = len(pauli_qubits) rng = np.random.default_rng(seed) params = rng.uniform(-1, 1, size=(num_qubits, 3)) init_circ = QuantumCircuit(num_qubits) for i, par in enumerate(params): init_circ.u3(*par, i) rho = DensityMatrix.from_instruction(init_circ) op = Operator.from_label(pauli) target = np.trace(Operator(rho).compose(op, pauli_qubits).data) qc = init_circ.copy() qc.snapshot_expectation_value('final', [(1, pauli)], pauli_qubits) qobj = assemble(qc) result = self.SIMULATOR.run( qobj, backend_options=self.BACKEND_OPTS).result() self.assertTrue(getattr(result, 'success', False)) snapshots = result.data(0).get('snapshots', {}) self.assertIn('expectation_value', snapshots) self.assertIn('final', snapshots['expectation_value']) expval = snapshots.get('expectation_value', {})['final'][0]['value'] self.assertAlmostEqual(expval, target) def test_pauli1(self): seed = 100 for tup in ['I', 'X', 'Y', 'Z']: pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) def test_pauli2(self): seed = 100 for tup in it.product(['I', 'X', 'Y', 'Z'], repeat=2): pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) def test_pauli3(self): seed = 100 for tup in it.product(['I', 'X', 'Y', 'Z'], repeat=3): pauli = ''.join(reversed(tup)) with self.subTest(msg='Pauli {}'.format(pauli)): self.general_test(pauli, num_qubits=3, seed=seed) class QasmSnapshotExpValMatrixTests: SIMULATOR = QasmSimulator() SUPPORTED_QASM_METHODS = [ 'automatic', 'statevector', 'statevector_gpu', 'statevector_thrust', 'matrix_product_state' ] BACKEND_OPTS = {} @staticmethod def expval_snapshots(data, labels): output = {} for label in labels: snaps = data.get("snapshots", {}).get("expectation_value", {}).get(label, []) inner = {} for snap_dict in snaps: inner[snap_dict['memory']] = snap_dict['value'] output[label] = inner return output def test_snapshot_expval_matrix_pre_measure(self): shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_pre_meas_values() circuits = snapshot_expval_circuits(pauli=False, post_measure=False) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7) def test_snapshot_expval_matrix_post_measure(self): shots = 1000 labels = snapshot_expval_labels() counts_targets = snapshot_expval_counts(shots) value_targets = snapshot_expval_post_meas_values() circuits = snapshot_expval_circuits(pauli=False, post_measure=True) qobj = assemble(circuits, self.SIMULATOR, shots=shots) job = self.SIMULATOR.run(qobj, backend_options=self.BACKEND_OPTS) result = job.result() success = getattr(result, 'success', False) method = self.BACKEND_OPTS.get('method', 'automatic') if method not in QasmSnapshotExpValMatrixTests.SUPPORTED_QASM_METHODS: self.assertFalse(success) else: self.assertTrue(success) self.compare_counts(result, circuits, counts_targets, delta=0.1 * shots) for j, circuit in enumerate(circuits): data = result.data(circuit) all_snapshots = self.expval_snapshots(data, labels) for label in labels: snaps = all_snapshots.get(label, {}) self.assertTrue(len(snaps), 1) for memory, value in snaps.items(): target = value_targets[j].get(label, {}).get(memory, {}) self.assertAlmostEqual(value, target, delta=1e-7)
true
true
f717dab74980995a0147cf18b683ffeabedb256b
9,351
py
Python
samoyed_ts/nmt.py
oshiooshi/cirneco
f71f1cd583bf6e290d7b8e74f148f06cadd39d63
[ "MIT" ]
null
null
null
samoyed_ts/nmt.py
oshiooshi/cirneco
f71f1cd583bf6e290d7b8e74f148f06cadd39d63
[ "MIT" ]
null
null
null
samoyed_ts/nmt.py
oshiooshi/cirneco
f71f1cd583bf6e290d7b8e74f148f06cadd39d63
[ "MIT" ]
13
2021-07-01T07:58:30.000Z
2021-09-09T16:52:22.000Z
import torch # import torchtext import torch.nn as nn # from torchtext.vocab import Vocab, build_vocab_from_iterator # from torchtext.utils import unicode_csv_reader # from torchtext.data.datasets_utils import _RawTextIterableDataset from torch import Tensor from typing import Iterable, List # import sentencepiece as spm # import io import math import vocab SEED = 1234 torch.manual_seed(SEED) torch.cuda.manual_seed(SEED) # 特殊トークンの定義 UNK_IDX, PAD_IDX, SOS_IDX, EOS_IDX = 0, 1, 2, 3 special_symbols = ['<unk>', '<pad>', '<sos>', '<eos>', '<blk>', '</blk>', '<sep>'] MAX_LEN=80 # sp = spm.SentencePieceProcessor(model_file='corpus_Python-JPN/p3/p3.model') # def jpn_tokenizer(text): # ss = [tok.replace('▁', '') for tok in sp.encode(text, out_type=str)][:MAX_LEN] # return [s for s in ss if len(s) != 0] # def py_tokenizer(text): # return [tok for tok in text.split()][:MAX_LEN] from torch.nn.utils.rnn import pad_sequence # 連続した操作をまとめて行うためのヘルパー関数 def sequential_transforms(*transforms): def func(txt_input): for transform in transforms: txt_input = transform(txt_input) return txt_input return func # SOS/EOSトークンを追加し、入力配列のインデックス用のテンソルを作成 def tensor_transform(token_ids: List[int]): return torch.cat((torch.tensor([SOS_IDX]), torch.tensor(token_ids), torch.tensor([EOS_IDX]))) ## Transformer の定義 from torch.nn import (TransformerEncoder, TransformerDecoder, TransformerEncoderLayer, TransformerDecoderLayer) class PositionalEncoding(nn.Module): def __init__(self, emb_size: int, dropout: float, maxlen: int = 5000): super(PositionalEncoding, self).__init__() den = torch.exp(- torch.arange(0, emb_size, 2) * math.log(10000) / emb_size) pos = torch.arange(0, maxlen).reshape(maxlen, 1) pos_embedding = torch.zeros((maxlen, emb_size)) pos_embedding[:, 0::2] = torch.sin(pos * den) pos_embedding[:, 1::2] = torch.cos(pos * den) pos_embedding = pos_embedding.unsqueeze(-2) self.dropout = nn.Dropout(dropout) self.register_buffer('pos_embedding', pos_embedding) def forward(self, token_embedding: Tensor): return self.dropout(token_embedding + self.pos_embedding[:token_embedding.size(0),:]) class TokenEmbedding(nn.Module): def __init__(self, vocab_size: int, emb_size): super(TokenEmbedding, self).__init__() self.embedding = nn.Embedding(vocab_size, emb_size) self.emb_size = emb_size def forward(self, tokens: Tensor): return self.embedding(tokens.long()) * math.sqrt(self.emb_size) class Seq2SeqTransformer(nn.Module): def __init__(self, num_encoder_layers: int, num_decoder_layers: int, emb_size: int, nhead: int, src_vocab_size: int, tgt_vocab_size: int, dim_feedforward: int = 512, dropout: float = 0.1): super(Seq2SeqTransformer, self).__init__() encoder_layer = TransformerEncoderLayer(d_model=emb_size, nhead=nhead, dim_feedforward=dim_feedforward) self.transformer_encoder = TransformerEncoder(encoder_layer, num_layers=num_encoder_layers) decoder_layer = TransformerDecoderLayer(d_model=emb_size, nhead=nhead, dim_feedforward=dim_feedforward) self.transformer_decoder = TransformerDecoder(decoder_layer, num_layers=num_decoder_layers) self.generator = nn.Linear(emb_size, tgt_vocab_size) self.src_tok_emb = TokenEmbedding(src_vocab_size, emb_size) self.tgt_tok_emb = TokenEmbedding(tgt_vocab_size, emb_size) self.positional_encoding = PositionalEncoding(emb_size, dropout=dropout) def forward(self, src: Tensor, tgt: Tensor, src_mask: Tensor, tgt_mask: Tensor, src_padding_mask: Tensor, tgt_padding_mask: Tensor, memory_key_padding_mask: Tensor): src_emb = self.positional_encoding(self.src_tok_emb(src)) tgt_emb = self.positional_encoding(self.tgt_tok_emb(tgt)) memory = self.transformer_encoder(src_emb, src_mask, src_padding_mask) outs = self.transformer_decoder(tgt_emb, memory, tgt_mask, None, tgt_padding_mask, memory_key_padding_mask) return self.generator(outs) def encode(self, src: Tensor, src_mask: Tensor): return self.transformer_encoder(self.positional_encoding( self.src_tok_emb(src)), src_mask) def decode(self, tgt: Tensor, memory: Tensor, tgt_mask: Tensor): return self.transformer_decoder(self.positional_encoding( self.tgt_tok_emb(tgt)), memory, tgt_mask) DEVICE = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu') ### Masking ## 異なるマスク処理を行う2つの関数を定義 # モデルが予測を行う際に、未来の単語を見ないようにするためのマスク def generate_square_subsequent_mask(sz): mask = (torch.triu(torch.ones((sz, sz), device=DEVICE)) == 1).transpose(0, 1) mask = mask.float().masked_fill(mask == 0, float('-inf')).masked_fill(mask == 1, float(0.0)) return mask # ソースとターゲットのパディングトークンを隠すためのマスク def create_mask(src, tgt): src_seq_len = src.shape[0] tgt_seq_len = tgt.shape[0] tgt_mask = generate_square_subsequent_mask(tgt_seq_len) src_mask = torch.zeros((src_seq_len, src_seq_len), device=DEVICE).type(torch.bool) src_padding_mask = (src == PAD_IDX).transpose(0, 1) tgt_padding_mask = (tgt == PAD_IDX).transpose(0, 1) return src_mask, tgt_mask, src_padding_mask, tgt_padding_mask def greedy_decode(model, src, src_mask, max_len, beamsize, start_symbol): src = src.to(DEVICE) src_mask = src_mask.to(DEVICE) memory = model.encode(src, src_mask) ys = torch.ones(1, 1).fill_(start_symbol).type(torch.long).to(DEVICE) for i in range(max_len-1): memory = memory.to(DEVICE) tgt_mask = (generate_square_subsequent_mask(ys.size(0)) .type(torch.bool)).to(DEVICE) out = model.decode(ys, memory, tgt_mask) out = out.transpose(0, 1) prob = model.generator(out[:, -1]) # prob.size() の実行結果 : torch.Size([1, 1088]) => 1088 はTGT のVOCAV_SIZE next_prob, next_word = prob.topk(k=beamsize, dim=1) # print(next_word) # print(next_prob) next_word = next_word[:, 0] # greedy なので、もっとも確率が高いものを選ぶ next_word = next_word.item() # 要素の値を取得 (int に変換) ys = torch.cat([ys, torch.ones(1, 1).type_as(src.data).fill_(next_word)], dim=0) if next_word == EOS_IDX: break return ys class NMT(object): src_vocab: object tgt_vocab: object def __init__(self, src_vocab='kujira', tgt_vocab='python'): self.src_vocab = vocab.load_vocab(src_vocab) self.tgt_vocab = vocab.load_vocab(tgt_vocab) tokenizer = vocab.tokenizer_from_vocab(self.src_vocab) self.src_transform = sequential_transforms(tokenizer, #Tokenization self.src_vocab, #Numericalization tensor_transform) # Add SOS/EOS and create tensor # パラメータの定義 self.SRC_VOCAB_SIZE = len(self.src_vocab) self.TGT_VOCAB_SIZE = len(self.tgt_vocab) self.EMB_SIZE = 512 # BERT の次元に揃えれば良いよ self.NHEAD = 8 self.FFN_HID_DIM = 512 self.BATCH_SIZE = 128 self.NUM_ENCODER_LAYERS = 3 self.NUM_DECODER_LAYERS = 3 # インスタンスの作成 self.transformer = Seq2SeqTransformer(self.NUM_ENCODER_LAYERS, self.NUM_DECODER_LAYERS, self.EMB_SIZE, self.NHEAD, self.SRC_VOCAB_SIZE, self.TGT_VOCAB_SIZE, self.FFN_HID_DIM) # TODO: ? for p in self.transformer.parameters(): if p.dim() > 1: nn.init.xavier_uniform_(p) # デバイスの設定 self.transformer = self.transformer.to(DEVICE) # 損失関数の定義 (クロスエントロピー) self.loss_fn = torch.nn.CrossEntropyLoss(ignore_index=PAD_IDX) # オプティマイザの定義 (Adam) self.optimizer = torch.optim.Adam(self.transformer.parameters(), lr=0.0001, betas=(0.9, 0.98), eps=1e-9) def load(self, filename='all-model.pt'): self.transformer.load_state_dict(torch.load(filename, map_location=DEVICE)) def translate(self, src_sentence: str): self.transformer.eval() src = self.src_transform(src_sentence).view(-1, 1) num_tokens = src.shape[0] src_mask = (torch.zeros(num_tokens, num_tokens)).type(torch.bool) tgt_tokens = greedy_decode( self.transformer, src, src_mask, max_len=num_tokens + 5, beamsize=5, start_symbol=SOS_IDX).flatten() return " ".join(self.tgt_vocab.lookup_tokens(list(tgt_tokens.cpu().numpy()))).replace("<sos>", "").replace("<eos>", "") if __name__ == '__main__': nmt = NMT() nmt.load('./all-model.pt') pred = nmt.translate('もし<A>が偶数のとき') print('pred:', pred)
39.791489
127
0.634371
import torch import torch.nn as nn from torch import Tensor from typing import Iterable, List import math import vocab SEED = 1234 torch.manual_seed(SEED) torch.cuda.manual_seed(SEED) UNK_IDX, PAD_IDX, SOS_IDX, EOS_IDX = 0, 1, 2, 3 special_symbols = ['<unk>', '<pad>', '<sos>', '<eos>', '<blk>', '</blk>', '<sep>'] MAX_LEN=80 from torch.nn.utils.rnn import pad_sequence def sequential_transforms(*transforms): def func(txt_input): for transform in transforms: txt_input = transform(txt_input) return txt_input return func def tensor_transform(token_ids: List[int]): return torch.cat((torch.tensor([SOS_IDX]), torch.tensor(token_ids), torch.tensor([EOS_IDX]))) mport (TransformerEncoder, TransformerDecoder, TransformerEncoderLayer, TransformerDecoderLayer) class PositionalEncoding(nn.Module): def __init__(self, emb_size: int, dropout: float, maxlen: int = 5000): super(PositionalEncoding, self).__init__() den = torch.exp(- torch.arange(0, emb_size, 2) * math.log(10000) / emb_size) pos = torch.arange(0, maxlen).reshape(maxlen, 1) pos_embedding = torch.zeros((maxlen, emb_size)) pos_embedding[:, 0::2] = torch.sin(pos * den) pos_embedding[:, 1::2] = torch.cos(pos * den) pos_embedding = pos_embedding.unsqueeze(-2) self.dropout = nn.Dropout(dropout) self.register_buffer('pos_embedding', pos_embedding) def forward(self, token_embedding: Tensor): return self.dropout(token_embedding + self.pos_embedding[:token_embedding.size(0),:]) class TokenEmbedding(nn.Module): def __init__(self, vocab_size: int, emb_size): super(TokenEmbedding, self).__init__() self.embedding = nn.Embedding(vocab_size, emb_size) self.emb_size = emb_size def forward(self, tokens: Tensor): return self.embedding(tokens.long()) * math.sqrt(self.emb_size) class Seq2SeqTransformer(nn.Module): def __init__(self, num_encoder_layers: int, num_decoder_layers: int, emb_size: int, nhead: int, src_vocab_size: int, tgt_vocab_size: int, dim_feedforward: int = 512, dropout: float = 0.1): super(Seq2SeqTransformer, self).__init__() encoder_layer = TransformerEncoderLayer(d_model=emb_size, nhead=nhead, dim_feedforward=dim_feedforward) self.transformer_encoder = TransformerEncoder(encoder_layer, num_layers=num_encoder_layers) decoder_layer = TransformerDecoderLayer(d_model=emb_size, nhead=nhead, dim_feedforward=dim_feedforward) self.transformer_decoder = TransformerDecoder(decoder_layer, num_layers=num_decoder_layers) self.generator = nn.Linear(emb_size, tgt_vocab_size) self.src_tok_emb = TokenEmbedding(src_vocab_size, emb_size) self.tgt_tok_emb = TokenEmbedding(tgt_vocab_size, emb_size) self.positional_encoding = PositionalEncoding(emb_size, dropout=dropout) def forward(self, src: Tensor, tgt: Tensor, src_mask: Tensor, tgt_mask: Tensor, src_padding_mask: Tensor, tgt_padding_mask: Tensor, memory_key_padding_mask: Tensor): src_emb = self.positional_encoding(self.src_tok_emb(src)) tgt_emb = self.positional_encoding(self.tgt_tok_emb(tgt)) memory = self.transformer_encoder(src_emb, src_mask, src_padding_mask) outs = self.transformer_decoder(tgt_emb, memory, tgt_mask, None, tgt_padding_mask, memory_key_padding_mask) return self.generator(outs) def encode(self, src: Tensor, src_mask: Tensor): return self.transformer_encoder(self.positional_encoding( self.src_tok_emb(src)), src_mask) def decode(self, tgt: Tensor, memory: Tensor, tgt_mask: Tensor): return self.transformer_decoder(self.positional_encoding( self.tgt_tok_emb(tgt)), memory, tgt_mask) DEVICE = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu') sz): mask = (torch.triu(torch.ones((sz, sz), device=DEVICE)) == 1).transpose(0, 1) mask = mask.float().masked_fill(mask == 0, float('-inf')).masked_fill(mask == 1, float(0.0)) return mask def create_mask(src, tgt): src_seq_len = src.shape[0] tgt_seq_len = tgt.shape[0] tgt_mask = generate_square_subsequent_mask(tgt_seq_len) src_mask = torch.zeros((src_seq_len, src_seq_len), device=DEVICE).type(torch.bool) src_padding_mask = (src == PAD_IDX).transpose(0, 1) tgt_padding_mask = (tgt == PAD_IDX).transpose(0, 1) return src_mask, tgt_mask, src_padding_mask, tgt_padding_mask def greedy_decode(model, src, src_mask, max_len, beamsize, start_symbol): src = src.to(DEVICE) src_mask = src_mask.to(DEVICE) memory = model.encode(src, src_mask) ys = torch.ones(1, 1).fill_(start_symbol).type(torch.long).to(DEVICE) for i in range(max_len-1): memory = memory.to(DEVICE) tgt_mask = (generate_square_subsequent_mask(ys.size(0)) .type(torch.bool)).to(DEVICE) out = model.decode(ys, memory, tgt_mask) out = out.transpose(0, 1) prob = model.generator(out[:, -1]) next_prob, next_word = prob.topk(k=beamsize, dim=1) next_word = next_word[:, 0] next_word = next_word.item() ys = torch.cat([ys, torch.ones(1, 1).type_as(src.data).fill_(next_word)], dim=0) if next_word == EOS_IDX: break return ys class NMT(object): src_vocab: object tgt_vocab: object def __init__(self, src_vocab='kujira', tgt_vocab='python'): self.src_vocab = vocab.load_vocab(src_vocab) self.tgt_vocab = vocab.load_vocab(tgt_vocab) tokenizer = vocab.tokenizer_from_vocab(self.src_vocab) self.src_transform = sequential_transforms(tokenizer, self.src_vocab, tensor_transform) self.SRC_VOCAB_SIZE = len(self.src_vocab) self.TGT_VOCAB_SIZE = len(self.tgt_vocab) self.EMB_SIZE = 512 self.NHEAD = 8 self.FFN_HID_DIM = 512 self.BATCH_SIZE = 128 self.NUM_ENCODER_LAYERS = 3 self.NUM_DECODER_LAYERS = 3 self.transformer = Seq2SeqTransformer(self.NUM_ENCODER_LAYERS, self.NUM_DECODER_LAYERS, self.EMB_SIZE, self.NHEAD, self.SRC_VOCAB_SIZE, self.TGT_VOCAB_SIZE, self.FFN_HID_DIM) for p in self.transformer.parameters(): if p.dim() > 1: nn.init.xavier_uniform_(p) self.transformer = self.transformer.to(DEVICE) self.loss_fn = torch.nn.CrossEntropyLoss(ignore_index=PAD_IDX) self.optimizer = torch.optim.Adam(self.transformer.parameters(), lr=0.0001, betas=(0.9, 0.98), eps=1e-9) def load(self, filename='all-model.pt'): self.transformer.load_state_dict(torch.load(filename, map_location=DEVICE)) def translate(self, src_sentence: str): self.transformer.eval() src = self.src_transform(src_sentence).view(-1, 1) num_tokens = src.shape[0] src_mask = (torch.zeros(num_tokens, num_tokens)).type(torch.bool) tgt_tokens = greedy_decode( self.transformer, src, src_mask, max_len=num_tokens + 5, beamsize=5, start_symbol=SOS_IDX).flatten() return " ".join(self.tgt_vocab.lookup_tokens(list(tgt_tokens.cpu().numpy()))).replace("<sos>", "").replace("<eos>", "") if __name__ == '__main__': nmt = NMT() nmt.load('./all-model.pt') pred = nmt.translate('もし<A>が偶数のとき') print('pred:', pred)
true
true
f717dbb4e70ad1e7af8abf4d384448d348389a87
9,237
py
Python
neo/io/pynnio.py
lkoelman/python-neo
6b0454519b4ead6605d3ce4100a07c33f57df830
[ "BSD-3-Clause" ]
null
null
null
neo/io/pynnio.py
lkoelman/python-neo
6b0454519b4ead6605d3ce4100a07c33f57df830
[ "BSD-3-Clause" ]
8
2018-06-02T11:46:10.000Z
2018-09-04T15:51:45.000Z
src/neo/neo/io/pynnio.py
grg2rsr/SeqPeelSort
58a207976fb33a50ea8e42b70d7da73b03474f42
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- """ Module for reading/writing data from/to legacy PyNN formats. PyNN is available at http://neuralensemble.org/PyNN Classes: PyNNNumpyIO PyNNTextIO Supported: Read/Write Authors: Andrew Davison, Pierre Yger """ from itertools import chain import numpy import quantities as pq import warnings from neo.io.baseio import BaseIO from neo.core import Segment, AnalogSignal, SpikeTrain try: unicode PY2 = True except NameError: PY2 = False UNITS_MAP = { 'spikes': pq.ms, 'v': pq.mV, 'gsyn': pq.UnitQuantity('microsiemens', 1e-6 * pq.S, 'uS', 'µS'), # checked } class BasePyNNIO(BaseIO): """ Base class for PyNN IO classes """ is_readable = True is_writable = True has_header = True is_streameable = False # TODO - correct spelling to "is_streamable" supported_objects = [Segment, AnalogSignal, SpikeTrain] readable_objects = supported_objects writeable_objects = supported_objects mode = 'file' def __init__(self, filename=None, **kargs): BaseIO.__init__(self, filename, *kargs) warnings.warn("PyNNTextIO and PyNNNumpyIO will be removed in Neo 0.7.0. " + "Please contact the Neo developers if this will cause you problems.", DeprecationWarning) def _read_file_contents(self): raise NotImplementedError def _extract_array(self, data, channel_index): idx = numpy.where(data[:, 1] == channel_index)[0] return data[idx, 0] def _determine_units(self, metadata): if 'units' in metadata: return metadata['units'] elif 'variable' in metadata and metadata['variable'] in UNITS_MAP: return UNITS_MAP[metadata['variable']] else: raise IOError("Cannot determine units") def _extract_signals(self, data, metadata): arr = numpy.vstack(self._extract_array(data, channel_index) for channel_index in range(metadata['first_index'], metadata['last_index'] + 1)) if len(arr) > 0: signal = AnalogSignal(arr.T, units=self._determine_units(metadata), sampling_period=metadata['dt'] * pq.ms) signal.annotate(label=metadata["label"], variable=metadata["variable"]) return signal def _extract_spikes(self, data, metadata, channel_index): spiketrain = None spike_times = self._extract_array(data, channel_index) if len(spike_times) > 0: spiketrain = SpikeTrain(spike_times, units=pq.ms, t_stop=spike_times.max()) spiketrain.annotate(label=metadata["label"], channel_index=channel_index, dt=metadata["dt"]) return spiketrain def _write_file_contents(self, data, metadata): raise NotImplementedError def read_segment(self, lazy=False): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() annotations = dict((k, metadata.get(k, 'unknown')) for k in ("label", "variable", "first_id", "last_id")) seg = Segment(**annotations) if metadata['variable'] == 'spikes': for i in range(metadata['first_index'], metadata['last_index'] + 1): spiketrain = self._extract_spikes(data, metadata, i) if spiketrain is not None: seg.spiketrains.append(spiketrain) # store dt for SpikeTrains only, as can be retrieved from sampling_period for AnalogSignal seg.annotate(dt=metadata['dt']) else: signal = self._extract_signals(data, metadata) if signal is not None: seg.analogsignals.append(signal) seg.create_many_to_one_relationship() return seg def write_segment(self, segment): source = segment.analogsignals or segment.spiketrains assert len(source) > 0, "Segment contains neither analog signals nor spike trains." metadata = segment.annotations.copy() s0 = source[0] if isinstance(s0, AnalogSignal): if len(source) > 1: warnings.warn("Cannot handle multiple analog signals. Writing only the first.") source = s0.T metadata['size'] = s0.shape[1] n = source.size else: metadata['size'] = len(source) n = sum(s.size for s in source) metadata['first_index'] = 0 metadata['last_index'] = metadata['size'] - 1 if 'label' not in metadata: metadata['label'] = 'unknown' if 'dt' not in metadata: # dt not included in annotations if Segment contains only AnalogSignals metadata['dt'] = s0.sampling_period.rescale(pq.ms).magnitude metadata['n'] = n data = numpy.empty((n, 2)) # if the 'variable' annotation is a standard one from PyNN, we rescale # to use standard PyNN units # we take the units from the first element of source and scale all # the signals to have the same units if 'variable' in segment.annotations: units = UNITS_MAP.get(segment.annotations['variable'], source[0].dimensionality) else: units = source[0].dimensionality metadata['variable'] = 'unknown' try: metadata['units'] = units.unicode except AttributeError: metadata['units'] = units.u_symbol start = 0 for i, signal in enumerate(source): # here signal may be AnalogSignal or SpikeTrain end = start + signal.size data[start:end, 0] = numpy.array(signal.rescale(units)) data[start:end, 1] = i * numpy.ones((signal.size,), dtype=float) start = end self._write_file_contents(data, metadata) def read_analogsignal(self, lazy=False): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() if metadata['variable'] == 'spikes': raise TypeError("File contains spike data, not analog signals") else: signal = self._extract_signals(data, metadata) if signal is None: raise IndexError("File does not contain a signal") else: return signal def read_spiketrain(self, lazy=False, channel_index=0): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() if metadata['variable'] != 'spikes': raise TypeError("File contains analog signals, not spike data") else: spiketrain = self._extract_spikes(data, metadata, channel_index) if spiketrain is None: raise IndexError( "File does not contain any spikes with channel index %d" % channel_index) else: return spiketrain class PyNNNumpyIO(BasePyNNIO): """ (DEPRECATED) Reads/writes data from/to PyNN NumpyBinaryFile format """ name = "PyNN NumpyBinaryFile" extensions = ['npz'] def _read_file_contents(self): contents = numpy.load(self.filename) data = contents["data"] metadata = {} for name, value in contents['metadata']: try: metadata[name] = eval(value) except Exception: metadata[name] = value return data, metadata def _write_file_contents(self, data, metadata): # we explicitly set the dtype to ensure roundtrips preserve file contents exactly max_metadata_length = max(chain([len(k) for k in metadata.keys()], [len(str(v)) for v in metadata.values()])) if PY2: dtype = "S%d" % max_metadata_length else: dtype = "U%d" % max_metadata_length metadata_array = numpy.array(sorted(metadata.items()), dtype) numpy.savez(self.filename, data=data, metadata=metadata_array) class PyNNTextIO(BasePyNNIO): """ (DEPRECATED) Reads/writes data from/to PyNN StandardTextFile format """ name = "PyNN StandardTextFile" extensions = ['v', 'ras', 'gsyn'] def _read_metadata(self): metadata = {} with open(self.filename) as f: for line in f: if line[0] == "#": name, value = line[1:].strip().split("=") name = name.strip() try: metadata[name] = eval(value) except Exception: metadata[name] = value.strip() else: break return metadata def _read_file_contents(self): data = numpy.loadtxt(self.filename) metadata = self._read_metadata() return data, metadata def _write_file_contents(self, data, metadata): with open(self.filename, 'wb') as f: for item in sorted(metadata.items()): f.write(("# %s = %s\n" % item).encode('utf8')) numpy.savetxt(f, data)
36.800797
105
0.590343
from itertools import chain import numpy import quantities as pq import warnings from neo.io.baseio import BaseIO from neo.core import Segment, AnalogSignal, SpikeTrain try: unicode PY2 = True except NameError: PY2 = False UNITS_MAP = { 'spikes': pq.ms, 'v': pq.mV, 'gsyn': pq.UnitQuantity('microsiemens', 1e-6 * pq.S, 'uS', 'µS'), } class BasePyNNIO(BaseIO): is_readable = True is_writable = True has_header = True is_streameable = False supported_objects = [Segment, AnalogSignal, SpikeTrain] readable_objects = supported_objects writeable_objects = supported_objects mode = 'file' def __init__(self, filename=None, **kargs): BaseIO.__init__(self, filename, *kargs) warnings.warn("PyNNTextIO and PyNNNumpyIO will be removed in Neo 0.7.0. " + "Please contact the Neo developers if this will cause you problems.", DeprecationWarning) def _read_file_contents(self): raise NotImplementedError def _extract_array(self, data, channel_index): idx = numpy.where(data[:, 1] == channel_index)[0] return data[idx, 0] def _determine_units(self, metadata): if 'units' in metadata: return metadata['units'] elif 'variable' in metadata and metadata['variable'] in UNITS_MAP: return UNITS_MAP[metadata['variable']] else: raise IOError("Cannot determine units") def _extract_signals(self, data, metadata): arr = numpy.vstack(self._extract_array(data, channel_index) for channel_index in range(metadata['first_index'], metadata['last_index'] + 1)) if len(arr) > 0: signal = AnalogSignal(arr.T, units=self._determine_units(metadata), sampling_period=metadata['dt'] * pq.ms) signal.annotate(label=metadata["label"], variable=metadata["variable"]) return signal def _extract_spikes(self, data, metadata, channel_index): spiketrain = None spike_times = self._extract_array(data, channel_index) if len(spike_times) > 0: spiketrain = SpikeTrain(spike_times, units=pq.ms, t_stop=spike_times.max()) spiketrain.annotate(label=metadata["label"], channel_index=channel_index, dt=metadata["dt"]) return spiketrain def _write_file_contents(self, data, metadata): raise NotImplementedError def read_segment(self, lazy=False): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() annotations = dict((k, metadata.get(k, 'unknown')) for k in ("label", "variable", "first_id", "last_id")) seg = Segment(**annotations) if metadata['variable'] == 'spikes': for i in range(metadata['first_index'], metadata['last_index'] + 1): spiketrain = self._extract_spikes(data, metadata, i) if spiketrain is not None: seg.spiketrains.append(spiketrain) seg.annotate(dt=metadata['dt']) else: signal = self._extract_signals(data, metadata) if signal is not None: seg.analogsignals.append(signal) seg.create_many_to_one_relationship() return seg def write_segment(self, segment): source = segment.analogsignals or segment.spiketrains assert len(source) > 0, "Segment contains neither analog signals nor spike trains." metadata = segment.annotations.copy() s0 = source[0] if isinstance(s0, AnalogSignal): if len(source) > 1: warnings.warn("Cannot handle multiple analog signals. Writing only the first.") source = s0.T metadata['size'] = s0.shape[1] n = source.size else: metadata['size'] = len(source) n = sum(s.size for s in source) metadata['first_index'] = 0 metadata['last_index'] = metadata['size'] - 1 if 'label' not in metadata: metadata['label'] = 'unknown' if 'dt' not in metadata: metadata['dt'] = s0.sampling_period.rescale(pq.ms).magnitude metadata['n'] = n data = numpy.empty((n, 2)) if 'variable' in segment.annotations: units = UNITS_MAP.get(segment.annotations['variable'], source[0].dimensionality) else: units = source[0].dimensionality metadata['variable'] = 'unknown' try: metadata['units'] = units.unicode except AttributeError: metadata['units'] = units.u_symbol start = 0 for i, signal in enumerate(source): end = start + signal.size data[start:end, 0] = numpy.array(signal.rescale(units)) data[start:end, 1] = i * numpy.ones((signal.size,), dtype=float) start = end self._write_file_contents(data, metadata) def read_analogsignal(self, lazy=False): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() if metadata['variable'] == 'spikes': raise TypeError("File contains spike data, not analog signals") else: signal = self._extract_signals(data, metadata) if signal is None: raise IndexError("File does not contain a signal") else: return signal def read_spiketrain(self, lazy=False, channel_index=0): assert not lazy, 'Do not support lazy' data, metadata = self._read_file_contents() if metadata['variable'] != 'spikes': raise TypeError("File contains analog signals, not spike data") else: spiketrain = self._extract_spikes(data, metadata, channel_index) if spiketrain is None: raise IndexError( "File does not contain any spikes with channel index %d" % channel_index) else: return spiketrain class PyNNNumpyIO(BasePyNNIO): name = "PyNN NumpyBinaryFile" extensions = ['npz'] def _read_file_contents(self): contents = numpy.load(self.filename) data = contents["data"] metadata = {} for name, value in contents['metadata']: try: metadata[name] = eval(value) except Exception: metadata[name] = value return data, metadata def _write_file_contents(self, data, metadata): max_metadata_length = max(chain([len(k) for k in metadata.keys()], [len(str(v)) for v in metadata.values()])) if PY2: dtype = "S%d" % max_metadata_length else: dtype = "U%d" % max_metadata_length metadata_array = numpy.array(sorted(metadata.items()), dtype) numpy.savez(self.filename, data=data, metadata=metadata_array) class PyNNTextIO(BasePyNNIO): name = "PyNN StandardTextFile" extensions = ['v', 'ras', 'gsyn'] def _read_metadata(self): metadata = {} with open(self.filename) as f: for line in f: if line[0] == "#": name, value = line[1:].strip().split("=") name = name.strip() try: metadata[name] = eval(value) except Exception: metadata[name] = value.strip() else: break return metadata def _read_file_contents(self): data = numpy.loadtxt(self.filename) metadata = self._read_metadata() return data, metadata def _write_file_contents(self, data, metadata): with open(self.filename, 'wb') as f: for item in sorted(metadata.items()): f.write(("# %s = %s\n" % item).encode('utf8')) numpy.savetxt(f, data)
true
true
f717dc0c491e0e926111b82f4d9a35f3ae57502b
397
py
Python
class_book/wsgi.py
3crabs/class-book
f5de12be816aa9be889d8413007be8eb4abdf45f
[ "WTFPL" ]
1
2020-11-19T14:49:41.000Z
2020-11-19T14:49:41.000Z
class_book/wsgi.py
3crabs/class-book
f5de12be816aa9be889d8413007be8eb4abdf45f
[ "WTFPL" ]
null
null
null
class_book/wsgi.py
3crabs/class-book
f5de12be816aa9be889d8413007be8eb4abdf45f
[ "WTFPL" ]
null
null
null
""" WSGI config for class_book project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.0/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'class_book.settings') application = get_wsgi_application()
23.352941
78
0.788413
import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'class_book.settings') application = get_wsgi_application()
true
true