code
stringlengths
75
104k
docstring
stringlengths
1
46.9k
def getfile(self, project_id, file_path, ref): """ Allows you to receive information about file in repository like name, size, content. Note that file content is Base64 encoded. :param project_id: project_id :param file_path: Full path to file. Ex. lib/class.rb :param ref: The name of branch, tag or commit :return: """ data = {'file_path': file_path, 'ref': ref} request = requests.get( '{0}/{1}/repository/files'.format(self.projects_url, project_id), headers=self.headers, data=data, verify=self.verify_ssl, auth=self.auth, timeout=self.timeout) if request.status_code == 200: return request.json() else: return False
Allows you to receive information about file in repository like name, size, content. Note that file content is Base64 encoded. :param project_id: project_id :param file_path: Full path to file. Ex. lib/class.rb :param ref: The name of branch, tag or commit :return:
def getUserAuthorizations(self, login, user): """ Parameters: - login - user """ self.send_getUserAuthorizations(login, user) return self.recv_getUserAuthorizations()
Parameters: - login - user
def execute(options): """execute the tool with given options.""" # Load the key in PKCS 12 format that you downloaded from the Google APIs # Console when you created your Service account. package_name = options['<package>'] source_directory = options['<output_dir>'] if options['upload'] is True: upstream = True else: upstream = False sub_tasks = {'images': options['--images'], 'listings': options['--listings'], 'inapp': options['--inapp']} if sub_tasks == {'images': False, 'listings': False, 'inapp': False}: sub_tasks = {'images': True, 'listings': True, 'inapp': True} credentials = create_credentials(credentials_file=options['--credentials'], service_email=options['--service-email'], service_key=options['--key']) command = SyncCommand( package_name, source_directory, upstream, credentials, **sub_tasks) command.execute()
execute the tool with given options.
def apply_grad_zmat_tensor(grad_C, construction_table, cart_dist): """Apply the gradient for transformation to Zmatrix space onto cart_dist. Args: grad_C (:class:`numpy.ndarray`): A ``(3, n, n, 3)`` array. The mathematical details of the index layout is explained in :meth:`~chemcoord.Cartesian.get_grad_zmat()`. construction_table (pandas.DataFrame): Explained in :meth:`~chemcoord.Cartesian.get_construction_table()`. cart_dist (:class:`~chemcoord.Cartesian`): Distortions in cartesian space. Returns: :class:`Zmat`: Distortions in Zmatrix space. """ if (construction_table.index != cart_dist.index).any(): message = "construction_table and cart_dist must use the same index" raise ValueError(message) X_dist = cart_dist.loc[:, ['x', 'y', 'z']].values.T C_dist = np.tensordot(grad_C, X_dist, axes=([3, 2], [0, 1])).T if C_dist.dtype == np.dtype('i8'): C_dist = C_dist.astype('f8') try: C_dist[:, [1, 2]] = np.rad2deg(C_dist[:, [1, 2]]) except AttributeError: C_dist[:, [1, 2]] = sympy.deg(C_dist[:, [1, 2]]) from chemcoord.internal_coordinates.zmat_class_main import Zmat cols = ['atom', 'b', 'bond', 'a', 'angle', 'd', 'dihedral'] dtypes = ['O', 'i8', 'f8', 'i8', 'f8', 'i8', 'f8'] new = pd.DataFrame(data=np.zeros((len(construction_table), 7)), index=cart_dist.index, columns=cols, dtype='f8') new = new.astype(dict(zip(cols, dtypes))) new.loc[:, ['b', 'a', 'd']] = construction_table new.loc[:, 'atom'] = cart_dist.loc[:, 'atom'] new.loc[:, ['bond', 'angle', 'dihedral']] = C_dist return Zmat(new, _metadata={'last_valid_cartesian': cart_dist})
Apply the gradient for transformation to Zmatrix space onto cart_dist. Args: grad_C (:class:`numpy.ndarray`): A ``(3, n, n, 3)`` array. The mathematical details of the index layout is explained in :meth:`~chemcoord.Cartesian.get_grad_zmat()`. construction_table (pandas.DataFrame): Explained in :meth:`~chemcoord.Cartesian.get_construction_table()`. cart_dist (:class:`~chemcoord.Cartesian`): Distortions in cartesian space. Returns: :class:`Zmat`: Distortions in Zmatrix space.
def load_csv_data(resource_name): # type: (str) -> List[str] """ Loads first column of specified CSV file from package data. """ data_bytes = pkgutil.get_data('clkhash', 'data/{}'.format(resource_name)) if data_bytes is None: raise ValueError("No data resource found with name {}".format(resource_name)) else: data = data_bytes.decode('utf8') reader = csv.reader(data.splitlines()) next(reader, None) # skip the headers return [row[0] for row in reader]
Loads first column of specified CSV file from package data.
def encode_dataset_coordinates(dataset): """Encode coordinates on the given dataset object into variable specific and global attributes. When possible, this is done according to CF conventions. Parameters ---------- dataset : Dataset Object to encode. Returns ------- variables : dict attrs : dict """ non_dim_coord_names = set(dataset.coords) - set(dataset.dims) return _encode_coordinates(dataset._variables, dataset.attrs, non_dim_coord_names=non_dim_coord_names)
Encode coordinates on the given dataset object into variable specific and global attributes. When possible, this is done according to CF conventions. Parameters ---------- dataset : Dataset Object to encode. Returns ------- variables : dict attrs : dict
def partition_read( self, session, table, key_set, transaction=None, index=None, columns=None, partition_options=None, retry=google.api_core.gapic_v1.method.DEFAULT, timeout=google.api_core.gapic_v1.method.DEFAULT, metadata=None, ): """ Creates a set of partition tokens that can be used to execute a read operation in parallel. Each of the returned partition tokens can be used by ``StreamingRead`` to specify a subset of the read result to read. The same session and read-only transaction must be used by the PartitionReadRequest used to create the partition tokens and the ReadRequests that use the partition tokens. There are no ordering guarantees on rows returned among the returned partition tokens, or even within each individual StreamingRead call issued with a partition\_token. Partition tokens become invalid when the session used to create them is deleted, is idle for too long, begins a new transaction, or becomes too old. When any of these happen, it is not possible to resume the read, and the whole operation must be restarted from the beginning. Example: >>> from google.cloud import spanner_v1 >>> >>> client = spanner_v1.SpannerClient() >>> >>> session = client.session_path('[PROJECT]', '[INSTANCE]', '[DATABASE]', '[SESSION]') >>> >>> # TODO: Initialize `table`: >>> table = '' >>> >>> # TODO: Initialize `key_set`: >>> key_set = {} >>> >>> response = client.partition_read(session, table, key_set) Args: session (str): Required. The session used to create the partitions. table (str): Required. The name of the table in the database to be read. key_set (Union[dict, ~google.cloud.spanner_v1.types.KeySet]): Required. ``key_set`` identifies the rows to be yielded. ``key_set`` names the primary keys of the rows in ``table`` to be yielded, unless ``index`` is present. If ``index`` is present, then ``key_set`` instead names index keys in ``index``. It is not an error for the ``key_set`` to name rows that do not exist in the database. Read yields nothing for nonexistent rows. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.KeySet` transaction (Union[dict, ~google.cloud.spanner_v1.types.TransactionSelector]): Read only snapshot transactions are supported, read/write and single use transactions are not. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.TransactionSelector` index (str): If non-empty, the name of an index on ``table``. This index is used instead of the table primary key when interpreting ``key_set`` and sorting result rows. See ``key_set`` for further information. columns (list[str]): The columns of ``table`` to be returned for each row matching this request. partition_options (Union[dict, ~google.cloud.spanner_v1.types.PartitionOptions]): Additional options that affect how many partitions are created. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.PartitionOptions` retry (Optional[google.api_core.retry.Retry]): A retry object used to retry requests. If ``None`` is specified, requests will not be retried. timeout (Optional[float]): The amount of time, in seconds, to wait for the request to complete. Note that if ``retry`` is specified, the timeout applies to each individual attempt. metadata (Optional[Sequence[Tuple[str, str]]]): Additional metadata that is provided to the method. Returns: A :class:`~google.cloud.spanner_v1.types.PartitionResponse` instance. Raises: google.api_core.exceptions.GoogleAPICallError: If the request failed for any reason. google.api_core.exceptions.RetryError: If the request failed due to a retryable error and retry attempts failed. ValueError: If the parameters are invalid. """ # Wrap the transport method to add retry and timeout logic. if "partition_read" not in self._inner_api_calls: self._inner_api_calls[ "partition_read" ] = google.api_core.gapic_v1.method.wrap_method( self.transport.partition_read, default_retry=self._method_configs["PartitionRead"].retry, default_timeout=self._method_configs["PartitionRead"].timeout, client_info=self._client_info, ) request = spanner_pb2.PartitionReadRequest( session=session, table=table, key_set=key_set, transaction=transaction, index=index, columns=columns, partition_options=partition_options, ) if metadata is None: metadata = [] metadata = list(metadata) try: routing_header = [("session", session)] except AttributeError: pass else: routing_metadata = google.api_core.gapic_v1.routing_header.to_grpc_metadata( routing_header ) metadata.append(routing_metadata) return self._inner_api_calls["partition_read"]( request, retry=retry, timeout=timeout, metadata=metadata )
Creates a set of partition tokens that can be used to execute a read operation in parallel. Each of the returned partition tokens can be used by ``StreamingRead`` to specify a subset of the read result to read. The same session and read-only transaction must be used by the PartitionReadRequest used to create the partition tokens and the ReadRequests that use the partition tokens. There are no ordering guarantees on rows returned among the returned partition tokens, or even within each individual StreamingRead call issued with a partition\_token. Partition tokens become invalid when the session used to create them is deleted, is idle for too long, begins a new transaction, or becomes too old. When any of these happen, it is not possible to resume the read, and the whole operation must be restarted from the beginning. Example: >>> from google.cloud import spanner_v1 >>> >>> client = spanner_v1.SpannerClient() >>> >>> session = client.session_path('[PROJECT]', '[INSTANCE]', '[DATABASE]', '[SESSION]') >>> >>> # TODO: Initialize `table`: >>> table = '' >>> >>> # TODO: Initialize `key_set`: >>> key_set = {} >>> >>> response = client.partition_read(session, table, key_set) Args: session (str): Required. The session used to create the partitions. table (str): Required. The name of the table in the database to be read. key_set (Union[dict, ~google.cloud.spanner_v1.types.KeySet]): Required. ``key_set`` identifies the rows to be yielded. ``key_set`` names the primary keys of the rows in ``table`` to be yielded, unless ``index`` is present. If ``index`` is present, then ``key_set`` instead names index keys in ``index``. It is not an error for the ``key_set`` to name rows that do not exist in the database. Read yields nothing for nonexistent rows. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.KeySet` transaction (Union[dict, ~google.cloud.spanner_v1.types.TransactionSelector]): Read only snapshot transactions are supported, read/write and single use transactions are not. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.TransactionSelector` index (str): If non-empty, the name of an index on ``table``. This index is used instead of the table primary key when interpreting ``key_set`` and sorting result rows. See ``key_set`` for further information. columns (list[str]): The columns of ``table`` to be returned for each row matching this request. partition_options (Union[dict, ~google.cloud.spanner_v1.types.PartitionOptions]): Additional options that affect how many partitions are created. If a dict is provided, it must be of the same form as the protobuf message :class:`~google.cloud.spanner_v1.types.PartitionOptions` retry (Optional[google.api_core.retry.Retry]): A retry object used to retry requests. If ``None`` is specified, requests will not be retried. timeout (Optional[float]): The amount of time, in seconds, to wait for the request to complete. Note that if ``retry`` is specified, the timeout applies to each individual attempt. metadata (Optional[Sequence[Tuple[str, str]]]): Additional metadata that is provided to the method. Returns: A :class:`~google.cloud.spanner_v1.types.PartitionResponse` instance. Raises: google.api_core.exceptions.GoogleAPICallError: If the request failed for any reason. google.api_core.exceptions.RetryError: If the request failed due to a retryable error and retry attempts failed. ValueError: If the parameters are invalid.
def write(self): """ Write result to the file. The output file is specified by ``file``. """ writer = csv.writer(self.file) for f, b in zip(self.gb.result["forward"], self.gb.result["backward"]): f = f._asdict() b = b._asdict() if not self.check_same(f, b): raise AssertionError() args_info = ", ".join(["{}: {}".format(k, v) for k, v in f["args_info"]]) out = [f["parameter_scope"], f["function_name"], f["inputs_shape"], args_info, f["mean_time"], b["mean_time"], f["n_run"], b["n_run"]] writer.writerow(out) writer.writerow([]) writer.writerow(["forward all", self.gb.result["forward_all"]]) writer.writerow( ["forward_all_n_run", self.gb.result["n_run_forward_all"]]) writer.writerow([]) writer.writerow(["backward all", self.gb.result["backward_all"]]) writer.writerow( ["backward_all_n_run", self.gb.result["n_run_backward_all"]]) if set(self.gb.result.keys()) >= {"training", "n_run_training"}: writer.writerow([]) writer.writerow( ["training(forward + backward + update)", self.gb.result["training"]]) writer.writerow( ["training_n_run", self.gb.result["n_run_training"]])
Write result to the file. The output file is specified by ``file``.
def get_knownGene_hg19(self): """ Get UCSC knownGene table for Build 37. Returns ------- pandas.DataFrame knownGene table if loading was successful, else None """ if self._knownGene_hg19 is None: self._knownGene_hg19 = self._load_knownGene(self._get_path_knownGene_hg19()) return self._knownGene_hg19
Get UCSC knownGene table for Build 37. Returns ------- pandas.DataFrame knownGene table if loading was successful, else None
def Stichlmair_dry(Vg, rhog, mug, voidage, specific_area, C1, C2, C3, H=1.): r'''Calculates dry pressure drop across a packed column, using the Stichlmair [1]_ correlation. Uses three regressed constants for each type of packing, and voidage and specific area. Pressure drop is given by: .. math:: \Delta P_{dry} = \frac{3}{4} f_0 \frac{1-\epsilon}{\epsilon^{4.65}} \rho_G \frac{H}{d_p}V_g^2 .. math:: f_0 = \frac{C_1}{Re_g} + \frac{C_2}{Re_g^{0.5}} + C_3 .. math:: d_p = \frac{6(1-\epsilon)}{a} Parameters ---------- Vg : float Superficial velocity of gas, Q/A [m/s] rhog : float Density of gas [kg/m^3] mug : float Viscosity of gas [Pa*s] voidage : float Voidage of bed of packing material [] specific_area : float Specific area of the packing material [m^2/m^3] C1 : float Packing-specific constant [] C2 : float Packing-specific constant [] C3 : float Packing-specific constant [] H : float, optional Height of packing [m] Returns ------- dP_dry : float Pressure drop across dry packing [Pa] Notes ----- This model is used by most process simulation tools. If H is not provided, it defaults to 1. If Z is not provided, it defaults to 1. Examples -------- >>> Stichlmair_dry(Vg=0.4, rhog=5., mug=5E-5, voidage=0.68, ... specific_area=260., C1=32., C2=7, C3=1) 236.80904286559885 References ---------- .. [1] Stichlmair, J., J. L. Bravo, and J. R. Fair. "General Model for Prediction of Pressure Drop and Capacity of Countercurrent Gas/liquid Packed Columns." Gas Separation & Purification 3, no. 1 (March 1989): 19-28. doi:10.1016/0950-4214(89)80016-7. ''' dp = 6*(1-voidage)/specific_area Re = Vg*rhog*dp/mug f0 = C1/Re + C2/Re**0.5 + C3 return 3/4.*f0*(1-voidage)/voidage**4.65*rhog*H/dp*Vg**2
r'''Calculates dry pressure drop across a packed column, using the Stichlmair [1]_ correlation. Uses three regressed constants for each type of packing, and voidage and specific area. Pressure drop is given by: .. math:: \Delta P_{dry} = \frac{3}{4} f_0 \frac{1-\epsilon}{\epsilon^{4.65}} \rho_G \frac{H}{d_p}V_g^2 .. math:: f_0 = \frac{C_1}{Re_g} + \frac{C_2}{Re_g^{0.5}} + C_3 .. math:: d_p = \frac{6(1-\epsilon)}{a} Parameters ---------- Vg : float Superficial velocity of gas, Q/A [m/s] rhog : float Density of gas [kg/m^3] mug : float Viscosity of gas [Pa*s] voidage : float Voidage of bed of packing material [] specific_area : float Specific area of the packing material [m^2/m^3] C1 : float Packing-specific constant [] C2 : float Packing-specific constant [] C3 : float Packing-specific constant [] H : float, optional Height of packing [m] Returns ------- dP_dry : float Pressure drop across dry packing [Pa] Notes ----- This model is used by most process simulation tools. If H is not provided, it defaults to 1. If Z is not provided, it defaults to 1. Examples -------- >>> Stichlmair_dry(Vg=0.4, rhog=5., mug=5E-5, voidage=0.68, ... specific_area=260., C1=32., C2=7, C3=1) 236.80904286559885 References ---------- .. [1] Stichlmair, J., J. L. Bravo, and J. R. Fair. "General Model for Prediction of Pressure Drop and Capacity of Countercurrent Gas/liquid Packed Columns." Gas Separation & Purification 3, no. 1 (March 1989): 19-28. doi:10.1016/0950-4214(89)80016-7.
def send_message(self, subject=None, text=None, markdown=None, message_dict=None): """ Helper function to send a message to a group """ message = FiestaMessage(self.api, self, subject, text, markdown, message_dict) return message.send()
Helper function to send a message to a group
def getTraceCombosByIds(self, trace_ids, adjust): """ Not content with just one of traces, summaries or timelines? Want it all? This is the method for you. Parameters: - trace_ids - adjust """ self.send_getTraceCombosByIds(trace_ids, adjust) return self.recv_getTraceCombosByIds()
Not content with just one of traces, summaries or timelines? Want it all? This is the method for you. Parameters: - trace_ids - adjust
def _make_annulus_path(patch_inner, patch_outer): """ Defines a matplotlib annulus path from two patches. This preserves the cubic Bezier curves (CURVE4) of the aperture paths. # This is borrowed from photutils aperture. """ import matplotlib.path as mpath path_inner = patch_inner.get_path() transform_inner = patch_inner.get_transform() path_inner = transform_inner.transform_path(path_inner) path_outer = patch_outer.get_path() transform_outer = patch_outer.get_transform() path_outer = transform_outer.transform_path(path_outer) verts_inner = path_inner.vertices[:-1][::-1] verts_inner = np.concatenate((verts_inner, [verts_inner[-1]])) verts = np.vstack((path_outer.vertices, verts_inner)) codes = np.hstack((path_outer.codes, path_inner.codes)) return mpath.Path(verts, codes)
Defines a matplotlib annulus path from two patches. This preserves the cubic Bezier curves (CURVE4) of the aperture paths. # This is borrowed from photutils aperture.
def cli(env, package_keyname, keyword, category): """List package items used for ordering. The item keyNames listed can be used with `slcli order place` to specify the items that are being ordered in the package. .. Note:: Items with a numbered category, like disk0 or gpu0, can be included multiple times in an order to match how many of the item you want to order. :: # List all items in the VSI package slcli order item-list CLOUD_SERVER # List Ubuntu OSes from the os category of the Bare Metal package slcli order item-list BARE_METAL_SERVER --category os --keyword ubuntu """ table = formatting.Table(COLUMNS) manager = ordering.OrderingManager(env.client) _filter = {'items': {}} if keyword: _filter['items']['description'] = {'operation': '*= %s' % keyword} if category: _filter['items']['categories'] = {'categoryCode': {'operation': '_= %s' % category}} items = manager.list_items(package_keyname, filter=_filter) sorted_items = sort_items(items) categories = sorted_items.keys() for catname in sorted(categories): for item in sorted_items[catname]: table.add_row([catname, item['keyName'], item['description'], get_price(item)]) env.fout(table)
List package items used for ordering. The item keyNames listed can be used with `slcli order place` to specify the items that are being ordered in the package. .. Note:: Items with a numbered category, like disk0 or gpu0, can be included multiple times in an order to match how many of the item you want to order. :: # List all items in the VSI package slcli order item-list CLOUD_SERVER # List Ubuntu OSes from the os category of the Bare Metal package slcli order item-list BARE_METAL_SERVER --category os --keyword ubuntu
def absent(name, user, enc='ssh-rsa', comment='', source='', options=None, config='.ssh/authorized_keys', fingerprint_hash_type=None): ''' Verifies that the specified SSH key is absent name The SSH key to manage user The user who owns the SSH authorized keys file to modify enc Defines what type of key is being used; can be ed25519, ecdsa, ssh-rsa or ssh-dss comment The comment to be placed with the SSH public key options The options passed to the key, pass a list object source The source file for the key(s). Can contain any number of public keys, in standard "authorized_keys" format. If this is set, comment, enc and options will be ignored. .. versionadded:: 2015.8.0 config The location of the authorized keys file relative to the user's home directory, defaults to ".ssh/authorized_keys". Token expansion %u and %h for username and home path supported. fingerprint_hash_type The public key fingerprint hash type that the public key fingerprint was originally hashed with. This defaults to ``sha256`` if not specified. .. versionadded:: 2016.11.7 ''' ret = {'name': name, 'changes': {}, 'result': True, 'comment': ''} if __opts__['test']: ret['result'], ret['comment'] = _absent_test( user, name, enc, comment, options or [], source, config, fingerprint_hash_type) return ret # Extract Key from file if source is present if source != '': key = __salt__['cp.get_file_str']( source, saltenv=__env__) filehasoptions = False # check if this is of form {options} {enc} {key} {comment} sshre = re.compile(r'^(ssh\-|ecds).*') key = key.rstrip().split('\n') for keyline in key: filehasoptions = sshre.match(keyline) if not filehasoptions: ret['comment'] = __salt__['ssh.rm_auth_key_from_file'](user, source, config, saltenv=__env__, fingerprint_hash_type=fingerprint_hash_type) else: # Split keyline to get key keyline = keyline.split(' ') ret['comment'] = __salt__['ssh.rm_auth_key'](user, keyline[1], config=config, fingerprint_hash_type=fingerprint_hash_type) else: # Get just the key sshre = re.compile(r'^(.*?)\s?((?:ssh\-|ecds)[\w-]+\s.+)$') fullkey = sshre.search(name) # if it is {key} [comment] if not fullkey: key_and_comment = name.split(None, 1) name = key_and_comment[0] if len(key_and_comment) == 2: comment = key_and_comment[1] else: # if there are options, set them if fullkey.group(1): options = fullkey.group(1).split(',') # key is of format: {enc} {key} [comment] comps = fullkey.group(2).split() enc = comps[0] name = comps[1] if len(comps) == 3: comment = comps[2] ret['comment'] = __salt__['ssh.rm_auth_key'](user, name, config=config, fingerprint_hash_type=fingerprint_hash_type) if ret['comment'] == 'User authorized keys file not present': ret['result'] = False return ret elif ret['comment'] == 'Key removed': ret['changes'][name] = 'Removed' return ret
Verifies that the specified SSH key is absent name The SSH key to manage user The user who owns the SSH authorized keys file to modify enc Defines what type of key is being used; can be ed25519, ecdsa, ssh-rsa or ssh-dss comment The comment to be placed with the SSH public key options The options passed to the key, pass a list object source The source file for the key(s). Can contain any number of public keys, in standard "authorized_keys" format. If this is set, comment, enc and options will be ignored. .. versionadded:: 2015.8.0 config The location of the authorized keys file relative to the user's home directory, defaults to ".ssh/authorized_keys". Token expansion %u and %h for username and home path supported. fingerprint_hash_type The public key fingerprint hash type that the public key fingerprint was originally hashed with. This defaults to ``sha256`` if not specified. .. versionadded:: 2016.11.7
def load_dic28(): """DIC28 Dataset from Pajek. This network represents connections among English words in a dictionary. It was generated from Knuth's dictionary. Two words are connected by an edge if we can reach one from the other by - changing a single character (e. g., work - word) - adding / removing a single character (e. g., ever - fever). There exist 52,652 words (vertices in a network) having 2 up to 8 characters in the dictionary. The obtained network has 89038 edges. """ dataset_path = _load('dic28') X = _load_csv(dataset_path, 'data') y = X.pop('label').values graph1 = nx.Graph(nx.read_gml(os.path.join(dataset_path, 'graph1.gml'))) graph2 = nx.Graph(nx.read_gml(os.path.join(dataset_path, 'graph2.gml'))) graph = graph1.copy() graph.add_nodes_from(graph2.nodes(data=True)) graph.add_edges_from(graph2.edges) graph.add_edges_from(X[['graph1', 'graph2']].values) graphs = { 'graph1': graph1, 'graph2': graph2, } return Dataset(load_dic28.__doc__, X, y, accuracy_score, stratify=True, graph=graph, graphs=graphs)
DIC28 Dataset from Pajek. This network represents connections among English words in a dictionary. It was generated from Knuth's dictionary. Two words are connected by an edge if we can reach one from the other by - changing a single character (e. g., work - word) - adding / removing a single character (e. g., ever - fever). There exist 52,652 words (vertices in a network) having 2 up to 8 characters in the dictionary. The obtained network has 89038 edges.
def mouse(table, day=None): """Handler for showing mouse statistics for specified type and day.""" where = (("day", day),) if day else () events = db.fetch(table, where=where, order="day") for e in events: e["dt"] = datetime.datetime.fromtimestamp(e["stamp"]) stats, positions, events = stats_mouse(events, table) days, input = db.fetch("counts", order="day", type=table), "mouse" return bottle.template("heatmap.tpl", locals(), conf=conf)
Handler for showing mouse statistics for specified type and day.
def set_domain_workgroup(workgroup): ''' Set the domain or workgroup the computer belongs to. .. versionadded:: 2019.2.0 Returns: bool: ``True`` if successful, otherwise ``False`` CLI Example: .. code-block:: bash salt 'minion-id' system.set_domain_workgroup LOCAL ''' if six.PY2: workgroup = _to_unicode(workgroup) # Initialize COM with salt.utils.winapi.Com(): # Grab the first Win32_ComputerSystem object from wmi conn = wmi.WMI() comp = conn.Win32_ComputerSystem()[0] # Now we can join the new workgroup res = comp.JoinDomainOrWorkgroup(Name=workgroup.upper()) return True if not res[0] else False
Set the domain or workgroup the computer belongs to. .. versionadded:: 2019.2.0 Returns: bool: ``True`` if successful, otherwise ``False`` CLI Example: .. code-block:: bash salt 'minion-id' system.set_domain_workgroup LOCAL
def values(self): "Returns all values this object can return via keys." return tuple(set(self.new.values()).union(self.old.values()))
Returns all values this object can return via keys.
def get_task_instances(self, state=None, session=None): """ Returns the task instances for this dag run """ from airflow.models.taskinstance import TaskInstance # Avoid circular import tis = session.query(TaskInstance).filter( TaskInstance.dag_id == self.dag_id, TaskInstance.execution_date == self.execution_date, ) if state: if isinstance(state, six.string_types): tis = tis.filter(TaskInstance.state == state) else: # this is required to deal with NULL values if None in state: tis = tis.filter( or_(TaskInstance.state.in_(state), TaskInstance.state.is_(None)) ) else: tis = tis.filter(TaskInstance.state.in_(state)) if self.dag and self.dag.partial: tis = tis.filter(TaskInstance.task_id.in_(self.dag.task_ids)) return tis.all()
Returns the task instances for this dag run
def write_bel_namespace(self, file: TextIO, use_names: bool = False) -> None: """Write as a BEL namespace file.""" if not self.is_populated(): self.populate() if use_names and not self.has_names: raise ValueError values = ( self._get_namespace_name_to_encoding(desc='writing names') if use_names else self._get_namespace_identifier_to_encoding(desc='writing identifiers') ) write_namespace( namespace_name=self._get_namespace_name(), namespace_keyword=self._get_namespace_keyword(), namespace_query_url=self.identifiers_url, values=values, file=file, )
Write as a BEL namespace file.
def _clear_screen(): """ http://stackoverflow.com/questions/18937058/python-clear-screen-in-shell """ if platform.system() == "Windows": tmp = os.system('cls') #for window else: tmp = os.system('clear') #for Linux return True
http://stackoverflow.com/questions/18937058/python-clear-screen-in-shell
def create_prefetch(self, addresses): """Create futures needed before starting the process of reading the address's value from the merkle tree. Args: addresses (list of str): addresses in the txn's inputs that aren't in any base context (or any in the chain). """ with self._lock: for add in addresses: self._state[add] = _ContextFuture(address=add, wait_for_tree=True)
Create futures needed before starting the process of reading the address's value from the merkle tree. Args: addresses (list of str): addresses in the txn's inputs that aren't in any base context (or any in the chain).
def get_float_relative(strings: Sequence[str], prefix1: str, delta: int, prefix2: str, ignoreleadingcolon: bool = False) -> Optional[float]: """ Fetches a float parameter via :func:`get_string_relative`. """ return get_float_raw(get_string_relative( strings, prefix1, delta, prefix2, ignoreleadingcolon=ignoreleadingcolon))
Fetches a float parameter via :func:`get_string_relative`.
async def send_message(self, message, *, end=False): """Coroutine to send message to the server. If client sends UNARY request, then you should call this coroutine only once. If client sends STREAM request, then you can call this coroutine as many times as you need. .. warning:: It is important to finally end stream from the client-side when you finished sending messages. You can do this in two ways: - specify ``end=True`` argument while sending last message - and last DATA frame will include END_STREAM flag; - call :py:meth:`end` coroutine after sending last message - and extra HEADERS frame with END_STREAM flag will be sent. First approach is preferred, because it doesn't require sending additional HTTP/2 frame. """ if not self._send_request_done: await self.send_request() if end and self._end_done: raise ProtocolError('Stream was already ended') with self._wrapper: message, = await self._dispatch.send_message(message) await send_message(self._stream, self._codec, message, self._send_type, end=end) self._send_message_count += 1 if end: self._end_done = True
Coroutine to send message to the server. If client sends UNARY request, then you should call this coroutine only once. If client sends STREAM request, then you can call this coroutine as many times as you need. .. warning:: It is important to finally end stream from the client-side when you finished sending messages. You can do this in two ways: - specify ``end=True`` argument while sending last message - and last DATA frame will include END_STREAM flag; - call :py:meth:`end` coroutine after sending last message - and extra HEADERS frame with END_STREAM flag will be sent. First approach is preferred, because it doesn't require sending additional HTTP/2 frame.
def get_candidate_config(self, merge=False, formal=False): """ Retrieve the configuration loaded as candidate config in your configuration session. :param merge: Merge candidate config with running config to return the complete configuration including all changed :param formal: Return configuration in IOS-XR formal config format """ command = "show configuration" if merge: command += " merge" if formal: command += " formal" response = self._execute_config_show(command) match = re.search(".*(!! IOS XR Configuration.*)$", response, re.DOTALL) if match is not None: response = match.group(1) return response
Retrieve the configuration loaded as candidate config in your configuration session. :param merge: Merge candidate config with running config to return the complete configuration including all changed :param formal: Return configuration in IOS-XR formal config format
def assert_raises_errno(exception, errno, msg_fmt="{msg}"): """Fail unless an exception with a specific errno is raised with the context. >>> with assert_raises_errno(OSError, 42): ... raise OSError(42, "OS Error") ... >>> with assert_raises_errno(OSError, 44): ... raise OSError(17, "OS Error") ... Traceback (most recent call last): ... AssertionError: wrong errno: 44 != 17 The following msg_fmt arguments are supported: * msg - the default error message * exc_type - exception type that is expected * exc_name - expected exception type name * expected_errno - * actual_errno - raised errno or None if no matching exception was raised """ def check_errno(exc): if errno != exc.errno: msg = "wrong errno: {!r} != {!r}".format(errno, exc.errno) fail( msg_fmt.format( msg=msg, exc_type=exception, exc_name=exception.__name__, expected_errno=errno, actual_errno=exc.errno, ) ) context = AssertRaisesErrnoContext(exception, errno, msg_fmt) context.add_test(check_errno) return context
Fail unless an exception with a specific errno is raised with the context. >>> with assert_raises_errno(OSError, 42): ... raise OSError(42, "OS Error") ... >>> with assert_raises_errno(OSError, 44): ... raise OSError(17, "OS Error") ... Traceback (most recent call last): ... AssertionError: wrong errno: 44 != 17 The following msg_fmt arguments are supported: * msg - the default error message * exc_type - exception type that is expected * exc_name - expected exception type name * expected_errno - * actual_errno - raised errno or None if no matching exception was raised
def typevalue(self, key, value): """Given a parameter identified by ``key`` and an untyped string, convert that string to the type that our version of key has. """ def listconvert(value): # this function might be called with both string # represenations of entire lists and simple (unquoted) # strings. String representations come in two flavours, # the (legacy/deprecated) python literal (eg "['foo', # 'bar']") and the simple (eg "foo, bar") The # ast.literal_eval handles the first case, and if the # value can't be parsed as a python expression, the second # way is attempted. If both fail, it is returned verbatim # (not wrapped in a list, for reasons) try: return ast.literal_eval(value) except (SyntaxError, ValueError): if "," in value: return [x.strip() for x in value.split(",")] else: return value # self.get(key) should never fail default = self.get(key) # if type(default) == type: if inspect.isclass(default): # print("Using class for %s" % key) t = default else: # print("Using instance for %s" % key) t = type(default) if t == bool: t = LayeredConfig.boolconvert elif t == list: t = listconvert elif t == date: t = LayeredConfig.dateconvert elif t == datetime: t = LayeredConfig.datetimeconvert # print("Converting %r to %r" % (value,t(value))) return t(value)
Given a parameter identified by ``key`` and an untyped string, convert that string to the type that our version of key has.
def inform(self, reading): """Inform strategy creator of the sensor status.""" try: self._inform_callback(self._sensor, reading) except Exception: log.exception('Unhandled exception trying to send {!r} ' 'for sensor {!r} of type {!r}' .format(reading, self._sensor.name, self._sensor.type))
Inform strategy creator of the sensor status.
def files_read(self, path, offset=0, count=None, **kwargs): """Reads a file stored in the MFS. .. code-block:: python >>> c.files_read("/bla/file") b'hi' Parameters ---------- path : str Filepath within the MFS offset : int Byte offset at which to begin reading at count : int Maximum number of bytes to read Returns ------- str : MFS file contents """ opts = {"offset": offset} if count is not None: opts["count"] = count kwargs.setdefault("opts", opts) args = (path,) return self._client.request('/files/read', args, **kwargs)
Reads a file stored in the MFS. .. code-block:: python >>> c.files_read("/bla/file") b'hi' Parameters ---------- path : str Filepath within the MFS offset : int Byte offset at which to begin reading at count : int Maximum number of bytes to read Returns ------- str : MFS file contents
def compare(testsuite, gold, select='i-id i-input mrs'): """ Compare two [incr tsdb()] profiles. Args: testsuite (str, TestSuite): path to the test [incr tsdb()] testsuite or a :class:`TestSuite` object gold (str, TestSuite): path to the gold [incr tsdb()] testsuite or a :class:`TestSuite` object select: TSQL query to select (id, input, mrs) triples (default: `i-id i-input mrs`) Yields: dict: Comparison results as:: {"id": "item identifier", "input": "input sentence", "test": number_of_unique_results_in_test, "shared": number_of_shared_results, "gold": number_of_unique_results_in_gold} """ from delphin.mrs import simplemrs, compare as mrs_compare if not isinstance(testsuite, itsdb.TestSuite): if isinstance(testsuite, itsdb.ItsdbProfile): testsuite = testsuite.root testsuite = itsdb.TestSuite(testsuite) if not isinstance(gold, itsdb.TestSuite): if isinstance(gold, itsdb.ItsdbProfile): gold = gold.root gold = itsdb.TestSuite(gold) queryobj = tsql.inspect_query('select ' + select) if len(queryobj['projection']) != 3: raise ValueError('select does not return 3 fields: ' + select) input_select = '{} {}'.format(queryobj['projection'][0], queryobj['projection'][1]) i_inputs = dict(tsql.select(input_select, testsuite)) matched_rows = itsdb.match_rows( tsql.select(select, testsuite), tsql.select(select, gold), 0) for (key, testrows, goldrows) in matched_rows: (test_unique, shared, gold_unique) = mrs_compare.compare_bags( [simplemrs.loads_one(row[2]) for row in testrows], [simplemrs.loads_one(row[2]) for row in goldrows]) yield {'id': key, 'input': i_inputs[key], 'test': test_unique, 'shared': shared, 'gold': gold_unique}
Compare two [incr tsdb()] profiles. Args: testsuite (str, TestSuite): path to the test [incr tsdb()] testsuite or a :class:`TestSuite` object gold (str, TestSuite): path to the gold [incr tsdb()] testsuite or a :class:`TestSuite` object select: TSQL query to select (id, input, mrs) triples (default: `i-id i-input mrs`) Yields: dict: Comparison results as:: {"id": "item identifier", "input": "input sentence", "test": number_of_unique_results_in_test, "shared": number_of_shared_results, "gold": number_of_unique_results_in_gold}
def find_path(name, path=None, exact=False): """ Search for a file or directory on your local filesystem by name (file must be in a directory specified in a PATH environment variable) Args: fname (PathLike or str): file name to match. If exact is False this may be a glob pattern path (str or Iterable[PathLike]): list of directories to search either specified as an os.pathsep separated string or a list of directories. Defaults to environment PATH. exact (bool): if True, only returns exact matches. Default False. Notes: For recursive behavior set `path=(d for d, _, _ in os.walk('.'))`, where '.' might be replaced by the root directory of interest. Example: >>> list(find_path('ping', exact=True)) >>> list(find_path('bin')) >>> list(find_path('bin')) >>> list(find_path('*cc*')) >>> list(find_path('cmake*')) Example: >>> import ubelt as ub >>> from os.path import dirname >>> path = dirname(dirname(ub.util_platform.__file__)) >>> res = sorted(find_path('ubelt/util_*.py', path=path)) >>> assert len(res) >= 10 >>> res = sorted(find_path('ubelt/util_platform.py', path=path, exact=True)) >>> print(res) >>> assert len(res) == 1 """ path = os.environ.get('PATH', os.defpath) if path is None else path dpaths = path.split(os.pathsep) if isinstance(path, six.string_types) else path candidates = (join(dpath, name) for dpath in dpaths) if exact: if WIN32: # nocover pathext = [''] + os.environ.get('PATHEXT', '').split(os.pathsep) candidates = (p + ext for p in candidates for ext in pathext) candidates = filter(exists, candidates) else: import glob candidates = it.chain.from_iterable( glob.glob(pattern) for pattern in candidates) return candidates
Search for a file or directory on your local filesystem by name (file must be in a directory specified in a PATH environment variable) Args: fname (PathLike or str): file name to match. If exact is False this may be a glob pattern path (str or Iterable[PathLike]): list of directories to search either specified as an os.pathsep separated string or a list of directories. Defaults to environment PATH. exact (bool): if True, only returns exact matches. Default False. Notes: For recursive behavior set `path=(d for d, _, _ in os.walk('.'))`, where '.' might be replaced by the root directory of interest. Example: >>> list(find_path('ping', exact=True)) >>> list(find_path('bin')) >>> list(find_path('bin')) >>> list(find_path('*cc*')) >>> list(find_path('cmake*')) Example: >>> import ubelt as ub >>> from os.path import dirname >>> path = dirname(dirname(ub.util_platform.__file__)) >>> res = sorted(find_path('ubelt/util_*.py', path=path)) >>> assert len(res) >= 10 >>> res = sorted(find_path('ubelt/util_platform.py', path=path, exact=True)) >>> print(res) >>> assert len(res) == 1
def lrange(self, key, start, stop): """Emulate lrange.""" redis_list = self._get_list(key, 'LRANGE') start, stop = self._translate_range(len(redis_list), start, stop) return redis_list[start:stop + 1]
Emulate lrange.
def make_exttrig_file(cp, ifos, sci_seg, out_dir): ''' Make an ExtTrig xml file containing information on the external trigger Parameters ---------- cp : pycbc.workflow.configuration.WorkflowConfigParser object The parsed configuration options of a pycbc.workflow.core.Workflow. ifos : str String containing the analysis interferometer IDs. sci_seg : ligo.segments.segment The science segment for the analysis run. out_dir : str The output directory, destination for xml file. Returns ------- xml_file : pycbc.workflow.File object The xml file with external trigger information. ''' # Initialise objects xmldoc = ligolw.Document() xmldoc.appendChild(ligolw.LIGO_LW()) tbl = lsctables.New(lsctables.ExtTriggersTable) cols = tbl.validcolumns xmldoc.childNodes[-1].appendChild(tbl) row = tbl.appendRow() # Add known attributes for this GRB setattr(row, "event_ra", float(cp.get("workflow", "ra"))) setattr(row, "event_dec", float(cp.get("workflow", "dec"))) setattr(row, "start_time", int(cp.get("workflow", "trigger-time"))) setattr(row, "event_number_grb", str(cp.get("workflow", "trigger-name"))) # Fill in all empty rows for entry in cols.keys(): if not hasattr(row, entry): if cols[entry] in ['real_4','real_8']: setattr(row,entry,0.) elif cols[entry] == 'int_4s': setattr(row,entry,0) elif cols[entry] == 'lstring': setattr(row,entry,'') elif entry == 'process_id': row.process_id = ilwd.ilwdchar("external_trigger:process_id:0") elif entry == 'event_id': row.event_id = ilwd.ilwdchar("external_trigger:event_id:0") else: print("Column %s not recognized" %(entry), file=sys.stderr) raise ValueError # Save file xml_file_name = "triggerGRB%s.xml" % str(cp.get("workflow", "trigger-name")) xml_file_path = os.path.join(out_dir, xml_file_name) utils.write_filename(xmldoc, xml_file_path) xml_file_url = urlparse.urljoin("file:", urllib.pathname2url(xml_file_path)) xml_file = File(ifos, xml_file_name, sci_seg, file_url=xml_file_url) xml_file.PFN(xml_file_url, site="local") return xml_file
Make an ExtTrig xml file containing information on the external trigger Parameters ---------- cp : pycbc.workflow.configuration.WorkflowConfigParser object The parsed configuration options of a pycbc.workflow.core.Workflow. ifos : str String containing the analysis interferometer IDs. sci_seg : ligo.segments.segment The science segment for the analysis run. out_dir : str The output directory, destination for xml file. Returns ------- xml_file : pycbc.workflow.File object The xml file with external trigger information.
def to_google(self, type, label, issuer, counter=None): """Generate the otpauth protocal string for Google Authenticator. .. deprecated:: 0.2.0 Use :func:`to_uri` instead. """ warnings.warn('deprecated, use to_uri instead', DeprecationWarning) return self.to_uri(type, label, issuer, counter)
Generate the otpauth protocal string for Google Authenticator. .. deprecated:: 0.2.0 Use :func:`to_uri` instead.
def get_bestnr(self, index=4.0, nhigh=3.0, null_snr_threshold=4.25,\ null_grad_thresh=20., null_grad_val = 1./5.): """ Return the BestNR statistic for this row. """ # weight SNR by chisq bestnr = self.get_new_snr(index=index, nhigh=nhigh, column="chisq") if len(self.get_ifos()) < 3: return bestnr # recontour null SNR threshold for higher SNRs if self.snr > null_grad_thresh: null_snr_threshold += (self.snr - null_grad_thresh) * null_grad_val # weight SNR by null SNR if self.get_null_snr() > null_snr_threshold: bestnr /= 1 + self.get_null_snr() - null_snr_threshold return bestnr
Return the BestNR statistic for this row.
def as_bel(self) -> str: """Return this protein modification variant as a BEL string.""" return 'pmod({}{})'.format( str(self[IDENTIFIER]), ''.join(', {}'.format(self[x]) for x in PMOD_ORDER[2:] if x in self) )
Return this protein modification variant as a BEL string.
def numSegments(self, cell=None): """ Returns the number of segments. :param cell: (int) Optional parameter to get the number of segments on a cell. :returns: (int) Number of segments on all cells if cell is not specified, or on a specific specified cell """ if cell is not None: return len(self._cells[cell]._segments) return self._nextFlatIdx - len(self._freeFlatIdxs)
Returns the number of segments. :param cell: (int) Optional parameter to get the number of segments on a cell. :returns: (int) Number of segments on all cells if cell is not specified, or on a specific specified cell
def patch_namespaced_custom_object(self, group, version, namespace, plural, name, body, **kwargs): """ patch the specified namespace scoped custom object This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_custom_object(group, version, namespace, plural, name, body, async_req=True) >>> result = thread.get() :param async_req bool :param str group: the custom resource's group (required) :param str version: the custom resource's version (required) :param str namespace: The custom resource's namespace (required) :param str plural: the custom resource's plural name. For TPRs this would be lowercase plural kind. (required) :param str name: the custom object's name (required) :param object body: The JSON schema of the Resource to patch. (required) :return: object If the method is called asynchronously, returns the request thread. """ kwargs['_return_http_data_only'] = True if kwargs.get('async_req'): return self.patch_namespaced_custom_object_with_http_info(group, version, namespace, plural, name, body, **kwargs) else: (data) = self.patch_namespaced_custom_object_with_http_info(group, version, namespace, plural, name, body, **kwargs) return data
patch the specified namespace scoped custom object This method makes a synchronous HTTP request by default. To make an asynchronous HTTP request, please pass async_req=True >>> thread = api.patch_namespaced_custom_object(group, version, namespace, plural, name, body, async_req=True) >>> result = thread.get() :param async_req bool :param str group: the custom resource's group (required) :param str version: the custom resource's version (required) :param str namespace: The custom resource's namespace (required) :param str plural: the custom resource's plural name. For TPRs this would be lowercase plural kind. (required) :param str name: the custom object's name (required) :param object body: The JSON schema of the Resource to patch. (required) :return: object If the method is called asynchronously, returns the request thread.
def _delete_reminders_from_list( self, listName): """* delete reminders from list* **Key Arguments:** - ``listName`` -- the name of the reminders list """ self.log.info('starting the ``_delete_reminders_from_list`` method') applescript = """ tell application "Reminders" -- Loop thru reminders in the targeted reminder.app list tell list "%(listName)s" delete every reminder end tell end tell """ % locals() cmd = "\n".join(["osascript << EOT", applescript, "EOT"]) p = Popen(cmd, stdout=PIPE, stderr=PIPE, shell=True) stdout, stderr = p.communicate() if len(stderr): self.log.error(stderr) sys.exit(0) self.log.info('completed the ``_delete_reminders_from_list`` method') return None
* delete reminders from list* **Key Arguments:** - ``listName`` -- the name of the reminders list
def get_image(self, component_info=None, data=None, component_position=None): """Get image.""" components = [] append_components = components.append for _ in range(component_info.image_count): component_position, image_info = QRTPacket._get_exact( RTImage, data, component_position ) append_components((image_info, data[component_position:-1])) return components
Get image.
def render(self, rect, data): """Draws the cells in grid.""" size = self.get_minimum_size(data) # Find how much extra space we have. extra_width = rect.w - size.x extra_height = rect.h - size.y # Distribute the extra space into the correct rows and columns. if self.scaling_col is None or not 0 <= self.scaling_col < self.cols: width_per_col = extra_width / float(self.cols) col_widths = [ width + width_per_col for width in self.col_widths ] else: col_widths = self.col_widths[:] col_widths[self.scaling_col] += extra_width if self.scaling_row is None or not 0 <= self.scaling_row < self.rows: height_per_row = extra_height / float(self.rows) row_heights = [ height + height_per_row for height in self.row_heights ] else: row_heights = self.row_heights[:] row_heights[self.scaling_row] += extra_height # Find the (start, end) positions of each row and column. col_xs = [] last_x = rect.left + self.outside_margin for width in col_widths: col_xs.append((last_x, last_x + width)) last_x += width + self.margin row_ys = [] last_y = rect.top - self.outside_margin for height in row_heights: row_ys.append((last_y, last_y - height)) last_y -= height + self.margin # Now we can loop over the elements and have them rendered. for col, row, cols, rows, element in self.elements: x_start = col_xs[col][0] y_start = row_ys[row][0] x_end = col_xs[col+cols-1][1] y_end = row_ys[row+rows-1][1] element.render(datatypes.Rectangle( x_start, y_end, x_end-x_start, y_start-y_end ), data) # And finally we can draw the rules def _get_value(array, index, sign): """Returns the value of the index in the given array, where the array (like col_xs and row_ys), consists of start-end pairs of values.""" if index <= 0: # Special case, it is the start of the first range return array[0][0]-self.outside_margin*sign elif index >= len(array): # Special case, it is the end of the last range return array[-1][1]+self.outside_margin*sign else: # Otherwise it is the blend of a start and end. return (array[index-1][1] + array[index][0])*0.5 for start_col, start_row, end_col, end_row, width, color in self.rules: x_start = _get_value(col_xs, start_col, 1) y_start = _get_value(row_ys, start_row, -1) x_end = _get_value(col_xs, end_col, 1) y_end = _get_value(row_ys, end_row, -1) data['output'].line( x_start, y_start, x_end, y_end, stroke=color, stroke_width=width )
Draws the cells in grid.
def password_option(*param_decls, **attrs): """Shortcut for password prompts. This is equivalent to decorating a function with :func:`option` with the following parameters:: @click.command() @click.option('--password', prompt=True, confirmation_prompt=True, hide_input=True) def changeadmin(password): pass """ def decorator(f): attrs.setdefault('prompt', True) attrs.setdefault('confirmation_prompt', True) attrs.setdefault('hide_input', True) return option(*(param_decls or ('--password',)), **attrs)(f) return decorator
Shortcut for password prompts. This is equivalent to decorating a function with :func:`option` with the following parameters:: @click.command() @click.option('--password', prompt=True, confirmation_prompt=True, hide_input=True) def changeadmin(password): pass
def _zip_files(files, root): """Generates a ZIP file in-memory from a list of files. Files will be stored in the archive with relative names, and have their UNIX permissions forced to 755 or 644 (depending on whether they are user-executable in the source filesystem). Args: files (list[str]): file names to add to the archive, relative to ``root``. root (str): base directory to retrieve files from. Returns: str: content of the ZIP file as a byte string. str: A calculated hash of all the files. """ zip_data = StringIO() with ZipFile(zip_data, 'w', ZIP_DEFLATED) as zip_file: for fname in files: zip_file.write(os.path.join(root, fname), fname) # Fix file permissions to avoid any issues - only care whether a file # is executable or not, choosing between modes 755 and 644 accordingly. for zip_entry in zip_file.filelist: perms = (zip_entry.external_attr & ZIP_PERMS_MASK) >> 16 if perms & stat.S_IXUSR != 0: new_perms = 0o755 else: new_perms = 0o644 if new_perms != perms: logger.debug("lambda: fixing perms: %s: %o => %o", zip_entry.filename, perms, new_perms) new_attr = ((zip_entry.external_attr & ~ZIP_PERMS_MASK) | (new_perms << 16)) zip_entry.external_attr = new_attr contents = zip_data.getvalue() zip_data.close() content_hash = _calculate_hash(files, root) return contents, content_hash
Generates a ZIP file in-memory from a list of files. Files will be stored in the archive with relative names, and have their UNIX permissions forced to 755 or 644 (depending on whether they are user-executable in the source filesystem). Args: files (list[str]): file names to add to the archive, relative to ``root``. root (str): base directory to retrieve files from. Returns: str: content of the ZIP file as a byte string. str: A calculated hash of all the files.
def key_binding(self, keydef, mode='force'): """Function decorator to register a low-level key binding. The callback function signature is ``fun(key_state, key_name)`` where ``key_state`` is either ``'U'`` for "key up" or ``'D'`` for "key down". The keydef format is: ``[Shift+][Ctrl+][Alt+][Meta+]<key>`` where ``<key>`` is either the literal character the key produces (ASCII or Unicode character), or a symbolic name (as printed by ``mpv --input-keylist``). To unregister the callback function, you can call its ``unregister_mpv_key_bindings`` attribute:: player = mpv.MPV() @player.key_binding('Q') def binding(state, name): print('blep') binding.unregister_mpv_key_bindings() WARNING: For a single keydef only a single callback/command can be registered at the same time. If you register a binding multiple times older bindings will be overwritten and there is a possibility of references leaking. So don't do that. BIG FAT WARNING: mpv's key binding mechanism is pretty powerful. This means, you essentially get arbitrary code exectution through key bindings. This interface makes some limited effort to sanitize the keydef given in the first parameter, but YOU SHOULD NOT RELY ON THIS IN FOR SECURITY. If your input comes from config files, this is completely fine--but, if you are about to pass untrusted input into this parameter, better double-check whether this is secure in your case. """ def register(fun): fun.mpv_key_bindings = getattr(fun, 'mpv_key_bindings', []) + [keydef] def unregister_all(): for keydef in fun.mpv_key_bindings: self.unregister_key_binding(keydef) fun.unregister_mpv_key_bindings = unregister_all self.register_key_binding(keydef, fun, mode) return fun return register
Function decorator to register a low-level key binding. The callback function signature is ``fun(key_state, key_name)`` where ``key_state`` is either ``'U'`` for "key up" or ``'D'`` for "key down". The keydef format is: ``[Shift+][Ctrl+][Alt+][Meta+]<key>`` where ``<key>`` is either the literal character the key produces (ASCII or Unicode character), or a symbolic name (as printed by ``mpv --input-keylist``). To unregister the callback function, you can call its ``unregister_mpv_key_bindings`` attribute:: player = mpv.MPV() @player.key_binding('Q') def binding(state, name): print('blep') binding.unregister_mpv_key_bindings() WARNING: For a single keydef only a single callback/command can be registered at the same time. If you register a binding multiple times older bindings will be overwritten and there is a possibility of references leaking. So don't do that. BIG FAT WARNING: mpv's key binding mechanism is pretty powerful. This means, you essentially get arbitrary code exectution through key bindings. This interface makes some limited effort to sanitize the keydef given in the first parameter, but YOU SHOULD NOT RELY ON THIS IN FOR SECURITY. If your input comes from config files, this is completely fine--but, if you are about to pass untrusted input into this parameter, better double-check whether this is secure in your case.
def send(self, message, *args, **kwargs): ''' Sends provided message to all listeners. Message is only added to queue and will be processed on next tick. :param Message message: Message to send. ''' self._messages.put((message, args, kwargs), False)
Sends provided message to all listeners. Message is only added to queue and will be processed on next tick. :param Message message: Message to send.
def get_resource_url(self): """ Get resource complete url """ name = self.__class__.resource_name url = self.__class__.rest_base_url() return "%s/%s" % (url, name)
Get resource complete url
def start(self, listen_ip=LISTEN_IP, listen_port=0): """Start discovery task.""" coro = self.loop.create_datagram_endpoint( lambda: self, local_addr=(listen_ip, listen_port)) self.task = self.loop.create_task(coro) return self.task
Start discovery task.
def argmin(self): """ Return the co-ordinates of the bin centre containing the minimum value. Same as numpy.argmin(), converting the indexes to bin co-ordinates. """ return tuple(centres[index] for centres, index in zip(self.centres(), numpy.unravel_index(self.array.argmin(), self.array.shape)))
Return the co-ordinates of the bin centre containing the minimum value. Same as numpy.argmin(), converting the indexes to bin co-ordinates.
def __fetch_crate_owner_user(self, crate_id): """Get crate user owners""" raw_owner_user = self.client.crate_attribute(crate_id, 'owner_user') owner_user = json.loads(raw_owner_user) return owner_user
Get crate user owners
def create_from_string(self, string, context=EMPTY_CONTEXT, *args, **kwargs): """ Deserializes a new instance from a string. This is a convenience method that creates a StringIO object and calls create_instance_from_stream(). """ if not PY2 and not isinstance(string, bytes): raise TypeError("string should be an instance of bytes in Python 3") io = StringIO(string) instance = self.create_from_stream(io, context, *args, **kwargs) io.close() return instance
Deserializes a new instance from a string. This is a convenience method that creates a StringIO object and calls create_instance_from_stream().
def execute_cql_query(self, query, compression): """ Executes a CQL (Cassandra Query Language) statement and returns a CqlResult containing the results. Parameters: - query - compression """ self._seqid += 1 d = self._reqs[self._seqid] = defer.Deferred() self.send_execute_cql_query(query, compression) return d
Executes a CQL (Cassandra Query Language) statement and returns a CqlResult containing the results. Parameters: - query - compression
def _DeserializeAttributeContainer(self, container_type, serialized_data): """Deserializes an attribute container. Args: container_type (str): attribute container type. serialized_data (bytes): serialized attribute container data. Returns: AttributeContainer: attribute container or None. Raises: IOError: if the serialized data cannot be decoded. OSError: if the serialized data cannot be decoded. """ if not serialized_data: return None if self._serializers_profiler: self._serializers_profiler.StartTiming(container_type) try: serialized_string = serialized_data.decode('utf-8') except UnicodeDecodeError as exception: raise IOError('Unable to decode serialized data: {0!s}'.format( exception)) attribute_container = self._serializer.ReadSerialized(serialized_string) if self._serializers_profiler: self._serializers_profiler.StopTiming(container_type) return attribute_container
Deserializes an attribute container. Args: container_type (str): attribute container type. serialized_data (bytes): serialized attribute container data. Returns: AttributeContainer: attribute container or None. Raises: IOError: if the serialized data cannot be decoded. OSError: if the serialized data cannot be decoded.
def label_from_lists(self, train_labels:Iterator, valid_labels:Iterator, label_cls:Callable=None, **kwargs)->'LabelList': "Use the labels in `train_labels` and `valid_labels` to label the data. `label_cls` will overwrite the default." label_cls = self.train.get_label_cls(train_labels, label_cls) self.train = self.train._label_list(x=self.train, y=label_cls(train_labels, **kwargs)) self.valid = self.valid._label_list(x=self.valid, y=self.train.y.new(valid_labels, **kwargs)) self.__class__ = LabelLists self.process() return self
Use the labels in `train_labels` and `valid_labels` to label the data. `label_cls` will overwrite the default.
def stack(recs, fields=None): """Stack common fields in multiple record arrays (concatenate them). Parameters ---------- recs : list List of NumPy record arrays fields : list of strings, optional (default=None) The list of fields to include in the stacked array. If None, then include the fields in common to all the record arrays. Returns ------- rec : NumPy record array The stacked array. """ if fields is None: fields = list(set.intersection( *[set(rec.dtype.names) for rec in recs])) # preserve order of fields wrt first record array if set(fields) == set(recs[0].dtype.names): fields = list(recs[0].dtype.names) return np.hstack([rec[fields] for rec in recs])
Stack common fields in multiple record arrays (concatenate them). Parameters ---------- recs : list List of NumPy record arrays fields : list of strings, optional (default=None) The list of fields to include in the stacked array. If None, then include the fields in common to all the record arrays. Returns ------- rec : NumPy record array The stacked array.
def computeISI(spikeTrains): """ Estimates the inter-spike interval from a spike train matrix. @param spikeTrains (array) matrix of spike trains @return isi (array) matrix with the inter-spike interval obtained from the spike train. Each entry in this matrix represents the number of time-steps in-between 2 spikes as the algorithm scans the spike train matrix. """ zeroCount = 0 isi = [] cells = 0 for i in range(np.shape(spikeTrains)[0]): if cells > 0 and cells % 250 == 0: print str(cells) + " cells processed" for j in range(np.shape(spikeTrains)[1]): if spikeTrains[i][j] == 0: zeroCount += 1 elif zeroCount > 0: isi.append(zeroCount) zeroCount = 0 zeroCount = 0 cells += 1 print "**All cells processed**" return isi
Estimates the inter-spike interval from a spike train matrix. @param spikeTrains (array) matrix of spike trains @return isi (array) matrix with the inter-spike interval obtained from the spike train. Each entry in this matrix represents the number of time-steps in-between 2 spikes as the algorithm scans the spike train matrix.
def button_clicked(self, button): """Action when button was clicked. Parameters ---------- button : instance of QPushButton which button was pressed """ if button is self.idx_ok: chans = self.get_channels() group = self.one_grp cycle = self.get_cycles() stage = self.idx_stage.selectedItems() params = {k: v.get_value() for k, v in self.index.items()} name = self.name.get_value() if None in [params['f1'], params['f2']]: self.parent.statusBar().showMessage( 'Specify bandpass frequencies') return if params['max_dur'] is None: self.parent.statusBar().showMessage('Specify maximum duration') return elif params['max_dur'] >= 30: self.parent.statusBar().showMessage( 'Maximum duration must be below 30 seconds.') return if stage == []: stage = None else: stage = [x.text() for x in self.idx_stage.selectedItems()] chan_full = None reject_artf = False if params['excl_event'] == 'channel-specific': chan_full = [i + ' (' + self.idx_group.currentText() + '' ')' for i in chans] chans = None reject_artf = True elif params['excl_event'] == 'from any channel': reject_artf = True data = fetch(self.parent.info.dataset, self.parent.notes.annot, cat=(1, 1, 1, 0), stage=stage, cycle=cycle, chan_full=chan_full, min_dur=params['min_seg_dur'], reject_epoch=params['excl_epoch'], reject_artf=reject_artf) if not data.segments: msg = 'No valid signal found.' error_dialog = QErrorMessage(self) error_dialog.setWindowTitle('Error fetching data') error_dialog.showMessage(msg) return ding = data.read_data(chans, group['ref_chan'], group['name'], parent=self) if not ding: self.parent.statusBar().showMessage('Process interrupted.') return data = data[0]['data'] # ============================================================================= # if params['resample']: # rs_freq = params['rs_freq'] # ratio = rs_freq / data.s_freq # up = Fraction(ratio).numerator # dn = Fraction(ratio).denominator # # rs_dat = zeros((data.number_of('chan')[0], # int(data.number_of('time')[0] * ratio))) # # for i in range(data.number_of('chan')[0]): # rs_dat[i, :] = resample_poly(data.data[0][i], up, dn) # # data.data[0] = rs_dat # data.s_freq = rs_freq # start_time = data.axis['time'][0][0] # end_time = data.axis['time'][0][-1] # data.axis['time'][0] = arange(start_time, end_time, # 1 / rs_freq) # lg.info(str(data.data.shape)) # lg.info(str(data.data[0].shape)) # lg.info(str(data.data[0][0].shape)) # lg.info(str(data.axis['time'].shape)) # lg.info(str(data.s_freq)) # ============================================================================= if params['prep_filt']: low_cut = params['prep_lc'] high_cut = params['prep_hc'] data = filter_(data, axis='time', low_cut=low_cut, high_cut=high_cut) if params['detrend']: data = math(data, operator_name='detrend', axis='time') self.parent.notes.detect_events(data, self.method, params, label=name) self.accept() if button is self.idx_cancel: self.reject()
Action when button was clicked. Parameters ---------- button : instance of QPushButton which button was pressed
def telegram(): '''Install Telegram desktop client for linux (x64). More infos: https://telegram.org https://desktop.telegram.org/ ''' if not exists('~/bin/Telegram', msg='Download and install Telegram:'): run('mkdir -p /tmp/telegram') run('cd /tmp/telegram && wget https://telegram.org/dl/desktop/linux') run('cd /tmp/telegram && tar xf linux') with warn_only(): run('mv /tmp/telegram/Telegram ~/bin') run('rm -rf /tmp/telegram') else: print('skip download, dir ~/bin/Telegram already exists') run('ln -snf ~/bin/Telegram/Telegram ~/bin/telegram', msg="\nCreate executable 'telegram':")
Install Telegram desktop client for linux (x64). More infos: https://telegram.org https://desktop.telegram.org/
def connect_text(instance, prop, widget): """ Connect a string callback property with a Qt widget containing text. Parameters ---------- instance : object The class instance that the callback property is attached to prop : str The name of the callback property widget : QtWidget The Qt widget to connect. This should implement the ``setText`` and ``text`` methods as well optionally the ``editingFinished`` signal. """ def update_prop(): val = widget.text() setattr(instance, prop, val) def update_widget(val): if hasattr(widget, 'editingFinished'): widget.blockSignals(True) widget.setText(val) widget.blockSignals(False) widget.editingFinished.emit() else: widget.setText(val) add_callback(instance, prop, update_widget) try: widget.editingFinished.connect(update_prop) except AttributeError: pass update_widget(getattr(instance, prop))
Connect a string callback property with a Qt widget containing text. Parameters ---------- instance : object The class instance that the callback property is attached to prop : str The name of the callback property widget : QtWidget The Qt widget to connect. This should implement the ``setText`` and ``text`` methods as well optionally the ``editingFinished`` signal.
def do_WhoIsRequest(self, apdu): """Respond to a Who-Is request.""" if _debug: WhoIsIAmServices._debug("do_WhoIsRequest %r", apdu) # ignore this if there's no local device if not self.localDevice: if _debug: WhoIsIAmServices._debug(" - no local device") return # extract the parameters low_limit = apdu.deviceInstanceRangeLowLimit high_limit = apdu.deviceInstanceRangeHighLimit # check for consistent parameters if (low_limit is not None): if (high_limit is None): raise MissingRequiredParameter("deviceInstanceRangeHighLimit required") if (low_limit < 0) or (low_limit > 4194303): raise ParameterOutOfRange("deviceInstanceRangeLowLimit out of range") if (high_limit is not None): if (low_limit is None): raise MissingRequiredParameter("deviceInstanceRangeLowLimit required") if (high_limit < 0) or (high_limit > 4194303): raise ParameterOutOfRange("deviceInstanceRangeHighLimit out of range") # see we should respond if (low_limit is not None): if (self.localDevice.objectIdentifier[1] < low_limit): return if (high_limit is not None): if (self.localDevice.objectIdentifier[1] > high_limit): return # generate an I-Am self.i_am(address=apdu.pduSource)
Respond to a Who-Is request.
def get_admin_url(obj, page=None): """Return the URL to admin pages for this object.""" if obj is None: return None if page is None: page = "change" if page not in ADMIN_ALL_PAGES: raise ValueError("Invalid page name '{}'. Available pages are: {}.".format(page, ADMIN_ALL_PAGES)) app_label = obj.__class__._meta.app_label object_name = obj.__class__._meta.object_name.lower() if page in ADMIN_GLOBAL_PAGES: url_name = page else: url_name = "{}_{}_{}".format(app_label, object_name, page) if page == "app_list": url_args = (app_label,) elif page == "view_on_site": content_type = ContentType.objects.get_for_model(obj.__class__) url_args = (content_type, obj._get_pk_val()) elif page in ADMIN_DETAIL_PAGES: url_args = (obj._get_pk_val(),) else: url_args = None return reverse("admin:{}".format(url_name), args=url_args)
Return the URL to admin pages for this object.
async def start(self, remoteParameters): """ Start DTLS transport negotiation with the parameters of the remote DTLS transport. :param: remoteParameters: An :class:`RTCDtlsParameters`. """ assert self._state == State.NEW assert len(remoteParameters.fingerprints) if self.transport.role == 'controlling': self._role = 'server' lib.SSL_set_accept_state(self.ssl) else: self._role = 'client' lib.SSL_set_connect_state(self.ssl) self._set_state(State.CONNECTING) try: while not self.encrypted: result = lib.SSL_do_handshake(self.ssl) await self._write_ssl() if result > 0: self.encrypted = True break error = lib.SSL_get_error(self.ssl, result) if error == lib.SSL_ERROR_WANT_READ: await self._recv_next() else: self.__log_debug('x DTLS handshake failed (error %d)', error) for info in get_error_queue(): self.__log_debug('x %s', ':'.join(info)) self._set_state(State.FAILED) return except ConnectionError: self.__log_debug('x DTLS handshake failed (connection error)') self._set_state(State.FAILED) return # check remote fingerprint x509 = lib.SSL_get_peer_certificate(self.ssl) remote_fingerprint = certificate_digest(x509) fingerprint_is_valid = False for f in remoteParameters.fingerprints: if f.algorithm.lower() == 'sha-256' and f.value.lower() == remote_fingerprint.lower(): fingerprint_is_valid = True break if not fingerprint_is_valid: self.__log_debug('x DTLS handshake failed (fingerprint mismatch)') self._set_state(State.FAILED) return # generate keying material buf = ffi.new('unsigned char[]', 2 * (SRTP_KEY_LEN + SRTP_SALT_LEN)) extractor = b'EXTRACTOR-dtls_srtp' _openssl_assert(lib.SSL_export_keying_material( self.ssl, buf, len(buf), extractor, len(extractor), ffi.NULL, 0, 0) == 1) view = ffi.buffer(buf) if self._role == 'server': srtp_tx_key = get_srtp_key_salt(view, 1) srtp_rx_key = get_srtp_key_salt(view, 0) else: srtp_tx_key = get_srtp_key_salt(view, 0) srtp_rx_key = get_srtp_key_salt(view, 1) rx_policy = Policy(key=srtp_rx_key, ssrc_type=Policy.SSRC_ANY_INBOUND) rx_policy.allow_repeat_tx = True rx_policy.window_size = 1024 self._rx_srtp = Session(rx_policy) tx_policy = Policy(key=srtp_tx_key, ssrc_type=Policy.SSRC_ANY_OUTBOUND) tx_policy.allow_repeat_tx = True tx_policy.window_size = 1024 self._tx_srtp = Session(tx_policy) # start data pump self.__log_debug('- DTLS handshake complete') self._set_state(State.CONNECTED) self._task = asyncio.ensure_future(self.__run())
Start DTLS transport negotiation with the parameters of the remote DTLS transport. :param: remoteParameters: An :class:`RTCDtlsParameters`.
def qteImportModule(self, fileName: str): """ Import ``fileName`` at run-time. If ``fileName`` has no path prefix then it must be in the standard Python module path. Relative path names are possible. |Args| * ``fileName`` (**str**): file name (with full path) of module to import. |Returns| * **module**: the imported Python module, or **None** if an error occurred. |Raises| * **None** """ # Split the absolute file name into the path- and file name. path, name = os.path.split(fileName) name, ext = os.path.splitext(name) # If the file name has a path prefix then search there, otherwise # search the default paths for Python. if path == '': path = sys.path else: path = [path] # Try to locate the module. try: fp, pathname, desc = imp.find_module(name, path) except ImportError: msg = 'Could not find module <b>{}</b>.'.format(fileName) self.qteLogger.error(msg) return None # Try to import the module. try: mod = imp.load_module(name, fp, pathname, desc) return mod except ImportError: msg = 'Could not import module <b>{}</b>.'.format(fileName) self.qteLogger.error(msg) return None finally: # According to the imp documentation the file pointer # should always be closed explicitly. if fp: fp.close()
Import ``fileName`` at run-time. If ``fileName`` has no path prefix then it must be in the standard Python module path. Relative path names are possible. |Args| * ``fileName`` (**str**): file name (with full path) of module to import. |Returns| * **module**: the imported Python module, or **None** if an error occurred. |Raises| * **None**
def unindent(self): """ Unindents the document text under cursor. :return: Method success. :rtype: bool """ cursor = self.textCursor() if not cursor.hasSelection(): cursor.movePosition(QTextCursor.StartOfBlock) line = foundations.strings.to_string(self.document().findBlockByNumber(cursor.blockNumber()).text()) indent_marker = re.match(r"({0})".format(self.__indent_marker), line) if indent_marker: foundations.common.repeat(cursor.deleteChar, len(indent_marker.group(1))) else: block = self.document().findBlock(cursor.selectionStart()) while True: block_cursor = self.textCursor() block_cursor.setPosition(block.position()) indent_marker = re.match(r"({0})".format(self.__indent_marker), block.text()) if indent_marker: foundations.common.repeat(block_cursor.deleteChar, len(indent_marker.group(1))) if block.contains(cursor.selectionEnd()): break block = block.next() return True
Unindents the document text under cursor. :return: Method success. :rtype: bool
async def request_resource(self, type: Type[T_Resource], name: str = 'default') -> T_Resource: """ Look up a resource in the chain of contexts. This is like :meth:`get_resource` except that if the resource is not already available, it will wait for one to become available. :param type: type of the requested resource :param name: name of the requested resource :return: the requested resource """ # First try to locate an existing resource in this context and its parents value = self.get_resource(type, name) if value is not None: return value # Wait until a matching resource or resource factory is available signals = [ctx.resource_added for ctx in self.context_chain] await wait_event( signals, lambda event: event.resource_name == name and type in event.resource_types) return self.require_resource(type, name)
Look up a resource in the chain of contexts. This is like :meth:`get_resource` except that if the resource is not already available, it will wait for one to become available. :param type: type of the requested resource :param name: name of the requested resource :return: the requested resource
def _run_module_as_main(mod_name, alter_argv=True): """Runs the designated module in the __main__ namespace Note that the executed module will have full access to the __main__ namespace. If this is not desirable, the run_module() function should be used to run the module code in a fresh namespace. At the very least, these variables in __main__ will be overwritten: __name__ __file__ __loader__ __package__ """ try: if alter_argv or mod_name != "__main__": # i.e. -m switch mod_name, loader, code, fname = _get_module_details(mod_name) else: # i.e. directory or zipfile execution mod_name, loader, code, fname = _get_main_module_details() except ImportError as exc: msg = "%s: %s" % (sys.executable, str(exc)) sys.exit(msg) pkg_name = mod_name.rpartition('.')[0] main_globals = sys.modules["__main__"].__dict__ if alter_argv: sys.argv[0] = fname return _run_code(code, main_globals, None, "__main__", fname, loader, pkg_name)
Runs the designated module in the __main__ namespace Note that the executed module will have full access to the __main__ namespace. If this is not desirable, the run_module() function should be used to run the module code in a fresh namespace. At the very least, these variables in __main__ will be overwritten: __name__ __file__ __loader__ __package__
def delete(path, dryrun=False, recursive=True, verbose=None, print_exists=True, ignore_errors=True): """ Removes a file, directory, or symlink """ if verbose is None: verbose = VERBOSE if not QUIET: verbose = 1 if verbose > 0: print('[util_path] Deleting path=%r' % path) exists_flag = exists(path) link_flag = islink(path) if not exists_flag and not link_flag: if print_exists and verbose: print('..does not exist!') flag = False else: rmargs = dict(verbose=verbose > 1, ignore_errors=ignore_errors, dryrun=dryrun) if islink(path): os.unlink(path) flag = True elif isdir(path): # First remove everything in the directory flag = remove_files_in_dir(path, recursive=recursive, **rmargs) # Then remove the directory itself flag = flag and remove_dirs(path, **rmargs) elif isfile(path): flag = remove_file(path, **rmargs) else: raise ValueError('Unknown type of path=%r' % (path,)) if verbose > 0: print('[util_path] Finished deleting path=%r' % path) return flag
Removes a file, directory, or symlink
def retrieveVals(self): """Retrieve values for graphs.""" ntpinfo = NTPinfo() stats = ntpinfo.getPeerStats() if stats: if self.hasGraph('ntp_peer_stratum'): self.setGraphVal('ntp_peer_stratum', 'stratum', stats.get('stratum')) if self.hasGraph('ntp_peer_stats'): self.setGraphVal('ntp_peer_stats', 'offset', stats.get('offset')) self.setGraphVal('ntp_peer_stats', 'delay', stats.get('delay')) self.setGraphVal('ntp_peer_stats', 'jitter', stats.get('jitter'))
Retrieve values for graphs.
def Serialize(self, val, info): """ Serialize an object """ self._Serialize(val, info, self.defaultNS)
Serialize an object
def is_compatible(self): """Check if package name is matched by compatible_patterns""" for pattern in OPTIONS['compatible_patterns']: if fnmatch(self.package.lower(), pattern): return True return False
Check if package name is matched by compatible_patterns
def SetAuth(self, style, user=None, password=None): '''Change auth style, return object to user. ''' self.auth_style, self.auth_user, self.auth_pass = \ style, user, password return self
Change auth style, return object to user.
def get_storage_conn(storage_account=None, storage_key=None, opts=None): ''' .. versionadded:: 2015.8.0 Return a storage_conn object for the storage account ''' if opts is None: opts = {} if not storage_account: storage_account = opts.get('storage_account', None) if not storage_key: storage_key = opts.get('storage_key', None) return azure.storage.BlobService(storage_account, storage_key)
.. versionadded:: 2015.8.0 Return a storage_conn object for the storage account
def absdir(path): """Return absolute, normalized path to directory, if it exists; None otherwise. """ if not os.path.isabs(path): path = os.path.normpath(os.path.abspath(os.path.join(os.getcwd(), path))) if path is None or not os.path.isdir(path): return None return path
Return absolute, normalized path to directory, if it exists; None otherwise.
def get_update_object(self, form): """ Retrieves the target object based on the update form's ``pk`` and the table's queryset. """ pk = form.cleaned_data['pk'] queryset = self.get_queryset() try: obj = queryset.get(pk=pk) except queryset.model.DoesNotExist: obj = None return obj
Retrieves the target object based on the update form's ``pk`` and the table's queryset.
def cached_property(prop): """ A replacement for the property decorator that will only compute the attribute's value on the first call and serve a cached copy from then on. """ def cache_wrapper(self): if not hasattr(self, "_cache"): self._cache = {} if prop.__name__ not in self._cache: return_value = prop(self) if isgenerator(return_value): return_value = tuple(return_value) self._cache[prop.__name__] = return_value return self._cache[prop.__name__] return property(cache_wrapper)
A replacement for the property decorator that will only compute the attribute's value on the first call and serve a cached copy from then on.
def run_cmd(cmd, out=os.path.devnull, err=os.path.devnull): """Runs an external command :param list cmd: Command to run. :param str out: Output file :param str err: Error file :raises: RuntimeError """ logger.debug(' '.join(cmd)) with open(out, 'w') as hout: proc = subprocess.Popen(cmd, stdout=hout, stderr=subprocess.PIPE) err_msg = proc.communicate()[1].decode() with open(err, 'w') as herr: herr.write(str(err_msg)) msg = '({}) {}'.format(' '.join(cmd), err_msg) if proc.returncode != 0: logger.error(msg) raise RuntimeError(msg)
Runs an external command :param list cmd: Command to run. :param str out: Output file :param str err: Error file :raises: RuntimeError
def dynacRepresentation(self): """ Return the Pynac representation of this Set4DAperture instance. """ details = [ self.energyDefnFlag.val, self.energy.val, self.phase.val, self.x.val, self.y.val, self.radius.val, ] return ['REJECT', [details]]
Return the Pynac representation of this Set4DAperture instance.
def _make_datablock(self): '''Make a data_block and sections list as required by DataWrapper''' section_ids = sorted(self.sections) # create all insertion id's, this needs to be done ahead of time # as some of the children may have a lower id than their parents id_to_insert_id = {} row_count = 0 for section_id in section_ids: row_count += len(self.sections[section_id].points) id_to_insert_id[section_id] = row_count - 1 datablock = np.empty((row_count, COLS.COL_COUNT), dtype=np.float) datablock[:, COLS.ID] = np.arange(len(datablock)) datablock[:, COLS.P] = datablock[:, COLS.ID] - 1 sections = [] insert_index = 0 for id_ in section_ids: sec = self.sections[id_] points, section_type, parent_id = sec.points, sec.section_type, sec.parent_id idx = slice(insert_index, insert_index + len(points)) datablock[idx, COLS.XYZR] = points datablock[idx, COLS.TYPE] = section_type datablock[idx.start, COLS.P] = id_to_insert_id.get(parent_id, ROOT_ID) sections.append(DataBlockSection(idx, section_type, parent_id)) insert_index = idx.stop return datablock, sections
Make a data_block and sections list as required by DataWrapper
async def create_turn_endpoint(protocol_factory, server_addr, username, password, lifetime=600, ssl=False, transport='udp'): """ Create datagram connection relayed over TURN. """ loop = asyncio.get_event_loop() if transport == 'tcp': _, inner_protocol = await loop.create_connection( lambda: TurnClientTcpProtocol(server_addr, username=username, password=password, lifetime=lifetime), host=server_addr[0], port=server_addr[1], ssl=ssl) else: _, inner_protocol = await loop.create_datagram_endpoint( lambda: TurnClientUdpProtocol(server_addr, username=username, password=password, lifetime=lifetime), remote_addr=server_addr) protocol = protocol_factory() transport = TurnTransport(protocol, inner_protocol) await transport._connect() return transport, protocol
Create datagram connection relayed over TURN.
def copyMakeBorder(src, top, bot, left, right, border_type=cv2.BORDER_CONSTANT, value=0): """Pad image border Wrapper for cv2.copyMakeBorder that uses mx.nd.NDArray Parameters ---------- src : NDArray Image in (width, height, channels). Others are the same with cv2.copyMakeBorder Returns ------- img : NDArray padded image """ hdl = NDArrayHandle() check_call(_LIB.MXCVcopyMakeBorder(src.handle, ctypes.c_int(top), ctypes.c_int(bot), ctypes.c_int(left), ctypes.c_int(right), ctypes.c_int(border_type), ctypes.c_double(value), ctypes.byref(hdl))) return mx.nd.NDArray(hdl)
Pad image border Wrapper for cv2.copyMakeBorder that uses mx.nd.NDArray Parameters ---------- src : NDArray Image in (width, height, channels). Others are the same with cv2.copyMakeBorder Returns ------- img : NDArray padded image
def _set_cdn_defaults(self): """Sets all the CDN-related attributes to default values.""" if self._cdn_enabled is FAULT: self._cdn_enabled = False self._cdn_uri = None self._cdn_ttl = DEFAULT_CDN_TTL self._cdn_ssl_uri = None self._cdn_streaming_uri = None self._cdn_ios_uri = None self._cdn_log_retention = False
Sets all the CDN-related attributes to default values.
def get_epithet_index(): """Return dict of epithets (key) to a set of all author ids of that epithet (value). """ _dict = {} for k, v in AUTHOR_EPITHET.items(): _dict[k] = set(v) return _dict
Return dict of epithets (key) to a set of all author ids of that epithet (value).
def map_seqprop_resnums_to_seqprop_resnums(self, resnums, seqprop1, seqprop2): """Map a residue number in any SeqProp to another SeqProp using the pairwise alignment information. Args: resnums (int, list): Residue numbers in seqprop1 seqprop1 (SeqProp): SeqProp object the resnums match to seqprop2 (SeqProp): SeqProp object you want to map the resnums to Returns: dict: Mapping of seqprop1 residue numbers to seqprop2 residue numbers. If mappings don't exist in this dictionary, that means the residue number cannot be mapped according to alignment! """ resnums = ssbio.utils.force_list(resnums) alignment = self._get_seqprop_to_seqprop_alignment(seqprop1=seqprop1, seqprop2=seqprop2) mapped = ssbio.protein.sequence.utils.alignment.map_resnum_a_to_resnum_b(resnums=resnums, a_aln=alignment[0], b_aln=alignment[1]) return mapped
Map a residue number in any SeqProp to another SeqProp using the pairwise alignment information. Args: resnums (int, list): Residue numbers in seqprop1 seqprop1 (SeqProp): SeqProp object the resnums match to seqprop2 (SeqProp): SeqProp object you want to map the resnums to Returns: dict: Mapping of seqprop1 residue numbers to seqprop2 residue numbers. If mappings don't exist in this dictionary, that means the residue number cannot be mapped according to alignment!
def create_java_executor(self, dist=None): """Create java executor that uses this task's ng daemon, if allowed. Call only in execute() or later. TODO: Enforce this. """ dist = dist or self.dist if self.execution_strategy == self.NAILGUN: classpath = os.pathsep.join(self.tool_classpath('nailgun-server')) return NailgunExecutor(self._identity, self._executor_workdir, classpath, dist, startup_timeout=self.get_options().nailgun_subprocess_startup_timeout, connect_timeout=self.get_options().nailgun_timeout_seconds, connect_attempts=self.get_options().nailgun_connect_attempts) else: return SubprocessExecutor(dist)
Create java executor that uses this task's ng daemon, if allowed. Call only in execute() or later. TODO: Enforce this.
def register_text_type(content_type, default_encoding, dumper, loader): """ Register handling for a text-based content type. :param str content_type: content type to register the hooks for :param str default_encoding: encoding to use if none is present in the request :param dumper: called to decode a string into a dictionary. Calling convention: ``dumper(obj_dict).encode(encoding) -> bytes`` :param loader: called to encode a dictionary to a string. Calling convention: ``loader(obj_bytes.decode(encoding)) -> dict`` The decoding of a text content body takes into account decoding the binary request body into a string before calling the underlying dump/load routines. """ content_type = headers.parse_content_type(content_type) content_type.parameters.clear() key = str(content_type) _content_types[key] = content_type handler = _content_handlers.setdefault(key, _ContentHandler(key)) handler.dict_to_string = dumper handler.string_to_dict = loader handler.default_encoding = default_encoding or handler.default_encoding
Register handling for a text-based content type. :param str content_type: content type to register the hooks for :param str default_encoding: encoding to use if none is present in the request :param dumper: called to decode a string into a dictionary. Calling convention: ``dumper(obj_dict).encode(encoding) -> bytes`` :param loader: called to encode a dictionary to a string. Calling convention: ``loader(obj_bytes.decode(encoding)) -> dict`` The decoding of a text content body takes into account decoding the binary request body into a string before calling the underlying dump/load routines.
def MAPGenoToTrans(parsedGTF,feature): """ Gets all positions of all bases in an exon :param df: a Pandas dataframe with 'start','end', and 'strand' information for each entry. df must contain 'seqname','feature','start','end','strand','frame','gene_id', 'transcript_id','exon_id','exon_number'] :param feature: feature upon wich to generate the map, eg. 'exon' or 'transcript' :returns: a string with the comma separated positions of all bases in the exon """ GenTransMap=parsedGTF[parsedGTF["feature"]==feature] def getExonsPositions(df): start=int(df["start"]) stop=int(df["end"]) strand=df["strand"] r=range(start,stop+1) if strand=="-": r.sort(reverse=True) r=[ str(s) for s in r] return ",".join(r) GenTransMap["feature_bases"]=GenTransMap.apply(getExonsPositions, axis=1) GenTransMap=GenTransMap.sort_values(by=["transcript_id","exon_number"],ascending=True) def CombineExons(df): return pd.Series(dict( feature_bases = ','.join(df['feature_bases']) ) ) GenTransMap=GenTransMap.groupby("transcript_id").apply(CombineExons) GenTransMap=GenTransMap.to_dict().get("feature_bases") return GenTransMap
Gets all positions of all bases in an exon :param df: a Pandas dataframe with 'start','end', and 'strand' information for each entry. df must contain 'seqname','feature','start','end','strand','frame','gene_id', 'transcript_id','exon_id','exon_number'] :param feature: feature upon wich to generate the map, eg. 'exon' or 'transcript' :returns: a string with the comma separated positions of all bases in the exon
def count(self, filter=None, session=None, **kwargs): """**DEPRECATED** - Get the number of documents in this collection. The :meth:`count` method is deprecated and **not** supported in a transaction. Please use :meth:`count_documents` or :meth:`estimated_document_count` instead. All optional count parameters should be passed as keyword arguments to this method. Valid options include: - `skip` (int): The number of matching documents to skip before returning results. - `limit` (int): The maximum number of documents to count. A limit of 0 (the default) is equivalent to setting no limit. - `maxTimeMS` (int): The maximum amount of time to allow the count command to run, in milliseconds. - `collation` (optional): An instance of :class:`~pymongo.collation.Collation`. This option is only supported on MongoDB 3.4 and above. - `hint` (string or list of tuples): The index to use. Specify either the index name as a string or the index specification as a list of tuples (e.g. [('a', pymongo.ASCENDING), ('b', pymongo.ASCENDING)]). The :meth:`count` method obeys the :attr:`read_preference` of this :class:`Collection`. .. note:: When migrating from :meth:`count` to :meth:`count_documents` the following query operators must be replaced: +-------------+-------------------------------------+ | Operator | Replacement | +=============+=====================================+ | $where | `$expr`_ | +-------------+-------------------------------------+ | $near | `$geoWithin`_ with `$center`_ | +-------------+-------------------------------------+ | $nearSphere | `$geoWithin`_ with `$centerSphere`_ | +-------------+-------------------------------------+ $expr requires MongoDB 3.6+ :Parameters: - `filter` (optional): A query document that selects which documents to count in the collection. - `session` (optional): a :class:`~pymongo.client_session.ClientSession`. - `**kwargs` (optional): See list of options above. .. versionchanged:: 3.7 Deprecated. .. versionchanged:: 3.6 Added ``session`` parameter. .. versionchanged:: 3.4 Support the `collation` option. .. _$expr: https://docs.mongodb.com/manual/reference/operator/query/expr/ .. _$geoWithin: https://docs.mongodb.com/manual/reference/operator/query/geoWithin/ .. _$center: https://docs.mongodb.com/manual/reference/operator/query/center/#op._S_center .. _$centerSphere: https://docs.mongodb.com/manual/reference/operator/query/centerSphere/#op._S_centerSphere """ warnings.warn("count is deprecated. Use estimated_document_count or " "count_documents instead. Please note that $where must " "be replaced by $expr, $near must be replaced by " "$geoWithin with $center, and $nearSphere must be " "replaced by $geoWithin with $centerSphere", DeprecationWarning, stacklevel=2) cmd = SON([("count", self.__name)]) if filter is not None: if "query" in kwargs: raise ConfigurationError("can't pass both filter and query") kwargs["query"] = filter if "hint" in kwargs and not isinstance(kwargs["hint"], string_type): kwargs["hint"] = helpers._index_document(kwargs["hint"]) collation = validate_collation_or_none(kwargs.pop('collation', None)) cmd.update(kwargs) return self._count(cmd, collation, session)
**DEPRECATED** - Get the number of documents in this collection. The :meth:`count` method is deprecated and **not** supported in a transaction. Please use :meth:`count_documents` or :meth:`estimated_document_count` instead. All optional count parameters should be passed as keyword arguments to this method. Valid options include: - `skip` (int): The number of matching documents to skip before returning results. - `limit` (int): The maximum number of documents to count. A limit of 0 (the default) is equivalent to setting no limit. - `maxTimeMS` (int): The maximum amount of time to allow the count command to run, in milliseconds. - `collation` (optional): An instance of :class:`~pymongo.collation.Collation`. This option is only supported on MongoDB 3.4 and above. - `hint` (string or list of tuples): The index to use. Specify either the index name as a string or the index specification as a list of tuples (e.g. [('a', pymongo.ASCENDING), ('b', pymongo.ASCENDING)]). The :meth:`count` method obeys the :attr:`read_preference` of this :class:`Collection`. .. note:: When migrating from :meth:`count` to :meth:`count_documents` the following query operators must be replaced: +-------------+-------------------------------------+ | Operator | Replacement | +=============+=====================================+ | $where | `$expr`_ | +-------------+-------------------------------------+ | $near | `$geoWithin`_ with `$center`_ | +-------------+-------------------------------------+ | $nearSphere | `$geoWithin`_ with `$centerSphere`_ | +-------------+-------------------------------------+ $expr requires MongoDB 3.6+ :Parameters: - `filter` (optional): A query document that selects which documents to count in the collection. - `session` (optional): a :class:`~pymongo.client_session.ClientSession`. - `**kwargs` (optional): See list of options above. .. versionchanged:: 3.7 Deprecated. .. versionchanged:: 3.6 Added ``session`` parameter. .. versionchanged:: 3.4 Support the `collation` option. .. _$expr: https://docs.mongodb.com/manual/reference/operator/query/expr/ .. _$geoWithin: https://docs.mongodb.com/manual/reference/operator/query/geoWithin/ .. _$center: https://docs.mongodb.com/manual/reference/operator/query/center/#op._S_center .. _$centerSphere: https://docs.mongodb.com/manual/reference/operator/query/centerSphere/#op._S_centerSphere
def solve(self, b_any, b, check_finite=True, p=None): """ solve A \ b """ #assert b.shape[:2]==(len(self.solver),self.dof_any) if self.schur_solver is None and self.A_any_solver is None: assert ( (b is None) or (b.shape[0]==0) ) and ( (b_any is None) or (b_any.shape[0]==0) ), "shape missmatch" return b, b_any elif self.schur_solver is None: assert (b is None) or (b.shape[0]==0), "shape missmatch" solution_any = self.A_any_solver.solve(b=b_any,p=p) return b,solution_any elif self.A_any_solver is None: assert (b_any is None) or (b_any.shape[0]==0), "shape missmatch" solution = self.schur_solver.solve(b=b, check_finite=check_finite) return solution, b_any else: assert p is None, "p is not None" cross_term = np.tensordot(self.DinvC,b_any,axes=([0,1],[0,1])) solution = self.schur_solver.solve(b=(b - cross_term), check_finite=check_finite) solution_any = self.A_any_solver.solve(b=b_any, check_finite=check_finite, p=p) solution_any -= self.DinvC.dot(solution) return solution, solution_any
solve A \ b
def visit(self, node): """Returns a generator that walks all children recursively.""" for child in node: yield child for subchild in self.visit(child): yield subchild
Returns a generator that walks all children recursively.
def setup_size(self, width, height): """Set the width and height for one cell in the tooltip This is inderectly acomplished by setting the iconsizes for the buttons. :param width: the width of one cell, min. is 7 -> icon width = 0 :type width: int :param height: the height of one cell, min. is 6 -> icon height = 0 :type height: int :returns: None :rtype: None :raises: None """ self._iconw = max(0, width - 7) self._iconh = max(0, height - 6) self.update_all_buttons()
Set the width and height for one cell in the tooltip This is inderectly acomplished by setting the iconsizes for the buttons. :param width: the width of one cell, min. is 7 -> icon width = 0 :type width: int :param height: the height of one cell, min. is 6 -> icon height = 0 :type height: int :returns: None :rtype: None :raises: None
def writelines(self, lines, fmt): """ Write `lines` with given `format`. """ if isinstance(fmt, basestring): fmt = [fmt] * len(lines) for f, line in zip(fmt, lines): self.writeline(f, line, self.endian)
Write `lines` with given `format`.
def resolve_variable(provided_variable, blueprint_name): """Resolve a provided variable value against the variable definition. This acts as a subset of resolve_variable logic in the base module, leaving out everything that doesn't apply to CFN parameters. Args: provided_variable (:class:`stacker.variables.Variable`): The variable value provided to the blueprint. blueprint_name (str): The name of the blueprint that the variable is being applied to. Returns: object: The resolved variable string value. Raises: UnresolvedVariable: Raised when the provided variable is not already resolved. """ value = None if provided_variable: if not provided_variable.resolved: raise UnresolvedVariable(blueprint_name, provided_variable) value = provided_variable.value return value
Resolve a provided variable value against the variable definition. This acts as a subset of resolve_variable logic in the base module, leaving out everything that doesn't apply to CFN parameters. Args: provided_variable (:class:`stacker.variables.Variable`): The variable value provided to the blueprint. blueprint_name (str): The name of the blueprint that the variable is being applied to. Returns: object: The resolved variable string value. Raises: UnresolvedVariable: Raised when the provided variable is not already resolved.
def xpathCompareValues(self, inf, strict): """Implement the compare operation on XPath objects: @arg1 < @arg2 (1, 1, ... @arg1 <= @arg2 (1, 0, ... @arg1 > @arg2 (0, 1, ... @arg1 >= @arg2 (0, 0, ... When neither object to be compared is a node-set and the operator is <=, <, >=, >, then the objects are compared by converted both objects to numbers and comparing the numbers according to IEEE 754. The < comparison will be true if and only if the first number is less than the second number. The <= comparison will be true if and only if the first number is less than or equal to the second number. The > comparison will be true if and only if the first number is greater than the second number. The >= comparison will be true if and only if the first number is greater than or equal to the second number. """ ret = libxml2mod.xmlXPathCompareValues(self._o, inf, strict) return ret
Implement the compare operation on XPath objects: @arg1 < @arg2 (1, 1, ... @arg1 <= @arg2 (1, 0, ... @arg1 > @arg2 (0, 1, ... @arg1 >= @arg2 (0, 0, ... When neither object to be compared is a node-set and the operator is <=, <, >=, >, then the objects are compared by converted both objects to numbers and comparing the numbers according to IEEE 754. The < comparison will be true if and only if the first number is less than the second number. The <= comparison will be true if and only if the first number is less than or equal to the second number. The > comparison will be true if and only if the first number is greater than the second number. The >= comparison will be true if and only if the first number is greater than or equal to the second number.
def get_runtime_vars(varset, experiment, token): '''get_runtime_vars will return the urlparsed string of one or more runtime variables. If None are present, None is returned. Parameters ========== varset: the variable set, a dictionary lookup with exp_id, token, vars experiment: the exp_id to look up token: the participant id (or token) that must be defined. Returns ======= url: the variable portion of the url to be passed to experiment, e.g, '?words=at the thing&color=red&globalname=globalvalue' ''' url = '' if experiment in varset: variables = dict() # Participant set variables if token in varset[experiment]: for k,v in varset[experiment][token].items(): variables[k] = v # Global set variables if "*" in varset[experiment]: for k,v in varset[experiment]['*'].items(): # Only add the variable if not already defined if k not in variables: variables[k] = v # Join together, the first ? is added by calling function varlist = ["%s=%s" %(k,v) for k,v in variables.items()] url = '&'.join(varlist) bot.debug('Parsed url: %s' %url) return url
get_runtime_vars will return the urlparsed string of one or more runtime variables. If None are present, None is returned. Parameters ========== varset: the variable set, a dictionary lookup with exp_id, token, vars experiment: the exp_id to look up token: the participant id (or token) that must be defined. Returns ======= url: the variable portion of the url to be passed to experiment, e.g, '?words=at the thing&color=red&globalname=globalvalue'
def password_reset_email_handler(notification): """Password reset email handler.""" base_subject = _('{domain} password reset').format(domain=notification.site.domain) subject = getattr(settings, 'DUM_PASSWORD_RESET_SUBJECT', base_subject) notification.email_subject = subject email_handler(notification, password_reset_email_context)
Password reset email handler.
def _make_result(cls, values, now, timezone): """ Makes a date or datetime or time object from a map of component values :param values: the component values :param now: the current now :param timezone: the current timezone :return: the date, datetime, time or none if values are invalid """ date = None time = None if Component.MONTH in values: year = cls._year_from_2digits(values.get(Component.YEAR, now.year), now.year) month = values[Component.MONTH] day = values.get(Component.DAY, 1) try: date = datetime.date(year, month, day) except ValueError: return None # not a valid date if (Component.HOUR in values and Component.MINUTE in values) or Component.HOUR_AND_MINUTE in values: if Component.HOUR_AND_MINUTE in values: combined = values[Component.HOUR_AND_MINUTE] hour = combined // 100 minute = combined - (hour * 100) second = 0 nano = 0 else: hour = values[Component.HOUR] minute = values[Component.MINUTE] second = values.get(Component.SECOND, 0) nano = values.get(Component.NANO, 0) if hour < 12 and values.get(Component.AM_PM) == cls.PM: hour += 12 elif hour == 12 and values.get(Component.AM_PM) == cls.AM: hour -= 12 try: time = datetime.time(hour, minute, second, microsecond=nano // 1000) except ValueError: return None # not a valid time if Component.OFFSET in values: timezone = pytz.FixedOffset(values[Component.OFFSET] // 60) if date is not None and time is not None: return timezone.localize(datetime.datetime.combine(date, time)) elif date is not None: return date elif time is not None: return time else: return None
Makes a date or datetime or time object from a map of component values :param values: the component values :param now: the current now :param timezone: the current timezone :return: the date, datetime, time or none if values are invalid
def parse_string(xml): """ Returns a slash-formatted string from the given XML representation. The return value is a TokenString (for MBSP) or TaggedString (for Pattern). """ string = "" # Traverse all the <sentence> elements in the XML. dom = XML(xml) for sentence in dom(XML_SENTENCE): _anchors.clear() # Populated by calling _parse_tokens(). _attachments.clear() # Populated by calling _parse_tokens(). # Parse the language from <sentence language="">. language = sentence.get(XML_LANGUAGE, "en") # Parse the token tag format from <sentence token="">. # This information is returned in TokenString.tags, # so the format and order of the token tags is retained when exporting/importing as XML. format = sentence.get(XML_TOKEN, [WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]) format = not isinstance(format, basestring) and format or format.replace(" ","").split(",") # Traverse all <chunk> and <chink> elements in the sentence. # Find the <word> elements inside and create tokens. tokens = [] for chunk in sentence: tokens.extend(_parse_tokens(chunk, format)) # Attach PNP's to their anchors. # Keys in _anchors have linked anchor chunks (each chunk is a list of tokens). # The keys correspond to the keys in _attachments, which have linked PNP chunks. if ANCHOR in format: A, P, a, i = _anchors, _attachments, 1, format.index(ANCHOR) for id in sorted(A.keys()): for token in A[id]: token[i] += "-"+"-".join(["A"+str(a+p) for p in range(len(P[id]))]) token[i] = token[i].strip("O-") for p, pnp in enumerate(P[id]): for token in pnp: token[i] += "-"+"P"+str(a+p) token[i] = token[i].strip("O-") a += len(P[id]) # Collapse the tokens to string. # Separate multiple sentences with a new line. tokens = ["/".join([tag for tag in token]) for token in tokens] tokens = " ".join(tokens) string += tokens + "\n" # Return a TokenString, which is a unicode string that transforms easily # into a plain str, a list of tokens, or a Sentence. try: if MBSP: from mbsp import TokenString return TokenString(string.strip(), tags=format, language=language) except: return TaggedString(string.strip(), tags=format, language=language)
Returns a slash-formatted string from the given XML representation. The return value is a TokenString (for MBSP) or TaggedString (for Pattern).
def activations(self): """Iterate over the Activations in the Agenda.""" activation = lib.EnvGetNextActivation(self._env, ffi.NULL) while activation != ffi.NULL: yield Activation(self._env, activation) activation = lib.EnvGetNextActivation(self._env, activation)
Iterate over the Activations in the Agenda.
def set_own_module(self, path): """ This is provided so the calling process can arrange for processing to be stopped and a LegionReset exception raised when any part of the program's own module tree changes. """ log = self._params.get('log', self._discard) self._name = path self.module_add(event_target(self, 'legion_reset', key=path, log=log), path)
This is provided so the calling process can arrange for processing to be stopped and a LegionReset exception raised when any part of the program's own module tree changes.
def show(self): """ Display the visualizer. """ self.iren.Initialize() self.ren_win.SetSize(800, 800) self.ren_win.SetWindowName(self.title) self.ren_win.Render() self.iren.Start()
Display the visualizer.