_id
stringlengths
2
7
title
stringlengths
1
88
partition
stringclasses
3 values
text
stringlengths
75
19.8k
language
stringclasses
1 value
meta_information
dict
q262200
calculate_concordance_probability_by_annotation
validation
def calculate_concordance_probability_by_annotation(graph, annotation, key, cutoff=None, permutations=None, percentage=None, use_ambiguous=False): """Returns the results of concordance analysis on each subgraph, stratified by the given annotation. :param pybel.BELGraph graph: A BEL graph :param str annotation: The annotation to group by. :param str key: The node data dictionary key storing the logFC :param float cutoff: The optional logFC cutoff for significance :param int permutations: The number of random permutations to test. Defaults to 500 :param float percentage: The percentage of the graph's edges to maintain. Defaults to 0.9 :param bool use_ambiguous: Compare to ambiguous edges as well :rtype: dict[str,tuple] """ result = [ (value, calculate_concordance_probability( subgraph, key, cutoff=cutoff, permutations=permutations, percentage=percentage, use_ambiguous=use_ambiguous, )) for value, subgraph in get_subgraphs_by_annotation(graph, annotation).items() ] return dict(result)
python
{ "resource": "" }
q262201
_get_drug_target_interactions
validation
def _get_drug_target_interactions(manager: Optional['bio2bel_drugbank.manager'] = None) -> Mapping[str, List[str]]: """Get a mapping from drugs to their list of gene.""" if manager is None: import bio2bel_drugbank manager = bio2bel_drugbank.Manager() if not manager.is_populated(): manager.populate() return manager.get_drug_to_hgnc_symbols()
python
{ "resource": "" }
q262202
multi_run_epicom
validation
def multi_run_epicom(graphs: Iterable[BELGraph], path: Union[None, str, TextIO]) -> None: """Run EpiCom analysis on many graphs.""" if isinstance(path, str): with open(path, 'w') as file: _multi_run_helper_file_wrapper(graphs, file) else: _multi_run_helper_file_wrapper(graphs, path)
python
{ "resource": "" }
q262203
main
validation
def main(): """Convert the Alzheimer's and Parkinson's disease NeuroMMSig excel sheets to BEL.""" logging.basicConfig(level=logging.INFO) log.setLevel(logging.INFO) bms_base = get_bms_base() neurommsig_base = get_neurommsig_base() neurommsig_excel_dir = os.path.join(neurommsig_base, 'resources', 'excels', 'neurommsig') nift_values = get_nift_values() log.info('Starting Alzheimers') ad_path = os.path.join(neurommsig_excel_dir, 'alzheimers', 'alzheimers.xlsx') ad_df = preprocess(ad_path) with open(os.path.join(bms_base, 'aetionomy', 'alzheimers', 'neurommsigdb_ad.bel'), 'w') as ad_file: write_neurommsig_bel(ad_file, ad_df, mesh_alzheimer, nift_values) log.info('Starting Parkinsons') pd_path = os.path.join(neurommsig_excel_dir, 'parkinsons', 'parkinsons.xlsx') pd_df = preprocess(pd_path) with open(os.path.join(bms_base, 'aetionomy', 'parkinsons', 'neurommsigdb_pd.bel'), 'w') as pd_file: write_neurommsig_bel(pd_file, pd_df, mesh_parkinson, nift_values)
python
{ "resource": "" }
q262204
remove_inconsistent_edges
validation
def remove_inconsistent_edges(graph: BELGraph) -> None: """Remove all edges between node pairs with inconsistent edges. This is the all-or-nothing approach. It would be better to do more careful investigation of the evidences during curation. """ for u, v in get_inconsistent_edges(graph): edges = [(u, v, k) for k in graph[u][v]] graph.remove_edges_from(edges)
python
{ "resource": "" }
q262205
get_walks_exhaustive
validation
def get_walks_exhaustive(graph, node, length): """Gets all walks under a given length starting at a given node :param networkx.Graph graph: A graph :param node: Starting node :param int length: The length of walks to get :return: A list of paths :rtype: list[tuple] """ if 0 == length: return (node,), return tuple( (node, key) + path for neighbor in graph.edge[node] for path in get_walks_exhaustive(graph, neighbor, length - 1) if node not in path for key in graph.edge[node][neighbor] )
python
{ "resource": "" }
q262206
match_simple_metapath
validation
def match_simple_metapath(graph, node, simple_metapath): """Matches a simple metapath starting at the given node :param pybel.BELGraph graph: A BEL graph :param tuple node: A BEL node :param list[str] simple_metapath: A list of BEL Functions :return: An iterable over paths from the node matching the metapath :rtype: iter[tuple] """ if 0 == len(simple_metapath): yield node, else: for neighbor in graph.edges[node]: if graph.nodes[neighbor][FUNCTION] == simple_metapath[0]: for path in match_simple_metapath(graph, neighbor, simple_metapath[1:]): if node not in path: yield (node,) + path
python
{ "resource": "" }
q262207
build_database
validation
def build_database(manager: pybel.Manager, annotation_url: Optional[str] = None) -> None: """Build a database of scores for NeuroMMSig annotated graphs. 1. Get all networks that use the Subgraph annotation 2. run on each """ annotation_url = annotation_url or NEUROMMSIG_DEFAULT_URL annotation = manager.get_namespace_by_url(annotation_url) if annotation is None: raise RuntimeError('no graphs in database with given annotation') networks = get_networks_using_annotation(manager, annotation) dtis = ... for network in networks: graph = network.as_bel() scores = epicom_on_graph(graph, dtis) for (drug_name, subgraph_name), score in scores.items(): drug_model = get_drug_model(manager, drug_name) subgraph_model = manager.get_annotation_entry(annotation_url, subgraph_name) score_model = Score( network=network, annotation=subgraph_model, drug=drug_model, score=score ) manager.session.add(score_model) t = time.time() logger.info('committing scores') manager.session.commit() logger.info('committed scores in %.2f seconds', time.time() - t)
python
{ "resource": "" }
q262208
calculate_average_scores_on_graph
validation
def calculate_average_scores_on_graph( graph: BELGraph, key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, use_tqdm: bool = False, ): """Calculate the scores over all biological processes in the sub-graph. As an implementation, it simply computes the sub-graphs then calls :func:`calculate_average_scores_on_subgraphs` as described in that function's documentation. :param graph: A BEL graph with heats already on the nodes :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param use_tqdm: Should there be a progress bar for runners? :return: A dictionary of {pybel node tuple: results tuple} :rtype: dict[tuple, tuple] Suggested usage with :mod:`pandas`: >>> import pandas as pd >>> from pybel_tools.analysis.heat import calculate_average_scores_on_graph >>> graph = ... # load graph and data >>> scores = calculate_average_scores_on_graph(graph) >>> pd.DataFrame.from_items(scores.items(), orient='index', columns=RESULT_LABELS) """ subgraphs = generate_bioprocess_mechanisms(graph, key=key) scores = calculate_average_scores_on_subgraphs( subgraphs, key=key, tag=tag, default_score=default_score, runs=runs, use_tqdm=use_tqdm ) return scores
python
{ "resource": "" }
q262209
calculate_average_scores_on_subgraphs
validation
def calculate_average_scores_on_subgraphs( subgraphs: Mapping[H, BELGraph], key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, use_tqdm: bool = False, tqdm_kwargs: Optional[Mapping[str, Any]] = None, ) -> Mapping[H, Tuple[float, float, float, float, int, int]]: """Calculate the scores over precomputed candidate mechanisms. :param subgraphs: A dictionary of keys to their corresponding subgraphs :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param use_tqdm: Should there be a progress bar for runners? :return: A dictionary of keys to results tuples Example Usage: >>> import pandas as pd >>> from pybel_tools.generation import generate_bioprocess_mechanisms >>> from pybel_tools.analysis.heat import calculate_average_scores_on_subgraphs >>> # load graph and data >>> graph = ... >>> candidate_mechanisms = generate_bioprocess_mechanisms(graph) >>> scores = calculate_average_scores_on_subgraphs(candidate_mechanisms) >>> pd.DataFrame.from_items(scores.items(), orient='index', columns=RESULT_LABELS) """ results = {} log.info('calculating results for %d candidate mechanisms using %d permutations', len(subgraphs), runs) it = subgraphs.items() if use_tqdm: _tqdm_kwargs = dict(total=len(subgraphs), desc='Candidate mechanisms') if tqdm_kwargs: _tqdm_kwargs.update(tqdm_kwargs) it = tqdm(it, **_tqdm_kwargs) for node, subgraph in it: number_first_neighbors = subgraph.in_degree(node) number_first_neighbors = 0 if isinstance(number_first_neighbors, dict) else number_first_neighbors mechanism_size = subgraph.number_of_nodes() runners = workflow(subgraph, node, key=key, tag=tag, default_score=default_score, runs=runs) scores = [runner.get_final_score() for runner in runners] if 0 == len(scores): results[node] = ( None, None, None, None, number_first_neighbors, mechanism_size, ) continue scores = np.array(scores) average_score = np.average(scores) score_std = np.std(scores) med_score = np.median(scores) chi_2_stat, norm_p = stats.normaltest(scores) results[node] = ( average_score, score_std, norm_p, med_score, number_first_neighbors, mechanism_size, ) return results
python
{ "resource": "" }
q262210
workflow
validation
def workflow( graph: BELGraph, node: BaseEntity, key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, minimum_nodes: int = 1, ) -> List['Runner']: """Generate candidate mechanisms and run the heat diffusion workflow. :param graph: A BEL graph :param node: The BEL node that is the focus of this analysis :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param minimum_nodes: The minimum number of nodes a sub-graph needs to try running heat diffusion :return: A list of runners """ subgraph = generate_mechanism(graph, node, key=key) if subgraph.number_of_nodes() <= minimum_nodes: return [] runners = multirun(subgraph, node, key=key, tag=tag, default_score=default_score, runs=runs) return list(runners)
python
{ "resource": "" }
q262211
workflow_aggregate
validation
def workflow_aggregate(graph: BELGraph, node: BaseEntity, key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, aggregator: Optional[Callable[[Iterable[float]], float]] = None, ) -> Optional[float]: """Get the average score over multiple runs. This function is very simple, and can be copied to do more interesting statistics over the :class:`Runner` instances. To iterate over the runners themselves, see :func:`workflow` :param graph: A BEL graph :param node: The BEL node that is the focus of this analysis :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param aggregator: A function that aggregates a list of scores. Defaults to :func:`numpy.average`. Could also use: :func:`numpy.mean`, :func:`numpy.median`, :func:`numpy.min`, :func:`numpy.max` :return: The average score for the target node """ runners = workflow(graph, node, key=key, tag=tag, default_score=default_score, runs=runs) scores = [runner.get_final_score() for runner in runners] if not scores: log.warning('Unable to run the heat diffusion workflow for %s', node) return if aggregator is None: return np.average(scores) return aggregator(scores)
python
{ "resource": "" }
q262212
workflow_all
validation
def workflow_all(graph: BELGraph, key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, ) -> Mapping[BaseEntity, List[Runner]]: """Run the heat diffusion workflow and get runners for every possible candidate mechanism 1. Get all biological processes 2. Get candidate mechanism induced two level back from each biological process 3. Heat diffusion workflow for each candidate mechanism for multiple runs 4. Return all runner results :param graph: A BEL graph :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :return: A dictionary of {node: list of runners} """ results = {} for node in get_nodes_by_function(graph, BIOPROCESS): results[node] = workflow(graph, node, key=key, tag=tag, default_score=default_score, runs=runs) return results
python
{ "resource": "" }
q262213
workflow_all_aggregate
validation
def workflow_all_aggregate(graph: BELGraph, key: Optional[str] = None, tag: Optional[str] = None, default_score: Optional[float] = None, runs: Optional[int] = None, aggregator: Optional[Callable[[Iterable[float]], float]] = None, ): """Run the heat diffusion workflow to get average score for every possible candidate mechanism. 1. Get all biological processes 2. Get candidate mechanism induced two level back from each biological process 3. Heat diffusion workflow on each candidate mechanism for multiple runs 4. Report average scores for each candidate mechanism :param graph: A BEL graph :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param tag: The key for the nodes' data dictionaries where the scores will be put. Defaults to 'score' :param default_score: The initial score for all nodes. This number can go up or down. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param aggregator: A function that aggregates a list of scores. Defaults to :func:`numpy.average`. Could also use: :func:`numpy.mean`, :func:`numpy.median`, :func:`numpy.min`, :func:`numpy.max` :return: A dictionary of {node: upstream causal subgraph} """ results = {} bioprocess_nodes = list(get_nodes_by_function(graph, BIOPROCESS)) for bioprocess_node in tqdm(bioprocess_nodes): subgraph = generate_mechanism(graph, bioprocess_node, key=key) try: results[bioprocess_node] = workflow_aggregate( graph=subgraph, node=bioprocess_node, key=key, tag=tag, default_score=default_score, runs=runs, aggregator=aggregator ) except Exception: log.exception('could not run on %', bioprocess_node) return results
python
{ "resource": "" }
q262214
calculate_average_score_by_annotation
validation
def calculate_average_score_by_annotation( graph: BELGraph, annotation: str, key: Optional[str] = None, runs: Optional[int] = None, use_tqdm: bool = False, ) -> Mapping[str, float]: """For each sub-graph induced over the edges matching the annotation, calculate the average score for all of the contained biological processes Assumes you haven't done anything yet 1. Generates biological process upstream candidate mechanistic sub-graphs with :func:`generate_bioprocess_mechanisms` 2. Calculates scores for each sub-graph with :func:`calculate_average_scores_on_sub-graphs` 3. Overlays data with pbt.integration.overlay_data 4. Calculates averages with pbt.selection.group_nodes.average_node_annotation :param graph: A BEL graph :param annotation: A BEL annotation :param key: The key in the node data dictionary representing the experimental data. Defaults to :data:`pybel_tools.constants.WEIGHT`. :param runs: The number of times to run the heat diffusion workflow. Defaults to 100. :param use_tqdm: Should there be a progress bar for runners? :return: A dictionary from {str annotation value: tuple scores} Example Usage: >>> import pybel >>> from pybel_tools.integration import overlay_data >>> from pybel_tools.analysis.heat import calculate_average_score_by_annotation >>> graph = pybel.from_path(...) >>> scores = calculate_average_score_by_annotation(graph, 'subgraph') """ candidate_mechanisms = generate_bioprocess_mechanisms(graph, key=key) #: {bp tuple: list of scores} scores: Mapping[BaseEntity, Tuple] = calculate_average_scores_on_subgraphs( subgraphs=candidate_mechanisms, key=key, runs=runs, use_tqdm=use_tqdm, ) subgraph_bp: Mapping[str, List[BaseEntity]] = defaultdict(list) subgraphs: Mapping[str, BELGraph] = get_subgraphs_by_annotation(graph, annotation) for annotation_value, subgraph in subgraphs.items(): subgraph_bp[annotation_value].extend(get_nodes_by_function(subgraph, BIOPROCESS)) #: Pick the average by slicing with 0. Refer to :func:`calculate_average_score_on_subgraphs` return { annotation_value: np.average(scores[bp][0] for bp in bps) for annotation_value, bps in subgraph_bp.items() }
python
{ "resource": "" }
q262215
Runner.iter_leaves
validation
def iter_leaves(self) -> Iterable[BaseEntity]: """Return an iterable over all nodes that are leaves. A node is a leaf if either: - it doesn't have any predecessors, OR - all of its predecessors have a score in their data dictionaries """ for node in self.graph: if self.tag in self.graph.nodes[node]: continue if not any(self.tag not in self.graph.nodes[p] for p in self.graph.predecessors(node)): yield node
python
{ "resource": "" }
q262216
Runner.unscored_nodes_iter
validation
def unscored_nodes_iter(self) -> BaseEntity: """Iterate over all nodes without a score.""" for node, data in self.graph.nodes(data=True): if self.tag not in data: yield node
python
{ "resource": "" }
q262217
Runner.remove_random_edge_until_has_leaves
validation
def remove_random_edge_until_has_leaves(self) -> None: """Remove random edges until there is at least one leaf node.""" while True: leaves = set(self.iter_leaves()) if leaves: return self.remove_random_edge()
python
{ "resource": "" }
q262218
Runner.score_leaves
validation
def score_leaves(self) -> Set[BaseEntity]: """Calculate the score for all leaves. :return: The set of leaf nodes that were scored """ leaves = set(self.iter_leaves()) if not leaves: log.warning('no leaves.') return set() for leaf in leaves: self.graph.nodes[leaf][self.tag] = self.calculate_score(leaf) log.log(5, 'chomping %s', leaf) return leaves
python
{ "resource": "" }
q262219
Runner.run_with_graph_transformation
validation
def run_with_graph_transformation(self) -> Iterable[BELGraph]: """Calculate scores for all leaves until there are none, removes edges until there are, and repeats until all nodes have been scored. Also, yields the current graph at every step so you can make a cool animation of how the graph changes throughout the course of the algorithm :return: An iterable of BEL graphs """ yield self.get_remaining_graph() while not self.done_chomping(): while not list(self.iter_leaves()): self.remove_random_edge() yield self.get_remaining_graph() self.score_leaves() yield self.get_remaining_graph()
python
{ "resource": "" }
q262220
Runner.done_chomping
validation
def done_chomping(self) -> bool: """Determines if the algorithm is complete by checking if the target node of this analysis has been scored yet. Because the algorithm removes edges when it gets stuck until it is un-stuck, it is always guaranteed to finish. :return: Is the algorithm done running? """ return self.tag in self.graph.nodes[self.target_node]
python
{ "resource": "" }
q262221
Runner.get_final_score
validation
def get_final_score(self) -> float: """Return the final score for the target node. :return: The final score for the target node """ if not self.done_chomping(): raise ValueError('algorithm has not yet completed') return self.graph.nodes[self.target_node][self.tag]
python
{ "resource": "" }
q262222
Runner.calculate_score
validation
def calculate_score(self, node: BaseEntity) -> float: """Calculate the new score of the given node.""" score = ( self.graph.nodes[node][self.tag] if self.tag in self.graph.nodes[node] else self.default_score ) for predecessor, _, d in self.graph.in_edges(node, data=True): if d[RELATION] in CAUSAL_INCREASE_RELATIONS: score += self.graph.nodes[predecessor][self.tag] elif d[RELATION] in CAUSAL_DECREASE_RELATIONS: score -= self.graph.nodes[predecessor][self.tag] return score
python
{ "resource": "" }
q262223
microcanonical_statistics_dtype
validation
def microcanonical_statistics_dtype(spanning_cluster=True): """ Return the numpy structured array data type for sample states Helper function Parameters ---------- spanning_cluster : bool, optional Whether to detect a spanning cluster or not. Defaults to ``True``. Returns ------- ret : list of pairs of str A list of tuples of field names and data types to be used as ``dtype`` argument in numpy ndarray constructors See Also -------- http://docs.scipy.org/doc/numpy/user/basics.rec.html canonical_statistics_dtype """ fields = list() fields.extend([ ('n', 'uint32'), ('edge', 'uint32'), ]) if spanning_cluster: fields.extend([ ('has_spanning_cluster', 'bool'), ]) fields.extend([ ('max_cluster_size', 'uint32'), ('moments', '(5,)uint64'), ]) return _ndarray_dtype(fields)
python
{ "resource": "" }
q262224
canonical_statistics_dtype
validation
def canonical_statistics_dtype(spanning_cluster=True): """ The NumPy Structured Array type for canonical statistics Helper function Parameters ---------- spanning_cluster : bool, optional Whether to detect a spanning cluster or not. Defaults to ``True``. Returns ------- ret : list of pairs of str A list of tuples of field names and data types to be used as ``dtype`` argument in numpy ndarray constructors See Also -------- http://docs.scipy.org/doc/numpy/user/basics.rec.html microcanoncial_statistics_dtype canonical_averages_dtype """ fields = list() if spanning_cluster: fields.extend([ ('percolation_probability', 'float64'), ]) fields.extend([ ('max_cluster_size', 'float64'), ('moments', '(5,)float64'), ]) return _ndarray_dtype(fields)
python
{ "resource": "" }
q262225
bond_canonical_statistics
validation
def bond_canonical_statistics( microcanonical_statistics, convolution_factors, **kwargs ): """ canonical cluster statistics for a single run and a single probability Parameters ---------- microcanonical_statistics : ndarray Return value of `bond_microcanonical_statistics` convolution_factors : 1-D array_like The coefficients of the convolution for the given probabilty ``p`` and for each occupation number ``n``. Returns ------- ret : ndarray of size ``1`` Structured array with dtype as returned by `canonical_statistics_dtype` ret['percolation_probability'] : ndarray of float The "percolation probability" of this run at the value of ``p``. Only exists if `microcanonical_statistics` argument has the ``has_spanning_cluster`` field. ret['max_cluster_size'] : ndarray of int Weighted size of the largest cluster (absolute number of sites) ret['moments'] : 1-D :py:class:`numpy.ndarray` of float Array of size ``5``. The ``k``-th entry is the weighted ``k``-th raw moment of the (absolute) cluster size distribution, with ``k`` ranging from ``0`` to ``4``. See Also -------- bond_microcanonical_statistics canonical_statistics_dtype """ # initialize return array spanning_cluster = ( 'has_spanning_cluster' in microcanonical_statistics.dtype.names ) ret = np.empty(1, dtype=canonical_statistics_dtype(spanning_cluster)) # compute percolation probability if spanning_cluster: ret['percolation_probability'] = np.sum( convolution_factors * microcanonical_statistics['has_spanning_cluster'] ) # convolve maximum cluster size ret['max_cluster_size'] = np.sum( convolution_factors * microcanonical_statistics['max_cluster_size'] ) # convolve moments ret['moments'] = np.sum( convolution_factors[:, np.newaxis] * microcanonical_statistics['moments'], axis=0, ) # return convolved cluster statistics return ret
python
{ "resource": "" }
q262226
canonical_averages_dtype
validation
def canonical_averages_dtype(spanning_cluster=True): """ The NumPy Structured Array type for canonical averages over several runs Helper function Parameters ---------- spanning_cluster : bool, optional Whether to detect a spanning cluster or not. Defaults to ``True``. Returns ------- ret : list of pairs of str A list of tuples of field names and data types to be used as ``dtype`` argument in numpy ndarray constructors See Also -------- http://docs.scipy.org/doc/numpy/user/basics.rec.html canonical_statistics_dtype finalized_canonical_averages_dtype """ fields = list() fields.extend([ ('number_of_runs', 'uint32'), ]) if spanning_cluster: fields.extend([ ('percolation_probability_mean', 'float64'), ('percolation_probability_m2', 'float64'), ]) fields.extend([ ('max_cluster_size_mean', 'float64'), ('max_cluster_size_m2', 'float64'), ('moments_mean', '(5,)float64'), ('moments_m2', '(5,)float64'), ]) return _ndarray_dtype(fields)
python
{ "resource": "" }
q262227
bond_initialize_canonical_averages
validation
def bond_initialize_canonical_averages( canonical_statistics, **kwargs ): """ Initialize the canonical averages from a single-run cluster statistics Parameters ---------- canonical_statistics : 1-D structured ndarray Typically contains the canonical statistics for a range of values of the occupation probability ``p``. The dtype is the result of `canonical_statistics_dtype`. Returns ------- ret : structured ndarray The dype is the result of `canonical_averages_dtype`. ret['number_of_runs'] : 1-D ndarray of int Equals ``1`` (initial run). ret['percolation_probability_mean'] : 1-D array of float Equals ``canonical_statistics['percolation_probability']`` (if ``percolation_probability`` is present) ret['percolation_probability_m2'] : 1-D array of float Each entry is ``0.0`` ret['max_cluster_size_mean'] : 1-D array of float Equals ``canonical_statistics['max_cluster_size']`` ret['max_cluster_size_m2'] : 1-D array of float Each entry is ``0.0`` ret['moments_mean'] : 2-D array of float Equals ``canonical_statistics['moments']`` ret['moments_m2'] : 2-D array of float Each entry is ``0.0`` See Also -------- canonical_averages_dtype bond_canonical_statistics """ # initialize return array spanning_cluster = ( 'percolation_probability' in canonical_statistics.dtype.names ) # array should have the same size as the input array ret = np.empty_like( canonical_statistics, dtype=canonical_averages_dtype(spanning_cluster=spanning_cluster), ) ret['number_of_runs'] = 1 # initialize percolation probability mean and sum of squared differences if spanning_cluster: ret['percolation_probability_mean'] = ( canonical_statistics['percolation_probability'] ) ret['percolation_probability_m2'] = 0.0 # initialize maximum cluster size mean and sum of squared differences ret['max_cluster_size_mean'] = ( canonical_statistics['max_cluster_size'] ) ret['max_cluster_size_m2'] = 0.0 # initialize moments means and sums of squared differences ret['moments_mean'] = canonical_statistics['moments'] ret['moments_m2'] = 0.0 return ret
python
{ "resource": "" }
q262228
bond_reduce
validation
def bond_reduce(row_a, row_b): """ Reduce the canonical averages over several runs This is a "true" reducer. It is associative and commutative. This is a wrapper around `simoa.stats.online_variance`. Parameters ---------- row_a, row_b : structured ndarrays Output of this function, or initial input from `bond_initialize_canonical_averages` Returns ------- ret : structured ndarray Array is of dtype as returned by `canonical_averages_dtype` See Also -------- bond_initialize_canonical_averages canonical_averages_dtype simoa.stats.online_variance """ spanning_cluster = ( 'percolation_probability_mean' in row_a.dtype.names and 'percolation_probability_mean' in row_b.dtype.names and 'percolation_probability_m2' in row_a.dtype.names and 'percolation_probability_m2' in row_b.dtype.names ) # initialize return array ret = np.empty_like(row_a) def _reducer(key, transpose=False): mean_key = '{}_mean'.format(key) m2_key = '{}_m2'.format(key) res = simoa.stats.online_variance(*[ ( row['number_of_runs'], row[mean_key].T if transpose else row[mean_key], row[m2_key].T if transpose else row[m2_key], ) for row in [row_a, row_b] ]) ( ret[mean_key], ret[m2_key], ) = ( res[1].T, res[2].T, ) if transpose else res[1:] if spanning_cluster: _reducer('percolation_probability') _reducer('max_cluster_size') _reducer('moments', transpose=True) ret['number_of_runs'] = row_a['number_of_runs'] + row_b['number_of_runs'] return ret
python
{ "resource": "" }
q262229
finalized_canonical_averages_dtype
validation
def finalized_canonical_averages_dtype(spanning_cluster=True): """ The NumPy Structured Array type for finalized canonical averages over several runs Helper function Parameters ---------- spanning_cluster : bool, optional Whether to detect a spanning cluster or not. Defaults to ``True``. Returns ------- ret : list of pairs of str A list of tuples of field names and data types to be used as ``dtype`` argument in numpy ndarray constructors See Also -------- http://docs.scipy.org/doc/numpy/user/basics.rec.html canonical_averages_dtype """ fields = list() fields.extend([ ('number_of_runs', 'uint32'), ('p', 'float64'), ('alpha', 'float64'), ]) if spanning_cluster: fields.extend([ ('percolation_probability_mean', 'float64'), ('percolation_probability_std', 'float64'), ('percolation_probability_ci', '(2,)float64'), ]) fields.extend([ ('percolation_strength_mean', 'float64'), ('percolation_strength_std', 'float64'), ('percolation_strength_ci', '(2,)float64'), ('moments_mean', '(5,)float64'), ('moments_std', '(5,)float64'), ('moments_ci', '(5,2)float64'), ]) return _ndarray_dtype(fields)
python
{ "resource": "" }
q262230
finalize_canonical_averages
validation
def finalize_canonical_averages( number_of_nodes, ps, canonical_averages, alpha, ): """ Finalize canonical averages """ spanning_cluster = ( ( 'percolation_probability_mean' in canonical_averages.dtype.names ) and 'percolation_probability_m2' in canonical_averages.dtype.names ) # append values of p as an additional field ret = np.empty_like( canonical_averages, dtype=finalized_canonical_averages_dtype( spanning_cluster=spanning_cluster ), ) n = canonical_averages['number_of_runs'] sqrt_n = np.sqrt(canonical_averages['number_of_runs']) ret['number_of_runs'] = n ret['p'] = ps ret['alpha'] = alpha def _transform( original_key, final_key=None, normalize=False, transpose=False, ): if final_key is None: final_key = original_key keys_mean = [ '{}_mean'.format(key) for key in [original_key, final_key] ] keys_std = [ '{}_m2'.format(original_key), '{}_std'.format(final_key), ] key_ci = '{}_ci'.format(final_key) # calculate sample mean ret[keys_mean[1]] = canonical_averages[keys_mean[0]] if normalize: ret[keys_mean[1]] /= number_of_nodes # calculate sample standard deviation array = canonical_averages[keys_std[0]] result = np.sqrt( (array.T if transpose else array) / (n - 1) ) ret[keys_std[1]] = ( result.T if transpose else result ) if normalize: ret[keys_std[1]] /= number_of_nodes # calculate standard normal confidence interval array = ret[keys_std[1]] scale = (array.T if transpose else array) / sqrt_n array = ret[keys_mean[1]] mean = (array.T if transpose else array) result = scipy.stats.t.interval( 1 - alpha, df=n - 1, loc=mean, scale=scale, ) ( ret[key_ci][..., 0], ret[key_ci][..., 1] ) = ([my_array.T for my_array in result] if transpose else result) if spanning_cluster: _transform('percolation_probability') _transform('max_cluster_size', 'percolation_strength', normalize=True) _transform('moments', normalize=True, transpose=True) return ret
python
{ "resource": "" }
q262231
compare
validation
def compare(graph: BELGraph, annotation: str = 'Subgraph') -> Mapping[str, Mapping[str, float]]: """Compare generated mechanisms to actual ones. 1. Generates candidate mechanisms for each biological process 2. Gets sub-graphs for all NeuroMMSig signatures 3. Make tanimoto similarity comparison for all sets :return: A dictionary table comparing the canonical subgraphs to generated ones """ canonical_mechanisms = get_subgraphs_by_annotation(graph, annotation) canonical_nodes = _transform_graph_dict_to_node_dict(canonical_mechanisms) candidate_mechanisms = generate_bioprocess_mechanisms(graph) candidate_nodes = _transform_graph_dict_to_node_dict(candidate_mechanisms) results: Dict[str, Dict[str, float]] = defaultdict(dict) it = itt.product(canonical_nodes.items(), candidate_nodes.items()) for (canonical_name, canonical_graph), (candidate_bp, candidate_graph) in it: tanimoto = tanimoto_set_similarity(candidate_nodes, canonical_nodes) results[canonical_name][candidate_bp] = tanimoto return dict(results)
python
{ "resource": "" }
q262232
summarize_node_filter
validation
def summarize_node_filter(graph: BELGraph, node_filters: NodePredicates) -> None: """Print a summary of the number of nodes passing a given set of filters. :param graph: A BEL graph :param node_filters: A node filter or list/tuple of node filters """ passed = count_passed_node_filter(graph, node_filters) print('{}/{} nodes passed'.format(passed, graph.number_of_nodes()))
python
{ "resource": "" }
q262233
node_inclusion_filter_builder
validation
def node_inclusion_filter_builder(nodes: Iterable[BaseEntity]) -> NodePredicate: """Build a filter that only passes on nodes in the given list. :param nodes: An iterable of BEL nodes """ node_set = set(nodes) def inclusion_filter(_: BELGraph, node: BaseEntity) -> bool: """Pass only for a node that is in the enclosed node list. :return: If the node is contained within the enclosed node list """ return node in node_set return inclusion_filter
python
{ "resource": "" }
q262234
node_exclusion_filter_builder
validation
def node_exclusion_filter_builder(nodes: Iterable[BaseEntity]) -> NodePredicate: """Build a filter that fails on nodes in the given list.""" node_set = set(nodes) def exclusion_filter(_: BELGraph, node: BaseEntity) -> bool: """Pass only for a node that isn't in the enclosed node list :return: If the node isn't contained within the enclosed node list """ return node not in node_set return exclusion_filter
python
{ "resource": "" }
q262235
function_namespace_inclusion_builder
validation
def function_namespace_inclusion_builder(func: str, namespace: Strings) -> NodePredicate: """Build a filter function for matching the given BEL function with the given namespace or namespaces. :param func: A BEL function :param namespace: The namespace to search by """ if isinstance(namespace, str): def function_namespaces_filter(_: BELGraph, node: BaseEntity) -> bool: """Pass only for nodes that have the enclosed function and enclosed namespace.""" if func != node[FUNCTION]: return False return NAMESPACE in node and node[NAMESPACE] == namespace elif isinstance(namespace, Iterable): namespaces = set(namespace) def function_namespaces_filter(_: BELGraph, node: BaseEntity) -> bool: """Pass only for nodes that have the enclosed function and namespace in the enclose set.""" if func != node[FUNCTION]: return False return NAMESPACE in node and node[NAMESPACE] in namespaces else: raise ValueError('Invalid type for argument: {}'.format(namespace)) return function_namespaces_filter
python
{ "resource": "" }
q262236
data_contains_key_builder
validation
def data_contains_key_builder(key: str) -> NodePredicate: # noqa: D202 """Build a filter that passes only on nodes that have the given key in their data dictionary. :param key: A key for the node's data dictionary """ def data_contains_key(_: BELGraph, node: BaseEntity) -> bool: """Pass only for a node that contains the enclosed key in its data dictionary. :return: If the node contains the enclosed key in its data dictionary """ return key in node return data_contains_key
python
{ "resource": "" }
q262237
variants_of
validation
def variants_of( graph: BELGraph, node: Protein, modifications: Optional[Set[str]] = None, ) -> Set[Protein]: """Returns all variants of the given node.""" if modifications: return _get_filtered_variants_of(graph, node, modifications) return { v for u, v, key, data in graph.edges(keys=True, data=True) if ( u == node and data[RELATION] == HAS_VARIANT and pybel.struct.has_protein_modification(v) ) }
python
{ "resource": "" }
q262238
get_variants_to_controllers
validation
def get_variants_to_controllers( graph: BELGraph, node: Protein, modifications: Optional[Set[str]] = None, ) -> Mapping[Protein, Set[Protein]]: """Get a mapping from variants of the given node to all of its upstream controllers.""" rv = defaultdict(set) variants = variants_of(graph, node, modifications) for controller, variant, data in graph.in_edges(variants, data=True): if data[RELATION] in CAUSAL_RELATIONS: rv[variant].add(controller) return rv
python
{ "resource": "" }
q262239
group_dict_set
validation
def group_dict_set(iterator: Iterable[Tuple[A, B]]) -> Mapping[A, Set[B]]: """Make a dict that accumulates the values for each key in an iterator of doubles.""" d = defaultdict(set) for key, value in iterator: d[key].add(value) return dict(d)
python
{ "resource": "" }
q262240
count_unique_relations
validation
def count_unique_relations(graph: BELGraph) -> Counter: """Return a histogram of the different types of relations present in a graph. Note: this operation only counts each type of edge once for each pair of nodes """ return Counter(itt.chain.from_iterable(get_edge_relations(graph).values()))
python
{ "resource": "" }
q262241
count_annotation_values
validation
def count_annotation_values(graph: BELGraph, annotation: str) -> Counter: """Count in how many edges each annotation appears in a graph :param graph: A BEL graph :param annotation: The annotation to count :return: A Counter from {annotation value: frequency} """ return Counter(iter_annotation_values(graph, annotation))
python
{ "resource": "" }
q262242
count_annotation_values_filtered
validation
def count_annotation_values_filtered(graph: BELGraph, annotation: str, source_predicate: Optional[NodePredicate] = None, target_predicate: Optional[NodePredicate] = None, ) -> Counter: """Count in how many edges each annotation appears in a graph, but filter out source nodes and target nodes. See :func:`pybel_tools.utils.keep_node` for a basic filter. :param graph: A BEL graph :param annotation: The annotation to count :param source_predicate: A predicate (graph, node) -> bool for keeping source nodes :param target_predicate: A predicate (graph, node) -> bool for keeping target nodes :return: A Counter from {annotation value: frequency} """ if source_predicate and target_predicate: return Counter( data[ANNOTATIONS][annotation] for u, v, data in graph.edges(data=True) if edge_has_annotation(data, annotation) and source_predicate(graph, u) and target_predicate(graph, v) ) elif source_predicate: return Counter( data[ANNOTATIONS][annotation] for u, v, data in graph.edges(data=True) if edge_has_annotation(data, annotation) and source_predicate(graph, u) ) elif target_predicate: return Counter( data[ANNOTATIONS][annotation] for u, v, data in graph.edges(data=True) if edge_has_annotation(data, annotation) and target_predicate(graph, u) ) else: return Counter( data[ANNOTATIONS][annotation] for u, v, data in graph.edges(data=True) if edge_has_annotation(data, annotation) )
python
{ "resource": "" }
q262243
pair_is_consistent
validation
def pair_is_consistent(graph: BELGraph, u: BaseEntity, v: BaseEntity) -> Optional[str]: """Return if the edges between the given nodes are consistent, meaning they all have the same relation. :return: If the edges aren't consistent, return false, otherwise return the relation type """ relations = {data[RELATION] for data in graph[u][v].values()} if 1 != len(relations): return return list(relations)[0]
python
{ "resource": "" }
q262244
infer_missing_two_way_edges
validation
def infer_missing_two_way_edges(graph): """Add edges to the graph when a two way edge exists, and the opposite direction doesn't exist. Use: two way edges from BEL definition and/or axiomatic inverses of membership relations :param pybel.BELGraph graph: A BEL graph """ for u, v, k, d in graph.edges(data=True, keys=True): if d[RELATION] in TWO_WAY_RELATIONS: infer_missing_backwards_edge(graph, u, v, k)
python
{ "resource": "" }
q262245
infer_missing_backwards_edge
validation
def infer_missing_backwards_edge(graph, u, v, k): """Add the same edge, but in the opposite direction if not already present. :type graph: pybel.BELGraph :type u: tuple :type v: tuple :type k: int """ if u in graph[v]: for attr_dict in graph[v][u].values(): if attr_dict == graph[u][v][k]: return graph.add_edge(v, u, key=k, **graph[u][v][k])
python
{ "resource": "" }
q262246
enrich_internal_unqualified_edges
validation
def enrich_internal_unqualified_edges(graph, subgraph): """Add the missing unqualified edges between entities in the subgraph that are contained within the full graph. :param pybel.BELGraph graph: The full BEL graph :param pybel.BELGraph subgraph: The query BEL subgraph """ for u, v in itt.combinations(subgraph, 2): if not graph.has_edge(u, v): continue for k in graph[u][v]: if k < 0: subgraph.add_edge(u, v, key=k, **graph[u][v][k])
python
{ "resource": "" }
q262247
boilerplate
validation
def boilerplate(name, contact, description, pmids, version, copyright, authors, licenses, disclaimer, output): """Build a template BEL document with the given PubMed identifiers.""" from .document_utils import write_boilerplate write_boilerplate( name=name, version=version, description=description, authors=authors, contact=contact, copyright=copyright, licenses=licenses, disclaimer=disclaimer, pmids=pmids, file=output, )
python
{ "resource": "" }
q262248
get_pmids
validation
def get_pmids(graph: BELGraph, output: TextIO): """Output PubMed identifiers from a graph to a stream.""" for pmid in get_pubmed_identifiers(graph): click.echo(pmid, file=output)
python
{ "resource": "" }
q262249
Table.getrowcount
validation
def getrowcount(self, window_name, object_name): """ Get count of rows in table object. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @return: Number of rows. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) return len(object_handle.AXRows)
python
{ "resource": "" }
q262250
Table.multiselect
validation
def multiselect(self, window_name, object_name, row_text_list, partial_match=False): """ Select multiple row @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_text_list: Row list with matching text to select @type row_text: string @return: 1 on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) object_handle.activate() selected = False try: window = self._get_front_most_window() except (IndexError,): window = self._get_any_window() for row_text in row_text_list: selected = False for cell in object_handle.AXRows: parent_cell = cell cell = self._getfirstmatchingchild(cell, "(AXTextField|AXStaticText)") if not cell: continue if re.match(row_text, cell.AXValue): selected = True if not parent_cell.AXSelected: x, y, width, height = self._getobjectsize(parent_cell) window.clickMouseButtonLeftWithMods((x + width / 2, y + height / 2), ['<command_l>']) # Following selection doesn't work # parent_cell.AXSelected=True self.wait(0.5) else: # Selected pass break if not selected: raise LdtpServerException(u"Unable to select row: %s" % row_text) if not selected: raise LdtpServerException(u"Unable to select any row") return 1
python
{ "resource": "" }
q262251
Table.selectrowpartialmatch
validation
def selectrowpartialmatch(self, window_name, object_name, row_text): """ Select row partial match @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_text: Row text to select @type row_text: string @return: 1 on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) for cell in object_handle.AXRows: if re.search(row_text, cell.AXChildren[0].AXValue): if not cell.AXSelected: object_handle.activate() cell.AXSelected = True else: # Selected pass return 1 raise LdtpServerException(u"Unable to select row: %s" % row_text)
python
{ "resource": "" }
q262252
Table.selectrowindex
validation
def selectrowindex(self, window_name, object_name, row_index): """ Select row index @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_index: Row index to select @type row_index: integer @return: 1 on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) count = len(object_handle.AXRows) if row_index < 0 or row_index > count: raise LdtpServerException('Row index out of range: %d' % row_index) cell = object_handle.AXRows[row_index] if not cell.AXSelected: object_handle.activate() cell.AXSelected = True else: # Selected pass return 1
python
{ "resource": "" }
q262253
Table.selectlastrow
validation
def selectlastrow(self, window_name, object_name): """ Select last row @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) cell = object_handle.AXRows[-1] if not cell.AXSelected: object_handle.activate() cell.AXSelected = True else: # Selected pass return 1
python
{ "resource": "" }
q262254
Table.getcellvalue
validation
def getcellvalue(self, window_name, object_name, row_index, column=0): """ Get cell value @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_index: Row index to get @type row_index: integer @param column: Column index to get, default value 0 @type column: integer @return: cell value on success. @rtype: string """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) count = len(object_handle.AXRows) if row_index < 0 or row_index > count: raise LdtpServerException('Row index out of range: %d' % row_index) cell = object_handle.AXRows[row_index] count = len(cell.AXChildren) if column < 0 or column > count: raise LdtpServerException('Column index out of range: %d' % column) obj = cell.AXChildren[column] if not re.search("AXColumn", obj.AXRole): obj = cell.AXChildren[column] return obj.AXValue
python
{ "resource": "" }
q262255
Table.gettablerowindex
validation
def gettablerowindex(self, window_name, object_name, row_text): """ Get table row index matching given text @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_text: Row text to select @type row_text: string @return: row index matching the text on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) index = 0 for cell in object_handle.AXRows: if re.match(row_text, cell.AXChildren[0].AXValue): return index index += 1 raise LdtpServerException(u"Unable to find row: %s" % row_text)
python
{ "resource": "" }
q262256
Table.verifypartialtablecell
validation
def verifypartialtablecell(self, window_name, object_name, row_index, column_index, row_text): """ Verify partial table cell value @param window_name: Window name to type in, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to type in, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param row_index: Row index to get @type row_index: integer @param column_index: Column index to get, default value 0 @type column_index: integer @param row_text: Row text to match @type string @return: 1 on success 0 on failure. @rtype: integer """ try: value = getcellvalue(window_name, object_name, row_index, column_index) if re.searchmatch(row_text, value): return 1 except LdtpServerException: pass return 0
python
{ "resource": "" }
q262257
Core.getapplist
validation
def getapplist(self): """ Get all accessibility application name that are currently running @return: list of appliction name of string type on success. @rtype: list """ app_list = [] # Update apps list, before parsing the list self._update_apps() for gui in self._running_apps: name = gui.localizedName() # default type was objc.pyobjc_unicode # convert to Unicode, else exception is thrown # TypeError: "cannot marshal <type 'objc.pyobjc_unicode'> objects" try: name = unicode(name) except NameError: name = str(name) except UnicodeEncodeError: pass app_list.append(name) # Return unique application list return list(set(app_list))
python
{ "resource": "" }
q262258
Core.startprocessmonitor
validation
def startprocessmonitor(self, process_name, interval=2): """ Start memory and CPU monitoring, with the time interval between each process scan @param process_name: Process name, ex: firefox-bin. @type process_name: string @param interval: Time interval between each process scan @type interval: double @return: 1 on success @rtype: integer """ if process_name in self._process_stats: # Stop previously running instance # At any point, only one process name can be tracked # If an instance already exist, then stop it self._process_stats[process_name].stop() # Create an instance of process stat self._process_stats[process_name] = ProcessStats(process_name, interval) # start monitoring the process self._process_stats[process_name].start() return 1
python
{ "resource": "" }
q262259
Core.stopprocessmonitor
validation
def stopprocessmonitor(self, process_name): """ Stop memory and CPU monitoring @param process_name: Process name, ex: firefox-bin. @type process_name: string @return: 1 on success @rtype: integer """ if process_name in self._process_stats: # Stop monitoring process self._process_stats[process_name].stop() return 1
python
{ "resource": "" }
q262260
Core.getcpustat
validation
def getcpustat(self, process_name): """ get CPU stat for the give process name @param process_name: Process name, ex: firefox-bin. @type process_name: string @return: cpu stat list on success, else empty list If same process name, running multiple instance, get the stat of all the process CPU usage @rtype: list """ # Create an instance of process stat _stat_inst = ProcessStats(process_name) _stat_list = [] for p in _stat_inst.get_cpu_memory_stat(): try: _stat_list.append(p.get_cpu_percent()) except psutil.AccessDenied: pass return _stat_list
python
{ "resource": "" }
q262261
Core.getmemorystat
validation
def getmemorystat(self, process_name): """ get memory stat @param process_name: Process name, ex: firefox-bin. @type process_name: string @return: memory stat list on success, else empty list If same process name, running multiple instance, get the stat of all the process memory usage @rtype: list """ # Create an instance of process stat _stat_inst = ProcessStats(process_name) _stat_list = [] for p in _stat_inst.get_cpu_memory_stat(): # Memory percent returned with 17 decimal values # ex: 0.16908645629882812, round it to 2 decimal values # as 0.03 try: _stat_list.append(round(p.get_memory_percent(), 2)) except psutil.AccessDenied: pass return _stat_list
python
{ "resource": "" }
q262262
Core.getobjectlist
validation
def getobjectlist(self, window_name): """ Get list of items in given GUI. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @return: list of items in LDTP naming convention. @rtype: list """ try: window_handle, name, app = self._get_window_handle(window_name, True) object_list = self._get_appmap(window_handle, name, True) except atomac._a11y.ErrorInvalidUIElement: # During the test, when the window closed and reopened # ErrorInvalidUIElement exception will be thrown self._windows = {} # Call the method again, after updating apps window_handle, name, app = self._get_window_handle(window_name, True) object_list = self._get_appmap(window_handle, name, True) return object_list.keys()
python
{ "resource": "" }
q262263
Core.getobjectinfo
validation
def getobjectinfo(self, window_name, object_name): """ Get object properties. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: list of properties @rtype: list """ try: obj_info = self._get_object_map(window_name, object_name, wait_for_object=False) except atomac._a11y.ErrorInvalidUIElement: # During the test, when the window closed and reopened # ErrorInvalidUIElement exception will be thrown self._windows = {} # Call the method again, after updating apps obj_info = self._get_object_map(window_name, object_name, wait_for_object=False) props = [] if obj_info: for obj_prop in obj_info.keys(): if not obj_info[obj_prop] or obj_prop == "obj": # Don't add object handle to the list continue props.append(obj_prop) return props
python
{ "resource": "" }
q262264
Core.getobjectproperty
validation
def getobjectproperty(self, window_name, object_name, prop): """ Get object property value. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param prop: property name. @type prop: string @return: property @rtype: string """ try: obj_info = self._get_object_map(window_name, object_name, wait_for_object=False) except atomac._a11y.ErrorInvalidUIElement: # During the test, when the window closed and reopened # ErrorInvalidUIElement exception will be thrown self._windows = {} # Call the method again, after updating apps obj_info = self._get_object_map(window_name, object_name, wait_for_object=False) if obj_info and prop != "obj" and prop in obj_info: if prop == "class": # ldtp_class_type are compatible with Linux and Windows class name # If defined class name exist return that, # else return as it is return ldtp_class_type.get(obj_info[prop], obj_info[prop]) else: return obj_info[prop] raise LdtpServerException('Unknown property "%s" in %s' % \ (prop, object_name))
python
{ "resource": "" }
q262265
Core.getchild
validation
def getchild(self, window_name, child_name='', role='', parent=''): """ Gets the list of object available in the window, which matches component name or role name or both. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param child_name: Child name to search for. @type child_name: string @param role: role name to search for, or an empty string for wildcard. @type role: string @param parent: parent name to search for, or an empty string for wildcard. @type role: string @return: list of matched children names @rtype: list """ matches = [] if role: role = re.sub(' ', '_', role) self._windows = {} if parent and (child_name or role): _window_handle, _window_name = \ self._get_window_handle(window_name)[0:2] if not _window_handle: raise LdtpServerException('Unable to find window "%s"' % \ window_name) appmap = self._get_appmap(_window_handle, _window_name) obj = self._get_object_map(window_name, parent) def _get_all_children_under_obj(obj, child_list): if role and obj['class'] == role: child_list.append(obj['label']) elif child_name and self._match_name_to_appmap(child_name, obj): child_list.append(obj['label']) if obj: children = obj['children'] if not children: return child_list for child in children.split(): return _get_all_children_under_obj( \ appmap[child], child_list) matches = _get_all_children_under_obj(obj, []) if not matches: if child_name: _name = 'name "%s" ' % child_name if role: _role = 'role "%s" ' % role if parent: _parent = 'parent "%s"' % parent exception = 'Could not find a child %s%s%s' % (_name, _role, _parent) raise LdtpServerException(exception) return matches _window_handle, _window_name = \ self._get_window_handle(window_name)[0:2] if not _window_handle: raise LdtpServerException('Unable to find window "%s"' % \ window_name) appmap = self._get_appmap(_window_handle, _window_name) for name in appmap.keys(): obj = appmap[name] # When only role arg is passed if role and not child_name and obj['class'] == role: matches.append(name) # When parent and child_name arg is passed if parent and child_name and not role and \ self._match_name_to_appmap(parent, obj): matches.append(name) # When only child_name arg is passed if child_name and not role and \ self._match_name_to_appmap(child_name, obj): return name matches.append(name) # When role and child_name args are passed if role and child_name and obj['class'] == role and \ self._match_name_to_appmap(child_name, obj): matches.append(name) if not matches: _name = '' _role = '' _parent = '' if child_name: _name = 'name "%s" ' % child_name if role: _role = 'role "%s" ' % role if parent: _parent = 'parent "%s"' % parent exception = 'Could not find a child %s%s%s' % (_name, _role, _parent) raise LdtpServerException(exception) return matches
python
{ "resource": "" }
q262266
Core.launchapp
validation
def launchapp(self, cmd, args=[], delay=0, env=1, lang="C"): """ Launch application. @param cmd: Command line string to execute. @type cmd: string @param args: Arguments to the application @type args: list @param delay: Delay after the application is launched @type delay: int @param env: GNOME accessibility environment to be set or not @type env: int @param lang: Application language to be used @type lang: string @return: 1 on success @rtype: integer @raise LdtpServerException: When command fails """ try: atomac.NativeUIElement.launchAppByBundleId(cmd) return 1 except RuntimeError: if atomac.NativeUIElement.launchAppByBundlePath(cmd, args): # Let us wait so that the application launches try: time.sleep(int(delay)) except ValueError: time.sleep(5) return 1 else: raise LdtpServerException(u"Unable to find app '%s'" % cmd)
python
{ "resource": "" }
q262267
Core.activatewindow
validation
def activatewindow(self, window_name): """ Activate window. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @return: 1 on success. @rtype: integer """ window_handle = self._get_window_handle(window_name) self._grabfocus(window_handle) return 1
python
{ "resource": "" }
q262268
Core.click
validation
def click(self, window_name, object_name): """ Click item. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success. @rtype: integer """ object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) size = self._getobjectsize(object_handle) self._grabfocus(object_handle) self.wait(0.5) # If object doesn't support Press, trying clicking with the object # coordinates, where size=(x, y, width, height) # click on center of the widget # Noticed this issue on clicking AXImage # click('Instruments*', 'Automation') self.generatemouseevent(size[0] + size[2] / 2, size[1] + size[3] / 2, "b1c") return 1
python
{ "resource": "" }
q262269
Core.getallstates
validation
def getallstates(self, window_name, object_name): """ Get all states of given object @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: list of string on success. @rtype: list """ object_handle = self._get_object_handle(window_name, object_name) _obj_states = [] if object_handle.AXEnabled: _obj_states.append("enabled") if object_handle.AXFocused: _obj_states.append("focused") else: try: if object_handle.AXFocused: _obj_states.append("focusable") except: pass if re.match("AXCheckBox", object_handle.AXRole, re.M | re.U | re.L) or \ re.match("AXRadioButton", object_handle.AXRole, re.M | re.U | re.L): if object_handle.AXValue: _obj_states.append("checked") return _obj_states
python
{ "resource": "" }
q262270
Core.getobjectsize
validation
def getobjectsize(self, window_name, object_name=None): """ Get object size @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @return: x, y, width, height on success. @rtype: list """ if not object_name: handle, name, app = self._get_window_handle(window_name) else: handle = self._get_object_handle(window_name, object_name) return self._getobjectsize(handle)
python
{ "resource": "" }
q262271
Core.grabfocus
validation
def grabfocus(self, window_name, object_name=None): """ Grab focus. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success. @rtype: integer """ if not object_name: handle, name, app = self._get_window_handle(window_name) else: handle = self._get_object_handle(window_name, object_name) return self._grabfocus(handle)
python
{ "resource": "" }
q262272
Core.waittillguiexist
validation
def waittillguiexist(self, window_name, object_name='', guiTimeOut=30, state=''): """ Wait till a window or component exists. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param guiTimeOut: Wait timeout in seconds @type guiTimeOut: integer @param state: Object state used only when object_name is provided. @type object_name: string @return: 1 if GUI was found, 0 if not. @rtype: integer """ timeout = 0 while timeout < guiTimeOut: if self.guiexist(window_name, object_name): return 1 # Wait 1 second before retrying time.sleep(1) timeout += 1 # Object and/or window doesn't appear within the timeout period return 0
python
{ "resource": "" }
q262273
Core.waittillguinotexist
validation
def waittillguinotexist(self, window_name, object_name='', guiTimeOut=30): """ Wait till a window does not exist. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @param guiTimeOut: Wait timeout in seconds @type guiTimeOut: integer @return: 1 if GUI has gone away, 0 if not. @rtype: integer """ timeout = 0 while timeout < guiTimeOut: if not self.guiexist(window_name, object_name): return 1 # Wait 1 second before retrying time.sleep(1) timeout += 1 # Object and/or window still appears within the timeout period return 0
python
{ "resource": "" }
q262274
Core.stateenabled
validation
def stateenabled(self, window_name, object_name): """ Check whether an object state is enabled or not @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success 0 on failure. @rtype: integer """ try: object_handle = self._get_object_handle(window_name, object_name) if object_handle.AXEnabled: return 1 except LdtpServerException: pass return 0
python
{ "resource": "" }
q262275
Core.check
validation
def check(self, window_name, object_name): """ Check item. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success. @rtype: integer """ # FIXME: Check for object type object_handle = self._get_object_handle(window_name, object_name) if not object_handle.AXEnabled: raise LdtpServerException(u"Object %s state disabled" % object_name) if object_handle.AXValue == 1: # Already checked return 1 # AXPress doesn't work with Instruments # So did the following work around self._grabfocus(object_handle) x, y, width, height = self._getobjectsize(object_handle) # Mouse left click on the object # Note: x + width/2, y + height / 2 doesn't work self.generatemouseevent(x + width / 2, y + height / 2, "b1c") return 1
python
{ "resource": "" }
q262276
Core.verifycheck
validation
def verifycheck(self, window_name, object_name): """ Verify check item. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. @type object_name: string @return: 1 on success 0 on failure. @rtype: integer """ try: object_handle = self._get_object_handle(window_name, object_name, wait_for_object=False) if object_handle.AXValue == 1: return 1 except LdtpServerException: pass return 0
python
{ "resource": "" }
q262277
Core.getaccesskey
validation
def getaccesskey(self, window_name, object_name): """ Get access key of given object @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @return: access key in string format on success, else LdtpExecutionError on failure. @rtype: string """ # Used http://www.danrodney.com/mac/ as reference for # mapping keys with specific control # In Mac noticed (in accessibility inspector) only menu had access keys # so, get the menu_handle of given object and # return the access key menu_handle = self._get_menu_handle(window_name, object_name) key = menu_handle.AXMenuItemCmdChar modifiers = menu_handle.AXMenuItemCmdModifiers glpyh = menu_handle.AXMenuItemCmdGlyph virtual_key = menu_handle.AXMenuItemCmdVirtualKey modifiers_type = "" if modifiers == 0: modifiers_type = "<command>" elif modifiers == 1: modifiers_type = "<shift><command>" elif modifiers == 2: modifiers_type = "<option><command>" elif modifiers == 3: modifiers_type = "<option><shift><command>" elif modifiers == 4: modifiers_type = "<ctrl><command>" elif modifiers == 6: modifiers_type = "<ctrl><option><command>" # Scroll up if virtual_key == 115 and glpyh == 102: modifiers = "<option>" key = "<cursor_left>" # Scroll down elif virtual_key == 119 and glpyh == 105: modifiers = "<option>" key = "<right>" # Page up elif virtual_key == 116 and glpyh == 98: modifiers = "<option>" key = "<up>" # Page down elif virtual_key == 121 and glpyh == 107: modifiers = "<option>" key = "<down>" # Line up elif virtual_key == 126 and glpyh == 104: key = "<up>" # Line down elif virtual_key == 125 and glpyh == 106: key = "<down>" # Noticed in Google Chrome navigating next tab elif virtual_key == 124 and glpyh == 101: key = "<right>" # Noticed in Google Chrome navigating previous tab elif virtual_key == 123 and glpyh == 100: key = "<left>" # List application in a window to Force Quit elif virtual_key == 53 and glpyh == 27: key = "<escape>" # FIXME: # * Instruments Menu View->Run Browser # modifiers==12 virtual_key==48 glpyh==2 # * Terminal Menu Edit->Start Dictation # fn fn - glpyh==148 modifiers==24 # * Menu Chrome->Clear Browsing Data in Google Chrome # virtual_key==51 glpyh==23 [Delete Left (like Backspace on a PC)] if not key: raise LdtpServerException("No access key associated") return modifiers_type + key
python
{ "resource": "" }
q262278
Clipboard.clearContents
validation
def clearContents(cls): """Clear contents of general pasteboard. Future enhancement can include specifying which clipboard to clear Returns: True on success; caller should expect to catch exceptions, probably from AppKit (ValueError) """ log_msg = 'Request to clear contents of pasteboard: general' logging.debug(log_msg) pb = AppKit.NSPasteboard.generalPasteboard() pb.clearContents() return True
python
{ "resource": "" }
q262279
Clipboard.isEmpty
validation
def isEmpty(cls, datatype=None): """Method to test if the general pasteboard is empty or not with respect to the type of object you want. Parameters: datatype (defaults to strings) Returns: Boolean True (empty) / False (has contents); Raises exception (passes any raised up) """ if not datatype: datatype = AppKit.NSString if not isinstance(datatype, types.ListType): datatype = [datatype] pp = pprint.PrettyPrinter() logging.debug('Desired datatypes: %s' % pp.pformat(datatype)) opt_dict = {} logging.debug('Results filter is: %s' % pp.pformat(opt_dict)) try: log_msg = 'Request to verify pasteboard is empty' logging.debug(log_msg) pb = AppKit.NSPasteboard.generalPasteboard() # canReadObjectForClasses_options_() seems to return an int (> 0 if # True) # Need to negate to get the sense we want (True if can not read the # data type from the pasteboard) its_empty = not bool(pb.canReadObjectForClasses_options_(datatype, opt_dict)) except ValueError as error: logging.error(error) raise return bool(its_empty)
python
{ "resource": "" }
q262280
Utils._ldtpize_accessible
validation
def _ldtpize_accessible(self, acc): """ Get LDTP format accessibile name @param acc: Accessible handle @type acc: object @return: object type, stripped object name (associated / direct), associated label @rtype: tuple """ actual_role = self._get_role(acc) label = self._get_title(acc) if re.match("AXWindow", actual_role, re.M | re.U | re.L): # Strip space and new line from window title strip = r"( |\n)" else: # Strip space, colon, dot, underscore and new line from # all other object types strip = r"( |:|\.|_|\n)" if label: # Return the role type (if, not in the know list of roles, # return ukn - unknown), strip the above characters from name # also return labely_by string label = re.sub(strip, u"", label) role = abbreviated_roles.get(actual_role, "ukn") if self._ldtp_debug and role == "ukn": print(actual_role, acc) return role, label
python
{ "resource": "" }
q262281
Utils._glob_match
validation
def _glob_match(self, pattern, string): """ Match given string, by escaping regex characters """ # regex flags Multi-line, Unicode, Locale return bool(re.match(fnmatch.translate(pattern), string, re.M | re.U | re.L))
python
{ "resource": "" }
q262282
Menu.doesmenuitemexist
validation
def doesmenuitemexist(self, window_name, object_name): """ Check a menu item exist. @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @param strict_hierarchy: Mandate menu hierarchy if set to True @type object_name: boolean @return: 1 on success. @rtype: integer """ try: menu_handle = self._get_menu_handle(window_name, object_name, False) return 1 except LdtpServerException: return 0
python
{ "resource": "" }
q262283
Menu.menuitemenabled
validation
def menuitemenabled(self, window_name, object_name): """ Verify a menu item is enabled @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @return: 1 on success. @rtype: integer """ try: menu_handle = self._get_menu_handle(window_name, object_name, False) if menu_handle.AXEnabled: return 1 except LdtpServerException: pass return 0
python
{ "resource": "" }
q262284
Menu.verifymenucheck
validation
def verifymenucheck(self, window_name, object_name): """ Verify a menu item is checked @param window_name: Window name to look for, either full name, LDTP's name convention, or a Unix glob. @type window_name: string @param object_name: Object name to look for, either full name, LDTP's name convention, or a Unix glob. Or menu heirarchy @type object_name: string @return: 1 on success. @rtype: integer """ try: menu_handle = self._get_menu_handle(window_name, object_name, False) try: if menu_handle.AXMenuItemMarkChar: # Checked return 1 except atomac._a11y.Error: pass except LdtpServerException: pass return 0
python
{ "resource": "" }
q262285
BaseAXUIElement._getRunningApps
validation
def _getRunningApps(cls): """Get a list of the running applications.""" def runLoopAndExit(): AppHelper.stopEventLoop() AppHelper.callLater(1, runLoopAndExit) AppHelper.runConsoleEventLoop() # Get a list of running applications ws = AppKit.NSWorkspace.sharedWorkspace() apps = ws.runningApplications() return apps
python
{ "resource": "" }
q262286
BaseAXUIElement.getFrontmostApp
validation
def getFrontmostApp(cls): """Get the current frontmost application. Raise a ValueError exception if no GUI applications are found. """ # Refresh the runningApplications list apps = cls._getRunningApps() for app in apps: pid = app.processIdentifier() ref = cls.getAppRefByPid(pid) try: if ref.AXFrontmost: return ref except (_a11y.ErrorUnsupported, _a11y.ErrorCannotComplete, _a11y.ErrorAPIDisabled, _a11y.ErrorNotImplemented): # Some applications do not have an explicit GUI # and so will not have an AXFrontmost attribute # Trying to read attributes from Google Chrome Helper returns # ErrorAPIDisabled for some reason - opened radar bug 12837995 pass raise ValueError('No GUI application found.')
python
{ "resource": "" }
q262287
BaseAXUIElement.getAnyAppWithWindow
validation
def getAnyAppWithWindow(cls): """Get a random app that has windows. Raise a ValueError exception if no GUI applications are found. """ # Refresh the runningApplications list apps = cls._getRunningApps() for app in apps: pid = app.processIdentifier() ref = cls.getAppRefByPid(pid) if hasattr(ref, 'windows') and len(ref.windows()) > 0: return ref raise ValueError('No GUI application found.')
python
{ "resource": "" }
q262288
BaseAXUIElement.launchAppByBundleId
validation
def launchAppByBundleId(bundleID): """Launch the application with the specified bundle ID""" # NSWorkspaceLaunchAllowingClassicStartup does nothing on any # modern system that doesn't have the classic environment installed. # Encountered a bug when passing 0 for no options on 10.6 PyObjC. ws = AppKit.NSWorkspace.sharedWorkspace() # Sorry about the length of the following line r = ws.launchAppWithBundleIdentifier_options_additionalEventParamDescriptor_launchIdentifier_( bundleID, AppKit.NSWorkspaceLaunchAllowingClassicStartup, AppKit.NSAppleEventDescriptor.nullDescriptor(), None) # On 10.6, this returns a tuple - first element bool result, second is # a number. Let's use the bool result. if not r[0]: raise RuntimeError('Error launching specified application.')
python
{ "resource": "" }
q262289
BaseAXUIElement.launchAppByBundlePath
validation
def launchAppByBundlePath(bundlePath, arguments=None): """Launch app with a given bundle path. Return True if succeed. """ if arguments is None: arguments = [] bundleUrl = NSURL.fileURLWithPath_(bundlePath) workspace = AppKit.NSWorkspace.sharedWorkspace() arguments_strings = list(map(lambda a: NSString.stringWithString_(str(a)), arguments)) arguments = NSDictionary.dictionaryWithDictionary_({ AppKit.NSWorkspaceLaunchConfigurationArguments: NSArray.arrayWithArray_( arguments_strings) }) return workspace.launchApplicationAtURL_options_configuration_error_( bundleUrl, AppKit.NSWorkspaceLaunchAllowingClassicStartup, arguments, None)
python
{ "resource": "" }
q262290
BaseAXUIElement.terminateAppByBundleId
validation
def terminateAppByBundleId(bundleID): """Terminate app with a given bundle ID. Requires 10.6. Return True if succeed. """ ra = AppKit.NSRunningApplication if getattr(ra, "runningApplicationsWithBundleIdentifier_"): appList = ra.runningApplicationsWithBundleIdentifier_(bundleID) if appList and len(appList) > 0: app = appList[0] return app and app.terminate() and True or False return False
python
{ "resource": "" }
q262291
BaseAXUIElement._queueEvent
validation
def _queueEvent(self, event, args): """Private method to queue events to run. Each event in queue is a tuple (event call, args to event call). """ if not hasattr(self, 'eventList'): self.eventList = deque([(event, args)]) return self.eventList.append((event, args))
python
{ "resource": "" }
q262292
BaseAXUIElement._addKeyToQueue
validation
def _addKeyToQueue(self, keychr, modFlags=0, globally=False): """Add keypress to queue. Parameters: key character or constant referring to a non-alpha-numeric key (e.g. RETURN or TAB) modifiers global or app specific Returns: None or raise ValueError exception. """ # Awkward, but makes modifier-key-only combinations possible # (since sendKeyWithModifiers() calls this) if not keychr: return if not hasattr(self, 'keyboard'): self.keyboard = AXKeyboard.loadKeyboard() if keychr in self.keyboard['upperSymbols'] and not modFlags: self._sendKeyWithModifiers(keychr, [AXKeyCodeConstants.SHIFT], globally) return if keychr.isupper() and not modFlags: self._sendKeyWithModifiers( keychr.lower(), [AXKeyCodeConstants.SHIFT], globally ) return if keychr not in self.keyboard: self._clearEventQueue() raise ValueError('Key %s not found in keyboard layout' % keychr) # Press the key keyDown = Quartz.CGEventCreateKeyboardEvent(None, self.keyboard[keychr], True) # Release the key keyUp = Quartz.CGEventCreateKeyboardEvent(None, self.keyboard[keychr], False) # Set modflags on keyDown (default None): Quartz.CGEventSetFlags(keyDown, modFlags) # Set modflags on keyUp: Quartz.CGEventSetFlags(keyUp, modFlags) # Post the event to the given app if not globally: # To direct output to the correct application need the PSN (macOS <=10.10) or PID(macOS > 10.10): macVer, _, _ = platform.mac_ver() macVer = int(macVer.split('.')[1]) if macVer > 10: appPid = self._getPid() self._queueEvent(Quartz.CGEventPostToPid, (appPid, keyDown)) self._queueEvent(Quartz.CGEventPostToPid, (appPid, keyUp)) else: appPsn = self._getPsnForPid(self._getPid()) self._queueEvent(Quartz.CGEventPostToPSN, (appPsn, keyDown)) self._queueEvent(Quartz.CGEventPostToPSN, (appPsn, keyUp)) else: self._queueEvent(Quartz.CGEventPost, (0, keyDown)) self._queueEvent(Quartz.CGEventPost, (0, keyUp))
python
{ "resource": "" }
q262293
BaseAXUIElement._sendKey
validation
def _sendKey(self, keychr, modFlags=0, globally=False): """Send one character with no modifiers. Parameters: key character or constant referring to a non-alpha-numeric key (e.g. RETURN or TAB) modifier flags, global or app specific Returns: None or raise ValueError exception """ escapedChrs = { '\n': AXKeyCodeConstants.RETURN, '\r': AXKeyCodeConstants.RETURN, '\t': AXKeyCodeConstants.TAB, } if keychr in escapedChrs: keychr = escapedChrs[keychr] self._addKeyToQueue(keychr, modFlags, globally=globally) self._postQueuedEvents()
python
{ "resource": "" }
q262294
BaseAXUIElement._isSingleCharacter
validation
def _isSingleCharacter(keychr): """Check whether given keyboard character is a single character. Parameters: key character which will be checked. Returns: True when given key character is a single character. """ if not keychr: return False # Regular character case. if len(keychr) == 1: return True # Tagged character case. return keychr.count('<') == 1 and keychr.count('>') == 1 and \ keychr[0] == '<' and keychr[-1] == '>'
python
{ "resource": "" }
q262295
BaseAXUIElement._sendKeyWithModifiers
validation
def _sendKeyWithModifiers(self, keychr, modifiers, globally=False): """Send one character with the given modifiers pressed. Parameters: key character, list of modifiers, global or app specific Returns: None or raise ValueError exception """ if not self._isSingleCharacter(keychr): raise ValueError('Please provide only one character to send') if not hasattr(self, 'keyboard'): self.keyboard = AXKeyboard.loadKeyboard() modFlags = self._pressModifiers(modifiers, globally=globally) # Press the non-modifier key self._sendKey(keychr, modFlags, globally=globally) # Release the modifiers self._releaseModifiers(modifiers, globally=globally) # Post the queued keypresses: self._postQueuedEvents()
python
{ "resource": "" }
q262296
BaseAXUIElement._queueMouseButton
validation
def _queueMouseButton(self, coord, mouseButton, modFlags, clickCount=1, dest_coord=None): """Private method to handle generic mouse button clicking. Parameters: coord (x, y) to click, mouseButton (e.g., kCGMouseButtonLeft), modFlags set (int) Optional: clickCount (default 1; set to 2 for double-click; 3 for triple-click on host) Returns: None """ # For now allow only left and right mouse buttons: mouseButtons = { Quartz.kCGMouseButtonLeft: 'LeftMouse', Quartz.kCGMouseButtonRight: 'RightMouse', } if mouseButton not in mouseButtons: raise ValueError('Mouse button given not recognized') eventButtonDown = getattr(Quartz, 'kCGEvent%sDown' % mouseButtons[mouseButton]) eventButtonUp = getattr(Quartz, 'kCGEvent%sUp' % mouseButtons[mouseButton]) eventButtonDragged = getattr(Quartz, 'kCGEvent%sDragged' % mouseButtons[ mouseButton]) # Press the button buttonDown = Quartz.CGEventCreateMouseEvent(None, eventButtonDown, coord, mouseButton) # Set modflags (default None) on button down: Quartz.CGEventSetFlags(buttonDown, modFlags) # Set the click count on button down: Quartz.CGEventSetIntegerValueField(buttonDown, Quartz.kCGMouseEventClickState, int(clickCount)) if dest_coord: # Drag and release the button buttonDragged = Quartz.CGEventCreateMouseEvent(None, eventButtonDragged, dest_coord, mouseButton) # Set modflags on the button dragged: Quartz.CGEventSetFlags(buttonDragged, modFlags) buttonUp = Quartz.CGEventCreateMouseEvent(None, eventButtonUp, dest_coord, mouseButton) else: # Release the button buttonUp = Quartz.CGEventCreateMouseEvent(None, eventButtonUp, coord, mouseButton) # Set modflags on the button up: Quartz.CGEventSetFlags(buttonUp, modFlags) # Set the click count on button up: Quartz.CGEventSetIntegerValueField(buttonUp, Quartz.kCGMouseEventClickState, int(clickCount)) # Queue the events self._queueEvent(Quartz.CGEventPost, (Quartz.kCGSessionEventTap, buttonDown)) if dest_coord: self._queueEvent(Quartz.CGEventPost, (Quartz.kCGHIDEventTap, buttonDragged)) self._queueEvent(Quartz.CGEventPost, (Quartz.kCGSessionEventTap, buttonUp))
python
{ "resource": "" }
q262297
BaseAXUIElement._leftMouseDragged
validation
def _leftMouseDragged(self, stopCoord, strCoord, speed): """Private method to handle generic mouse left button dragging and dropping. Parameters: stopCoord(x,y) drop point Optional: strCoord (x, y) drag point, default (0,0) get current mouse position speed (int) 1 to unlimit, simulate mouse moving action from some special requirement Returns: None """ # To direct output to the correct application need the PSN: appPid = self._getPid() # Get current position as start point if strCoord not given if strCoord == (0, 0): loc = AppKit.NSEvent.mouseLocation() strCoord = (loc.x, Quartz.CGDisplayPixelsHigh(0) - loc.y) # To direct output to the correct application need the PSN: appPid = self._getPid() # Press left button down pressLeftButton = Quartz.CGEventCreateMouseEvent( None, Quartz.kCGEventLeftMouseDown, strCoord, Quartz.kCGMouseButtonLeft ) # Queue the events Quartz.CGEventPost(Quartz.CoreGraphics.kCGHIDEventTap, pressLeftButton) # Wait for reponse of system, a fuzzy icon appears time.sleep(5) # Simulate mouse moving speed, k is slope speed = round(1 / float(speed), 2) xmoved = stopCoord[0] - strCoord[0] ymoved = stopCoord[1] - strCoord[1] if ymoved == 0: raise ValueError('Not support horizontal moving') else: k = abs(ymoved / xmoved) if xmoved != 0: for xpos in range(int(abs(xmoved))): if xmoved > 0 and ymoved > 0: currcoord = (strCoord[0] + xpos, strCoord[1] + xpos * k) elif xmoved > 0 and ymoved < 0: currcoord = (strCoord[0] + xpos, strCoord[1] - xpos * k) elif xmoved < 0 and ymoved < 0: currcoord = (strCoord[0] - xpos, strCoord[1] - xpos * k) elif xmoved < 0 and ymoved > 0: currcoord = (strCoord[0] - xpos, strCoord[1] + xpos * k) # Drag with left button dragLeftButton = Quartz.CGEventCreateMouseEvent( None, Quartz.kCGEventLeftMouseDragged, currcoord, Quartz.kCGMouseButtonLeft ) Quartz.CGEventPost(Quartz.CoreGraphics.kCGHIDEventTap, dragLeftButton) # Wait for reponse of system time.sleep(speed) else: raise ValueError('Not support vertical moving') upLeftButton = Quartz.CGEventCreateMouseEvent( None, Quartz.kCGEventLeftMouseUp, stopCoord, Quartz.kCGMouseButtonLeft ) # Wait for reponse of system, a plus icon appears time.sleep(5) # Up left button up Quartz.CGEventPost(Quartz.CoreGraphics.kCGHIDEventTap, upLeftButton)
python
{ "resource": "" }
q262298
BaseAXUIElement._waitFor
validation
def _waitFor(self, timeout, notification, **kwargs): """Wait for a particular UI event to occur; this can be built upon in NativeUIElement for specific convenience methods. """ callback = self._matchOther retelem = None callbackArgs = None callbackKwargs = None # Allow customization of the callback, though by default use the basic # _match() method if 'callback' in kwargs: callback = kwargs['callback'] del kwargs['callback'] # Deal with these only if callback is provided: if 'args' in kwargs: if not isinstance(kwargs['args'], tuple): errStr = 'Notification callback args not given as a tuple' raise TypeError(errStr) # If args are given, notification will pass back the returned # element in the first positional arg callbackArgs = kwargs['args'] del kwargs['args'] if 'kwargs' in kwargs: if not isinstance(kwargs['kwargs'], dict): errStr = 'Notification callback kwargs not given as a dict' raise TypeError(errStr) callbackKwargs = kwargs['kwargs'] del kwargs['kwargs'] # If kwargs are not given as a dictionary but individually listed # need to update the callbackKwargs dict with the remaining items in # kwargs if kwargs: if callbackKwargs: callbackKwargs.update(kwargs) else: callbackKwargs = kwargs else: callbackArgs = (retelem,) # Pass the kwargs to the default callback callbackKwargs = kwargs return self._setNotification(timeout, notification, callback, callbackArgs, callbackKwargs)
python
{ "resource": "" }
q262299
BaseAXUIElement._getActions
validation
def _getActions(self): """Retrieve a list of actions supported by the object.""" actions = _a11y.AXUIElement._getActions(self) # strip leading AX from actions - help distinguish them from attributes return [action[2:] for action in actions]
python
{ "resource": "" }