code
stringlengths
51
2.34k
docstring
stringlengths
11
171
def iter_rows(self, start=None, end=None): start = start or 0 end = end or self.nrows for i in range(start, end): yield self.iloc[i, :]
Iterate each of the Region rows in this region
def debug(self, key): return (not self.quiet and not self.debug_none and (self.debug_all or getattr(self, "debug_%s" % key)))
Returns True if the debug setting is enabled.
def logs(self): for record in self.history: if (record["when"] >= self.options.since.date and record["when"] < self.options.until.date): for change in record["changes"]: yield record["who"], change
Return relevant who-did-what pairs from the bug history
def _remap_tag_to_tag_id(cls, tag, new_data): try: value = new_data[tag] except: return tag_id = tag + '_id' try: new_data[tag_id] = value['id'] except: try: new_data[tag_id] = value.id except AttributeError: new_data[tag_id] = value del new_data[tag]
Remaps a given changed field from tag to tag_id.
def task_detail(job_id=None, task_name=None): jobs = get_jobs() job = [job for job in jobs if str(job['job_id']) == job_id][0] return render_template('task_detail.html', job=job, task_name=task_name, task=[task for task in job['tasks'] if task['name'] == task_name][0])
Show a detailed description of a specific task.
def run(self, matches): def _run(matches): group_starting_pos = 0 for current_pos, (group_length, group_function) in enumerate(zip(self.group_lengths, self.group_functions)): start_pos = current_pos + group_starting_pos end_pos = current_pos + group_starting_pos + group_length yield group_function(matches[start_pos:end_pos]) group_starting_pos += group_length - 1 return self.final_function(list(_run(matches)))
Run group functions over matches
def value_for_key(membersuite_object_data, key): key_value_dicts = { d['Key']: d['Value'] for d in membersuite_object_data["Fields"]["KeyValueOfstringanyType"]} return key_value_dicts[key]
Return the value for `key` of membersuite_object_data.
def rotateX(self, angle): rad = angle * math.pi / 180 cosa = math.cos(rad) sina = math.sin(rad) y = self.y * cosa - self.z * sina z = self.y * sina + self.z * cosa return Point3D(self.x, y, z)
Rotates the point around the X axis by the given angle in degrees.
def send_invitation(request, project_id=None): project = None if project_id is not None: project = get_object_or_404(Project, id=project_id) if project is None: if not is_admin(request): return HttpResponseForbidden('<h1>Access Denied</h1>') else: if not project.can_edit(request): return HttpResponseForbidden('<h1>Access Denied</h1>') return _send_invitation(request, project)
The logged in project leader wants to invite somebody to their project.
def iter_components(self): names = self.list_components() for name in names: yield self.get_component(name)
Iterate over all defined components yielding IOTile objects.
def CollectionItemToClientPath(item, client_id=None): if isinstance(item, rdf_flows.GrrMessage): client_id = item.source item = item.payload elif isinstance(item, rdf_flow_objects.FlowResult): client_id = item.client_id item = item.payload if client_id is None: raise ValueError("Could not determine client_id.") elif isinstance(client_id, rdfvalue.RDFURN): client_id = client_id.Basename() if isinstance(item, rdf_client_fs.StatEntry): return db.ClientPath.FromPathSpec(client_id, item.pathspec) elif isinstance(item, rdf_file_finder.FileFinderResult): return db.ClientPath.FromPathSpec(client_id, item.stat_entry.pathspec) elif isinstance(item, collectors.ArtifactFilesDownloaderResult): if item.HasField("downloaded_file"): return db.ClientPath.FromPathSpec(client_id, item.downloaded_file.pathspec) raise ItemNotExportableError(item)
Converts given RDFValue to a ClientPath of a file to be downloaded.
def choices(klass): _choices = [] for attr in user_attributes(klass.Meta): val = getattr(klass.Meta, attr) setattr(klass, attr, val[0]) _choices.append((val[0], val[1])) setattr(klass, 'CHOICES', tuple(_choices)) return klass
Decorator to set `CHOICES` and other attributes
def _ratio_scores(parameters_value, clusteringmodel_gmm_good, clusteringmodel_gmm_bad): ratio = clusteringmodel_gmm_good.score([parameters_value]) / clusteringmodel_gmm_bad.score([parameters_value]) sigma = 0 return ratio, sigma
The ratio is smaller the better
def check_cups_allowed(func): @wraps(func) def decorator(*args, **kwargs): cups = kwargs.get('cups') if (cups and current_user.is_authenticated() and not current_user.allowed(cups, 'cups')): return current_app.login_manager.unauthorized() return func(*args, **kwargs) return decorator
Check if CUPS is allowd by token
def Flow(self, flow_id): return flow.FlowRef( client_id=self.client_id, flow_id=flow_id, context=self._context)
Return a reference to a flow with a given id on this client.
def _validate_args(self, args): assert(args.bucket) if args.subscribers: for _subscriber in args.subscribers: assert(isinstance(_subscriber, AsperaBaseSubscriber)) if (args.transfer_config): assert(isinstance(args.transfer_config, AsperaConfig)) if args.transfer_config.multi_session > self._config.ascp_max_concurrent: raise ValueError("Max sessions is %d" % self._config.ascp_max_concurrent) for _pair in args.file_pair_list: if not _pair.key or not _pair.fileobj: raise ValueError("Invalid file pair")
validate the user arguments
def spawn_managed_host(config_file, manager, connect_on_start=True): data = manager.request_host_status(config_file) is_running = data['started'] if is_running: host_status = json.loads(data['host']['output']) logfile = data['host']['logfile'] else: data = manager.start_host(config_file) host_status = json.loads(data['output']) logfile = data['logfile'] host = JSHost( status=host_status, logfile=logfile, config_file=config_file, manager=manager ) if not is_running and settings.VERBOSITY >= verbosity.PROCESS_START: print('Started {}'.format(host.get_name())) if connect_on_start: host.connect() return host
Spawns a managed host, if it is not already running
def input_to_phase(data, rate, data_type): if data_type == "phase": return data elif data_type == "freq": return frequency2phase(data, rate) else: raise Exception("unknown data_type: " + data_type)
Take either phase or frequency as input and return phase
def setnx(self, key, value): return self.set(key, value, nx=True)
Set the value of ``key`` to ``value`` if key doesn't exist
def close(self): close_command = StandardSend(self._address, COMMAND_LIGHT_OFF_0X13_0X00) self._send_method(close_command, self._close_message_received)
Send CLOSE command to device.
def path_exists(self, path, follow_symlinks=True): "test if path exists" return ( self.file_exists(path, follow_symlinks) or self.symlink_exists(path, follow_symlinks) or self.directory_exists(path, follow_symlinks) )
test if path exists
def end_request(self): if not self._chunked: return trailers = [(n, get_header(self._headers, n)) for n in self._trailer] \ if self._trailer else None ending = create_chunked_body_end(trailers) self._protocol.writer.write(ending)
End the request body.
def diffpow(self, x, rot=0): N = len(x) if rot: x = rotate(x) return sum(np.abs(x)**(2. + 4.*np.arange(N) / (N - 1.)))**0.5
Diffpow test objective function
def postprocessor(prediction): prediction = prediction.data.numpy()[0] top_predictions = prediction.argsort()[-3:][::-1] return [labels[prediction] for prediction in top_predictions]
Map prediction tensor to labels.
def order_module_dependencies(modules, parser): result = [] for modk in modules: if modk not in result: result.append(modk) recursed = list(result) for i in range(len(result)): module = result[i] _process_module_order(parser, module, i, recursed) return recursed
Orders the specified list of modules based on their inter-dependencies.
def remove_spin(self): for site in self.sites: new_sp = collections.defaultdict(float) for sp, occu in site.species.items(): oxi_state = getattr(sp, "oxi_state", None) new_sp[Specie(sp.symbol, oxidation_state=oxi_state)] += occu site.species = new_sp
Removes spin states from a structure.
def find_lt(a, x): i = bisect.bisect_left(a, x) if i: return a[i-1] raise ValueError
Find rightmost value less than x
def build_tree(self, name, props, resource_name=None): n = Node(name, props, resource_name) prop_type_list = self._get_type_list(props) if not prop_type_list: return n prop_type_list = sorted(prop_type_list) for prop_name in prop_type_list: if prop_name == 'Tag': continue child = self.build_tree(prop_name, self.properties[prop_name]) if child is not None: n.add_child(child) return n
Build a tree of non-primitive typed dependency order.
def after_scenario(context, scenario): os.chdir(context.old_cwd) context.directory.cleanup()
Leave the environment fresh after each scenario.
async def list(source): result = [] async with streamcontext(source) as streamer: async for item in streamer: result.append(item) yield result
Generate a single list from an asynchronous sequence.
def unpause_trial(self, trial): assert trial.status == Trial.PAUSED, trial.status self.set_status(trial, Trial.PENDING)
Sets PAUSED trial to pending to allow scheduler to start.
def _wmi_to_ts(self, wmi_ts): year, month, day, hour, minute, second, microsecond, tz = to_time(wmi_ts) tz_delta = timedelta(minutes=int(tz)) if '+' in wmi_ts: tz_delta = -tz_delta dt = ( datetime(year=year, month=month, day=day, hour=hour, minute=minute, second=second, microsecond=microsecond) + tz_delta ) return int(calendar.timegm(dt.timetuple()))
Convert a wmi formatted timestamp into an epoch.
def find_link(cls, plot, link=None): registry = Link.registry.items() for source in plot.link_sources: if link is None: links = [ l for src, links in registry for l in links if src is source or (src._plot_id is not None and src._plot_id == source._plot_id)] if links: return (plot, links) else: if ((link.target is source) or (link.target is not None and link.target._plot_id is not None and link.target._plot_id == source._plot_id)): return (plot, [link])
Searches a GenericElementPlot for a Link.
def identify(self, obj): hash_ = hash(obj) type_id = self.to_id(type(obj)) return (hash_, TypeId(type_id))
Return an Ident-shaped tuple for the given object.
def umi_cycles(self) -> int: return sum((int(re.sub(r'\D', '', op)) for op in self.umi_tokens))
The number of cycles dedicated to UMI.
def read(filename, loader=None, implicit_tuple=True, allow_errors=False): with open(filename, 'r') as f: return reads(f.read(), filename=filename, loader=loader, implicit_tuple=implicit_tuple, allow_errors=allow_errors)
Load but don't evaluate a GCL expression from a file.
def sort_candidate_pairs(pairs, ice_controlling): def pair_priority(pair): return -candidate_pair_priority(pair.local_candidate, pair.remote_candidate, ice_controlling) pairs.sort(key=pair_priority)
Sort a list of candidate pairs.
def finite_datetimes(self, finite_start, finite_stop): date_start = datetime(finite_start.year, finite_start.month, finite_start.day) dates = [] for i in itertools.count(): t = date_start + timedelta(days=i) if t >= finite_stop: return dates if t >= finite_start: dates.append(t)
Simply returns the points in time that correspond to turn of day.
def info(self, i: int=None) -> str: head = "[" + colors.blue("info") + "]" if i is not None: head = str(i) + " " + head return head
Returns an info message
def _add_custom_headers(self, dct): if self.client_id is None: self.client_id = os.environ.get("CLOUD_QUEUES_ID") if self.client_id: dct["Client-ID"] = self.client_id
Add the Client-ID header required by Cloud Queues
def models_of_config(config): resources = resources_of_config(config) models = [] for resource in resources: if not hasattr(resource, '__table__') and hasattr(resource, 'model'): models.append(resource.model) else: models.append(resource) return models
Return list of models from all resources in config.
def run(host: Optional[str] = None, port: Optional[int] = None, *args, **kwargs) -> None: get_bot().run(host=host, port=port, *args, **kwargs)
Run the NoneBot instance.
def cast_to_swimlane(self, value): value = super(ListField, self).cast_to_swimlane(value) if not value: return None value_ids = deepcopy(self._initial_value_to_ids_map) return [self._build_list_item(item, value_ids[item].pop(0) if value_ids[item] else None) for item in value]
Restore swimlane format, attempting to keep initial IDs for any previously existing values
def generateProjectFiles(self, dir=os.getcwd(), args=[]): if os.path.exists(os.path.join(dir, 'Source')) == False: Utility.printStderr('Pure Blueprint project, nothing to generate project files for.') return genScript = self.getGenerateScript() projectFile = self.getProjectDescriptor(dir) Utility.run([genScript, '-project=' + projectFile, '-game', '-engine'] + args, cwd=os.path.dirname(genScript), raiseOnError=True)
Generates IDE project files for the Unreal project in the specified directory
def transcribe(decoder, audio_file, libdir=None): decoder = get_decoder() decoder.start_utt() stream = open(audio_file, 'rb') while True: buf = stream.read(1024) if buf: decoder.process_raw(buf, False, False) else: break decoder.end_utt() return evaluate_results(decoder)
Decode streaming audio data from raw binary file on disk.
def _strip(self, tokens, criteria, prop, rev=False): num = len(tokens) res = [] for i, token in enumerate(reversed(tokens) if rev else tokens): if criteria(token) and num > i + 1: res.insert(0, tokens.pop()) if rev else res.append(tokens.pop(0)) else: break if res: self[prop] = self._clean(' '.join(res)) return tokens
Strip off contiguous tokens from the start or end of the list that meet the criteria.
def annotate_diamond(fasta_path: 'path to fasta input', diamond_path: 'path to Diamond taxonomic classification output'): records = tictax.parse_seqs(fasta_path) annotated_records = tictax.annotate_diamond(records, diamond_path) SeqIO.write(annotated_records, sys.stdout, 'fasta')
Annotate fasta headers with taxonomy information from Diamond
def convert_timestamp(timestamp): timestamp = timestamp.strip() chunk, millis = timestamp.split(',') hours, minutes, seconds = chunk.split(':') hours = int(hours) minutes = int(minutes) seconds = int(seconds) seconds = seconds + hours * 60 * 60 + minutes * 60 + float(millis) / 1000 return seconds
Convert an srt timestamp into seconds.
def _parse_date_greek(dateString): m = _greek_date_format_re.match(dateString) if not m: return wday = _greek_wdays[m.group(1)] month = _greek_months[m.group(3)] rfc822date = '%(wday)s, %(day)s %(month)s %(year)s %(hour)s:%(minute)s:%(second)s %(zonediff)s' % \ {'wday': wday, 'day': m.group(2), 'month': month, 'year': m.group(4),\ 'hour': m.group(5), 'minute': m.group(6), 'second': m.group(7),\ 'zonediff': m.group(8)} return _parse_date_rfc822(rfc822date)
Parse a string according to a Greek 8-bit date format.
def from_list(cls, database, key, data, clear=False): arr = cls(database, key) if clear: arr.clear() arr.extend(data) return arr
Create and populate an Array object from a data dictionary.
def load(self, filename=None): DataFile.load(self, filename) self.spectrum.filename = filename
Method was overriden to set spectrum.filename as well
def selecttrue(table, field, complement=False): return select(table, field, lambda v: bool(v), complement=complement)
Select rows where the given field evaluates `True`.
def catch_timeout(f): def new_f(self, *args, **kwargs): try: return f(self, *args, **kwargs) except (requests.exceptions.ReadTimeout, requests.packages.urllib3.exceptions.ReadTimeoutError) as e: log.warning("caught read timeout: %s", e) self.connect() return f(self, *args, **kwargs) return new_f
A decorator to handle read timeouts from Twitter.
def find_target_migration_file(database=DEFAULT_DB_ALIAS, changelog_file=None): if not database: database = DEFAULT_DB_ALIAS if not changelog_file: changelog_file = get_changelog_file_for_database(database) try: doc = minidom.parse(changelog_file) except ExpatError as ex: raise InvalidChangelogFile( 'Could not parse XML file %s: %s' % (changelog_file, ex)) try: dbchglog = doc.getElementsByTagName('databaseChangeLog')[0] except IndexError: raise InvalidChangelogFile( 'Missing <databaseChangeLog> node in file %s' % ( changelog_file)) else: nodes = list(filter(lambda x: x.nodeType is x.ELEMENT_NODE, dbchglog.childNodes)) if not nodes: return changelog_file last_node = nodes[-1] if last_node.tagName == 'include': last_file = last_node.attributes.get('file').firstChild.data return find_target_migration_file( database=database, changelog_file=last_file) else: return changelog_file
Finds best matching target migration file
def mavset(self, mav, name, value, retries=3): got_ack = False while retries > 0 and not got_ack: retries -= 1 mav.param_set_send(name.upper(), float(value)) tstart = time.time() while time.time() - tstart < 1: ack = mav.recv_match(type='PARAM_VALUE', blocking=False) if ack == None: time.sleep(0.1) continue if str(name).upper() == str(ack.param_id).upper(): got_ack = True self.__setitem__(name, float(value)) break if not got_ack: print("timeout setting %s to %f" % (name, float(value))) return False return True
set a parameter on a mavlink connection
def fail(self, cmd, title=None, message=None): if message is None: message = self.handle_exc() else: message = escape(message) self.db.send( 'Echo|%s' % dump({ 'for': escape(title or '%s failed' % cmd), 'val': message }) )
Send back captured exceptions
def hasBidAsk(self) -> bool: return ( self.bid != -1 and not isNan(self.bid) and self.bidSize > 0 and self.ask != -1 and not isNan(self.ask) and self.askSize > 0)
See if this ticker has a valid bid and ask.
def config(conf, confdefs): conf = conf.copy() for name, info in confdefs: conf.setdefault(name, info.get('defval')) return conf
Initialize a config dict using the given confdef tuples.
def _get_alignment(self, pys_style): alignment_styles = ["justification", "vertical_align", "angle"] if not any(astyle in pys_style for astyle in alignment_styles): return def angle2xfrotation(angle): if 0 <= angle <= 90: return angle elif -90 <= angle < 0: return 90 - angle return 0 justification2xfalign = { "left": 1, "center": 2, "right": 3, } vertical_align2xfalign = { "top": 0, "middle": 1, "bottom": 2, } alignment = xlwt.Alignment() try: alignment.horz = justification2xfalign[pys_style["justification"]] except KeyError: pass try: alignment.vert = \ vertical_align2xfalign[pys_style["vertical_align"]] except KeyError: pass try: alignment.rota = angle2xfrotation(pys_style["angle"]) except KeyError: pass return alignment
Returns xlwt.Alignment for pyspread style
def envelope(**kwargs): e_oaipmh = Element(etree.QName(NS_OAIPMH, 'OAI-PMH'), nsmap=NSMAP) e_oaipmh.set(etree.QName(NS_XSI, 'schemaLocation'), '{0} {1}'.format(NS_OAIPMH, NS_OAIPMH_XSD)) e_tree = ElementTree(element=e_oaipmh) if current_app.config['OAISERVER_XSL_URL']: e_oaipmh.addprevious(etree.ProcessingInstruction( 'xml-stylesheet', 'type="text/xsl" href="{0}"' .format(current_app.config['OAISERVER_XSL_URL']))) e_responseDate = SubElement( e_oaipmh, etree.QName( NS_OAIPMH, 'responseDate')) e_responseDate.text = datetime_to_datestamp(datetime.utcnow()) e_request = SubElement(e_oaipmh, etree.QName(NS_OAIPMH, 'request')) for key, value in kwargs.items(): if key == 'from_' or key == 'until': value = datetime_to_datestamp(value) elif key == 'resumptionToken': value = value['token'] e_request.set(key, value) e_request.text = url_for('invenio_oaiserver.response', _external=True) return e_tree, e_oaipmh
Create OAI-PMH envelope for response.
def tci_path(self): tci_paths = [ path for path in self.dataset._product_metadata.xpath( ".//Granule[@granuleIdentifier='%s']/IMAGE_FILE/text()" % self.granule_identifier ) if path.endswith('TCI') ] try: tci_path = tci_paths[0] except IndexError: return None return os.path.join( self.dataset._zip_root if self.dataset.is_zip else self.dataset.path, tci_path ) + '.jp2'
Return the path to the granules TrueColorImage.
def format_maxtime(self, maxtime): minutes = maxtime.get("minutes", "0") hours = maxtime.get("hours", "0") days = maxtime.get("days", "0") return u"{}: {} {}: {} {}: {}".format( _("days"), days, _("hours"), hours, _("minutes"), minutes)
Formats the max time record to a days, hours, minutes string
def _copy_database_data_clientside(self, tables, source, destination): rows = self.get_database_rows(tables, source) cols = self.get_database_columns(tables, source) for r in list(rows.keys()): assert r in tables for c in list(cols.keys()): assert c in tables self.change_db(destination) insert_queries = self._get_insert_commands(rows, cols) self._execute_insert_commands(insert_queries)
Copy the data from a table into another table.
def getrefnames(idf, objname): iddinfo = idf.idd_info dtls = idf.model.dtls index = dtls.index(objname) fieldidds = iddinfo[index] for fieldidd in fieldidds: if 'field' in fieldidd: if fieldidd['field'][0].endswith('Name'): if 'reference' in fieldidd: return fieldidd['reference'] else: return []
get the reference names for this object
def format_decimal(self, altitude=None): coordinates = [str(self.latitude), str(self.longitude)] if altitude is None: altitude = bool(self.altitude) if altitude: if not isinstance(altitude, string_compare): altitude = 'km' coordinates.append(self.format_altitude(altitude)) return ", ".join(coordinates)
Format decimal degrees with altitude
def zlines(f = None, sep = "\0", osep = None, size = 8192): if f is None: f = sys.stdin if osep is None: osep = sep buf = "" while True: chars = f.read(size) if not chars: break buf += chars; lines = buf.split(sep); buf = lines.pop() for line in lines: yield line + osep if buf: yield buf
File iterator that uses alternative line terminators.
def _clear_xauth(self): os.remove(self._xauth_filename) for varname in ['AUTHFILE', 'XAUTHORITY']: if self._old_xauth[varname] is None: del os.environ[varname] else: os.environ[varname] = self._old_xauth[varname] self._old_xauth = None
Clear the Xauthority file and restore the environment variables.
def smallest_signed_angle(source, target): dth = target - source dth = (dth + np.pi) % (2.0 * np.pi) - np.pi return dth
Find the smallest angle going from angle `source` to angle `target`.
def write_byte_data(self, address, register, value): LOGGER.debug("Writing byte data %s to register %s on device %s", bin(value), hex(register), hex(address)) return self.driver.write_byte_data(address, register, value)
Write a byte value to a device's register.
def inline(self): tp_lines, tp_decl = self.get_decl_pair() tp_lines = " ".join(tp_lines) if tp_decl is None: return tp_lines else: return "%s %s" % (tp_lines, tp_decl)
Return the declarator as a single line.
def _calculate_sv_bins_gatk(data, cnv_group, size_calc_fn): if dd.get_background_cnv_reference(data, "gatk-cnv"): target_bed = gatkcnv.pon_to_bed(dd.get_background_cnv_reference(data, "gatk-cnv"), cnv_group.work_dir, data) else: target_bed = gatkcnv.prepare_intervals(data, cnv_group.region_file, cnv_group.work_dir) gc_annotated_tsv = gatkcnv.annotate_intervals(target_bed, data) return target_bed, None, gc_annotated_tsv
Calculate structural variant bins using GATK4 CNV callers region or even intervals approach.
def _server_error_message(url, message): msg = _error_message.format(url=url, message=message) log.error(msg) return msg
Log and return a server error message.
def element_not_contains(self, element_id, value): elem = world.browser.find_elements_by_xpath(str( 'id("{id}")[contains(., "{value}")]'.format( id=element_id, value=value))) assert not elem, \ "Expected element not to contain the given text."
Assert provided content is not contained within an element found by ``id``.
def protocol_names(self): l = self.protocols() retval = [str(k.name) for k in l] return retval
Returns all registered protocol names
def remove(self, tab_index): _id = id(self.editor.tabs.widget(tab_index)) if _id in self.history: self.history.remove(_id)
Remove the widget at the corresponding tab_index.
def policy_error_scope(error, data): err_path = list(error.absolute_path) if err_path[0] != 'policies': return error pdata = data['policies'][err_path[1]] pdata.get('name', 'unknown') error.message = "Error on policy:{} resource:{}\n".format( pdata.get('name', 'unknown'), pdata.get('resource', 'unknown')) + error.message return error
Scope a schema error to its policy name and resource.
def post_register_hook(self, verbosity=1): if not getattr(settings, 'FLOW_DOCKER_DONT_PULL', False): call_command('list_docker_images', pull=True, verbosity=verbosity)
Pull Docker images needed by processes after registering.
def _filter_or_exclude(self, negate, *args, **kwargs): from .models import Translation new_args = self.get_cleaned_args(args) new_kwargs = self.get_cleaned_kwargs(kwargs) translation_args = self.get_translation_args(args) translation_kwargs = self.get_translation_kwargs(kwargs) has_linguist_args = self.has_linguist_args(args) has_linguist_kwargs = self.has_linguist_kwargs(kwargs) if translation_args or translation_kwargs: ids = list( set( Translation.objects.filter( *translation_args, **translation_kwargs ).values_list("object_id", flat=True) ) ) if ids: new_kwargs["id__in"] = ids has_kwargs = has_linguist_kwargs and not (new_kwargs or new_args) has_args = has_linguist_args and not (new_args or new_kwargs) if has_kwargs or has_args: return self._clone().none() return super(QuerySetMixin, self)._filter_or_exclude( negate, *new_args, **new_kwargs )
Overrides default behavior to handle linguist fields.
def _print_convergence(self): if self.verbosity: if (self.num_acquisitions == self.max_iter) and (not self.initial_iter): print(' ** Maximum number of iterations reached **') return 1 elif (self._distance_last_evaluations() < self.eps) and (not self.initial_iter): print(' ** Two equal location selected **') return 1 elif (self.max_time < self.cum_time) and not (self.initial_iter): print(' ** Evaluation time reached **') return 0 if self.initial_iter: print('** GPyOpt Bayesian Optimization class initialized successfully **') self.initial_iter = False
Prints the reason why the optimization stopped.
def getNodeByName(node, name): if node is None: raise Exception( "Cannot search for a child '%s' in a None object" % (name,) ) if not name: raise Exception("Unspecified name to find node for.") try: childNode = node.xpath("*[local-name() = '%s']" % name)[0] except: return None return childNode
Get the first child node matching a given local name
def pluralize(data_type): known = { u"address": u"addresses", u"company": u"companies" } if data_type in known.keys(): return known[data_type] else: return u"%ss" % data_type
adds s to the data type or the correct english plural form
def _find_new_partners(self): open_channels = views.get_channelstate_open( chain_state=views.state_from_raiden(self.raiden), payment_network_id=self.registry_address, token_address=self.token_address, ) known = set(channel_state.partner_state.address for channel_state in open_channels) known.add(self.BOOTSTRAP_ADDR) known.add(self.raiden.address) participants_addresses = views.get_participants_addresses( views.state_from_raiden(self.raiden), self.registry_address, self.token_address, ) available = participants_addresses - known available = list(available) shuffle(available) new_partners = available log.debug( 'Found partners', node=pex(self.raiden.address), number_of_partners=len(available), ) return new_partners
Search the token network for potential channel partners.
def parse_command_line_args(): parser = argparse.ArgumentParser(description='PipApp. {}'.format(DESCRIPTION)) parser.add_argument( '-d', '--dir', metavar='DIR', help='Root directory where to create new project files and dirs. Default is current directory.' ) parser.add_argument( '-v,', '--version', action='version', version='{} v{}'.format(PROGRAMNAME, VERSION) ) parser.add_argument( "project_name", metavar='PROJECTNAME', help="Name of the generated Project. Has to be a valid Python identifier." ) return parser.parse_args()
parse command line args
def collect_manifest_dependencies(manifest_data, lockfile_data): output = {} for dependencyName, dependencyConstraint in manifest_data.items(): output[dependencyName] = { 'source': 'example-package-manager', 'constraint': dependencyConstraint, 'available': [ {'name': '2.0.0'}, ], } return output
Convert the manifest format to the dependencies schema
def save(self, vs, filetype): 'Copy rows to the system clipboard.' with tempfile.NamedTemporaryFile(suffix='.'+filetype) as temp: saveSheets(temp.name, vs) sync(1) p = subprocess.Popen( self.command, stdin=open(temp.name, 'r', encoding=options.encoding), stdout=subprocess.DEVNULL, close_fds=True) p.communicate()
Copy rows to the system clipboard.
def _check_index(self): assert 0 <= self._index <= len(self._history) - 1 assert len(self._history) >= 1
Check that the index is without the bounds of _history.
def descendents(self): for c in self.children: yield c for d in c.descendents: yield d
Iterate over all descendent terms
def distance_measure_common(A, func, alpha, R, k, epsilon): x = relaxation_vectors(A, R, k, alpha) d = func(x) (rows, cols) = A.nonzero() weak = np.where(rows == cols)[0] d[weak] = 0 C = sparse.csr_matrix((d, (rows, cols)), shape=A.shape) C.eliminate_zeros() amg_core.apply_distance_filter(C.shape[0], epsilon, C.indptr, C.indices, C.data) C.eliminate_zeros() C.data = 1.0 / C.data C = C + sparse.eye(C.shape[0], C.shape[1], format='csr') C = scale_rows_by_largest_entry(C) return C
Create strength of connection matrixfrom a function applied to relaxation vectors.
def show_metering_label_rule(self, metering_label_rule, **_params): return self.get(self.metering_label_rule_path % (metering_label_rule), params=_params)
Fetches information of a certain metering label rule.
def generate_kubernetes(self): example = {} example['spec'] = {} example['spec']['containers'] = [] example['spec']['containers'].append({"name": '', "image": '', "env": []}) for key, value in self.spec.items(): if value['type'] in (dict, list): kvalue = f"\'{json.dumps(value.get('example', ''))}\'" else: kvalue = f"{value.get('example', '')}" entry = {"name": f"{self.env_prefix}_{key.upper()}", "value": kvalue} example['spec']['containers'][0]['env'].append(entry) print(yaml.dump(example, default_flow_style=False))
Generate a sample kubernetes
async def _get_protocol(self): if self._protocol is None: self._protocol = asyncio.Task(Context.create_client_context( loop=self._loop)) return (await self._protocol)
Get the protocol for the request.
def symbols(): symbols = [] for line in symbols_stream(): symbols.append(line.decode('utf-8').strip()) return symbols
Return a list of symbols.
def open_datasets(path, backend_kwargs={}, no_warn=False, **kwargs): if not no_warn: warnings.warn("open_datasets is an experimental API, DO NOT RELY ON IT!", FutureWarning) fbks = [] datasets = [] try: datasets.append(open_dataset(path, backend_kwargs=backend_kwargs, **kwargs)) except DatasetBuildError as ex: fbks.extend(ex.args[2]) for fbk in fbks: bks = backend_kwargs.copy() bks['filter_by_keys'] = fbk datasets.extend(open_datasets(path, backend_kwargs=bks, no_warn=True, **kwargs)) return datasets
Open a GRIB file groupping incompatible hypercubes to different datasets via simple heuristics.
def reload_webservers(): if env.verbosity: print env.host, "RELOADING apache2" with settings(warn_only=True): a = sudo("/etc/init.d/apache2 reload") if env.verbosity: print '',a if env.verbosity: print env.host,"RELOADING nginx" with settings(warn_only=True): s = run("/etc/init.d/nginx status") if 'running' in s: n = sudo("/etc/init.d/nginx reload") else: n = sudo("/etc/init.d/nginx start") if env.verbosity: print ' *',n return True
Reload apache2 and nginx
def extract_image_size(self): width, _ = self._extract_alternative_fields( ['Image ImageWidth', 'EXIF ExifImageWidth'], -1, int) height, _ = self._extract_alternative_fields( ['Image ImageLength', 'EXIF ExifImageLength'], -1, int) return width, height
Extract image height and width
def lookup(img, **kwargs): luts = np.array(kwargs['luts'], dtype=np.float32) / 255.0 def func(band_data, luts=luts, index=-1): lut = luts[:, index] if len(luts.shape) == 2 else luts band_data = band_data.clip(0, lut.size - 1).astype(np.uint8) new_delay = dask.delayed(_lookup_delayed)(lut, band_data) new_data = da.from_delayed(new_delay, shape=band_data.shape, dtype=luts.dtype) return new_data return apply_enhancement(img.data, func, separate=True, pass_dask=True)
Assign values to channels based on a table.
def extract_cookies(self, response, request): _debug("extract_cookies: %s", response.info()) self._cookies_lock.acquire() try: self._policy._now = self._now = int(time.time()) for cookie in self.make_cookies(response, request): if self._policy.set_ok(cookie, request): _debug(" setting cookie: %s", cookie) self.set_cookie(cookie) finally: self._cookies_lock.release()
Extract cookies from response, where allowable given the request.
def on(self, group): asyncio.ensure_future(self._send_led_on_off_request(group, 1), loop=self._loop)
Turn the LED on for a group.
def _list_superclasses(class_def): superclasses = class_def.get('superClasses', []) if superclasses: return list(superclasses) sup = class_def.get('superClass', None) if sup: return [sup] else: return []
Return a list of the superclasses of the given class
def mark_good(self, proxy): if proxy not in self.proxies: logger.warn("Proxy <%s> was not found in proxies list" % proxy) return if proxy not in self.good: logger.debug("Proxy <%s> is GOOD" % proxy) self.unchecked.discard(proxy) self.dead.discard(proxy) self.good.add(proxy) self.proxies[proxy].failed_attempts = 0
Mark a proxy as good