_id
stringlengths
2
7
title
stringlengths
1
88
partition
stringclasses
3 values
text
stringlengths
31
13.1k
language
stringclasses
1 value
meta_information
dict
q264900
Nemo.register
validation
def register(self): """ Register the app using Blueprint :return: Nemo blueprint :rtype: flask.Blueprint """ if self.app is not None: if not self.blueprint: self.blueprint = self.create_blueprint()
python
{ "resource": "" }
q264901
Nemo.register_filters
validation
def register_filters(self): """ Register filters for Jinja to use .. note:: Extends the dictionary filters of jinja_env using self._filters list """ for _filter, instance in self._filters: if not instance: self.app.jinja_env.filters[ _filter.replace("f_", "") ] = getattr(flask_nemo.filters, _filter) else:
python
{ "resource": "" }
q264902
Nemo.register_plugins
validation
def register_plugins(self): """ Register plugins in Nemo instance - Clear routes first if asked by one plugin - Clear assets if asked by one plugin and replace by the last plugin registered static_folder - Register each plugin - Append plugin routes to registered routes - Append plugin filters to registered filters - Append templates directory to given namespaces - Append assets (CSS, JS, statics) to given resources - Append render view (if exists) to Nemo.render stack """ if len([plugin for plugin in self.__plugins__.values() if plugin.clear_routes]) > 0: # Clear current routes self._urls = list() self.cached = list() clear_assets = [plugin for plugin in self.__plugins__.values() if plugin.clear_assets] if len(clear_assets) > 0 and not self.prevent_plugin_clearing_assets: # Clear current Assets self.__assets__ = copy(type(self).ASSETS) static_path = [plugin.static_folder for plugin in clear_assets if plugin.static_folder] if len(static_path) > 0: self.static_folder = static_path[-1] for plugin in self.__plugins__.values(): self._urls.extend([(url, function, methods, plugin) for url, function, methods in plugin.routes]) self._filters.extend([(filt, plugin) for filt in plugin.filters])
python
{ "resource": "" }
q264903
Nemo.chunk
validation
def chunk(self, text, reffs): """ Handle a list of references depending on the text identifier using the chunker dictionary. :param text: Text object from which comes the references :type text: MyCapytains.resources.texts.api.Text :param reffs: List of references to transform :type reffs: References :return: Transformed
python
{ "resource": "" }
q264904
add_tag
validation
def add_tag(): """ Obtains the data from the pipe and appends the given tag. """ if len(sys.argv) > 1: tag = sys.argv[1] doc_mapper = DocMapper() if doc_mapper.is_pipe: count = 0 for obj in doc_mapper.get_pipe(): obj.add_tag(tag) obj.update(tags=obj.tags) count += 1 print_success("Added tag
python
{ "resource": "" }
q264905
Config.set
validation
def set(self, section, key, value): """ Creates the section value if it does not exists and sets the value. Use write_config to actually set the value. """ if not
python
{ "resource": "" }
q264906
Config.get
validation
def get(self, section, key): """ This function tries to retrieve the value from the configfile otherwise will return a default. """ try: return self.config.get(section, key)
python
{ "resource": "" }
q264907
Config.config_dir
validation
def config_dir(self): """ Returns the configuration directory """ home
python
{ "resource": "" }
q264908
Config.write_config
validation
def write_config(self, initialize_indices=False): """ Write the current config to disk to store them. """ if not os.path.exists(self.config_dir): os.mkdir(self.config_dir) with open(self.config_file, 'w') as configfile: self.config.write(configfile) if initialize_indices: index = self.get('jackal', 'index') from jackal import Host, Range, Service, User, Credential, Log from jackal.core import create_connection create_connection(self) Host.init(index="{}-hosts".format(index))
python
{ "resource": "" }
q264909
ensure_remote_branch_is_tracked
validation
def ensure_remote_branch_is_tracked(branch): """Track the specified remote branch if it is not already tracked.""" if branch == MASTER_BRANCH: # We don't need to explicitly track the master branch, so we're done. return # Ensure the specified branch is in the local branch list. output = subprocess.check_output(['git', 'branch', '--list']) for line in output.split('\n'): if line.strip() == branch: # We are already tracking the remote branch break else: # We are not tracking
python
{ "resource": "" }
q264910
main
validation
def main(branch): """Checkout, update and branch from the specified branch.""" try: # Ensure that we're in a git repository. This command is silent unless # you're not actually in a git repository, in which case, you receive a # "Not a git repository" error message. output = subprocess.check_output(['git', 'rev-parse']).decode('utf-8') sys.stdout.write(output) except subprocess.CalledProcessError: # Bail if we're not in a git repository. return # This behavior ensures a better user experience for those that aren't # intimately familiar with git. ensure_remote_branch_is_tracked(branch) # Switch to the specified branch and update it. subprocess.check_call(['git', 'checkout', '--quiet', branch]) # Pulling is always safe here, because we never commit to this branch. subprocess.check_call(['git', 'pull', '--quiet']) # Checkout the top commit in the branch, effectively going "untracked." subprocess.check_call(['git',
python
{ "resource": "" }
q264911
get_interface_name
validation
def get_interface_name(): """ Returns the interface name of the first not link_local and not loopback interface. """ interface_name = '' interfaces = psutil.net_if_addrs() for name, details in interfaces.items(): for detail in details: if detail.family == socket.AF_INET:
python
{ "resource": "" }
q264912
Spoofing.load_targets
validation
def load_targets(self): """ load_targets will load the services with smb signing disabled and if ldap is enabled the services with the ldap port open. """ ldap_services = [] if self.ldap:
python
{ "resource": "" }
q264913
Spoofing.write_targets
validation
def write_targets(self): """ write_targets will write the contents of ips and ldap_strings to the targets_file. """ if len(self.ldap_strings) == 0 and len(self.ips) == 0:
python
{ "resource": "" }
q264914
Spoofing.start_processes
validation
def start_processes(self): """ Starts the ntlmrelayx.py and responder processes. Assumes you have these programs in your path. """ self.relay = subprocess.Popen(['ntlmrelayx.py', '-6', '-tf', self.targets_file, '-w', '-l', self.directory,
python
{ "resource": "" }
q264915
Spoofing.callback
validation
def callback(self, event): """ Function that gets called on each event from pyinotify. """ # IN_CLOSE_WRITE -> 0x00000008 if event.mask == 0x00000008: if event.name.endswith('.json'): print_success("Ldapdomaindump file found") if event.name in ['domain_groups.json', 'domain_users.json']: if event.name == 'domain_groups.json': self.domain_groups_file = event.pathname if event.name == 'domain_users.json': self.domain_users_file = event.pathname if self.domain_groups_file and self.domain_users_file: print_success("Importing users") subprocess.Popen(['jk-import-domaindump', self.domain_groups_file, self.domain_users_file]) elif event.name == 'domain_computers.json': print_success("Importing computers") subprocess.Popen(['jk-import-domaindump', event.pathname]) # Ldap has been dumped, so remove the ldap targets.
python
{ "resource": "" }
q264916
Spoofing.watch
validation
def watch(self): """ Watches directory for changes """ wm = pyinotify.WatchManager() self.notifier = pyinotify.Notifier(wm, default_proc_fun=self.callback) wm.add_watch(self.directory, pyinotify.ALL_EVENTS) try: self.notifier.loop()
python
{ "resource": "" }
q264917
Spoofing.terminate_processes
validation
def terminate_processes(self): """ Terminate the processes. """ if self.relay:
python
{ "resource": "" }
q264918
Spoofing.wait
validation
def wait(self): """ This function waits for the relay and responding processes to exit. Captures KeyboardInterrupt to shutdown these processes. """ try: self.relay.wait()
python
{ "resource": "" }
q264919
QueryPrototype.getAnnotations
validation
def getAnnotations(self, targets, wildcard=".", include=None, exclude=None, limit=None, start=1, expand=False, **kwargs): """ Retrieve annotations from the query provider :param targets: The CTS URN(s) to query as the target of annotations :type targets: [MyCapytain.common.reference.URN], URN or None :param wildcard: Wildcard specifier for how to match the URN :type wildcard: str :param include: URI(s) of Annotation types to include in the results :type include: list(str) :param exclude: URI(s) of Annotation types to include in the results :type exclude: list(str) :param limit: The max number of results to return (Default is None for no limit) :type limit: int :param start: the starting record to return (Default is 1) :type start: int :param expand: Flag to state whether Annotations are expanded (Default is False) :type expand: bool :return: Tuple representing the query results. The first element The first
python
{ "resource": "" }
q264920
Breadcrumb.render
validation
def render(self, **kwargs): """ Make breadcrumbs for a route :param kwargs: dictionary of named arguments used to construct the view :type kwargs: dict :return: List of dict items the view can use to construct the link. :rtype: {str: list({ "link": str, "title", str, "args", dict})} """ breadcrumbs = [] # this is the list of items we want to accumulate in the breadcrumb trail. # item[0] is the key into the kwargs["url"] object and item[1] is the name of the route # setting a route name to None means that it's needed to construct the route of the next item in the list # but shouldn't be included in the list itself (this is currently the case for work -- # at some point we probably should include work in the navigation) breadcrumbs = [] if "collections" in kwargs: breadcrumbs = [{ "title": "Text Collections", "link": ".r_collections", "args": {} }] if "parents" in kwargs["collections"]: breadcrumbs += [ { "title": parent["label"], "link": ".r_collection_semantic", "args": {
python
{ "resource": "" }
q264921
main
validation
def main(): """ This function obtains hosts from core and starts a nessus scan on these hosts. The nessus tag is appended to the host tags. """ config = Config() core = HostSearch() hosts = core.get_hosts(tags=['!nessus'], up=True) hosts = [host for host in hosts] host_ips = ",".join([str(host.address) for host in hosts]) url = config.get('nessus', 'host') access = config.get('nessus', 'access_key') secret = config.get('nessus', 'secret_key') template_name = config.get('nessus', 'template_name')
python
{ "resource": "" }
q264922
Nessus.get_template_uuid
validation
def get_template_uuid(self): """ Retrieves the uuid of the given template name. """ response = requests.get(self.url + 'editor/scan/templates', headers=self.headers, verify=False) templates = json.loads(response.text)
python
{ "resource": "" }
q264923
Nessus.create_scan
validation
def create_scan(self, host_ips): """ Creates a scan with the given host ips Returns the scan id of the created object. """ now = datetime.datetime.now() data = { "uuid": self.get_template_uuid(), "settings": { "name": "jackal-" + now.strftime("%Y-%m-%d %H:%M"), "text_targets": host_ips }
python
{ "resource": "" }
q264924
Nessus.start_scan
validation
def start_scan(self, scan_id): """ Starts the scan identified by the scan_id.s """
python
{ "resource": "" }
q264925
cmpToDataStore_uri
validation
def cmpToDataStore_uri(base, ds1, ds2): '''Bases the comparison of the datastores on URI alone.''' ret
python
{ "resource": "" }
q264926
JackalDoc.add_tag
validation
def add_tag(self, tag): """ Adds a tag to the list of tags and makes sure
python
{ "resource": "" }
q264927
JackalDoc.remove_tag
validation
def remove_tag(self, tag): """ Removes a tag from this object """
python
{ "resource": "" }
q264928
JackalDoc.to_dict
validation
def to_dict(self, include_meta=False): """ Returns the result as a dictionary, provide the include_meta flag to als show information like index and doctype.
python
{ "resource": "" }
q264929
AnnotationsApiPlugin.r_annotations
validation
def r_annotations(self): """ Route to retrieve annotations by target :param target_urn: The CTS URN for which to retrieve annotations :type target_urn: str :return: a JSON string containing count and list of resources :rtype: {str: Any} """ target = request.args.get("target", None) wildcard = request.args.get("wildcard", ".", type=str) include = request.args.get("include") exclude = request.args.get("exclude") limit = request.args.get("limit", None, type=int) start = request.args.get("start", 1, type=int) expand = request.args.get("expand", False, type=bool) if target: try: urn = MyCapytain.common.reference.URN(target) except ValueError: return "invalid urn", 400 count, annotations = self.__queryinterface__.getAnnotations(urn, wildcard=wildcard, include=include,
python
{ "resource": "" }
q264930
Enum.lookup
validation
def lookup(cls, key, get=False): """Returns the label for a given Enum key""" if get: item = cls._item_dict.get(key)
python
{ "resource": "" }
q264931
Enum.verbose
validation
def verbose(cls, key=False, default=''): """Returns the verbose name for a given enum value""" if key is False: items = cls._item_dict.values() return [(x.key, x.value)
python
{ "resource": "" }
q264932
get_configured_dns
validation
def get_configured_dns(): """ Returns the configured DNS servers with the use f nmcli. """ ips = [] try: output = subprocess.check_output(['nmcli', 'device', 'show']) output = output.decode('utf-8') for line in output.split('\n'): if 'DNS' in line:
python
{ "resource": "" }
q264933
zone_transfer
validation
def zone_transfer(address, dns_name): """ Tries to perform a zone transfer. """ ips = [] try: print_notification("Attempting dns zone transfer for {} on {}".format(dns_name, address)) z = dns.zone.from_xfr(dns.query.xfr(address, dns_name)) except dns.exception.FormError: print_notification("Zone transfer not allowed") return ips names = z.nodes.keys() print_success("Zone transfer successfull for {}, found {} entries".format(address, len(names)))
python
{ "resource": "" }
q264934
resolve_domains
validation
def resolve_domains(domains, disable_zone=False): """ Resolves the list of domains and returns the ips. """ dnsresolver = dns.resolver.Resolver() ips = [] for domain in domains: print_notification("Resolving {}".format(domain))
python
{ "resource": "" }
q264935
parse_ips
validation
def parse_ips(ips, netmask, include_public): """ Parses the list of ips, turns these into ranges based on the netmask given. Set include_public to True to include public IP adresses. """ hs = HostSearch() rs = RangeSearch() ranges = [] ips = list(set(ips)) included_ips = [] print_success("Found {} ips".format(len(ips))) for ip in ips: ip_address = ipaddress.ip_address(ip) if include_public or ip_address.is_private: # To stop the screen filling with ranges. if len(ips) < 15:
python
{ "resource": "" }
q264936
create_connection
validation
def create_connection(conf): """ Creates a connection based upon the given configuration object. """ host_config = {} host_config['hosts'] = [conf.get('jackal', 'host')] if int(conf.get('jackal', 'use_ssl')): host_config['use_ssl'] = True if conf.get('jackal', 'ca_certs'):
python
{ "resource": "" }
q264937
CoreSearch.search
validation
def search(self, number=None, *args, **kwargs): """ Searches the elasticsearch instance to retrieve the requested documents. """ search = self.create_search(*args, **kwargs) try: if number: response = search[0:number] else: args, _ = self.core_parser.parse_known_args() if args.number: response = search[0:args.number] else:
python
{ "resource": "" }
q264938
CoreSearch.argument_search
validation
def argument_search(self): """ Uses the command line arguments to fill the search function and call it. """
python
{ "resource": "" }
q264939
CoreSearch.count
validation
def count(self, *args, **kwargs): """ Returns the number of results after filtering with the given arguments. """ search = self.create_search(*args, **kwargs) try: return search.count() except NotFoundError:
python
{ "resource": "" }
q264940
CoreSearch.argument_count
validation
def argument_count(self): """ Uses the command line arguments to fill the count function and call it. """
python
{ "resource": "" }
q264941
CoreSearch.get_pipe
validation
def get_pipe(self, object_type): """ Returns a generator that maps the input of the pipe to an elasticsearch object. Will call id_to_object if it cannot serialize the data from json.
python
{ "resource": "" }
q264942
RangeSearch.id_to_object
validation
def id_to_object(self, line): """ Resolves an ip adres to a range object, creating it if it doesn't exists.
python
{ "resource": "" }
q264943
RangeSearch.argparser
validation
def argparser(self): """ Argparser option with search functionality specific for ranges. """ core_parser = self.core_parser
python
{ "resource": "" }
q264944
ServiceSearch.object_to_id
validation
def object_to_id(self, obj): """ Searches elasticsearch for objects with the same address, protocol, port and state. """ search = Service.search() search = search.filter("term", address=obj.address)
python
{ "resource": "" }
q264945
UserSearch.id_to_object
validation
def id_to_object(self, line): """ Resolves the given id to a user object, if it doesn't exists it will be
python
{ "resource": "" }
q264946
UserSearch.get_domains
validation
def get_domains(self): """ Retrieves the domains of the users from elastic. """ search = User.search() search.aggs.bucket('domains', 'terms', field='domain', order={'_count': 'desc'}, size=100)
python
{ "resource": "" }
q264947
DocMapper.get_pipe
validation
def get_pipe(self): """ Returns a list that maps the input of the pipe to an elasticsearch object. Will call id_to_object if it cannot serialize the data from json.
python
{ "resource": "" }
q264948
Protocol.tree2commands
validation
def tree2commands(self, adapter, session, lastcmds, xsync): '''Consumes an ET protocol tree and converts it to state.Command commands''' # do some preliminary sanity checks... # todo: do i really want to be using assert statements?... assert xsync.tag == constants.NODE_SYNCML assert len(xsync) == 2 assert xsync[0].tag == constants.CMD_SYNCHDR assert xsync[1].tag == constants.NODE_SYNCBODY version = xsync[0].findtext('VerProto') if version != constants.SYNCML_VERSION_1_2: raise common.FeatureNotSupported('unsupported SyncML version "%s" (expected "%s")' \ % (version, constants.SYNCML_VERSION_1_2)) verdtd = xsync[0].findtext('VerDTD') if verdtd != constants.SYNCML_DTD_VERSION_1_2: raise common.FeatureNotSupported('unsupported SyncML DTD version "%s" (expected "%s")' \ % (verdtd, constants.SYNCML_DTD_VERSION_1_2)) ret = self.initialize(adapter, session, xsync) hdrcmd = ret[0] if session.isServer: log.debug('received request SyncML message from "%s" (s%s.m%s)', hdrcmd.target, hdrcmd.sessionID, hdrcmd.msgID) else: log.debug('received response SyncML message from "%s" (s%s.m%s)', lastcmds[0].target, lastcmds[0].sessionID, lastcmds[0].msgID) try: return self._tree2commands(adapter, session, lastcmds, xsync, ret) except Exception, e: if not session.isServer: raise # TODO: make this configurable as to whether or not any error #
python
{ "resource": "" }
q264949
initialize_indices
validation
def initialize_indices(): """ Initializes the indices """ Host.init()
python
{ "resource": "" }
q264950
parse_single_computer
validation
def parse_single_computer(entry): """ Parse the entry into a computer object. """ computer = Computer(dns_hostname=get_field(entry, 'dNSHostName'), description=get_field( entry, 'description'), os=get_field(entry, 'operatingSystem'), group_id=get_field(entry, 'primaryGroupID')) try: ip =
python
{ "resource": "" }
q264951
parse_domain_computers
validation
def parse_domain_computers(filename): """ Parse the file and extract the computers, import the computers that resolve into jackal. """ with open(filename) as f: data = json.loads(f.read()) hs = HostSearch() count = 0 entry_count = 0 print_notification("Parsing {} entries".format(len(data))) for system in data: entry_count += 1 parsed = parse_single_computer(system) if parsed.ip:
python
{ "resource": "" }
q264952
parse_user
validation
def parse_user(entry, domain_groups): """ Parses a single entry from the domaindump """ result = {} distinguished_name = get_field(entry, 'distinguishedName') result['domain'] = ".".join(distinguished_name.split(',DC=')[1:]) result['name'] = get_field(entry, 'name') result['username'] = get_field(entry, 'sAMAccountName') result['description'] = get_field(entry, 'description') result['sid'] = get_field(entry, 'objectSid').split('-')[-1] primary_group = get_field(entry, 'primaryGroupID') member_of = entry['attributes'].get('memberOf', []) groups = [] for member in member_of: for e in member.split(','): if e.startswith('CN='):
python
{ "resource": "" }
q264953
parse_domain_users
validation
def parse_domain_users(domain_users_file, domain_groups_file): """ Parses the domain users and groups files. """ with open(domain_users_file) as f: users = json.loads(f.read()) domain_groups = {} if domain_groups_file: with open(domain_groups_file) as f: groups = json.loads(f.read()) for group in groups: sid = get_field(group, 'objectSid') domain_groups[int(sid.split('-')[-1])] = get_field(group, 'cn') user_search = UserSearch() count = 0 total = len(users) print_notification("Importing {} users".format(total)) for entry in users: result = parse_user(entry, domain_groups) user = user_search.id_to_object(result['username']) user.name = result['name']
python
{ "resource": "" }
q264954
import_domaindump
validation
def import_domaindump(): """ Parses ldapdomaindump files and stores hosts and users in elasticsearch. """ parser = argparse.ArgumentParser( description="Imports users, groups and computers result files from the ldapdomaindump tool, will resolve the names from domain_computers output for IPs") parser.add_argument("files", nargs='+', help="The domaindump files to import") arguments = parser.parse_args() domain_users_file = '' domain_groups_file = '' computer_count = 0 user_count = 0 stats = {} for filename in arguments.files: if filename.endswith('domain_computers.json'): print_notification('Parsing domain computers') computer_count = parse_domain_computers(filename) if computer_count: stats['hosts'] = computer_count
python
{ "resource": "" }
q264955
autocomplete
validation
def autocomplete(query, country=None, hurricanes=False, cities=True, timeout=5): """Make an autocomplete API request This can be used to find cities and/or hurricanes by name :param string query: city :param string country: restrict search to a specific country. Must be a two letter country code :param boolean hurricanes: whether to search for hurricanes or not
python
{ "resource": "" }
q264956
request
validation
def request(key, features, query, timeout=5): """Make an API request :param string key: API key to use :param list features: features to request. It must be a subset of :data:`FEATURES` :param string query: query to send :param integer timeout: timeout of the request :returns: result of the API request :rtype: dict """ data = {}
python
{ "resource": "" }
q264957
_unicode
validation
def _unicode(string): """Try to convert a string to unicode using different encodings""" for encoding in ['utf-8', 'latin1']: try: result = unicode(string, encoding) return
python
{ "resource": "" }
q264958
http_get_provider
validation
def http_get_provider(provider, request_url, params, token_secret, token_cookie = None): '''Handle HTTP GET requests on an authentication endpoint. Authentication flow begins when ``params`` has a ``login`` key with a value of ``start``. For instance, ``/auth/twitter?login=start``. :param str provider: An provider to obtain a user ID from. :param str request_url: The authentication endpoint/callback. :param dict params: GET parameters from the query string. :param str token_secret: An app secret to encode/decode JSON web tokens. :param str token_cookie: The current JSON web token, if available. :return: A dict containing any of the following possible keys: ``status``: an HTTP status code the server should sent
python
{ "resource": "" }
q264959
Target.to_json
validation
def to_json(self): """ Method to call to get a serializable object for json.dump or jsonify based on the target :return: dict """ if self.subreference is not None: return { "source": self.objectId, "selector": { "type": "FragmentSelector",
python
{ "resource": "" }
q264960
AnnotationResource.read
validation
def read(self): """ Read the contents of the Annotation Resource :return: the contents of the resource :rtype: str or bytes or flask.response
python
{ "resource": "" }
q264961
build_index_and_mapping
validation
def build_index_and_mapping(triples): """index all triples into indexes and return their mappings""" ents = bidict() rels = bidict() ent_id = 0 rel_id = 0 collected = [] for t in triples: for e in (t.head, t.tail): if e not in ents: ents[e] = ent_id ent_id += 1 if t.relation not in rels:
python
{ "resource": "" }
q264962
recover_triples_from_mapping
validation
def recover_triples_from_mapping(indexes, ents: bidict, rels: bidict): """recover triples from mapping.""" triples = [] for t in indexes:
python
{ "resource": "" }
q264963
_transform_triple_numpy
validation
def _transform_triple_numpy(x): """Transform triple index into a 1-D numpy array."""
python
{ "resource": "" }
q264964
pack_triples_numpy
validation
def pack_triples_numpy(triples): """Packs a list of triple indexes into a 2D numpy array.""" if len(triples) == 0: return np.array([],
python
{ "resource": "" }
q264965
remove_near_duplicate_relation
validation
def remove_near_duplicate_relation(triples, threshold=0.97): """If entity pairs in a relation is as close as another relations, only keep one relation of such set.""" logging.debug("remove duplicate") _assert_threshold(threshold) duplicate_rel_counter = defaultdict(list) relations = set() for t in triples: duplicate_rel_counter[t.relation].append(f"{t.head} {t.tail}") relations.add(t.relation) relations = list(relations) num_triples = len(triples) removal_relation_set = set() for rel, values in duplicate_rel_counter.items(): duplicate_rel_counter[rel] = Superminhash(values) for i in relations: for j in relations: if i == j or i in removal_relation_set or j in removal_relation_set: continue
python
{ "resource": "" }
q264966
remove_direct_link_triples
validation
def remove_direct_link_triples(train, valid, test): """Remove direct links in the training sets.""" pairs = set() merged = valid + test for t in merged:
python
{ "resource": "" }
q264967
Indexer.shrink_indexes_in_place
validation
def shrink_indexes_in_place(self, triples): """Uses a union find to find segment.""" _ent_roots = self.UnionFind(self._ent_id) _rel_roots = self.UnionFind(self._rel_id) for t in triples: _ent_roots.add(t.head) _ent_roots.add(t.tail) _rel_roots.add(t.relation) for i, t in enumerate(triples): h = _ent_roots.find(t.head) r = _rel_roots.find(t.relation) t = _ent_roots.find(t.tail) triples[i] = kgedata.TripleIndex(h, r, t) ents = bidict() available_ent_idx = 0 for previous_idx, ent_exist in enumerate(_ent_roots.roots()): if not ent_exist: self._ents.inverse.pop(previous_idx) else: ents[self._ents.inverse[previous_idx]] = available_ent_idx available_ent_idx += 1 rels = bidict() available_rel_idx = 0
python
{ "resource": "" }
q264968
IndexBuilder.freeze
validation
def freeze(self): """Create a usable data structure for serializing.""" data = super(IndexBuilder, self).freeze() try: # Sphinx >= 1.5 format # Due to changes from github.com/sphinx-doc/sphinx/pull/2454 base_file_names = data['docnames'] except KeyError: # Sphinx < 1.5 format base_file_names = data['filenames'] store = {} c = itertools.count() for prefix, items in iteritems(data['objects']): for name, (index, typeindex, _, shortanchor) in iteritems(items): objtype = data['objtypes'][typeindex] if objtype.startswith('cpp:'): split = name.rsplit('::', 1) if len(split) != 2: warnings.warn("What's up with %s?" % str((prefix, name, objtype))) continue
python
{ "resource": "" }
q264969
log_operation
validation
def log_operation(entities, operation_name, params=None): """Logs an operation done on an entity, possibly with other arguments """ if isinstance(entities, (list, tuple)): entities = list(entities) else: entities = [entities] p =
python
{ "resource": "" }
q264970
log_state
validation
def log_state(entity, state): """Logs a new state of an entity """
python
{ "resource": "" }
q264971
log_update
validation
def log_update(entity, update): """Logs an update done on an
python
{ "resource": "" }
q264972
log_error
validation
def log_error(error, result): """Logs an error """ p =
python
{ "resource": "" }
q264973
dict_cursor
validation
def dict_cursor(func): """ Decorator that provides a dictionary cursor to the calling function Adds the cursor as the second argument to the calling functions Requires that the function being decorated is an instance of a class or object that yields a cursor from a get_cursor(cursor_type=CursorType.DICT) coroutine or provides such an object as the first argument in its signature Yields:
python
{ "resource": "" }
q264974
cursor
validation
def cursor(func): """ Decorator that provides a cursor to the calling function Adds the cursor as the second argument to the calling functions Requires that the function being decorated is an instance of a class or object that yields a cursor from a get_cursor() coroutine or provides such an object as the first argument in its signature Yields:
python
{ "resource": "" }
q264975
nt_cursor
validation
def nt_cursor(func): """ Decorator that provides a namedtuple cursor to the calling function Adds the cursor as the second argument to the calling functions Requires that the function being decorated is an instance of a class or object that yields a cursor from a get_cursor(cursor_type=CursorType.NAMEDTUPLE) coroutine or provides such an object as the first argument in its signature Yields: A client-side
python
{ "resource": "" }
q264976
transaction
validation
def transaction(func): """ Provides a transacted cursor which will run in autocommit=false mode For any exception the transaction will be rolled back. Requires that the function being decorated is an instance of a class or object that yields a cursor from a get_cursor(cursor_type=CursorType.NAMEDTUPLE) coroutine or provides such an object as the first argument in its signature Yields: A client-side transacted named cursor """ @wraps(func) def wrapper(cls, *args, **kwargs): with (yield from cls.get_cursor(_CursorType.NAMEDTUPLE)) as c: try:
python
{ "resource": "" }
q264977
PostgresStore.count
validation
def count(cls, cur, table:str, where_keys: list=None): """ gives the number of records in the table Args: table: a string indicating the name of the table Returns: an integer indicating the number of records in the table
python
{ "resource": "" }
q264978
PostgresStore.insert
validation
def insert(cls, cur, table: str, values: dict): """ Creates an insert statement with only chosen fields Args: table: a string indicating the name of the table values: a dict of fields and values to be inserted
python
{ "resource": "" }
q264979
PostgresStore.update
validation
def update(cls, cur, table: str, values: dict, where_keys: list) -> tuple: """ Creates an update query with only chosen fields Supports only a single field where clause Args: table: a string indicating the name of the table values: a dict of fields and values to be inserted where_keys: list of dictionary example of where keys: [{'name':('>', 'cip'),'url':('=', 'cip.com'},{'type':{'<=', 'manufacturer'}}] where_clause will look like ((name>%s and url=%s) or (type <= %s)) items within each dictionary get 'AND'-ed and dictionaries themselves get 'OR'-ed Returns: an integer indicating count
python
{ "resource": "" }
q264980
PostgresStore.delete
validation
def delete(cls, cur, table: str, where_keys: list): """ Creates a delete query with where keys Supports multiple where clause with and or or both Args: table: a string indicating the name of the table where_keys: list of dictionary example of where keys: [{'name':('>', 'cip'),'url':('=', 'cip.com'},{'type':{'<=', 'manufacturer'}}]
python
{ "resource": "" }
q264981
PostgresStore.select
validation
def select(cls, cur, table: str, order_by: str, columns: list=None, where_keys: list=None, limit=100, offset=0): """ Creates a select query for selective columns with where keys Supports multiple where claus with and or or both Args: table: a string indicating the name of the table order_by: a string indicating column name to order the results on columns: list of columns to select from where_keys: list of dictionary limit: the limit on the number of results offset: offset on the results example of where keys: [{'name':('>', 'cip'),'url':('=', 'cip.com'},{'type':{'<=', 'manufacturer'}}] where_clause will look like ((name>%s and url=%s) or (type <= %s)) items within each dictionary get 'AND'-ed and across dictionaries get 'OR'-ed Returns: A list of 'Record' object with table columns as properties """ if columns: columns_string = cls._COMMA.join(columns) if where_keys: where_clause, values = cls._get_where_clause_with_values(where_keys) query = cls._select_selective_column_with_condition.format(columns_string, table, where_clause,
python
{ "resource": "" }
q264982
PostgresStore.raw_sql
validation
def raw_sql(cls, cur, query: str, values: tuple): """ Run a raw sql query Args: query : query string to execute values : tuple of values to be used with the query Returns:
python
{ "resource": "" }
q264983
serialize_text
validation
def serialize_text(out, text): """This method is used to append content of the `text` argument to the `out` argument. Depending on how many lines in the text, a padding can be added to all lines except the first one. Concatenation result is appended to the `out` argument. """ padding = len(out)
python
{ "resource": "" }
q264984
format_value
validation
def format_value(value): """This function should return unicode representation of the value """ value_id = id(value) if value_id in recursion_breaker.processed: return u'<recursion>' recursion_breaker.processed.add(value_id) try: if isinstance(value, six.binary_type): # suppose, all byte strings are in unicode # don't know if everybody in the world uses anything else? return u"'{0}'".format(value.decode('utf-8')) elif isinstance(value, six.text_type): return u"u'{0}'".format(value) elif isinstance(value, (list, tuple)): # long lists or lists with multiline items # will be shown vertically values = list(map(format_value, value)) result = serialize_list(u'[', values, delimiter=u',') + u']' return force_unicode(result) elif isinstance(value, dict): items = six.iteritems(value) # format each key/value pair as a text, # calling format_value recursively items = (tuple(map(format_value, item))
python
{ "resource": "" }
q264985
traverse
validation
def traverse(element, query, deep=False): """ Helper function to traverse an element tree rooted at element, yielding nodes matching the query. """ # Grab the next part of the query (it will be chopped from the front each iteration). part = query[0] if not part: # If the part is blank, we encountered a //, meaning search all sub-nodes. query = query[1:] part = query[0] deep = True # Parse out any predicate (tag[pred]) from this part of the query. part, predicate = xpath_re.match(query[0]).groups() for c in element._children: if part in ('*', c.tagname) and c._match(predicate): # A potential matching branch: this child matches the next query part (and predicate). if len(query) == 1: # If this is the last part of the query, we found a matching element, yield it. yield c else: #
python
{ "resource": "" }
q264986
parse_query
validation
def parse_query(query): """ Given a simplified XPath query string, returns an array of normalized query parts. """ parts = query.split('/') norm = [] for p in parts: p = p.strip()
python
{ "resource": "" }
q264987
XmlElement.insert
validation
def insert(self, before, name, attrs=None, data=None): """ Inserts a new element as a child of this element, before the specified index or sibling. :param before: An :class:`XmlElement` or a numeric index to insert the new node before :param name: The tag name to add :param attrs: Attributes for the new tag :param data: CDATA for the new tag :returns: The newly-created element :rtype: :class:`XmlElement` """ if isinstance(before, self.__class__): if before.parent != self: raise ValueError('Cannot insert before an element with a different parent.')
python
{ "resource": "" }
q264988
XmlElement.children
validation
def children(self, name=None, reverse=False): """ A generator yielding children of this node. :param name: If specified, only consider elements with this tag name :param reverse: If ``True``, children will be yielded in reverse declaration order """ elems = self._children
python
{ "resource": "" }
q264989
XmlElement._match
validation
def _match(self, pred): """ Helper function to determine if this node matches the given predicate. """ if not pred: return True # Strip off the [ and ] pred = pred[1:-1] if pred.startswith('@'): # An attribute predicate checks the existence (and optionally value) of an attribute on this tag. pred = pred[1:] if '=' in pred: attr, value = pred.split('=', 1) if value[0] in ('"', "'"): value = value[1:] if value[-1] in ('"', "'"): value = value[:-1] return self.attrs.get(attr) == value else: return pred in self.attrs elif num_re.match(pred): # An index predicate checks
python
{ "resource": "" }
q264990
XmlElement.path
validation
def path(self, include_root=False): """ Returns a canonical path to this element, relative to the root node. :param include_root: If ``True``, include the root node in the path. Defaults to ``False``. """ path = '%s[%d]' % (self.tagname, self.index or 0)
python
{ "resource": "" }
q264991
XmlElement.iter
validation
def iter(self, name=None): """ Recursively find any descendants of this node with the given tag name. If a tag name is omitted, this will yield every descendant node. :param name: If specified, only consider elements with this tag name :returns: A generator yielding descendants of this node """
python
{ "resource": "" }
q264992
XmlElement.last
validation
def last(self, name=None): """ Returns the last child of this node. :param name: If specified, only
python
{ "resource": "" }
q264993
XmlElement.parents
validation
def parents(self, name=None): """ Yields all parents of this element, back to the root element. :param name: If specified, only consider elements with this tag name """ p = self.parent while
python
{ "resource": "" }
q264994
XmlElement.next
validation
def next(self, name=None): """ Returns the next sibling of this node. :param name: If specified, only consider elements with this tag name :rtype: :class:`XmlElement` """ if self.parent is None or self.index is None:
python
{ "resource": "" }
q264995
XmlElement.prev
validation
def prev(self, name=None): """ Returns the previous sibling of this node. :param name: If specified, only consider elements with this tag name :rtype: :class:`XmlElement` """ if self.parent is None or self.index is None:
python
{ "resource": "" }
q264996
WebObsResultsParser.get_observations
validation
def get_observations(self): """ Parses the HTML table into a list of dictionaries, each of which represents a single observation. """ if self.empty: return [] rows = list(self.tbody) observations = [] for row_observation, row_details in zip(rows[::2], rows[1::2]): data = {} cells = OBSERVATION_XPATH(row_observation) data['name'] = _clean_cell(cells[0]) data['date'] = _clean_cell(cells[1]) data['magnitude'] = _clean_cell(cells[3]) data['obscode'] = _clean_cell(cells[6]) cells
python
{ "resource": "" }
q264997
get_cache_key
validation
def get_cache_key(prefix, *args, **kwargs): """ Calculates cache key based on `args` and `kwargs`. `args` and `kwargs` must be instances of hashable types. """ hash_args_kwargs
python
{ "resource": "" }
q264998
cache_func
validation
def cache_func(prefix, method=False): """ Cache result of function execution into the django cache backend. Calculate cache key based on `prefix`, `args` and `kwargs` of the function. For using like object method set `method=True`. """ def decorator(func): @wraps(func) def wrapper(*args, **kwargs): cache_args = args if method: cache_args = args[1:] cache_key = get_cache_key(prefix, *cache_args, **kwargs)
python
{ "resource": "" }
q264999
get_or_default
validation
def get_or_default(func=None, default=None): """ Wrapper around Django's ORM `get` functionality. Wrap anything that raises ObjectDoesNotExist exception and provide the default value if necessary. `default` by default is None. `default` can be any callable, if it is callable it will be called when ObjectDoesNotExist exception will be raised. """ def decorator(func): @wraps(func) def wrapper(*args, **kwargs): try: return func(*args, **kwargs)
python
{ "resource": "" }