_id
stringlengths
2
7
title
stringlengths
1
88
partition
stringclasses
3 values
text
stringlengths
31
13.1k
language
stringclasses
1 value
meta_information
dict
q260700
SimilarityDistance.predict_proba
validation
def predict_proba(self, X): """Returns the value of the nearest neighbor from the training set. Parameters ---------- X : array-like or sparse matrix, shape (n_samples, n_features) The input samples. Internally, it will be converted to
python
{ "resource": "" }
q260701
Leverage.fit
validation
def fit(self, X, y=None): """Learning is to find the inverse matrix for X and calculate the threshold. Parameters ---------- X : array-like or sparse matrix, shape (n_samples, n_features) The input samples. Use ``dtype=np.float32`` for maximum efficiency. y : array-like, shape = [n_samples] or [n_samples, n_outputs] The target values (real numbers in regression). Returns ------- self : object """ # Check that X have correct shape X = check_array(X) self.inverse_influence_matrix = self.__make_inverse_matrix(X) if self.threshold == 'auto': self.threshold_value = 3 * (1 + X.shape[1]) / X.shape[0] elif self.threshold == 'cv': if y is None: raise ValueError("Y must be specified to find the optimal threshold.") y = check_array(y, accept_sparse='csc', ensure_2d=False, dtype=None) self.threshold_value = 0 score =
python
{ "resource": "" }
q260702
Leverage.predict_proba
validation
def predict_proba(self, X): """Predict the distances for X to center of the training set. Parameters ---------- X : array-like or sparse matrix, shape (n_samples, n_features) The input samples. Internally, it will be converted to ``dtype=np.float32`` and if a sparse matrix is provided to a sparse ``csr_matrix``. Returns ------- leverages: array of shape = [n_samples]
python
{ "resource": "" }
q260703
Leverage.predict
validation
def predict(self, X): """Predict inside or outside AD for X. Parameters ---------- X : array-like or sparse matrix, shape (n_samples, n_features) The input samples. Internally, it will be converted to ``dtype=np.float32`` and if a sparse matrix is provided to a sparse ``csr_matrix``. Returns ------- ad : array of shape = [n_samples] Array contains True (reaction in AD) and False (reaction residing outside AD).
python
{ "resource": "" }
q260704
Box.fit
validation
def fit(self, X, y=None): """Find min and max values of every feature. Parameters ---------- X : {array-like, sparse matrix}, shape (n_samples, n_features) The training input samples. y : Ignored not used, present for API consistency by convention. Returns ------- self : object """ # Check that X have correct shape X = check_array(X)
python
{ "resource": "" }
q260705
CIMtoolsTransformerMixin.fit
validation
def fit(self, x, y=None): """Do nothing and return the estimator unchanged This method is just there to implement the usual API and hence work in pipelines. """ if self._dtype is not None:
python
{ "resource": "" }
q260706
Fragmentor.finalize
validation
def finalize(self): """ finalize partial fitting procedure """ if self.__head_less: warn(f'{self.__class__.__name__} configured to head less mode. finalize unusable') elif not self.__head_generate: warn(f'{self.__class__.__name__} already finalized or fitted') elif not self.__head_dict: raise NotFittedError(f'{self.__class__.__name__} instance is not fitted yet')
python
{ "resource": "" }
q260707
Fragmentor.fit
validation
def fit(self, x, y=None): """Compute the header. """ x = iter2array(x, dtype=(MoleculeContainer, CGRContainer)) if self.__head_less:
python
{ "resource": "" }
q260708
ReactionTypeControl.fit
validation
def fit(self, X): """Fit structure-based AD. The training model memorizes the unique set of reaction signature. Parameters ----------
python
{ "resource": "" }
q260709
_self_referential_fk
validation
def _self_referential_fk(klass_model): """ Return whether this model has a self ref FK, and the name for the field
python
{ "resource": "" }
q260710
SyncableModel.serialize
validation
def serialize(self): """All concrete fields of the ``SyncableModel`` subclass, except for those specifically blacklisted, are returned in a dict.""" # NOTE: code adapted from https://github.com/django/django/blob/master/django/forms/models.py#L75 opts = self._meta data = {} for f in opts.concrete_fields: if f.attname in self.morango_fields_not_to_serialize: continue if f.attname in self._morango_internal_fields_not_to_serialize: continue # case if model is morango mptt
python
{ "resource": "" }
q260711
SyncableModel.deserialize
validation
def deserialize(cls, dict_model): """Returns an unsaved class object based on the valid properties passed in."""
python
{ "resource": "" }
q260712
UUIDField.get_default
validation
def get_default(self): """ Returns the default value for this field. """ if self.has_default(): if callable(self.default): default = self.default()
python
{ "resource": "" }
q260713
UUIDModelMixin.calculate_uuid
validation
def calculate_uuid(self): """Should return a 32-digit hex string for a UUID that is calculated as a function of a set of fields from the model.""" # raise an error if no inputs to the UUID calculation were specified if self.uuid_input_fields is None: raise NotImplementedError("""You must define either a 'uuid_input_fields' attribute (with a tuple of field names) or override the 'calculate_uuid' method, on models that inherit from UUIDModelMixin. If you want a fully random UUID, you can set 'uuid_input_fields' to the string 'RANDOM'.""")
python
{ "resource": "" }
q260714
add_to_deleted_models
validation
def add_to_deleted_models(sender, instance=None, *args, **kwargs): """ Whenever a model is deleted, we record its ID in a separate model for tracking purposes. During serialization, we will mark the model as deleted in
python
{ "resource": "" }
q260715
APIWrapper._with_error_handling
validation
def _with_error_handling(resp, error, mode, response_format): """ Static method for error handling. :param resp - API response :param error - Error thrown :param mode - Error mode :param response_format - XML or json """ def safe_parse(r): try: return APIWrapper._parse_resp(r, response_format) except (ValueError, SyntaxError) as ex: log.error(ex) r.parsed = None return r if isinstance(error, requests.HTTPError): if resp.status_code == 400: # It means that request parameters were rejected by the server, # so we need to enrich standard error message # with 'ValidationErrors' # from the response resp = safe_parse(resp) if resp.parsed is not None: parsed_resp = resp.parsed messages = [] if response_format == 'xml' and\ parsed_resp.find('./ValidationErrors') is not None: messages = [e.find('./Message').text for e in parsed_resp.findall('./ValidationErrors/ValidationErrorDto')] elif response_format == 'json' and 'ValidationErrors' in parsed_resp: messages = [e['Message'] for e in parsed_resp['ValidationErrors']] error = requests.HTTPError( '%s: %s' % (error, '\n\t'.join(messages)), response=resp) elif resp.status_code == 429: error = requests.HTTPError('%sToo many requests in the last minute.' % error,
python
{ "resource": "" }
q260716
APIWrapper._default_poll_callback
validation
def _default_poll_callback(self, poll_resp): """ Checks the condition in poll response to determine if it is complete and no subsequent poll requests should be done. """ if poll_resp.parsed is None: return False success_list = ['UpdatesComplete', True, 'COMPLETE']
python
{ "resource": "" }
q260717
_fsic_queuing_calc
validation
def _fsic_queuing_calc(fsic1, fsic2): """ We set the lower counter between two same instance ids. If an instance_id exists in one fsic but not the other we want to give that counter a value of 0. :param fsic1: dictionary containing (instance_id, counter) pairs :param fsic2: dictionary containing (instance_id, counter) pairs
python
{ "resource": "" }
q260718
_deserialize_from_store
validation
def _deserialize_from_store(profile): """ Takes data from the store and integrates into the application. """ # we first serialize to avoid deserialization merge conflicts _serialize_into_store(profile) fk_cache = {} with transaction.atomic(): syncable_dict = _profile_models[profile] excluded_list = [] # iterate through classes which are in foreign key dependency order for model_name, klass_model in six.iteritems(syncable_dict): # handle cases where a class has a single FK reference to itself self_ref_fk = _self_referential_fk(klass_model) query = Q(model_name=klass_model.morango_model_name) for klass in klass_model.morango_model_dependencies: query |= Q(model_name=klass.morango_model_name) if self_ref_fk: clean_parents = Store.objects.filter(dirty_bit=False, profile=profile).filter(query).char_ids_list() dirty_children = Store.objects.filter(dirty_bit=True, profile=profile) \ .filter(Q(_self_ref_fk__in=clean_parents) | Q(_self_ref_fk='')).filter(query) # keep iterating until size of dirty_children is 0 while len(dirty_children) > 0: for store_model in dirty_children: try: app_model = store_model._deserialize_store_model(fk_cache) if app_model: with mute_signals(signals.pre_save, signals.post_save): app_model.save(update_dirty_bit_to=False) # we update a store model after we have deserialized it to be able to mark it as a clean parent store_model.dirty_bit = False store_model.save(update_fields=['dirty_bit']) except exceptions.ValidationError: # if the app model did not validate, we leave the store dirty bit set excluded_list.append(store_model.id) # update lists with new clean parents and dirty children clean_parents = Store.objects.filter(dirty_bit=False, profile=profile).filter(query).char_ids_list() dirty_children = Store.objects.filter(dirty_bit=True, profile=profile, _self_ref_fk__in=clean_parents).filter(query) else: # array for holding db values from the fields of each model for this
python
{ "resource": "" }
q260719
_dequeue_into_store
validation
def _dequeue_into_store(transfersession): """ Takes data from the buffers and merges into the store and record max counters. """ with connection.cursor() as cursor: DBBackend._dequeuing_delete_rmcb_records(cursor, transfersession.id) DBBackend._dequeuing_delete_buffered_records(cursor, transfersession.id) current_id = InstanceIDModel.get_current_instance_and_increment_counter() DBBackend._dequeuing_merge_conflict_buffer(cursor, current_id, transfersession.id) DBBackend._dequeuing_merge_conflict_rmcb(cursor, transfersession.id) DBBackend._dequeuing_update_rmcs_last_saved_by(cursor, current_id, transfersession.id) DBBackend._dequeuing_delete_mc_rmcb(cursor, transfersession.id) DBBackend._dequeuing_delete_mc_buffer(cursor, transfersession.id)
python
{ "resource": "" }
q260720
max_parameter_substitution
validation
def max_parameter_substitution(): """ SQLite has a limit on the max number of variables allowed for parameter substitution. This limit is usually 999, but can be compiled to a different number. This function calculates what the max is for the sqlite version running on the device. We use the calculated value to chunk our SQL bulk insert statements when deserializing from the store to the app layer. """ if os.path.isfile(SQLITE_VARIABLE_FILE_CACHE): return conn = sqlite3.connect(':memory:') low = 1 high = 1000 # hard limit for SQLITE_MAX_VARIABLE_NUMBER <http://www.sqlite.org/limits.html> conn.execute('CREATE TABLE T1 (id C1)') while low < high - 1: guess = (low + high) // 2 try:
python
{ "resource": "" }
q260721
BasicMultiArgumentAuthentication.authenticate_credentials
validation
def authenticate_credentials(self, userargs, password, request=None): """ Authenticate the userargs and password against Django auth backends. The "userargs" string may be just the username, or a querystring-encoded set of params. """ credentials = { 'password': password } if "=" not in userargs: # if it doesn't seem to be in querystring format, just use it as the username credentials[get_user_model().USERNAME_FIELD] = userargs else: # parse out the user args from querystring format into the credentials dict for arg in userargs.split("&"):
python
{ "resource": "" }
q260722
_multiple_self_ref_fk_check
validation
def _multiple_self_ref_fk_check(class_model): """ We check whether a class has more than 1 FK reference to itself. """ self_fk = []
python
{ "resource": "" }
q260723
add_syncable_models
validation
def add_syncable_models(): """ Per profile, adds each model to a dictionary mapping the morango model name to its model class. We sort by ForeignKey dependencies to safely sync data. """ import django.apps from morango.models import SyncableModel from morango.manager import SyncableModelManager from morango.query import SyncableModelQuerySet model_list = [] for model_class in django.apps.apps.get_models(): # several validation checks to assert models will be syncing correctly if issubclass(model_class, SyncableModel): name = model_class.__name__ if _multiple_self_ref_fk_check(model_class): raise InvalidMorangoModelConfiguration("Syncing models with more than 1 self referential ForeignKey is not supported.") try: from mptt import models from morango.utils.morango_mptt import MorangoMPTTModel, MorangoMPTTTreeManager, MorangoTreeQuerySet # mptt syncable model checks if issubclass(model_class, models.MPTTModel): if not issubclass(model_class, MorangoMPTTModel): raise InvalidMorangoModelConfiguration("{} that inherits from MPTTModel, should instead inherit from MorangoMPTTModel.".format(name)) if not isinstance(model_class.objects, MorangoMPTTTreeManager):
python
{ "resource": "" }
q260724
NetworkSyncConnection._request
validation
def _request(self, endpoint, method="GET", lookup=None, data={}, params={}, userargs=None, password=None): """ Generic request method designed to handle any morango endpoint. :param endpoint: constant representing which morango endpoint we are querying :param method: HTTP verb/method for request :param lookup: the pk value for the specific object we are querying :param data: dict that will be form-encoded in request :param params: dict to be sent as part of URL's query string :param userargs: Authorization credentials :param password: :return: ``Response`` object from request """ # convert user arguments into query str for passing to auth layer
python
{ "resource": "" }
q260725
TokenGenerator.create_access_token
validation
def create_access_token(self, valid_in_hours=1, data=None): """ Creates an access token. TODO: check valid in hours TODO: maybe specify how often a token can be used """ data = data or {}
python
{ "resource": "" }
q260726
MongodbServiceStore.save_service
validation
def save_service(self, service, overwrite=True): """ Stores an OWS service in mongodb. """ name = namesgenerator.get_sane_name(service.name) if not name: name = namesgenerator.get_random_name() if self.collection.count_documents({'name': name}) > 0: name = namesgenerator.get_random_name(retry=True) # check if service is already registered if self.collection.count_documents({'name': name}) > 0: if overwrite: self.collection.delete_one({'name': name}) else:
python
{ "resource": "" }
q260727
MongodbServiceStore.list_services
validation
def list_services(self): """ Lists all services in mongodb storage. """ my_services = []
python
{ "resource": "" }
q260728
MongodbServiceStore.fetch_by_name
validation
def fetch_by_name(self, name): """ Gets service for given ``name`` from mongodb storage. """ service = self.collection.find_one({'name':
python
{ "resource": "" }
q260729
MongodbServiceStore.fetch_by_url
validation
def fetch_by_url(self, url): """ Gets service for given ``url`` from mongodb storage. """ service = self.collection.find_one({'url': url})
python
{ "resource": "" }
q260730
owsproxy_delegate
validation
def owsproxy_delegate(request): """ Delegates owsproxy request to external twitcher service. """ twitcher_url = request.registry.settings.get('twitcher.url') protected_path = request.registry.settings.get('twitcher.ows_proxy_protected_path', '/ows') url = twitcher_url + protected_path + '/proxy' if request.matchdict.get('service_name'): url += '/' + request.matchdict.get('service_name') if request.matchdict.get('access_token'): url += '/' + request.matchdict.get('service_name') url += '?' + urlparse.urlencode(request.params)
python
{ "resource": "" }
q260731
ows_security_tween_factory
validation
def ows_security_tween_factory(handler, registry): """A tween factory which produces a tween which raises an exception if access to OWS service is not allowed.""" security = owssecurity_factory(registry) def ows_security_tween(request):
python
{ "resource": "" }
q260732
includeme
validation
def includeme(config): """ The callable makes it possible to include rpcinterface in a Pyramid application. Calling ``config.include(twitcher.rpcinterface)`` will result in this callable being called. Arguments: * ``config``: the ``pyramid.config.Configurator`` object. """ settings = config.registry.settings if asbool(settings.get('twitcher.rpcinterface', True)): LOGGER.debug('Twitcher XML-RPC Interface enabled.') # include twitcher config config.include('twitcher.config') # using basic auth config.include('twitcher.basicauth') # pyramid xml-rpc # http://docs.pylonsproject.org/projects/pyramid-rpc/en/latest/xmlrpc.html config.include('pyramid_rpc.xmlrpc') config.include('twitcher.db') config.add_xmlrpc_endpoint('api', '/RPC2') # register xmlrpc methods config.add_xmlrpc_method(RPCInterface, attr='generate_token', endpoint='api', method='generate_token') config.add_xmlrpc_method(RPCInterface, attr='revoke_token', endpoint='api', method='revoke_token') config.add_xmlrpc_method(RPCInterface, attr='revoke_all_tokens', endpoint='api', method='revoke_all_tokens')
python
{ "resource": "" }
q260733
MemoryServiceStore.save_service
validation
def save_service(self, service, overwrite=True): """ Store an OWS service in database. """ name = namesgenerator.get_sane_name(service.name) if not name: name = namesgenerator.get_random_name() if name in self.name_index: name = namesgenerator.get_random_name(retry=True) # check if service is already registered if name in self.name_index: if overwrite: self._delete(name=name) else:
python
{ "resource": "" }
q260734
MemoryServiceStore.list_services
validation
def list_services(self): """ Lists all services in memory storage. """ my_services = [] for service
python
{ "resource": "" }
q260735
MemoryServiceStore.fetch_by_name
validation
def fetch_by_name(self, name): """ Get service for given ``name`` from memory storage. """ service
python
{ "resource": "" }
q260736
ESGFAccessManager._retrieve_certificate
validation
def _retrieve_certificate(self, access_token, timeout=3): """ Generates a new private key and certificate request, submits the request to be signed by the SLCS CA and returns the certificate. """ logger.debug("Retrieve certificate with token.") # Generate a new key pair key_pair = crypto.PKey() key_pair.generate_key(crypto.TYPE_RSA, 2048) private_key = crypto.dump_privatekey(crypto.FILETYPE_PEM, key_pair).decode("utf-8") # Generate a certificate request using that key-pair cert_request = crypto.X509Req() # Create public key object cert_request.set_pubkey(key_pair)
python
{ "resource": "" }
q260737
Get._get_param
validation
def _get_param(self, param, allowed_values=None, optional=False): """Get parameter in GET request.""" request_params = self._request_params() if param in request_params: value = request_params[param].lower() if allowed_values is not None: if value in allowed_values: self.params[param] = value else:
python
{ "resource": "" }
q260738
Get._get_version
validation
def _get_version(self): """Find requested version in GET request.""" version = self._get_param(param="version", allowed_values=allowed_versions[self.params['service']], optional=True)
python
{ "resource": "" }
q260739
Post._get_service
validation
def _get_service(self): """Check mandatory service name parameter in POST request.""" if "service" in self.document.attrib: value = self.document.attrib["service"].lower() if value in allowed_service_types: self.params["service"] = value else:
python
{ "resource": "" }
q260740
Post._get_request_type
validation
def _get_request_type(self): """Find requested request type in POST request.""" value = self.document.tag.lower() if value in allowed_request_types[self.params['service']]: self.params["request"] = value else:
python
{ "resource": "" }
q260741
Post._get_version
validation
def _get_version(self): """Find requested version in POST request.""" if "version" in self.document.attrib: value = self.document.attrib["version"].lower() if value in allowed_versions[self.params['service']]: self.params["version"] = value else: raise OWSInvalidParameterValue("Version %s is not
python
{ "resource": "" }
q260742
localize_datetime
validation
def localize_datetime(dt, tz_name='UTC'): """Provide a timzeone-aware object for a given datetime and timezone name """ tz_aware_dt = dt if dt.tzinfo is None: utc = pytz.timezone('UTC') aware = utc.localize(dt) timezone = pytz.timezone(tz_name)
python
{ "resource": "" }
q260743
baseurl
validation
def baseurl(url): """ return baseurl of given url """ parsed_url = urlparse.urlparse(url) if not parsed_url.netloc or parsed_url.scheme not in ("http", "https"): raise ValueError('bad url')
python
{ "resource": "" }
q260744
Service.verify
validation
def verify(self): """Verify ssl service certificate.""" value = self.get('verify', 'true') if isinstance(value, bool): verify = value elif value.lower() == 'true':
python
{ "resource": "" }
q260745
get_egg_info
validation
def get_egg_info(cfg, verbose=False): """Call 'setup egg_info' and return the parsed meta-data.""" result = Bunch() setup_py = cfg.rootjoin('setup.py') if not os.path.exists(setup_py): return result egg_info = shell.capture("python {} egg_info".format(setup_py), echo=True if verbose else None) for info_line in egg_info.splitlines(): if info_line.endswith('PKG-INFO'): pkg_info_file = info_line.split(None, 1)[1] result['__file__'] = pkg_info_file with io.open(pkg_info_file, encoding='utf-8') as handle: lastkey = None for line in handle: if line.lstrip() != line: assert lastkey, "Bad continuation in PKG-INFO file '{}': {}".format(pkg_info_file, line) result[lastkey] += '\n' + line else: lastkey, value = line.split(':', 1)
python
{ "resource": "" }
q260746
bump
validation
def bump(ctx, verbose=False, pypi=False): """Bump a development version.""" cfg = config.load() scm = scm_provider(cfg.project_root, commit=False, ctx=ctx) # Check for uncommitted changes if not scm.workdir_is_clean(): notify.warning("You have uncommitted changes, will create a time-stamped version!") pep440 = scm.pep440_dev_version(verbose=verbose, non_local=pypi) # Rewrite 'setup.cfg' TODO: refactor to helper, see also release-prep # with util.rewrite_file(cfg.rootjoin('setup.cfg')) as lines: # ... setup_cfg = cfg.rootjoin('setup.cfg') if not pep440: notify.info("Working directory contains a release version!") elif os.path.exists(setup_cfg): with io.open(setup_cfg, encoding='utf-8') as handle: data = handle.readlines() changed = False for i, line in enumerate(data): if re.match(r"#? *tag_build *= *.*", line): verb, _ = data[i].split('=', 1) data[i] = '{}= {}\n'.format(verb, pep440) changed = True if changed: notify.info("Rewriting 'setup.cfg'...") with io.open(setup_cfg, 'w', encoding='utf-8')
python
{ "resource": "" }
q260747
dist
validation
def dist(ctx, devpi=False, egg=False, wheel=False, auto=True): """Distribute the project.""" config.load() cmd = ["python", "setup.py", "sdist"] # Automatically create wheels if possible if auto: egg = sys.version_info.major == 2 try: import wheel as _ wheel = True except ImportError: wheel = False if
python
{ "resource": "" }
q260748
prep
validation
def prep(ctx, commit=True): """Prepare for a release.""" cfg = config.load() scm = scm_provider(cfg.project_root, commit=commit, ctx=ctx) # Check for uncommitted changes if not scm.workdir_is_clean(): notify.failure("You have uncommitted changes, please commit or stash them!") # TODO Check that changelog entry carries the current date # Rewrite 'setup.cfg' setup_cfg = cfg.rootjoin('setup.cfg') if os.path.exists(setup_cfg): with io.open(setup_cfg, encoding='utf-8') as handle: data = handle.readlines() changed = False for i, line in enumerate(data): if any(line.startswith(i) for i in ('tag_build', 'tag_date')): data[i] = '#' + data[i] changed = True if changed and commit: notify.info("Rewriting 'setup.cfg'...") with io.open(setup_cfg, 'w', encoding='utf-8') as handle: handle.write(''.join(data)) scm.add_file('setup.cfg') elif changed: notify.warning("WOULD rewrite 'setup.cfg', but --no-commit was passed") else: notify.warning("Cannot rewrite 'setup.cfg', none found!") # Update metadata and command stubs ctx.run('python setup.py -q develop -U') # Build a clean dist and check version number
python
{ "resource": "" }
q260749
pylint
validation
def pylint(ctx, skip_tests=False, skip_root=False, reports=False): """Perform source code checks via pylint.""" cfg = config.load() add_dir2pypath(cfg.project_root) if not os.path.exists(cfg.testjoin('__init__.py')): add_dir2pypath(cfg.testjoin()) namelist = set() for package in cfg.project.get('packages', []): if '.' not in package: namelist.add(cfg.srcjoin(package)) for module in cfg.project.get('py_modules', []): namelist.add(module + '.py') if not skip_tests: test_py = antglob.FileSet(cfg.testdir, '**/*.py') test_py = [cfg.testjoin(i) for i in test_py] if test_py: namelist |= set(test_py) if not skip_root: root_py = antglob.FileSet('.', '*.py') if root_py: namelist |= set(root_py) namelist = set([i[len(os.getcwd())+1:] if i.startswith(os.getcwd() + os.sep) else i for i in namelist]) cmd = 'pylint' cmd += ' "{}"'.format('" "'.join(sorted(namelist))) cmd += ' --reports={0}'.format('y' if reports else 'n') for cfgfile in ('.pylintrc', 'pylint.rc', 'pylint.cfg', 'project.d/pylint.cfg'): if os.path.exists(cfgfile): cmd += ' --rcfile={0}'.format(cfgfile) break try: shell.run(cmd, report_error=False, runner=ctx.run) notify.info("OK - No problems found by pylint.") except exceptions.Failure as exc: # Check bit flags within pylint return code
python
{ "resource": "" }
q260750
GitProvider.workdir_is_clean
validation
def workdir_is_clean(self, quiet=False): """ Check for uncommitted changes, return `True` if everything is clean. Inspired by http://stackoverflow.com/questions/3878624/. """ # Update the index self.run('git update-index -q --ignore-submodules --refresh', **RUN_KWARGS) unchanged = True # Disallow unstaged changes in the working tree try: self.run('git diff-files --quiet --ignore-submodules --', report_error=False, **RUN_KWARGS) except exceptions.Failure: unchanged = False if not quiet: notify.warning('You have unstaged changes!')
python
{ "resource": "" }
q260751
description
validation
def description(_dummy_ctx, markdown=False): """Dump project metadata for Jenkins Description Setter Plugin.""" cfg = config.load() markup = 'md' if markdown else 'html' description_file = cfg.rootjoin("build/project.{}".format(markup)) notify.banner("Creating {} file for Jenkins...".format(description_file)) long_description = cfg.project.long_description long_description = long_description.replace('\n\n', '</p>\n<p>') long_description = re.sub(r'(\W)``([^`]+)``(\W)', r'\1<tt>\2</tt>\3', long_description) text = DESCRIPTION_TEMPLATES[markup].format( keywords=', '.join(cfg.project.keywords),
python
{ "resource": "" }
q260752
capture
validation
def capture(cmd, **kw): """Run a command and return its stripped captured output.""" kw = kw.copy() kw['hide'] = 'out' if not kw.get('echo', False): kw['echo'] = False ignore_failures = kw.pop('ignore_failures', False) try: return invoke_run(cmd,
python
{ "resource": "" }
q260753
run
validation
def run(cmd, **kw): """Run a command and flush its output.""" kw = kw.copy() kw.setdefault('warn', False) # make extra sure errors don't get silenced report_error = kw.pop('report_error', True) runner = kw.pop('runner', invoke_run) try: return runner(cmd, **kw) except exceptions.Failure as exc: sys.stdout.flush()
python
{ "resource": "" }
q260754
auto_detect
validation
def auto_detect(workdir): """ Return string signifying the SCM used in the given directory. Currently, 'git' is supported. Anything else returns 'unknown'. """ # Any additions here also need a change
python
{ "resource": "" }
q260755
provider
validation
def provider(workdir, commit=True, **kwargs): """Factory for the correct SCM provider in `workdir`.""" return
python
{ "resource": "" }
q260756
fail
validation
def fail(message, exitcode=1): """Exit with error code and message."""
python
{ "resource": "" }
q260757
get_pypi_auth
validation
def get_pypi_auth(configfile='~/.pypirc'): """Read auth from pip config.""" pypi_cfg = ConfigParser() if pypi_cfg.read(os.path.expanduser(configfile)): try: user = pypi_cfg.get('pypi', 'username') pwd = pypi_cfg.get('pypi', 'password') return user, pwd
python
{ "resource": "" }
q260758
confluence
validation
def confluence(ctx, no_publish=False, clean=False, opts=''): """Build Sphinx docs and publish to Confluence.""" cfg = config.load() if clean: ctx.run("invoke clean --docs") cmd = ['sphinx-build', '-b', 'confluence'] cmd.extend(['-E', '-a']) # force a full rebuild if opts: cmd.append(opts) cmd.extend(['.', ctx.rituals.docs.build + '_cf'])
python
{ "resource": "" }
q260759
DocsUploader._zipped
validation
def _zipped(self, docs_base): """ Provide a zipped stream of the docs tree.""" with pushd(docs_base): with tempfile.NamedTemporaryFile(prefix='pythonhosted-', delete=False) as ziphandle: pass zip_name = shutil.make_archive(ziphandle.name, 'zip') notify.info("Uploading {:.1f} MiB from '{}' to '{}'..."
python
{ "resource": "" }
q260760
DocsUploader._to_pypi
validation
def _to_pypi(self, docs_base, release): """Upload to PyPI.""" url = None with self._zipped(docs_base) as handle: reply = requests.post(self.params['url'], auth=get_pypi_auth(), allow_redirects=False, files=dict(content=(self.cfg.project.name + '.zip', handle, 'application/zip')), data={':action': 'doc_upload', 'name': self.cfg.project.name}) if reply.status_code in range(200, 300): notify.info("{status_code} {reason}".format(**vars(reply))) elif reply.status_code ==
python
{ "resource": "" }
q260761
DocsUploader._to_webdav
validation
def _to_webdav(self, docs_base, release): """Upload to WebDAV store.""" try: git_path = subprocess.check_output('git remote get-url origin 2>/dev/null', shell=True) except subprocess.CalledProcessError: git_path = '' else: git_path = git_path.decode('ascii').strip() git_path = git_path.replace('http://', '').replace('https://', '').replace('ssh://', '') git_path = re.search(r'[^:/]+?[:/](.+)', git_path) git_path = git_path.group(1).replace('.git', '') if git_path else '' url = None with self._zipped(docs_base) as handle: url_ns = dict(name=self.cfg.project.name, version=release, git_path=git_path) reply = requests.put(self.params['url'].format(**url_ns), data=handle.read(), headers={'Accept': 'application/json'}) if reply.status_code in range(200, 300): notify.info("{status_code} {reason}".format(**vars(reply))) try: data = reply.json() except ValueError as
python
{ "resource": "" }
q260762
DocsUploader.upload
validation
def upload(self, docs_base, release): """Upload docs in ``docs_base`` to
python
{ "resource": "" }
q260763
search_file_upwards
validation
def search_file_upwards(name, base=None): """ Search for a file named `name` from cwd or given directory to root. Return None if nothing's found.
python
{ "resource": "" }
q260764
pushd
validation
def pushd(path): """ A context that enters a given directory and restores the old state on exit. The original directory is returned as the context variable. """ saved =
python
{ "resource": "" }
q260765
url_as_file
validation
def url_as_file(url, ext=None): """ Context manager that GETs a given `url` and provides it as a local file. The file is in a closed state upon entering the context, and removed when leaving it, if still there. To give the file name a specific extension, use `ext`; the extension can optionally include a separating dot, otherwise it will be added. Parameters: url (str): URL to retrieve. ext (str, optional): Extension for the generated filename. Yields: str: The path to a temporary file with the content of the URL. Raises: requests.RequestException: Base exception of ``requests``, see its
python
{ "resource": "" }
q260766
ProviderBase.run
validation
def run(self, cmd, *args, **kwargs): """Run a command."""
python
{ "resource": "" }
q260767
ProviderBase.run_elective
validation
def run_elective(self, cmd, *args, **kwargs): """Run a command, or just echo it, depending on `commit`.""" if self._commit: return self.run(cmd, *args,
python
{ "resource": "" }
q260768
info
validation
def info(msg): """Emit a normal message.""" _flush()
python
{ "resource": "" }
q260769
warning
validation
def warning(msg): """Emit a warning message.""" _flush()
python
{ "resource": "" }
q260770
error
validation
def error(msg): """Emit an error message to stderr.""" _flush()
python
{ "resource": "" }
q260771
get_devpi_url
validation
def get_devpi_url(ctx): """Get currently used 'devpi' base URL.""" cmd = 'devpi use --urls' lines = ctx.run(cmd, hide='out', echo=False).stdout.splitlines() for line in lines: try: line, base_url = line.split(':', 1)
python
{ "resource": "" }
q260772
get_project_root
validation
def get_project_root(): """ Determine location of `tasks.py`.""" try: tasks_py = sys.modules['tasks'] except KeyError: return None
python
{ "resource": "" }
q260773
load
validation
def load(): """ Load and return configuration as a ``Bunch``. Values are based on ``DEFAULTS``, and metadata from ``setup.py``. """ cfg = Bunch(DEFAULTS) # TODO: override with contents of [rituals] section in setup.cfg cfg.project_root = get_project_root() if not cfg.project_root: raise RuntimeError("No tasks module is imported, cannot determine project root") cfg.rootjoin = lambda *names: os.path.join(cfg.project_root, *names) cfg.srcjoin = lambda *names: cfg.rootjoin(cfg.srcdir, *names) cfg.testjoin = lambda *names: cfg.rootjoin(cfg.testdir, *names) cfg.cwd = os.getcwd() os.chdir(cfg.project_root)
python
{ "resource": "" }
q260774
glob2re
validation
def glob2re(part): """Convert a path part to regex syntax.""" return "[^/]*".join( re.escape(bit).replace(r'\[\^',
python
{ "resource": "" }
q260775
parse_glob
validation
def parse_glob(pattern): """Generate parts of regex transformed from glob pattern.""" if not pattern: return bits = pattern.split("/")
python
{ "resource": "" }
q260776
compile_glob
validation
def compile_glob(spec): """Convert the given glob `spec` to a compiled regex.""" parsed = "".join(parse_glob(spec))
python
{ "resource": "" }
q260777
FileSet.included
validation
def included(self, path, is_dir=False): """Check patterns in order, last match that includes or excludes `path` wins. Return `None` on undecided.""" inclusive = None for pattern in self.patterns: if pattern.is_dir == is_dir and pattern.matches(path):
python
{ "resource": "" }
q260778
FileSet.walk
validation
def walk(self, **kwargs): """ Like `os.walk` and taking the same keyword arguments, but generating paths relative to the root. Starts in the fileset's root and filters based on its patterns. If ``with_root=True`` is passed in, the generated paths include the root path. """ lead = '' if 'with_root' in kwargs and kwargs.pop('with_root'): lead = self.root.rstrip(os.sep) + os.sep for base, dirs, files in os.walk(self.root, **kwargs): prefix = base[len(self.root):].lstrip(os.sep) bits = prefix.split(os.sep) if prefix else [] for dirname in dirs[:]:
python
{ "resource": "" }
q260779
build
validation
def build(ctx, dput='', opts=''): """Build a DEB package.""" # Get package metadata with io.open('debian/changelog', encoding='utf-8') as changes: metadata = re.match(r'^([^ ]+) \(([^)]+)\) ([^;]+); urgency=(.+)$', changes.readline().rstrip()) if not metadata: notify.failure('Badly formatted top entry in changelog') name, version, _, _ = metadata.groups() # Build package ctx.run('dpkg-buildpackage {} {}'.format(ctx.rituals.deb.build.opts, opts)) # Move created artifacts into "dist" if not os.path.exists('dist'): os.makedirs('dist')
python
{ "resource": "" }
q260780
clean
validation
def clean(_dummy_ctx, docs=False, backups=False, bytecode=False, dist=False, # pylint: disable=redefined-outer-name all=False, venv=False, tox=False, extra=''): # pylint: disable=redefined-builtin """Perform house-keeping.""" cfg = config.load() notify.banner("Cleaning up project files") # Add patterns based on given parameters venv_dirs = ['bin', 'include', 'lib', 'share', 'local', '.venv'] patterns = ['build/', 'pip-selfcheck.json'] excludes = ['.git/', '.hg/', '.svn/', 'debian/*/'] if docs or all: patterns.extend(['docs/_build/', 'doc/_build/']) if dist or all: patterns.append('dist/') if backups or all: patterns.extend(['**/*~']) if bytecode
python
{ "resource": "" }
q260781
build
validation
def build(ctx, docs=False): """Build the project.""" cfg = config.load() ctx.run("python setup.py build") if docs: for doc_path in ('docs', 'doc'): if os.path.exists(cfg.rootjoin(doc_path, 'conf.py')): break else:
python
{ "resource": "" }
q260782
freeze
validation
def freeze(ctx, local=False): """Freeze currently installed requirements.""" cmd = 'pip --disable-pip-version-check freeze{}'.format(' --local' if local else '') frozen = ctx.run(cmd, hide='out').stdout.replace('\x1b', '#') with io.open('frozen-requirements.txt', 'w', encoding='ascii') as out:
python
{ "resource": "" }
q260783
isodate
validation
def isodate(datestamp=None, microseconds=False): """Return current or given time formatted according to ISO-8601.""" datestamp = datestamp or datetime.datetime.now() if not microseconds: usecs
python
{ "resource": "" }
q260784
_get_registered_executable
validation
def _get_registered_executable(exe_name): """Windows allow application paths to be registered in the registry.""" registered = None if sys.platform.startswith('win'): if os.path.splitext(exe_name)[1].lower() != '.exe': exe_name += '.exe' import _winreg # pylint: disable=import-error try: key = "SOFTWARE\\Microsoft\\Windows\\CurrentVersion\\App Paths\\" + exe_name
python
{ "resource": "" }
q260785
whichgen
validation
def whichgen(command, path=None, verbose=0, exts=None): # pylint: disable=too-many-branches, too-many-statements """Return a generator of full paths to the given command. "command" is a the name of the executable to search for. "path" is an optional alternate path list to search. The default it to use the PATH environment variable. "verbose", if true, will cause a 2-tuple to be returned for each match. The second element is a textual description of where the match was found. "exts" optionally allows one to specify a list of extensions to use instead of the standard list for this system. This can effectively be used as an optimization to, for example, avoid stat's of "foo.vbs" when searching for "foo" and you know it is not a VisualBasic script but ".vbs" is on PATHEXT. This option is only supported on Windows. This method returns a generator which yields either full paths to the given command or, if verbose, tuples of the form (<path to command>, <where path found>). """ matches = [] if path is None: using_given_path = 0 path = os.environ.get("PATH", "").split(os.pathsep) if sys.platform.startswith("win"): path.insert(0, os.curdir) # implied by Windows shell else: using_given_path = 1 # Windows has the concept of a list of extensions (PATHEXT env var). if sys.platform.startswith("win"): if exts is None: exts = os.environ.get("PATHEXT", "").split(os.pathsep) # If '.exe' is not in exts then obviously this is Win9x and # or a bogus PATHEXT, then use a reasonable default. for ext in exts: if ext.lower() == ".exe": break else: exts = ['.COM', '.EXE', '.BAT'] elif not isinstance(exts, list): raise TypeError("'exts' argument must be a list or None") else: if exts is not None: raise WhichError("'exts' argument is not supported on platform '%s'" % sys.platform) exts = [] # File name cannot have path separators because PATH lookup does not # work that way. if os.sep in command or os.altsep and os.altsep in command: pass
python
{ "resource": "" }
q260786
which
validation
def which(command, path=None, verbose=0, exts=None): """Return the full path to the first match of the given command on the path. "command" is a the name of the executable to search for. "path" is an optional alternate path list to search. The default it to use the PATH environment variable. "verbose", if true, will cause a 2-tuple to be returned. The second element is a textual description of where the match was found. "exts" optionally allows one to specify a list of extensions to use instead of the standard list for this system. This can effectively be used as an optimization to, for example, avoid stat's of "foo.vbs" when searching for "foo" and you know it is
python
{ "resource": "" }
q260787
SymmetricKeyRatchet.step
validation
def step(self, key, chain): """ Perform a rachted step, replacing one of the internally managed chains with a new one. :param key: A bytes-like object encoding the key to initialize the replacement chain with. :param chain: The chain to
python
{ "resource": "" }
q260788
DoubleRatchet.decryptMessage
validation
def decryptMessage(self, ciphertext, header, ad = None): """ Decrypt a message using this double ratchet session. :param ciphertext: A bytes-like object encoding the message to decrypt. :param header: An instance of the Header class. This should have been sent together with the ciphertext. :param ad: A bytes-like object encoding the associated data to use for message authentication. Pass None to use the associated data set during construction. :returns: The plaintext. :raises AuthenticationFailedException: If checking the authentication for this message failed. :raises NotInitializedException: If this double ratchet session is not yet initialized with a key pair, thus not prepared to decrypt an incoming message. :raises TooManySavedMessageKeysException: If more than message_key_store_max have to be stored to decrypt this message. """ if ad == None: ad = self.__ad # Try to decrypt the message using a previously saved message key plaintext = self.__decryptSavedMessage(ciphertext, header, ad) if plaintext:
python
{ "resource": "" }
q260789
DoubleRatchet.encryptMessage
validation
def encryptMessage(self, message, ad = None): """ Encrypt a message using this double ratchet session. :param message: A bytes-like object encoding the message to encrypt. :param ad: A bytes-like object encoding the associated data to use for message authentication. Pass None to use the associated data set during construction. :returns: A dictionary containing the message header and ciphertext. The header is required to synchronize the double ratchet of the receiving party. Send it along with the ciphertext. The returned dictionary consists of two keys: "header", which includes an instance of the Header class and "ciphertext", which includes the encrypted message encoded as a bytes-like object. :raises NotInitializedException: If this double ratchet session is not yet initialized with the other parties public key, thus not ready to encrypt a message to that party. """ if ad == None: ad = self.__ad
python
{ "resource": "" }
q260790
DHRatchet.step
validation
def step(self, other_pub): """ Perform a rachted step, calculating a new shared secret from the public key and deriving new chain keys from this secret. New Diffie-Hellman calculations are only performed if the public key is different
python
{ "resource": "" }
q260791
KDFChain.next
validation
def next(self, data): """ Derive a new set of internal and output data from given input data and the data stored internally. Use the key derivation function to derive new data. The kdf gets supplied with the current key and the data passed to this method. :param data: A bytes-like object encoding the data to pass to the key derivation
python
{ "resource": "" }
q260792
Mesh.connect_to
validation
def connect_to(self, other_mesh): """Create a connection to an other mesh. .. warning:: Both meshes need to be disconnected and one needs to be a consumed and the other a produced mesh. You
python
{ "resource": "" }
q260793
Mesh.can_connect_to
validation
def can_connect_to(self, other): """Whether a connection can be established between those two meshes.""" assert other.is_mesh() disconnected = not other.is_connected() and not self.is_connected()
python
{ "resource": "" }
q260794
new_knitting_pattern_set_loader
validation
def new_knitting_pattern_set_loader(specification=DefaultSpecification()): """Create a loader for a knitting pattern set. :param specification: a :class:`specification <knittingpattern.ParsingSpecification.ParsingSpecification>` for the knitting pattern set, default
python
{ "resource": "" }
q260795
walk
validation
def walk(knitting_pattern): """Walk the knitting pattern in a right-to-left fashion. :return: an iterable to walk the rows :rtype: list :param knittingpattern.KnittingPattern.KnittingPattern knitting_pattern: a knitting pattern to take the rows from """ rows_before = {} # key consumes from values free_rows = [] walk = [] for row in knitting_pattern.rows: rows_before_ = row.rows_before[:] if rows_before_: rows_before[row] = rows_before_ else: free_rows.append(row) assert free_rows while free_rows: # print("free rows:", free_rows)
python
{ "resource": "" }
q260796
ContentDumper.file
validation
def file(self, file=None): """Saves the dump in a file-like object in text mode. :param file: :obj:`None` or a file-like object. :return: a file-like object If :paramref:`file` is :obj:`None`, a new :class:`io.StringIO` is returned. If :paramref:`file` is not :obj:`None` it should be a file-like object.
python
{ "resource": "" }
q260797
ContentDumper._file
validation
def _file(self, file): """Dump the content to a `file`. """ if not self.__text_is_expected: file
python
{ "resource": "" }
q260798
ContentDumper._binary_file
validation
def _binary_file(self, file): """Dump the ocntent into the `file` in binary mode.
python
{ "resource": "" }
q260799
ContentDumper._path
validation
def _path(self, path): """Saves the dump in a file named `path`.""" mode, encoding = self._mode_and_encoding_for_open()
python
{ "resource": "" }