id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
13,109
def get_version(package): init_py = open(os.path.join(package, '__init__.py')).read() return re.search('__version__ = [\'"]([^\'"]+)[\'"]', init_py).group(1)
[ "def", "get_version", "(", "package", ")", ":", "init_py", "=", "open", "(", "os", ".", "path", ".", "join", "(", "package", ",", "'__init__.py'", ")", ")", ".", "read", "(", ")", "return", "re", ".", "search", "(", "'__version__ = [\\'\"]([^\\'\"]+)[\\'\"...
parse __init__ .
train
true
13,110
def get_image(conn, vm_): vm_image = config.get_cloud_config_value('image', vm_, __opts__, default='').encode('ascii', 'salt-cloud-force-ascii') if (not vm_image): log.debug('No image set, must be boot from volume') return None image_list = conn.image_list() for img in image_list: if (vm_image in (image_list[img]['id'], img)): return image_list[img]['id'] try: image = conn.image_show(vm_image) return image['id'] except novaclient.exceptions.NotFound as exc: raise SaltCloudNotFound("The specified image, '{0}', could not be found: {1}".format(vm_image, str(exc)))
[ "def", "get_image", "(", "conn", ",", "vm_", ")", ":", "vm_image", "=", "config", ".", "get_cloud_config_value", "(", "'image'", ",", "vm_", ",", "__opts__", ",", "default", "=", "''", ")", ".", "encode", "(", "'ascii'", ",", "'salt-cloud-force-ascii'", ")...
return the image object to use .
train
true
13,111
def get_xmlsec_binary(paths=None): if (os.name == 'posix'): bin_name = ['xmlsec1'] elif (os.name == 'nt'): bin_name = ['xmlsec.exe', 'xmlsec1.exe'] else: bin_name = ['xmlsec1'] if paths: for bname in bin_name: for path in paths: fil = os.path.join(path, bname) try: if os.lstat(fil): return fil except OSError: pass for path in os.environ['PATH'].split(os.pathsep): for bname in bin_name: fil = os.path.join(path, bname) try: if os.lstat(fil): return fil except OSError: pass raise SigverError(("Can't find %s" % bin_name))
[ "def", "get_xmlsec_binary", "(", "paths", "=", "None", ")", ":", "if", "(", "os", ".", "name", "==", "'posix'", ")", ":", "bin_name", "=", "[", "'xmlsec1'", "]", "elif", "(", "os", ".", "name", "==", "'nt'", ")", ":", "bin_name", "=", "[", "'xmlsec...
tries to find the xmlsec1 binary .
train
true
13,112
def dnsrepr2names(x): res = [] cur = '' while x: l = ord(x[0]) x = x[1:] if (l == 0): if (cur and (cur[(-1)] == '.')): cur = cur[:(-1)] res.append(cur) cur = '' if (x and (ord(x[0]) == 0)): x = x[1:] continue if (l & 192): raise Exception("DNS message can't be compressed at this point!") else: cur += (x[:l] + '.') x = x[l:] return res
[ "def", "dnsrepr2names", "(", "x", ")", ":", "res", "=", "[", "]", "cur", "=", "''", "while", "x", ":", "l", "=", "ord", "(", "x", "[", "0", "]", ")", "x", "=", "x", "[", "1", ":", "]", "if", "(", "l", "==", "0", ")", ":", "if", "(", "...
take as input a dns encoded string and returns a list of dns names contained in it .
train
false
13,113
def loggers(): root = logging.root existing = root.manager.loggerDict.keys() return [logging.getLogger(name) for name in existing]
[ "def", "loggers", "(", ")", ":", "root", "=", "logging", ".", "root", "existing", "=", "root", ".", "manager", ".", "loggerDict", ".", "keys", "(", ")", "return", "[", "logging", ".", "getLogger", "(", "name", ")", "for", "name", "in", "existing", "]...
get list of all loggers .
train
false
13,114
def get_course_versions(course_key): course_locator = CourseKey.from_string(course_key) store = modulestore()._get_modulestore_for_courselike(course_locator) index_entry = store.get_course_index(course_locator) if (index_entry is not None): return {'draft-branch': index_entry['versions']['draft-branch'], 'published-branch': index_entry['versions']['published-branch']} return None
[ "def", "get_course_versions", "(", "course_key", ")", ":", "course_locator", "=", "CourseKey", ".", "from_string", "(", "course_key", ")", "store", "=", "modulestore", "(", ")", ".", "_get_modulestore_for_courselike", "(", "course_locator", ")", "index_entry", "=", ...
fetches the latest course versions .
train
false
13,115
def extract_first_line(paragraph): lines = paragraph.splitlines() first = lines[0].strip() spaces = list(re.finditer('(\\s+)', first)) if spaces: longest = max(spaces, key=(lambda m: (m.span()[1] - m.span()[0]))) if (longest and (longest.start() > 1) and ((longest.end() - longest.start()) > 1)): first = first[:longest.start()] return first
[ "def", "extract_first_line", "(", "paragraph", ")", ":", "lines", "=", "paragraph", ".", "splitlines", "(", ")", "first", "=", "lines", "[", "0", "]", ".", "strip", "(", ")", "spaces", "=", "list", "(", "re", ".", "finditer", "(", "'(\\\\s+)'", ",", ...
extract the first line information from the function code text if available .
train
false
13,117
def _parse_valid_types_from_validator(validator): if (not isinstance(validator, list)): return anglicize_json_type(validator) if (len(validator) == 1): return anglicize_json_type(validator[0]) return u'{}, or {}'.format(u', '.join(([anglicize_json_type(validator[0])] + validator[1:(-1)])), anglicize_json_type(validator[(-1)]))
[ "def", "_parse_valid_types_from_validator", "(", "validator", ")", ":", "if", "(", "not", "isinstance", "(", "validator", ",", "list", ")", ")", ":", "return", "anglicize_json_type", "(", "validator", ")", "if", "(", "len", "(", "validator", ")", "==", "1", ...
a validator value can be either an array of valid types or a string of a valid type .
train
false
13,118
def get_idx_from_sent(sent, word_idx_map, max_l=51, k=300, filter_h=5): x = [] pad = (filter_h - 1) for i in xrange(pad): x.append(0) words = sent.split() for word in words: if (word in word_idx_map): x.append(word_idx_map[word]) while (len(x) < (max_l + (2 * pad))): x.append(0) return x
[ "def", "get_idx_from_sent", "(", "sent", ",", "word_idx_map", ",", "max_l", "=", "51", ",", "k", "=", "300", ",", "filter_h", "=", "5", ")", ":", "x", "=", "[", "]", "pad", "=", "(", "filter_h", "-", "1", ")", "for", "i", "in", "xrange", "(", "...
transforms sentence into a list of indices .
train
false
13,119
def decodeHexValue(value, raw=False): retVal = value def _(value): retVal = value if (value and isinstance(value, basestring) and ((len(value) % 2) == 0)): retVal = hexdecode(retVal) if ((not kb.binaryField) and (not raw)): if (Backend.isDbms(DBMS.MSSQL) and value.startswith('0x')): try: retVal = retVal.decode('utf-16-le') except UnicodeDecodeError: pass elif Backend.isDbms(DBMS.HSQLDB): try: retVal = retVal.decode('utf-16-be') except UnicodeDecodeError: pass if (not isinstance(retVal, unicode)): retVal = getUnicode(retVal, 'utf8') return retVal try: retVal = applyFunctionRecursively(value, _) except: singleTimeWarnMessage(("there was a problem decoding value '%s' from expected hexadecimal form" % value)) return retVal
[ "def", "decodeHexValue", "(", "value", ",", "raw", "=", "False", ")", ":", "retVal", "=", "value", "def", "_", "(", "value", ")", ":", "retVal", "=", "value", "if", "(", "value", "and", "isinstance", "(", "value", ",", "basestring", ")", "and", "(", ...
returns value decoded from dbms specific hexadecimal representation .
train
false
13,121
def gpu_flatten(x, outdim=1): x = as_cuda_ndarray_variable(x) if (outdim > 1): dims = (tuple(x.shape[:(outdim - 1)]) + ((-1),)) else: dims = ((-1),) return GpuReshape(outdim)(x, dims)
[ "def", "gpu_flatten", "(", "x", ",", "outdim", "=", "1", ")", ":", "x", "=", "as_cuda_ndarray_variable", "(", "x", ")", "if", "(", "outdim", ">", "1", ")", ":", "dims", "=", "(", "tuple", "(", "x", ".", "shape", "[", ":", "(", "outdim", "-", "1...
implement flatten on the gpu .
train
false
13,123
@register.function def get_compat_counts(addon): return CompatReport.get_counts(addon.guid)
[ "@", "register", ".", "function", "def", "get_compat_counts", "(", "addon", ")", ":", "return", "CompatReport", ".", "get_counts", "(", "addon", ".", "guid", ")" ]
get counts for add-on compatibility reports .
train
false
13,124
def break_simplesub(ctext, startkey=None): ctext = re.sub('[^A-Z]', '', ctext.upper()) (parentkey, parentscore) = ((startkey or list('ABCDEFGHIJKLMNOPQRSTUVWXYZ')), (-9.9e+100)) if (not startkey): random.shuffle(parentkey) parentscore = fitness.score(sub_decipher(ctext, parentkey)) count = 0 while (count < 1000): a = random.randint(0, 25) b = random.randint(0, 25) child = parentkey[:] (child[a], child[b]) = (child[b], child[a]) score = fitness.score(sub_decipher(ctext, child)) if (score > parentscore): (parentscore, parentkey) = (score, child[:]) count = 0 count += 1 return (parentscore, parentkey)
[ "def", "break_simplesub", "(", "ctext", ",", "startkey", "=", "None", ")", ":", "ctext", "=", "re", ".", "sub", "(", "'[^A-Z]'", ",", "''", ",", "ctext", ".", "upper", "(", ")", ")", "(", "parentkey", ",", "parentscore", ")", "=", "(", "(", "startk...
perform hill-climbing with a single start .
train
false
13,125
def _getWriters(reactor): if IReactorFDSet.providedBy(reactor): return reactor.getWriters() elif ('IOCP' in reactor.__class__.__name__): return reactor.handles else: raise Exception(('Cannot find writers on %r' % (reactor,)))
[ "def", "_getWriters", "(", "reactor", ")", ":", "if", "IReactorFDSet", ".", "providedBy", "(", "reactor", ")", ":", "return", "reactor", ".", "getWriters", "(", ")", "elif", "(", "'IOCP'", "in", "reactor", ".", "__class__", ".", "__name__", ")", ":", "re...
like l{ireactorfdset .
train
false
13,126
@require_context def mapped_array(shape, dtype=np.float, strides=None, order='C', stream=0, portable=False, wc=False): (shape, strides, dtype) = _prepare_shape_strides_dtype(shape, strides, dtype, order) bytesize = driver.memory_size_from_info(shape, strides, dtype.itemsize) buffer = current_context().memhostalloc(bytesize, mapped=True) npary = np.ndarray(shape=shape, strides=strides, dtype=dtype, order=order, buffer=buffer) mappedview = np.ndarray.view(npary, type=devicearray.MappedNDArray) mappedview.device_setup(buffer, stream=stream) return mappedview
[ "@", "require_context", "def", "mapped_array", "(", "shape", ",", "dtype", "=", "np", ".", "float", ",", "strides", "=", "None", ",", "order", "=", "'C'", ",", "stream", "=", "0", ",", "portable", "=", "False", ",", "wc", "=", "False", ")", ":", "(...
mapped_array allocate a mapped ndarray with a buffer that is pinned and mapped on to the device .
train
false
13,128
def tailProbability(x, distributionParams): if (('mean' not in distributionParams) or ('stdev' not in distributionParams)): raise RuntimeError('Insufficient parameters to specify the distribution.') if (x < distributionParams['mean']): xp = ((2 * distributionParams['mean']) - x) return tailProbability(xp, distributionParams) z = ((x - distributionParams['mean']) / distributionParams['stdev']) return (0.5 * math.erfc((z / 1.4142)))
[ "def", "tailProbability", "(", "x", ",", "distributionParams", ")", ":", "if", "(", "(", "'mean'", "not", "in", "distributionParams", ")", "or", "(", "'stdev'", "not", "in", "distributionParams", ")", ")", ":", "raise", "RuntimeError", "(", "'Insufficient para...
given the normal distribution specified by the mean and standard deviation in distributionparams .
train
true
13,129
def test_coil_trans(): rng = np.random.RandomState(0) x = rng.randn(4, 4) x[3] = [0, 0, 0, 1] assert_allclose(_loc_to_coil_trans(_coil_trans_to_loc(x)), x) x = rng.randn(12) assert_allclose(_coil_trans_to_loc(_loc_to_coil_trans(x)), x)
[ "def", "test_coil_trans", "(", ")", ":", "rng", "=", "np", ".", "random", ".", "RandomState", "(", "0", ")", "x", "=", "rng", ".", "randn", "(", "4", ",", "4", ")", "x", "[", "3", "]", "=", "[", "0", ",", "0", ",", "0", ",", "1", "]", "as...
test loc<->coil_trans functions .
train
false
13,130
@task() @timeit def send_award_notification(award): @email_utils.safe_translation def _make_mail(locale, context, email): subject = _(u"You were awarded the '{title}' badge!").format(title=pgettext('DB: badger.Badge.title', award.badge.title)) mail = email_utils.make_mail(subject=subject, text_template='kbadge/email/award_notification.ltxt', html_template='kbadge/email/award_notification.html', context_vars=context, from_email=settings.DEFAULT_FROM_EMAIL, to_email=email) return mail msg = _make_mail(locale=award.user.profile.locale, context={'host': Site.objects.get_current().domain, 'award': award, 'badge': award.badge}, email=award.user.email) email_utils.send_messages([msg])
[ "@", "task", "(", ")", "@", "timeit", "def", "send_award_notification", "(", "award", ")", ":", "@", "email_utils", ".", "safe_translation", "def", "_make_mail", "(", "locale", ",", "context", ",", "email", ")", ":", "subject", "=", "_", "(", "u\"You were ...
sends the award notification email :arg award: the django-badger award instance .
train
false
13,131
def _job_dir(): return os.path.join(__opts__['cachedir'], 'jobs')
[ "def", "_job_dir", "(", ")", ":", "return", "os", ".", "path", ".", "join", "(", "__opts__", "[", "'cachedir'", "]", ",", "'jobs'", ")" ]
return root of the jobs cache directory .
train
false
13,132
def artifactory_maven(registry, xml_parent, data): artifactory = XML.SubElement(xml_parent, 'org.jfrog.hudson.maven3.ArtifactoryMaven3NativeConfigurator') details = XML.SubElement(artifactory, 'details') artifactory_common_details(details, data) if ('repo-key' in data): XML.SubElement(details, 'downloadRepositoryKey').text = data['repo-key'] else: XML.SubElement(details, 'downloadSnapshotRepositoryKey').text = data.get('snapshot-repo-key', '') XML.SubElement(details, 'downloadReleaseRepositoryKey').text = data.get('release-repo-key', '')
[ "def", "artifactory_maven", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "artifactory", "=", "XML", ".", "SubElement", "(", "xml_parent", ",", "'org.jfrog.hudson.maven3.ArtifactoryMaven3NativeConfigurator'", ")", "details", "=", "XML", ".", "SubElement",...
yaml: artifactory-maven wrapper for non-maven projects .
train
false
13,133
def _log_multivariate_normal_density_spherical(X, means, covars): cv = covars.copy() if (covars.ndim == 1): cv = cv[:, np.newaxis] if (cv.shape[1] == 1): cv = np.tile(cv, (1, X.shape[(-1)])) return _log_multivariate_normal_density_diag(X, means, cv)
[ "def", "_log_multivariate_normal_density_spherical", "(", "X", ",", "means", ",", "covars", ")", ":", "cv", "=", "covars", ".", "copy", "(", ")", "if", "(", "covars", ".", "ndim", "==", "1", ")", ":", "cv", "=", "cv", "[", ":", ",", "np", ".", "new...
compute gaussian log-density at x for a spherical model .
train
true
13,134
def _size_map(size): try: if (not isinstance(size, int)): if re.search('[Kk]', size): size = (1024 * float(re.sub('[Kk]', '', size))) elif re.search('[Mm]', size): size = ((1024 ** 2) * float(re.sub('[Mm]', '', size))) size = int(size) return size except: return None
[ "def", "_size_map", "(", "size", ")", ":", "try", ":", "if", "(", "not", "isinstance", "(", "size", ",", "int", ")", ")", ":", "if", "re", ".", "search", "(", "'[Kk]'", ",", "size", ")", ":", "size", "=", "(", "1024", "*", "float", "(", "re", ...
map bcaches size strings to real bytes .
train
false
13,135
def check_exclusive_options(**kwargs): if (not kwargs): return pretty_keys = kwargs.pop('pretty_keys', True) exclusive_options = {} for (k, v) in kwargs.items(): if (v is not None): exclusive_options[k] = True if (len(exclusive_options) > 1): if pretty_keys: names = [k.replace('_', ' ') for k in kwargs.keys()] else: names = kwargs.keys() names = ', '.join(sorted(names)) msg = (_('May specify only one of %s') % names) raise exception.InvalidInput(reason=msg)
[ "def", "check_exclusive_options", "(", "**", "kwargs", ")", ":", "if", "(", "not", "kwargs", ")", ":", "return", "pretty_keys", "=", "kwargs", ".", "pop", "(", "'pretty_keys'", ",", "True", ")", "exclusive_options", "=", "{", "}", "for", "(", "k", ",", ...
checks that only one of the provided options is actually not-none .
train
false
13,136
def validate_config_against_schema(config_schema, config_object, config_path, pack_name=None): pack_name = (pack_name or 'unknown') schema = util_schema.get_schema_for_resource_parameters(parameters_schema=config_schema, allow_additional_properties=True) instance = config_object try: cleaned = util_schema.validate(instance=instance, schema=schema, cls=util_schema.CustomValidator, use_default=True, allow_default_none=True) except jsonschema.ValidationError as e: attribute = getattr(e, 'path', []) attribute = '.'.join(attribute) msg = ('Failed validating attribute "%s" in config for pack "%s" (%s): %s' % (attribute, pack_name, config_path, str(e))) raise jsonschema.ValidationError(msg) return cleaned
[ "def", "validate_config_against_schema", "(", "config_schema", ",", "config_object", ",", "config_path", ",", "pack_name", "=", "None", ")", ":", "pack_name", "=", "(", "pack_name", "or", "'unknown'", ")", "schema", "=", "util_schema", ".", "get_schema_for_resource_...
validate provided config dictionary against the provided config schema dictionary .
train
false
13,137
def sfilter(pred, brule): def filtered_brl(expr): for x in filter(pred, brule(expr)): (yield x) return filtered_brl
[ "def", "sfilter", "(", "pred", ",", "brule", ")", ":", "def", "filtered_brl", "(", "expr", ")", ":", "for", "x", "in", "filter", "(", "pred", ",", "brule", "(", "expr", ")", ")", ":", "(", "yield", "x", ")", "return", "filtered_brl" ]
yield only those results which satisfy the predicate .
train
false
13,139
def _prepare_write_tfr(tfr, condition): return (condition, dict(times=tfr.times, freqs=tfr.freqs, data=tfr.data, info=tfr.info, nave=tfr.nave, comment=tfr.comment, method=tfr.method))
[ "def", "_prepare_write_tfr", "(", "tfr", ",", "condition", ")", ":", "return", "(", "condition", ",", "dict", "(", "times", "=", "tfr", ".", "times", ",", "freqs", "=", "tfr", ".", "freqs", ",", "data", "=", "tfr", ".", "data", ",", "info", "=", "t...
aux function .
train
false
13,140
def apply_matrix_norm(m, v): (a, b, c, d, e, f) = m (p, q) = v return (((a * p) + (c * q)), ((b * p) + (d * q)))
[ "def", "apply_matrix_norm", "(", "m", ",", "v", ")", ":", "(", "a", ",", "b", ",", "c", ",", "d", ",", "e", ",", "f", ")", "=", "m", "(", "p", ",", "q", ")", "=", "v", "return", "(", "(", "(", "a", "*", "p", ")", "+", "(", "c", "*", ...
equivalent to apply_matrix_pt(m .
train
true
13,141
def handle_file_upload(path, file, site): uploadedfile = None try: file_path = os.path.join(path, file.name) uploadedfile = site.storage.save(file_path, file) except Exception as inst: raise inst return uploadedfile
[ "def", "handle_file_upload", "(", "path", ",", "file", ",", "site", ")", ":", "uploadedfile", "=", "None", "try", ":", "file_path", "=", "os", ".", "path", ".", "join", "(", "path", ",", "file", ".", "name", ")", "uploadedfile", "=", "site", ".", "st...
handle file upload .
train
false
13,145
def rc_file_defaults(): rcParams.update(rcParamsOrig)
[ "def", "rc_file_defaults", "(", ")", ":", "rcParams", ".", "update", "(", "rcParamsOrig", ")" ]
restore the default rc params from the original matplotlib rc that was loaded .
train
false
13,146
def get_channel_max_user_count(channel=14, **kwargs): access = get_user_access(channel=channel, uid=1, **kwargs) return access['channel_info']['max_user_count']
[ "def", "get_channel_max_user_count", "(", "channel", "=", "14", ",", "**", "kwargs", ")", ":", "access", "=", "get_user_access", "(", "channel", "=", "channel", ",", "uid", "=", "1", ",", "**", "kwargs", ")", "return", "access", "[", "'channel_info'", "]",...
get max users in channel .
train
true
13,147
def get_model_root(model): if model._meta.parents: parent_model = list(model._meta.parents.items())[0][0] return get_model_root(parent_model) return model
[ "def", "get_model_root", "(", "model", ")", ":", "if", "model", ".", "_meta", ".", "parents", ":", "parent_model", "=", "list", "(", "model", ".", "_meta", ".", "parents", ".", "items", "(", ")", ")", "[", "0", "]", "[", "0", "]", "return", "get_mo...
this function finds the root model for any given model .
train
false
13,148
def remove_cert(name, thumbprint, context=_DEFAULT_CONTEXT, store=_DEFAULT_STORE): ret = {'name': name, 'changes': dict(), 'comment': str(), 'result': None} store_path = 'Cert:\\{0}\\{1}'.format(context, store) current_certs = __salt__['win_pki.get_certs'](context=context, store=store) if (thumbprint not in current_certs): ret['comment'] = "Certificate '{0}' already removed from store: {1}".format(thumbprint, store_path) ret['result'] = True elif __opts__['test']: ret['comment'] = "Certificate '{0}' will be removed from store: {1}".format(thumbprint, store_path) ret['changes'] = {'old': thumbprint, 'new': None} else: ret['changes'] = {'old': thumbprint, 'new': None} ret['result'] = __salt__['win_pki.remove_cert'](thumbprint=thumbprint, context=context, store=store) if ret['result']: ret['comment'] = "Certificate '{0}' removed from store: {1}".format(thumbprint, store_path) else: ret['comment'] = "Certificate '{0}' unable to be removed from store: {1}".format(thumbprint, store_path) return ret
[ "def", "remove_cert", "(", "name", ",", "thumbprint", ",", "context", "=", "_DEFAULT_CONTEXT", ",", "store", "=", "_DEFAULT_STORE", ")", ":", "ret", "=", "{", "'name'", ":", "name", ",", "'changes'", ":", "dict", "(", ")", ",", "'comment'", ":", "str", ...
remove the certificate from the given certificate store .
train
false
13,149
def getTagBracketedProcedure(procedure): return getTagBracketedLine('procedureName', procedure)
[ "def", "getTagBracketedProcedure", "(", "procedure", ")", ":", "return", "getTagBracketedLine", "(", "'procedureName'", ",", "procedure", ")" ]
get line with a begin procedure tag .
train
false
13,151
def _expand_integer(ty): if isinstance(ty, types.Integer): if ty.signed: return max(types.intp, ty) else: return max(types.uintp, ty) elif isinstance(ty, types.Boolean): return types.intp else: return ty
[ "def", "_expand_integer", "(", "ty", ")", ":", "if", "isinstance", "(", "ty", ",", "types", ".", "Integer", ")", ":", "if", "ty", ".", "signed", ":", "return", "max", "(", "types", ".", "intp", ",", "ty", ")", "else", ":", "return", "max", "(", "...
if *ty* is an integer .
train
false
13,152
def imsiDetachIndication(): a = TpPd(pd=5) b = MessageType(mesType=1) c = MobileStationClassmark1() d = MobileId() packet = (((a / b) / c) / d) return packet
[ "def", "imsiDetachIndication", "(", ")", ":", "a", "=", "TpPd", "(", "pd", "=", "5", ")", "b", "=", "MessageType", "(", "mesType", "=", "1", ")", "c", "=", "MobileStationClassmark1", "(", ")", "d", "=", "MobileId", "(", ")", "packet", "=", "(", "("...
imsi detach indication section 9 .
train
true
13,156
def save_config_value(request, response, key, value): request.session[key] = value response.set_cookie(key, value, expires=one_year_from_now()) return response
[ "def", "save_config_value", "(", "request", ",", "response", ",", "key", ",", "value", ")", ":", "request", ".", "session", "[", "key", "]", "=", "value", "response", ".", "set_cookie", "(", "key", ",", "value", ",", "expires", "=", "one_year_from_now", ...
sets value of key key to value in both session and cookies .
train
true
13,157
def test_hist_flush_with_store_stdout(hist, xonsh_builtins): hf = hist.flush() assert (hf is None) xonsh_builtins.__xonsh_env__['HISTCONTROL'] = set() xonsh_builtins.__xonsh_env__['XONSH_STORE_STDOUT'] = True hist.append({'inp': 'still alive?', 'rtn': 0, 'out': 'yes'}) hf = hist.flush() assert (hf is not None) while hf.is_alive(): pass with LazyJSON(hist.filename) as lj: assert (len(lj['cmds']) == 1) assert (lj['cmds'][0]['inp'] == 'still alive?') assert (lj['cmds'][0]['out'].strip() == 'yes')
[ "def", "test_hist_flush_with_store_stdout", "(", "hist", ",", "xonsh_builtins", ")", ":", "hf", "=", "hist", ".", "flush", "(", ")", "assert", "(", "hf", "is", "None", ")", "xonsh_builtins", ".", "__xonsh_env__", "[", "'HISTCONTROL'", "]", "=", "set", "(", ...
verify explicit flushing of the history works .
train
false
13,158
def _failhard(): raise FileserverConfigError('Failed to load hg fileserver backend')
[ "def", "_failhard", "(", ")", ":", "raise", "FileserverConfigError", "(", "'Failed to load hg fileserver backend'", ")" ]
fatal fileserver configuration issue .
train
false
13,159
def truncate_too_long_number(numobj): if is_valid_number(numobj): return True numobj_copy = PhoneNumber() numobj_copy.merge_from(numobj) national_number = numobj.national_number while (not is_valid_number(numobj_copy)): national_number = (national_number // 10) numobj_copy.national_number = national_number validation_result = is_possible_number_with_reason(numobj_copy) if ((validation_result == ValidationResult.TOO_SHORT) or (national_number == 0)): return False numobj.national_number = national_number return True
[ "def", "truncate_too_long_number", "(", "numobj", ")", ":", "if", "is_valid_number", "(", "numobj", ")", ":", "return", "True", "numobj_copy", "=", "PhoneNumber", "(", ")", "numobj_copy", ".", "merge_from", "(", "numobj", ")", "national_number", "=", "numobj", ...
truncate a number object that is too long .
train
true
13,161
def disable_dhcp(): current = network() if (current['Network Settings']['DHCP_ENABLE']['VALUE'] == 'N'): return True _xml = '<RIBCL VERSION="2.0">\n <LOGIN USER_LOGIN="adminname" PASSWORD="password">\n <RIB_INFO MODE="write">\n <MOD_NETWORK_SETTINGS>\n <DHCP_ENABLE value="No"/>\n </MOD_NETWORK_SETTINGS>\n </RIB_INFO>\n </LOGIN>\n </RIBCL>' return __execute_cmd('Disable_DHCP', _xml)
[ "def", "disable_dhcp", "(", ")", ":", "current", "=", "network", "(", ")", "if", "(", "current", "[", "'Network Settings'", "]", "[", "'DHCP_ENABLE'", "]", "[", "'VALUE'", "]", "==", "'N'", ")", ":", "return", "True", "_xml", "=", "'<RIBCL VERSION=\"2.0\">...
disable dhcp cli example: .
train
false
13,162
def reload_loadbalancers(group, load_balancers, exclude=None): exclude = (exclude or []) id_list = grouputils.get_member_refids(group, exclude=exclude) for (name, lb) in six.iteritems(load_balancers): props = copy.copy(lb.properties.data) if ('Instances' in lb.properties_schema): props['Instances'] = id_list elif ('members' in lb.properties_schema): props['members'] = id_list else: raise exception.Error((_("Unsupported resource '%s' in LoadBalancerNames") % name)) lb_defn = rsrc_defn.ResourceDefinition(lb.name, lb.type(), properties=props, metadata=lb.t.metadata(), deletion_policy=lb.t.deletion_policy()) scheduler.TaskRunner(lb.update, lb_defn)()
[ "def", "reload_loadbalancers", "(", "group", ",", "load_balancers", ",", "exclude", "=", "None", ")", ":", "exclude", "=", "(", "exclude", "or", "[", "]", ")", "id_list", "=", "grouputils", ".", "get_member_refids", "(", "group", ",", "exclude", "=", "excl...
notify the loadbalancer to reload its config .
train
false
13,164
def powerset(iterable, include_empty=True): s = list(iterable) i = chain.from_iterable((combinations(s, r) for r in range((len(s) + 1)))) if (not include_empty): next(i) return i
[ "def", "powerset", "(", "iterable", ",", "include_empty", "=", "True", ")", ":", "s", "=", "list", "(", "iterable", ")", "i", "=", "chain", ".", "from_iterable", "(", "(", "combinations", "(", "s", ",", "r", ")", "for", "r", "in", "range", "(", "("...
generates all subsets of a set or sequence u .
train
false
13,165
def load_opts_from_mrjob_confs(runner_alias, conf_paths=None): if (conf_paths is None): return load_opts_from_mrjob_conf(runner_alias) else: already_loaded = [] results = [] for path in reversed(conf_paths): results = (load_opts_from_mrjob_conf(runner_alias, path, already_loaded=already_loaded) + results) return results
[ "def", "load_opts_from_mrjob_confs", "(", "runner_alias", ",", "conf_paths", "=", "None", ")", ":", "if", "(", "conf_paths", "is", "None", ")", ":", "return", "load_opts_from_mrjob_conf", "(", "runner_alias", ")", "else", ":", "already_loaded", "=", "[", "]", ...
load a list of dictionaries representing the options in a given list of mrjob config files for a specific runner .
train
false
13,166
def instance_metadata_delete(context, instance_uuid, key): IMPL.instance_metadata_delete(context, instance_uuid, key)
[ "def", "instance_metadata_delete", "(", "context", ",", "instance_uuid", ",", "key", ")", ":", "IMPL", ".", "instance_metadata_delete", "(", "context", ",", "instance_uuid", ",", "key", ")" ]
delete the given metadata item .
train
false
13,169
def zpkfreqd(z, p, k, worN=None): if ((worN is None) or isinstance(worN, int)): N = (worN or 512) ws = [((mpmath.pi * mpmath.mpf(j)) / N) for j in range(N)] else: ws = worN h = [] for wk in ws: zm1 = mpmath.exp((1j * wk)) numer = _prod([(zm1 - t) for t in z]) denom = _prod([(zm1 - t) for t in p]) hk = ((k * numer) / denom) h.append(hk) return (ws, h)
[ "def", "zpkfreqd", "(", "z", ",", "p", ",", "k", ",", "worN", "=", "None", ")", ":", "if", "(", "(", "worN", "is", "None", ")", "or", "isinstance", "(", "worN", ",", "int", ")", ")", ":", "N", "=", "(", "worN", "or", "512", ")", "ws", "=", ...
frequency response of a filter in zpk format .
train
false
13,170
def plot_layout(layout, show=True): import matplotlib.pyplot as plt fig = plt.figure() ax = fig.add_subplot(111) fig.subplots_adjust(left=0, bottom=0, right=1, top=1, wspace=None, hspace=None) ax.set_xticks([]) ax.set_yticks([]) pos = [((p[0] + (p[2] / 2.0)), (p[1] + (p[3] / 2.0))) for p in layout.pos] (pos, outlines) = _check_outlines(pos, 'head') _draw_outlines(ax, outlines) for (ii, (this_pos, ch_id)) in enumerate(zip(pos, layout.names)): ax.annotate(ch_id, xy=this_pos[:2], horizontalalignment='center', verticalalignment='center', size='x-small') plt_show(show) return fig
[ "def", "plot_layout", "(", "layout", ",", "show", "=", "True", ")", ":", "import", "matplotlib", ".", "pyplot", "as", "plt", "fig", "=", "plt", ".", "figure", "(", ")", "ax", "=", "fig", ".", "add_subplot", "(", "111", ")", "fig", ".", "subplots_adju...
plot the sensor positions .
train
false
13,172
def initialize_events(article_generator): del events[:] localized_events.clear()
[ "def", "initialize_events", "(", "article_generator", ")", ":", "del", "events", "[", ":", "]", "localized_events", ".", "clear", "(", ")" ]
clears the events list before generating articles to properly support plugins with multiple generation passes like i18n_subsites .
train
false
13,173
def stroke_path(path, pen): stroker = QPainterPathStroker() stroker.setCapStyle(pen.capStyle()) stroker.setJoinStyle(pen.joinStyle()) stroker.setMiterLimit(pen.miterLimit()) stroker.setWidth(max(pen.widthF(), 1e-09)) return stroker.createStroke(path)
[ "def", "stroke_path", "(", "path", ",", "pen", ")", ":", "stroker", "=", "QPainterPathStroker", "(", ")", "stroker", ".", "setCapStyle", "(", "pen", ".", "capStyle", "(", ")", ")", "stroker", ".", "setJoinStyle", "(", "pen", ".", "joinStyle", "(", ")", ...
create a qpainterpath stroke from the path drawn with pen .
train
false
13,174
def run_ninja(): signal.signal(signal.SIGINT, signal.SIG_DFL) if ((not settings.IS_WINDOWS) and (not settings.IS_MAC_OS)): try: import ctypes libc = ctypes.CDLL('libc.so.6') procname = 'ninja-ide' libc.prctl(15, ('%s\x00' % procname), 0, 0, 0) except: print "The process couldn't be renamed'" (filenames, projects_path, extra_plugins, linenos, log_level, log_file) = cliparser.parse() resources.create_home_dir_structure() from ninja_ide.tools.logger import NinjaLogger NinjaLogger.argparse(log_level, log_file) settings.load_settings() app = QApplication(sys.argv) from ninja_ide import gui gui.start_ide(app, filenames, projects_path, extra_plugins, linenos) sys.exit(app.exec_())
[ "def", "run_ninja", "(", ")", ":", "signal", ".", "signal", "(", "signal", ".", "SIGINT", ",", "signal", ".", "SIG_DFL", ")", "if", "(", "(", "not", "settings", ".", "IS_WINDOWS", ")", "and", "(", "not", "settings", ".", "IS_MAC_OS", ")", ")", ":", ...
first obtain the execution args and create the resources folder .
train
false
13,176
def tenant_list(profile=None, **connection_args): kstone = auth(profile, **connection_args) ret = {} for tenant in getattr(kstone, _TENANTS, None).list(): ret[tenant.name] = dict(((value, getattr(tenant, value)) for value in dir(tenant) if ((not value.startswith('_')) and isinstance(getattr(tenant, value), (six.text_type, dict, bool, str))))) return ret
[ "def", "tenant_list", "(", "profile", "=", "None", ",", "**", "connection_args", ")", ":", "kstone", "=", "auth", "(", "profile", ",", "**", "connection_args", ")", "ret", "=", "{", "}", "for", "tenant", "in", "getattr", "(", "kstone", ",", "_TENANTS", ...
return a list of available tenants cli example: .
train
true
13,180
def SearchClients(query_str, token=None, limit=1000): client_schema = aff4.AFF4Object.classes['VFSGRRClient'].SchemaCls index = client_index.CreateClientIndex(token=token) client_list = index.LookupClients([query_str]) result_set = aff4.FACTORY.MultiOpen(client_list, token=token) results = [] for result in result_set: results.append((result, str(result.Get(client_schema.HOSTNAME)), str(result.Get(client_schema.OS_VERSION)), str(result.Get(client_schema.PING)))) if (len(results) >= limit): break return results
[ "def", "SearchClients", "(", "query_str", ",", "token", "=", "None", ",", "limit", "=", "1000", ")", ":", "client_schema", "=", "aff4", ".", "AFF4Object", ".", "classes", "[", "'VFSGRRClient'", "]", ".", "SchemaCls", "index", "=", "client_index", ".", "Cre...
search indexes for clients .
train
true
13,181
def setDefaultClock(clock): global defaultClock defaultClock = clock
[ "def", "setDefaultClock", "(", "clock", ")", ":", "global", "defaultClock", "defaultClock", "=", "clock" ]
set the default clock to be used to reference all logging times .
train
false
13,182
def _bit_size(number): if (number < 0): raise ValueError(('Only nonnegative numbers possible: %s' % number)) if (number == 0): return 0 bits = 0 while number: bits += 1 number >>= 1 return bits
[ "def", "_bit_size", "(", "number", ")", ":", "if", "(", "number", "<", "0", ")", ":", "raise", "ValueError", "(", "(", "'Only nonnegative numbers possible: %s'", "%", "number", ")", ")", "if", "(", "number", "==", "0", ")", ":", "return", "0", "bits", ...
returns the number of bits required to hold a specific long number .
train
false
13,183
def gmean(a, axis=0, dtype=None): if (not isinstance(a, np.ndarray)): log_a = np.log(np.array(a, dtype=dtype)) elif dtype: if isinstance(a, np.ma.MaskedArray): log_a = np.log(np.ma.asarray(a, dtype=dtype)) else: log_a = np.log(np.asarray(a, dtype=dtype)) else: log_a = np.log(a) return np.exp(log_a.mean(axis=axis))
[ "def", "gmean", "(", "a", ",", "axis", "=", "0", ",", "dtype", "=", "None", ")", ":", "if", "(", "not", "isinstance", "(", "a", ",", "np", ".", "ndarray", ")", ")", ":", "log_a", "=", "np", ".", "log", "(", "np", ".", "array", "(", "a", ","...
compute the geometric mean along the specified axis .
train
false
13,184
def load_inline_module(): from cffi import FFI defs = '\n double _numba_test_sin(double x);\n double _numba_test_cos(double x);\n double _numba_test_funcptr(double (*func)(double));\n ' ffi = FFI() ffi.cdef(defs) from numba import _helperlib return (ffi, ffi.dlopen(_helperlib.__file__))
[ "def", "load_inline_module", "(", ")", ":", "from", "cffi", "import", "FFI", "defs", "=", "'\\n double _numba_test_sin(double x);\\n double _numba_test_cos(double x);\\n double _numba_test_funcptr(double (*func)(double));\\n '", "ffi", "=", "FFI", "(", ")", "ffi", "."...
create an inline module .
train
false
13,185
def send_query_failure_email(recipient_id, query_id, query_params): email_subject = ('Query %s has failed' % query_id) email_body_template = 'Hi %s,<br>Your query with id %s has failed due to error during execution. Please check the query parameters and submit query again.<br><br>Thanks!<br><br>Best wishes,<br>The Oppia Team<br><br>%s' recipient_user_settings = user_services.get_user_settings(recipient_id) email_body = (email_body_template % (recipient_user_settings.username, query_id, EMAIL_FOOTER.value)) _send_email(recipient_id, feconf.SYSTEM_COMMITTER_ID, feconf.EMAIL_INTENT_QUERY_STATUS_NOTIFICATION, email_subject, email_body, feconf.NOREPLY_EMAIL_ADDRESS) admin_email_subject = 'Query job has failed.' admin_email_body_template = 'Query job with %s query id has failed in its execution.\nQuery parameters:\n\n' for key in sorted(query_params): admin_email_body_template += ('%s: %s\n' % (key, query_params[key])) admin_email_body = (admin_email_body_template % query_id) send_mail_to_admin(admin_email_subject, admin_email_body)
[ "def", "send_query_failure_email", "(", "recipient_id", ",", "query_id", ",", "query_params", ")", ":", "email_subject", "=", "(", "'Query %s has failed'", "%", "query_id", ")", "email_body_template", "=", "'Hi %s,<br>Your query with id %s has failed due to error during executi...
send an email to the initiator of a failed bulk email query .
train
false
13,186
def adjust(color, attribute, percent): (r, g, b, a, type) = parse_color(color) (r, g, b) = hsl_to_rgb(*_adjust(rgb_to_hsl(r, g, b), attribute, percent)) return unparse_color(r, g, b, a, type)
[ "def", "adjust", "(", "color", ",", "attribute", ",", "percent", ")", ":", "(", "r", ",", "g", ",", "b", ",", "a", ",", "type", ")", "=", "parse_color", "(", "color", ")", "(", "r", ",", "g", ",", "b", ")", "=", "hsl_to_rgb", "(", "*", "_adju...
adjust an attribute of color by a percent .
train
true
13,188
def _other_endian(typ): try: return getattr(typ, _OTHER_ENDIAN) except AttributeError: if (type(typ) == _array_type): return (_other_endian(typ._type_) * typ._length_) raise TypeError(('This type does not support other endian: %s' % typ))
[ "def", "_other_endian", "(", "typ", ")", ":", "try", ":", "return", "getattr", "(", "typ", ",", "_OTHER_ENDIAN", ")", "except", "AttributeError", ":", "if", "(", "type", "(", "typ", ")", "==", "_array_type", ")", ":", "return", "(", "_other_endian", "(",...
return the type with the other byte order .
train
true
13,189
def load_json(filename): try: if PY2: args = 'rb' else: args = 'r' with open(filename, args) as fid: data = json.load(fid) return (data, None) except Exception as err: return (None, str(err))
[ "def", "load_json", "(", "filename", ")", ":", "try", ":", "if", "PY2", ":", "args", "=", "'rb'", "else", ":", "args", "=", "'r'", "with", "open", "(", "filename", ",", "args", ")", "as", "fid", ":", "data", "=", "json", ".", "load", "(", "fid", ...
load data from a json file parameters filename : str filename to load data from .
train
true
13,190
def readlineAvailable(): return (readline._readline is not None)
[ "def", "readlineAvailable", "(", ")", ":", "return", "(", "readline", ".", "_readline", "is", "not", "None", ")" ]
check if the readline is available .
train
false
13,194
def format_isodate(value, timezone=None): if (not value): return '' dt = dateutil.parser.parse(str(value)) if timezone: dt = dt.astimezone(TZ(timezone)) value = format_dt(dt) return value
[ "def", "format_isodate", "(", "value", ",", "timezone", "=", "None", ")", ":", "if", "(", "not", "value", ")", ":", "return", "''", "dt", "=", "dateutil", ".", "parser", ".", "parse", "(", "str", "(", "value", ")", ")", "if", "timezone", ":", "dt",...
make a iso date time string human friendly .
train
false
13,196
def organizations_with(username, number=(-1), etag=None): return gh.organizations_with(username, number, etag)
[ "def", "organizations_with", "(", "username", ",", "number", "=", "(", "-", "1", ")", ",", "etag", "=", "None", ")", ":", "return", "gh", ".", "organizations_with", "(", "username", ",", "number", ",", "etag", ")" ]
list the organizations with username as a member .
train
false
13,197
def check_session_cookie(response): for cookie in request.cookies: if ((cookie == u'ckan') and (not getattr(g, u'user', None))): is_valid_cookie_data = False for (key, value) in session.items(): if ((not key.startswith(u'_')) and value): is_valid_cookie_data = True break if (not is_valid_cookie_data): if session.id: log.debug(u'No valid session data - deleting session') log.debug(u'Session: %r', session.items()) session.delete() else: log.debug(u'No session id - deleting session cookie') response.delete_cookie(cookie) elif ((cookie == u'auth_tkt') and (not session.id)): response.delete_cookie(cookie) return response
[ "def", "check_session_cookie", "(", "response", ")", ":", "for", "cookie", "in", "request", ".", "cookies", ":", "if", "(", "(", "cookie", "==", "u'ckan'", ")", "and", "(", "not", "getattr", "(", "g", ",", "u'user'", ",", "None", ")", ")", ")", ":", ...
the cookies for auth and session are separate .
train
false
13,198
def custom(command, user=None, conf_file=None, bin_env=None): ret = __salt__['cmd.run_all'](_ctl_cmd(command, None, conf_file, bin_env), runas=user, python_shell=False) return _get_return(ret)
[ "def", "custom", "(", "command", ",", "user", "=", "None", ",", "conf_file", "=", "None", ",", "bin_env", "=", "None", ")", ":", "ret", "=", "__salt__", "[", "'cmd.run_all'", "]", "(", "_ctl_cmd", "(", "command", ",", "None", ",", "conf_file", ",", "...
return a custom composite of status data and info for this minion .
train
true
13,200
def turn_on_internet(verbose=False): global INTERNET_OFF global _orig_opener if (not INTERNET_OFF): return INTERNET_OFF = False if verbose: print(u'Internet access enabled') urllib.request.install_opener(_orig_opener) socket.create_connection = socket_create_connection socket.socket.bind = socket_bind socket.socket.connect = socket_connect return socket
[ "def", "turn_on_internet", "(", "verbose", "=", "False", ")", ":", "global", "INTERNET_OFF", "global", "_orig_opener", "if", "(", "not", "INTERNET_OFF", ")", ":", "return", "INTERNET_OFF", "=", "False", "if", "verbose", ":", "print", "(", "u'Internet access enab...
restore internet access .
train
false
13,201
def quickshift(image, ratio=1.0, kernel_size=5, max_dist=10, return_tree=False, sigma=0, convert2lab=True, random_seed=42): image = img_as_float(np.atleast_3d(image)) if convert2lab: if (image.shape[2] != 3): ValueError('Only RGB images can be converted to Lab space.') image = rgb2lab(image) if (kernel_size < 1): raise ValueError('`kernel_size` should be >= 1.') image = ndi.gaussian_filter(image, [sigma, sigma, 0]) image = np.ascontiguousarray((image * ratio)) segment_mask = _quickshift_cython(image, kernel_size=kernel_size, max_dist=max_dist, return_tree=return_tree, random_seed=random_seed) return segment_mask
[ "def", "quickshift", "(", "image", ",", "ratio", "=", "1.0", ",", "kernel_size", "=", "5", ",", "max_dist", "=", "10", ",", "return_tree", "=", "False", ",", "sigma", "=", "0", ",", "convert2lab", "=", "True", ",", "random_seed", "=", "42", ")", ":",...
segments image using quickshift clustering in color- space .
train
false
13,202
def from_html(html_code, **kwargs): parser = TableHandler(**kwargs) parser.feed(html_code) return parser.tables
[ "def", "from_html", "(", "html_code", ",", "**", "kwargs", ")", ":", "parser", "=", "TableHandler", "(", "**", "kwargs", ")", "parser", ".", "feed", "(", "html_code", ")", "return", "parser", ".", "tables" ]
generates a list of prettytables from a string of html code .
train
true
13,204
def get_key(postfix=''): domain = getattr(settings, 'CURRENT_DOMAIN', 'default') key = ('treeio_%s_chat_%s' % (domain, postfix)) return key
[ "def", "get_key", "(", "postfix", "=", "''", ")", ":", "domain", "=", "getattr", "(", "settings", ",", "'CURRENT_DOMAIN'", ",", "'default'", ")", "key", "=", "(", "'treeio_%s_chat_%s'", "%", "(", "domain", ",", "postfix", ")", ")", "return", "key" ]
returns the ssh private key for vm access .
train
false
13,205
def assert_config_change(actual_result, expected_result): change_diffs = assert_config_change_dict(actual_result, expected_result) for file_change in change_diffs.values(): for line_change in file_change: if (len(line_change) != 0): return False return True
[ "def", "assert_config_change", "(", "actual_result", ",", "expected_result", ")", ":", "change_diffs", "=", "assert_config_change_dict", "(", "actual_result", ",", "expected_result", ")", "for", "file_change", "in", "change_diffs", ".", "values", "(", ")", ":", "for...
wrapper of the upper method returning boolean true if no config changes were detected .
train
false
13,206
@scope.define def callpipe1(fn_list, arg): for f in fn_list: arg = f(arg) return arg
[ "@", "scope", ".", "define", "def", "callpipe1", "(", "fn_list", ",", "arg", ")", ":", "for", "f", "in", "fn_list", ":", "arg", "=", "f", "(", "arg", ")", "return", "arg" ]
fn_list: a list lambdas that return either pyll expressions or python values arg: the argument to the first function in the list return: fn_list[-1]( .
train
false
13,208
def test_aligned_mem(): a = arange(804, dtype=np.uint8) z = np.frombuffer(a.data, offset=4, count=100, dtype=float) z.shape = (10, 10) eig(z, overwrite_a=True) eig(z.T, overwrite_a=True)
[ "def", "test_aligned_mem", "(", ")", ":", "a", "=", "arange", "(", "804", ",", "dtype", "=", "np", ".", "uint8", ")", "z", "=", "np", ".", "frombuffer", "(", "a", ".", "data", ",", "offset", "=", "4", ",", "count", "=", "100", ",", "dtype", "="...
check linalg works with non-aligned memory .
train
false
13,209
def group_ranges(ranges): return foldr(_combine, ranges, ())
[ "def", "group_ranges", "(", "ranges", ")", ":", "return", "foldr", "(", "_combine", ",", "ranges", ",", "(", ")", ")" ]
group any overlapping ranges into a single range .
train
false
13,210
def magma(): rc(u'image', cmap=u'magma') im = gci() if (im is not None): im.set_cmap(cm.magma)
[ "def", "magma", "(", ")", ":", "rc", "(", "u'image'", ",", "cmap", "=", "u'magma'", ")", "im", "=", "gci", "(", ")", "if", "(", "im", "is", "not", "None", ")", ":", "im", ".", "set_cmap", "(", "cm", ".", "magma", ")" ]
set the default colormap to magma and apply to current image if any .
train
false
13,211
def s_size(block_name, offset=0, length=4, endian='<', format='binary', inclusive=False, signed=False, math=None, fuzzable=False, name=None): if (block_name in blocks.CURRENT.block_stack): raise sex.SullyRuntimeError('CAN NOT ADD A SIZE FOR A BLOCK CURRENTLY IN THE STACK') size = blocks.size(block_name, blocks.CURRENT, offset, length, endian, format, inclusive, signed, math, fuzzable, name) blocks.CURRENT.push(size)
[ "def", "s_size", "(", "block_name", ",", "offset", "=", "0", ",", "length", "=", "4", ",", "endian", "=", "'<'", ",", "format", "=", "'binary'", ",", "inclusive", "=", "False", ",", "signed", "=", "False", ",", "math", "=", "None", ",", "fuzzable", ...
create a sizer block bound to the block with the specified name .
train
false
13,213
def deduplication(arg): if (type(arg) is types.ListType): return list(set(arg)) elif (type(arg) is types.TupleType): return tuple(set(arg)) return arg
[ "def", "deduplication", "(", "arg", ")", ":", "if", "(", "type", "(", "arg", ")", "is", "types", ".", "ListType", ")", ":", "return", "list", "(", "set", "(", "arg", ")", ")", "elif", "(", "type", "(", "arg", ")", "is", "types", ".", "TupleType",...
deduplication the arg .
train
false
13,215
def create_warning(path, error_message, skip_file=True): print_string = 'warning: ' if skip_file: print_string = (((print_string + 'Skipping file ') + path) + '. ') print_string = (print_string + error_message) warning_message = WarningResult(message=print_string, error=False, warning=True) return warning_message
[ "def", "create_warning", "(", "path", ",", "error_message", ",", "skip_file", "=", "True", ")", ":", "print_string", "=", "'warning: '", "if", "skip_file", ":", "print_string", "=", "(", "(", "(", "print_string", "+", "'Skipping file '", ")", "+", "path", ")...
this creates a printtask for whenever a warning is to be thrown .
train
false
13,218
def remove_hop_by_hop_headers(headers): headers[:] = [(key, value) for (key, value) in headers if (not is_hop_by_hop_header(key))]
[ "def", "remove_hop_by_hop_headers", "(", "headers", ")", ":", "headers", "[", ":", "]", "=", "[", "(", "key", ",", "value", ")", "for", "(", "key", ",", "value", ")", "in", "headers", "if", "(", "not", "is_hop_by_hop_header", "(", "key", ")", ")", "]...
remove all http/1 .
train
true
13,219
def make_blob_public(bucket_name, blob_name): storage_client = storage.Client() bucket = storage_client.get_bucket(bucket_name) blob = bucket.blob(blob_name) blob.make_public() print 'Blob {} is publicly accessible at {}'.format(blob.name, blob.public_url)
[ "def", "make_blob_public", "(", "bucket_name", ",", "blob_name", ")", ":", "storage_client", "=", "storage", ".", "Client", "(", ")", "bucket", "=", "storage_client", ".", "get_bucket", "(", "bucket_name", ")", "blob", "=", "bucket", ".", "blob", "(", "blob_...
makes a blob publicly accessible .
train
false
13,220
def _fd(f): return (f.fileno() if hasattr(f, 'fileno') else f)
[ "def", "_fd", "(", "f", ")", ":", "return", "(", "f", ".", "fileno", "(", ")", "if", "hasattr", "(", "f", ",", "'fileno'", ")", "else", "f", ")" ]
get a filedescriptor from something which could be a file or an fd .
train
false
13,221
def make_full_schema(data, schema): flattened_schema = flatten_schema(schema) key_combinations = get_all_key_combinations(data, flattened_schema) full_schema = {} for combination in key_combinations: sub_schema = schema for key in combination[::2]: sub_schema = sub_schema[key] for (key, value) in sub_schema.iteritems(): if isinstance(value, list): full_schema[(combination + (key,))] = value return full_schema
[ "def", "make_full_schema", "(", "data", ",", "schema", ")", ":", "flattened_schema", "=", "flatten_schema", "(", "schema", ")", "key_combinations", "=", "get_all_key_combinations", "(", "data", ",", "flattened_schema", ")", "full_schema", "=", "{", "}", "for", "...
make schema by getting all valid combinations and making sure that all keys are available .
train
false
13,222
def write_theme(): return theme_xml
[ "def", "write_theme", "(", ")", ":", "return", "theme_xml" ]
write the theme xml .
train
false
13,223
@lower_builtin(types.NumberClass, types.Any) def number_constructor(context, builder, sig, args): if isinstance(sig.return_type, types.Array): impl = context.get_function(np.array, sig) return impl(builder, args) else: [val] = args [valty] = sig.args return context.cast(builder, val, valty, sig.return_type)
[ "@", "lower_builtin", "(", "types", ".", "NumberClass", ",", "types", ".", "Any", ")", "def", "number_constructor", "(", "context", ",", "builder", ",", "sig", ",", "args", ")", ":", "if", "isinstance", "(", "sig", ".", "return_type", ",", "types", ".", ...
call a number class .
train
false
13,224
def test_bootstrap_random_seed(): data = rs.randn(50) seed = 42 boots1 = algo.bootstrap(data, random_seed=seed) boots2 = algo.bootstrap(data, random_seed=seed) assert_array_equal(boots1, boots2)
[ "def", "test_bootstrap_random_seed", "(", ")", ":", "data", "=", "rs", ".", "randn", "(", "50", ")", "seed", "=", "42", "boots1", "=", "algo", ".", "bootstrap", "(", "data", ",", "random_seed", "=", "seed", ")", "boots2", "=", "algo", ".", "bootstrap",...
test that we can get reproducible resamples by seeding the rng .
train
false
13,225
def translation(language): global _translations t = _translations.get(language, None) if (t is not None): return t from django.conf import settings globalpath = os.path.join(os.path.dirname(upath(sys.modules[settings.__module__].__file__)), u'locale') def _fetch(lang, fallback=None): global _translations res = _translations.get(lang, None) if (res is not None): return res loc = to_locale(lang) def _translation(path): try: t = gettext_module.translation(u'django', path, [loc], DjangoTranslation) t.set_language(lang) return t except IOError: return None res = _translation(globalpath) base_lang = (lambda x: x.split(u'-', 1)[0]) if (base_lang(lang) in [base_lang(trans) for trans in _translations]): res._info = res._info.copy() res._catalog = res._catalog.copy() def _merge(path): t = _translation(path) if (t is not None): if (res is None): return t else: res.merge(t) return res for appname in reversed(settings.INSTALLED_APPS): app = import_module(appname) apppath = os.path.join(os.path.dirname(upath(app.__file__)), u'locale') if os.path.isdir(apppath): res = _merge(apppath) for localepath in reversed(settings.LOCALE_PATHS): if os.path.isdir(localepath): res = _merge(localepath) if (res is None): if (fallback is not None): res = fallback else: return gettext_module.NullTranslations() _translations[lang] = res return res default_translation = _fetch(settings.LANGUAGE_CODE) current_translation = _fetch(language, fallback=default_translation) return current_translation
[ "def", "translation", "(", "language", ")", ":", "global", "_translations", "t", "=", "_translations", ".", "get", "(", "language", ",", "None", ")", "if", "(", "t", "is", "not", "None", ")", ":", "return", "t", "from", "django", ".", "conf", "import",...
create an array with a translation matrix .
train
false
13,227
def search_ohloh_project(project_name): results = get_ohloh_api_request(u'https://www.ohloh.net/p.xml', api_key, {u'query': project_name, u'sort': u'relevance'}) if (results.find(u'result/project/id') is None): raise Exception((u'Could not find project %s on Ohloh' % project_name)) project = results.find(u'result/project') return {u'id': project.findtext(u'id'), u'name': project.findtext(u'name'), u'description': project.findtext(u'description'), u'analysis': project.findtext(u'analysis_id'), u'tags': [tag.text for tag in project.iterfind(u'tags/tag')]}
[ "def", "search_ohloh_project", "(", "project_name", ")", ":", "results", "=", "get_ohloh_api_request", "(", "u'https://www.ohloh.net/p.xml'", ",", "api_key", ",", "{", "u'query'", ":", "project_name", ",", "u'sort'", ":", "u'relevance'", "}", ")", "if", "(", "resu...
searches for an ohloh project by name keyword arguments: project_name -- the project name to search for .
train
false
13,228
def convert_rgb_to_bokehrgba(img_data, downsample=1): if (img_data.dtype != np.uint8): raise NotImplementedError if (img_data.ndim != 3): raise NotImplementedError img_data = img_data[::(- downsample), ::downsample, :] (img_h, img_w, C) = img_data.shape bokeh_img = np.dstack([img_data, (255 * np.ones((img_h, img_w), np.uint8))]) final_image = bokeh_img.reshape(img_h, (img_w * (C + 1))).view(np.uint32) return final_image
[ "def", "convert_rgb_to_bokehrgba", "(", "img_data", ",", "downsample", "=", "1", ")", ":", "if", "(", "img_data", ".", "dtype", "!=", "np", ".", "uint8", ")", ":", "raise", "NotImplementedError", "if", "(", "img_data", ".", "ndim", "!=", "3", ")", ":", ...
convert rgb image to two-dimensional array of rgba values .
train
false
13,229
def dict_with_str_keys(dct, encoding='utf-8'): return {force_str(key, encoding): value for (key, value) in six.iteritems(dct)}
[ "def", "dict_with_str_keys", "(", "dct", ",", "encoding", "=", "'utf-8'", ")", ":", "return", "{", "force_str", "(", "key", ",", "encoding", ")", ":", "value", "for", "(", "key", ",", "value", ")", "in", "six", ".", "iteritems", "(", "dct", ")", "}" ...
applies force_str on the keys of a dict .
train
false
13,231
def test_xml_filters_round_trip(): plot = Bar() plot.add('A', [60, 75, 80, 78, 83, 90]) plot.add('B', [92, 87, 81, 73, 68, 55]) before = plot.render() plot.add_xml_filter((lambda T: T)) after = plot.render() assert (before == after)
[ "def", "test_xml_filters_round_trip", "(", ")", ":", "plot", "=", "Bar", "(", ")", "plot", ".", "add", "(", "'A'", ",", "[", "60", ",", "75", ",", "80", ",", "78", ",", "83", ",", "90", "]", ")", "plot", ".", "add", "(", "'B'", ",", "[", "92"...
ensure doing nothing does nothing .
train
false
13,233
def rescan_if_missed_blocks(store): (bad,) = store.selectrow('\n SELECT COUNT(1)\n FROM block\n LEFT JOIN chain_candidate USING (block_id)\n WHERE chain_id IS NULL\n ') if (bad > 0): store.sql('UPDATE datadir SET blkfile_number = 1, blkfile_offset = 0')
[ "def", "rescan_if_missed_blocks", "(", "store", ")", ":", "(", "bad", ",", ")", "=", "store", ".", "selectrow", "(", "'\\n SELECT COUNT(1)\\n FROM block\\n LEFT JOIN chain_candidate USING (block_id)\\n WHERE chain_id IS NULL\\n '", ")", "if", "...
due to a bug .
train
false
13,234
@verbose def lcmv_epochs(epochs, forward, noise_cov, data_cov, reg=0.01, label=None, pick_ori=None, return_generator=False, picks=None, rank=None, verbose=None): _check_reference(epochs) info = epochs.info tmin = epochs.times[0] picks = _setup_picks(picks, info, forward, noise_cov) data = epochs.get_data()[:, picks, :] stcs = _apply_lcmv(data=data, info=info, tmin=tmin, forward=forward, noise_cov=noise_cov, data_cov=data_cov, reg=reg, label=label, picks=picks, rank=rank, pick_ori=pick_ori) if (not return_generator): stcs = [s for s in stcs] return stcs
[ "@", "verbose", "def", "lcmv_epochs", "(", "epochs", ",", "forward", ",", "noise_cov", ",", "data_cov", ",", "reg", "=", "0.01", ",", "label", "=", "None", ",", "pick_ori", "=", "None", ",", "return_generator", "=", "False", ",", "picks", "=", "None", ...
linearly constrained minimum variance beamformer .
train
false
13,235
def showInfo(text, parent=False, help='', type='info', title='Anki'): if (parent is False): parent = (aqt.mw.app.activeWindow() or aqt.mw) if (type == 'warning'): icon = QMessageBox.Warning elif (type == 'critical'): icon = QMessageBox.Critical else: icon = QMessageBox.Information mb = QMessageBox(parent) mb.setText(text) mb.setIcon(icon) mb.setWindowModality(Qt.WindowModal) mb.setWindowTitle(title) b = mb.addButton(QMessageBox.Ok) b.setDefault(True) if help: b = mb.addButton(QMessageBox.Help) b.clicked.connect((lambda : openHelp(help))) b.setAutoDefault(False) return mb.exec_()
[ "def", "showInfo", "(", "text", ",", "parent", "=", "False", ",", "help", "=", "''", ",", "type", "=", "'info'", ",", "title", "=", "'Anki'", ")", ":", "if", "(", "parent", "is", "False", ")", ":", "parent", "=", "(", "aqt", ".", "mw", ".", "ap...
show a small info window with an ok button .
train
false
13,236
def _unit_file_changed(name): return ("'systemctl daemon-reload'" in _systemctl_status(name)['stdout'].lower())
[ "def", "_unit_file_changed", "(", "name", ")", ":", "return", "(", "\"'systemctl daemon-reload'\"", "in", "_systemctl_status", "(", "name", ")", "[", "'stdout'", "]", ".", "lower", "(", ")", ")" ]
returns true if systemctl reports that the unit file has changed .
train
false
13,237
def print_all_links(res, rules): r = [] for host in res: host_name = _apply_rules(host, rules) print ('%s::esxhostname=%s' % (host_name, host_name)) print ('%s::isesxhost=1' % host_name) for vm in res[host]: vm_name = _apply_rules(vm, rules) print ('%s::vmname=%s' % (vm_name, vm_name)) print ('%s::isesxvm=1' % vm_name) print ('%s::esxhost=%s' % (vm_name, host_name)) return r
[ "def", "print_all_links", "(", "res", ",", "rules", ")", ":", "r", "=", "[", "]", "for", "host", "in", "res", ":", "host_name", "=", "_apply_rules", "(", "host", ",", "rules", ")", "print", "(", "'%s::esxhostname=%s'", "%", "(", "host_name", ",", "host...
create all tuples of the links for the hosts .
train
false
13,238
def list_directories(): tenant = show_tenant() tenant_id = tenant.get('href', '').split('/')[(-1)] (status, result) = _query(action='tenants', command='{0}/directories'.format(tenant_id)) return result
[ "def", "list_directories", "(", ")", ":", "tenant", "=", "show_tenant", "(", ")", "tenant_id", "=", "tenant", ".", "get", "(", "'href'", ",", "''", ")", ".", "split", "(", "'/'", ")", "[", "(", "-", "1", ")", "]", "(", "status", ",", "result", ")...
show all directories .
train
false
13,239
def makeExtension(*args, **kwargs): return HeaderAnchorExtension(*args, **kwargs)
[ "def", "makeExtension", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "HeaderAnchorExtension", "(", "*", "args", ",", "**", "kwargs", ")" ]
make extension .
train
false
13,240
def find_libs(): from pyamf.util import get_module types = [] mapping = {} for mod in ETREE_MODULES: try: etree = get_module(mod) except ImportError: continue t = _get_etree_type(etree) types.append(t) mapping[t] = etree return (tuple(types), mapping)
[ "def", "find_libs", "(", ")", ":", "from", "pyamf", ".", "util", "import", "get_module", "types", "=", "[", "]", "mapping", "=", "{", "}", "for", "mod", "in", "ETREE_MODULES", ":", "try", ":", "etree", "=", "get_module", "(", "mod", ")", "except", "I...
run through l{etree_modules} and find c{elementtree} implementations so that any type can be encoded .
train
true
13,242
def notnull(values): return (~ isnull(values))
[ "def", "notnull", "(", "values", ")", ":", "return", "(", "~", "isnull", "(", "values", ")", ")" ]
replacement for numpy .
train
false
13,243
def autoload_server(model, app_path='/', session_id=None, url='default'): coords = _SessionCoordinates(dict(url=url, session_id=session_id, app_path=app_path)) elementid = make_id() model_id = '' if (model is not None): model_id = model._id if (model_id and (session_id is None)): raise ValueError("A specific model was passed to autoload_server() but no session_id; this doesn't work because the server will generate a fresh session which won't have the model in it.") src_path = (((coords.server_url + '/autoload.js') + '?bokeh-autoload-element=') + elementid) if (coords.session_id_allowing_none is not None): src_path = ((src_path + '&bokeh-session-id=') + session_id) tag = AUTOLOAD_TAG.render(src_path=src_path, elementid=elementid, modelid=model_id) return encode_utf8(tag)
[ "def", "autoload_server", "(", "model", ",", "app_path", "=", "'/'", ",", "session_id", "=", "None", ",", "url", "=", "'default'", ")", ":", "coords", "=", "_SessionCoordinates", "(", "dict", "(", "url", "=", "url", ",", "session_id", "=", "session_id", ...
return a script tag that embeds the given model from a bokeh server session .
train
false
13,244
def is_object_mutable(context, object): if context.is_admin: return True if (context.owner is None): return False return (object.namespace.owner == context.owner)
[ "def", "is_object_mutable", "(", "context", ",", "object", ")", ":", "if", "context", ".", "is_admin", ":", "return", "True", "if", "(", "context", ".", "owner", "is", "None", ")", ":", "return", "False", "return", "(", "object", ".", "namespace", ".", ...
return true if the object is mutable in this context .
train
false
13,245
def _function_pprint(obj, p, cycle): name = _safe_getattr(obj, '__qualname__', obj.__name__) mod = obj.__module__ if (mod and (mod not in ('__builtin__', 'builtins', 'exceptions'))): name = ((mod + '.') + name) p.text(('<function %s>' % name))
[ "def", "_function_pprint", "(", "obj", ",", "p", ",", "cycle", ")", ":", "name", "=", "_safe_getattr", "(", "obj", ",", "'__qualname__'", ",", "obj", ".", "__name__", ")", "mod", "=", "obj", ".", "__module__", "if", "(", "mod", "and", "(", "mod", "no...
base pprint for all functions and builtin functions .
train
false