id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
50,617
def data_url(mimetype, data): b64 = base64.b64encode(data).decode('ascii') url = QUrl('data:{};base64,{}'.format(mimetype, b64)) qtutils.ensure_valid(url) return url
[ "def", "data_url", "(", "mimetype", ",", "data", ")", ":", "b64", "=", "base64", ".", "b64encode", "(", "data", ")", ".", "decode", "(", "'ascii'", ")", "url", "=", "QUrl", "(", "'data:{};base64,{}'", ".", "format", "(", "mimetype", ",", "b64", ")", ...
get a data: qurl for the given data .
train
false
50,618
def test_prewitt_h_horizontal(): (i, j) = np.mgrid[(-5):6, (-5):6] image = (i >= 0).astype(float) result = filters.prewitt_h(image) i[(np.abs(j) == 5)] = 10000 assert np.all((result[(i == 0)] == 1)) assert_allclose(result[(np.abs(i) > 1)], 0, atol=1e-10)
[ "def", "test_prewitt_h_horizontal", "(", ")", ":", "(", "i", ",", "j", ")", "=", "np", ".", "mgrid", "[", "(", "-", "5", ")", ":", "6", ",", "(", "-", "5", ")", ":", "6", "]", "image", "=", "(", "i", ">=", "0", ")", ".", "astype", "(", "f...
horizontal prewitt on an edge should be a horizontal line .
train
false
50,621
def _ssl_dn_extract_info(dn_string): search_string = re.search('/emailAddress=(.*)@([^/]+)', dn_string) if search_string: user = search_string.group(1) email = ('%s@%s' % (user, search_string.group(2))) else: raise ValueError search_string = re.search('/CN=([^/]+)/', dn_string) if search_string: fullname = search_string.group(1) else: raise ValueError return (user, email, fullname)
[ "def", "_ssl_dn_extract_info", "(", "dn_string", ")", ":", "search_string", "=", "re", ".", "search", "(", "'/emailAddress=(.*)@([^/]+)'", ",", "dn_string", ")", "if", "search_string", ":", "user", "=", "search_string", ".", "group", "(", "1", ")", "email", "=...
extract username .
train
false
50,622
def libvlc_log_set(cb, data, p_instance): f = (_Cfunctions.get('libvlc_log_set', None) or _Cfunction('libvlc_log_set', ((1,), (1,), (1,)), None, None, Instance, LogCb, ctypes.c_void_p)) return f(cb, data, p_instance)
[ "def", "libvlc_log_set", "(", "cb", ",", "data", ",", "p_instance", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_log_set'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_log_set'", ",", "(", "(", "1", ",", ")", ",", "(", "1...
sets the logging callback for a libvlc instance .
train
true
50,623
def get_kinds(start=None, end=None): q = Kind.query() if ((start is not None) and (start != '')): q = q.filter((Kind.key >= Kind.key_for_kind(start))) if (end is not None): if (end == ''): return [] q = q.filter((Kind.key < Kind.key_for_kind(end))) return [x.kind_name for x in q]
[ "def", "get_kinds", "(", "start", "=", "None", ",", "end", "=", "None", ")", ":", "q", "=", "Kind", ".", "query", "(", ")", "if", "(", "(", "start", "is", "not", "None", ")", "and", "(", "start", "!=", "''", ")", ")", ":", "q", "=", "q", "....
return all kinds in the specified range .
train
true
50,624
def _qnwtrap1(n, a, b): if (n < 1): raise ValueError('n must be at least one') nodes = np.linspace(a, b, n) dx = (nodes[1] - nodes[0]) weights = (dx * np.ones(n)) weights[0] *= 0.5 weights[(-1)] *= 0.5 return (nodes, weights)
[ "def", "_qnwtrap1", "(", "n", ",", "a", ",", "b", ")", ":", "if", "(", "n", "<", "1", ")", ":", "raise", "ValueError", "(", "'n must be at least one'", ")", "nodes", "=", "np", ".", "linspace", "(", "a", ",", "b", ",", "n", ")", "dx", "=", "(",...
compute univariate trapezoid rule quadrature nodes and weights parameters n : int the number of nodes a : int the lower endpoint b : int the upper endpoint returns nodes : np .
train
true
50,625
def block_until_ssh_open(ipstring, wait_time=10, timeout=120): reached = False t_elapsed = 0 while ((not reached) and (t_elapsed < timeout)): try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.connect((ipstring, 22)) reached = True except socket.error as err: time.sleep(wait_time) t_elapsed += wait_time sock.close()
[ "def", "block_until_ssh_open", "(", "ipstring", ",", "wait_time", "=", "10", ",", "timeout", "=", "120", ")", ":", "reached", "=", "False", "t_elapsed", "=", "0", "while", "(", "(", "not", "reached", ")", "and", "(", "t_elapsed", "<", "timeout", ")", "...
blocks until server at ipstring has an open port 22 .
train
false
50,627
def GetEnvironFallback(var_list, default): for var in var_list: if (var in os.environ): return os.environ[var] return default
[ "def", "GetEnvironFallback", "(", "var_list", ",", "default", ")", ":", "for", "var", "in", "var_list", ":", "if", "(", "var", "in", "os", ".", "environ", ")", ":", "return", "os", ".", "environ", "[", "var", "]", "return", "default" ]
look up a key in the environment .
train
false
50,628
def align_iterators(func, *iterables): class myiter: def __init__(self, it): self.it = it self.key = self.value = None self.iternext() def iternext(self): try: self.value = next(self.it) self.key = func(self.value) except StopIteration: self.value = self.key = None def __call__(self, key): retval = None if (key == self.key): retval = self.value self.iternext() elif (self.key and (key > self.key)): raise ValueError(u'Iterator has been left behind') return retval iters = [myiter(it) for it in iterables] minvals = minkey = True while True: minvals = [_f for _f in [it.key for it in iters] if _f] if minvals: minkey = min(minvals) (yield (minkey, [it(minkey) for it in iters])) else: break
[ "def", "align_iterators", "(", "func", ",", "*", "iterables", ")", ":", "class", "myiter", ":", "def", "__init__", "(", "self", ",", "it", ")", ":", "self", ".", "it", "=", "it", "self", ".", "key", "=", "self", ".", "value", "=", "None", "self", ...
this generator takes a bunch of iterables that are ordered by func it sends out ordered tuples:: (func .
train
false
50,630
def parse_filename(filename): filename = parse_string(filename) result = resource_find(filename) if (result is None): Logger.error(('Resource: unable to find <%s>' % filename)) return (result or filename)
[ "def", "parse_filename", "(", "filename", ")", ":", "filename", "=", "parse_string", "(", "filename", ")", "result", "=", "resource_find", "(", "filename", ")", "if", "(", "result", "is", "None", ")", ":", "Logger", ".", "error", "(", "(", "'Resource: unab...
parse a filename and search for it using resource_find() .
train
false
50,633
def list_tiles(query, zoom_adjust): resolution = (sum(query.resolution) / 2) diff = ((_log(resolution) / _log(2)) - zoom_adjust) zoom = round((meter_zoom + diff)) scale = (2 ** zoom) mincol = int((scale * ((query.bbox.minx / diameter) + 0.5))) maxcol = int((scale * ((query.bbox.maxx / diameter) + 0.5))) minrow = int((scale * (0.5 - (query.bbox.maxy / diameter)))) maxrow = int((scale * (0.5 - (query.bbox.miny / diameter)))) (cols, rows) = (range(mincol, (maxcol + 1)), range(minrow, (maxrow + 1))) return [dict(z=zoom, x=col, y=row) for (col, row) in product(cols, rows)]
[ "def", "list_tiles", "(", "query", ",", "zoom_adjust", ")", ":", "resolution", "=", "(", "sum", "(", "query", ".", "resolution", ")", "/", "2", ")", "diff", "=", "(", "(", "_log", "(", "resolution", ")", "/", "_log", "(", "2", ")", ")", "-", "zoo...
get a list of tile coordinates .
train
false
50,634
def getgoal(path, opts=None): cmd = 'mfsgetgoal' ret = {} if opts: cmd += (' -' + opts) else: opts = '' cmd += (' ' + path) out = __salt__['cmd.run_all'](cmd, python_shell=False) output = out['stdout'].splitlines() if ('r' not in opts): goal = output[0].split(': ') ret = {'goal': goal[1]} else: for line in output: if (not line): continue if (path in line): continue comps = line.split() keytext = (comps[0] + ' with goal') if (keytext not in ret): ret[keytext] = {} ret[keytext][comps[3]] = comps[5] return ret
[ "def", "getgoal", "(", "path", ",", "opts", "=", "None", ")", ":", "cmd", "=", "'mfsgetgoal'", "ret", "=", "{", "}", "if", "opts", ":", "cmd", "+=", "(", "' -'", "+", "opts", ")", "else", ":", "opts", "=", "''", "cmd", "+=", "(", "' '", "+", ...
return goal(s) for a file or directory cli example: .
train
true
50,635
def initial_seed(): return default_generator.initial_seed()
[ "def", "initial_seed", "(", ")", ":", "return", "default_generator", ".", "initial_seed", "(", ")" ]
returns the initial seed for generating random numbers as a python long .
train
false
50,637
def _convert_warp_input(image, preserve_range): if preserve_range: image = image.astype(np.double) else: image = img_as_float(image) return image
[ "def", "_convert_warp_input", "(", "image", ",", "preserve_range", ")", ":", "if", "preserve_range", ":", "image", "=", "image", ".", "astype", "(", "np", ".", "double", ")", "else", ":", "image", "=", "img_as_float", "(", "image", ")", "return", "image" ]
convert input image to double image with the appropriate range .
train
false
50,638
@ignore_warnings def test_sensitivity_specificity_support_errors(): (y_true, y_pred, _) = make_prediction(binary=True) assert_raises(ValueError, sensitivity_specificity_support, y_true, y_pred, pos_label=2, average='binary') assert_raises(ValueError, sensitivity_specificity_support, [0, 1, 2], [1, 2, 0], average='mega')
[ "@", "ignore_warnings", "def", "test_sensitivity_specificity_support_errors", "(", ")", ":", "(", "y_true", ",", "y_pred", ",", "_", ")", "=", "make_prediction", "(", "binary", "=", "True", ")", "assert_raises", "(", "ValueError", ",", "sensitivity_specificity_suppo...
test either if an error is raised depending on parameters .
train
false
50,639
def _HashPassword(password, version, salt): if (version == _PASSWORD_VERSION_MD5): m = hashlib.md5() m.update(password) m.update(secrets.GetSecret('cookie_secret')) hashed = m.digest() elif (version == _PASSWORD_VERSION_PBKDF2): hashed = PBKDF2(password, base64.b64decode(salt), count=10000) return base64.b32encode(hashed)
[ "def", "_HashPassword", "(", "password", ",", "version", ",", "salt", ")", ":", "if", "(", "version", "==", "_PASSWORD_VERSION_MD5", ")", ":", "m", "=", "hashlib", ".", "md5", "(", ")", "m", ".", "update", "(", "password", ")", "m", ".", "update", "(...
hashes the provided password according to the specified versions policy .
train
false
50,640
def process_show_version(net_device): show_ver = net_device.show_version (net_device.vendor, net_device.model) = obtain_vendor_model(show_ver) net_device.os_version = obtain_os_version(show_ver) net_device.uptime = obtain_uptime(show_ver) net_device.hostname = obtain_hostname(show_ver) net_device.serial_number = obtain_serial_number(show_ver) net_device.device_type = obtain_device_type(net_device.model)
[ "def", "process_show_version", "(", "net_device", ")", ":", "show_ver", "=", "net_device", ".", "show_version", "(", "net_device", ".", "vendor", ",", "net_device", ".", "model", ")", "=", "obtain_vendor_model", "(", "show_ver", ")", "net_device", ".", "os_versi...
process the show_version output for net_device assign the following attributes to the net_device object hostname device_type # router .
train
false
50,641
def get_new_coords(init_loc, distance, bearing): origin = geopy.Point(init_loc[0], init_loc[1]) destination = geopy.distance.distance(kilometers=distance).destination(origin, bearing) return (destination.latitude, destination.longitude)
[ "def", "get_new_coords", "(", "init_loc", ",", "distance", ",", "bearing", ")", ":", "origin", "=", "geopy", ".", "Point", "(", "init_loc", "[", "0", "]", ",", "init_loc", "[", "1", "]", ")", "destination", "=", "geopy", ".", "distance", ".", "distance...
given an initial lat/lng .
train
false
50,642
def customClass(classes, argStr): (cname, args, kwargs) = splitArgs(argStr) cls = classes.get(cname, None) if (not cls): raise Exception(('error: %s is unknown - please specify one of %s' % (cname, classes.keys()))) if ((not args) and (not kwargs)): return cls return specialClass(cls, append=args, defaults=kwargs)
[ "def", "customClass", "(", "classes", ",", "argStr", ")", ":", "(", "cname", ",", "args", ",", "kwargs", ")", "=", "splitArgs", "(", "argStr", ")", "cls", "=", "classes", ".", "get", "(", "cname", ",", "None", ")", "if", "(", "not", "cls", ")", "...
return customized class based on argstr the args and key/val pairs in argstr will be automatically applied when the generated class is later used .
train
false
50,643
def _immutable_fields_error(result_pipeline_definition): for e in result_pipeline_definition['error']: for e2 in e['errors']: if ('can not be changed' in e2): return True return False
[ "def", "_immutable_fields_error", "(", "result_pipeline_definition", ")", ":", "for", "e", "in", "result_pipeline_definition", "[", "'error'", "]", ":", "for", "e2", "in", "e", "[", "'errors'", "]", ":", "if", "(", "'can not be changed'", "in", "e2", ")", ":",...
return true if update pipeline failed due to immutable fields some fields cannot be changed after a pipeline has been activated .
train
false
50,644
def read_file_contents(path): with open(path) as file_handle: return file_handle.read()
[ "def", "read_file_contents", "(", "path", ")", ":", "with", "open", "(", "path", ")", "as", "file_handle", ":", "return", "file_handle", ".", "read", "(", ")" ]
reads the contents of the given file .
train
false
50,646
def preprocess_form(): BoundField.label_tag = add_required_label_tag(BoundField.label_tag)
[ "def", "preprocess_form", "(", ")", ":", "BoundField", ".", "label_tag", "=", "add_required_label_tag", "(", "BoundField", ".", "label_tag", ")" ]
add asterisk to field labels .
train
false
50,647
def _parse_relative_url(relative_url): if (not relative_url): raise _RelativeUrlError('Relative URL is empty') (scheme, netloc, path, query, fragment) = urlparse.urlsplit(relative_url) if (scheme or netloc): raise _RelativeUrlError('Relative URL may not have a scheme or location') if fragment: raise _RelativeUrlError('Relative URL may not specify a fragment') if ((not path) or (path[0] != '/')): raise _RelativeUrlError('Relative URL path must start with "/"') return (path, query)
[ "def", "_parse_relative_url", "(", "relative_url", ")", ":", "if", "(", "not", "relative_url", ")", ":", "raise", "_RelativeUrlError", "(", "'Relative URL is empty'", ")", "(", "scheme", ",", "netloc", ",", "path", ",", "query", ",", "fragment", ")", "=", "u...
parses a relative url and splits it into its path and query string .
train
false
50,648
def _ToSearchError(error): if (error.application_error in _ERROR_MAP): return _ERROR_MAP[error.application_error](error.error_detail) return error
[ "def", "_ToSearchError", "(", "error", ")", ":", "if", "(", "error", ".", "application_error", "in", "_ERROR_MAP", ")", ":", "return", "_ERROR_MAP", "[", "error", ".", "application_error", "]", "(", "error", ".", "error_detail", ")", "return", "error" ]
translate an application error to a search error .
train
false
50,651
@click.command(u'trigger-scheduler-event') @click.argument(u'event') @pass_context def trigger_scheduler_event(context, event): import frappe.utils.scheduler for site in context.sites: try: frappe.init(site=site) frappe.connect() frappe.utils.scheduler.trigger(site, event, now=True) finally: frappe.destroy()
[ "@", "click", ".", "command", "(", "u'trigger-scheduler-event'", ")", "@", "click", ".", "argument", "(", "u'event'", ")", "@", "pass_context", "def", "trigger_scheduler_event", "(", "context", ",", "event", ")", ":", "import", "frappe", ".", "utils", ".", "...
trigger a scheduler event .
train
false
50,652
def write_dig_points(fid, dig, block=False, coord_frame=None): if (dig is not None): data_size = (5 * 4) if block: start_block(fid, FIFF.FIFFB_ISOTRAK) if (coord_frame is not None): write_int(fid, FIFF.FIFF_MNE_COORD_FRAME, coord_frame) for d in dig: fid.write(np.array(FIFF.FIFF_DIG_POINT, '>i4').tostring()) fid.write(np.array(FIFF.FIFFT_DIG_POINT_STRUCT, '>i4').tostring()) fid.write(np.array(data_size, dtype='>i4').tostring()) fid.write(np.array(FIFF.FIFFV_NEXT_SEQ, '>i4').tostring()) fid.write(np.array(d['kind'], '>i4').tostring()) fid.write(np.array(d['ident'], '>i4').tostring()) fid.write(np.array(d['r'][:3], '>f4').tostring()) if block: end_block(fid, FIFF.FIFFB_ISOTRAK)
[ "def", "write_dig_points", "(", "fid", ",", "dig", ",", "block", "=", "False", ",", "coord_frame", "=", "None", ")", ":", "if", "(", "dig", "is", "not", "None", ")", ":", "data_size", "=", "(", "5", "*", "4", ")", "if", "block", ":", "start_block",...
write a set of digitizer data points into a fif file .
train
false
50,654
def delta_resolution(dt, delta): delta = timedelta_seconds(delta) resolutions = ((3, (lambda x: (x / 86400))), (4, (lambda x: (x / 3600))), (5, (lambda x: (x / 60)))) args = (dt.year, dt.month, dt.day, dt.hour, dt.minute, dt.second) for (res, predicate) in resolutions: if (predicate(delta) >= 1.0): return datetime(*args[:res]) return dt
[ "def", "delta_resolution", "(", "dt", ",", "delta", ")", ":", "delta", "=", "timedelta_seconds", "(", "delta", ")", "resolutions", "=", "(", "(", "3", ",", "(", "lambda", "x", ":", "(", "x", "/", "86400", ")", ")", ")", ",", "(", "4", ",", "(", ...
round a datetime to the resolution of a timedelta .
train
false
50,655
def _logs_exist(fs, path): try: return fs.exists(path) except IOError: return None
[ "def", "_logs_exist", "(", "fs", ",", "path", ")", ":", "try", ":", "return", "fs", ".", "exists", "(", "path", ")", "except", "IOError", ":", "return", "None" ]
do fs .
train
false
50,658
def graph_laplacian(csgraph, normed=False, return_diag=False): if ((csgraph.ndim != 2) or (csgraph.shape[0] != csgraph.shape[1])): raise ValueError('csgraph must be a square matrix or array') if (normed and (np.issubdtype(csgraph.dtype, np.int) or np.issubdtype(csgraph.dtype, np.uint))): csgraph = check_array(csgraph, dtype=np.float64, accept_sparse=True) if sparse.isspmatrix(csgraph): return _laplacian_sparse(csgraph, normed=normed, return_diag=return_diag) else: return _laplacian_dense(csgraph, normed=normed, return_diag=return_diag)
[ "def", "graph_laplacian", "(", "csgraph", ",", "normed", "=", "False", ",", "return_diag", "=", "False", ")", ":", "if", "(", "(", "csgraph", ".", "ndim", "!=", "2", ")", "or", "(", "csgraph", ".", "shape", "[", "0", "]", "!=", "csgraph", ".", "sha...
return the laplacian matrix of a directed graph .
train
false
50,659
def expand_otu_ids(otu_map, otus_to_expand, ignore_missing=False): result = [] for o in otus_to_expand: otu_id = o.split()[0] try: result += otu_map[otu_id] except KeyError: if ignore_missing: continue else: raise KeyError(('OTU id not in OTU map: %s' % o.split()[0])) return result
[ "def", "expand_otu_ids", "(", "otu_map", ",", "otus_to_expand", ",", "ignore_missing", "=", "False", ")", ":", "result", "=", "[", "]", "for", "o", "in", "otus_to_expand", ":", "otu_id", "=", "o", ".", "split", "(", ")", "[", "0", "]", "try", ":", "r...
from otu map and otu ids .
train
false
50,660
def esc(s, esc_chars): if (not s): return u'' for c in esc_chars: esc_str = (u'\\' + c) s = s.replace(c, esc_str) return s
[ "def", "esc", "(", "s", ",", "esc_chars", ")", ":", "if", "(", "not", "s", ")", ":", "return", "u''", "for", "c", "in", "esc_chars", ":", "esc_str", "=", "(", "u'\\\\'", "+", "c", ")", "s", "=", "s", ".", "replace", "(", "c", ",", "esc_str", ...
escape special characters .
train
false
50,661
def igcd(*args): if (len(args) < 2): raise TypeError(('igcd() takes at least 2 arguments (%s given)' % len(args))) if (1 in args): a = 1 k = 0 else: a = abs(as_int(args[0])) k = 1 if (a != 1): while (k < len(args)): b = args[k] k += 1 try: a = _gcdcache[(a, b)] except KeyError: b = as_int(b) if (not b): continue if (b == 1): a = 1 break if (b < 0): b = (- b) t = (a, b) while b: (a, b) = (b, (a % b)) _gcdcache[t] = _gcdcache[(t[1], t[0])] = a while (k < len(args)): ok = as_int(args[k]) k += 1 return a
[ "def", "igcd", "(", "*", "args", ")", ":", "if", "(", "len", "(", "args", ")", "<", "2", ")", ":", "raise", "TypeError", "(", "(", "'igcd() takes at least 2 arguments (%s given)'", "%", "len", "(", "args", ")", ")", ")", "if", "(", "1", "in", "args",...
computes nonnegative integer greatest common divisor .
train
false
50,662
def penalty_string(distance, limit=None): penalties = [] for key in distance.keys(): key = key.replace('album_', '') key = key.replace('track_', '') key = key.replace('_', ' ') penalties.append(key) if penalties: if (limit and (len(penalties) > limit)): penalties = (penalties[:limit] + ['...']) return ui.colorize('text_warning', (u'(%s)' % ', '.join(penalties)))
[ "def", "penalty_string", "(", "distance", ",", "limit", "=", "None", ")", ":", "penalties", "=", "[", "]", "for", "key", "in", "distance", ".", "keys", "(", ")", ":", "key", "=", "key", ".", "replace", "(", "'album_'", ",", "''", ")", "key", "=", ...
returns a colorized string that indicates all the penalties applied to a distance object .
train
false
50,663
def conv_input_length(output_length, filter_size, stride, pad=0): if (output_length is None): return None if (pad == 'valid'): pad = 0 elif (pad == 'full'): pad = (filter_size - 1) elif (pad == 'same'): pad = (filter_size // 2) if (not isinstance(pad, int)): raise ValueError('Invalid pad: {0}'.format(pad)) return ((((output_length - 1) * stride) - (2 * pad)) + filter_size)
[ "def", "conv_input_length", "(", "output_length", ",", "filter_size", ",", "stride", ",", "pad", "=", "0", ")", ":", "if", "(", "output_length", "is", "None", ")", ":", "return", "None", "if", "(", "pad", "==", "'valid'", ")", ":", "pad", "=", "0", "...
helper function to compute the input size of a convolution operation this function computes the length along a single axis .
train
false
50,664
def libvlc_log_get_context(ctx): f = (_Cfunctions.get('libvlc_log_get_context', None) or _Cfunction('libvlc_log_get_context', ((1,), (2,), (2,), (2,)), None, None, Log_ptr, ListPOINTER(ctypes.c_char_p), ListPOINTER(ctypes.c_char_p), ctypes.POINTER(ctypes.c_uint))) return f(ctx)
[ "def", "libvlc_log_get_context", "(", "ctx", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_log_get_context'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_log_get_context'", ",", "(", "(", "1", ",", ")", ",", "(", "2", ",", ")"...
gets debugging informations about a log message: the name of the vlc module emitting the message and the message location within the source code .
train
true
50,665
def match_anywhere(needles, haystack, ignore_case=False): regex_needle = (('.*' + '.*'.join(imap(re.escape, needles))) + '.*') regex_flags = ((re.IGNORECASE | re.UNICODE) if ignore_case else re.UNICODE) found = (lambda haystack: re.search(regex_needle, haystack.path, flags=regex_flags)) return ifilter(found, haystack)
[ "def", "match_anywhere", "(", "needles", ",", "haystack", ",", "ignore_case", "=", "False", ")", ":", "regex_needle", "=", "(", "(", "'.*'", "+", "'.*'", ".", "join", "(", "imap", "(", "re", ".", "escape", ",", "needles", ")", ")", ")", "+", "'.*'", ...
matches needles anywhere in the path as long as theyre in the same order .
train
false
50,666
def test_server_error_exception(exception_app): (request, response) = sanic_endpoint_test(exception_app, uri='/error') assert (response.status == 500)
[ "def", "test_server_error_exception", "(", "exception_app", ")", ":", "(", "request", ",", "response", ")", "=", "sanic_endpoint_test", "(", "exception_app", ",", "uri", "=", "'/error'", ")", "assert", "(", "response", ".", "status", "==", "500", ")" ]
test the built-in servererror exception works .
train
false
50,668
def chloginclass(name, loginclass, root=None): if (__grains__['kernel'] != 'OpenBSD'): return False if (loginclass == get_loginclass(name)): return True cmd = ['usermod', '-L', '{0}'.format(loginclass), '{0}'.format(name)] if (root is not None): cmd.extend(('-R', root)) __salt__['cmd.run'](cmd, python_shell=False) return (get_loginclass(name) == loginclass)
[ "def", "chloginclass", "(", "name", ",", "loginclass", ",", "root", "=", "None", ")", ":", "if", "(", "__grains__", "[", "'kernel'", "]", "!=", "'OpenBSD'", ")", ":", "return", "False", "if", "(", "loginclass", "==", "get_loginclass", "(", "name", ")", ...
change the default login class of the user .
train
true
50,669
def find_staff_lock_source(xblock): if xblock.fields['visible_to_staff_only'].is_set_on(xblock): return xblock if (xblock.category == 'chapter'): return None parent_location = modulestore().get_parent_location(xblock.location, revision=ModuleStoreEnum.RevisionOption.draft_preferred) if (not parent_location): return None parent = modulestore().get_item(parent_location) return find_staff_lock_source(parent)
[ "def", "find_staff_lock_source", "(", "xblock", ")", ":", "if", "xblock", ".", "fields", "[", "'visible_to_staff_only'", "]", ".", "is_set_on", "(", "xblock", ")", ":", "return", "xblock", "if", "(", "xblock", ".", "category", "==", "'chapter'", ")", ":", ...
returns the xblock responsible for setting this xblocks staff lock .
train
false
50,670
def try_run(commands): null = open(os.devnull, 'w') for command in commands: retcode = subprocess.call(command, stdout=null, shell=True, stderr=subprocess.STDOUT) if (retcode != 0): return False return True
[ "def", "try_run", "(", "commands", ")", ":", "null", "=", "open", "(", "os", ".", "devnull", ",", "'w'", ")", "for", "command", "in", "commands", ":", "retcode", "=", "subprocess", ".", "call", "(", "command", ",", "stdout", "=", "null", ",", "shell"...
run a series of commands and only return true if all ran fine .
train
false
50,672
def dict_of_key_value_pairs(arg): lexer = shlex.shlex(str(arg)) lexer.wordchars += '/.+-():' tokens = list(lexer) tokens_len = len(tokens) D = {} i = 0 while (i < tokens_len): k_eq_v = tokens[i:(i + 3)] if ((len(k_eq_v) != 3) or (k_eq_v[1] != '=')): raise ValueError(("Unexpected end of key/value pairs in value '%s'" % arg)) D[k_eq_v[0]] = k_eq_v[2].strip('\'"') i += 4 return D
[ "def", "dict_of_key_value_pairs", "(", "arg", ")", ":", "lexer", "=", "shlex", ".", "shlex", "(", "str", "(", "arg", ")", ")", "lexer", ".", "wordchars", "+=", "'/.+-():'", "tokens", "=", "list", "(", "lexer", ")", "tokens_len", "=", "len", "(", "token...
parse key=val .
train
false
50,674
def get_api_client(api_config, student): client_name = api_config.OAUTH2_CLIENT_NAME try: client = Client.objects.get(name=client_name) except Client.DoesNotExist: raise ImproperlyConfigured('OAuth2 Client with name [{}] does not exist.'.format(client_name)) scopes = ['email', 'profile'] expires_in = settings.OAUTH_ID_TOKEN_EXPIRATION jwt = JwtBuilder(student, secret=client.client_secret).build_token(scopes, expires_in, aud=client.client_id) return EdxRestApiClient(api_config.internal_api_url, jwt=jwt)
[ "def", "get_api_client", "(", "api_config", ",", "student", ")", ":", "client_name", "=", "api_config", ".", "OAUTH2_CLIENT_NAME", "try", ":", "client", "=", "Client", ".", "objects", ".", "get", "(", "name", "=", "client_name", ")", "except", "Client", ".",...
create and configure an api client for authenticated http requests .
train
false
50,675
def in_string(state, text, i, formats, user_data): q = (u'"' if (state.parse == IN_DQS) else u"'") pos = text.find(q, i) if (pos == (-1)): if (text[(-1)] == u'\\'): return [((len(text) - i), formats[u'string'])] state.parse = (NORMAL if (state.blocks < 1) else IN_CONTENT) return [((len(text) - i), formats[u'unterminated-string'])] state.parse = (NORMAL if (state.blocks < 1) else IN_CONTENT) return [(((pos - i) + len(q)), formats[u'string'])]
[ "def", "in_string", "(", "state", ",", "text", ",", "i", ",", "formats", ",", "user_data", ")", ":", "q", "=", "(", "u'\"'", "if", "(", "state", ".", "parse", "==", "IN_DQS", ")", "else", "u\"'\"", ")", "pos", "=", "text", ".", "find", "(", "q", ...
inside a string .
train
false
50,676
def py_conv(img, kern, mode, subsample): if isinstance(mode, int): mode = (mode, mode) if isinstance(mode, tuple): (pad_h, pad_w) = map(int, mode) img = py_conv_pad_img(img, pad_h, pad_w) mode = 'valid' if imported_scipy_convolve2d: return py_conv_scipy(img, kern, mode, subsample) elif (mode == 'valid'): return py_conv_valid_numpy(img, kern)[:, :, ::subsample[0], ::subsample[1]] elif (mode == 'full'): return py_conv_full_numpy(img, kern)[:, :, ::subsample[0], ::subsample[1]] else: raise Exception("Can't execute this kernel.")
[ "def", "py_conv", "(", "img", ",", "kern", ",", "mode", ",", "subsample", ")", ":", "if", "isinstance", "(", "mode", ",", "int", ")", ":", "mode", "=", "(", "mode", ",", "mode", ")", "if", "isinstance", "(", "mode", ",", "tuple", ")", ":", "(", ...
use a scipy or numpy implementation depending is scipy is available .
train
false
50,677
def requote_uri(uri): safe_with_percent = "!#$%&'()*+,/:;=?@[]~" safe_without_percent = "!#$&'()*+,/:;=?@[]~" try: return quote(unquote_unreserved(uri), safe=safe_with_percent) except InvalidURL: return quote(uri, safe=safe_without_percent)
[ "def", "requote_uri", "(", "uri", ")", ":", "safe_with_percent", "=", "\"!#$%&'()*+,/:;=?@[]~\"", "safe_without_percent", "=", "\"!#$&'()*+,/:;=?@[]~\"", "try", ":", "return", "quote", "(", "unquote_unreserved", "(", "uri", ")", ",", "safe", "=", "safe_with_percent", ...
re-quote the given uri .
train
true
50,678
def _compile_firstset(info, fs): reverse = bool((info.flags & REVERSE)) fs = _check_firstset(info, reverse, fs) if (not fs): return [] return fs.compile(reverse)
[ "def", "_compile_firstset", "(", "info", ",", "fs", ")", ":", "reverse", "=", "bool", "(", "(", "info", ".", "flags", "&", "REVERSE", ")", ")", "fs", "=", "_check_firstset", "(", "info", ",", "reverse", ",", "fs", ")", "if", "(", "not", "fs", ")", ...
compiles the firstset for the pattern .
train
false
50,680
def _get_index_name_by_column(table, column_name): protected_name = metadata.protect_name(column_name) possible_index_values = [protected_name, ('values(%s)' % protected_name)] for index_metadata in table.indexes.values(): options = dict(index_metadata.index_options) if (options.get('target') in possible_index_values): return index_metadata.name
[ "def", "_get_index_name_by_column", "(", "table", ",", "column_name", ")", ":", "protected_name", "=", "metadata", ".", "protect_name", "(", "column_name", ")", "possible_index_values", "=", "[", "protected_name", ",", "(", "'values(%s)'", "%", "protected_name", ")"...
find the index name for a given table and column .
train
true
50,684
def disjoint_union_all(graphs): graphs = iter(graphs) U = next(graphs) for H in graphs: U = nx.disjoint_union(U, H) return U
[ "def", "disjoint_union_all", "(", "graphs", ")", ":", "graphs", "=", "iter", "(", "graphs", ")", "U", "=", "next", "(", "graphs", ")", "for", "H", "in", "graphs", ":", "U", "=", "nx", ".", "disjoint_union", "(", "U", ",", "H", ")", "return", "U" ]
return the disjoint union of all graphs .
train
false
50,685
def _get_my_ip(): csock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) csock.connect(('8.8.8.8', 80)) (addr, _port) = csock.getsockname() csock.close() return addr
[ "def", "_get_my_ip", "(", ")", ":", "csock", "=", "socket", ".", "socket", "(", "socket", ".", "AF_INET", ",", "socket", ".", "SOCK_DGRAM", ")", "csock", ".", "connect", "(", "(", "'8.8.8.8'", ",", "80", ")", ")", "(", "addr", ",", "_port", ")", "=...
returns the actual ip of the local machine .
train
false
50,686
def int_from_bool_as_string(subject): return ((bool_from_string(subject) and 1) or 0)
[ "def", "int_from_bool_as_string", "(", "subject", ")", ":", "return", "(", "(", "bool_from_string", "(", "subject", ")", "and", "1", ")", "or", "0", ")" ]
interpret a string as a boolean and return either 1 or 0 .
train
false
50,687
def bridge_exists(br): cmd = 'ovs-vsctl br-exists {0}'.format(br) result = __salt__['cmd.run_all'](cmd) retcode = result['retcode'] return _retcode_to_bool(retcode)
[ "def", "bridge_exists", "(", "br", ")", ":", "cmd", "=", "'ovs-vsctl br-exists {0}'", ".", "format", "(", "br", ")", "result", "=", "__salt__", "[", "'cmd.run_all'", "]", "(", "cmd", ")", "retcode", "=", "result", "[", "'retcode'", "]", "return", "_retcode...
tests whether bridge exists as a real or fake bridge .
train
true
50,688
def clean_output(s): return re.sub('\n( *\n)+', '\n', s).replace(' ', ' ')
[ "def", "clean_output", "(", "s", ")", ":", "return", "re", ".", "sub", "(", "'\\n( *\\n)+'", ",", "'\\n'", ",", "s", ")", ".", "replace", "(", "' '", ",", "' '", ")" ]
remove double newlines with whitespace in between and reduce the level of indentation .
train
false
50,689
@login_required def diff_submit(request): data = {'svn_diff_form': forms.DiffForm(request.user.username), 'svn_diff_error_message': ''} if (request.method == 'POST'): temp_svn_directory = tempfile.mkdtemp() form = forms.DiffForm(request.user.username, temp_svn_directory, request.POST) if form.is_valid(): form.commit_diff() view_helpers.set_mission_completed(request.user.get_profile(), 'svn_diff') return HttpResponseRedirect(reverse('svn_diff')) data['svn_diff_form'] = form shutil.rmtree(temp_svn_directory) request.method = 'GET' return Diff.as_view()(request, extra_context_data=data)
[ "@", "login_required", "def", "diff_submit", "(", "request", ")", ":", "data", "=", "{", "'svn_diff_form'", ":", "forms", ".", "DiffForm", "(", "request", ".", "user", ".", "username", ")", ",", "'svn_diff_error_message'", ":", "''", "}", "if", "(", "reque...
handle submitting the results of an svn diff to the mission 1 .
train
false
50,691
@frappe.whitelist() def create_new_folder(file_name, folder): file = frappe.new_doc(u'File') file.file_name = file_name file.is_folder = 1 file.folder = folder file.insert()
[ "@", "frappe", ".", "whitelist", "(", ")", "def", "create_new_folder", "(", "file_name", ",", "folder", ")", ":", "file", "=", "frappe", ".", "new_doc", "(", "u'File'", ")", "file", ".", "file_name", "=", "file_name", "file", ".", "is_folder", "=", "1", ...
create new folder under current parent folder .
train
false
50,693
@pytest.fixture def project_store_resources0(project0, subdir0): from pootle_project.models import ProjectResource from pootle_store.models import Store store = subdir0.child_stores.live().first() resources = Store.objects.live().filter(name=store.name, parent__name=subdir0.name, translation_project__project=project0) return ProjectResource(resources, ('/projects/%s/%s/%s' % (project0.code, subdir0.name, store.name)))
[ "@", "pytest", ".", "fixture", "def", "project_store_resources0", "(", "project0", ",", "subdir0", ")", ":", "from", "pootle_project", ".", "models", "import", "ProjectResource", "from", "pootle_store", ".", "models", "import", "Store", "store", "=", "subdir0", ...
returns a projectresource object for a store .
train
false
50,695
def get_language_objects(site_id=None): return list(get_languages(site_id))
[ "def", "get_language_objects", "(", "site_id", "=", "None", ")", ":", "return", "list", "(", "get_languages", "(", "site_id", ")", ")" ]
returns list of all language objects filled up by default values .
train
false
50,696
def angle_between_vectors(v0, v1, directed=True, axis=0): v0 = numpy.array(v0, dtype=numpy.float64, copy=False) v1 = numpy.array(v1, dtype=numpy.float64, copy=False) dot = numpy.sum((v0 * v1), axis=axis) dot /= (vector_norm(v0, axis=axis) * vector_norm(v1, axis=axis)) return numpy.arccos((dot if directed else numpy.fabs(dot)))
[ "def", "angle_between_vectors", "(", "v0", ",", "v1", ",", "directed", "=", "True", ",", "axis", "=", "0", ")", ":", "v0", "=", "numpy", ".", "array", "(", "v0", ",", "dtype", "=", "numpy", ".", "float64", ",", "copy", "=", "False", ")", "v1", "=...
return angle between vectors .
train
true
50,697
def _requireSSL(decoratee): if (SSL is None): @wraps(decoratee) def raiseNotImplemented(*a, **kw): '\n pyOpenSSL is not available.\n\n @param a: The positional arguments for C{decoratee}.\n\n @param kw: The keyword arguments for C{decoratee}.\n\n @raise NotImplementedError: Always.\n ' raise NotImplementedError('SSL support unavailable') return raiseNotImplemented return decoratee
[ "def", "_requireSSL", "(", "decoratee", ")", ":", "if", "(", "SSL", "is", "None", ")", ":", "@", "wraps", "(", "decoratee", ")", "def", "raiseNotImplemented", "(", "*", "a", ",", "**", "kw", ")", ":", "raise", "NotImplementedError", "(", "'SSL support un...
the decorated method requires pyopenssl to be present .
train
false
50,698
def document_collection_object(section, collection_model, include_signature=True): if include_signature: section.style.start_sphinx_py_attr(collection_model.name) section.include_doc_string(('A collection of %s resources' % collection_model.resource.type))
[ "def", "document_collection_object", "(", "section", ",", "collection_model", ",", "include_signature", "=", "True", ")", ":", "if", "include_signature", ":", "section", ".", "style", ".", "start_sphinx_py_attr", "(", "collection_model", ".", "name", ")", "section",...
documents a collection resource object .
train
false
50,699
def normalize_known_hosts_key(key): k = key.strip() k = key.split() d = dict() if (k[0][0] == '@'): d['options'] = k[0] d['host'] = k[1] d['type'] = k[2] d['key'] = k[3] else: d['host'] = k[0] d['type'] = k[1] d['key'] = k[2] return d
[ "def", "normalize_known_hosts_key", "(", "key", ")", ":", "k", "=", "key", ".", "strip", "(", ")", "k", "=", "key", ".", "split", "(", ")", "d", "=", "dict", "(", ")", "if", "(", "k", "[", "0", "]", "[", "0", "]", "==", "'@'", ")", ":", "d"...
transform a key .
train
false
50,700
def copy_strip(): src = join(abspath(dirname(dirname(dirname(__file__)))), 'Doc', 'build', 'html', 'library', 'idle.html') dst = join(abspath(dirname(__file__)), 'help.html') with open(src, 'rb') as inn: with open(dst, 'wb') as out: for line in inn: out.write((line.rstrip() + '\n')) print 'idle.html copied to help.html'
[ "def", "copy_strip", "(", ")", ":", "src", "=", "join", "(", "abspath", "(", "dirname", "(", "dirname", "(", "dirname", "(", "__file__", ")", ")", ")", ")", ",", "'Doc'", ",", "'build'", ",", "'html'", ",", "'library'", ",", "'idle.html'", ")", "dst"...
copy idle .
train
false
50,701
def get_stored_content_length(headers): length = headers.get('x-goog-stored-content-length') if (length is None): length = headers.get('content-length') return length
[ "def", "get_stored_content_length", "(", "headers", ")", ":", "length", "=", "headers", ".", "get", "(", "'x-goog-stored-content-length'", ")", "if", "(", "length", "is", "None", ")", ":", "length", "=", "headers", ".", "get", "(", "'content-length'", ")", "...
return the content length of the object as stored in gcs .
train
true
50,702
def flac2wav(path, keep=True): flac_path = _getFlacPath() flac_files = [] if path.endswith('.flac'): flac_files = [path] elif ((type(path) == str) and os.path.isdir(path)): flac_files = glob.glob(os.path.join(path, '*.flac')) if (len(flac_files) == 0): logging.warn(('failed to find .flac file(s) from %s' % path)) return None wav_files = [] for flacfile in flac_files: wavname = (flacfile.strip('.flac') + '.wav') flac_cmd = [flac_path, '-d', '--totally-silent', '-f', '-o', wavname, flacfile] (_junk, se) = core.shellCall(flac_cmd, stderr=True) if se: logging.error(se) if (not keep): os.unlink(flacfile) wav_files.append(wavname) if (len(wav_files) == 1): return wav_files[0] else: return wav_files
[ "def", "flac2wav", "(", "path", ",", "keep", "=", "True", ")", ":", "flac_path", "=", "_getFlacPath", "(", ")", "flac_files", "=", "[", "]", "if", "path", ".", "endswith", "(", "'.flac'", ")", ":", "flac_files", "=", "[", "path", "]", "elif", "(", ...
uncompress: convert .
train
false
50,703
def is_leaf(cluster): return (len(cluster) == 1)
[ "def", "is_leaf", "(", "cluster", ")", ":", "return", "(", "len", "(", "cluster", ")", "==", "1", ")" ]
a cluster is a leaf if it has length 1 .
train
false
50,704
@profiler.trace def remove_group_role(request, role, group, domain=None, project=None): manager = keystoneclient(request, admin=True).roles return manager.revoke(role=role, group=group, project=project, domain=domain)
[ "@", "profiler", ".", "trace", "def", "remove_group_role", "(", "request", ",", "role", ",", "group", ",", "domain", "=", "None", ",", "project", "=", "None", ")", ":", "manager", "=", "keystoneclient", "(", "request", ",", "admin", "=", "True", ")", "...
removes a given single role for a group from a domain or project .
train
true
50,707
def _get_tab_registry(win_id, tab_id): if (tab_id is None): raise ValueError('Got tab_id None (win_id {})'.format(win_id)) if ((tab_id == 'current') and (win_id is None)): app = get('app') window = app.activeWindow() if ((window is None) or (not hasattr(window, 'win_id'))): raise RegistryUnavailableError('tab') win_id = window.win_id elif (win_id is not None): window = window_registry[win_id] else: raise TypeError('window is None with scope tab!') if (tab_id == 'current'): tabbed_browser = get('tabbed-browser', scope='window', window=win_id) tab = tabbed_browser.currentWidget() if (tab is None): raise RegistryUnavailableError('window') tab_id = tab.tab_id tab_registry = get('tab-registry', scope='window', window=win_id) try: return tab_registry[tab_id].registry except AttributeError: raise RegistryUnavailableError('tab')
[ "def", "_get_tab_registry", "(", "win_id", ",", "tab_id", ")", ":", "if", "(", "tab_id", "is", "None", ")", ":", "raise", "ValueError", "(", "'Got tab_id None (win_id {})'", ".", "format", "(", "win_id", ")", ")", "if", "(", "(", "tab_id", "==", "'current'...
get the registry of a tab .
train
false
50,708
def numToHex(num, numBytes): hexString = '' if (not isinstance(num, int)): return ((-1), 'Bad number') try: hexNumber = hex(num)[2:] if ((len(hexNumber) % 2) != 0): hexNumber = ('0' + hexNumber) for i in range(0, (len(hexNumber) - 1), 2): hexString += chr(int((hexNumber[i] + hexNumber[(i + 1)]), 16)) hexString = (('\x00' * (numBytes - len(hexString))) + hexString) except: return ((-1), 'Error in hexadecimal conversion') return (0, hexString)
[ "def", "numToHex", "(", "num", ",", "numBytes", ")", ":", "hexString", "=", "''", "if", "(", "not", "isinstance", "(", "num", ",", "int", ")", ")", ":", "return", "(", "(", "-", "1", ")", ",", "'Bad number'", ")", "try", ":", "hexNumber", "=", "h...
given a number returns its hexadecimal format with the specified length .
train
false
50,709
def detrend_none(x): return x
[ "def", "detrend_none", "(", "x", ")", ":", "return", "x" ]
return x: no detrending .
train
false
50,710
def requires_auth(f): @wraps(f) def decorated(*args, **kwargs): message = 'Authentication credentials not found' success = False try: auth_jwt() success = True except JWTError as e: if ((e.headers is not None) and ('WWW-Authenticate' not in e.headers)): raise NotAuthorizedError(message=((e.error + ': ') + e.description)) if (not success): results = auth_basic() if (not results[0]): if results[1]: raise NotAuthorizedError(message=results[1]) else: success = True if (not success): if login.current_user.is_authenticated: g.user = UserModel.query.get(login.current_user.id) success = True else: g.user.update_lat() if success: return f(*args, **kwargs) else: raise NotAuthorizedError(message=message) return decorated
[ "def", "requires_auth", "(", "f", ")", ":", "@", "wraps", "(", "f", ")", "def", "decorated", "(", "*", "args", ",", "**", "kwargs", ")", ":", "message", "=", "'Authentication credentials not found'", "success", "=", "False", "try", ":", "auth_jwt", "(", ...
decorator for :class:tvdbclient methods that require authentication .
train
false
50,711
def getSimState(): try: status = '?' mContext = autoclass('android.content.Context') pythonActivity = autoclass('org.renpy.android.PythonService') TelephonyManager = autoclass('android.telephony.TelephonyManager') telephonyManager = cast('android.telephony.TelephonyManager', pythonActivity.mService.getSystemService(mContext.TELEPHONY_SERVICE)) simState = telephonyManager.getSimState() if (simState == TelephonyManager.SIM_STATE_UNKNOWN): status = 'unknown' elif (simState == TelephonyManager.SIM_STATE_ABSENT): status = 'absent' elif (simState == TelephonyManager.SIM_STATE_PIN_REQUIRED): status = 'pin_required' elif (simState == TelephonyManager.SIM_STATE_PUK_REQUIRED): status = 'puk_required' elif (simState == TelephonyManager.SIM_STATE_NETWORK_LOCKED): status = 'network_locked' elif (simState == TelephonyManager.SIM_STATE_READY): status = 'ready' return status except Exception as e: return None
[ "def", "getSimState", "(", ")", ":", "try", ":", "status", "=", "'?'", "mContext", "=", "autoclass", "(", "'android.content.Context'", ")", "pythonActivity", "=", "autoclass", "(", "'org.renpy.android.PythonService'", ")", "TelephonyManager", "=", "autoclass", "(", ...
returns a string indicating the state of the default sim card returns none if an error .
train
false
50,713
def test_csv_table_read(): lines = ['# a, b', '1, 2', '3, 4'] t = ascii.read(lines) assert (t.colnames == ['a', 'b'])
[ "def", "test_csv_table_read", "(", ")", ":", "lines", "=", "[", "'# a, b'", ",", "'1, 2'", ",", "'3, 4'", "]", "t", "=", "ascii", ".", "read", "(", "lines", ")", "assert", "(", "t", ".", "colnames", "==", "[", "'a'", ",", "'b'", "]", ")" ]
check for a regression introduced by #1935 .
train
false
50,715
def has_results(errors=None, failures=None, skipped=None, expected_failures=None, unexpected_successes=None, tests_run=None): if (errors is None): errors = Equals([]) if (failures is None): failures = Equals([]) if (skipped is None): skipped = Equals([]) if (expected_failures is None): expected_failures = Equals([]) if (unexpected_successes is None): unexpected_successes = Equals([]) if (tests_run is None): tests_run = Equals(0) return MatchesStructure(errors=errors, failures=failures, skipped=skipped, expectedFailures=expected_failures, unexpectedSuccesses=unexpected_successes, testsRun=tests_run)
[ "def", "has_results", "(", "errors", "=", "None", ",", "failures", "=", "None", ",", "skipped", "=", "None", ",", "expected_failures", "=", "None", ",", "unexpected_successes", "=", "None", ",", "tests_run", "=", "None", ")", ":", "if", "(", "errors", "i...
return a matcher on test results .
train
false
50,716
def _check_reference(inst): if _needs_eeg_average_ref_proj(inst.info): raise ValueError('EEG average reference is mandatory for inverse modeling, use set_eeg_reference method.') if inst.info['custom_ref_applied']: raise ValueError('Custom EEG reference is not allowed for inverse modeling.')
[ "def", "_check_reference", "(", "inst", ")", ":", "if", "_needs_eeg_average_ref_proj", "(", "inst", ".", "info", ")", ":", "raise", "ValueError", "(", "'EEG average reference is mandatory for inverse modeling, use set_eeg_reference method.'", ")", "if", "inst", ".", "info...
check for eeg ref .
train
false
50,717
def _fix_related_fields(srv, data, service_ids): if (srv[0] not in RELATED_FIELDS): return data for field in RELATED_FIELDS[srv[0]]: if (field[0] not in data): data[field[1]] = None continue old_value = data[field[0]] if (type(old_value) == list): ls = [] for i in old_value: old_id = i['id'] new_id = service_ids[field[2]][old_id] ls += [new_id] del data[field[0]] data[field[1]] = ls else: if (type(old_value) == dict): old_id = old_value['id'] else: old_id = old_value del data[field[0]] if (old_id is None): data[field[1]] = None else: data[field[1]] = service_ids[field[2]][old_id] return data
[ "def", "_fix_related_fields", "(", "srv", ",", "data", ",", "service_ids", ")", ":", "if", "(", "srv", "[", "0", "]", "not", "in", "RELATED_FIELDS", ")", ":", "return", "data", "for", "field", "in", "RELATED_FIELDS", "[", "srv", "[", "0", "]", "]", "...
fixes the ids services which are related to others .
train
false
50,718
def select_random_ports(n): ports = [] sockets = [] for i in range(n): sock = socket.socket() sock.bind(('', 0)) ports.append(sock.getsockname()[1]) sockets.append(sock) for sock in sockets: sock.close() return ports
[ "def", "select_random_ports", "(", "n", ")", ":", "ports", "=", "[", "]", "sockets", "=", "[", "]", "for", "i", "in", "range", "(", "n", ")", ":", "sock", "=", "socket", ".", "socket", "(", ")", "sock", ".", "bind", "(", "(", "''", ",", "0", ...
select and return n random ports that are available .
train
false
50,719
@verbose def compare_fiff(fname_1, fname_2, fname_out=None, show=True, indent=' ', read_limit=np.inf, max_str=30, verbose=None): file_1 = show_fiff(fname_1, output=list, indent=indent, read_limit=read_limit, max_str=max_str) file_2 = show_fiff(fname_2, output=list, indent=indent, read_limit=read_limit, max_str=max_str) diff = difflib.HtmlDiff().make_file(file_1, file_2, fname_1, fname_2) if (fname_out is not None): f = open(fname_out, 'wb') else: f = tempfile.NamedTemporaryFile('wb', delete=False, suffix='.html') fname_out = f.name with f as fid: fid.write(diff.encode('utf-8')) if (show is True): webbrowser.open_new_tab(fname_out) return fname_out
[ "@", "verbose", "def", "compare_fiff", "(", "fname_1", ",", "fname_2", ",", "fname_out", "=", "None", ",", "show", "=", "True", ",", "indent", "=", "' '", ",", "read_limit", "=", "np", ".", "inf", ",", "max_str", "=", "30", ",", "verbose", "=", "N...
compare the contents of two fiff files using diff and show_fiff .
train
false
50,721
def _get_size_linux(): def ioctl_GWINSZ(fd): '\n Attempt to discover the dimensions of a terminal window, using IOCTL.\n ' try: import fcntl, termios cr = struct.unpack('hh', fcntl.ioctl(fd, termios.TIOCGWINSZ, '1234')) except: return None return cr cr = (ioctl_GWINSZ(0) or ioctl_GWINSZ(1) or ioctl_GWINSZ(2)) if (not cr): try: fd = os.open(os.ctermid(), os.O_RDONLY) cr = ioctl_GWINSZ(fd) os.close(fd) except: pass if (not cr): try: cr = (env['LINES'], env['COLUMNS']) except: return None return (int(cr[1]), int(cr[0]))
[ "def", "_get_size_linux", "(", ")", ":", "def", "ioctl_GWINSZ", "(", "fd", ")", ":", "try", ":", "import", "fcntl", ",", "termios", "cr", "=", "struct", ".", "unpack", "(", "'hh'", ",", "fcntl", ".", "ioctl", "(", "fd", ",", "termios", ".", "TIOCGWIN...
attempt to discover the dimensions of a terminal window .
train
false
50,722
def primary_collator(): global _primary_collator if (_primary_collator is None): _primary_collator = collator().clone() _primary_collator.strength = _icu.UCOL_PRIMARY return _primary_collator
[ "def", "primary_collator", "(", ")", ":", "global", "_primary_collator", "if", "(", "_primary_collator", "is", "None", ")", ":", "_primary_collator", "=", "collator", "(", ")", ".", "clone", "(", ")", "_primary_collator", ".", "strength", "=", "_icu", ".", "...
ignores case differences and accented characters .
train
false
50,723
def getFirstValue(gcodeText, word): for line in archive.getTextLines(gcodeText): splitLine = gcodec.getSplitLineBeforeBracketSemicolon(line) if (gcodec.getFirstWord(splitLine) == word): return splitLine[1] return ''
[ "def", "getFirstValue", "(", "gcodeText", ",", "word", ")", ":", "for", "line", "in", "archive", ".", "getTextLines", "(", "gcodeText", ")", ":", "splitLine", "=", "gcodec", ".", "getSplitLineBeforeBracketSemicolon", "(", "line", ")", "if", "(", "gcodec", "....
get the value from the first line which starts with the given word .
train
false
50,724
def mergedicts_copy(d1, d2): ret = d1.copy() _setmerged(ret, d2) for k in d2: if ((k in d1) and isinstance(d1[k], dict) and isinstance(d2[k], dict)): ret[k] = mergedicts_copy(d1[k], d2[k]) else: ret[k] = d2[k] return ret
[ "def", "mergedicts_copy", "(", "d1", ",", "d2", ")", ":", "ret", "=", "d1", ".", "copy", "(", ")", "_setmerged", "(", "ret", ",", "d2", ")", "for", "k", "in", "d2", ":", "if", "(", "(", "k", "in", "d1", ")", "and", "isinstance", "(", "d1", "[...
recursively merge two dictionaries .
train
false
50,725
def config_option_list(context, data_dict): return {'success': False}
[ "def", "config_option_list", "(", "context", ",", "data_dict", ")", ":", "return", "{", "'success'", ":", "False", "}" ]
list runtime-editable configuration options .
train
false
50,727
def count_seqs_in_filepaths(fasta_filepaths, seq_counter=count_seqs): total = 0 counts = [] inaccessible_filepaths = [] for fasta_filepath in fasta_filepaths: if (fasta_filepath.endswith('.fastq') or fasta_filepath.endswith('.fq')): parser = partial(parse_fastq, enforce_qual_range=False) elif (fasta_filepath.endswith('.tre') or fasta_filepath.endswith('.ph') or fasta_filepath.endswith('.ntree')): def parser(f): t = DndParser(f, constructor=PhyloNode) return zip(t.iterTips(), repeat('')) else: parser = parse_fasta try: current_count = seq_counter(fasta_filepath, parser=parser) counts.append((current_count, fasta_filepath)) total += current_count[0] except IOError: inaccessible_filepaths.append(fasta_filepath) return (counts, total, inaccessible_filepaths)
[ "def", "count_seqs_in_filepaths", "(", "fasta_filepaths", ",", "seq_counter", "=", "count_seqs", ")", ":", "total", "=", "0", "counts", "=", "[", "]", "inaccessible_filepaths", "=", "[", "]", "for", "fasta_filepath", "in", "fasta_filepaths", ":", "if", "(", "f...
wrapper to apply seq_counter to fasta_filepaths fasta_filepaths: list of one or more fasta filepaths seq_counter: a function which takes a single filepath and returns the count of the number of sequences -- this is parameterized to facilitate unit testing .
train
false
50,728
def provider_handlers(providers): return _load_provider_feature('default_handlers', providers)
[ "def", "provider_handlers", "(", "providers", ")", ":", "return", "_load_provider_feature", "(", "'default_handlers'", ",", "providers", ")" ]
load tornado url handlers from an ordered list of dotted-notation modules which contain a default_handlers function default_handlers should accept a list of handlers and returns an augmented list of handlers: this allows the addition of .
train
false
50,730
def test_fit_sample_auto(): ratio = 'auto' cc = ClusterCentroids(ratio=ratio, random_state=RND_SEED) (X_resampled, y_resampled) = cc.fit_sample(X, Y) X_gt = np.array([[0.92923648, 0.76103773], [0.47104475, 0.44386323], [0.13347175, 0.12167502], [0.06738818, (-0.529627)], [0.17901516, 0.69860992], [0.094035, (-2.55298982)]]) y_gt = np.array([0, 0, 0, 1, 1, 1]) assert_allclose(X_resampled, X_gt, rtol=R_TOL) assert_array_equal(y_resampled, y_gt)
[ "def", "test_fit_sample_auto", "(", ")", ":", "ratio", "=", "'auto'", "cc", "=", "ClusterCentroids", "(", "ratio", "=", "ratio", ",", "random_state", "=", "RND_SEED", ")", "(", "X_resampled", ",", "y_resampled", ")", "=", "cc", ".", "fit_sample", "(", "X",...
test fit and sample routines with auto ratio .
train
false
50,732
def libvlc_media_player_set_time(p_mi, i_time): f = (_Cfunctions.get('libvlc_media_player_set_time', None) or _Cfunction('libvlc_media_player_set_time', ((1,), (1,)), None, None, MediaPlayer, ctypes.c_longlong)) return f(p_mi, i_time)
[ "def", "libvlc_media_player_set_time", "(", "p_mi", ",", "i_time", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_media_player_set_time'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_media_player_set_time'", ",", "(", "(", "1", ",", ...
set the movie time .
train
true
50,733
def requires_application(backend=None, has=(), capable=()): (good, msg) = has_application(backend, has, capable) dec_backend = np.testing.dec.skipif((not good), ('Skipping test: %s' % msg)) try: import pytest except Exception: return dec_backend dec_app = pytest.mark.vispy_app_test return composed(dec_app, dec_backend)
[ "def", "requires_application", "(", "backend", "=", "None", ",", "has", "=", "(", ")", ",", "capable", "=", "(", ")", ")", ":", "(", "good", ",", "msg", ")", "=", "has_application", "(", "backend", ",", "has", ",", "capable", ")", "dec_backend", "=",...
return a decorator for tests that require an application .
train
false
50,734
def get_v1_mutation_key_and_entity(v1_mutation): if v1_mutation.HasField('delete'): return (v1_mutation.delete, None) else: v1_entity = getattr(v1_mutation, v1_mutation.WhichOneof('operation')) return (v1_entity.key, v1_entity)
[ "def", "get_v1_mutation_key_and_entity", "(", "v1_mutation", ")", ":", "if", "v1_mutation", ".", "HasField", "(", "'delete'", ")", ":", "return", "(", "v1_mutation", ".", "delete", ",", "None", ")", "else", ":", "v1_entity", "=", "getattr", "(", "v1_mutation",...
returns the v1 key and entity for a v1 mutation proto .
train
false
50,736
def _guess_node_id(node): if (node is None): return __salt__['grains.get']('id') return node
[ "def", "_guess_node_id", "(", "node", ")", ":", "if", "(", "node", "is", "None", ")", ":", "return", "__salt__", "[", "'grains.get'", "]", "(", "'id'", ")", "return", "node" ]
try to guess kube node id using salt minion id .
train
false
50,737
def errfunc(*args): raise ValueError
[ "def", "errfunc", "(", "*", "args", ")", ":", "raise", "ValueError" ]
test function that raises an error .
train
false
50,738
def dirichlet_expectation(alpha): if (len(alpha.shape) == 1): result = (psi(alpha) - psi(np.sum(alpha))) else: result = (psi(alpha) - psi(np.sum(alpha, 1))[:, np.newaxis]) return result.astype(alpha.dtype)
[ "def", "dirichlet_expectation", "(", "alpha", ")", ":", "if", "(", "len", "(", "alpha", ".", "shape", ")", "==", "1", ")", ":", "result", "=", "(", "psi", "(", "alpha", ")", "-", "psi", "(", "np", ".", "sum", "(", "alpha", ")", ")", ")", "else"...
for a vector theta~dir .
train
false
50,740
def setTitle(template, title, chapterNumber): if (numberer.getNumberSections() and chapterNumber): titleNode = dom.Text() titleNode.ownerDocument = template.ownerDocument titleNode.data = ('%s. ' % (chapterNumber,)) title.insert(0, titleNode) for nodeList in (domhelpers.findNodesNamed(template, 'title'), domhelpers.findElementsWithAttribute(template, 'class', 'title')): if nodeList: for titleNode in title: nodeList[0].appendChild(titleNode.cloneNode(True))
[ "def", "setTitle", "(", "template", ",", "title", ",", "chapterNumber", ")", ":", "if", "(", "numberer", ".", "getNumberSections", "(", ")", "and", "chapterNumber", ")", ":", "titleNode", "=", "dom", ".", "Text", "(", ")", "titleNode", ".", "ownerDocument"...
add title and chapter number information to the template document .
train
false
50,741
def _bdecode(s): if (not s): return s return base64.decodestring(s)
[ "def", "_bdecode", "(", "s", ")", ":", "if", "(", "not", "s", ")", ":", "return", "s", "return", "base64", ".", "decodestring", "(", "s", ")" ]
decodes a base64 string .
train
false
50,744
@csrf_exempt def sql_explain(request): form = SQLSelectForm((request.POST or None)) if form.is_valid(): sql = form.cleaned_data[u'raw_sql'] params = form.cleaned_data[u'params'] vendor = form.connection.vendor cursor = form.cursor if (vendor == u'sqlite'): cursor.execute((u'EXPLAIN QUERY PLAN %s' % (sql,)), params) elif (vendor == u'postgresql'): cursor.execute((u'EXPLAIN ANALYZE %s' % (sql,)), params) else: cursor.execute((u'EXPLAIN %s' % (sql,)), params) headers = [d[0] for d in cursor.description] result = cursor.fetchall() cursor.close() context = {u'result': result, u'sql': form.reformat_sql(), u'duration': form.cleaned_data[u'duration'], u'headers': headers, u'alias': form.cleaned_data[u'alias']} return render_to_response(u'debug_toolbar/panels/sql_explain.html', context) return HttpResponseBadRequest(u'Form errors')
[ "@", "csrf_exempt", "def", "sql_explain", "(", "request", ")", ":", "form", "=", "SQLSelectForm", "(", "(", "request", ".", "POST", "or", "None", ")", ")", "if", "form", ".", "is_valid", "(", ")", ":", "sql", "=", "form", ".", "cleaned_data", "[", "u...
returns the output of the sql explain on the given query .
train
false
50,745
def health_state(consul_url=None, state=None, **kwargs): ret = {} query_params = {} if (not consul_url): consul_url = _get_config() if (not consul_url): log.error('No Consul URL found.') ret['message'] = 'No Consul URL found.' ret['res'] = False return ret if (not state): raise SaltInvocationError('Required argument "state" is missing.') if ('dc' in kwargs): query_params['dc'] = kwargs['dc'] if (state not in ('any', 'unknown', 'passing', 'warning', 'critical')): ret['message'] = 'State must be any, unknown, passing, warning, or critical.' ret['res'] = False return ret function = 'health/state/{0}'.format(state) ret = _query(consul_url=consul_url, function=function, query_params=query_params) return ret
[ "def", "health_state", "(", "consul_url", "=", "None", ",", "state", "=", "None", ",", "**", "kwargs", ")", ":", "ret", "=", "{", "}", "query_params", "=", "{", "}", "if", "(", "not", "consul_url", ")", ":", "consul_url", "=", "_get_config", "(", ")"...
returns the checks in the state provided on the path .
train
true
50,746
def maybe_schedule(s, relative=False, app=None): if (s is not None): if isinstance(s, numbers.Number): s = timedelta(seconds=s) if isinstance(s, timedelta): return schedule(s, relative, app=app) else: s.app = app return s
[ "def", "maybe_schedule", "(", "s", ",", "relative", "=", "False", ",", "app", "=", "None", ")", ":", "if", "(", "s", "is", "not", "None", ")", ":", "if", "isinstance", "(", "s", ",", "numbers", ".", "Number", ")", ":", "s", "=", "timedelta", "(",...
return schedule from number .
train
false
50,747
def _get_css_imports_regex(data): urls = [] for pattern in _CSS_URL_PATTERNS: for match in pattern.finditer(data): url = match.group('url') if url: urls.append(url) return urls
[ "def", "_get_css_imports_regex", "(", "data", ")", ":", "urls", "=", "[", "]", "for", "pattern", "in", "_CSS_URL_PATTERNS", ":", "for", "match", "in", "pattern", ".", "finditer", "(", "data", ")", ":", "url", "=", "match", ".", "group", "(", "'url'", "...
return all assets that are referenced in the given css document .
train
false
50,748
def to_7L5M(value): return ((value & 127), ((value >> 7) & 31))
[ "def", "to_7L5M", "(", "value", ")", ":", "return", "(", "(", "value", "&", "127", ")", ",", "(", "(", "value", ">>", "7", ")", "&", "31", ")", ")" ]
returns a list with the 7 lower bits of the value followed by the 5 higher bits .
train
false
50,749
@login_required(redirect_field_name='redirect_to') def login_protected_view_changed_redirect(request): t = Template('This is a login protected test. Username is {{ user.username }}.', name='Login Template') c = Context({'user': request.user}) return HttpResponse(t.render(c))
[ "@", "login_required", "(", "redirect_field_name", "=", "'redirect_to'", ")", "def", "login_protected_view_changed_redirect", "(", "request", ")", ":", "t", "=", "Template", "(", "'This is a login protected test. Username is {{ user.username }}.'", ",", "name", "=", "'Login...
a simple view that is login protected with a custom redirect field set .
train
false
50,750
def get_np_state_ptr(context, builder): return get_state_ptr(context, builder, 'np')
[ "def", "get_np_state_ptr", "(", "context", ",", "builder", ")", ":", "return", "get_state_ptr", "(", "context", ",", "builder", ",", "'np'", ")" ]
get a pointer to the thread-local numpy random state .
train
false
50,751
def _CreateClassFromElementTree(target_class, tree, namespace=None, tag=None): if (namespace is None): namespace = target_class._namespace if (tag is None): tag = target_class._tag if (tree.tag == ('{%s}%s' % (namespace, tag))): target = target_class() target._HarvestElementTree(tree) return target else: return None
[ "def", "_CreateClassFromElementTree", "(", "target_class", ",", "tree", ",", "namespace", "=", "None", ",", "tag", "=", "None", ")", ":", "if", "(", "namespace", "is", "None", ")", ":", "namespace", "=", "target_class", ".", "_namespace", "if", "(", "tag",...
instantiates the class and populates members according to the tree .
train
false