id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
45,412
def load_tool_elements_from_path(path, load_exception_handler=load_exception_handler, recursive=False, register_load_errors=False): return _load_tools_from_path(path, load_exception_handler=load_exception_handler, recursive=recursive, register_load_errors=register_load_errors, loader_func=loader.load_tool, enable_beta_formats=False)
[ "def", "load_tool_elements_from_path", "(", "path", ",", "load_exception_handler", "=", "load_exception_handler", ",", "recursive", "=", "False", ",", "register_load_errors", "=", "False", ")", ":", "return", "_load_tools_from_path", "(", "path", ",", "load_exception_ha...
walk a directory and load tool xml elements .
train
false
45,413
def _permute_observations(x, y, num_perms): vals = hstack([array(x), array(y)]) lenx = len(x) vals.sort() inds = arange(vals.size) (xs, ys) = ([], []) for i in range(num_perms): shuffle(inds) xs.append(vals[inds[:lenx]]) ys.append(vals[inds[lenx:]]) return (xs, ys)
[ "def", "_permute_observations", "(", "x", ",", "y", ",", "num_perms", ")", ":", "vals", "=", "hstack", "(", "[", "array", "(", "x", ")", ",", "array", "(", "y", ")", "]", ")", "lenx", "=", "len", "(", "x", ")", "vals", ".", "sort", "(", ")", ...
return num_perms pairs of permuted vectors x .
train
false
45,414
def stream_list(lst): def sublist_iterator(start, stop): return iter(lst[start:stop]) start = 0 while True: stop = len(lst) (yield sublist_iterator(start, stop)) start = stop
[ "def", "stream_list", "(", "lst", ")", ":", "def", "sublist_iterator", "(", "start", ",", "stop", ")", ":", "return", "iter", "(", "lst", "[", "start", ":", "stop", "]", ")", "start", "=", "0", "while", "True", ":", "stop", "=", "len", "(", "lst", ...
given a list .
train
false
45,417
def getLoopsFromLoopsDirection(isWiddershins, loops): directionalLoops = [] for loop in loops: if (euclidean.isWiddershins(loop) == isWiddershins): directionalLoops.append(loop) return directionalLoops
[ "def", "getLoopsFromLoopsDirection", "(", "isWiddershins", ",", "loops", ")", ":", "directionalLoops", "=", "[", "]", "for", "loop", "in", "loops", ":", "if", "(", "euclidean", ".", "isWiddershins", "(", "loop", ")", "==", "isWiddershins", ")", ":", "directi...
get the loops going round in a given direction .
train
false
45,418
def write_trunc_qual(trunc_qual_scores, qual_out_fp, seq_order): qual_line_size = 60 qual_out = open(qual_out_fp, 'w') for label in seq_order: trunc_label = label.split()[0].strip() current_trunc_qual_scores = trunc_qual_scores[trunc_label] qual_out.write(('>%s\n' % label)) current_qual_scores_lines = [] for slice in range(0, len(trunc_qual_scores[trunc_label]), qual_line_size): current_segment = current_trunc_qual_scores[slice:(slice + qual_line_size)] current_qual_scores_lines.append(' '.join(current_segment)) qual_out.write('\n'.join(current_qual_scores_lines)) qual_out.write('\n')
[ "def", "write_trunc_qual", "(", "trunc_qual_scores", ",", "qual_out_fp", ",", "seq_order", ")", ":", "qual_line_size", "=", "60", "qual_out", "=", "open", "(", "qual_out_fp", ",", "'w'", ")", "for", "label", "in", "seq_order", ":", "trunc_label", "=", "label",...
writes truncated quality score files out in proper format trunc_qual_scores: dict of seq label: numpy array of scores as ints qual_out_fp: output filepath to write truncated quality scores to seq_order: list of full fasta labels to write to output filepath and maintain the same order as input quality file .
train
false
45,419
def _resize(masked, new_size): new_array = ma.zeros((new_size,), dtype=masked.dtype) length = min(len(masked), new_size) try: new_array.data[:length] = masked.data[:length] except TypeError: new_array[:length] = masked[:length] else: if (length != 0): new_array.mask[:length] = masked.mask[:length] return new_array
[ "def", "_resize", "(", "masked", ",", "new_size", ")", ":", "new_array", "=", "ma", ".", "zeros", "(", "(", "new_size", ",", ")", ",", "dtype", "=", "masked", ".", "dtype", ")", "length", "=", "min", "(", "len", "(", "masked", ")", ",", "new_size",...
masked arrays can not be resized inplace .
train
false
45,420
def _compute_multivariate_sample_pacf(endog, maxlag): sample_autocovariances = _compute_multivariate_sample_acovf(endog, maxlag) return _compute_multivariate_pacf_from_autocovariances(sample_autocovariances)
[ "def", "_compute_multivariate_sample_pacf", "(", "endog", ",", "maxlag", ")", ":", "sample_autocovariances", "=", "_compute_multivariate_sample_acovf", "(", "endog", ",", "maxlag", ")", "return", "_compute_multivariate_pacf_from_autocovariances", "(", "sample_autocovariances", ...
computer multivariate sample partial autocorrelations parameters endog : array_like sample data on which to compute sample autocovariances .
train
false
45,421
def search_collections(query, limit, sort=None, cursor=None): return search_services.search(query, SEARCH_INDEX_COLLECTIONS, cursor, limit, sort, ids_only=True)
[ "def", "search_collections", "(", "query", ",", "limit", ",", "sort", "=", "None", ",", "cursor", "=", "None", ")", ":", "return", "search_services", ".", "search", "(", "query", ",", "SEARCH_INDEX_COLLECTIONS", ",", "cursor", ",", "limit", ",", "sort", ",...
searches through the available collections .
train
false
45,422
def truncate_week(dt, measure): return truncate_day(dt, (measure * 7))
[ "def", "truncate_week", "(", "dt", ",", "measure", ")", ":", "return", "truncate_day", "(", "dt", ",", "(", "measure", "*", "7", ")", ")" ]
truncate by weeks .
train
false
45,424
def makeColor(color): return make_color(color)
[ "def", "makeColor", "(", "color", ")", ":", "return", "make_color", "(", "color", ")" ]
an old name for the make_color function .
train
false
45,425
def _get_plural_forms(js_translations): plural = None n_plural = 2 if ('' in js_translations._catalog): for l in js_translations._catalog[''].split('\n'): if l.startswith('Plural-Forms:'): plural = l.split(':', 1)[1].strip() print ('plural is %s' % plural) if (plural is not None): for raw_element in plural.split(';'): element = raw_element.strip() if element.startswith('nplurals='): n_plural = int(element.split('=', 1)[1]) elif element.startswith('plural='): plural = element.split('=', 1)[1] print ('plural is now %s' % plural) else: n_plural = 2 plural = '(n == 1) ? 0 : 1' return (plural, n_plural)
[ "def", "_get_plural_forms", "(", "js_translations", ")", ":", "plural", "=", "None", "n_plural", "=", "2", "if", "(", "''", "in", "js_translations", ".", "_catalog", ")", ":", "for", "l", "in", "js_translations", ".", "_catalog", "[", "''", "]", ".", "sp...
extracts the parameters for what constitutes a plural .
train
false
45,426
def init(mpstate): return SerialModule(mpstate)
[ "def", "init", "(", "mpstate", ")", ":", "return", "SerialModule", "(", "mpstate", ")" ]
initialise module .
train
false
45,427
def split_name_version(name): prefix = '' if (name[0] in ['-', '~', '+']): prefix = name[0] name = name[1:] version_check = re.compile('^(.*?)((?:<|>|<=|>=|=)[0-9.-]*)?$') try: reres = version_check.match(name) (name, version) = reres.groups() return (prefix, name, version) except: return (prefix, name, None)
[ "def", "split_name_version", "(", "name", ")", ":", "prefix", "=", "''", "if", "(", "name", "[", "0", "]", "in", "[", "'-'", ",", "'~'", ",", "'+'", "]", ")", ":", "prefix", "=", "name", "[", "0", "]", "name", "=", "name", "[", "1", ":", "]",...
splits of the package name and desired version example formats: - docker>=1 .
train
false
45,428
def test_quantize_api(method, prec, exprange, restricted_range, itr, stat): for op in all_binary(prec, restricted_range, itr): for rounding in RoundModes: c = randcontext(exprange) quantizeop = (op[0], op[1], rounding, c) t = TestSet(method, quantizeop) try: if (not convert(t)): continue callfuncs(t) verify(t, stat) except VerifyError as err: log(err)
[ "def", "test_quantize_api", "(", "method", ",", "prec", ",", "exprange", ",", "restricted_range", ",", "itr", ",", "stat", ")", ":", "for", "op", "in", "all_binary", "(", "prec", ",", "restricted_range", ",", "itr", ")", ":", "for", "rounding", "in", "Ro...
iterate the quantize method through many test cases .
train
false
45,431
def _get_flags(**kwargs): flags = kwargs.get('flags', __salt__['config.option']('service.flags', default='')) return flags
[ "def", "_get_flags", "(", "**", "kwargs", ")", ":", "flags", "=", "kwargs", ".", "get", "(", "'flags'", ",", "__salt__", "[", "'config.option'", "]", "(", "'service.flags'", ",", "default", "=", "''", ")", ")", "return", "flags" ]
return an integer appropriate for use as a flag for the re module from a list of human-readable strings .
train
false
45,433
def chunks_from_arrays(arrays): if (not arrays): return () result = [] dim = 0 def shape(x): try: return x.shape except AttributeError: return (1,) while isinstance(arrays, (list, tuple)): result.append(tuple((shape(deepfirst(a))[dim] for a in arrays))) arrays = arrays[0] dim += 1 return tuple(result)
[ "def", "chunks_from_arrays", "(", "arrays", ")", ":", "if", "(", "not", "arrays", ")", ":", "return", "(", ")", "result", "=", "[", "]", "dim", "=", "0", "def", "shape", "(", "x", ")", ":", "try", ":", "return", "x", ".", "shape", "except", "Attr...
chunks tuple from nested list of arrays .
train
false
45,435
def update_firmware(filename, host=None, admin_username=None, admin_password=None): if os.path.exists(filename): return _update_firmware('update -f {0}'.format(filename), host=None, admin_username=None, admin_password=None) else: raise CommandExecutionError('Unable to find firmware file {0}'.format(filename))
[ "def", "update_firmware", "(", "filename", ",", "host", "=", "None", ",", "admin_username", "=", "None", ",", "admin_password", "=", "None", ")", ":", "if", "os", ".", "path", ".", "exists", "(", "filename", ")", ":", "return", "_update_firmware", "(", "...
updates firmware using local firmware file .
train
true
45,436
def _get_textbox(text, renderer): projected_xs = [] projected_ys = [] theta = np.deg2rad(text.get_rotation()) tr = mtransforms.Affine2D().rotate((- theta)) (_, parts, d) = text._get_layout(renderer) for (t, wh, x, y) in parts: (w, h) = wh (xt1, yt1) = tr.transform_point((x, y)) yt1 -= d (xt2, yt2) = ((xt1 + w), (yt1 + h)) projected_xs.extend([xt1, xt2]) projected_ys.extend([yt1, yt2]) (xt_box, yt_box) = (min(projected_xs), min(projected_ys)) (w_box, h_box) = ((max(projected_xs) - xt_box), (max(projected_ys) - yt_box)) tr = mtransforms.Affine2D().rotate(theta) (x_box, y_box) = tr.transform_point((xt_box, yt_box)) return (x_box, y_box, w_box, h_box)
[ "def", "_get_textbox", "(", "text", ",", "renderer", ")", ":", "projected_xs", "=", "[", "]", "projected_ys", "=", "[", "]", "theta", "=", "np", ".", "deg2rad", "(", "text", ".", "get_rotation", "(", ")", ")", "tr", "=", "mtransforms", ".", "Affine2D",...
calculate the bounding box of the text .
train
false
45,437
def rq_workers_are_running(): if redis_is_running(): queue = get_queue() if len(queue.connection.smembers(Worker.redis_workers_keys)): return True return False
[ "def", "rq_workers_are_running", "(", ")", ":", "if", "redis_is_running", "(", ")", ":", "queue", "=", "get_queue", "(", ")", "if", "len", "(", "queue", ".", "connection", ".", "smembers", "(", "Worker", ".", "redis_workers_keys", ")", ")", ":", "return", ...
checks if there are any rq workers running :returns: true if there are rq workers running .
train
false
45,440
def getTransferredPaths(insides, loop): transferredPaths = [] for insideIndex in xrange((len(insides) - 1), (-1), (-1)): inside = insides[insideIndex] if isPathInsideLoop(loop, inside): transferredPaths.append(inside) del insides[insideIndex] return transferredPaths
[ "def", "getTransferredPaths", "(", "insides", ",", "loop", ")", ":", "transferredPaths", "=", "[", "]", "for", "insideIndex", "in", "xrange", "(", "(", "len", "(", "insides", ")", "-", "1", ")", ",", "(", "-", "1", ")", ",", "(", "-", "1", ")", "...
get transferred paths from inside paths .
train
false
45,444
def get_programs_for_credentials(user, programs_credentials): certified_programs = [] programs = get_programs(user) for program in programs: for credential in programs_credentials: if (program[u'uuid'] == credential[u'credential'][u'program_uuid']): program[u'credential_url'] = credential[u'certificate_url'] certified_programs.append(program) return certified_programs
[ "def", "get_programs_for_credentials", "(", "user", ",", "programs_credentials", ")", ":", "certified_programs", "=", "[", "]", "programs", "=", "get_programs", "(", "user", ")", "for", "program", "in", "programs", ":", "for", "credential", "in", "programs_credent...
given a user and an iterable of credentials .
train
false
45,446
def dup_gcd(f, g, K): return dup_inner_gcd(f, g, K)[0]
[ "def", "dup_gcd", "(", "f", ",", "g", ",", "K", ")", ":", "return", "dup_inner_gcd", "(", "f", ",", "g", ",", "K", ")", "[", "0", "]" ]
computes polynomial gcd of f and g in k[x] .
train
false
45,448
def _authn_context_decl_ref(decl_ref, authn_auth=None): return factory(saml.AuthnContext, authn_context_decl_ref=decl_ref, authenticating_authority=factory(saml.AuthenticatingAuthority, text=authn_auth))
[ "def", "_authn_context_decl_ref", "(", "decl_ref", ",", "authn_auth", "=", "None", ")", ":", "return", "factory", "(", "saml", ".", "AuthnContext", ",", "authn_context_decl_ref", "=", "decl_ref", ",", "authenticating_authority", "=", "factory", "(", "saml", ".", ...
construct the authn context with a authn context declaration reference .
train
true
45,449
def to36(q): if (q < 0): raise ValueError('must supply a positive integer') letters = '0123456789abcdefghijklmnopqrstuvwxyz' converted = [] while (q != 0): (q, r) = divmod(q, 36) converted.insert(0, letters[r]) return (''.join(converted) or '0')
[ "def", "to36", "(", "q", ")", ":", "if", "(", "q", "<", "0", ")", ":", "raise", "ValueError", "(", "'must supply a positive integer'", ")", "letters", "=", "'0123456789abcdefghijklmnopqrstuvwxyz'", "converted", "=", "[", "]", "while", "(", "q", "!=", "0", ...
converts an integer to base 36 .
train
false
45,450
@hook.command('steamcalc', 'steamdb') def steamcalc(text): user = text.strip().lower() try: data = get_data(user) except SteamError as e: return '{}'.format(e) data['short_url'] = web.try_shorten(data['url']) return '\x02{name}\x02 has \x02{count:,}\x02 games with a total value of \x02{value}\x02 (\x02{value_sales}\x02 during sales). \x02{count_unplayed:,}\x02 (\x02{percent_unplayed}%\x02) have never been played - {short_url}'.format(**data)
[ "@", "hook", ".", "command", "(", "'steamcalc'", ",", "'steamdb'", ")", "def", "steamcalc", "(", "text", ")", ":", "user", "=", "text", ".", "strip", "(", ")", ".", "lower", "(", ")", "try", ":", "data", "=", "get_data", "(", "user", ")", "except",...
steamcalc <username> - gets value of steam account .
train
false
45,451
def show_most_common_types(limit=10, objects=None, shortnames=True): stats = most_common_types(limit, objects, shortnames=shortnames) width = max((len(name) for (name, count) in stats)) for (name, count) in stats: print ('%-*s %i' % (width, name, count))
[ "def", "show_most_common_types", "(", "limit", "=", "10", ",", "objects", "=", "None", ",", "shortnames", "=", "True", ")", ":", "stats", "=", "most_common_types", "(", "limit", ",", "objects", ",", "shortnames", "=", "shortnames", ")", "width", "=", "max"...
print the table of types of most common instances .
train
false
45,452
def get_answer(aphorism): time.sleep(0.0) return aphorisms.get(aphorism, 'Error: unknown aphorism.')
[ "def", "get_answer", "(", "aphorism", ")", ":", "time", ".", "sleep", "(", "0.0", ")", "return", "aphorisms", ".", "get", "(", "aphorism", ",", "'Error: unknown aphorism.'", ")" ]
return the string response to a particular zen-of-python aphorism .
train
false
45,453
def image_comparison(baseline_images=None, extensions=None, tol=0, freetype_version=None, remove_text=False, savefig_kwarg=None, style=u'classic'): if (baseline_images is None): raise ValueError(u'baseline_images must be specified') if (extensions is None): extensions = [u'png', u'pdf', u'svg'] if (savefig_kwarg is None): savefig_kwarg = dict() return ImageComparisonDecorator(baseline_images=baseline_images, extensions=extensions, tol=tol, freetype_version=freetype_version, remove_text=remove_text, savefig_kwargs=savefig_kwarg, style=style)
[ "def", "image_comparison", "(", "baseline_images", "=", "None", ",", "extensions", "=", "None", ",", "tol", "=", "0", ",", "freetype_version", "=", "None", ",", "remove_text", "=", "False", ",", "savefig_kwarg", "=", "None", ",", "style", "=", "u'classic'", ...
compare images generated by the test with those specified in *baseline_images* .
train
false
45,454
def p_type_name(t): pass
[ "def", "p_type_name", "(", "t", ")", ":", "pass" ]
type_name : specifier_qualifier_list abstract_declarator_opt .
train
false
45,455
def create_option_values(option_values): return _FakeOptionValues(option_values)
[ "def", "create_option_values", "(", "option_values", ")", ":", "return", "_FakeOptionValues", "(", "option_values", ")" ]
create a fake optionvaluecontainer object for testing .
train
false
45,456
def metrop_select(mr, q, q0): if (np.isfinite(mr) and (np.log(uniform()) < mr)): return q else: return q0
[ "def", "metrop_select", "(", "mr", ",", "q", ",", "q0", ")", ":", "if", "(", "np", ".", "isfinite", "(", "mr", ")", "and", "(", "np", ".", "log", "(", "uniform", "(", ")", ")", "<", "mr", ")", ")", ":", "return", "q", "else", ":", "return", ...
perform rejection/acceptance step for metropolis class samplers .
train
false
45,457
def is_trusted_ip(request): ip_addr = get_real_ip(request) return (ip_addr and any(((ip_addr in trusted_network) for trusted_network in request.app[KEY_TRUSTED_NETWORKS])))
[ "def", "is_trusted_ip", "(", "request", ")", ":", "ip_addr", "=", "get_real_ip", "(", "request", ")", "return", "(", "ip_addr", "and", "any", "(", "(", "(", "ip_addr", "in", "trusted_network", ")", "for", "trusted_network", "in", "request", ".", "app", "["...
test if request is from a trusted ip .
train
false
45,458
def read_raw_edf(input_fname, montage=None, eog=None, misc=None, stim_channel=(-1), annot=None, annotmap=None, exclude=(), preload=False, verbose=None): return RawEDF(input_fname=input_fname, montage=montage, eog=eog, misc=misc, stim_channel=stim_channel, annot=annot, annotmap=annotmap, exclude=exclude, preload=preload, verbose=verbose)
[ "def", "read_raw_edf", "(", "input_fname", ",", "montage", "=", "None", ",", "eog", "=", "None", ",", "misc", "=", "None", ",", "stim_channel", "=", "(", "-", "1", ")", ",", "annot", "=", "None", ",", "annotmap", "=", "None", ",", "exclude", "=", "...
reader function for edf+ .
train
false
45,459
def getNewMouseTool(): return ViewpointRotate()
[ "def", "getNewMouseTool", "(", ")", ":", "return", "ViewpointRotate", "(", ")" ]
get a new mouse tool .
train
false
45,460
def recursive_repr(func): repr_running = set() @wraps(func) def wrapper(self): key = (id(self), get_ident()) if (key in repr_running): return '...' repr_running.add(key) try: return func(self) finally: repr_running.discard(key) return wrapper
[ "def", "recursive_repr", "(", "func", ")", ":", "repr_running", "=", "set", "(", ")", "@", "wraps", "(", "func", ")", "def", "wrapper", "(", "self", ")", ":", "key", "=", "(", "id", "(", "self", ")", ",", "get_ident", "(", ")", ")", "if", "(", ...
decorator to make a repr function return fillvalue for a recursive call .
train
true
45,461
def get_cache(): return requests.Session().cache
[ "def", "get_cache", "(", ")", ":", "return", "requests", ".", "Session", "(", ")", ".", "cache" ]
returns internal cache object .
train
false
45,462
def get_declared_fields(bases, attrs, with_base_fields=True): fields = [(field_name, attrs.pop(field_name)) for (field_name, obj) in list(six.iteritems(attrs)) if isinstance(obj, Field)] fields.sort(key=(lambda x: x[1].creation_counter)) if with_base_fields: for base in bases[::(-1)]: if hasattr(base, u'base_fields'): fields = (list(six.iteritems(base.base_fields)) + fields) else: for base in bases[::(-1)]: if hasattr(base, u'declared_fields'): fields = (list(six.iteritems(base.declared_fields)) + fields) return SortedDict(fields)
[ "def", "get_declared_fields", "(", "bases", ",", "attrs", ",", "with_base_fields", "=", "True", ")", ":", "fields", "=", "[", "(", "field_name", ",", "attrs", ".", "pop", "(", "field_name", ")", ")", "for", "(", "field_name", ",", "obj", ")", "in", "li...
create a list of form field instances from the passed in attrs .
train
false
45,466
def mac_address(mac, for_item=True): if (not isinstance(mac, basestring)): raise CX('Invalid input, mac must be a string') else: mac = mac.lower().strip() if (for_item is True): if (mac == 'random'): return mac if (not netaddr.valid_mac(mac)): raise CX(('Invalid mac address format (%s)' % mac)) return mac
[ "def", "mac_address", "(", "mac", ",", "for_item", "=", "True", ")", ":", "if", "(", "not", "isinstance", "(", "mac", ",", "basestring", ")", ")", ":", "raise", "CX", "(", "'Invalid input, mac must be a string'", ")", "else", ":", "mac", "=", "mac", ".",...
validate as an eternet mac address .
train
false
45,467
def DrillURL(urlstats): drill = Drill() drill.reqsummary = Summary(urlstats) drill.rpcsummaries = SortedRPCSummaries(urlstats, 0.9) drill.rpc_variation = RPCVariation(drill.reqsummary, drill.rpcsummaries) groupcounts = urlstats.EntityGroupCount() (drill.groupcounts, drill.maxgroupcount) = SplitByKind(groupcounts) entitycounts = urlstats.EntityCount() (drill.entitycounts, drill.maxentitycount) = SplitByKind(entitycounts) drill_json = drill._ToJsonDrill() return drill_json
[ "def", "DrillURL", "(", "urlstats", ")", ":", "drill", "=", "Drill", "(", ")", "drill", ".", "reqsummary", "=", "Summary", "(", "urlstats", ")", "drill", ".", "rpcsummaries", "=", "SortedRPCSummaries", "(", "urlstats", ",", "0.9", ")", "drill", ".", "rpc...
analyzes url statistics and generates data for drill page .
train
false
45,468
def get_cdrom_attach_config_spec(client_factory, datastore, file_path, controller_key, cdrom_unit_number): config_spec = client_factory.create('ns0:VirtualMachineConfigSpec') device_config_spec = [] virtual_device_config_spec = create_virtual_cdrom_spec(client_factory, datastore, controller_key, file_path, cdrom_unit_number) device_config_spec.append(virtual_device_config_spec) config_spec.deviceChange = device_config_spec return config_spec
[ "def", "get_cdrom_attach_config_spec", "(", "client_factory", ",", "datastore", ",", "file_path", ",", "controller_key", ",", "cdrom_unit_number", ")", ":", "config_spec", "=", "client_factory", ".", "create", "(", "'ns0:VirtualMachineConfigSpec'", ")", "device_config_spe...
builds and returns the cdrom attach config spec .
train
false
45,469
def on_failure(exc_info): pass
[ "def", "on_failure", "(", "exc_info", ")", ":", "pass" ]
called if the function fails .
train
false
45,471
def describe_enum_value(enum_value): enum_value_descriptor = EnumValueDescriptor() enum_value_descriptor.name = unicode(enum_value.name) enum_value_descriptor.number = enum_value.number return enum_value_descriptor
[ "def", "describe_enum_value", "(", "enum_value", ")", ":", "enum_value_descriptor", "=", "EnumValueDescriptor", "(", ")", "enum_value_descriptor", ".", "name", "=", "unicode", "(", "enum_value", ".", "name", ")", "enum_value_descriptor", ".", "number", "=", "enum_va...
build descriptor for enum instance .
train
false
45,472
def score_from_local(directory): results = gym.monitoring.monitor.load_results(directory) if (results is None): return None episode_lengths = results['episode_lengths'] episode_rewards = results['episode_rewards'] episode_types = results['episode_types'] timestamps = results['timestamps'] initial_reset_timestamp = results['initial_reset_timestamp'] spec = gym.spec(results['env_info']['env_id']) return score_from_merged(episode_lengths, episode_rewards, episode_types, timestamps, initial_reset_timestamp, spec.trials, spec.reward_threshold)
[ "def", "score_from_local", "(", "directory", ")", ":", "results", "=", "gym", ".", "monitoring", ".", "monitor", ".", "load_results", "(", "directory", ")", "if", "(", "results", "is", "None", ")", ":", "return", "None", "episode_lengths", "=", "results", ...
calculate score from a local results directory .
train
false
45,473
def equateCylindrical(point, returnValue): point = evaluate.getVector3ByFloatList(returnValue, point) azimuthComplex = (euclidean.getWiddershinsUnitPolar(math.radians(point.y)) * point.x) point.x = azimuthComplex.real point.y = azimuthComplex.imag
[ "def", "equateCylindrical", "(", "point", ",", "returnValue", ")", ":", "point", "=", "evaluate", ".", "getVector3ByFloatList", "(", "returnValue", ",", "point", ")", "azimuthComplex", "=", "(", "euclidean", ".", "getWiddershinsUnitPolar", "(", "math", ".", "rad...
get equation for cylindrical .
train
false
45,474
def from_class_path(cls_path): (module_path, cls_name) = cls_path.split(':') module = __import__(module_path, fromlist=[cls_name]) return getattr(module, cls_name)
[ "def", "from_class_path", "(", "cls_path", ")", ":", "(", "module_path", ",", "cls_name", ")", "=", "cls_path", ".", "split", "(", "':'", ")", "module", "=", "__import__", "(", "module_path", ",", "fromlist", "=", "[", "cls_name", "]", ")", "return", "ge...
returns the class takes a class path and returns the class for it .
train
false
45,475
def make_message_identifier(command_id): return (MESSAGE_START + (command_id,))
[ "def", "make_message_identifier", "(", "command_id", ")", ":", "return", "(", "MESSAGE_START", "+", "(", "command_id", ",", ")", ")" ]
return the unique initial part of the sysex message without any arguments .
train
false
45,476
def make_group_frame(master, name=None, label=None, fill=Y, side=None, expand=None, font=None): font = (font or '-*-helvetica-medium-r-normal-*-*-100-*-*-*-*-*-*') outer = Frame(master, borderwidth=2, relief=GROOVE) outer.pack(expand=expand, fill=fill, side=side) if label: Label(outer, text=label, font=font, anchor=W).pack(fill=X) inner = Frame(master, borderwidth='1m', name=name) inner.pack(expand=1, fill=BOTH, in_=outer) inner.forget = outer.forget return inner
[ "def", "make_group_frame", "(", "master", ",", "name", "=", "None", ",", "label", "=", "None", ",", "fill", "=", "Y", ",", "side", "=", "None", ",", "expand", "=", "None", ",", "font", "=", "None", ")", ":", "font", "=", "(", "font", "or", "'-*-h...
create nested frames with a border and optional label .
train
false
45,477
def _reverse_and_conj(x): reverse = ([slice(None, None, (-1))] * x.ndim) return x[reverse].conj()
[ "def", "_reverse_and_conj", "(", "x", ")", ":", "reverse", "=", "(", "[", "slice", "(", "None", ",", "None", ",", "(", "-", "1", ")", ")", "]", "*", "x", ".", "ndim", ")", "return", "x", "[", "reverse", "]", ".", "conj", "(", ")" ]
reverse array x in all dimensions and perform the complex conjugate .
train
false
45,478
@receiver(SUBSECTION_SCORE_CHANGED) def recalculate_course_grade(sender, course, course_structure, user, **kwargs): CourseGradeFactory().update(user, course, course_structure)
[ "@", "receiver", "(", "SUBSECTION_SCORE_CHANGED", ")", "def", "recalculate_course_grade", "(", "sender", ",", "course", ",", "course_structure", ",", "user", ",", "**", "kwargs", ")", ":", "CourseGradeFactory", "(", ")", ".", "update", "(", "user", ",", "cours...
updates a saved course grade .
train
false
45,481
def test_import_vispy_app2(): allmodnames = loaded_vispy_modules('vispy.app', 2, True) assert_not_in('PySide', allmodnames) assert_not_in('PyQt4', allmodnames) assert_not_in('pyglet', allmodnames)
[ "def", "test_import_vispy_app2", "(", ")", ":", "allmodnames", "=", "loaded_vispy_modules", "(", "'vispy.app'", ",", "2", ",", "True", ")", "assert_not_in", "(", "'PySide'", ",", "allmodnames", ")", "assert_not_in", "(", "'PyQt4'", ",", "allmodnames", ")", "asse...
importing vispy .
train
false
45,482
@pytest.mark.parametrize(u'testframe', totest_frames) def test_icrs_altaz_moonish(testframe): (earth_pv_helio, earth_pv_bary) = epv00(*get_jd12(testframe.obstime, u'tdb')) earth_icrs_xyz = (earth_pv_bary[0] * u.au) moonoffset = ([0, 0, MOONDIST.value] * MOONDIST.unit) moonish_icrs = ICRS(CartesianRepresentation((earth_icrs_xyz + moonoffset))) moonaa = moonish_icrs.transform_to(testframe) assert ((1000 * u.km) < np.abs((moonaa.distance - MOONDIST)).to(u.au) < (7000 * u.km))
[ "@", "pytest", ".", "mark", ".", "parametrize", "(", "u'testframe'", ",", "totest_frames", ")", "def", "test_icrs_altaz_moonish", "(", "testframe", ")", ":", "(", "earth_pv_helio", ",", "earth_pv_bary", ")", "=", "epv00", "(", "*", "get_jd12", "(", "testframe"...
check that something expressed in *icrs* as being moon-like goes to the right altaz distance .
train
false
45,483
def comp_matrix(coefs): (p, k, k2) = coefs.shape assert (k == k2) kp = (k * p) result = np.zeros((kp, kp)) result[:k] = np.concatenate(coefs, axis=1) if (p > 1): result[(np.arange(k, kp), np.arange((kp - k)))] = 1 return result
[ "def", "comp_matrix", "(", "coefs", ")", ":", "(", "p", ",", "k", ",", "k2", ")", "=", "coefs", ".", "shape", "assert", "(", "k", "==", "k2", ")", "kp", "=", "(", "k", "*", "p", ")", "result", "=", "np", ".", "zeros", "(", "(", "kp", ",", ...
return compansion matrix for the var(1) representation for a var(p) process a = [a_1 a_2 .
train
false
45,484
def p_type_qualifier_list_1(t): pass
[ "def", "p_type_qualifier_list_1", "(", "t", ")", ":", "pass" ]
type_qualifier_list : type_qualifier .
train
false
45,487
def refusing_port(): (server_socket, port) = bind_unused_port() server_socket.setblocking(1) client_socket = socket.socket() client_socket.connect(('127.0.0.1', port)) (conn, client_addr) = server_socket.accept() conn.close() server_socket.close() return (client_socket.close, client_addr[1])
[ "def", "refusing_port", "(", ")", ":", "(", "server_socket", ",", "port", ")", "=", "bind_unused_port", "(", ")", "server_socket", ".", "setblocking", "(", "1", ")", "client_socket", "=", "socket", ".", "socket", "(", ")", "client_socket", ".", "connect", ...
returns a local port number that will refuse all connections .
train
false
45,488
@decorators.api_view(['GET']) @decorators.permission_classes((permissions.AllowAny,)) @decorators.renderer_classes((JSONRenderer,)) def docurl(request): project = request.GET.get('project') version = request.GET.get('version', LATEST) doc = request.GET.get('doc', 'index') if (project is None): return Response({'error': 'Need project and doc'}, status=status.HTTP_400_BAD_REQUEST) project = get_object_or_404(Project, slug=project) version = get_object_or_404(Version.objects.public(request.user, project=project, only_active=False), slug=version) return Response({'url': make_document_url(project=project, version=version.slug, page=doc)})
[ "@", "decorators", ".", "api_view", "(", "[", "'GET'", "]", ")", "@", "decorators", ".", "permission_classes", "(", "(", "permissions", ".", "AllowAny", ",", ")", ")", "@", "decorators", ".", "renderer_classes", "(", "(", "JSONRenderer", ",", ")", ")", "...
get the url that a slug resolves to .
train
false
45,489
def test_loads_sum_steps(): world.ran = False @before.each_step def assert_is_fine(step): world.ran = True runner = Runner(join(abspath(dirname(__file__)), 'simple_features', '2nd_feature_dir'), verbosity=0) runner.run() assert world.ran
[ "def", "test_loads_sum_steps", "(", ")", ":", "world", ".", "ran", "=", "False", "@", "before", ".", "each_step", "def", "assert_is_fine", "(", "step", ")", ":", "world", ".", "ran", "=", "True", "runner", "=", "Runner", "(", "join", "(", "abspath", "(...
can load step definitions from step_definitions folder .
train
false
45,490
def create_secret_key(): import random import string secret_key = list(((string.letters + string.digits) + string.punctuation).replace('\\', '').replace("'", '"').replace('{', '_').replace('}', '-')) random.shuffle(secret_key) secret_key = ''.join(secret_key[:40]) return secret_key
[ "def", "create_secret_key", "(", ")", ":", "import", "random", "import", "string", "secret_key", "=", "list", "(", "(", "(", "string", ".", "letters", "+", "string", ".", "digits", ")", "+", "string", ".", "punctuation", ")", ".", "replace", "(", "'\\\\'...
randomly create the secret key for the settings file .
train
false
45,492
def getMinimumRadius(beginComplexSegmentLength, endComplexSegmentLength, radius): return min(abs(radius), (0.5 * min(beginComplexSegmentLength, endComplexSegmentLength)))
[ "def", "getMinimumRadius", "(", "beginComplexSegmentLength", ",", "endComplexSegmentLength", ",", "radius", ")", ":", "return", "min", "(", "abs", "(", "radius", ")", ",", "(", "0.5", "*", "min", "(", "beginComplexSegmentLength", ",", "endComplexSegmentLength", ")...
get minimum radius .
train
false
45,495
def create_zipfile(directory, files, zipname='test'): zipfile_path = ((directory / zipname) + '.zip') with zipfile.ZipFile(str(zipfile_path), 'w') as new_zipfile: for file_path in files: new_zipfile.write(str(file_path), arcname=os.path.basename(str(file_path))) return str(zipfile_path)
[ "def", "create_zipfile", "(", "directory", ",", "files", ",", "zipname", "=", "'test'", ")", ":", "zipfile_path", "=", "(", "(", "directory", "/", "zipname", ")", "+", "'.zip'", ")", "with", "zipfile", ".", "ZipFile", "(", "str", "(", "zipfile_path", ")"...
return a path to a newly created zip file .
train
false
45,496
def _load_properties(property_name, config_option, set_default=False, default=None): if (not property_name): log.debug('No property specified in function, trying to load from salt configuration') try: options = __salt__['config.option']('cassandra') except BaseException as e: log.error('Failed to get cassandra config options. Reason: {0}'.format(str(e))) raise loaded_property = options.get(config_option) if (not loaded_property): if set_default: log.debug('Setting default Cassandra {0} to {1}'.format(config_option, default)) loaded_property = default else: log.error('No cassandra {0} specified in the configuration or passed to the module.'.format(config_option)) raise CommandExecutionError('ERROR: Cassandra {0} cannot be empty.'.format(config_option)) return loaded_property return property_name
[ "def", "_load_properties", "(", "property_name", ",", "config_option", ",", "set_default", "=", "False", ",", "default", "=", "None", ")", ":", "if", "(", "not", "property_name", ")", ":", "log", ".", "debug", "(", "'No property specified in function, trying to lo...
load properties for the cassandra module from config or pillar .
train
true
45,497
def linear_congruence(a, b, m): from sympy.polys.polytools import gcdex if ((a % m) == 0): if ((b % m) == 0): return list(range(m)) else: return [] (r, _, g) = gcdex(a, m) if ((b % g) != 0): return [] return [((((r * b) // g) + ((t * m) // g)) % m) for t in range(g)]
[ "def", "linear_congruence", "(", "a", ",", "b", ",", "m", ")", ":", "from", "sympy", ".", "polys", ".", "polytools", "import", "gcdex", "if", "(", "(", "a", "%", "m", ")", "==", "0", ")", ":", "if", "(", "(", "b", "%", "m", ")", "==", "0", ...
returns the values of x satisfying a*x congruent b mod(m) here m is positive integer and a .
train
false
45,500
def serve_download(path, name=None): return serve_file(path, 'application/x-download', 'attachment', name)
[ "def", "serve_download", "(", "path", ",", "name", "=", "None", ")", ":", "return", "serve_file", "(", "path", ",", "'application/x-download'", ",", "'attachment'", ",", "name", ")" ]
serve path as an application/x-download attachment .
train
false
45,501
def quote(string, safe=u'/'): if (sys.version_info.major < 3): if isinstance(string, unicode): string = string.encode(u'utf8') string = urllib.quote(string, safe.encode(u'utf8')) else: string = urllib.parse.quote(str(string), safe) return string
[ "def", "quote", "(", "string", ",", "safe", "=", "u'/'", ")", ":", "if", "(", "sys", ".", "version_info", ".", "major", "<", "3", ")", ":", "if", "isinstance", "(", "string", ",", "unicode", ")", ":", "string", "=", "string", ".", "encode", "(", ...
returns quoted po term string .
train
false
45,502
def getchunks(im, **params): class collector(object, ): data = [] def write(self, data): pass def append(self, chunk): self.data.append(chunk) def append(fp, cid, *data): data = ''.join(data) (hi, lo) = Image.core.crc32(data, Image.core.crc32(cid)) crc = (o16(hi) + o16(lo)) fp.append((cid, data, crc)) fp = collector() try: im.encoderinfo = params _save(im, fp, None, append) finally: del im.encoderinfo return fp.data
[ "def", "getchunks", "(", "im", ",", "**", "params", ")", ":", "class", "collector", "(", "object", ",", ")", ":", "data", "=", "[", "]", "def", "write", "(", "self", ",", "data", ")", ":", "pass", "def", "append", "(", "self", ",", "chunk", ")", ...
return a list of png chunks representing this image .
train
false
45,503
def search_ctr(start_date, end_date): ctr = {} request = _build_request() date = start_date while (date <= end_date): date_str = str(date) metric_name = 'ga:goal11ConversionRate' ctr_str = request.get(ids=('ga:' + profile_id), start_date=date_str, end_date=date_str, metrics=metric_name).execute()['rows'][0][0] ctr[date_str] = float(ctr_str) date += timedelta(days=1) return ctr
[ "def", "search_ctr", "(", "start_date", ",", "end_date", ")", ":", "ctr", "=", "{", "}", "request", "=", "_build_request", "(", ")", "date", "=", "start_date", "while", "(", "date", "<=", "end_date", ")", ":", "date_str", "=", "str", "(", "date", ")", ...
return search click through rate based on goal 11 in google analytics .
train
false
45,504
def _parse_task_stderr(lines): task_error = None stack_trace_start_line = None for (line_num, line) in enumerate(lines): line = line.rstrip('\r\n') if _SUBPROCESS_FAILED_STACK_TRACE_START.match(line): stack_trace_start_line = line_num continue if (stack_trace_start_line is not None): if (line.lstrip() != line): continue else: stack_trace_start_line = None if _TASK_STDERR_IGNORE_RE.match(line): if (task_error and ('num_lines' not in task_error)): task_error['num_lines'] = (line_num - task_error['start_line']) continue elif ((not task_error) or line.startswith('+ ')): task_error = dict(message=line, start_line=line_num) else: task_error['message'] += ('\n' + line) if task_error: if ('num_lines' not in task_error): end_line = (stack_trace_start_line or (line_num + 1)) task_error['num_lines'] = (end_line - task_error['start_line']) return task_error else: return None
[ "def", "_parse_task_stderr", "(", "lines", ")", ":", "task_error", "=", "None", "stack_trace_start_line", "=", "None", "for", "(", "line_num", ",", "line", ")", "in", "enumerate", "(", "lines", ")", ":", "line", "=", "line", ".", "rstrip", "(", "'\\r\\n'",...
attempt to explain any error in task stderr .
train
false
45,505
def get_link_suffix(symlink): with settings(warn_only=True): if (not exists(('~/%s' % symlink))): return None target = run(('readlink ~/%s' % symlink)) if (target.return_code != 0): return None suffix = get_file_suffix(symlink, target) assert (suffix is not None), ('Could not determine suffix from filename %s' % target) return suffix
[ "def", "get_link_suffix", "(", "symlink", ")", ":", "with", "settings", "(", "warn_only", "=", "True", ")", ":", "if", "(", "not", "exists", "(", "(", "'~/%s'", "%", "symlink", ")", ")", ")", ":", "return", "None", "target", "=", "run", "(", "(", "...
follow symlink in ~/ and determine the suffix for the target (of the form <symlink> .
train
false
45,508
def decode_file_iter(file_obj, codec_options=DEFAULT_CODEC_OPTIONS): while True: size_data = file_obj.read(4) if (len(size_data) == 0): break elif (len(size_data) != 4): raise InvalidBSON('cut off in middle of objsize') obj_size = (_UNPACK_INT(size_data)[0] - 4) elements = (size_data + file_obj.read(obj_size)) (yield _bson_to_dict(elements, codec_options))
[ "def", "decode_file_iter", "(", "file_obj", ",", "codec_options", "=", "DEFAULT_CODEC_OPTIONS", ")", ":", "while", "True", ":", "size_data", "=", "file_obj", ".", "read", "(", "4", ")", "if", "(", "len", "(", "size_data", ")", "==", "0", ")", ":", "break...
decode bson data from a file to multiple documents as a generator .
train
true
45,509
def hashable(a): if isinstance(a, dict): return hashable(a.items()) try: return tuple(map(hashable, a)) except: return a
[ "def", "hashable", "(", "a", ")", ":", "if", "isinstance", "(", "a", ",", "dict", ")", ":", "return", "hashable", "(", "a", ".", "items", "(", ")", ")", "try", ":", "return", "tuple", "(", "map", "(", "hashable", ",", "a", ")", ")", "except", "...
turn some unhashable objects into hashable ones .
train
false
45,510
def container_configuration_response(application, node): result = {'node_uuid': unicode(node), 'name': application.name} result.update(ApplicationMarshaller(application).convert()) if (u'volume' in result): volume = result.pop(u'volume') result[u'volumes'] = [{u'dataset_id': volume[u'dataset_id'], u'mountpoint': volume[u'mountpoint']}] if (application.cpu_shares is not None): result['cpu_shares'] = application.cpu_shares if (application.memory_limit is not None): result['memory_limit'] = application.memory_limit if (application.command_line is not None): result['command_line'] = list(application.command_line) return result
[ "def", "container_configuration_response", "(", "application", ",", "node", ")", ":", "result", "=", "{", "'node_uuid'", ":", "unicode", "(", "node", ")", ",", "'name'", ":", "application", ".", "name", "}", "result", ".", "update", "(", "ApplicationMarshaller...
return a container dict which confirms to /v1/endpoints .
train
false
45,512
def get_logout_url(*args, **kwargs): return get_client().get_logout_url(*args, **kwargs)
[ "def", "get_logout_url", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "get_client", "(", ")", ".", "get_logout_url", "(", "*", "args", ",", "**", "kwargs", ")" ]
convenience function for getting a logout url for a service .
train
false
45,515
def debug_error_message(msg): action = config.compute_test_value assert (action != 'off') if (action in ['raise', 'ignore']): raise ValueError(msg) else: assert (action == 'warn') warnings.warn(msg, stacklevel=2)
[ "def", "debug_error_message", "(", "msg", ")", ":", "action", "=", "config", ".", "compute_test_value", "assert", "(", "action", "!=", "'off'", ")", "if", "(", "action", "in", "[", "'raise'", ",", "'ignore'", "]", ")", ":", "raise", "ValueError", "(", "m...
displays a message saying that an error was found in some test_values .
train
false
45,516
def CDLMORNINGSTAR(barDs, count, penetration=(-4e+37)): return call_talib_with_ohlc(barDs, count, talib.CDLMORNINGSTAR, penetration)
[ "def", "CDLMORNINGSTAR", "(", "barDs", ",", "count", ",", "penetration", "=", "(", "-", "4e+37", ")", ")", ":", "return", "call_talib_with_ohlc", "(", "barDs", ",", "count", ",", "talib", ".", "CDLMORNINGSTAR", ",", "penetration", ")" ]
morning star .
train
false
45,517
def get_cohort(user, course_key, assign=True, use_cached=False): request_cache = RequestCache.get_request_cache() cache_key = u'cohorts.get_cohort.{}.{}'.format(user.id, course_key) if (use_cached and (cache_key in request_cache.data)): return request_cache.data[cache_key] request_cache.data.pop(cache_key, None) course_cohort_settings = get_course_cohort_settings(course_key) if (not course_cohort_settings.is_cohorted): return request_cache.data.setdefault(cache_key, None) try: membership = CohortMembership.objects.get(course_id=course_key, user_id=user.id) return request_cache.data.setdefault(cache_key, membership.course_user_group) except CohortMembership.DoesNotExist: if (not assign): return None membership = CohortMembership.objects.create(user=user, course_user_group=get_random_cohort(course_key)) return request_cache.data.setdefault(cache_key, membership.course_user_group)
[ "def", "get_cohort", "(", "user", ",", "course_key", ",", "assign", "=", "True", ",", "use_cached", "=", "False", ")", ":", "request_cache", "=", "RequestCache", ".", "get_request_cache", "(", ")", "cache_key", "=", "u'cohorts.get_cohort.{}.{}'", ".", "format", ...
returns the users cohort for the specified course .
train
false
45,518
def apt_get_update(sudo=False): return _from_args(sudo)(['apt-get', 'update'])
[ "def", "apt_get_update", "(", "sudo", "=", "False", ")", ":", "return", "_from_args", "(", "sudo", ")", "(", "[", "'apt-get'", ",", "'update'", "]", ")" ]
update apts package metadata cache .
train
false
45,519
def sqlquote(a): if isinstance(a, list): return _sqllist(a) else: return sqlparam(a).sqlquery()
[ "def", "sqlquote", "(", "a", ")", ":", "if", "isinstance", "(", "a", ",", "list", ")", ":", "return", "_sqllist", "(", "a", ")", "else", ":", "return", "sqlparam", "(", "a", ")", ".", "sqlquery", "(", ")" ]
ensures a is quoted properly for use in a sql query .
train
false
45,521
@with_device def proc_exe(pid): with context.quiet: io = process(['realpath', ('/proc/%d/exe' % pid)]) data = io.recvall().strip() return data
[ "@", "with_device", "def", "proc_exe", "(", "pid", ")", ":", "with", "context", ".", "quiet", ":", "io", "=", "process", "(", "[", "'realpath'", ",", "(", "'/proc/%d/exe'", "%", "pid", ")", "]", ")", "data", "=", "io", ".", "recvall", "(", ")", "."...
returns the full path of the executable for the provided pid .
train
false
45,522
def _get_reason(cluster_or_step): return getattr(getattr(cluster_or_step.status, 'statechangereason', ''), 'message', '').rstrip()
[ "def", "_get_reason", "(", "cluster_or_step", ")", ":", "return", "getattr", "(", "getattr", "(", "cluster_or_step", ".", "status", ",", "'statechangereason'", ",", "''", ")", ",", "'message'", ",", "''", ")", ".", "rstrip", "(", ")" ]
extract statechangereason .
train
false
45,523
def build_from_document(service, base, future=None, http=None, developerKey=None, model=None, requestBuilder=HttpRequest): service = simplejson.loads(service) base = urlparse.urljoin(base, service['basePath']) if future: future = simplejson.loads(future) auth_discovery = future.get('auth', {}) else: future = {} auth_discovery = {} schema = Schemas(service) if (model is None): features = service.get('features', []) model = JsonModel(('dataWrapper' in features)) resource = createResource(http, base, model, requestBuilder, developerKey, service, future, schema) def auth_method(): 'Discovery information about the authentication the API uses.' return auth_discovery setattr(resource, 'auth_discovery', auth_method) return resource
[ "def", "build_from_document", "(", "service", ",", "base", ",", "future", "=", "None", ",", "http", "=", "None", ",", "developerKey", "=", "None", ",", "model", "=", "None", ",", "requestBuilder", "=", "HttpRequest", ")", ":", "service", "=", "simplejson",...
create a resource for interacting with an api .
train
false
45,524
def _set_vhostuser_settings(vif, obj): obj.mode = vif['details'].get(model.VIF_DETAILS_VHOSTUSER_MODE, 'server') path = vif['details'].get(model.VIF_DETAILS_VHOSTUSER_SOCKET, None) if path: obj.path = path else: raise exception.VifDetailsMissingVhostuserSockPath(vif_id=vif['id'])
[ "def", "_set_vhostuser_settings", "(", "vif", ",", "obj", ")", ":", "obj", ".", "mode", "=", "vif", "[", "'details'", "]", ".", "get", "(", "model", ".", "VIF_DETAILS_VHOSTUSER_MODE", ",", "'server'", ")", "path", "=", "vif", "[", "'details'", "]", ".", ...
set vhostuser socket mode and path .
train
false
45,525
def getPillarsOutput(loopLists): pillarsOutput = [] for loopList in loopLists: pillarsOutput.append(getPillarOutput(loopList)) return getUnifiedOutput(pillarsOutput)
[ "def", "getPillarsOutput", "(", "loopLists", ")", ":", "pillarsOutput", "=", "[", "]", "for", "loopList", "in", "loopLists", ":", "pillarsOutput", ".", "append", "(", "getPillarOutput", "(", "loopList", ")", ")", "return", "getUnifiedOutput", "(", "pillarsOutput...
get pillars output .
train
false
45,527
def mt_services(request): machine_services = list(MACHINE_TRANSLATION_SERVICES.keys()) return JsonResponse(data=machine_services, safe=False)
[ "def", "mt_services", "(", "request", ")", ":", "machine_services", "=", "list", "(", "MACHINE_TRANSLATION_SERVICES", ".", "keys", "(", ")", ")", "return", "JsonResponse", "(", "data", "=", "machine_services", ",", "safe", "=", "False", ")" ]
generates list of installed machine translation services in json .
train
false
45,528
def _get_distset(tgt): tgtattrs = tgt.split('-') if (tgtattrs[0] == 'amzn'): distset = '--define "dist .{0}1"'.format(tgtattrs[0]) elif (tgtattrs[1] in ['5', '6', '7']): distset = '--define "dist .el{0}"'.format(tgtattrs[1]) else: distset = '' return distset
[ "def", "_get_distset", "(", "tgt", ")", ":", "tgtattrs", "=", "tgt", ".", "split", "(", "'-'", ")", "if", "(", "tgtattrs", "[", "0", "]", "==", "'amzn'", ")", ":", "distset", "=", "'--define \"dist .{0}1\"'", ".", "format", "(", "tgtattrs", "[", "0", ...
get the distribution string for use with rpmbuild and mock .
train
true
45,529
def _mminion(): global MMINION if (MMINION is None): MMINION = salt.minion.MasterMinion(__opts__) return MMINION
[ "def", "_mminion", "(", ")", ":", "global", "MMINION", "if", "(", "MMINION", "is", "None", ")", ":", "MMINION", "=", "salt", ".", "minion", ".", "MasterMinion", "(", "__opts__", ")", "return", "MMINION" ]
create a single mminion for this module to use .
train
false
45,530
def categorical_sample(prob_n, np_random): prob_n = np.asarray(prob_n) csprob_n = np.cumsum(prob_n) return (csprob_n > np_random.rand()).argmax()
[ "def", "categorical_sample", "(", "prob_n", ",", "np_random", ")", ":", "prob_n", "=", "np", ".", "asarray", "(", "prob_n", ")", "csprob_n", "=", "np", ".", "cumsum", "(", "prob_n", ")", "return", "(", "csprob_n", ">", "np_random", ".", "rand", "(", ")...
sample from categorical distribution each row specifies class probabilities .
train
false
45,532
def add_null_handler_to_root_logger(): log = logging.getLogger() if (NULL_HANDLER not in log.handlers): log.addHandler(NULL_HANDLER)
[ "def", "add_null_handler_to_root_logger", "(", ")", ":", "log", "=", "logging", ".", "getLogger", "(", ")", "if", "(", "NULL_HANDLER", "not", "in", "log", ".", "handlers", ")", ":", "log", ".", "addHandler", "(", "NULL_HANDLER", ")" ]
add a nullhandler to the root logger .
train
false
45,533
def monitor_copy_global(sock, orig_name, new_name): return communicate(sock, ('__copy_global__("%s", "%s")' % (orig_name, new_name)))
[ "def", "monitor_copy_global", "(", "sock", ",", "orig_name", ",", "new_name", ")", ":", "return", "communicate", "(", "sock", ",", "(", "'__copy_global__(\"%s\", \"%s\")'", "%", "(", "orig_name", ",", "new_name", ")", ")", ")" ]
copy global variable *orig_name* to *new_name* .
train
false
45,534
def unique_resource_id(delimiter='_'): build_id = os.getenv('TRAVIS_BUILD_ID', os.getenv('CIRCLE_BUILD_NUM', '')) if (build_id == ''): return ('%s%d' % (delimiter, (1000 * time.time()))) else: return ('%s%s%s%d' % (delimiter, build_id, delimiter, time.time()))
[ "def", "unique_resource_id", "(", "delimiter", "=", "'_'", ")", ":", "build_id", "=", "os", ".", "getenv", "(", "'TRAVIS_BUILD_ID'", ",", "os", ".", "getenv", "(", "'CIRCLE_BUILD_NUM'", ",", "''", ")", ")", "if", "(", "build_id", "==", "''", ")", ":", ...
a unique identifier for a resource .
train
false
45,535
def dup_to_raw_dict(f, K=None, zero=False): if ((not f) and zero): return {0: K.zero} (n, result) = ((len(f) - 1), {}) for k in range(0, (n + 1)): if f[(n - k)]: result[k] = f[(n - k)] return result
[ "def", "dup_to_raw_dict", "(", "f", ",", "K", "=", "None", ",", "zero", "=", "False", ")", ":", "if", "(", "(", "not", "f", ")", "and", "zero", ")", ":", "return", "{", "0", ":", "K", ".", "zero", "}", "(", "n", ",", "result", ")", "=", "("...
convert a k[x] polynomial to a raw dict .
train
false
45,537
def makeGauss(x, mean=0.0, sd=1.0, gain=1.0, base=0.0): simpleGauss = numpy.exp(((- numpy.power((mean - x), 2)) / (2 * (sd ** 2)))) return (base + (gain * simpleGauss))
[ "def", "makeGauss", "(", "x", ",", "mean", "=", "0.0", ",", "sd", "=", "1.0", ",", "gain", "=", "1.0", ",", "base", "=", "0.0", ")", ":", "simpleGauss", "=", "numpy", ".", "exp", "(", "(", "(", "-", "numpy", ".", "power", "(", "(", "mean", "-...
return the gaussian distribution for a given set of x-vals .
train
false
45,538
def libvlc_media_list_player_new(p_instance): f = (_Cfunctions.get('libvlc_media_list_player_new', None) or _Cfunction('libvlc_media_list_player_new', ((1,),), class_result(MediaListPlayer), ctypes.c_void_p, Instance)) return f(p_instance)
[ "def", "libvlc_media_list_player_new", "(", "p_instance", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_media_list_player_new'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_media_list_player_new'", ",", "(", "(", "1", ",", ")", ",", ...
create new media_list_player .
train
true
45,541
def build_treeprocessors(md_instance, **kwargs): treeprocessors = odict.OrderedDict() treeprocessors[u'inline'] = InlineProcessor(md_instance) treeprocessors[u'prettify'] = PrettifyTreeprocessor(md_instance) return treeprocessors
[ "def", "build_treeprocessors", "(", "md_instance", ",", "**", "kwargs", ")", ":", "treeprocessors", "=", "odict", ".", "OrderedDict", "(", ")", "treeprocessors", "[", "u'inline'", "]", "=", "InlineProcessor", "(", "md_instance", ")", "treeprocessors", "[", "u'pr...
build the default treeprocessors for markdown .
train
false
45,542
def getScreen(screen): screens = getScreens() if (screen > (len(screens) - 1)): msg = 'Requested refresh rate of screen %i, but only have %i screens.' raise IndexError((msg % (screen, len(screens)))) return getScreens()[screen]
[ "def", "getScreen", "(", "screen", ")", ":", "screens", "=", "getScreens", "(", ")", "if", "(", "screen", ">", "(", "len", "(", "screens", ")", "-", "1", ")", ")", ":", "msg", "=", "'Requested refresh rate of screen %i, but only have %i screens.'", "raise", ...
select screen from getscreens() .
train
false
45,543
def _get_last_modified(last_modified=None): if (last_modified is None): dt = datetime.datetime.utcnow() else: dt = datetime.datetime.utcfromtimestamp((last_modified / 1000.0)) dt = dt.replace(microsecond=0) return (u'%sZ' % dt.isoformat())
[ "def", "_get_last_modified", "(", "last_modified", "=", "None", ")", ":", "if", "(", "last_modified", "is", "None", ")", ":", "dt", "=", "datetime", ".", "datetime", ".", "utcnow", "(", ")", "else", ":", "dt", "=", "datetime", ".", "datetime", ".", "ut...
formats last modified timestamp of a playlist for mpd .
train
false
45,545
def _validate_rpc_ip(rpc_server_ip): if (not is_valid_ipv4(rpc_server_ip)): raise NetworkControllerError(desc='Invalid rpc ip address.') return rpc_server_ip
[ "def", "_validate_rpc_ip", "(", "rpc_server_ip", ")", ":", "if", "(", "not", "is_valid_ipv4", "(", "rpc_server_ip", ")", ")", ":", "raise", "NetworkControllerError", "(", "desc", "=", "'Invalid rpc ip address.'", ")", "return", "rpc_server_ip" ]
validates given ip for use as rpc host bind address .
train
false
45,549
def loaddata(settings_module, fixtures, bin_env=None, database=None, pythonpath=None, env=None): kwargs = {} if database: kwargs['database'] = database return command(settings_module, 'loaddata', bin_env, pythonpath, env, *fixtures.split(','), **kwargs)
[ "def", "loaddata", "(", "settings_module", ",", "fixtures", ",", "bin_env", "=", "None", ",", "database", "=", "None", ",", "pythonpath", "=", "None", ",", "env", "=", "None", ")", ":", "kwargs", "=", "{", "}", "if", "database", ":", "kwargs", "[", "...
load fixture data fixtures: comma separated list of fixtures to load cli example: .
train
false
45,551
def stream_decode_response_unicode(iterator, r): if (r.encoding is None): for item in iterator: (yield item) return decoder = codecs.getincrementaldecoder(r.encoding)(errors='replace') for chunk in iterator: rv = decoder.decode(chunk) if rv: (yield rv) rv = decoder.decode('', final=True) if rv: (yield rv)
[ "def", "stream_decode_response_unicode", "(", "iterator", ",", "r", ")", ":", "if", "(", "r", ".", "encoding", "is", "None", ")", ":", "for", "item", "in", "iterator", ":", "(", "yield", "item", ")", "return", "decoder", "=", "codecs", ".", "getincrement...
stream decodes a iterator .
train
true
45,552
def delete_minion_cachedir(minion_id, provider, opts, base=None): if isinstance(opts, dict): __opts__.update(opts) if (__opts__.get('update_cachedir', False) is False): return if (base is None): base = __opts__['cachedir'] driver = next(six.iterkeys(__opts__['providers'][provider])) fname = '{0}.p'.format(minion_id) for cachedir in ('requested', 'active'): path = os.path.join(base, cachedir, driver, provider, fname) log.debug('path: {0}'.format(path)) if os.path.exists(path): os.remove(path)
[ "def", "delete_minion_cachedir", "(", "minion_id", ",", "provider", ",", "opts", ",", "base", "=", "None", ")", ":", "if", "isinstance", "(", "opts", ",", "dict", ")", ":", "__opts__", ".", "update", "(", "opts", ")", "if", "(", "__opts__", ".", "get",...
deletes a minions entry from the cloud cachedir .
train
true
45,554
def clean_image_extension(form_field): if form_field: if ('.' not in form_field.name): raise ValidationError(MSG_IMAGE_EXTENSION) (_, ext) = form_field.name.rsplit('.', 1) if (ext.lower() not in ALLOWED_IMAGE_EXTENSIONS): raise ValidationError(MSG_IMAGE_EXTENSION)
[ "def", "clean_image_extension", "(", "form_field", ")", ":", "if", "form_field", ":", "if", "(", "'.'", "not", "in", "form_field", ".", "name", ")", ":", "raise", "ValidationError", "(", "MSG_IMAGE_EXTENSION", ")", "(", "_", ",", "ext", ")", "=", "form_fie...
ensure only images of certain extensions can be uploaded .
train
false
45,555
def proxyconnect_sso(request): if request.user.is_authenticated(): return mysite.base.decorators.as_view(request, 'vanilla-proxy-connect-sso.txt', {}, 'proxyconnect-sso') return HttpResponse('')
[ "def", "proxyconnect_sso", "(", "request", ")", ":", "if", "request", ".", "user", ".", "is_authenticated", "(", ")", ":", "return", "mysite", ".", "base", ".", "decorators", ".", "as_view", "(", "request", ",", "'vanilla-proxy-connect-sso.txt'", ",", "{", "...
this function implements the proxyconnect single sign-on api described by vanilla forums .
train
false