id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
10,313
def needs_local_scope(func): func.needs_local_scope = True return func
[ "def", "needs_local_scope", "(", "func", ")", ":", "func", ".", "needs_local_scope", "=", "True", "return", "func" ]
decorator to mark magic functions which need to local scope to run .
train
false
10,314
def _api_restart_repair(name, output, kwargs): sabnzbd.request_repair() sabnzbd.trigger_restart() return report(output)
[ "def", "_api_restart_repair", "(", "name", ",", "output", ",", "kwargs", ")", ":", "sabnzbd", ".", "request_repair", "(", ")", "sabnzbd", ".", "trigger_restart", "(", ")", "return", "report", "(", "output", ")" ]
api: accepts output .
train
false
10,315
@verbose def followtoscreen_demo(limit=10): oauth = credsfromfile() client = Streamer(**oauth) client.register(TweetViewer(limit=limit)) client.statuses.filter(follow=USERIDS)
[ "@", "verbose", "def", "followtoscreen_demo", "(", "limit", "=", "10", ")", ":", "oauth", "=", "credsfromfile", "(", ")", "client", "=", "Streamer", "(", "**", "oauth", ")", "client", ".", "register", "(", "TweetViewer", "(", "limit", "=", "limit", ")", ...
using the streaming api .
train
false
10,318
def test_show_verbose_with_classifiers(script): result = script.pip('show', 'pip', '--verbose') lines = result.stdout.splitlines() assert ('Name: pip' in lines) assert re.search('Classifiers:\\n( .+\\n)+', result.stdout) assert ('Intended Audience :: Developers' in result.stdout)
[ "def", "test_show_verbose_with_classifiers", "(", "script", ")", ":", "result", "=", "script", ".", "pip", "(", "'show'", ",", "'pip'", ",", "'--verbose'", ")", "lines", "=", "result", ".", "stdout", ".", "splitlines", "(", ")", "assert", "(", "'Name: pip'",...
test that classifiers can be listed .
train
false
10,321
def worker_direct(hostname): if isinstance(hostname, Queue): return hostname return Queue(WORKER_DIRECT_QUEUE_FORMAT.format(hostname=hostname), WORKER_DIRECT_EXCHANGE, hostname)
[ "def", "worker_direct", "(", "hostname", ")", ":", "if", "isinstance", "(", "hostname", ",", "Queue", ")", ":", "return", "hostname", "return", "Queue", "(", "WORKER_DIRECT_QUEUE_FORMAT", ".", "format", "(", "hostname", "=", "hostname", ")", ",", "WORKER_DIREC...
return the :class:kombu .
train
false
10,322
def GetProgram(plist): try: return (("['%s']" % plist['Program']), HashFile(plist['Program'])) except KeyError: return (plist['ProgramArguments'], HashFile(plist['ProgramArguments']))
[ "def", "GetProgram", "(", "plist", ")", ":", "try", ":", "return", "(", "(", "\"['%s']\"", "%", "plist", "[", "'Program'", "]", ")", ",", "HashFile", "(", "plist", "[", "'Program'", "]", ")", ")", "except", "KeyError", ":", "return", "(", "plist", "[...
plists have either a program or programarguments key .
train
false
10,324
def force_escape(value): from django.utils.html import escape return mark_safe(escape(value))
[ "def", "force_escape", "(", "value", ")", ":", "from", "django", ".", "utils", ".", "html", "import", "escape", "return", "mark_safe", "(", "escape", "(", "value", ")", ")" ]
escapes a strings html .
train
false
10,325
def setClockFormat(clockFormat, **kwargs): if ((clockFormat != '12h') and (clockFormat != '24h')): return False _gsession = _GSettings(user=kwargs.get('user'), schema='org.gnome.desktop.interface', key='clock-format') return _gsession._set(clockFormat)
[ "def", "setClockFormat", "(", "clockFormat", ",", "**", "kwargs", ")", ":", "if", "(", "(", "clockFormat", "!=", "'12h'", ")", "and", "(", "clockFormat", "!=", "'24h'", ")", ")", ":", "return", "False", "_gsession", "=", "_GSettings", "(", "user", "=", ...
set the clock format .
train
true
10,326
def wrap_neg_index(index, dim): if ((index is not None) and (index < 0)): index %= dim return index
[ "def", "wrap_neg_index", "(", "index", ",", "dim", ")", ":", "if", "(", "(", "index", "is", "not", "None", ")", "and", "(", "index", "<", "0", ")", ")", ":", "index", "%=", "dim", "return", "index" ]
converts a negative index into a positive index .
train
false
10,327
def s_hex_dump(data, addr=0): dump = slice = '' for byte in data: if ((addr % 16) == 0): dump += ' ' for char in slice: if ((ord(char) >= 32) and (ord(char) <= 126)): dump += char else: dump += '.' dump += ('\n%04x: ' % addr) slice = '' dump += ('%02x ' % ord(byte)) slice += byte addr += 1 remainder = (addr % 16) if (remainder != 0): dump += ((' ' * (16 - remainder)) + ' ') for char in slice: if ((ord(char) >= 32) and (ord(char) <= 126)): dump += char else: dump += '.' return (dump + '\n')
[ "def", "s_hex_dump", "(", "data", ",", "addr", "=", "0", ")", ":", "dump", "=", "slice", "=", "''", "for", "byte", "in", "data", ":", "if", "(", "(", "addr", "%", "16", ")", "==", "0", ")", ":", "dump", "+=", "' '", "for", "char", "in", "slic...
return the hex dump of the supplied data starting at the offset address specified .
train
false
10,328
def get_empty_preference_message(preference_key): return "Preference '{preference_key}' cannot be set to an empty value.".format(preference_key=preference_key)
[ "def", "get_empty_preference_message", "(", "preference_key", ")", ":", "return", "\"Preference '{preference_key}' cannot be set to an empty value.\"", ".", "format", "(", "preference_key", "=", "preference_key", ")" ]
returns the validation message shown for an empty preference .
train
false
10,329
@app.route('/account/<subscription_id>/providers/<provider_namespace>/register', methods=['POST']) @auth.require_login def provider_register_post(subscription_id, provider_namespace): creds = _get_credentials() models.register_provider(subscription_id, creds, provider_namespace) return redirect(url_for('subscription_view', subscription_id=subscription_id))
[ "@", "app", ".", "route", "(", "'/account/<subscription_id>/providers/<provider_namespace>/register'", ",", "methods", "=", "[", "'POST'", "]", ")", "@", "auth", ".", "require_login", "def", "provider_register_post", "(", "subscription_id", ",", "provider_namespace", ")...
register provider request .
train
false
10,332
def build_encoder(tparams, options): opt_ret = dict() trng = RandomStreams(1234) x = tensor.matrix('x', dtype='int64') x_mask = tensor.matrix('x_mask', dtype='float32') n_timesteps = x.shape[0] n_samples = x.shape[1] emb = tparams['Wemb'][x.flatten()].reshape([n_timesteps, n_samples, options['dim_word']]) proj = get_layer(options['encoder'])[1](tparams, emb, None, options, prefix='encoder', mask=x_mask) ctx = proj[0][(-1)] return (trng, x, x_mask, ctx, emb)
[ "def", "build_encoder", "(", "tparams", ",", "options", ")", ":", "opt_ret", "=", "dict", "(", ")", "trng", "=", "RandomStreams", "(", "1234", ")", "x", "=", "tensor", ".", "matrix", "(", "'x'", ",", "dtype", "=", "'int64'", ")", "x_mask", "=", "tens...
computation graph .
train
false
10,334
def _shift_intercept(arr): arr = np.asarray(arr) side = int(np.sqrt(len(arr))) return np.roll(np.roll(arr.reshape(side, side), (-1), axis=1), (-1), axis=0)
[ "def", "_shift_intercept", "(", "arr", ")", ":", "arr", "=", "np", ".", "asarray", "(", "arr", ")", "side", "=", "int", "(", "np", ".", "sqrt", "(", "len", "(", "arr", ")", ")", ")", "return", "np", ".", "roll", "(", "np", ".", "roll", "(", "...
a convenience function to make the sas covariance matrix compatible with statsmodels .
train
false
10,335
def protected(callback=None): def wrapper(f): @functools.wraps(f) def inner(self, request, *args, **kwargs): request.assert_authenticated() if request.context.is_admin: LOG.warning(_LW('RBAC: Bypassing authorization')) elif (callback is not None): prep_info = {'f_name': f.__name__, 'input_attr': kwargs} callback(self, request, prep_info, *args, **kwargs) else: action = ('identity:%s' % f.__name__) creds = _build_policy_check_credentials(self, action, request.context_dict, kwargs) policy_dict = {} if (hasattr(self, 'get_member_from_driver') and (self.get_member_from_driver is not None)): key = ('%s_id' % self.member_name) if (key in kwargs): ref = self.get_member_from_driver(kwargs[key]) policy_dict['target'] = {self.member_name: ref} if (request.context_dict.get('subject_token_id') is not None): window_seconds = self._token_validation_window(request) token_ref = token_model.KeystoneToken(token_id=request.context_dict['subject_token_id'], token_data=self.token_provider_api.validate_token(request.context_dict['subject_token_id'], window_seconds=window_seconds)) policy_dict.setdefault('target', {}) policy_dict['target'].setdefault(self.member_name, {}) policy_dict['target'][self.member_name]['user_id'] = token_ref.user_id try: user_domain_id = token_ref.user_domain_id except exception.UnexpectedError: user_domain_id = None if user_domain_id: policy_dict['target'][self.member_name].setdefault('user', {}) policy_dict['target'][self.member_name]['user'].setdefault('domain', {}) policy_dict['target'][self.member_name]['user']['domain']['id'] = user_domain_id policy_dict.update(kwargs) self.policy_api.enforce(creds, action, utils.flatten_dict(policy_dict)) LOG.debug('RBAC: Authorization granted') return f(self, request, *args, **kwargs) return inner return wrapper
[ "def", "protected", "(", "callback", "=", "None", ")", ":", "def", "wrapper", "(", "f", ")", ":", "@", "functools", ".", "wraps", "(", "f", ")", "def", "inner", "(", "self", ",", "request", ",", "*", "args", ",", "**", "kwargs", ")", ":", "reques...
wraps api calls with role based access controls .
train
false
10,336
def get_plugin_folders(): if (not sys.platform.startswith('win')): return [] plugin_dir = os.path.join(distutils.sysconfig.get_python_lib(), 'PyQt5', 'plugins') folders = ['audio', 'iconengines', 'mediaservice', 'printsupport'] return [os.path.join(plugin_dir, folder) for folder in folders]
[ "def", "get_plugin_folders", "(", ")", ":", "if", "(", "not", "sys", ".", "platform", ".", "startswith", "(", "'win'", ")", ")", ":", "return", "[", "]", "plugin_dir", "=", "os", ".", "path", ".", "join", "(", "distutils", ".", "sysconfig", ".", "get...
get the plugin folders to copy to the output .
train
false
10,338
def _smooth(image, sigma, mode, cval): smoothed = np.empty(image.shape, dtype=np.double) if (image.ndim == 3): for dim in range(image.shape[2]): ndi.gaussian_filter(image[..., dim], sigma, output=smoothed[..., dim], mode=mode, cval=cval) else: ndi.gaussian_filter(image, sigma, output=smoothed, mode=mode, cval=cval) return smoothed
[ "def", "_smooth", "(", "image", ",", "sigma", ",", "mode", ",", "cval", ")", ":", "smoothed", "=", "np", ".", "empty", "(", "image", ".", "shape", ",", "dtype", "=", "np", ".", "double", ")", "if", "(", "image", ".", "ndim", "==", "3", ")", ":"...
return image with each channel smoothed by the gaussian filter .
train
false
10,340
def getPrecision(elementNode): return getCascadeFloatWithoutSelf((0.2 * getLayerHeight(elementNode)), elementNode, 'precision')
[ "def", "getPrecision", "(", "elementNode", ")", ":", "return", "getCascadeFloatWithoutSelf", "(", "(", "0.2", "*", "getLayerHeight", "(", "elementNode", ")", ")", ",", "elementNode", ",", "'precision'", ")" ]
get the cascade precision .
train
false
10,341
def get_spider_list(project, runner=None, pythonpath=None, version=''): if ('cache' not in get_spider_list.__dict__): get_spider_list.cache = UtilsCache() try: return get_spider_list.cache[project][version] except KeyError: pass if (runner is None): runner = Config().get('runner') env = os.environ.copy() env['PYTHONIOENCODING'] = 'UTF-8' env['SCRAPY_PROJECT'] = project if pythonpath: env['PYTHONPATH'] = pythonpath if version: env['SCRAPY_EGG_VERSION'] = version pargs = [sys.executable, '-m', runner, 'list'] proc = Popen(pargs, stdout=PIPE, stderr=PIPE, env=env) (out, err) = proc.communicate() if proc.returncode: msg = (err or out or '') msg = msg.decode('utf8') raise RuntimeError((msg.encode('unicode_escape') if six.PY2 else msg)) tmp = out.decode('utf-8').splitlines() try: project_cache = get_spider_list.cache[project] project_cache[version] = tmp except KeyError: project_cache = {version: tmp} get_spider_list.cache[project] = project_cache return tmp
[ "def", "get_spider_list", "(", "project", ",", "runner", "=", "None", ",", "pythonpath", "=", "None", ",", "version", "=", "''", ")", ":", "if", "(", "'cache'", "not", "in", "get_spider_list", ".", "__dict__", ")", ":", "get_spider_list", ".", "cache", "...
return the spider list from the given project .
train
false
10,342
def iddp_rsvd(eps, m, n, matvect, matvec): (k, iU, iV, iS, w, ier) = _id.iddp_rsvd(eps, m, n, matvect, matvec) if ier: raise _RETCODE_ERROR U = w[(iU - 1):((iU + (m * k)) - 1)].reshape((m, k), order='F') V = w[(iV - 1):((iV + (n * k)) - 1)].reshape((n, k), order='F') S = w[(iS - 1):((iS + k) - 1)] return (U, V, S)
[ "def", "iddp_rsvd", "(", "eps", ",", "m", ",", "n", ",", "matvect", ",", "matvec", ")", ":", "(", "k", ",", "iU", ",", "iV", ",", "iS", ",", "w", ",", "ier", ")", "=", "_id", ".", "iddp_rsvd", "(", "eps", ",", "m", ",", "n", ",", "matvect",...
compute svd of a real matrix to a specified relative precision using random matrix-vector multiplication .
train
false
10,343
def python_zip(file_list, gallery_path, extension='.py'): zipname = gallery_path.replace(os.path.sep, '_') zipname += ('_python' if (extension == '.py') else '_jupyter') zipname = os.path.join(gallery_path, (zipname + '.zip')) zipf = zipfile.ZipFile(zipname, mode='w') for fname in file_list: file_src = (os.path.splitext(fname)[0] + extension) zipf.write(file_src) zipf.close() return zipname
[ "def", "python_zip", "(", "file_list", ",", "gallery_path", ",", "extension", "=", "'.py'", ")", ":", "zipname", "=", "gallery_path", ".", "replace", "(", "os", ".", "path", ".", "sep", ",", "'_'", ")", "zipname", "+=", "(", "'_python'", "if", "(", "ex...
stores all files in file_list into an zip file parameters file_list : list of strings holds all the file names to be included in zip file gallery_path : string path to where the zipfile is stored extension : str .
train
false
10,345
def _parse_server_raw(server): parsed_server = {'addrs': set(), 'ssl': False, 'names': set()} apply_ssl_to_all_addrs = False for directive in server: if (not directive): continue if (directive[0] == 'listen'): addr = obj.Addr.fromstring(directive[1]) parsed_server['addrs'].add(addr) if addr.ssl: parsed_server['ssl'] = True elif (directive[0] == 'server_name'): parsed_server['names'].update(_get_servernames(directive[1])) elif ((directive[0] == 'ssl') and (directive[1] == 'on')): parsed_server['ssl'] = True apply_ssl_to_all_addrs = True if apply_ssl_to_all_addrs: for addr in parsed_server['addrs']: addr.ssl = True return parsed_server
[ "def", "_parse_server_raw", "(", "server", ")", ":", "parsed_server", "=", "{", "'addrs'", ":", "set", "(", ")", ",", "'ssl'", ":", "False", ",", "'names'", ":", "set", "(", ")", "}", "apply_ssl_to_all_addrs", "=", "False", "for", "directive", "in", "ser...
parses a list of server directives .
train
false
10,346
def gps_offset(lat, lon, east, north): bearing = math.degrees(math.atan2(east, north)) distance = math.sqrt(((east ** 2) + (north ** 2))) return gps_newpos(lat, lon, bearing, distance)
[ "def", "gps_offset", "(", "lat", ",", "lon", ",", "east", ",", "north", ")", ":", "bearing", "=", "math", ".", "degrees", "(", "math", ".", "atan2", "(", "east", ",", "north", ")", ")", "distance", "=", "math", ".", "sqrt", "(", "(", "(", "east",...
return new lat/lon after moving east/north by the given number of meters .
train
true
10,348
def _clip_boxes(boxes, im_shape): boxes[:, 0::4] = np.maximum(boxes[:, 0::4], 0) boxes[:, 1::4] = np.maximum(boxes[:, 1::4], 0) boxes[:, 2::4] = np.minimum(boxes[:, 2::4], (im_shape[1] - 1)) boxes[:, 3::4] = np.minimum(boxes[:, 3::4], (im_shape[0] - 1)) return boxes
[ "def", "_clip_boxes", "(", "boxes", ",", "im_shape", ")", ":", "boxes", "[", ":", ",", "0", ":", ":", "4", "]", "=", "np", ".", "maximum", "(", "boxes", "[", ":", ",", "0", ":", ":", "4", "]", ",", "0", ")", "boxes", "[", ":", ",", "1", "...
clip boxes to image boundaries .
train
false
10,349
@blueprint.route('/extension-static/<extension_type>/<extension_id>/<path:filename>') def extension_static(extension_type, extension_id, filename): extension = None if (extension_type == 'view'): extension = extensions.view.get_extension(extension_id) elif (extension_type == 'data'): extension = extensions.data.get_extension(extension_id) if (extension is None): raise ValueError(("Unknown extension '%s'" % extension_id)) digits_root = os.path.dirname(os.path.abspath(digits.__file__)) rootdir = os.path.join(digits_root, *['extensions', 'view', extension.get_dirname(), 'static']) return flask.send_from_directory(rootdir, filename)
[ "@", "blueprint", ".", "route", "(", "'/extension-static/<extension_type>/<extension_id>/<path:filename>'", ")", "def", "extension_static", "(", "extension_type", ",", "extension_id", ",", "filename", ")", ":", "extension", "=", "None", "if", "(", "extension_type", "=="...
returns static files from an extensions static directory .
train
false
10,350
def _find_method_hash(method_hash): for (hashname, asn1code) in HASH_ASN1.items(): if (not method_hash.startswith(asn1code)): continue return (hashname, method_hash[len(asn1code):]) raise VerificationError('Verification failed')
[ "def", "_find_method_hash", "(", "method_hash", ")", ":", "for", "(", "hashname", ",", "asn1code", ")", "in", "HASH_ASN1", ".", "items", "(", ")", ":", "if", "(", "not", "method_hash", ".", "startswith", "(", "asn1code", ")", ")", ":", "continue", "retur...
finds the hash method and the hash itself .
train
false
10,352
def change_SHOWUPDATE_HOUR(freq): sickbeard.SHOWUPDATE_HOUR = try_int(freq, sickbeard.DEFAULT_SHOWUPDATE_HOUR) if (sickbeard.SHOWUPDATE_HOUR > 23): sickbeard.SHOWUPDATE_HOUR = 0 elif (sickbeard.SHOWUPDATE_HOUR < 0): sickbeard.SHOWUPDATE_HOUR = 0 sickbeard.showUpdateScheduler.start_time = datetime.time(hour=sickbeard.SHOWUPDATE_HOUR)
[ "def", "change_SHOWUPDATE_HOUR", "(", "freq", ")", ":", "sickbeard", ".", "SHOWUPDATE_HOUR", "=", "try_int", "(", "freq", ",", "sickbeard", ".", "DEFAULT_SHOWUPDATE_HOUR", ")", "if", "(", "sickbeard", ".", "SHOWUPDATE_HOUR", ">", "23", ")", ":", "sickbeard", "...
change frequency of show updater thread .
train
false
10,353
def show_progress(iteration, loglikelihood): pass
[ "def", "show_progress", "(", "iteration", ",", "loglikelihood", ")", ":", "pass" ]
callback function to be used for training the model .
train
false
10,354
def set_restart_power_failure(enabled): state = salt.utils.mac_utils.validate_enabled(enabled) cmd = 'systemsetup -setrestartpowerfailure {0}'.format(state) salt.utils.mac_utils.execute_return_success(cmd) return salt.utils.mac_utils.confirm_updated(state, get_restart_power_failure)
[ "def", "set_restart_power_failure", "(", "enabled", ")", ":", "state", "=", "salt", ".", "utils", ".", "mac_utils", ".", "validate_enabled", "(", "enabled", ")", "cmd", "=", "'systemsetup -setrestartpowerfailure {0}'", ".", "format", "(", "state", ")", "salt", "...
set whether or not the computer will automatically restart after a power failure .
train
true
10,356
def less_shitty_error_messages(func): @wraps(func) def inner(self, sql, *args, **kwargs): try: return func(self, sql, *args, **kwargs) except Exception as e: exc_info = sys.exc_info() msg = '{}\nSQL: {}'.format(repr(e), sql) six.reraise(exc_info[0], exc_info[0](msg), exc_info[2]) return inner
[ "def", "less_shitty_error_messages", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "inner", "(", "self", ",", "sql", ",", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "return", "func", "(", "self", ",", "sql", ",", "*", ...
wraps functions where the first param is a sql statement and enforces any exceptions thrown will also contain the statement in the message .
train
false
10,357
def cidr_to_netmask(bits): v = ((1 << bits) - 1) v = (v << (32 - bits)) return IPAddr(v, networkOrder=False)
[ "def", "cidr_to_netmask", "(", "bits", ")", ":", "v", "=", "(", "(", "1", "<<", "bits", ")", "-", "1", ")", "v", "=", "(", "v", "<<", "(", "32", "-", "bits", ")", ")", "return", "IPAddr", "(", "v", ",", "networkOrder", "=", "False", ")" ]
takes a number of network bits .
train
false
10,358
def RunDebugModbusFrontend(server, port=8080): from bottle import run application = build_application(server) run(app=application, port=port)
[ "def", "RunDebugModbusFrontend", "(", "server", ",", "port", "=", "8080", ")", ":", "from", "bottle", "import", "run", "application", "=", "build_application", "(", "server", ")", "run", "(", "app", "=", "application", ",", "port", "=", "port", ")" ]
helper method to start the bottle server .
train
false
10,359
def text_to_int(ip): if (':' not in ip): return ipv4_to_int(ip) else: return ipv6_to_int(ip)
[ "def", "text_to_int", "(", "ip", ")", ":", "if", "(", "':'", "not", "in", "ip", ")", ":", "return", "ipv4_to_int", "(", "ip", ")", "else", ":", "return", "ipv6_to_int", "(", "ip", ")" ]
converts human readable ipv4 or ipv6 string to int type representation .
train
false
10,360
def readFeatureFile(font, text, prepend=True): writer = FilterFeatureWriter(set(font.keys())) if prepend: text += font.features.text else: text = (font.features.text + text) parser.parseFeatures(writer, text) font.features.text = writer.write()
[ "def", "readFeatureFile", "(", "font", ",", "text", ",", "prepend", "=", "True", ")", ":", "writer", "=", "FilterFeatureWriter", "(", "set", "(", "font", ".", "keys", "(", ")", ")", ")", "if", "prepend", ":", "text", "+=", "font", ".", "features", "....
incorporate valid definitions from feature text into font .
train
false
10,362
def meminfo(): def linux_meminfo(): '\n linux specific implementation of meminfo\n ' ret = {} try: with salt.utils.fopen('/proc/meminfo', 'r') as fp_: stats = fp_.read() except IOError: pass else: for line in stats.splitlines(): if (not line): continue comps = line.split() comps[0] = comps[0].replace(':', '') ret[comps[0]] = {'value': comps[1]} if (len(comps) > 2): ret[comps[0]]['unit'] = comps[2] return ret def freebsd_meminfo(): '\n freebsd specific implementation of meminfo\n ' sysctlvm = __salt__['cmd.run']('sysctl vm').splitlines() sysctlvm = [x for x in sysctlvm if x.startswith('vm')] sysctlvm = [x.split(':') for x in sysctlvm] sysctlvm = [[y.strip() for y in x] for x in sysctlvm] sysctlvm = [x for x in sysctlvm if x[1]] ret = {} for line in sysctlvm: ret[line[0]] = line[1] sysctlvmtot = __salt__['cmd.run']('sysctl -n vm.vmtotal').splitlines() sysctlvmtot = [x for x in sysctlvmtot if x] ret['vm.vmtotal'] = sysctlvmtot return ret get_version = {'Linux': linux_meminfo, 'FreeBSD': freebsd_meminfo} errmsg = 'This method is unsupported on the current operating system!' return get_version.get(__grains__['kernel'], (lambda : errmsg))()
[ "def", "meminfo", "(", ")", ":", "def", "linux_meminfo", "(", ")", ":", "ret", "=", "{", "}", "try", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "'/proc/meminfo'", ",", "'r'", ")", "as", "fp_", ":", "stats", "=", "fp_", ".", "read", "("...
get memory information from /proc/meminfo .
train
false
10,363
def test_scenario_ru_from_string(): ru = Language('ru') scenario = Scenario.from_string(SCENARIO, language=ru) assert_equals(scenario.name, u'\u0421\u043e\u0445\u0440\u0430\u043d\u0435\u043d\u0438\u0435 \u0431\u0430\u0437\u044b \u043a\u0443\u0440\u0441\u043e\u0432 \u0443\u043d\u0438\u0432\u0435\u0440\u0438\u0442\u0435\u0442\u0430 \u0432 \u0442\u0435\u043a\u0441\u0442\u043e\u0432\u044b\u0439 \u0444\u0430\u0439\u043b') assert_equals(scenario.steps[0].hashes, [{u'\u041d\u0430\u0437\u0432\u0430\u043d\u0438\u0435': u'\u041c\u0430\u0442\u0430\u043d', u'\u0414\u043b\u0438\u0442\u0435\u043b\u044c\u043d\u043e\u0441\u0442\u044c': u'2 \u0433\u043e\u0434\u0430'}, {u'\u041d\u0430\u0437\u0432\u0430\u043d\u0438\u0435': u'\u041e\u0441\u043d\u043e\u0432\u044b \u043f\u0440\u043e\u0433\u0440\u0430\u043c\u043c\u0438\u0440\u043e\u0432\u0430\u043d\u0438\u044f', u'\u0414\u043b\u0438\u0442\u0435\u043b\u044c\u043d\u043e\u0441\u0442\u044c': u'1 \u0433\u043e\u0434'}])
[ "def", "test_scenario_ru_from_string", "(", ")", ":", "ru", "=", "Language", "(", "'ru'", ")", "scenario", "=", "Scenario", ".", "from_string", "(", "SCENARIO", ",", "language", "=", "ru", ")", "assert_equals", "(", "scenario", ".", "name", ",", "u'\\u0421\\...
language: ru -> scenario .
train
false
10,364
def _block_size(sep): return (BLOCK_SIZE + (len(sep) * ((BLOCK_SIZE // Card.length) - 1)))
[ "def", "_block_size", "(", "sep", ")", ":", "return", "(", "BLOCK_SIZE", "+", "(", "len", "(", "sep", ")", "*", "(", "(", "BLOCK_SIZE", "//", "Card", ".", "length", ")", "-", "1", ")", ")", ")" ]
determine the size of a fits header block if a non-blank separator is used between cards .
train
false
10,365
def setup_input(pin, pull_mode): import Adafruit_BBIO.GPIO as GPIO GPIO.setup(pin, GPIO.IN, (GPIO.PUD_DOWN if (pull_mode == 'DOWN') else GPIO.PUD_UP))
[ "def", "setup_input", "(", "pin", ",", "pull_mode", ")", ":", "import", "Adafruit_BBIO", ".", "GPIO", "as", "GPIO", "GPIO", ".", "setup", "(", "pin", ",", "GPIO", ".", "IN", ",", "(", "GPIO", ".", "PUD_DOWN", "if", "(", "pull_mode", "==", "'DOWN'", "...
setup a gpio as input .
train
false
10,366
def Query(sorted=0, **kwarg): cmd = [] keys = kwarg.keys() if sorted: keys.sort() for k in keys: v = kwarg[k] k = k.upper() if ((k in _SIMPLE_BOOL) and v): cmd.append(k) elif (k == 'HEADER'): cmd.extend([k, v[0], ('"%s"' % (v[1],))]) elif (k not in _NO_QUOTES): cmd.extend([k, ('"%s"' % (v,))]) else: cmd.extend([k, ('%s' % (v,))]) if (len(cmd) > 1): return ('(%s)' % ' '.join(cmd)) else: return ' '.join(cmd)
[ "def", "Query", "(", "sorted", "=", "0", ",", "**", "kwarg", ")", ":", "cmd", "=", "[", "]", "keys", "=", "kwarg", ".", "keys", "(", ")", "if", "sorted", ":", "keys", ".", "sort", "(", ")", "for", "k", "in", "keys", ":", "v", "=", "kwarg", ...
create a query string among the accepted keywords are:: all : if set to a true value .
train
false
10,367
@lower_constant(types.EnumMember) def enum_constant(context, builder, ty, pyval): return context.get_constant_generic(builder, ty.dtype, pyval.value)
[ "@", "lower_constant", "(", "types", ".", "EnumMember", ")", "def", "enum_constant", "(", "context", ",", "builder", ",", "ty", ",", "pyval", ")", ":", "return", "context", ".", "get_constant_generic", "(", "builder", ",", "ty", ".", "dtype", ",", "pyval",...
return a llvm constant representing enum member *pyval* .
train
false
10,368
def file_html(models, resources, title=None, template=FILE, template_variables={}): models = _check_models(models) with _ModelInDocument(models): (docs_json, render_items) = _standalone_docs_json_and_render_items(models) title = _title_from_models(models, title) bundle = _bundle_for_objs_and_resources(models, resources) return _html_page_for_render_items(bundle, docs_json, render_items, title=title, template=template, template_variables=template_variables)
[ "def", "file_html", "(", "models", ",", "resources", ",", "title", "=", "None", ",", "template", "=", "FILE", ",", "template_variables", "=", "{", "}", ")", ":", "models", "=", "_check_models", "(", "models", ")", "with", "_ModelInDocument", "(", "models",...
return an html document that embeds bokeh model or document objects .
train
false
10,369
def is_pre_release(version): parsed_version = parse_version(version) return ((parsed_version.pre_release is not None) and (parsed_version.weekly_release is None) and (parsed_version.commit_count is None) and (parsed_version.dirty is None))
[ "def", "is_pre_release", "(", "version", ")", ":", "parsed_version", "=", "parse_version", "(", "version", ")", "return", "(", "(", "parsed_version", ".", "pre_release", "is", "not", "None", ")", "and", "(", "parsed_version", ".", "weekly_release", "is", "None...
return whether the version corresponds to a pre-release .
train
false
10,370
def decipher_elgamal(msg, key): (p, r, d) = key (c1, c2) = msg u = igcdex((c1 ** d), p)[0] return ((u * c2) % p)
[ "def", "decipher_elgamal", "(", "msg", ",", "key", ")", ":", "(", "p", ",", "r", ",", "d", ")", "=", "key", "(", "c1", ",", "c2", ")", "=", "msg", "u", "=", "igcdex", "(", "(", "c1", "**", "d", ")", ",", "p", ")", "[", "0", "]", "return",...
decrypt message with private key msg = key = according to extended eucliden theorem .
train
false
10,371
def exception_data(func, *args, **kwds): try: func(*args, **kwds) except Exception as detail: return (detail, detail.args, ('%s: %s' % (detail.__class__.__name__, detail)))
[ "def", "exception_data", "(", "func", ",", "*", "args", ",", "**", "kwds", ")", ":", "try", ":", "func", "(", "*", "args", ",", "**", "kwds", ")", "except", "Exception", "as", "detail", ":", "return", "(", "detail", ",", "detail", ".", "args", ",",...
execute func and return the resulting exception .
train
false
10,372
def _ImageProcessing(image_buffer, shape): image = tf.image.decode_png(image_buffer, channels=shape.depth) image.set_shape([shape.height, shape.width, shape.depth]) image = tf.cast(image, tf.float32) image = tf.sub(image, 128.0) image = tf.mul(image, (1 / 100.0)) return image
[ "def", "_ImageProcessing", "(", "image_buffer", ",", "shape", ")", ":", "image", "=", "tf", ".", "image", ".", "decode_png", "(", "image_buffer", ",", "channels", "=", "shape", ".", "depth", ")", "image", ".", "set_shape", "(", "[", "shape", ".", "height...
convert a png string into an input tensor .
train
false
10,373
def test_import_error_in_warning_logging(): class FakeModule(object, ): def __getattr__(self, attr): raise ImportError(u'_showwarning should ignore any exceptions here') log.enable_warnings_logging() sys.modules[u'<test fake module>'] = FakeModule() try: warnings.showwarning(AstropyWarning(u'Regression test for #2671'), AstropyWarning, u'<this is only a test>', 1) finally: del sys.modules[u'<test fake module>']
[ "def", "test_import_error_in_warning_logging", "(", ")", ":", "class", "FakeModule", "(", "object", ",", ")", ":", "def", "__getattr__", "(", "self", ",", "attr", ")", ":", "raise", "ImportError", "(", "u'_showwarning should ignore any exceptions here'", ")", "log",...
regression test for URL this test actually puts a goofy fake module into sys .
train
false
10,374
def _generate_states_report(sorted_data): states = [] for (state, data) in sorted_data: (module, stateid, name, function) = [x.rstrip('_').lstrip('-') for x in state.split('|')] module_function = '.'.join((module, function)) result = data.get('result', '') single = [{'function': module_function}, {'name': name}, {'result': result}, {'duration': data.get('duration', 0.0)}, {'comment': data.get('comment', '')}] if (not result): style = 'failed' else: changes = data.get('changes', {}) if (changes and isinstance(changes, dict)): single.append({'changes': _dict_to_name_value(changes)}) style = 'changed' else: style = 'unchanged' started = data.get('start_time', '') if started: single.append({'started': started}) states.append({stateid: single, '__style__': style}) return states
[ "def", "_generate_states_report", "(", "sorted_data", ")", ":", "states", "=", "[", "]", "for", "(", "state", ",", "data", ")", "in", "sorted_data", ":", "(", "module", ",", "stateid", ",", "name", ",", "function", ")", "=", "[", "x", ".", "rstrip", ...
generate states report .
train
true
10,376
def safe_enumerate(iterable): for (i, v) in enumerate(iterable): if (v is not None): (yield (i, v))
[ "def", "safe_enumerate", "(", "iterable", ")", ":", "for", "(", "i", ",", "v", ")", "in", "enumerate", "(", "iterable", ")", ":", "if", "(", "v", "is", "not", "None", ")", ":", "(", "yield", "(", "i", ",", "v", ")", ")" ]
enumerate which does not yield none values .
train
false
10,377
def build_initiator_target_map(connector, target_wwns, lookup_service): init_targ_map = {} initiator_wwns = connector['wwpns'] if lookup_service: dev_map = lookup_service.get_device_mapping_from_network(initiator_wwns, target_wwns) for fabric_name in dev_map: fabric = dev_map[fabric_name] for initiator in fabric['initiator_port_wwn_list']: init_targ_map[initiator] = fabric['target_port_wwn_list'] else: for initiator in initiator_wwns: init_targ_map[initiator] = target_wwns return init_targ_map
[ "def", "build_initiator_target_map", "(", "connector", ",", "target_wwns", ",", "lookup_service", ")", ":", "init_targ_map", "=", "{", "}", "initiator_wwns", "=", "connector", "[", "'wwpns'", "]", "if", "lookup_service", ":", "dev_map", "=", "lookup_service", ".",...
return a dictionary mapping server-wwns and lists of storage-wwns .
train
false
10,380
@register.inclusion_tag('addons/includes/collection_add_widget.html') @register.function @jinja2.contextfunction def collection_add_widget(context, addon, condensed=False): c = dict(context.items()) c.update(locals()) return c
[ "@", "register", ".", "inclusion_tag", "(", "'addons/includes/collection_add_widget.html'", ")", "@", "register", ".", "function", "@", "jinja2", ".", "contextfunction", "def", "collection_add_widget", "(", "context", ",", "addon", ",", "condensed", "=", "False", ")...
displays add to collection widget .
train
false
10,381
def element_to_extension_element(element): exel = ExtensionElement(element.c_tag, element.c_namespace, text=element.text) exel.attributes.update(element.extension_attributes) exel.children.extend(element.extension_elements) for (xml_attribute, (member_name, typ, req)) in element.c_attributes.iteritems(): member_value = getattr(element, member_name) if (member_value is not None): exel.attributes[xml_attribute] = member_value exel.children.extend([element_to_extension_element(c) for c in element.children_with_values()]) return exel
[ "def", "element_to_extension_element", "(", "element", ")", ":", "exel", "=", "ExtensionElement", "(", "element", ".", "c_tag", ",", "element", ".", "c_namespace", ",", "text", "=", "element", ".", "text", ")", "exel", ".", "attributes", ".", "update", "(", ...
convert an element into a extension element .
train
true
10,382
def _force_update_info(info_base, info_target): exclude_keys = ['chs', 'ch_names', 'nchan'] info_target = np.atleast_1d(info_target).ravel() all_infos = np.hstack([info_base, info_target]) for ii in all_infos: if (not isinstance(ii, Info)): raise ValueError(('Inputs must be of type Info. Found type %s' % type(ii))) for (key, val) in info_base.items(): if (key in exclude_keys): continue for i_targ in info_target: i_targ[key] = val
[ "def", "_force_update_info", "(", "info_base", ",", "info_target", ")", ":", "exclude_keys", "=", "[", "'chs'", ",", "'ch_names'", ",", "'nchan'", "]", "info_target", "=", "np", ".", "atleast_1d", "(", "info_target", ")", ".", "ravel", "(", ")", "all_infos",...
update target info objects with values from info base .
train
false
10,384
def timestamp_local(value): try: return dt_util.as_local(dt_util.utc_from_timestamp(value)).strftime(DATE_STR_FORMAT) except (ValueError, TypeError): return value
[ "def", "timestamp_local", "(", "value", ")", ":", "try", ":", "return", "dt_util", ".", "as_local", "(", "dt_util", ".", "utc_from_timestamp", "(", "value", ")", ")", ".", "strftime", "(", "DATE_STR_FORMAT", ")", "except", "(", "ValueError", ",", "TypeError"...
filter to convert given timestamp to local date/time .
train
false
10,385
def build_reverse_dictionary(word_to_id): reverse_dictionary = dict(zip(word_to_id.values(), word_to_id.keys())) return reverse_dictionary
[ "def", "build_reverse_dictionary", "(", "word_to_id", ")", ":", "reverse_dictionary", "=", "dict", "(", "zip", "(", "word_to_id", ".", "values", "(", ")", ",", "word_to_id", ".", "keys", "(", ")", ")", ")", "return", "reverse_dictionary" ]
given a dictionary for converting word to integer id .
train
true
10,387
def zy_basis_transform(self, format='sympy'): return matrix_cache.get_matrix('ZY', format)
[ "def", "zy_basis_transform", "(", "self", ",", "format", "=", "'sympy'", ")", ":", "return", "matrix_cache", ".", "get_matrix", "(", "'ZY'", ",", "format", ")" ]
transformation matrix from z to y basis .
train
false
10,388
@register.simple_tag def bootstrap_button(*args, **kwargs): return render_button(*args, **kwargs)
[ "@", "register", ".", "simple_tag", "def", "bootstrap_button", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "render_button", "(", "*", "args", ",", "**", "kwargs", ")" ]
render a button **tag name**:: bootstrap_button **parameters**: content the text to be displayed in the button button_type optional field defining what type of button this is .
train
false
10,389
def require_version(version): def check_require_version(f): version_elements = version.split('.') try: compare = tuple([int(v) for v in version_elements]) except ValueError: raise ValueError(('%s is not a correct version : should be X.Y[.Z].' % version)) current = sys.version_info[:3] if (current < compare): def new_f(self, *args, **kwargs): self.skipTest(('Need at least %s version of python. Current version is %s.' % (version, '.'.join([str(element) for element in current])))) new_f.__name__ = f.__name__ return new_f else: return f return check_require_version
[ "def", "require_version", "(", "version", ")", ":", "def", "check_require_version", "(", "f", ")", ":", "version_elements", "=", "version", ".", "split", "(", "'.'", ")", "try", ":", "compare", "=", "tuple", "(", "[", "int", "(", "v", ")", "for", "v", ...
compare version of python interpreter to the given one .
train
false
10,390
def _set_up_ready_watcher(): ready_watcher = UserFactory(email='ready@example.com') ReadyRevisionEvent.notify(ready_watcher) return ready_watcher
[ "def", "_set_up_ready_watcher", "(", ")", ":", "ready_watcher", "=", "UserFactory", "(", "email", "=", "'ready@example.com'", ")", "ReadyRevisionEvent", ".", "notify", "(", "ready_watcher", ")", "return", "ready_watcher" ]
make a user who watches for revision readiness .
train
false
10,392
def distance_to_current_waypoint(): nextwaypoint = vehicle.commands.next if (nextwaypoint == 0): return None missionitem = vehicle.commands[(nextwaypoint - 1)] lat = missionitem.x lon = missionitem.y alt = missionitem.z targetWaypointLocation = LocationGlobalRelative(lat, lon, alt) distancetopoint = get_distance_metres(vehicle.location.global_frame, targetWaypointLocation) return distancetopoint
[ "def", "distance_to_current_waypoint", "(", ")", ":", "nextwaypoint", "=", "vehicle", ".", "commands", ".", "next", "if", "(", "nextwaypoint", "==", "0", ")", ":", "return", "None", "missionitem", "=", "vehicle", ".", "commands", "[", "(", "nextwaypoint", "-...
gets distance in metres to the current waypoint .
train
true
10,393
def processChildNodesByIndexValue(elementNode, function, index, indexValue, value): if (indexValue.indexName != ''): function.localDictionary[indexValue.indexName] = index if (indexValue.valueName != ''): function.localDictionary[indexValue.valueName] = value function.processChildNodes(elementNode)
[ "def", "processChildNodesByIndexValue", "(", "elementNode", ",", "function", ",", "index", ",", "indexValue", ",", "value", ")", ":", "if", "(", "indexValue", ".", "indexName", "!=", "''", ")", ":", "function", ".", "localDictionary", "[", "indexValue", ".", ...
process childnodes by index value .
train
false
10,394
def jwt_get_user_id_from_payload_handler(payload): warnings.warn('The following will be removed in the future. Use `JWT_PAYLOAD_GET_USERNAME_HANDLER` instead.', DeprecationWarning) return payload.get('user_id')
[ "def", "jwt_get_user_id_from_payload_handler", "(", "payload", ")", ":", "warnings", ".", "warn", "(", "'The following will be removed in the future. Use `JWT_PAYLOAD_GET_USERNAME_HANDLER` instead.'", ",", "DeprecationWarning", ")", "return", "payload", ".", "get", "(", "'user_...
override this function if user_id is formatted differently in payload .
train
false
10,395
def parse_newick(lines, constructor=PhyloNode): return DndParser(lines, constructor=constructor, unescape_name=True)
[ "def", "parse_newick", "(", "lines", ",", "constructor", "=", "PhyloNode", ")", ":", "return", "DndParser", "(", "lines", ",", "constructor", "=", "constructor", ",", "unescape_name", "=", "True", ")" ]
return phylonode from newick file handle stripping quotes from tip names this function wraps cogent .
train
false
10,396
def _cmp_by_highest_wg(path1, path2): return None
[ "def", "_cmp_by_highest_wg", "(", "path1", ",", "path2", ")", ":", "return", "None" ]
selects a path with highest weight .
train
false
10,397
@mock_ec2 def test_igw_delete_attached(): conn = boto.connect_vpc(u'the_key', u'the_secret') igw = conn.create_internet_gateway() vpc = conn.create_vpc(VPC_CIDR) conn.attach_internet_gateway(igw.id, vpc.id) with assert_raises(EC2ResponseError) as cm: conn.delete_internet_gateway(igw.id) cm.exception.code.should.equal(u'DependencyViolation') cm.exception.status.should.equal(400) cm.exception.request_id.should_not.be.none
[ "@", "mock_ec2", "def", "test_igw_delete_attached", "(", ")", ":", "conn", "=", "boto", ".", "connect_vpc", "(", "u'the_key'", ",", "u'the_secret'", ")", "igw", "=", "conn", ".", "create_internet_gateway", "(", ")", "vpc", "=", "conn", ".", "create_vpc", "("...
internet gateway fail to delete attached .
train
false
10,398
def _get_skiprows(skiprows): if isinstance(skiprows, slice): return lrange((skiprows.start or 0), skiprows.stop, (skiprows.step or 1)) elif (isinstance(skiprows, numbers.Integral) or is_list_like(skiprows)): return skiprows elif (skiprows is None): return 0 raise TypeError(('%r is not a valid type for skipping rows' % type(skiprows).__name__))
[ "def", "_get_skiprows", "(", "skiprows", ")", ":", "if", "isinstance", "(", "skiprows", ",", "slice", ")", ":", "return", "lrange", "(", "(", "skiprows", ".", "start", "or", "0", ")", ",", "skiprows", ".", "stop", ",", "(", "skiprows", ".", "step", "...
get an iterator given an integer .
train
true
10,399
def parsemsg(s): prefix = '' trailing = [] if (not s): raise IRCBadMessage('Empty line.') if (s[0] == ':'): (prefix, s) = s[1:].split(' ', 1) if (s.find(' :') != (-1)): (s, trailing) = s.split(' :', 1) args = s.split() args.append(trailing) else: args = s.split() command = args.pop(0) return (prefix, command, args)
[ "def", "parsemsg", "(", "s", ")", ":", "prefix", "=", "''", "trailing", "=", "[", "]", "if", "(", "not", "s", ")", ":", "raise", "IRCBadMessage", "(", "'Empty line.'", ")", "if", "(", "s", "[", "0", "]", "==", "':'", ")", ":", "(", "prefix", ",...
breaks a message from an irc server into its prefix .
train
true
10,400
def find_first_level_groups(string, enclosing, blank_sep=None): groups = find_first_level_groups_span(string, enclosing) if blank_sep: for (start, end) in groups: string = str_replace(string, start, blank_sep) string = str_replace(string, (end - 1), blank_sep) return split_on_groups(string, groups)
[ "def", "find_first_level_groups", "(", "string", ",", "enclosing", ",", "blank_sep", "=", "None", ")", ":", "groups", "=", "find_first_level_groups_span", "(", "string", ",", "enclosing", ")", "if", "blank_sep", ":", "for", "(", "start", ",", "end", ")", "in...
return a list of groups that could be split because of explicit grouping .
train
false
10,401
def format_html_pre_18(format_string, *args, **kwargs): args_safe = map(conditional_escape, args) kwargs_safe = {} for k in kwargs: kwargs_safe[k] = conditional_escape(kwargs[k]) return mark_safe(format_string.format(*args_safe, **kwargs_safe))
[ "def", "format_html_pre_18", "(", "format_string", ",", "*", "args", ",", "**", "kwargs", ")", ":", "args_safe", "=", "map", "(", "conditional_escape", ",", "args", ")", "kwargs_safe", "=", "{", "}", "for", "k", "in", "kwargs", ":", "kwargs_safe", "[", "...
fake function to support format_html in django < 1 .
train
false
10,402
def listToStrValue(value): if isinstance(value, (set, tuple)): value = list(value) if isinstance(value, list): retVal = value.__str__().lstrip('[').rstrip(']') else: retVal = value return retVal
[ "def", "listToStrValue", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "(", "set", ",", "tuple", ")", ")", ":", "value", "=", "list", "(", "value", ")", "if", "isinstance", "(", "value", ",", "list", ")", ":", "retVal", "=", "value...
flattens list to a string value .
train
false
10,403
def extract_angular(fileobj, keywords, comment_tags, options): parser = AngularGettextHTMLParser() for line in fileobj: parser.feed(line) for string in parser.strings: (yield string)
[ "def", "extract_angular", "(", "fileobj", ",", "keywords", ",", "comment_tags", ",", "options", ")", ":", "parser", "=", "AngularGettextHTMLParser", "(", ")", "for", "line", "in", "fileobj", ":", "parser", ".", "feed", "(", "line", ")", "for", "string", "i...
extract messages from angular template files that use the angular-gettext translate directive as per URL .
train
true
10,405
def _create_course(test, course_key, course_data): course_url = get_url('course_handler', course_key, 'course_key_string') response = test.client.ajax_post(course_url, course_data) test.assertEqual(response.status_code, 200) data = parse_json(response) test.assertNotIn('ErrMsg', data) test.assertEqual(data['url'], course_url)
[ "def", "_create_course", "(", "test", ",", "course_key", ",", "course_data", ")", ":", "course_url", "=", "get_url", "(", "'course_handler'", ",", "course_key", ",", "'course_key_string'", ")", "response", "=", "test", ".", "client", ".", "ajax_post", "(", "co...
creates a course via an ajax request and verifies the url returned in the response .
train
false
10,407
@handle_response_format @treeio_login_required def widget_my_watchlist(request, response_format='html'): profile = request.user.profile query = ((_get_filter_query(profile, do_recipients=False) & Q(about__in=profile.subscriptions.all())) & (~ Q(author=profile))) updates = UpdateRecord.objects.filter(query).distinct() context = _get_default_context(request) context.update({'updates': updates, 'profile': profile}) return render_to_response('news/widgets/my_watchlist', context, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "def", "widget_my_watchlist", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "profile", "=", "request", ".", "user", ".", "profile", "query", "=", "(", "(", "_get_filter_query", "(", ...
displays news about all objects a user is subscribed to .
train
false
10,408
def document_generator(dir_path_pattern, count=None): for (running_count, item) in enumerate(glob.iglob(dir_path_pattern)): if (count and (running_count >= count)): raise StopIteration() doc_id = os.path.basename(item) with codecs.open(item, encoding='utf-8') as f: try: text = f.read() except UnicodeDecodeError: continue (yield Document(text, doc_id, item))
[ "def", "document_generator", "(", "dir_path_pattern", ",", "count", "=", "None", ")", ":", "for", "(", "running_count", ",", "item", ")", "in", "enumerate", "(", "glob", ".", "iglob", "(", "dir_path_pattern", ")", ")", ":", "if", "(", "count", "and", "("...
generator for the input movie documents .
train
false
10,410
def invalidate_pricing_cache(): PRICING_DATA['compute'] = {} PRICING_DATA['storage'] = {}
[ "def", "invalidate_pricing_cache", "(", ")", ":", "PRICING_DATA", "[", "'compute'", "]", "=", "{", "}", "PRICING_DATA", "[", "'storage'", "]", "=", "{", "}" ]
invalidate pricing cache for all the drivers .
train
false
10,411
def proj_transform(xs, ys, zs, M): vec = vec_pad_ones(xs, ys, zs) return proj_transform_vec(vec, M)
[ "def", "proj_transform", "(", "xs", ",", "ys", ",", "zs", ",", "M", ")", ":", "vec", "=", "vec_pad_ones", "(", "xs", ",", "ys", ",", "zs", ")", "return", "proj_transform_vec", "(", "vec", ",", "M", ")" ]
transform the points by the projection matrix .
train
false
10,412
def privacy_info_handle(info, anonymous, name=False): if anonymous: return ('A user' if name else '') return info
[ "def", "privacy_info_handle", "(", "info", ",", "anonymous", ",", "name", "=", "False", ")", ":", "if", "anonymous", ":", "return", "(", "'A user'", "if", "name", "else", "''", ")", "return", "info" ]
hide user info from api if anonymous .
train
false
10,413
def title_from_content(content): for end in (u'. ', u'?', u'!', u'<br />', u'\n', u'</p>'): if (end in content): content = (content.split(end)[0] + end) break return strip_tags(content)
[ "def", "title_from_content", "(", "content", ")", ":", "for", "end", "in", "(", "u'. '", ",", "u'?'", ",", "u'!'", ",", "u'<br />'", ",", "u'\\n'", ",", "u'</p>'", ")", ":", "if", "(", "end", "in", "content", ")", ":", "content", "=", "(", "content",...
try and extract the first sentence from a block of test to use as a title .
train
false
10,414
def geom_output(func, argtypes, offset=None): func.argtypes = argtypes if (not offset): func.restype = c_void_p func.errcheck = check_geom else: func.restype = c_int def geomerrcheck(result, func, cargs): return check_geom_offset(result, func, cargs, offset) func.errcheck = geomerrcheck return func
[ "def", "geom_output", "(", "func", ",", "argtypes", ",", "offset", "=", "None", ")", ":", "func", ".", "argtypes", "=", "argtypes", "if", "(", "not", "offset", ")", ":", "func", ".", "restype", "=", "c_void_p", "func", ".", "errcheck", "=", "check_geom...
for geos routines that return a geometry .
train
false
10,416
def findtemplate(template=None): if (MacOS.runtimemodel == 'macho'): return None if (not template): template = TEMPLATE for p in sys.path: file = os.path.join(p, template) try: (file, d1, d2) = Carbon.File.FSResolveAliasFile(file, 1) break except (Carbon.File.Error, ValueError): continue else: raise BuildError, ('Template %r not found on sys.path' % (template,)) file = file.as_pathname() return file
[ "def", "findtemplate", "(", "template", "=", "None", ")", ":", "if", "(", "MacOS", ".", "runtimemodel", "==", "'macho'", ")", ":", "return", "None", "if", "(", "not", "template", ")", ":", "template", "=", "TEMPLATE", "for", "p", "in", "sys", ".", "p...
locate the applet template along sys .
train
false
10,417
@aggregate(_(u'Sum of')) def aggregate_sum(items, col): return sum((getattr(item, col) for item in items))
[ "@", "aggregate", "(", "_", "(", "u'Sum of'", ")", ")", "def", "aggregate_sum", "(", "items", ",", "col", ")", ":", "return", "sum", "(", "(", "getattr", "(", "item", ",", "col", ")", "for", "item", "in", "items", ")", ")" ]
function to use on group by charts .
train
false
10,418
def test_ast_good_global(): can_compile(u'(global a)') can_compile(u'(global foo bar)')
[ "def", "test_ast_good_global", "(", ")", ":", "can_compile", "(", "u'(global a)'", ")", "can_compile", "(", "u'(global foo bar)'", ")" ]
make sure ast can compile valid global .
train
false
10,419
def _get_blkno_placements(blknos, blk_count, group=True): blknos = _ensure_int64(blknos) for (blkno, indexer) in lib.get_blkno_indexers(blknos, group): (yield (blkno, BlockPlacement(indexer)))
[ "def", "_get_blkno_placements", "(", "blknos", ",", "blk_count", ",", "group", "=", "True", ")", ":", "blknos", "=", "_ensure_int64", "(", "blknos", ")", "for", "(", "blkno", ",", "indexer", ")", "in", "lib", ".", "get_blkno_indexers", "(", "blknos", ",", ...
parameters blknos : array of int64 blk_count : int group : bool returns iterator yield .
train
false
10,420
@transaction.non_atomic_requests @ensure_csrf_cookie @cache_control(no_cache=True, no_store=True, must_revalidate=True) @require_global_staff @require_POST def generate_certificate_exceptions(request, course_id, generate_for=None): course_key = CourseKey.from_string(course_id) if (generate_for == 'all'): students = 'all_whitelisted' elif (generate_for == 'new'): students = 'whitelisted_not_generated' else: return JsonResponse({'success': False, 'message': _('Invalid data, generate_for must be "new" or "all".')}, status=400) lms.djangoapps.instructor_task.api.generate_certificates_for_students(request, course_key, student_set=students) response_payload = {'success': True, 'message': _('Certificate generation started for white listed students.')} return JsonResponse(response_payload)
[ "@", "transaction", ".", "non_atomic_requests", "@", "ensure_csrf_cookie", "@", "cache_control", "(", "no_cache", "=", "True", ",", "no_store", "=", "True", ",", "must_revalidate", "=", "True", ")", "@", "require_global_staff", "@", "require_POST", "def", "generat...
generate certificate for students in the certificate white list .
train
false
10,421
def platforms_to_releases(info, debug): output = [] temp_releases = {} platforms = info.get('platforms') for platform in platforms: for release in platforms[platform]: key = ('%s-%s' % (release['version'], release['url'])) if (key not in temp_releases): temp_releases[key] = {'sublime_text': '<3000', 'version': release['version'], 'date': info.get('last_modified', '2011-08-01 00:00:00'), 'url': update_url(release['url'], debug), 'platforms': []} if (platform == '*'): temp_releases[key]['platforms'] = ['*'] elif (temp_releases[key]['platforms'] != ['*']): temp_releases[key]['platforms'].append(platform) for key in temp_releases: release = temp_releases[key] if (release['platforms'] == ['windows', 'linux', 'osx']): release['platforms'] = ['*'] output.append(release) return output
[ "def", "platforms_to_releases", "(", "info", ",", "debug", ")", ":", "output", "=", "[", "]", "temp_releases", "=", "{", "}", "platforms", "=", "info", ".", "get", "(", "'platforms'", ")", "for", "platform", "in", "platforms", ":", "for", "release", "in"...
accepts a dict from a schema version 1 .
train
false
10,423
def do_get_relationship(parser, token): bits = token.contents.split() if (len(bits) == 3): return GetRelationship(bits[1], bits[2]) if (len(bits) == 5): return GetRelationship(bits[1], bits[2], bits[4]) if (len(bits) == 4): raise template.TemplateSyntaxError, ("The tag '%s' needs an 'as' as its third argument." % bits[0]) if (len(bits) < 3): raise template.TemplateSyntaxError, ("The tag '%s' takes two arguments" % bits[0])
[ "def", "do_get_relationship", "(", "parser", ",", "token", ")", ":", "bits", "=", "token", ".", "contents", ".", "split", "(", ")", "if", "(", "len", "(", "bits", ")", "==", "3", ")", ":", "return", "GetRelationship", "(", "bits", "[", "1", "]", ",...
get relationship between two users .
train
false
10,426
def keep_awake(): if (KERNEL32 or sleepless): if sabnzbd.cfg.keep_awake(): awake = False if (not sabnzbd.downloader.Downloader.do.paused): if ((not PostProcessor.do.empty()) or (not NzbQueue.do.is_empty())): awake = True if KERNEL32: KERNEL32.SetThreadExecutionState(ctypes.c_int(1)) else: sleepless.keep_awake(u'SABnzbd is busy downloading and/or post-processing') if ((not awake) and sleepless): sleepless.allow_sleep()
[ "def", "keep_awake", "(", ")", ":", "if", "(", "KERNEL32", "or", "sleepless", ")", ":", "if", "sabnzbd", ".", "cfg", ".", "keep_awake", "(", ")", ":", "awake", "=", "False", "if", "(", "not", "sabnzbd", ".", "downloader", ".", "Downloader", ".", "do"...
if we still have work to do .
train
false
10,427
def mc_sample_path(P, init=0, sample_size=1000, random_state=None): random_state = check_random_state(random_state) if isinstance(init, numbers.Integral): X_0 = init else: cdf0 = np.cumsum(init) u_0 = random_state.random_sample() X_0 = searchsorted(cdf0, u_0) mc = MarkovChain(P) return mc.simulate(ts_length=sample_size, init=X_0, random_state=random_state)
[ "def", "mc_sample_path", "(", "P", ",", "init", "=", "0", ",", "sample_size", "=", "1000", ",", "random_state", "=", "None", ")", ":", "random_state", "=", "check_random_state", "(", "random_state", ")", "if", "isinstance", "(", "init", ",", "numbers", "."...
generates one sample path from the markov chain represented by transition matrix p on state space s = {{0 .
train
true
10,428
def _ensure_no_filesystem(device, block_device_manager): if block_device_manager.has_filesystem(device): raise FilesystemExists(device)
[ "def", "_ensure_no_filesystem", "(", "device", ",", "block_device_manager", ")", ":", "if", "block_device_manager", ".", "has_filesystem", "(", "device", ")", ":", "raise", "FilesystemExists", "(", "device", ")" ]
raises an error if theres already a filesystem on device .
train
false
10,429
def for_unsigned_dtypes(name='dtype'): return for_dtypes(_unsigned_dtypes, name=name)
[ "def", "for_unsigned_dtypes", "(", "name", "=", "'dtype'", ")", ":", "return", "for_dtypes", "(", "_unsigned_dtypes", ",", "name", "=", "name", ")" ]
decorator that checks the fixture with all dtypes .
train
false
10,431
def path_from_name(name, ext=None, sep=u'|'): if ext: return fsencode((name.replace(os.sep, sep) + ext)) else: return fsencode(name.replace(os.sep, sep))
[ "def", "path_from_name", "(", "name", ",", "ext", "=", "None", ",", "sep", "=", "u'|'", ")", ":", "if", "ext", ":", "return", "fsencode", "(", "(", "name", ".", "replace", "(", "os", ".", "sep", ",", "sep", ")", "+", "ext", ")", ")", "else", ":...
convert name with optional extension to file path .
train
false
10,432
def xblock_studio_url(xblock, parent_xblock=None): if (not xblock_has_own_studio_page(xblock, parent_xblock)): return None category = xblock.category if (category == 'course'): return reverse_course_url('course_handler', xblock.location.course_key) elif (category in ('chapter', 'sequential')): return u'{url}?show={usage_key}'.format(url=reverse_course_url('course_handler', xblock.location.course_key), usage_key=urllib.quote(unicode(xblock.location))) elif (category == 'library'): library_key = xblock.location.course_key return reverse_library_url('library_handler', library_key) else: return reverse_usage_url('container_handler', xblock.location)
[ "def", "xblock_studio_url", "(", "xblock", ",", "parent_xblock", "=", "None", ")", ":", "if", "(", "not", "xblock_has_own_studio_page", "(", "xblock", ",", "parent_xblock", ")", ")", ":", "return", "None", "category", "=", "xblock", ".", "category", "if", "(...
returns the studio editing url for the specified xblock .
train
false
10,433
def ensure_soup(value, parser=None): if isinstance(value, BeautifulSoup): return value.find() if isinstance(value, Tag): return value if isinstance(value, list): return [ensure_soup(item, parser=parser) for item in value] parsed = BeautifulSoup(value, features=parser) return parsed.find()
[ "def", "ensure_soup", "(", "value", ",", "parser", "=", "None", ")", ":", "if", "isinstance", "(", "value", ",", "BeautifulSoup", ")", ":", "return", "value", ".", "find", "(", ")", "if", "isinstance", "(", "value", ",", "Tag", ")", ":", "return", "v...
coerce a value to tag .
train
true
10,436
def prism(): rc('image', cmap='prism') im = gci() if (im is not None): im.set_cmap(cm.prism) draw_if_interactive()
[ "def", "prism", "(", ")", ":", "rc", "(", "'image'", ",", "cmap", "=", "'prism'", ")", "im", "=", "gci", "(", ")", "if", "(", "im", "is", "not", "None", ")", ":", "im", ".", "set_cmap", "(", "cm", ".", "prism", ")", "draw_if_interactive", "(", ...
set the default colormap to prism and apply to current image if any .
train
false
10,437
def test_wrong_nn(): kind = 'borderline1' nn_m = 'rnd' nn_k = NearestNeighbors(n_neighbors=6) smote = SMOTE(random_state=RND_SEED, kind=kind, k_neighbors=nn_k, m_neighbors=nn_m) assert_raises(ValueError, smote.fit_sample, X, Y) nn_k = 'rnd' nn_m = NearestNeighbors(n_neighbors=10) smote = SMOTE(random_state=RND_SEED, kind=kind, k_neighbors=nn_k, m_neighbors=nn_m) assert_raises(ValueError, smote.fit_sample, X, Y) kind = 'regular' nn_k = 'rnd' smote = SMOTE(random_state=RND_SEED, kind=kind, k_neighbors=nn_k) assert_raises(ValueError, smote.fit_sample, X, Y)
[ "def", "test_wrong_nn", "(", ")", ":", "kind", "=", "'borderline1'", "nn_m", "=", "'rnd'", "nn_k", "=", "NearestNeighbors", "(", "n_neighbors", "=", "6", ")", "smote", "=", "SMOTE", "(", "random_state", "=", "RND_SEED", ",", "kind", "=", "kind", ",", "k_...
test either if an error is raised while passing a wrong nn object .
train
false
10,440
def cxBlend(ind1, ind2, alpha): for (i, (x1, x2)) in enumerate(zip(ind1, ind2)): gamma = (((1.0 + (2.0 * alpha)) * random.random()) - alpha) ind1[i] = (((1.0 - gamma) * x1) + (gamma * x2)) ind2[i] = ((gamma * x1) + ((1.0 - gamma) * x2)) return (ind1, ind2)
[ "def", "cxBlend", "(", "ind1", ",", "ind2", ",", "alpha", ")", ":", "for", "(", "i", ",", "(", "x1", ",", "x2", ")", ")", "in", "enumerate", "(", "zip", "(", "ind1", ",", "ind2", ")", ")", ":", "gamma", "=", "(", "(", "(", "1.0", "+", "(", ...
executes a blend crossover that modify in-place the input individuals .
train
false
10,443
def nlinks_file(path): if iswindows: return windows_nlinks(path) return os.stat(path).st_nlink
[ "def", "nlinks_file", "(", "path", ")", ":", "if", "iswindows", ":", "return", "windows_nlinks", "(", "path", ")", "return", "os", ".", "stat", "(", "path", ")", ".", "st_nlink" ]
return number of hardlinks to the file .
train
false
10,444
def school(): output = s3_rest_controller() return output
[ "def", "school", "(", ")", ":", "output", "=", "s3_rest_controller", "(", ")", "return", "output" ]
restful crud controller .
train
false
10,446
def get_units_with_due_date(course): units = [] def visit(node): '\n Visit a node. Checks to see if node has a due date and appends to\n `units` if it does. Otherwise recurses into children to search for\n nodes with due dates.\n ' if getattr(node, 'due', None): units.append(node) else: for child in node.get_children(): visit(child) visit(course) return units
[ "def", "get_units_with_due_date", "(", "course", ")", ":", "units", "=", "[", "]", "def", "visit", "(", "node", ")", ":", "if", "getattr", "(", "node", ",", "'due'", ",", "None", ")", ":", "units", ".", "append", "(", "node", ")", "else", ":", "for...
returns all top level units which have due dates .
train
false
10,447
def _step6(state): if (np.any(state.row_uncovered) and np.any(state.col_uncovered)): minval = np.min(state.C[state.row_uncovered], axis=0) minval = np.min(minval[state.col_uncovered]) state.C[(~ state.row_uncovered)] += minval state.C[:, state.col_uncovered] -= minval return _step4
[ "def", "_step6", "(", "state", ")", ":", "if", "(", "np", ".", "any", "(", "state", ".", "row_uncovered", ")", "and", "np", ".", "any", "(", "state", ".", "col_uncovered", ")", ")", ":", "minval", "=", "np", ".", "min", "(", "state", ".", "C", ...
add the value found in step 4 to every element of each covered row .
train
false
10,449
def pdbref(accessing_obj, accessed_obj, *args, **kwargs): return dbref(_to_player(accessing_obj), accessed_obj, *args, **kwargs)
[ "def", "pdbref", "(", "accessing_obj", ",", "accessed_obj", ",", "*", "args", ",", "**", "kwargs", ")", ":", "return", "dbref", "(", "_to_player", "(", "accessing_obj", ")", ",", "accessed_obj", ",", "*", "args", ",", "**", "kwargs", ")" ]
same as dbref .
train
false