id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
22,962
def chain_dot(*arrs): return reduce((lambda x, y: np.dot(y, x)), arrs[::(-1)])
[ "def", "chain_dot", "(", "*", "arrs", ")", ":", "return", "reduce", "(", "(", "lambda", "x", ",", "y", ":", "np", ".", "dot", "(", "y", ",", "x", ")", ")", ",", "arrs", "[", ":", ":", "(", "-", "1", ")", "]", ")" ]
returns the dot product of the given matrices .
train
false
22,963
def _entity_descriptor(entity, key): insp = inspection.inspect(entity) if insp.is_selectable: description = entity entity = insp.c elif insp.is_aliased_class: entity = insp.entity description = entity elif hasattr(insp, 'mapper'): description = entity = insp.mapper.class_ else: description = entity try: return getattr(entity, key) except AttributeError: raise sa_exc.InvalidRequestError(("Entity '%s' has no property '%s'" % (description, key)))
[ "def", "_entity_descriptor", "(", "entity", ",", "key", ")", ":", "insp", "=", "inspection", ".", "inspect", "(", "entity", ")", "if", "insp", ".", "is_selectable", ":", "description", "=", "entity", "entity", "=", "insp", ".", "c", "elif", "insp", ".", ...
return a class attribute given an entity and string name .
train
false
22,964
def sockaddr(host, port, network='ipv4'): address_family = {'ipv4': socket.AF_INET, 'ipv6': socket.AF_INET6}[network] for (family, _, _, _, ip) in socket.getaddrinfo(host, None, address_family): ip = ip[0] if (family == address_family): break else: log.error(('Could not find %s address for %r' % (network, host))) info = socket.getaddrinfo(host, None, address_family) host = socket.inet_pton(address_family, ip) sockaddr = p16(address_family) sockaddr += pack(port, word_size=16, endianness='big') length = 0 if (network == 'ipv4'): sockaddr += host length = 16 else: sockaddr += p32(4294967295) sockaddr += host length = (len(sockaddr) + 4) return (sockaddr, length, address_family)
[ "def", "sockaddr", "(", "host", ",", "port", ",", "network", "=", "'ipv4'", ")", ":", "address_family", "=", "{", "'ipv4'", ":", "socket", ".", "AF_INET", ",", "'ipv6'", ":", "socket", ".", "AF_INET6", "}", "[", "network", "]", "for", "(", "family", ...
sockaddr -> creates a sockaddr_in or sockaddr_in6 memory buffer for use in shellcode .
train
false
22,966
def text_to_bin(ip): if (':' not in ip): return ipv4_to_bin(ip) else: return ipv6_to_bin(ip)
[ "def", "text_to_bin", "(", "ip", ")", ":", "if", "(", "':'", "not", "in", "ip", ")", ":", "return", "ipv4_to_bin", "(", "ip", ")", "else", ":", "return", "ipv6_to_bin", "(", "ip", ")" ]
converts human readable ipv4 or ipv6 string to binary representation .
train
false
22,968
def str2dn(dn, flags=0): if (not dn): return [] return ldap.functions._ldap_function_call(None, _ldap.str2dn, dn, flags)
[ "def", "str2dn", "(", "dn", ",", "flags", "=", "0", ")", ":", "if", "(", "not", "dn", ")", ":", "return", "[", "]", "return", "ldap", ".", "functions", ".", "_ldap_function_call", "(", "None", ",", "_ldap", ".", "str2dn", ",", "dn", ",", "flags", ...
this function takes a dn as string as parameter and returns a decomposed dn .
train
false
22,969
def assert_clean(data): def _ensure_clean(value): if (value != bleach.clean(value)): raise ValueError return escape_html(data)
[ "def", "assert_clean", "(", "data", ")", ":", "def", "_ensure_clean", "(", "value", ")", ":", "if", "(", "value", "!=", "bleach", ".", "clean", "(", "value", ")", ")", ":", "raise", "ValueError", "return", "escape_html", "(", "data", ")" ]
ensure that data is cleaned :raise: assertionerror .
train
false
22,971
def ListBuckets(): return ObjectStore.ListInstances()
[ "def", "ListBuckets", "(", ")", ":", "return", "ObjectStore", ".", "ListInstances", "(", ")" ]
return the list of bucket names initialized in the object store .
train
false
22,972
def _check_rule(rule, _rule): if (_rule.get('from_port') is None): _rule['from_port'] = (-1) if (_rule.get('to_port') is None): _rule['to_port'] = (-1) if ((rule['ip_protocol'] == _rule['ip_protocol']) and (str(rule['from_port']) == str(_rule['from_port'])) and (str(rule['to_port']) == str(_rule['to_port']))): _cidr_ip = _rule.get('cidr_ip') if (_cidr_ip and (_cidr_ip == rule.get('cidr_ip'))): return True _owner_id = _rule.get('source_group_owner_id') if (_owner_id and (_owner_id == rule.get('source_group_owner_id'))): return True _group_id = _rule.get('source_group_group_id') if (_group_id and (_group_id == rule.get('source_group_group_id'))): return True _group_name = _rule.get('source_group_name') if (_group_name and (_group_id == rule.get('source_group_name'))): return True return False
[ "def", "_check_rule", "(", "rule", ",", "_rule", ")", ":", "if", "(", "_rule", ".", "get", "(", "'from_port'", ")", "is", "None", ")", ":", "_rule", "[", "'from_port'", "]", "=", "(", "-", "1", ")", "if", "(", "_rule", ".", "get", "(", "'to_port'...
check to see if two rules are the same .
train
false
22,973
def maybe_timedelta(delta): if isinstance(delta, numbers.Real): return timedelta(seconds=delta) return delta
[ "def", "maybe_timedelta", "(", "delta", ")", ":", "if", "isinstance", "(", "delta", ",", "numbers", ".", "Real", ")", ":", "return", "timedelta", "(", "seconds", "=", "delta", ")", "return", "delta" ]
convert integer to timedelta .
train
false
22,974
def scrub_dt_dn(dt, dn): (ndt, ndn) = (dt, dn) if (dt in lower_case_files_for): (ndt, ndn) = (scrub(dt), scrub(dn)) return (ndt, ndn)
[ "def", "scrub_dt_dn", "(", "dt", ",", "dn", ")", ":", "(", "ndt", ",", "ndn", ")", "=", "(", "dt", ",", "dn", ")", "if", "(", "dt", "in", "lower_case_files_for", ")", ":", "(", "ndt", ",", "ndn", ")", "=", "(", "scrub", "(", "dt", ")", ",", ...
returns in lowercase and code friendly names of doctype and name for certain types .
train
false
22,975
def prepare_earth_position_vel(time): from ..solar_system import get_body_barycentric, get_body_barycentric_posvel earth_pv = get_body_barycentric_posvel(u'earth', time) sun = get_body_barycentric(u'sun', time) earth_heliocentric = (earth_pv[0] - sun).get_xyz(xyz_axis=(-1)).to(u.au).value earth_pv = np.concatenate((earth_pv[0].get_xyz(xyz_axis=(-1)).to(u.au)[..., np.newaxis, :].value, earth_pv[1].get_xyz(xyz_axis=(-1)).to((u.au / u.d))[..., np.newaxis, :].value), axis=(-2)) return (earth_pv, earth_heliocentric)
[ "def", "prepare_earth_position_vel", "(", "time", ")", ":", "from", ".", ".", "solar_system", "import", "get_body_barycentric", ",", "get_body_barycentric_posvel", "earth_pv", "=", "get_body_barycentric_posvel", "(", "u'earth'", ",", "time", ")", "sun", "=", "get_body...
get barycentric position and velocity .
train
false
22,976
def get_browser(user_agent): match = re.search('(?i)(firefox|msie|chrome|safari|trident)', user_agent, re.IGNORECASE) if match: browser = match.group(1) else: browser = None return browser
[ "def", "get_browser", "(", "user_agent", ")", ":", "match", "=", "re", ".", "search", "(", "'(?i)(firefox|msie|chrome|safari|trident)'", ",", "user_agent", ",", "re", ".", "IGNORECASE", ")", "if", "match", ":", "browser", "=", "match", ".", "group", "(", "1"...
get browser name from user agent .
train
false
22,977
def commit(): connection._commit() set_clean()
[ "def", "commit", "(", ")", ":", "connection", ".", "_commit", "(", ")", "set_clean", "(", ")" ]
commit messages to transifex .
train
false
22,978
def set_owner_permissions(resource): if resource.polymorphic_ctype: if (resource.polymorphic_ctype.name == 'layer'): for perm in LAYER_ADMIN_PERMISSIONS: assign_perm(perm, resource.owner, resource.layer) set_geofence_owner(resource, str(resource.owner)) for perm in ADMIN_PERMISSIONS: assign_perm(perm, resource.owner, resource.get_self_resource())
[ "def", "set_owner_permissions", "(", "resource", ")", ":", "if", "resource", ".", "polymorphic_ctype", ":", "if", "(", "resource", ".", "polymorphic_ctype", ".", "name", "==", "'layer'", ")", ":", "for", "perm", "in", "LAYER_ADMIN_PERMISSIONS", ":", "assign_perm...
assign all admin permissions to the owner .
train
false
22,979
def decorate_logger_methods(logger): logger.findCaller = find_caller for key in LOGGER_KEYS: log_method = getattr(logger, key) log_method = decorate_log_method(log_method) setattr(logger, key, log_method) return logger
[ "def", "decorate_logger_methods", "(", "logger", ")", ":", "logger", ".", "findCaller", "=", "find_caller", "for", "key", "in", "LOGGER_KEYS", ":", "log_method", "=", "getattr", "(", "logger", ",", "key", ")", "log_method", "=", "decorate_log_method", "(", "lo...
decorate all the logger methods so all the keys in the extra dictionary are automatically prefixed with an underscore to avoid clashes with standard log record attributes .
train
false
22,980
@contextmanager def _temp_alembic_ini(db_url): with TemporaryDirectory() as td: alembic_ini = os.path.join(td, 'alembic.ini') write_alembic_ini(alembic_ini, db_url) (yield alembic_ini)
[ "@", "contextmanager", "def", "_temp_alembic_ini", "(", "db_url", ")", ":", "with", "TemporaryDirectory", "(", ")", "as", "td", ":", "alembic_ini", "=", "os", ".", "path", ".", "join", "(", "td", ",", "'alembic.ini'", ")", "write_alembic_ini", "(", "alembic_...
context manager for temporary jupyterhub alembic directory temporarily write an alembic .
train
false
22,981
def grab_doc(cmd, trap_error=True): proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, shell=True) (stdout, stderr) = proc.communicate() if (trap_error and proc.returncode): msg = (u'Attempting to run %s. Returned Error: %s' % (cmd, stderr)) raise IOError(msg) if stderr: return stderr return stdout
[ "def", "grab_doc", "(", "cmd", ",", "trap_error", "=", "True", ")", ":", "proc", "=", "subprocess", ".", "Popen", "(", "cmd", ",", "stdout", "=", "subprocess", ".", "PIPE", ",", "stderr", "=", "subprocess", ".", "PIPE", ",", "shell", "=", "True", ")"...
run cmd without args and grab documentation .
train
false
22,983
def check_doi(doi): if (not doi): return None doi_check = re.search('10\\.\\d{4}/\\S+', doi) if (doi_check is not None): return doi_check.group() return None
[ "def", "check_doi", "(", "doi", ")", ":", "if", "(", "not", "doi", ")", ":", "return", "None", "doi_check", "=", "re", ".", "search", "(", "'10\\\\.\\\\d{4}/\\\\S+'", ",", "doi", ")", "if", "(", "doi_check", "is", "not", "None", ")", ":", "return", "...
check if something that looks like a doi is present anywhere in the string .
train
false
22,984
def unix_path(path, is_win32=utils.is_win32): unix_path = path if is_win32(): first = path[0] second = path[1] if (second == u':'): unix_path = ((u'/' + first) + path[2:].replace(u'\\', u'/')) return unix_path
[ "def", "unix_path", "(", "path", ",", "is_win32", "=", "utils", ".", "is_win32", ")", ":", "unix_path", "=", "path", "if", "is_win32", "(", ")", ":", "first", "=", "path", "[", "0", "]", "second", "=", "path", "[", "1", "]", "if", "(", "second", ...
git for windows requires unix paths .
train
false
22,985
def make_field_value_list(headers, field, mdata): return sorted(set(mdata[:, headers.index(field)]))
[ "def", "make_field_value_list", "(", "headers", ",", "field", ",", "mdata", ")", ":", "return", "sorted", "(", "set", "(", "mdata", "[", ":", ",", "headers", ".", "index", "(", "field", ")", "]", ")", ")" ]
return sorted list of unique values field takes in mdata .
train
false
22,987
def _get_image_infos(image): status = base_status.copy() client = _get_client() try: infos = client.inspect_image(image) if infos: _valid(status, id_=infos['Id'], out=infos, comment='found') except Exception: pass if (not status['id']): _invalid(status) raise CommandExecutionError("ImageID '{0}' could not be resolved to an existing Image".format(image)) return status['out']
[ "def", "_get_image_infos", "(", "image", ")", ":", "status", "=", "base_status", ".", "copy", "(", ")", "client", "=", "_get_client", "(", ")", "try", ":", "infos", "=", "client", ".", "inspect_image", "(", "image", ")", "if", "infos", ":", "_valid", "...
verify that the image exists we will try to resolve either by: - name - image_id - tag image image name / image id / image tag returns the image id .
train
false
22,988
def create_floatingip(floating_network, port=None, profile=None): conn = _auth(profile) return conn.create_floatingip(floating_network, port)
[ "def", "create_floatingip", "(", "floating_network", ",", "port", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_auth", "(", "profile", ")", "return", "conn", ".", "create_floatingip", "(", "floating_network", ",", "port", ")" ]
creates a new floatingip cli example: .
train
true
22,989
def minmax_scale(X, feature_range=(0, 1), axis=0, copy=True): X = check_array(X, copy=False, ensure_2d=False, warn_on_dtype=True, dtype=FLOAT_DTYPES) original_ndim = X.ndim if (original_ndim == 1): X = X.reshape(X.shape[0], 1) s = MinMaxScaler(feature_range=feature_range, copy=copy) if (axis == 0): X = s.fit_transform(X) else: X = s.fit_transform(X.T).T if (original_ndim == 1): X = X.ravel() return X
[ "def", "minmax_scale", "(", "X", ",", "feature_range", "=", "(", "0", ",", "1", ")", ",", "axis", "=", "0", ",", "copy", "=", "True", ")", ":", "X", "=", "check_array", "(", "X", ",", "copy", "=", "False", ",", "ensure_2d", "=", "False", ",", "...
transforms features by scaling each feature to a given range .
train
false
22,990
@login_required def reset_api_key(request): request.user.auth_token.delete() Token.objects.create(user=request.user, key=get_random_string(40)) return redirect_profile(u'#api')
[ "@", "login_required", "def", "reset_api_key", "(", "request", ")", ":", "request", ".", "user", ".", "auth_token", ".", "delete", "(", ")", "Token", ".", "objects", ".", "create", "(", "user", "=", "request", ".", "user", ",", "key", "=", "get_random_st...
resets user api key .
train
false
22,992
def get_project_type_handler(project_type): global PROJECT_TYPES return PROJECT_TYPES.get(project_type)
[ "def", "get_project_type_handler", "(", "project_type", ")", ":", "global", "PROJECT_TYPES", "return", "PROJECT_TYPES", ".", "get", "(", "project_type", ")" ]
returns the handler for the given project_type .
train
false
22,993
def delete_file_system(filesystemid, keyid=None, key=None, profile=None, region=None, **kwargs): client = _get_conn(key=key, keyid=keyid, profile=profile, region=region) client.delete_file_system(FileSystemId=filesystemid)
[ "def", "delete_file_system", "(", "filesystemid", ",", "keyid", "=", "None", ",", "key", "=", "None", ",", "profile", "=", "None", ",", "region", "=", "None", ",", "**", "kwargs", ")", ":", "client", "=", "_get_conn", "(", "key", "=", "key", ",", "ke...
deletes a file system .
train
true
22,994
def list_contexts(): return _context_reg.keys()
[ "def", "list_contexts", "(", ")", ":", "return", "_context_reg", ".", "keys", "(", ")" ]
return an iterable of all the registered context names .
train
false
22,996
def list_groups(refresh=False): if (('group.list_groups' in __context__) and (not refresh)): return __context__['group.getent'] ret = [] pythoncom.CoInitialize() nt = win32com.client.Dispatch('AdsNameSpaces') results = nt.GetObject('', 'WinNT://.') results.Filter = ['group'] for result in results: ret.append(result.name) __context__['group.list_groups'] = ret return ret
[ "def", "list_groups", "(", "refresh", "=", "False", ")", ":", "if", "(", "(", "'group.list_groups'", "in", "__context__", ")", "and", "(", "not", "refresh", ")", ")", ":", "return", "__context__", "[", "'group.getent'", "]", "ret", "=", "[", "]", "python...
return a list of groups the named user belongs to .
train
false
22,997
def _node_func(G): if G.is_multigraph(): def sorted_node(u, v, key): return ((u, v, key) if (u <= v) else (v, u, key)) else: def sorted_node(u, v): return ((u, v) if (u <= v) else (v, u)) return sorted_node
[ "def", "_node_func", "(", "G", ")", ":", "if", "G", ".", "is_multigraph", "(", ")", ":", "def", "sorted_node", "(", "u", ",", "v", ",", "key", ")", ":", "return", "(", "(", "u", ",", "v", ",", "key", ")", "if", "(", "u", "<=", "v", ")", "el...
returns a function which returns a sorted node for line graphs .
train
false
22,998
@lru_cache(maxsize=1024) def s_esc(s): return s.replace('/', '\\/')
[ "@", "lru_cache", "(", "maxsize", "=", "1024", ")", "def", "s_esc", "(", "s", ")", ":", "return", "s", ".", "replace", "(", "'/'", ",", "'\\\\/'", ")" ]
equivalent to s .
train
false
22,999
def compat_input(prompt): sys.stdout.write(prompt) sys.stdout.flush() return raw_input()
[ "def", "compat_input", "(", "prompt", ")", ":", "sys", ".", "stdout", ".", "write", "(", "prompt", ")", "sys", ".", "stdout", ".", "flush", "(", ")", "return", "raw_input", "(", ")" ]
cygwins ptys are based on pipes .
train
false
23,001
@with_setup(setup, teardown) def test_show_samples(): rows = 10 cols = 10 m = (rows * cols) model = show_samples.load_model('dbm.pkl', m) dataset = yaml_parse.load(model.dataset_yaml_src) samples_viewer = show_samples.init_viewer(dataset, rows, cols) vis_batch = dataset.get_batch_topo(m) show_samples.update_viewer(dataset, samples_viewer, vis_batch, rows, cols)
[ "@", "with_setup", "(", "setup", ",", "teardown", ")", "def", "test_show_samples", "(", ")", ":", "rows", "=", "10", "cols", "=", "10", "m", "=", "(", "rows", "*", "cols", ")", "model", "=", "show_samples", ".", "load_model", "(", "'dbm.pkl'", ",", "...
test the samples update_viewer function .
train
false
23,003
def _smart_truncate(content, length=100, suffix='...'): if (len(content) <= length): return content return (content[:length].rsplit(' ', 1)[0] + suffix)
[ "def", "_smart_truncate", "(", "content", ",", "length", "=", "100", ",", "suffix", "=", "'...'", ")", ":", "if", "(", "len", "(", "content", ")", "<=", "length", ")", ":", "return", "content", "return", "(", "content", "[", ":", "length", "]", ".", ...
smart truncate .
train
false
23,004
def deltify_pack_objects(objects, window_size=None): if (window_size is None): window_size = DEFAULT_PACK_DELTA_WINDOW_SIZE magic = [] for (obj, path) in objects: magic.append((obj.type_num, path, (- obj.raw_length()), obj)) magic.sort() possible_bases = deque() for (type_num, path, neg_length, o) in magic: raw = o.as_raw_string() winner = raw winner_base = None for base in possible_bases: if (base.type_num != type_num): continue delta = create_delta(base.as_raw_string(), raw) if (len(delta) < len(winner)): winner_base = base.sha().digest() winner = delta (yield (type_num, o.sha().digest(), winner_base, winner)) possible_bases.appendleft(o) while (len(possible_bases) > window_size): possible_bases.pop()
[ "def", "deltify_pack_objects", "(", "objects", ",", "window_size", "=", "None", ")", ":", "if", "(", "window_size", "is", "None", ")", ":", "window_size", "=", "DEFAULT_PACK_DELTA_WINDOW_SIZE", "magic", "=", "[", "]", "for", "(", "obj", ",", "path", ")", "...
generate deltas for pack objects .
train
false
23,005
@hook.command(autohelp=False) def beats(text): if (text.lower() == 'wut'): return 'Instead of hours and minutes, the mean solar day is divided up into 1000 parts called ".beats". Each .beat lasts 1 minute and 26.4 seconds. Times are notated as a 3-digit number out of 1000 after midnight. So, @248 would indicate a time 248 .beats after midnight representing 248/1000 of a day, just over 5 hours and 57 minutes. There are no timezones.' elif (text.lower() == 'guide'): return '1 day = 1000 .beats, 1 hour = 41.666 .beats, 1 min = 0.6944 .beats, 1 second = 0.01157 .beats' t = time.gmtime() (h, m, s) = (t.tm_hour, t.tm_min, t.tm_sec) utc = (((3600 * h) + (60 * m)) + s) bmt = (utc + 3600) beat = (bmt / 86.4) if (beat > 1000): beat -= 1000 return ('Swatch Internet Time: @%06.2f' % beat)
[ "@", "hook", ".", "command", "(", "autohelp", "=", "False", ")", "def", "beats", "(", "text", ")", ":", "if", "(", "text", ".", "lower", "(", ")", "==", "'wut'", ")", ":", "return", "'Instead of hours and minutes, the mean solar day is divided up into 1000 parts...
shows the internet time in swatch beats .
train
false
23,006
def _get_all_files_in_directory(dir_path, excluded_glob_patterns): files_in_directory = [] for (_dir, _, files) in os.walk(dir_path): for file_name in files: filename = os.path.relpath(os.path.join(_dir, file_name), os.getcwd()) if (not any([fnmatch.fnmatch(filename, gp) for gp in excluded_glob_patterns])): files_in_directory.append(filename) return files_in_directory
[ "def", "_get_all_files_in_directory", "(", "dir_path", ",", "excluded_glob_patterns", ")", ":", "files_in_directory", "=", "[", "]", "for", "(", "_dir", ",", "_", ",", "files", ")", "in", "os", ".", "walk", "(", "dir_path", ")", ":", "for", "file_name", "i...
recursively collects all files in directory and subdirectories of specified path .
train
false
23,007
def actor_url(parser, token): bits = token.split_contents() if (len(bits) != 2): raise TemplateSyntaxError('Accepted format {% actor_url [actor_instance] %}') else: return DisplayActivityActorUrl(*bits[1:])
[ "def", "actor_url", "(", "parser", ",", "token", ")", ":", "bits", "=", "token", ".", "split_contents", "(", ")", "if", "(", "len", "(", "bits", ")", "!=", "2", ")", ":", "raise", "TemplateSyntaxError", "(", "'Accepted format {% actor_url [actor_instance] %}'"...
renders the url for a particular actor instance <a href="{% actor_url request .
train
true
23,008
def log_query_count(name_point): log.info((u'At point named `%s` total of %s queries were ran' % (name_point, query_count)))
[ "def", "log_query_count", "(", "name_point", ")", ":", "log", ".", "info", "(", "(", "u'At point named `%s` total of %s queries were ran'", "%", "(", "name_point", ",", "query_count", ")", ")", ")" ]
debugging purposes .
train
false
23,009
def list_runners(*args): run_ = salt.runner.Runner(__opts__) runners = set() if (not args): for func in run_.functions: runners.add(func.split('.')[0]) return sorted(runners) for module in args: if ('*' in module): for func in fnmatch.filter(run_.functions, module): runners.add(func.split('.')[0]) else: for func in run_.functions: mod_test = func.split('.')[0] if (mod_test == module): runners.add(mod_test) return sorted(runners)
[ "def", "list_runners", "(", "*", "args", ")", ":", "run_", "=", "salt", ".", "runner", ".", "Runner", "(", "__opts__", ")", "runners", "=", "set", "(", ")", "if", "(", "not", "args", ")", ":", "for", "func", "in", "run_", ".", "functions", ":", "...
list the runners loaded on the minion .
train
true
23,010
@task def test_admin(ctx): module = 'admin_tests/' module_fmt = (' '.join(module) if isinstance(module, list) else module) admin_tasks.manage(ctx, 'test {}'.format(module_fmt))
[ "@", "task", "def", "test_admin", "(", "ctx", ")", ":", "module", "=", "'admin_tests/'", "module_fmt", "=", "(", "' '", ".", "join", "(", "module", ")", "if", "isinstance", "(", "module", ",", "list", ")", "else", "module", ")", "admin_tasks", ".", "ma...
run the admin test suite .
train
false
23,011
def get_os_ca_bundle_path(): for path in POSSIBLE_CA_BUNDLE_PATHS: if os.path.exists(path): return path return None
[ "def", "get_os_ca_bundle_path", "(", ")", ":", "for", "path", "in", "POSSIBLE_CA_BUNDLE_PATHS", ":", "if", "os", ".", "path", ".", "exists", "(", "path", ")", ":", "return", "path", "return", "None" ]
try to pick an available ca certificate bundle provided by the os .
train
false
23,013
def qos_specs_get(context, qos_specs_id): return IMPL.qos_specs_get(context, qos_specs_id)
[ "def", "qos_specs_get", "(", "context", ",", "qos_specs_id", ")", ":", "return", "IMPL", ".", "qos_specs_get", "(", "context", ",", "qos_specs_id", ")" ]
get all specification for a given qos_specs .
train
false
23,014
def inch(value): return dpi2px(value, 'in')
[ "def", "inch", "(", "value", ")", ":", "return", "dpi2px", "(", "value", ",", "'in'", ")" ]
convert from inches to pixels .
train
false
23,015
@register.inclusion_tag('horizon/common/_progress_bar.html') def horizon_progress_bar(current_val, max_val): return {'current_val': current_val, 'max_val': max_val}
[ "@", "register", ".", "inclusion_tag", "(", "'horizon/common/_progress_bar.html'", ")", "def", "horizon_progress_bar", "(", "current_val", ",", "max_val", ")", ":", "return", "{", "'current_val'", ":", "current_val", ",", "'max_val'", ":", "max_val", "}" ]
renders a progress bar based on parameters passed to the tag .
train
false
23,016
def list_tasks(location='\\'): pythoncom.CoInitialize() task_service = win32com.client.Dispatch('Schedule.Service') task_service.Connect() task_folder = task_service.GetFolder(location) tasks = task_folder.GetTasks(0) ret = [] for task in tasks: ret.append(task.Name) return ret
[ "def", "list_tasks", "(", "location", "=", "'\\\\'", ")", ":", "pythoncom", ".", "CoInitialize", "(", ")", "task_service", "=", "win32com", ".", "client", ".", "Dispatch", "(", "'Schedule.Service'", ")", "task_service", ".", "Connect", "(", ")", "task_folder",...
list all tasks located in a specific location in the task scheduler .
train
false
23,017
def tag_for_tool(tool): return slugify(tool.name, delim='-')
[ "def", "tag_for_tool", "(", "tool", ")", ":", "return", "slugify", "(", "tool", ".", "name", ",", "delim", "=", "'-'", ")" ]
generate a reasonable biostar tag for a tool .
train
false
23,018
def update_api_model_schema(restApiId, modelName, schema, region=None, key=None, keyid=None, profile=None): try: schema_json = (json.dumps(schema) if isinstance(schema, dict) else schema) conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) response = _api_model_patch_replace(conn, restApiId, modelName, '/schema', schema_json) return {'updated': True, 'model': _convert_datetime_str(response)} except ClientError as e: return {'updated': False, 'error': salt.utils.boto3.get_error(e)}
[ "def", "update_api_model_schema", "(", "restApiId", ",", "modelName", ",", "schema", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "try", ":", "schema_json", "=", "(", "json", "."...
update the schema for the given model in the given restapiid cli example: .
train
true
23,019
def find_header(hdr): for dir in ['/usr/include', '/usr/local/include']: file = os.path.join(dir, hdr) if os.path.exists(file): return file raise ValueError(('Missing header: %s' % hdr))
[ "def", "find_header", "(", "hdr", ")", ":", "for", "dir", "in", "[", "'/usr/include'", ",", "'/usr/local/include'", "]", ":", "file", "=", "os", ".", "path", ".", "join", "(", "dir", ",", "hdr", ")", "if", "os", ".", "path", ".", "exists", "(", "fi...
find a given header in the system .
train
false
23,020
def is_nanpa_country(region_code): return (region_code in _NANPA_REGIONS)
[ "def", "is_nanpa_country", "(", "region_code", ")", ":", "return", "(", "region_code", "in", "_NANPA_REGIONS", ")" ]
checks if this region is a nanpa region .
train
false
23,021
def init_evolutions(app, created_models, **kwargs): if (FileDiff in created_models): return try: latest_version = django_evolution.Version.objects.latest(u'when') except django_evolution.Version.DoesNotExist: latest_version = None if latest_version: try: FileDiff.objects.filter(parent_diff64=u'') return except: django_evolution.Version.objects.all().delete() django_evolution.Evolution.objects.all().delete() call_command(u'loaddata', u'admin/fixtures/initial_evolution_schema.json', verbosity=0)
[ "def", "init_evolutions", "(", "app", ",", "created_models", ",", "**", "kwargs", ")", ":", "if", "(", "FileDiff", "in", "created_models", ")", ":", "return", "try", ":", "latest_version", "=", "django_evolution", ".", "Version", ".", "objects", ".", "latest...
attempt to initialize the django evolution schema signatures .
train
false
23,022
def clean_txt(txt): if isbytestring(txt): txt = txt.decode('utf-8', 'replace') txt = '\n'.join([line.rstrip() for line in txt.splitlines()]) txt = re.sub('(?m)(?<=^)([ ]{2,}| DCTB +)(?=.)', ('&nbsp;' * 4), txt) txt = re.sub('[ ]{2,}', ' ', txt) txt = re.sub('^\\s+(?=.)', '', txt) txt = re.sub('(?<=.)\\s+$', '', txt) txt = re.sub('\n{5,}', '\n\n\n\n', txt) txt = clean_ascii_chars(txt) return txt
[ "def", "clean_txt", "(", "txt", ")", ":", "if", "isbytestring", "(", "txt", ")", ":", "txt", "=", "txt", ".", "decode", "(", "'utf-8'", ",", "'replace'", ")", "txt", "=", "'\\n'", ".", "join", "(", "[", "line", ".", "rstrip", "(", ")", "for", "li...
run transformations on the text to put it into consistent state .
train
false
23,026
def check_bgp(net_connect, cmd='show run | inc router bgp'): output = net_connect.send_command_expect(cmd) return ('bgp' in output)
[ "def", "check_bgp", "(", "net_connect", ",", "cmd", "=", "'show run | inc router bgp'", ")", ":", "output", "=", "net_connect", ".", "send_command_expect", "(", "cmd", ")", "return", "(", "'bgp'", "in", "output", ")" ]
check whether bgp is currently configured on device .
train
false
23,028
def read_str(fid, count=1): dtype = np.dtype(('>S%i' % count)) string = fid.read(dtype.itemsize) data = np.fromstring(string, dtype=dtype)[0] bytestr = b('').join([data[0:(data.index(b('\x00')) if (b('\x00') in data) else count)]]) return str(bytestr.decode('ascii'))
[ "def", "read_str", "(", "fid", ",", "count", "=", "1", ")", ":", "dtype", "=", "np", ".", "dtype", "(", "(", "'>S%i'", "%", "count", ")", ")", "string", "=", "fid", ".", "read", "(", "dtype", ".", "itemsize", ")", "data", "=", "np", ".", "froms...
if a python string literal begins at the specified position in the given string .
train
false
23,029
def referer_str(request): referrer = request.headers.get('Referer') if (referrer is None): return referrer return to_native_str(referrer, errors='replace')
[ "def", "referer_str", "(", "request", ")", ":", "referrer", "=", "request", ".", "headers", ".", "get", "(", "'Referer'", ")", "if", "(", "referrer", "is", "None", ")", ":", "return", "referrer", "return", "to_native_str", "(", "referrer", ",", "errors", ...
return referer http header suitable for logging .
train
false
23,030
def config_set(key, value=None, multivar=None, cwd=None, user=None, password=None, ignore_retcode=False, **kwargs): kwargs = salt.utils.clean_kwargs(**kwargs) add_ = kwargs.pop('add', False) global_ = kwargs.pop('global', False) if kwargs: salt.utils.invalid_kwargs(kwargs) if (cwd is None): if (not global_): raise SaltInvocationError("'cwd' argument required unless global=True") else: cwd = _expand_path(cwd, user) if all(((x is not None) for x in (value, multivar))): raise SaltInvocationError("Only one of 'value' and 'multivar' is permitted") if (value is not None): if (not isinstance(value, six.string_types)): value = str(value) if (multivar is not None): if (not isinstance(multivar, list)): try: multivar = multivar.split(',') except AttributeError: multivar = str(multivar).split(',') else: new_multivar = [] for item in multivar: if isinstance(item, six.string_types): new_multivar.append(item) else: new_multivar.append(str(item)) multivar = new_multivar command_prefix = ['git', 'config'] if global_: command_prefix.append('--global') if (value is not None): command = copy.copy(command_prefix) if add_: command.append('--add') else: command.append('--replace-all') command.extend([key, value]) _git_run(command, cwd=cwd, user=user, password=password, ignore_retcode=ignore_retcode) else: for (idx, target) in enumerate(multivar): command = copy.copy(command_prefix) if (idx == 0): command.append('--replace-all') else: command.append('--add') command.extend([key, target]) _git_run(command, cwd=cwd, user=user, password=password, ignore_retcode=ignore_retcode) return config_get(key, user=user, password=password, cwd=cwd, ignore_retcode=ignore_retcode, **{'all': True, 'global': global_})
[ "def", "config_set", "(", "key", ",", "value", "=", "None", ",", "multivar", "=", "None", ",", "cwd", "=", "None", ",", "user", "=", "None", ",", "password", "=", "None", ",", "ignore_retcode", "=", "False", ",", "**", "kwargs", ")", ":", "kwargs", ...
set redis server configuration values cli example: .
train
true
23,032
def _read_complex_double(fid, tag, shape, rlims): if (shape is not None): shape = (shape[0], (shape[1] * 2)) d = _fromstring_rows(fid, tag.size, dtype='>f8', shape=shape, rlims=rlims) d = (d[::2] + (1j * d[1::2])) return d
[ "def", "_read_complex_double", "(", "fid", ",", "tag", ",", "shape", ",", "rlims", ")", ":", "if", "(", "shape", "is", "not", "None", ")", ":", "shape", "=", "(", "shape", "[", "0", "]", ",", "(", "shape", "[", "1", "]", "*", "2", ")", ")", "...
read complex double tag .
train
false
23,033
def new_local_dict(): d = {} _local_dict_stack.append(d) return d
[ "def", "new_local_dict", "(", ")", ":", "d", "=", "{", "}", "_local_dict_stack", ".", "append", "(", "d", ")", "return", "d" ]
initialize a new local dictionary & push it onto the stack .
train
false
23,034
def survey_answerlist_dataTable_post(r): current.response.s3.actions = [{'label': str(current.messages.UPDATE), '_class': 'action-btn edit', 'url': URL(c='survey', f='series', args=[r.id, 'complete', '[id]', 'update'])}]
[ "def", "survey_answerlist_dataTable_post", "(", "r", ")", ":", "current", ".", "response", ".", "s3", ".", "actions", "=", "[", "{", "'label'", ":", "str", "(", "current", ".", "messages", ".", "UPDATE", ")", ",", "'_class'", ":", "'action-btn edit'", ",",...
replace action buttons .
train
false
23,035
def dup_sqf_p(f, K): if (not f): return True else: return (not dup_degree(dup_gcd(f, dup_diff(f, 1, K), K)))
[ "def", "dup_sqf_p", "(", "f", ",", "K", ")", ":", "if", "(", "not", "f", ")", ":", "return", "True", "else", ":", "return", "(", "not", "dup_degree", "(", "dup_gcd", "(", "f", ",", "dup_diff", "(", "f", ",", "1", ",", "K", ")", ",", "K", ")",...
return true if f is a square-free polynomial in k[x] .
train
false
23,036
def dump_file(filename, head=None): if (head is None): log.info(('%s' % filename)) else: log.info(head) file = open(filename) try: log.info(file.read()) finally: file.close()
[ "def", "dump_file", "(", "filename", ",", "head", "=", "None", ")", ":", "if", "(", "head", "is", "None", ")", ":", "log", ".", "info", "(", "(", "'%s'", "%", "filename", ")", ")", "else", ":", "log", ".", "info", "(", "head", ")", "file", "=",...
dumps a file content into log .
train
false
23,037
def ip_bracket(addr): if (addr and (':' in addr) and (not addr.startswith('['))): return '[{0}]'.format(addr) return addr
[ "def", "ip_bracket", "(", "addr", ")", ":", "if", "(", "addr", "and", "(", "':'", "in", "addr", ")", "and", "(", "not", "addr", ".", "startswith", "(", "'['", ")", ")", ")", ":", "return", "'[{0}]'", ".", "format", "(", "addr", ")", "return", "ad...
convert ip address representation to zmq format .
train
false
23,039
def stubout_lookup_image(stubs): def f(_1, _2, _3, _4): raise Exception('Test Exception raised by fake lookup_image') stubs.Set(vm_utils, 'lookup_image', f)
[ "def", "stubout_lookup_image", "(", "stubs", ")", ":", "def", "f", "(", "_1", ",", "_2", ",", "_3", ",", "_4", ")", ":", "raise", "Exception", "(", "'Test Exception raised by fake lookup_image'", ")", "stubs", ".", "Set", "(", "vm_utils", ",", "'lookup_image...
simulates a failure in lookup image .
train
false
23,041
def should_write(write_opt=None): return _bool_fallback(write_opt, config['import']['write'].get(bool))
[ "def", "should_write", "(", "write_opt", "=", "None", ")", ":", "return", "_bool_fallback", "(", "write_opt", ",", "config", "[", "'import'", "]", "[", "'write'", "]", ".", "get", "(", "bool", ")", ")" ]
decide whether a command that updates metadata should also write tags .
train
false
23,042
def iterator(dictionary): if (PY_MAJOR_VERSION > 2): return dictionary.items() else: return dictionary.iteritems()
[ "def", "iterator", "(", "dictionary", ")", ":", "if", "(", "PY_MAJOR_VERSION", ">", "2", ")", ":", "return", "dictionary", ".", "items", "(", ")", "else", ":", "return", "dictionary", ".", "iteritems", "(", ")" ]
for cross compatibility between python 2 and python 3 dictionaries .
train
false
23,043
def _get_incdec_value(match, incdec, url, count): (pre, zeroes, number, post) = match.groups() val = int(number) if (incdec == 'decrement'): if (val <= 0): raise IncDecError("Can't decrement {}!".format(val), url) val -= count elif (incdec == 'increment'): val += count else: raise ValueError('Invalid value {} for indec!'.format(incdec)) if zeroes: if (len(number) < len(str(val))): zeroes = zeroes[1:] elif (len(number) > len(str(val))): zeroes += '0' return ''.join([pre, zeroes, str(val), post])
[ "def", "_get_incdec_value", "(", "match", ",", "incdec", ",", "url", ",", "count", ")", ":", "(", "pre", ",", "zeroes", ",", "number", ",", "post", ")", "=", "match", ".", "groups", "(", ")", "val", "=", "int", "(", "number", ")", "if", "(", "inc...
get an incremented/decremented url based on a url match .
train
false
23,044
def _empty_bucket(bucket): for blob in bucket.list_blobs(): try: blob.delete() except exceptions.NotFound: pass
[ "def", "_empty_bucket", "(", "bucket", ")", ":", "for", "blob", "in", "bucket", ".", "list_blobs", "(", ")", ":", "try", ":", "blob", ".", "delete", "(", ")", "except", "exceptions", ".", "NotFound", ":", "pass" ]
empty a bucket of all existing blobs .
train
false
23,045
def _insert_axis_in_shape(context, builder, orig_shape, ndim, axis): assert (len(orig_shape) == (ndim - 1)) ll_shty = ir.ArrayType(cgutils.intp_t, ndim) shapes = cgutils.alloca_once(builder, ll_shty) one = cgutils.intp_t(1) for dim in range((ndim - 1)): ll_dim = cgutils.intp_t(dim) after_axis = builder.icmp_signed('>=', ll_dim, axis) sh = orig_shape[dim] idx = builder.select(after_axis, builder.add(ll_dim, one), ll_dim) builder.store(sh, cgutils.gep_inbounds(builder, shapes, 0, idx)) builder.store(one, cgutils.gep_inbounds(builder, shapes, 0, axis)) return cgutils.unpack_tuple(builder, builder.load(shapes))
[ "def", "_insert_axis_in_shape", "(", "context", ",", "builder", ",", "orig_shape", ",", "ndim", ",", "axis", ")", ":", "assert", "(", "len", "(", "orig_shape", ")", "==", "(", "ndim", "-", "1", ")", ")", "ll_shty", "=", "ir", ".", "ArrayType", "(", "...
compute shape with the new axis inserted e .
train
false
23,046
def key_type_n(index): def transform_function(key): id_or_name = _key_id_or_name_n(key, index) if (id_or_name is None): return '' if isinstance(id_or_name, basestring): return KEY_TYPE_NAME return KEY_TYPE_ID return transform_function
[ "def", "key_type_n", "(", "index", ")", ":", "def", "transform_function", "(", "key", ")", ":", "id_or_name", "=", "_key_id_or_name_n", "(", "key", ",", "index", ")", "if", "(", "id_or_name", "is", "None", ")", ":", "return", "''", "if", "isinstance", "(...
pull out the nth key type from a key which has parents .
train
false
23,047
def getipaddrinfo(host): try: return [addrinfo for addrinfo in socket.getaddrinfo(host, None) if (((addrinfo[0] == socket.AF_INET) or (addrinfo[0] == socket.AF_INET6)) and isinstance(addrinfo[4][0], basestring))] except socket.error: return []
[ "def", "getipaddrinfo", "(", "host", ")", ":", "try", ":", "return", "[", "addrinfo", "for", "addrinfo", "in", "socket", ".", "getaddrinfo", "(", "host", ",", "None", ")", "if", "(", "(", "(", "addrinfo", "[", "0", "]", "==", "socket", ".", "AF_INET"...
filter out non-ip and bad ip addresses from getaddrinfo .
train
false
23,048
@cache_permission def can_lock_translation(user, project): return check_permission(user, project, 'trans.lock_translation')
[ "@", "cache_permission", "def", "can_lock_translation", "(", "user", ",", "project", ")", ":", "return", "check_permission", "(", "user", ",", "project", ",", "'trans.lock_translation'", ")" ]
checks whether user can lock translation .
train
false
23,049
def strip_bom(data): for bom in (codecs.BOM_UTF32_BE, codecs.BOM_UTF32_LE, codecs.BOM_UTF16_BE, codecs.BOM_UTF16_LE, codecs.BOM_UTF8): if data.startswith(bom): data = data[len(bom):] break return data
[ "def", "strip_bom", "(", "data", ")", ":", "for", "bom", "in", "(", "codecs", ".", "BOM_UTF32_BE", ",", "codecs", ".", "BOM_UTF32_LE", ",", "codecs", ".", "BOM_UTF16_BE", ",", "codecs", ".", "BOM_UTF16_LE", ",", "codecs", ".", "BOM_UTF8", ")", ":", "if",...
strip the bom from byte string data .
train
false
23,050
def remove_section(file_name, section, separator='='): inifile = _Ini.get_ini_file(file_name, separator=separator) section = inifile.pop(section, {}) inifile.flush() ret = {} for (key, value) in six.iteritems(section): if (key[0] != '#'): ret.update({key: value}) return ret
[ "def", "remove_section", "(", "file_name", ",", "section", ",", "separator", "=", "'='", ")", ":", "inifile", "=", "_Ini", ".", "get_ini_file", "(", "file_name", ",", "separator", "=", "separator", ")", "section", "=", "inifile", ".", "pop", "(", "section"...
remove a section in an ini file .
train
true
23,051
def _lookup_style(element, names): return (_STYLES.get(('_' + element), '') + ''.join([_STYLES.get(name, '') for name in names]))
[ "def", "_lookup_style", "(", "element", ",", "names", ")", ":", "return", "(", "_STYLES", ".", "get", "(", "(", "'_'", "+", "element", ")", ",", "''", ")", "+", "''", ".", "join", "(", "[", "_STYLES", ".", "get", "(", "name", ",", "''", ")", "f...
lookup style by either element name or the list of classes .
train
true
23,052
def gender(word, pos=NOUN): w = word.lower() if (pos == NOUN): if w.endswith(gender_masculine): return MASCULINE if w.endswith(gender_feminine): return FEMININE if w.endswith(gender_neuter): return NEUTER for g in gender_majority_vote: if w.endswith(gender_majority_vote[g]): return g
[ "def", "gender", "(", "word", ",", "pos", "=", "NOUN", ")", ":", "w", "=", "word", ".", "lower", "(", ")", "if", "(", "pos", "==", "NOUN", ")", ":", "if", "w", ".", "endswith", "(", "gender_masculine", ")", ":", "return", "MASCULINE", "if", "w", ...
returns the gender for nouns .
train
true
23,053
def _to_int(byte_string): return int(byte_string[::(-1)].encode('hex'), 16)
[ "def", "_to_int", "(", "byte_string", ")", ":", "return", "int", "(", "byte_string", "[", ":", ":", "(", "-", "1", ")", "]", ".", "encode", "(", "'hex'", ")", ",", "16", ")" ]
convert a string of bytes to int .
train
false
23,055
def HashPassword(password, salt): prf = (lambda p, s: HMAC.new(p, s, SHA512).digest()) return base64.b64encode(PBKDF2(escape.utf8(password), base64.b64decode(salt), count=1000, prf=prf))
[ "def", "HashPassword", "(", "password", ",", "salt", ")", ":", "prf", "=", "(", "lambda", "p", ",", "s", ":", "HMAC", ".", "new", "(", "p", ",", "s", ",", "SHA512", ")", ".", "digest", "(", ")", ")", "return", "base64", ".", "b64encode", "(", "...
computes the hash of the given password using 1000 sha512 iterations .
train
false
23,056
def publish_over_ssh(registry, xml_parent, data): ssh(registry, xml_parent, data)
[ "def", "publish_over_ssh", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "ssh", "(", "registry", ",", "xml_parent", ",", "data", ")" ]
yaml: publish-over-ssh send files or execute commands over ssh .
train
false
23,057
def get_systemd_os_info(filepath='/etc/os-release'): os_name = _get_systemd_os_release_var('ID', filepath=filepath) os_version = _get_systemd_os_release_var('VERSION_ID', filepath=filepath) return (os_name, os_version)
[ "def", "get_systemd_os_info", "(", "filepath", "=", "'/etc/os-release'", ")", ":", "os_name", "=", "_get_systemd_os_release_var", "(", "'ID'", ",", "filepath", "=", "filepath", ")", "os_version", "=", "_get_systemd_os_release_var", "(", "'VERSION_ID'", ",", "filepath"...
parse systemd /etc/os-release for distribution information .
train
false
23,058
def getNewRepository(): return ExportRepository()
[ "def", "getNewRepository", "(", ")", ":", "return", "ExportRepository", "(", ")" ]
get new repository .
train
false
23,061
def server_exists(s_name, ip=None, s_state=None, **connection_args): server = _server_get(s_name, **connection_args) if (server is None): return False if ((ip is not None) and (ip != server.get_ipaddress())): return False if ((s_state is not None) and (s_state.upper() != server.get_state())): return False return True
[ "def", "server_exists", "(", "s_name", ",", "ip", "=", "None", ",", "s_state", "=", "None", ",", "**", "connection_args", ")", ":", "server", "=", "_server_get", "(", "s_name", ",", "**", "connection_args", ")", "if", "(", "server", "is", "None", ")", ...
checks if a server exists cli example: .
train
true
23,062
def expand_power_exp(expr, deep=True): return sympify(expr).expand(deep=deep, complex=False, basic=False, log=False, mul=False, power_exp=True, power_base=False, multinomial=False)
[ "def", "expand_power_exp", "(", "expr", ",", "deep", "=", "True", ")", ":", "return", "sympify", "(", "expr", ")", ".", "expand", "(", "deep", "=", "deep", ",", "complex", "=", "False", ",", "basic", "=", "False", ",", "log", "=", "False", ",", "mu...
wrapper around expand that only uses the power_exp hint .
train
false
23,064
def removeH1(document): h1 = domhelpers.findNodesNamed(document, 'h1') empty = dom.Element('span') for node in h1: node.parentNode.replaceChild(empty, node)
[ "def", "removeH1", "(", "document", ")", ":", "h1", "=", "domhelpers", ".", "findNodesNamed", "(", "document", ",", "'h1'", ")", "empty", "=", "dom", ".", "Element", "(", "'span'", ")", "for", "node", "in", "h1", ":", "node", ".", "parentNode", ".", ...
replace all c{h1} nodes in the given document with empty c{span} nodes .
train
false
23,065
def zdt3(individual): g = (1.0 + ((9.0 * sum(individual[1:])) / (len(individual) - 1))) f1 = individual[0] f2 = (g * ((1 - sqrt((f1 / g))) - ((f1 / g) * sin(((10 * pi) * f1))))) return (f1, f2)
[ "def", "zdt3", "(", "individual", ")", ":", "g", "=", "(", "1.0", "+", "(", "(", "9.0", "*", "sum", "(", "individual", "[", "1", ":", "]", ")", ")", "/", "(", "len", "(", "individual", ")", "-", "1", ")", ")", ")", "f1", "=", "individual", ...
zdt3 multiobjective function .
train
false
23,066
@image_comparison(baseline_images=[u'EventCollection_plot__set_linewidth']) def test__EventCollection__set_linewidth(): (splt, coll, _) = generate_EventCollection_plot() new_linewidth = 5 coll.set_linewidth(new_linewidth) assert_equal(coll.get_linewidth(), new_linewidth) splt.set_title(u'EventCollection: set_linewidth')
[ "@", "image_comparison", "(", "baseline_images", "=", "[", "u'EventCollection_plot__set_linewidth'", "]", ")", "def", "test__EventCollection__set_linewidth", "(", ")", ":", "(", "splt", ",", "coll", ",", "_", ")", "=", "generate_EventCollection_plot", "(", ")", "new...
check to make sure set_linestyle works properly .
train
false
23,067
@contextmanager def target_cell(context, cell_mapping): original_db_connection = context.db_connection from nova import db db_connection_string = cell_mapping.database_connection context.db_connection = db.create_context_manager(db_connection_string) try: (yield context) finally: context.db_connection = original_db_connection
[ "@", "contextmanager", "def", "target_cell", "(", "context", ",", "cell_mapping", ")", ":", "original_db_connection", "=", "context", ".", "db_connection", "from", "nova", "import", "db", "db_connection_string", "=", "cell_mapping", ".", "database_connection", "contex...
adds database connection information to the context for communicating with the given target cell .
train
false
23,068
def convert_column_args(method): def column_wrapper(self, *args, **kwargs): try: if len(args): int(args[0]) except ValueError: (cell_1, cell_2) = [(col + '1') for col in args[0].split(':')] (_, col_1) = xl_cell_to_rowcol(cell_1) (_, col_2) = xl_cell_to_rowcol(cell_2) new_args = [col_1, col_2] new_args.extend(args[1:]) args = new_args return method(self, *args, **kwargs) return column_wrapper
[ "def", "convert_column_args", "(", "method", ")", ":", "def", "column_wrapper", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "if", "len", "(", "args", ")", ":", "int", "(", "args", "[", "0", "]", ")", "except", "ValueEr...
decorator function to convert a1 notation in columns method calls to the default row/col notation .
train
false
23,069
def has_vary_header(response, header_query): if (not response.has_header(u'Vary')): return False vary_headers = cc_delim_re.split(response[u'Vary']) existing_headers = set([header.lower() for header in vary_headers]) return (header_query.lower() in existing_headers)
[ "def", "has_vary_header", "(", "response", ",", "header_query", ")", ":", "if", "(", "not", "response", ".", "has_header", "(", "u'Vary'", ")", ")", ":", "return", "False", "vary_headers", "=", "cc_delim_re", ".", "split", "(", "response", "[", "u'Vary'", ...
checks to see if the response has a given header name in its vary header .
train
true
23,070
def is_program_installed(basename): for path in os.environ['PATH'].split(os.pathsep): abspath = osp.join(path, basename) if osp.isfile(abspath): return abspath
[ "def", "is_program_installed", "(", "basename", ")", ":", "for", "path", "in", "os", ".", "environ", "[", "'PATH'", "]", ".", "split", "(", "os", ".", "pathsep", ")", ":", "abspath", "=", "osp", ".", "join", "(", "path", ",", "basename", ")", "if", ...
return program absolute path if installed in path .
train
true
23,071
def malletmodel2ldamodel(mallet_model, gamma_threshold=0.001, iterations=50): model_gensim = LdaModel(id2word=mallet_model.id2word, num_topics=mallet_model.num_topics, alpha=mallet_model.alpha, iterations=iterations, gamma_threshold=gamma_threshold) model_gensim.expElogbeta[:] = mallet_model.wordtopics return model_gensim
[ "def", "malletmodel2ldamodel", "(", "mallet_model", ",", "gamma_threshold", "=", "0.001", ",", "iterations", "=", "50", ")", ":", "model_gensim", "=", "LdaModel", "(", "id2word", "=", "mallet_model", ".", "id2word", ",", "num_topics", "=", "mallet_model", ".", ...
function to convert mallet model to gensim ldamodel .
train
false
23,075
def defers(func): @wraps(func) def wrapped(*a, **kw): return defer.maybeDeferred(func, *a, **kw) return wrapped
[ "def", "defers", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "wrapped", "(", "*", "a", ",", "**", "kw", ")", ":", "return", "defer", ".", "maybeDeferred", "(", "func", ",", "*", "a", ",", "**", "kw", ")", "return", "wrapped" ]
decorator to make sure a function always returns a deferred .
train
false
23,076
def compute_edge_measures(ntwk): iflogger.info(u'Computing edge measures:') measures = {} return measures
[ "def", "compute_edge_measures", "(", "ntwk", ")", ":", "iflogger", ".", "info", "(", "u'Computing edge measures:'", ")", "measures", "=", "{", "}", "return", "measures" ]
these return edge-based measures .
train
false
23,077
def install_overlay(module, name, list_url=None): layman_conf = BareConfig(read_configfile=True) layman = init_layman(layman_conf) if layman.is_installed(name): return False if module.check_mode: mymsg = (("Would add layman repo '" + name) + "'") module.exit_json(changed=True, msg=mymsg) if (not layman.is_repo(name)): if (not list_url): raise ModuleError(("Overlay '%s' is not on the list of known overlays and URL of the remote list was not provided." % name)) overlay_defs = layman_conf.get_option('overlay_defs') dest = path.join(overlay_defs, (name + '.xml')) download_url(module, list_url, dest) layman = init_layman() if (not layman.add_repos(name)): raise ModuleError(layman.get_errors()) return True
[ "def", "install_overlay", "(", "module", ",", "name", ",", "list_url", "=", "None", ")", ":", "layman_conf", "=", "BareConfig", "(", "read_configfile", "=", "True", ")", "layman", "=", "init_layman", "(", "layman_conf", ")", "if", "layman", ".", "is_installe...
installs the overlay repository .
train
false
23,081
def errcheck(ret, func, args): if (ret == (-1)): try: ec = ctypes.get_errno() raise LabJackException(ec, ('Exodriver returned error number %s' % ec)) except AttributeError: raise LabJackException((-1), 'Exodriver returned an error, but LabJackPython is unable to read the error code. Upgrade to Python 2.6 for this functionality.') else: return ret
[ "def", "errcheck", "(", "ret", ",", "func", ",", "args", ")", ":", "if", "(", "ret", "==", "(", "-", "1", ")", ")", ":", "try", ":", "ec", "=", "ctypes", ".", "get_errno", "(", ")", "raise", "LabJackException", "(", "ec", ",", "(", "'Exodriver re...
whenever a function is called through ctypes .
train
false
23,082
def _ipset_cmd(): return salt.utils.which('ipset')
[ "def", "_ipset_cmd", "(", ")", ":", "return", "salt", ".", "utils", ".", "which", "(", "'ipset'", ")" ]
return correct command .
train
false
23,083
def visit_hcb_html(self, node): global HCB_COUNTER HCB_COUNTER += 1 try: self.visit_literal_block(node) except nodes.SkipNode: pass code_block = self.body[(-1)] fill_header = {'divname': 'hiddencodeblock{0}'.format(HCB_COUNTER), 'startdisplay': ('none' if node['starthidden'] else 'block'), 'label': node.get('label')} divheader = '<p><strong><a href="javascript:showhide(document.getElementById(\'{divname}\'))">{label}</a></strong></p><div id="{divname}" style="display: {startdisplay}">'.format(**fill_header) code_block = (((js_showhide + divheader) + code_block) + '</div>') self.body[(-1)] = code_block raise nodes.SkipNode
[ "def", "visit_hcb_html", "(", "self", ",", "node", ")", ":", "global", "HCB_COUNTER", "HCB_COUNTER", "+=", "1", "try", ":", "self", ".", "visit_literal_block", "(", "node", ")", "except", "nodes", ".", "SkipNode", ":", "pass", "code_block", "=", "self", "....
visit hidden code block .
train
false
23,084
def feature_hidden(feature_name): return ((feature_name is not None) and (feature_name in settings.OSCAR_HIDDEN_FEATURES))
[ "def", "feature_hidden", "(", "feature_name", ")", ":", "return", "(", "(", "feature_name", "is", "not", "None", ")", "and", "(", "feature_name", "in", "settings", ".", "OSCAR_HIDDEN_FEATURES", ")", ")" ]
test if a certain oscar feature is disabled .
train
false
23,085
def is_process_64_from_handle(hProcess): iswow64 = c_bool(False) if (IsWow64Process is None): return False if (not IsWow64Process(hProcess, byref(iswow64))): raise WinError() return (not iswow64.value)
[ "def", "is_process_64_from_handle", "(", "hProcess", ")", ":", "iswow64", "=", "c_bool", "(", "False", ")", "if", "(", "IsWow64Process", "is", "None", ")", ":", "return", "False", "if", "(", "not", "IsWow64Process", "(", "hProcess", ",", "byref", "(", "isw...
take a process handle .
train
false
23,086
def with_worker_threads(n_threads, dbname='foo', n_jobs=sys.maxsize, timeout=10.0): def newth(ii): return threading.Thread(target=_worker_thread_fn, args=(('hostname', ii), n_jobs, timeout, dbname)) def deco(f): def wrapper(*args, **kwargs): threads = list(map(newth, list(range(n_threads)))) [th.start() for th in threads] try: return f(*args, **kwargs) finally: [th.join() for th in threads] wrapper.__name__ = f.__name__ return wrapper return deco
[ "def", "with_worker_threads", "(", "n_threads", ",", "dbname", "=", "'foo'", ",", "n_jobs", "=", "sys", ".", "maxsize", ",", "timeout", "=", "10.0", ")", ":", "def", "newth", "(", "ii", ")", ":", "return", "threading", ".", "Thread", "(", "target", "="...
decorator that will run a test with some mongoworker threads in flight .
train
false
23,087
def pretty_iban(iban): return ' '.join([iban[i:(i + 4)] for i in range(0, len(iban), 4)])
[ "def", "pretty_iban", "(", "iban", ")", ":", "return", "' '", ".", "join", "(", "[", "iban", "[", "i", ":", "(", "i", "+", "4", ")", "]", "for", "i", "in", "range", "(", "0", ",", "len", "(", "iban", ")", ",", "4", ")", "]", ")" ]
return iban in groups of four characters separated by a single space .
train
false
23,088
def read_serial(console): data_bytes = console.inWaiting() if data_bytes: return console.read(data_bytes) else: return ''
[ "def", "read_serial", "(", "console", ")", ":", "data_bytes", "=", "console", ".", "inWaiting", "(", ")", "if", "data_bytes", ":", "return", "console", ".", "read", "(", "data_bytes", ")", "else", ":", "return", "''" ]
check if there is data waiting to be read read and return it .
train
false