id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
5,662 | def StemmingAnalyzer(expression=default_pattern, stoplist=STOP_WORDS, minsize=2, maxsize=None, gaps=False, stemfn=stem, ignore=None, cachesize=50000):
ret = RegexTokenizer(expression=expression, gaps=gaps)
chain = (ret | LowercaseFilter())
if (stoplist is not None):
chain = (chain | StopFilter(stoplist=stoplist, minsize=minsize, maxsize=maxsize))
return (chain | StemFilter(stemfn=stemfn, ignore=ignore, cachesize=cachesize))
| [
"def",
"StemmingAnalyzer",
"(",
"expression",
"=",
"default_pattern",
",",
"stoplist",
"=",
"STOP_WORDS",
",",
"minsize",
"=",
"2",
",",
"maxsize",
"=",
"None",
",",
"gaps",
"=",
"False",
",",
"stemfn",
"=",
"stem",
",",
"ignore",
"=",
"None",
",",
"cach... | composes a regextokenizer with a lower case filter . | train | false |
5,663 | def write_items(lib, query, pretend, force):
(items, albums) = _do_query(lib, query, False, False)
for item in items:
if (not os.path.exists(syspath(item.path))):
log.info(u'missing file: {0}'.format(util.displayable_path(item.path)))
continue
try:
clean_item = library.Item.from_path(item.path)
except library.ReadError as exc:
log.error(u'error reading {0}: {1}'.format(displayable_path(item.path), exc))
continue
changed = ui.show_model_changes(item, clean_item, library.Item._media_fields, force)
if ((changed or force) and (not pretend)):
item.try_sync()
| [
"def",
"write_items",
"(",
"lib",
",",
"query",
",",
"pretend",
",",
"force",
")",
":",
"(",
"items",
",",
"albums",
")",
"=",
"_do_query",
"(",
"lib",
",",
"query",
",",
"False",
",",
"False",
")",
"for",
"item",
"in",
"items",
":",
"if",
"(",
"... | write tag information from the database to the respective files in the filesystem . | train | false |
5,664 | def get_bracket_regions(settings, minimap):
styles = settings.get('bracket_styles', DEFAULT_STYLES)
icon_path = 'Packages/BracketHighlighter/icons'
for (key, value) in DEFAULT_STYLES.items():
if (key not in styles):
styles[key] = value
continue
for (k, v) in value.items():
if (k not in styles[key]):
styles[key][k] = v
default_settings = styles['default']
for (k, v) in styles.items():
(yield (k, StyleDefinition(k, v, default_settings, icon_path, minimap)))
| [
"def",
"get_bracket_regions",
"(",
"settings",
",",
"minimap",
")",
":",
"styles",
"=",
"settings",
".",
"get",
"(",
"'bracket_styles'",
",",
"DEFAULT_STYLES",
")",
"icon_path",
"=",
"'Packages/BracketHighlighter/icons'",
"for",
"(",
"key",
",",
"value",
")",
"i... | get styled regions for brackets to use . | train | false |
5,667 | def CopyTool(flavor, out_path, generator_flags={}):
prefix = {'aix': 'flock', 'solaris': 'flock', 'mac': 'mac', 'win': 'win'}.get(flavor, None)
if (not prefix):
return
source_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), ('%s_tool.py' % prefix))
with open(source_path) as source_file:
source = source_file.readlines()
header = '# Generated by gyp. Do not edit.\n'
mac_toolchain_dir = generator_flags.get('mac_toolchain_dir', None)
if ((flavor == 'mac') and mac_toolchain_dir):
header += ("import os;\nos.environ['DEVELOPER_DIR']='%s'\n" % mac_toolchain_dir)
tool_path = os.path.join(out_path, ('gyp-%s-tool' % prefix))
with open(tool_path, 'w') as tool_file:
tool_file.write(''.join(([source[0], header] + source[1:])))
os.chmod(tool_path, 493)
| [
"def",
"CopyTool",
"(",
"flavor",
",",
"out_path",
",",
"generator_flags",
"=",
"{",
"}",
")",
":",
"prefix",
"=",
"{",
"'aix'",
":",
"'flock'",
",",
"'solaris'",
":",
"'flock'",
",",
"'mac'",
":",
"'mac'",
",",
"'win'",
":",
"'win'",
"}",
".",
"get"... | finds _tool . | train | false |
5,668 | def broken_view(request):
raise KeyError('Oops! Looks like you wrote some bad code.')
| [
"def",
"broken_view",
"(",
"request",
")",
":",
"raise",
"KeyError",
"(",
"'Oops! Looks like you wrote some bad code.'",
")"
] | a view which just raises an exception . | train | false |
5,669 | def _setup(text):
root = text._root()
engine = searchengine.get(root)
if (not hasattr(engine, '_searchdialog')):
engine._searchdialog = SearchDialog(root, engine)
return engine._searchdialog
| [
"def",
"_setup",
"(",
"text",
")",
":",
"root",
"=",
"text",
".",
"_root",
"(",
")",
"engine",
"=",
"searchengine",
".",
"get",
"(",
"root",
")",
"if",
"(",
"not",
"hasattr",
"(",
"engine",
",",
"'_searchdialog'",
")",
")",
":",
"engine",
".",
"_se... | prepare and global state and updates it with command line options . | train | false |
5,670 | @pytest.mark.django_db
@pytest.mark.parametrize('show_all_pages', [True, False])
def test_page_links_plugin_hide_expired(show_all_pages):
context = get_jinja_context()
page = create_page(eternal=True, visible_in_menu=True)
another_page = create_page(eternal=True, visible_in_menu=True)
plugin = PageLinksPlugin({'pages': [page.pk, another_page.pk], 'show_all_pages': show_all_pages})
assert (page in plugin.get_context_data(context)['pages'])
page.available_from = None
page.available_to = None
page.save()
assert (page in plugin.get_context_data(context)['pages'])
plugin.config['hide_expired'] = True
pages_in_context = plugin.get_context_data(context)['pages']
assert (page not in pages_in_context)
assert (another_page in pages_in_context)
| [
"@",
"pytest",
".",
"mark",
".",
"django_db",
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'show_all_pages'",
",",
"[",
"True",
",",
"False",
"]",
")",
"def",
"test_page_links_plugin_hide_expired",
"(",
"show_all_pages",
")",
":",
"context",
"=",
"g... | make sure plugin correctly filters out expired pages based on plugin configuration . | train | false |
5,671 | def find_up(l_node, f_node):
if isinstance(l_node, gof.Apply):
l_outs = l_node.outputs
else:
l_outs = l_node
l_ins = gof.graph.inputs(l_outs)
nodes = gof.graph.io_toposort(l_ins, l_outs)
return (f_node in nodes)
| [
"def",
"find_up",
"(",
"l_node",
",",
"f_node",
")",
":",
"if",
"isinstance",
"(",
"l_node",
",",
"gof",
".",
"Apply",
")",
":",
"l_outs",
"=",
"l_node",
".",
"outputs",
"else",
":",
"l_outs",
"=",
"l_node",
"l_ins",
"=",
"gof",
".",
"graph",
".",
... | goes up in the graph and returns true if a node in nodes is found . | train | false |
5,672 | @require_POST
@ensure_csrf_cookie
@cache_control(no_cache=True, no_store=True, must_revalidate=True)
@require_level('staff')
def list_instructor_tasks(request, course_id):
course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)
problem_location_str = strip_if_string(request.POST.get('problem_location_str', False))
student = request.POST.get('unique_student_identifier', None)
if (student is not None):
student = get_student_from_identifier(student)
if (student and (not problem_location_str)):
return HttpResponseBadRequest('unique_student_identifier must accompany problem_location_str')
if problem_location_str:
try:
module_state_key = course_id.make_usage_key_from_deprecated_string(problem_location_str)
except InvalidKeyError:
return HttpResponseBadRequest()
if student:
tasks = lms.djangoapps.instructor_task.api.get_instructor_task_history(course_id, module_state_key, student)
else:
tasks = lms.djangoapps.instructor_task.api.get_instructor_task_history(course_id, module_state_key)
else:
tasks = lms.djangoapps.instructor_task.api.get_running_instructor_tasks(course_id)
response_payload = {'tasks': map(extract_task_features, tasks)}
return JsonResponse(response_payload)
| [
"@",
"require_POST",
"@",
"ensure_csrf_cookie",
"@",
"cache_control",
"(",
"no_cache",
"=",
"True",
",",
"no_store",
"=",
"True",
",",
"must_revalidate",
"=",
"True",
")",
"@",
"require_level",
"(",
"'staff'",
")",
"def",
"list_instructor_tasks",
"(",
"request",... | list instructor tasks . | train | false |
5,673 | def node_redundancy(G, nodes=None):
if (nodes is None):
nodes = G
if any(((len(G[v]) < 2) for v in nodes)):
raise NetworkXError('Cannot compute redundancy coefficient for a node that has fewer than two neighbors.')
return {v: _node_redundancy(G, v) for v in nodes}
| [
"def",
"node_redundancy",
"(",
"G",
",",
"nodes",
"=",
"None",
")",
":",
"if",
"(",
"nodes",
"is",
"None",
")",
":",
"nodes",
"=",
"G",
"if",
"any",
"(",
"(",
"(",
"len",
"(",
"G",
"[",
"v",
"]",
")",
"<",
"2",
")",
"for",
"v",
"in",
"nodes... | computes the node redundancy coefficients for the nodes in the bipartite graph g . | train | false |
5,674 | def task_cli_pkg_install(distribution, package_source=PackageSource()):
commands = task_package_install('clusterhq-flocker-cli', distribution, package_source)
return sequence([(Effect(Sudo(command=e.intent.command, log_command_filter=e.intent.log_command_filter)) if isinstance(e.intent, Run) else e) for e in commands.intent.effects])
| [
"def",
"task_cli_pkg_install",
"(",
"distribution",
",",
"package_source",
"=",
"PackageSource",
"(",
")",
")",
":",
"commands",
"=",
"task_package_install",
"(",
"'clusterhq-flocker-cli'",
",",
"distribution",
",",
"package_source",
")",
"return",
"sequence",
"(",
... | install the flocker cli package . | train | false |
5,675 | def pick_types_forward(orig, meg=True, eeg=False, ref_meg=True, seeg=False, ecog=False, include=[], exclude=[]):
info = orig['info']
sel = pick_types(info, meg, eeg, ref_meg=ref_meg, seeg=seeg, ecog=ecog, include=include, exclude=exclude)
if (len(sel) == 0):
raise ValueError('No valid channels found')
include_ch_names = [info['ch_names'][k] for k in sel]
return pick_channels_forward(orig, include_ch_names)
| [
"def",
"pick_types_forward",
"(",
"orig",
",",
"meg",
"=",
"True",
",",
"eeg",
"=",
"False",
",",
"ref_meg",
"=",
"True",
",",
"seeg",
"=",
"False",
",",
"ecog",
"=",
"False",
",",
"include",
"=",
"[",
"]",
",",
"exclude",
"=",
"[",
"]",
")",
":"... | pick by channel type and names from a forward operator . | train | false |
5,676 | def delete_role_policy(role_name, policy_name, region=None, key=None, keyid=None, profile=None):
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
_policy = get_role_policy(role_name, policy_name, region, key, keyid, profile)
if (not _policy):
return True
try:
conn.delete_role_policy(role_name, policy_name)
msg = 'Successfully deleted {0} policy for role {1}.'
log.info(msg.format(policy_name, role_name))
return True
except boto.exception.BotoServerError as e:
log.debug(e)
msg = 'Failed to delete {0} policy for role {1}.'
log.error(msg.format(policy_name, role_name))
return False
| [
"def",
"delete_role_policy",
"(",
"role_name",
",",
"policy_name",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key"... | delete a role policy . | train | true |
5,677 | @app.route('/cookies/delete')
def delete_cookies():
cookies = dict(request.args.items())
r = app.make_response(redirect(url_for('view_cookies')))
for (key, value) in cookies.items():
r.delete_cookie(key=key)
return r
| [
"@",
"app",
".",
"route",
"(",
"'/cookies/delete'",
")",
"def",
"delete_cookies",
"(",
")",
":",
"cookies",
"=",
"dict",
"(",
"request",
".",
"args",
".",
"items",
"(",
")",
")",
"r",
"=",
"app",
".",
"make_response",
"(",
"redirect",
"(",
"url_for",
... | deletes cookie(s) as provided by the query string and redirects to cookie list . | train | true |
5,678 | def _pretty_frame_relation_type(freltyp):
outstr = u'<frame relation type ({0.ID}): {0.superFrameName} -- {0.name} -> {0.subFrameName}>'.format(freltyp)
return outstr
| [
"def",
"_pretty_frame_relation_type",
"(",
"freltyp",
")",
":",
"outstr",
"=",
"u'<frame relation type ({0.ID}): {0.superFrameName} -- {0.name} -> {0.subFrameName}>'",
".",
"format",
"(",
"freltyp",
")",
"return",
"outstr"
] | helper function for pretty-printing a frame relation type . | train | false |
5,680 | def alter_subprocess_kwargs_by_platform(**kwargs):
kwargs.setdefault('close_fds', (os.name == 'posix'))
if (os.name == 'nt'):
CONSOLE_CREATION_FLAGS = 0
CREATE_NO_WINDOW = 134217728
CONSOLE_CREATION_FLAGS |= CREATE_NO_WINDOW
kwargs.setdefault('creationflags', CONSOLE_CREATION_FLAGS)
return kwargs
| [
"def",
"alter_subprocess_kwargs_by_platform",
"(",
"**",
"kwargs",
")",
":",
"kwargs",
".",
"setdefault",
"(",
"'close_fds'",
",",
"(",
"os",
".",
"name",
"==",
"'posix'",
")",
")",
"if",
"(",
"os",
".",
"name",
"==",
"'nt'",
")",
":",
"CONSOLE_CREATION_FL... | given a dict . | train | true |
5,684 | def norm_lls_grad(y, params):
(mu, sigma2) = params.T
dllsdmu = ((y - mu) / sigma2)
dllsdsigma2 = (((((y - mu) ** 2) / sigma2) - 1) / np.sqrt(sigma2))
return np.column_stack((dllsdmu, dllsdsigma2))
| [
"def",
"norm_lls_grad",
"(",
"y",
",",
"params",
")",
":",
"(",
"mu",
",",
"sigma2",
")",
"=",
"params",
".",
"T",
"dllsdmu",
"=",
"(",
"(",
"y",
"-",
"mu",
")",
"/",
"sigma2",
")",
"dllsdsigma2",
"=",
"(",
"(",
"(",
"(",
"(",
"y",
"-",
"mu",... | jacobian of normal loglikelihood wrt mean mu and variance sigma2 parameters y : array . | train | false |
5,685 | def decode_unicode_obj(obj):
if isinstance(obj, dict):
r = {}
for (k, v) in iteritems(obj):
r[decode_unicode_string(k)] = decode_unicode_obj(v)
return r
elif isinstance(obj, six.string_types):
return decode_unicode_string(obj)
elif isinstance(obj, (list, tuple)):
return [decode_unicode_obj(x) for x in obj]
else:
return obj
| [
"def",
"decode_unicode_obj",
"(",
"obj",
")",
":",
"if",
"isinstance",
"(",
"obj",
",",
"dict",
")",
":",
"r",
"=",
"{",
"}",
"for",
"(",
"k",
",",
"v",
")",
"in",
"iteritems",
"(",
"obj",
")",
":",
"r",
"[",
"decode_unicode_string",
"(",
"k",
")... | decode unicoded dict/list/tuple encoded by unicode_obj . | train | true |
5,686 | def approximate_taylor_polynomial(f, x, degree, scale, order=None):
if (order is None):
order = degree
n = (order + 1)
xs = ((scale * np.cos(np.linspace(0, np.pi, n, endpoint=(n % 1)))) + x)
P = KroghInterpolator(xs, f(xs))
d = P.derivatives(x, der=(degree + 1))
return np.poly1d((d / factorial(np.arange((degree + 1))))[::(-1)])
| [
"def",
"approximate_taylor_polynomial",
"(",
"f",
",",
"x",
",",
"degree",
",",
"scale",
",",
"order",
"=",
"None",
")",
":",
"if",
"(",
"order",
"is",
"None",
")",
":",
"order",
"=",
"degree",
"n",
"=",
"(",
"order",
"+",
"1",
")",
"xs",
"=",
"(... | estimate the taylor polynomial of f at x by polynomial fitting . | train | false |
5,687 | def aslinearoperator(A):
if isinstance(A, LinearOperator):
return A
elif (isinstance(A, np.ndarray) or isinstance(A, np.matrix)):
if (A.ndim > 2):
raise ValueError('array must have ndim <= 2')
A = np.atleast_2d(np.asarray(A))
return MatrixLinearOperator(A)
elif isspmatrix(A):
return MatrixLinearOperator(A)
elif (hasattr(A, 'shape') and hasattr(A, 'matvec')):
rmatvec = None
dtype = None
if hasattr(A, 'rmatvec'):
rmatvec = A.rmatvec
if hasattr(A, 'dtype'):
dtype = A.dtype
return LinearOperator(A.shape, A.matvec, rmatvec=rmatvec, dtype=dtype)
else:
raise TypeError('type not understood')
| [
"def",
"aslinearoperator",
"(",
"A",
")",
":",
"if",
"isinstance",
"(",
"A",
",",
"LinearOperator",
")",
":",
"return",
"A",
"elif",
"(",
"isinstance",
"(",
"A",
",",
"np",
".",
"ndarray",
")",
"or",
"isinstance",
"(",
"A",
",",
"np",
".",
"matrix",
... | return a as a linearoperator . | train | false |
5,688 | def _process_text_args(override, fontdict=None, **kwargs):
if (fontdict is not None):
override.update(fontdict)
override.update(kwargs)
return override
| [
"def",
"_process_text_args",
"(",
"override",
",",
"fontdict",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"fontdict",
"is",
"not",
"None",
")",
":",
"override",
".",
"update",
"(",
"fontdict",
")",
"override",
".",
"update",
"(",
"kwargs",
... | return an override dict . | train | false |
5,689 | def clear_time_override():
utcnow.override_time = None
| [
"def",
"clear_time_override",
"(",
")",
":",
"utcnow",
".",
"override_time",
"=",
"None"
] | remove the overridden time . | train | false |
5,691 | def find_executable_linenos(filename):
try:
prog = open(filename, 'rU').read()
except IOError as err:
print >>sys.stderr, ('Not printing coverage data for %r: %s' % (filename, err))
return {}
code = compile(prog, filename, 'exec')
strs = find_strings(filename)
return find_lines(code, strs)
| [
"def",
"find_executable_linenos",
"(",
"filename",
")",
":",
"try",
":",
"prog",
"=",
"open",
"(",
"filename",
",",
"'rU'",
")",
".",
"read",
"(",
")",
"except",
"IOError",
"as",
"err",
":",
"print",
">>",
"sys",
".",
"stderr",
",",
"(",
"'Not printing... | return dict where keys are line numbers in the line number table . | train | false |
5,692 | @ensure_csrf_cookie
@cache_control(no_cache=True, no_store=True, must_revalidate=True)
@require_level('staff')
@require_POST
def get_registration_codes(request, course_id):
course_id = SlashSeparatedCourseKey.from_deprecated_string(course_id)
registration_codes = CourseRegistrationCode.objects.filter(course_id=course_id).order_by('invoice_item__invoice__company_name')
company_name = request.POST['download_company_name']
if company_name:
registration_codes = registration_codes.filter(invoice_item__invoice__company_name=company_name)
csv_type = 'download'
return registration_codes_csv('Registration_Codes.csv', registration_codes, csv_type)
| [
"@",
"ensure_csrf_cookie",
"@",
"cache_control",
"(",
"no_cache",
"=",
"True",
",",
"no_store",
"=",
"True",
",",
"must_revalidate",
"=",
"True",
")",
"@",
"require_level",
"(",
"'staff'",
")",
"@",
"require_POST",
"def",
"get_registration_codes",
"(",
"request"... | respond with csv which contains a summary of all registration codes . | train | false |
5,693 | def version_is_compatible(imp_version, version):
version_parts = version.split('.')
imp_version_parts = imp_version.split('.')
if (int(version_parts[0]) != int(imp_version_parts[0])):
return False
if (int(version_parts[1]) > int(imp_version_parts[1])):
return False
return True
| [
"def",
"version_is_compatible",
"(",
"imp_version",
",",
"version",
")",
":",
"version_parts",
"=",
"version",
".",
"split",
"(",
"'.'",
")",
"imp_version_parts",
"=",
"imp_version",
".",
"split",
"(",
"'.'",
")",
"if",
"(",
"int",
"(",
"version_parts",
"[",... | determine whether versions are compatible . | train | false |
5,694 | def sanitize_redirect(host, redirect_to):
if redirect_to:
try:
netloc = (urlparse(redirect_to)[1] or host)
except (TypeError, AttributeError):
pass
else:
if (netloc == host):
return redirect_to
| [
"def",
"sanitize_redirect",
"(",
"host",
",",
"redirect_to",
")",
":",
"if",
"redirect_to",
":",
"try",
":",
"netloc",
"=",
"(",
"urlparse",
"(",
"redirect_to",
")",
"[",
"1",
"]",
"or",
"host",
")",
"except",
"(",
"TypeError",
",",
"AttributeError",
")"... | given the hostname and an untrusted url to redirect to . | train | false |
5,696 | def get_collection_by_alias(bus, alias):
service_obj = bus_get_object(bus, SS_PATH)
service_iface = dbus.Interface(service_obj, SERVICE_IFACE)
collection_path = service_iface.ReadAlias(alias, signature='s')
if (len(collection_path) <= 1):
raise ItemNotFoundException('No collection with such alias.')
return Collection(bus, collection_path)
| [
"def",
"get_collection_by_alias",
"(",
"bus",
",",
"alias",
")",
":",
"service_obj",
"=",
"bus_get_object",
"(",
"bus",
",",
"SS_PATH",
")",
"service_iface",
"=",
"dbus",
".",
"Interface",
"(",
"service_obj",
",",
"SERVICE_IFACE",
")",
"collection_path",
"=",
... | returns the collection with the given alias . | train | false |
5,698 | def LocalGroupEnum():
resume = 0
nmembers = 0
while 1:
(data, total, resume) = win32net.NetLocalGroupEnum(server, 1, resume)
for group in data:
verbose(('Found group %(name)s:%(comment)s ' % group))
memberresume = 0
while 1:
(memberdata, total, memberresume) = win32net.NetLocalGroupGetMembers(server, group['name'], 2, resume)
for member in memberdata:
(username, domain, type) = win32security.LookupAccountSid(server, member['sid'])
nmembers = (nmembers + 1)
verbose((' Member %s (%s)' % (username, member['domainandname'])))
if (memberresume == 0):
break
if (not resume):
break
assert nmembers, 'Couldnt find a single member in a single group!'
print 'Enumerated all the local groups'
| [
"def",
"LocalGroupEnum",
"(",
")",
":",
"resume",
"=",
"0",
"nmembers",
"=",
"0",
"while",
"1",
":",
"(",
"data",
",",
"total",
",",
"resume",
")",
"=",
"win32net",
".",
"NetLocalGroupEnum",
"(",
"server",
",",
"1",
",",
"resume",
")",
"for",
"group"... | enumerates all the local groups . | train | false |
5,699 | def quantity_allclose(a, b, rtol=1e-05, atol=None, **kwargs):
import numpy as np
return np.allclose(*_unquantify_allclose_arguments(a, b, rtol, atol), **kwargs)
| [
"def",
"quantity_allclose",
"(",
"a",
",",
"b",
",",
"rtol",
"=",
"1e-05",
",",
"atol",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"import",
"numpy",
"as",
"np",
"return",
"np",
".",
"allclose",
"(",
"*",
"_unquantify_allclose_arguments",
"(",
"a",
",... | returns true if two arrays are element-wise equal within a tolerance . | train | false |
5,700 | @command('(mv|sw)\\s*(\\d{1,4})\\s*[\\s,]\\s*(\\d{1,4})')
def songlist_mv_sw(action, a, b):
(i, j) = ((int(a) - 1), (int(b) - 1))
if (action == 'mv'):
g.model.songs.insert(j, g.model.songs.pop(i))
g.message = (util.F('song move') % (g.model[j].title, b))
elif (action == 'sw'):
(g.model[i], g.model[j]) = (g.model[j], g.model[i])
g.message = (util.F('song sw') % (min(a, b), max(a, b)))
g.content = content.generate_songlist_display()
| [
"@",
"command",
"(",
"'(mv|sw)\\\\s*(\\\\d{1,4})\\\\s*[\\\\s,]\\\\s*(\\\\d{1,4})'",
")",
"def",
"songlist_mv_sw",
"(",
"action",
",",
"a",
",",
"b",
")",
":",
"(",
"i",
",",
"j",
")",
"=",
"(",
"(",
"int",
"(",
"a",
")",
"-",
"1",
")",
",",
"(",
"int",... | move a song or swap two songs . | train | false |
5,701 | def _RemoveRedundantHandlers(handler_list):
no_duplicates = []
patterns_found_so_far = set()
for i in xrange(len(handler_list)):
current_handler = handler_list[i]
matched_by_later = False
for j in xrange((i + 1), len(handler_list)):
if current_handler.IsFullyHandledBy(handler_list[j]):
matched_by_later = True
break
if ((not matched_by_later) and (current_handler.pattern not in patterns_found_so_far)):
no_duplicates.append(current_handler)
patterns_found_so_far.add(current_handler.pattern)
return no_duplicates
| [
"def",
"_RemoveRedundantHandlers",
"(",
"handler_list",
")",
":",
"no_duplicates",
"=",
"[",
"]",
"patterns_found_so_far",
"=",
"set",
"(",
")",
"for",
"i",
"in",
"xrange",
"(",
"len",
"(",
"handler_list",
")",
")",
":",
"current_handler",
"=",
"handler_list",... | removes duplicated or unnecessary handlers from the list . | train | false |
5,702 | def remove_content_type(app_label, model_name):
try:
ct = ContentType.objects.get(app_label=app_label, model=model_name.lower())
ct.delete()
except ContentType.DoesNotExist:
pass
| [
"def",
"remove_content_type",
"(",
"app_label",
",",
"model_name",
")",
":",
"try",
":",
"ct",
"=",
"ContentType",
".",
"objects",
".",
"get",
"(",
"app_label",
"=",
"app_label",
",",
"model",
"=",
"model_name",
".",
"lower",
"(",
")",
")",
"ct",
".",
... | delete from the django contenttype table . | train | false |
5,705 | def stub_verify(conn, cert, errno, errdepth, code):
return True
| [
"def",
"stub_verify",
"(",
"conn",
",",
"cert",
",",
"errno",
",",
"errdepth",
",",
"code",
")",
":",
"return",
"True"
] | we dont verify the server when we attempt a mitm . | train | false |
5,707 | def server_message(response):
message = None
if (response.headers.get('content-type') and ('text/html' in response.headers.get('content-type'))):
try:
soup = BeautifulSoup(response.content, 'html5lib')
except Exception:
pass
message = soup.find('body')
elements = ('header', 'script', 'footer', 'nav', 'input', 'textarea')
for element in elements:
for tag in soup.find_all(element):
tag.replaceWith('')
message = (message.text if message else soup.text)
message = message.strip()
if (message is None):
message = response.content.strip()
if message:
if (len(message) > 150):
message = (message[:150] + '...')
logger.debug('Server responded with message: %s', message)
| [
"def",
"server_message",
"(",
"response",
")",
":",
"message",
"=",
"None",
"if",
"(",
"response",
".",
"headers",
".",
"get",
"(",
"'content-type'",
")",
"and",
"(",
"'text/html'",
"in",
"response",
".",
"headers",
".",
"get",
"(",
"'content-type'",
")",
... | extract server message from response and log in to logger with debug level . | train | false |
5,708 | def exceptionFromStanza(stanza):
children = []
condition = text = textLang = appCondition = type = code = None
for element in stanza.elements():
if ((element.name == 'error') and (element.uri == stanza.uri)):
code = element.getAttribute('code')
type = element.getAttribute('type')
error = _parseError(element, NS_XMPP_STANZAS)
condition = error['condition']
text = error['text']
textLang = error['textLang']
appCondition = error['appCondition']
if ((not condition) and code):
(condition, type) = CODES_TO_CONDITIONS[code]
text = _getText(stanza.error)
else:
children.append(element)
if (condition is None):
return StanzaError(None)
exception = StanzaError(condition, type, text, textLang, appCondition)
exception.children = children
exception.stanza = stanza
return exception
| [
"def",
"exceptionFromStanza",
"(",
"stanza",
")",
":",
"children",
"=",
"[",
"]",
"condition",
"=",
"text",
"=",
"textLang",
"=",
"appCondition",
"=",
"type",
"=",
"code",
"=",
"None",
"for",
"element",
"in",
"stanza",
".",
"elements",
"(",
")",
":",
"... | build an exception object from an error stanza . | train | false |
5,709 | def batch_shuffle(index_array, batch_size):
batch_count = int((len(index_array) / batch_size))
last_batch = index_array[(batch_count * batch_size):]
index_array = index_array[:(batch_count * batch_size)]
index_array = index_array.reshape((batch_count, batch_size))
np.random.shuffle(index_array)
index_array = index_array.flatten()
return np.append(index_array, last_batch)
| [
"def",
"batch_shuffle",
"(",
"index_array",
",",
"batch_size",
")",
":",
"batch_count",
"=",
"int",
"(",
"(",
"len",
"(",
"index_array",
")",
"/",
"batch_size",
")",
")",
"last_batch",
"=",
"index_array",
"[",
"(",
"batch_count",
"*",
"batch_size",
")",
":... | this shuffles an array in a batch-wise fashion . | train | false |
5,710 | def _cond(condition, then_lambda, else_lambda):
try:
cond_fn = tf.cond
except AttributeError:
from tensorflow.python.ops import control_flow_ops
cond_fn = control_flow_ops.cond
return cond_fn(condition, then_lambda, else_lambda)
| [
"def",
"_cond",
"(",
"condition",
",",
"then_lambda",
",",
"else_lambda",
")",
":",
"try",
":",
"cond_fn",
"=",
"tf",
".",
"cond",
"except",
"AttributeError",
":",
"from",
"tensorflow",
".",
"python",
".",
"ops",
"import",
"control_flow_ops",
"cond_fn",
"=",... | backwards compatible interface to tf . | train | false |
5,711 | def hrm_training_job_title(row):
try:
person_id = row.hrm_training.person_id
except AttributeError:
person_id = None
if person_id:
s3db = current.s3db
table = s3db.hrm_human_resource
jtable = s3db.hrm_job_title
query = (((table.person_id == person_id) & (table.status != 2)) & (table.job_title_id == jtable.id))
jobs = current.db(query).select(jtable.name, distinct=True, orderby=jtable.name)
if jobs:
output = ''
for job in jobs:
repr = job.name
if output:
output = ('%s, %s' % (output, repr))
else:
output = repr
return output
return current.messages['NONE']
| [
"def",
"hrm_training_job_title",
"(",
"row",
")",
":",
"try",
":",
"person_id",
"=",
"row",
".",
"hrm_training",
".",
"person_id",
"except",
"AttributeError",
":",
"person_id",
"=",
"None",
"if",
"person_id",
":",
"s3db",
"=",
"current",
".",
"s3db",
"table"... | which job titles(s) the person is active with . | train | false |
5,713 | def print_queries():
for query in connection.queries:
print (query['sql'] + ';\n')
| [
"def",
"print_queries",
"(",
")",
":",
"for",
"query",
"in",
"connection",
".",
"queries",
":",
"print",
"(",
"query",
"[",
"'sql'",
"]",
"+",
"';\\n'",
")"
] | print all sql queries executed so far . | train | false |
5,714 | def load_host_keys(filename):
from paramiko.hostkeys import HostKeys
return HostKeys(filename)
| [
"def",
"load_host_keys",
"(",
"filename",
")",
":",
"from",
"paramiko",
".",
"hostkeys",
"import",
"HostKeys",
"return",
"HostKeys",
"(",
"filename",
")"
] | read a file of known ssh host keys . | train | false |
5,716 | def cas(key, value, old_value):
store = load()
if (key not in store):
return False
if (store[key] != old_value):
return False
store[key] = value
dump(store)
return True
| [
"def",
"cas",
"(",
"key",
",",
"value",
",",
"old_value",
")",
":",
"store",
"=",
"load",
"(",
")",
"if",
"(",
"key",
"not",
"in",
"store",
")",
":",
"return",
"False",
"if",
"(",
"store",
"[",
"key",
"]",
"!=",
"old_value",
")",
":",
"return",
... | check and set a value in the minion datastore cli example: . | train | true |
5,717 | def find_lines_from_code(code, strs):
linenos = {}
for (_, lineno) in dis.findlinestarts(code):
if (lineno not in strs):
linenos[lineno] = 1
return linenos
| [
"def",
"find_lines_from_code",
"(",
"code",
",",
"strs",
")",
":",
"linenos",
"=",
"{",
"}",
"for",
"(",
"_",
",",
"lineno",
")",
"in",
"dis",
".",
"findlinestarts",
"(",
"code",
")",
":",
"if",
"(",
"lineno",
"not",
"in",
"strs",
")",
":",
"lineno... | return dict where keys are lines in the line number table . | train | false |
5,718 | def report_config_interaction(modified, modifiers):
if isinstance(modified, str):
modified = (modified,)
if isinstance(modifiers, str):
modifiers = (modifiers,)
for var in modified:
VAR_MODIFIERS.setdefault(var, set()).update(modifiers)
| [
"def",
"report_config_interaction",
"(",
"modified",
",",
"modifiers",
")",
":",
"if",
"isinstance",
"(",
"modified",
",",
"str",
")",
":",
"modified",
"=",
"(",
"modified",
",",
")",
"if",
"isinstance",
"(",
"modifiers",
",",
"str",
")",
":",
"modifiers",... | registers config option interaction to be checked by set_by_cli . | train | false |
5,719 | def _get_echo_exe_path():
if (sys.platform == 'win32'):
return os.path.join(utils.abs_datapath(), 'userscripts', 'echo.bat')
else:
return 'echo'
| [
"def",
"_get_echo_exe_path",
"(",
")",
":",
"if",
"(",
"sys",
".",
"platform",
"==",
"'win32'",
")",
":",
"return",
"os",
".",
"path",
".",
"join",
"(",
"utils",
".",
"abs_datapath",
"(",
")",
",",
"'userscripts'",
",",
"'echo.bat'",
")",
"else",
":",
... | return the path to an echo-like command . | train | false |
5,720 | def examplesquickTest(vm, prompt=Prompt):
installPexpect(vm, prompt)
vm.sendline('sudo -n python ~/mininet/examples/test/runner.py -v -quick')
| [
"def",
"examplesquickTest",
"(",
"vm",
",",
"prompt",
"=",
"Prompt",
")",
":",
"installPexpect",
"(",
"vm",
",",
"prompt",
")",
"vm",
".",
"sendline",
"(",
"'sudo -n python ~/mininet/examples/test/runner.py -v -quick'",
")"
] | quick test of mininet examples . | train | false |
5,723 | def posts(request, forum_slug, thread_id, form=None, post_preview=None, is_reply=False):
thread = get_object_or_404(Thread, pk=thread_id)
forum = thread.forum
if ((forum.slug != forum_slug) and (not is_reply)):
new_forum = get_object_or_404(Forum, slug=forum_slug)
if new_forum.allows_viewing_by(request.user):
return HttpResponseRedirect(thread.get_absolute_url())
raise Http404
elif (forum.slug != forum_slug):
raise Http404
if (not forum.allows_viewing_by(request.user)):
raise Http404
posts_ = thread.post_set.all()
count = posts_.count()
if count:
last_post = posts_[(count - 1)]
else:
last_post = None
posts_ = posts_.select_related('author', 'updated_by')
posts_ = posts_.extra(select={'author_post_count': 'SELECT COUNT(*) FROM forums_post WHERE forums_post.author_id = auth_user.id'})
posts_ = paginate(request, posts_, constants.POSTS_PER_PAGE, count=count)
if (not form):
form = ReplyForm()
feed_urls = ((reverse('forums.posts.feed', kwargs={'forum_slug': forum_slug, 'thread_id': thread_id}), PostsFeed().title(thread)),)
is_watching_thread = (request.user.is_authenticated() and NewPostEvent.is_notifying(request.user, thread))
return render(request, 'forums/posts.html', {'forum': forum, 'thread': thread, 'posts': posts_, 'form': form, 'count': count, 'last_post': last_post, 'post_preview': post_preview, 'is_watching_thread': is_watching_thread, 'feeds': feed_urls, 'forums': Forum.objects.all()})
| [
"def",
"posts",
"(",
"request",
",",
"forum_slug",
",",
"thread_id",
",",
"form",
"=",
"None",
",",
"post_preview",
"=",
"None",
",",
"is_reply",
"=",
"False",
")",
":",
"thread",
"=",
"get_object_or_404",
"(",
"Thread",
",",
"pk",
"=",
"thread_id",
")",... | view all the posts in a thread . | train | false |
5,725 | def reduce(fn, sequences, outputs_info, non_sequences=None, go_backwards=False, mode=None, name=None):
rval = scan(fn=fn, sequences=sequences, outputs_info=outputs_info, non_sequences=non_sequences, go_backwards=go_backwards, truncate_gradient=(-1), mode=mode, name=name)
if isinstance(rval[0], (list, tuple)):
return ([x[(-1)] for x in rval[0]], rval[1])
else:
return (rval[0][(-1)], rval[1])
| [
"def",
"reduce",
"(",
"fn",
",",
"sequences",
",",
"outputs_info",
",",
"non_sequences",
"=",
"None",
",",
"go_backwards",
"=",
"False",
",",
"mode",
"=",
"None",
",",
"name",
"=",
"None",
")",
":",
"rval",
"=",
"scan",
"(",
"fn",
"=",
"fn",
",",
"... | similar behaviour as pythons reduce . | train | false |
5,726 | def get_collections():
return _do_mb_query('collection', '')
| [
"def",
"get_collections",
"(",
")",
":",
"return",
"_do_mb_query",
"(",
"'collection'",
",",
"''",
")"
] | list the collections for the currently :func:authenticated <auth> user as a dict with a collection-list key . | train | false |
5,729 | def test_negative_and_float__sparktext():
'Test negative values'
chart = Line()
chart.add('_', [0.1, 0.2, 0.9, (-0.5)])
assert (chart.render_sparktext() == u('\xe2\x96\x81\xe2\x96\x82\xe2\x96\x88\xe2\x96\x81'))
| [
"def",
"test_negative_and_float__sparktext",
"(",
")",
":",
"chart",
"=",
"Line",
"(",
")",
"chart",
".",
"add",
"(",
"'_'",
",",
"[",
"0.1",
",",
"0.2",
",",
"0.9",
",",
"(",
"-",
"0.5",
")",
"]",
")",
"assert",
"(",
"chart",
".",
"render_sparktext"... | test negative values . | train | false |
5,730 | def pressure_network(flow_rates, Qtot, k):
P = (k * (flow_rates ** 2))
F = np.hstack(((P[1:] - P[0]), (flow_rates.sum() - Qtot)))
return F
| [
"def",
"pressure_network",
"(",
"flow_rates",
",",
"Qtot",
",",
"k",
")",
":",
"P",
"=",
"(",
"k",
"*",
"(",
"flow_rates",
"**",
"2",
")",
")",
"F",
"=",
"np",
".",
"hstack",
"(",
"(",
"(",
"P",
"[",
"1",
":",
"]",
"-",
"P",
"[",
"0",
"]",
... | evaluate non-linear equation system representing the pressures and flows in a system of n parallel pipes:: f_i = p_i - p_0 . | train | false |
5,731 | def get_tensor_children_placeholders(tensor):
placeholders_list = []
if (tensor.op.type == 'Placeholder'):
placeholders_list.append(tensor)
if tensor.op:
for t in tensor.op.outputs:
if (not ('read:0' in t.name)):
placeholders_list += get_tensor_children_placeholders(t)
return list(set(placeholders_list))
| [
"def",
"get_tensor_children_placeholders",
"(",
"tensor",
")",
":",
"placeholders_list",
"=",
"[",
"]",
"if",
"(",
"tensor",
".",
"op",
".",
"type",
"==",
"'Placeholder'",
")",
":",
"placeholders_list",
".",
"append",
"(",
"tensor",
")",
"if",
"tensor",
".",... | get all placeholders that is depending the given tensor . | train | false |
5,732 | def change_settings(new_settings={}, file=None):
gl = globals()
if (file is not None):
execfile(file)
gl.update(locals())
gl.update(new_settings)
| [
"def",
"change_settings",
"(",
"new_settings",
"=",
"{",
"}",
",",
"file",
"=",
"None",
")",
":",
"gl",
"=",
"globals",
"(",
")",
"if",
"(",
"file",
"is",
"not",
"None",
")",
":",
"execfile",
"(",
"file",
")",
"gl",
".",
"update",
"(",
"locals",
... | changes the value of configuration variables . | train | false |
5,733 | def getfullnameof(mod, xtrapath=None):
from distutils.sysconfig import get_python_lib
numpy_core_paths = [os.path.join(get_python_lib(), 'numpy', 'core')]
if is_venv:
numpy_core_paths.append(os.path.join(base_prefix, 'Lib', 'site-packages', 'numpy', 'core'))
epath = ((sys.path + numpy_core_paths) + winutils.get_system_path())
if (xtrapath is not None):
if (type(xtrapath) == type('')):
epath.insert(0, xtrapath)
else:
epath = (xtrapath + epath)
for p in epath:
npth = os.path.join(p, mod)
if (os.path.exists(npth) and matchDLLArch(npth)):
return npth
for p in epath:
npth = os.path.join(p, mod.lower())
if (os.path.exists(npth) and matchDLLArch(npth)):
return npth
return ''
| [
"def",
"getfullnameof",
"(",
"mod",
",",
"xtrapath",
"=",
"None",
")",
":",
"from",
"distutils",
".",
"sysconfig",
"import",
"get_python_lib",
"numpy_core_paths",
"=",
"[",
"os",
".",
"path",
".",
"join",
"(",
"get_python_lib",
"(",
")",
",",
"'numpy'",
",... | return the full path name of mod . | train | false |
5,734 | @register.simple_tag
def simple_unlimited_args_kwargs(one, two='hi', *args, **kwargs):
sorted_kwarg = sorted(kwargs.iteritems(), key=operator.itemgetter(0))
return ('simple_unlimited_args_kwargs - Expected result: %s / %s' % (', '.join([unicode(arg) for arg in ([one, two] + list(args))]), ', '.join([('%s=%s' % (k, v)) for (k, v) in sorted_kwarg])))
| [
"@",
"register",
".",
"simple_tag",
"def",
"simple_unlimited_args_kwargs",
"(",
"one",
",",
"two",
"=",
"'hi'",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"sorted_kwarg",
"=",
"sorted",
"(",
"kwargs",
".",
"iteritems",
"(",
")",
",",
"key",
"=",
... | expected simple_unlimited_args_kwargs __doc__ . | train | false |
5,735 | def image_meta_to_http_headers(image_meta):
headers = {}
for (k, v) in image_meta.items():
if (v is not None):
if (k == 'properties'):
for (pk, pv) in v.items():
if (pv is not None):
headers[('x-image-meta-property-%s' % pk.lower())] = unicode(pv)
else:
headers[('x-image-meta-%s' % k.lower())] = unicode(v)
return headers
| [
"def",
"image_meta_to_http_headers",
"(",
"image_meta",
")",
":",
"headers",
"=",
"{",
"}",
"for",
"(",
"k",
",",
"v",
")",
"in",
"image_meta",
".",
"items",
"(",
")",
":",
"if",
"(",
"v",
"is",
"not",
"None",
")",
":",
"if",
"(",
"k",
"==",
"'pr... | returns a set of image metadata into a dict of http headers that can be fed to either a webob request object or an httplib . | train | false |
5,737 | def cleanQuery(query):
retVal = query
for sqlStatements in SQL_STATEMENTS.values():
for sqlStatement in sqlStatements:
sqlStatementEsc = sqlStatement.replace('(', '\\(')
queryMatch = re.search(('(%s)' % sqlStatementEsc), query, re.I)
if (queryMatch and ('sys_exec' not in query)):
retVal = retVal.replace(queryMatch.group(1), sqlStatement.upper())
return retVal
| [
"def",
"cleanQuery",
"(",
"query",
")",
":",
"retVal",
"=",
"query",
"for",
"sqlStatements",
"in",
"SQL_STATEMENTS",
".",
"values",
"(",
")",
":",
"for",
"sqlStatement",
"in",
"sqlStatements",
":",
"sqlStatementEsc",
"=",
"sqlStatement",
".",
"replace",
"(",
... | switch all sql statement keywords to upper case . | train | false |
5,741 | def instance_decorator(decorator):
class Decorator(object, ):
def __init__(self, func=nop, *args, **kws):
self.__name__ = func.__name__
self.__doc__ = func.__doc__
self._data_name = ('%s_%d_decorated_instance' % (func.__name__, id(self)))
self._func = func
self._args = args
self._kws = kws
def __get__(self, obj, cls=None):
if (obj is None):
return
data_name = self._data_name
try:
return obj.__dict__[data_name]
except KeyError:
decorated = decorator(obj, self._func, *self._args, **self._kws)
obj.__dict__[data_name] = decorated
return decorated
return Decorator
| [
"def",
"instance_decorator",
"(",
"decorator",
")",
":",
"class",
"Decorator",
"(",
"object",
",",
")",
":",
"def",
"__init__",
"(",
"self",
",",
"func",
"=",
"nop",
",",
"*",
"args",
",",
"**",
"kws",
")",
":",
"self",
".",
"__name__",
"=",
"func",
... | meta-decorator to define decorators that decorate a method in a concrete instance . | train | false |
5,742 | def strftime(date, date_format):
def strip_zeros(x):
return (x.lstrip(u'0') or u'0')
c89_directives = u'aAbBcdfHIjmMpSUwWxXyYzZ%'
format_options = u'%[-]?.'
candidates = re.findall(format_options, date_format)
template = re.sub(format_options, u'%s', date_format)
(lang_code, enc) = locale.getlocale(locale.LC_TIME)
formatted_candidates = []
for candidate in candidates:
if (candidate[(-1)] in c89_directives):
if (len(candidate) == 3):
candidate = u'%{}'.format(candidate[(-1)])
conversion = strip_zeros
else:
conversion = None
if isinstance(date, SafeDatetime):
formatted = date.strftime(candidate, safe=False)
else:
formatted = date.strftime(candidate)
if ((not six.PY3) and (enc is not None)):
formatted = formatted.decode(enc)
if conversion:
formatted = conversion(formatted)
else:
formatted = candidate
formatted_candidates.append(formatted)
return (template % tuple(formatted_candidates))
| [
"def",
"strftime",
"(",
"date",
",",
"date_format",
")",
":",
"def",
"strip_zeros",
"(",
"x",
")",
":",
"return",
"(",
"x",
".",
"lstrip",
"(",
"u'0'",
")",
"or",
"u'0'",
")",
"c89_directives",
"=",
"u'aAbBcdfHIjmMpSUwWxXyYzZ%'",
"format_options",
"=",
"u'... | a version of strftime that returns unicode strings and tries to handle dates before 1900 . | train | false |
5,743 | @pytest.mark.parametrize('parallel', [True, False])
def test_commented_header(parallel, read_commented_header):
text = '\n # A B C\n 1 2 3\n 4 5 6\n'
t1 = read_commented_header(text, parallel=parallel)
expected = Table([[1, 4], [2, 5], [3, 6]], names=('A', 'B', 'C'))
assert_table_equal(t1, expected)
text = ('# first commented line\n # second commented line\n\n' + text)
t2 = read_commented_header(text, header_start=2, data_start=0, parallel=parallel)
assert_table_equal(t2, expected)
t3 = read_commented_header(text, header_start=(-1), data_start=0, parallel=parallel)
assert_table_equal(t3, expected)
text += '7 8 9'
t4 = read_commented_header(text, header_start=2, data_start=2, parallel=parallel)
expected = Table([[7], [8], [9]], names=('A', 'B', 'C'))
assert_table_equal(t4, expected)
with pytest.raises(ParameterError):
read_commented_header(text, header_start=(-1), data_start=(-1), parallel=parallel)
| [
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'parallel'",
",",
"[",
"True",
",",
"False",
"]",
")",
"def",
"test_commented_header",
"(",
"parallel",
",",
"read_commented_header",
")",
":",
"text",
"=",
"'\\n # A B C\\n 1 2 3\\n 4 5 6\\n'",
"t1",
"=",
... | the fastcommentedheader reader should mimic the behavior of the commentedheader by overriding the default header behavior of fastbasic . | train | false |
5,744 | def maxdists(Z):
Z = np.asarray(Z, order='c', dtype=np.double)
is_valid_linkage(Z, throw=True, name='Z')
n = (Z.shape[0] + 1)
MD = np.zeros(((n - 1),))
[Z] = _copy_arrays_if_base_present([Z])
_hierarchy.get_max_dist_for_each_cluster(Z, MD, int(n))
return MD
| [
"def",
"maxdists",
"(",
"Z",
")",
":",
"Z",
"=",
"np",
".",
"asarray",
"(",
"Z",
",",
"order",
"=",
"'c'",
",",
"dtype",
"=",
"np",
".",
"double",
")",
"is_valid_linkage",
"(",
"Z",
",",
"throw",
"=",
"True",
",",
"name",
"=",
"'Z'",
")",
"n",
... | returns the maximum distance between any non-singleton cluster . | train | false |
5,746 | def import_dynamic(modname):
if (sys.version_info >= (3, 3)):
import importlib
importlib.invalidate_caches()
__import__(modname)
return sys.modules[modname]
| [
"def",
"import_dynamic",
"(",
"modname",
")",
":",
"if",
"(",
"sys",
".",
"version_info",
">=",
"(",
"3",
",",
"3",
")",
")",
":",
"import",
"importlib",
"importlib",
".",
"invalidate_caches",
"(",
")",
"__import__",
"(",
"modname",
")",
"return",
"sys",... | import and return a module of the given name . | train | false |
5,747 | def _service_get(s_name, **connection_args):
nitro = _connect(**connection_args)
if (nitro is None):
return None
service = NSService()
service.set_name(s_name)
try:
service = NSService.get(nitro, service)
except NSNitroError as error:
log.debug('netscaler module error - NSService.get() failed: {0}'.format(error))
service = None
_disconnect(nitro)
return service
| [
"def",
"_service_get",
"(",
"s_name",
",",
"**",
"connection_args",
")",
":",
"nitro",
"=",
"_connect",
"(",
"**",
"connection_args",
")",
"if",
"(",
"nitro",
"is",
"None",
")",
":",
"return",
"None",
"service",
"=",
"NSService",
"(",
")",
"service",
"."... | returns a service ressource or none . | train | true |
5,748 | def validate_pre_plugin_load():
if (cfg.CONF.core_plugin is None):
msg = _('Neutron core_plugin not configured!')
return msg
| [
"def",
"validate_pre_plugin_load",
"(",
")",
":",
"if",
"(",
"cfg",
".",
"CONF",
".",
"core_plugin",
"is",
"None",
")",
":",
"msg",
"=",
"_",
"(",
"'Neutron core_plugin not configured!'",
")",
"return",
"msg"
] | checks if the configuration variables are valid . | train | false |
5,749 | def _get_models_to_patch():
return list(itertools.chain(*[app_config.get_models(include_auto_created=False) for app_config in apps.get_app_configs() if ((app_config.label == 'osf') or app_config.label.startswith('addons_'))]))
| [
"def",
"_get_models_to_patch",
"(",
")",
":",
"return",
"list",
"(",
"itertools",
".",
"chain",
"(",
"*",
"[",
"app_config",
".",
"get_models",
"(",
"include_auto_created",
"=",
"False",
")",
"for",
"app_config",
"in",
"apps",
".",
"get_app_configs",
"(",
")... | return all models from osf and addons . | train | false |
5,750 | def match_str(filter_str, dct):
return all((_match_one(filter_part, dct) for filter_part in filter_str.split(u'&')))
| [
"def",
"match_str",
"(",
"filter_str",
",",
"dct",
")",
":",
"return",
"all",
"(",
"(",
"_match_one",
"(",
"filter_part",
",",
"dct",
")",
"for",
"filter_part",
"in",
"filter_str",
".",
"split",
"(",
"u'&'",
")",
")",
")"
] | filter a dictionary with a simple string syntax . | train | false |
5,751 | def delete_api_deployment(restApiId, deploymentId, region=None, key=None, keyid=None, profile=None):
try:
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
conn.delete_deployment(restApiId=restApiId, deploymentId=deploymentId)
return {'deleted': True}
except ClientError as e:
return {'deleted': False, 'error': salt.utils.boto3.get_error(e)}
| [
"def",
"delete_api_deployment",
"(",
"restApiId",
",",
"deploymentId",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"try",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"reg... | deletes api deployment for a given restapiid and deploymentid cli example: . | train | false |
5,753 | def possible_filename(filename):
if isinstance(filename, six.text_type):
return True
elif isinstance(filename, six.binary_type):
return (not ((sys.platform == u'win32') and (sys.version_info[:2] >= (3, 3))))
return False
| [
"def",
"possible_filename",
"(",
"filename",
")",
":",
"if",
"isinstance",
"(",
"filename",
",",
"six",
".",
"text_type",
")",
":",
"return",
"True",
"elif",
"isinstance",
"(",
"filename",
",",
"six",
".",
"binary_type",
")",
":",
"return",
"(",
"not",
"... | determine if the filename argument is an allowable type for a filename . | train | false |
5,754 | def auto_through(field):
return ((not field.rel.through) or getattr(getattr(field.rel.through, '_meta', None), 'auto_created', False))
| [
"def",
"auto_through",
"(",
"field",
")",
":",
"return",
"(",
"(",
"not",
"field",
".",
"rel",
".",
"through",
")",
"or",
"getattr",
"(",
"getattr",
"(",
"field",
".",
"rel",
".",
"through",
",",
"'_meta'",
",",
"None",
")",
",",
"'auto_created'",
",... | returns if the m2m class passed in has an autogenerated through table or not . | train | false |
5,756 | def st_mode_to_octal(mode):
try:
return oct(mode)[(-4):]
except (TypeError, IndexError):
return ''
| [
"def",
"st_mode_to_octal",
"(",
"mode",
")",
":",
"try",
":",
"return",
"oct",
"(",
"mode",
")",
"[",
"(",
"-",
"4",
")",
":",
"]",
"except",
"(",
"TypeError",
",",
"IndexError",
")",
":",
"return",
"''"
] | convert the st_mode value from a stat(2) call (as returned from os . | train | false |
5,757 | def _req_json_rpc(url, method, *args, **kwargs):
data = json.dumps({'method': method, 'params': args})
try:
res = requests.post(url, data=data, timeout=5, **kwargs)
except requests.exceptions.Timeout:
_LOGGER.exception('Connection to the router timed out')
return
if (res.status_code == 200):
try:
result = res.json()
except ValueError:
_LOGGER.exception('Failed to parse response from luci')
return
try:
return result['result']
except KeyError:
_LOGGER.exception('No result in response from luci')
return
elif (res.status_code == 401):
_LOGGER.exception('Failed to authenticate, please check your username and password')
return
else:
_LOGGER.error('Invalid response from luci: %s', res)
| [
"def",
"_req_json_rpc",
"(",
"url",
",",
"method",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"data",
"=",
"json",
".",
"dumps",
"(",
"{",
"'method'",
":",
"method",
",",
"'params'",
":",
"args",
"}",
")",
"try",
":",
"res",
"=",
"requests",
... | perform one json rpc operation . | train | false |
5,758 | def _Replacement_write_data(writer, data, is_attrib=False):
data = data.replace('&', '&').replace('<', '<')
data = data.replace('"', '"').replace('>', '>')
if is_attrib:
data = data.replace('\r', '
').replace('\n', '
').replace(' DCTB ', '	')
writer.write(data)
| [
"def",
"_Replacement_write_data",
"(",
"writer",
",",
"data",
",",
"is_attrib",
"=",
"False",
")",
":",
"data",
"=",
"data",
".",
"replace",
"(",
"'&'",
",",
"'&'",
")",
".",
"replace",
"(",
"'<'",
",",
"'<'",
")",
"data",
"=",
"data",
".",
"r... | writes datachars to writer . | train | false |
5,759 | def _offset_or_limit_clause_asint(clause, attrname):
if (clause is None):
return None
try:
value = clause._limit_offset_value
except AttributeError:
raise exc.CompileError(('This SELECT structure does not use a simple integer value for %s' % attrname))
else:
return util.asint(value)
| [
"def",
"_offset_or_limit_clause_asint",
"(",
"clause",
",",
"attrname",
")",
":",
"if",
"(",
"clause",
"is",
"None",
")",
":",
"return",
"None",
"try",
":",
"value",
"=",
"clause",
".",
"_limit_offset_value",
"except",
"AttributeError",
":",
"raise",
"exc",
... | convert the "offset or limit" clause of a select construct to an integer . | train | false |
5,760 | def event_elapsed_time(evtstart, evtend):
msec = c_float()
driver.cuEventElapsedTime(byref(msec), evtstart.handle, evtend.handle)
return msec.value
| [
"def",
"event_elapsed_time",
"(",
"evtstart",
",",
"evtend",
")",
":",
"msec",
"=",
"c_float",
"(",
")",
"driver",
".",
"cuEventElapsedTime",
"(",
"byref",
"(",
"msec",
")",
",",
"evtstart",
".",
"handle",
",",
"evtend",
".",
"handle",
")",
"return",
"ms... | compute the elapsed time between two events in milliseconds . | train | false |
5,761 | def get_video_from_youtube_id(youtube_id):
for (channel, language) in available_content_databases():
video = _get_video_from_youtube_id(channel=channel, language=language, youtube_id=youtube_id)
if video:
return video
| [
"def",
"get_video_from_youtube_id",
"(",
"youtube_id",
")",
":",
"for",
"(",
"channel",
",",
"language",
")",
"in",
"available_content_databases",
"(",
")",
":",
"video",
"=",
"_get_video_from_youtube_id",
"(",
"channel",
"=",
"channel",
",",
"language",
"=",
"l... | this function is provided to ensure that the data migration 0029_set_video_id_for_realz in the main app is still able to be run if needed . | train | false |
5,762 | def _import_type_text(dest, elem, type=None):
if (type is None):
type = elem.attrib['type']
if _should_skip_elem(elem, type, dest):
return
dest[type] = _text(elem)
| [
"def",
"_import_type_text",
"(",
"dest",
",",
"elem",
",",
"type",
"=",
"None",
")",
":",
"if",
"(",
"type",
"is",
"None",
")",
":",
"type",
"=",
"elem",
".",
"attrib",
"[",
"'type'",
"]",
"if",
"_should_skip_elem",
"(",
"elem",
",",
"type",
",",
"... | conditionally import the elements inner text(s) into the dest dict . | train | false |
5,764 | def InstallModule(conf_module_name, params, options, log=(lambda *args: None)):
if (not hasattr(sys, 'frozen')):
conf_module_name = os.path.abspath(conf_module_name)
if (not os.path.isfile(conf_module_name)):
raise ConfigurationError(('%s does not exist' % (conf_module_name,)))
loader_dll = GetLoaderModuleName(conf_module_name)
_PatchParamsModule(params, loader_dll)
Install(params, options)
log(1, 'Installation complete.')
| [
"def",
"InstallModule",
"(",
"conf_module_name",
",",
"params",
",",
"options",
",",
"log",
"=",
"(",
"lambda",
"*",
"args",
":",
"None",
")",
")",
":",
"if",
"(",
"not",
"hasattr",
"(",
"sys",
",",
"'frozen'",
")",
")",
":",
"conf_module_name",
"=",
... | install the extension . | train | false |
5,765 | def getDiagonalFlippedLoop(loop):
diagonalFlippedLoop = []
for point in loop:
diagonalFlippedLoop.append(complex(point.imag, point.real))
return diagonalFlippedLoop
| [
"def",
"getDiagonalFlippedLoop",
"(",
"loop",
")",
":",
"diagonalFlippedLoop",
"=",
"[",
"]",
"for",
"point",
"in",
"loop",
":",
"diagonalFlippedLoop",
".",
"append",
"(",
"complex",
"(",
"point",
".",
"imag",
",",
"point",
".",
"real",
")",
")",
"return",... | get loop flipped over the dialogonal . | train | false |
5,767 | def auto_configure_disk(session, vdi_ref, new_gb):
with vdi_attached_here(session, vdi_ref, read_only=False) as dev:
partitions = _get_partitions(dev)
if (len(partitions) != 1):
return
(_num, start, old_sectors, ptype) = partitions[0]
if (ptype in ('ext3', 'ext4')):
new_sectors = ((((new_gb * 1024) * 1024) * 1024) / SECTOR_SIZE)
_resize_part_and_fs(dev, start, old_sectors, new_sectors)
| [
"def",
"auto_configure_disk",
"(",
"session",
",",
"vdi_ref",
",",
"new_gb",
")",
":",
"with",
"vdi_attached_here",
"(",
"session",
",",
"vdi_ref",
",",
"read_only",
"=",
"False",
")",
"as",
"dev",
":",
"partitions",
"=",
"_get_partitions",
"(",
"dev",
")",
... | partition and resize fs to match the size specified by instance_types . | train | false |
5,768 | def has_app_context():
return (_app_ctx_stack.top is not None)
| [
"def",
"has_app_context",
"(",
")",
":",
"return",
"(",
"_app_ctx_stack",
".",
"top",
"is",
"not",
"None",
")"
] | works like :func:has_request_context but for the application context . | train | false |
5,769 | def namignizer_iterator(names, counts, batch_size, num_steps, epoch_size):
name_distribution = (counts / counts.sum())
for i in range(epoch_size):
data = np.zeros(((batch_size * num_steps) + 1))
samples = np.random.choice(names, size=((batch_size * num_steps) // 2), replace=True, p=name_distribution)
data_index = 0
for sample in samples:
if (data_index >= (batch_size * num_steps)):
break
for letter in (map(_letter_to_number, sample) + [_EON]):
if (data_index >= (batch_size * num_steps)):
break
data[data_index] = letter
data_index += 1
x = data[:(batch_size * num_steps)].reshape((batch_size, num_steps))
y = data[1:((batch_size * num_steps) + 1)].reshape((batch_size, num_steps))
(yield (x, y))
| [
"def",
"namignizer_iterator",
"(",
"names",
",",
"counts",
",",
"batch_size",
",",
"num_steps",
",",
"epoch_size",
")",
":",
"name_distribution",
"=",
"(",
"counts",
"/",
"counts",
".",
"sum",
"(",
")",
")",
"for",
"i",
"in",
"range",
"(",
"epoch_size",
... | takes a list of names and counts like those output from read_names . | train | false |
5,771 | def convert_kvp_str_to_list(data):
kvp = [x.strip() for x in data.split('=', 1)]
if ((len(kvp) == 2) and kvp[0]):
return kvp
msg = (_("'%s' is not of the form <key>=[value]") % data)
raise q_exc.InvalidInput(error_message=msg)
| [
"def",
"convert_kvp_str_to_list",
"(",
"data",
")",
":",
"kvp",
"=",
"[",
"x",
".",
"strip",
"(",
")",
"for",
"x",
"in",
"data",
".",
"split",
"(",
"'='",
",",
"1",
")",
"]",
"if",
"(",
"(",
"len",
"(",
"kvp",
")",
"==",
"2",
")",
"and",
"kvp... | convert a value of the form key=value to [key . | train | false |
5,772 | def index_db(index_filename, filenames=None, format=None, alphabet=None, key_function=None):
if (not isinstance(index_filename, basestring)):
raise TypeError('Need a string for the index filename')
if isinstance(filenames, basestring):
filenames = [filenames]
if ((filenames is not None) and (not isinstance(filenames, list))):
raise TypeError('Need a list of filenames (as strings), or one filename')
if ((format is not None) and (not isinstance(format, basestring))):
raise TypeError('Need a string for the file format (lower case)')
if (format and (format != format.lower())):
raise ValueError(("Format string '%s' should be lower case" % format))
if ((alphabet is not None) and (not (isinstance(alphabet, Alphabet) or isinstance(alphabet, AlphabetEncoder)))):
raise ValueError(('Invalid alphabet, %r' % alphabet))
from ._index import _FormatToRandomAccess
from Bio.File import _SQLiteManySeqFilesDict
repr = ('SeqIO.index_db(%r, filenames=%r, format=%r, alphabet=%r, key_function=%r)' % (index_filename, filenames, format, alphabet, key_function))
def proxy_factory(format, filename=None):
'Given a filename returns proxy object, else boolean if format OK.'
if filename:
return _FormatToRandomAccess[format](filename, format, alphabet)
else:
return (format in _FormatToRandomAccess)
return _SQLiteManySeqFilesDict(index_filename, filenames, proxy_factory, format, key_function, repr)
| [
"def",
"index_db",
"(",
"index_filename",
",",
"filenames",
"=",
"None",
",",
"format",
"=",
"None",
",",
"alphabet",
"=",
"None",
",",
"key_function",
"=",
"None",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"index_filename",
",",
"basestring",
")",
... | indexes several search output files into an sqlite database . | train | false |
5,773 | def get_blocks_with_unallocated(module, cp_driver, lb_driver, network_domain):
total_unallocated_ips = 0
all_blocks = list_public_ip_blocks(module, cp_driver, network_domain)
unalloc_blocks = []
unalloc_addresses = []
for block in all_blocks:
d_blocks = get_block_allocation(module, cp_driver, lb_driver, network_domain, block)
i = 0
for addr in d_blocks['addresses']:
if (addr['allocated'] is False):
if (i == 0):
unalloc_blocks.append(d_blocks)
unalloc_addresses.append(addr['address'])
total_unallocated_ips += 1
i += 1
return {'unallocated_count': total_unallocated_ips, 'ip_blocks': unalloc_blocks, 'unallocated_addresses': unalloc_addresses}
| [
"def",
"get_blocks_with_unallocated",
"(",
"module",
",",
"cp_driver",
",",
"lb_driver",
",",
"network_domain",
")",
":",
"total_unallocated_ips",
"=",
"0",
"all_blocks",
"=",
"list_public_ip_blocks",
"(",
"module",
",",
"cp_driver",
",",
"network_domain",
")",
"una... | gets ip blocks with one or more unallocated ips . | train | false |
5,775 | def approx_jacobian(x, func, epsilon, *args):
x0 = asfarray(x)
f0 = atleast_1d(func(*((x0,) + args)))
jac = zeros([len(x0), len(f0)])
dx = zeros(len(x0))
for i in range(len(x0)):
dx[i] = epsilon
jac[i] = ((func(*(((x0 + dx),) + args)) - f0) / epsilon)
dx[i] = 0.0
return jac.transpose()
| [
"def",
"approx_jacobian",
"(",
"x",
",",
"func",
",",
"epsilon",
",",
"*",
"args",
")",
":",
"x0",
"=",
"asfarray",
"(",
"x",
")",
"f0",
"=",
"atleast_1d",
"(",
"func",
"(",
"*",
"(",
"(",
"x0",
",",
")",
"+",
"args",
")",
")",
")",
"jac",
"=... | approximate the jacobian matrix of a callable function . | train | false |
5,777 | def EncodeRspFileList(args):
if (not args):
return ''
if args[0].startswith('call '):
(call, program) = args[0].split(' ', 1)
program = ((call + ' ') + os.path.normpath(program))
else:
program = os.path.normpath(args[0])
return ((program + ' ') + ' '.join((QuoteForRspFile(arg) for arg in args[1:])))
| [
"def",
"EncodeRspFileList",
"(",
"args",
")",
":",
"if",
"(",
"not",
"args",
")",
":",
"return",
"''",
"if",
"args",
"[",
"0",
"]",
".",
"startswith",
"(",
"'call '",
")",
":",
"(",
"call",
",",
"program",
")",
"=",
"args",
"[",
"0",
"]",
".",
... | process a list of arguments using quotecmdexeargument . | train | false |
5,779 | def _compare_combo(raw, new, times, n_times):
for ti in times:
orig = raw[:, (ti % n_times)][0]
assert_allclose(orig, new[:, ti][0])
| [
"def",
"_compare_combo",
"(",
"raw",
",",
"new",
",",
"times",
",",
"n_times",
")",
":",
"for",
"ti",
"in",
"times",
":",
"orig",
"=",
"raw",
"[",
":",
",",
"(",
"ti",
"%",
"n_times",
")",
"]",
"[",
"0",
"]",
"assert_allclose",
"(",
"orig",
",",
... | compare data . | train | false |
5,780 | def _restore_service(service):
_apply_service(service, SonosDevice.restore)
| [
"def",
"_restore_service",
"(",
"service",
")",
":",
"_apply_service",
"(",
"service",
",",
"SonosDevice",
".",
"restore",
")"
] | restore a snapshot . | train | false |
5,781 | def pil_from_ndarray(ndarray):
try:
if ((ndarray.dtype == 'float32') or (ndarray.dtype == 'float64')):
assert (ndarray.min() >= 0.0)
assert (ndarray.max() <= 1.0)
ndarray = np.cast['uint8']((ndarray * 255))
if ((len(ndarray.shape) == 3) and (ndarray.shape[2] == 1)):
ndarray = ndarray[:, :, 0]
ensure_Image()
rval = Image.fromarray(ndarray)
return rval
except Exception as e:
logger.exception('original exception: ')
logger.exception(e)
logger.exception('ndarray.dtype: {0}'.format(ndarray.dtype))
logger.exception('ndarray.shape: {0}'.format(ndarray.shape))
raise
assert False
| [
"def",
"pil_from_ndarray",
"(",
"ndarray",
")",
":",
"try",
":",
"if",
"(",
"(",
"ndarray",
".",
"dtype",
"==",
"'float32'",
")",
"or",
"(",
"ndarray",
".",
"dtype",
"==",
"'float64'",
")",
")",
":",
"assert",
"(",
"ndarray",
".",
"min",
"(",
")",
... | converts an ndarray to a pil image . | train | false |
5,782 | @pytest.fixture
def config_tmpdir(monkeypatch, tmpdir):
confdir = (tmpdir / 'config')
path = str(confdir)
os.mkdir(path)
monkeypatch.setattr('qutebrowser.utils.standarddir.config', (lambda : path))
return confdir
| [
"@",
"pytest",
".",
"fixture",
"def",
"config_tmpdir",
"(",
"monkeypatch",
",",
"tmpdir",
")",
":",
"confdir",
"=",
"(",
"tmpdir",
"/",
"'config'",
")",
"path",
"=",
"str",
"(",
"confdir",
")",
"os",
".",
"mkdir",
"(",
"path",
")",
"monkeypatch",
".",
... | set tmpdir/config as the configdir . | train | false |
5,786 | def GetRegisteredHelpFile(helpDesc):
try:
return GetRegistryDefaultValue(((BuildDefaultPythonKey() + '\\Help\\') + helpDesc))
except win32api.error:
try:
return GetRegistryDefaultValue(((BuildDefaultPythonKey() + '\\Help\\') + helpDesc), win32con.HKEY_CURRENT_USER)
except win32api.error:
pass
return None
| [
"def",
"GetRegisteredHelpFile",
"(",
"helpDesc",
")",
":",
"try",
":",
"return",
"GetRegistryDefaultValue",
"(",
"(",
"(",
"BuildDefaultPythonKey",
"(",
")",
"+",
"'\\\\Help\\\\'",
")",
"+",
"helpDesc",
")",
")",
"except",
"win32api",
".",
"error",
":",
"try",... | given a description . | train | false |
5,787 | def block_comments_begin_with_a_space(physical_line, line_number):
MESSAGE = "K002 block comments should start with '# '"
if ((line_number == 1) and physical_line.startswith('#!')):
return
text = physical_line.strip()
if text.startswith('#'):
if ((len(text) > 1) and (not text[1].isspace())):
return (physical_line.index('#'), MESSAGE)
| [
"def",
"block_comments_begin_with_a_space",
"(",
"physical_line",
",",
"line_number",
")",
":",
"MESSAGE",
"=",
"\"K002 block comments should start with '# '\"",
"if",
"(",
"(",
"line_number",
"==",
"1",
")",
"and",
"physical_line",
".",
"startswith",
"(",
"'#!'",
")"... | there should be a space after the # of block comments . | train | false |
5,788 | def test_compare_fiff():
check_usage(mne_compare_fiff)
| [
"def",
"test_compare_fiff",
"(",
")",
":",
"check_usage",
"(",
"mne_compare_fiff",
")"
] | test mne compare_fiff . | train | false |
5,789 | def get_icon(name, default=None, resample=False):
icon_path = get_image_path(name, default=None)
if (icon_path is not None):
icon = QIcon(icon_path)
elif isinstance(default, QIcon):
icon = default
elif (default is None):
try:
icon = get_std_icon(name[:(-4)])
except AttributeError:
icon = QIcon(get_image_path(name, default))
else:
icon = QIcon(get_image_path(name, default))
if resample:
icon0 = QIcon()
for size in (16, 24, 32, 48, 96, 128, 256, 512):
icon0.addPixmap(icon.pixmap(size, size))
return icon0
else:
return icon
| [
"def",
"get_icon",
"(",
"name",
",",
"default",
"=",
"None",
",",
"resample",
"=",
"False",
")",
":",
"icon_path",
"=",
"get_image_path",
"(",
"name",
",",
"default",
"=",
"None",
")",
"if",
"(",
"icon_path",
"is",
"not",
"None",
")",
":",
"icon",
"=... | retrieve a qicon for the named image from the zip file if it exists . | train | true |
5,790 | def compileADF(expr, psets):
adfdict = {}
func = None
for (pset, subexpr) in reversed(zip(psets, expr)):
pset.context.update(adfdict)
func = compile(subexpr, pset)
adfdict.update({pset.name: func})
return func
| [
"def",
"compileADF",
"(",
"expr",
",",
"psets",
")",
":",
"adfdict",
"=",
"{",
"}",
"func",
"=",
"None",
"for",
"(",
"pset",
",",
"subexpr",
")",
"in",
"reversed",
"(",
"zip",
"(",
"psets",
",",
"expr",
")",
")",
":",
"pset",
".",
"context",
".",... | compile the expression represented by a list of trees . | train | false |
5,791 | def watch(no_compress):
setup()
import time
compile_less()
build(no_compress=True)
while True:
compile_less()
if files_dirty():
build(no_compress=True)
time.sleep(3)
| [
"def",
"watch",
"(",
"no_compress",
")",
":",
"setup",
"(",
")",
"import",
"time",
"compile_less",
"(",
")",
"build",
"(",
"no_compress",
"=",
"True",
")",
"while",
"True",
":",
"compile_less",
"(",
")",
"if",
"files_dirty",
"(",
")",
":",
"build",
"("... | add a file to the watch list . | train | false |
5,792 | @docfiller
def gaussian_gradient_magnitude(input, sigma, output=None, mode='reflect', cval=0.0, **kwargs):
input = numpy.asarray(input)
def derivative(input, axis, output, mode, cval, sigma, **kwargs):
order = ([0] * input.ndim)
order[axis] = 1
return gaussian_filter(input, sigma, order, output, mode, cval, **kwargs)
return generic_gradient_magnitude(input, derivative, output, mode, cval, extra_arguments=(sigma,), extra_keywords=kwargs)
| [
"@",
"docfiller",
"def",
"gaussian_gradient_magnitude",
"(",
"input",
",",
"sigma",
",",
"output",
"=",
"None",
",",
"mode",
"=",
"'reflect'",
",",
"cval",
"=",
"0.0",
",",
"**",
"kwargs",
")",
":",
"input",
"=",
"numpy",
".",
"asarray",
"(",
"input",
... | multidimensional gradient magnitude using gaussian derivatives . | train | false |
5,793 | def init_subsystem(subsystem_type, options=None):
init_subsystems([subsystem_type], options)
| [
"def",
"init_subsystem",
"(",
"subsystem_type",
",",
"options",
"=",
"None",
")",
":",
"init_subsystems",
"(",
"[",
"subsystem_type",
"]",
",",
"options",
")"
] | singular form of :func:pants_test . | train | false |
5,796 | def gjrconvertparams(self, params, nar, nma):
(p, q) = (nar, nma)
ar = np.concatenate(([1], params[:p]))
ma = np.zeros(((q + 1), 3))
ma[(0, 0)] = params[(-1)]
ma[:, 1] = np.concatenate(([0], params[p:(p + q)]))
ma[:, 2] = np.concatenate(([0], params[(p + q):(p + (2 * q))]))
mu = params[(-1)]
params2 = (ar, ma)
return paramsclass
| [
"def",
"gjrconvertparams",
"(",
"self",
",",
"params",
",",
"nar",
",",
"nma",
")",
":",
"(",
"p",
",",
"q",
")",
"=",
"(",
"nar",
",",
"nma",
")",
"ar",
"=",
"np",
".",
"concatenate",
"(",
"(",
"[",
"1",
"]",
",",
"params",
"[",
":",
"p",
... | flat to matrix notes needs to be overwritten by subclass . | train | false |
5,797 | def str(val):
return format('%.12g', val)
| [
"def",
"str",
"(",
"val",
")",
":",
"return",
"format",
"(",
"'%.12g'",
",",
"val",
")"
] | convert float to integer . | train | false |
5,798 | def get_network_timezone(network):
if (network is None):
return sr_timezone
try:
return (tz.gettz(network_dict[network]) or sr_timezone)
except Exception:
return sr_timezone
| [
"def",
"get_network_timezone",
"(",
"network",
")",
":",
"if",
"(",
"network",
"is",
"None",
")",
":",
"return",
"sr_timezone",
"try",
":",
"return",
"(",
"tz",
".",
"gettz",
"(",
"network_dict",
"[",
"network",
"]",
")",
"or",
"sr_timezone",
")",
"excep... | get a timezone of a network from a given network dict . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.