id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
55,343 | def _is_astropy_source(path=None):
if (path is None):
path = os.path.join(os.path.dirname(__file__), os.pardir)
elif os.path.isfile(path):
path = os.path.dirname(path)
source_dir = os.path.abspath(path)
return os.path.exists(os.path.join(source_dir, '.astropy-root'))
| [
"def",
"_is_astropy_source",
"(",
"path",
"=",
"None",
")",
":",
"if",
"(",
"path",
"is",
"None",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
",",
"os",
".",
"pardir",
")",
... | returns whether the source for this module is directly in an astropy source distribution or checkout . | train | false |
55,344 | def _ValidateClientId(client_id):
if (not isinstance(client_id, basestring)):
raise InvalidChannelClientIdError(('"%s" is not a string.' % client_id))
if isinstance(client_id, unicode):
client_id = client_id.encode('utf-8')
if (len(client_id) > MAXIMUM_CLIENT_ID_LENGTH):
msg = ('Client id length %d is greater than max length %d' % (len(client_id), MAXIMUM_CLIENT_ID_LENGTH))
raise InvalidChannelClientIdError(msg)
return client_id
| [
"def",
"_ValidateClientId",
"(",
"client_id",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"client_id",
",",
"basestring",
")",
")",
":",
"raise",
"InvalidChannelClientIdError",
"(",
"(",
"'\"%s\" is not a string.'",
"%",
"client_id",
")",
")",
"if",
"isinstan... | validates a client id . | train | false |
55,345 | def make_fna(sff_fp, output_fp, use_sfftools=False, no_trim=False):
if use_sfftools:
_fail_on_gzipped_sff(sff_fp)
check_sffinfo()
if no_trim:
_check_call(['sffinfo', '-notrim', '-s', sff_fp], stdout=open(output_fp, 'w'))
else:
_check_call(['sffinfo', '-s', sff_fp], stdout=open(output_fp, 'w'))
else:
try:
format_binary_sff_as_fna(qiime_open(sff_fp, 'rb'), open(output_fp, 'w'))
except:
raise IOError(('Could not parse SFF %s' % sff_fp))
| [
"def",
"make_fna",
"(",
"sff_fp",
",",
"output_fp",
",",
"use_sfftools",
"=",
"False",
",",
"no_trim",
"=",
"False",
")",
":",
"if",
"use_sfftools",
":",
"_fail_on_gzipped_sff",
"(",
"sff_fp",
")",
"check_sffinfo",
"(",
")",
"if",
"no_trim",
":",
"_check_cal... | makes fna file from sff file . | train | false |
55,346 | def perm2tensor(t, g, canon_bp=False):
if (not isinstance(t, TensExpr)):
return t
new_tids = get_tids(t).perm2tensor(g, canon_bp)
coeff = get_coeff(t)
if (g[(-1)] != (len(g) - 1)):
coeff = (- coeff)
res = TensMul.from_TIDS(coeff, new_tids, is_canon_bp=canon_bp)
return res
| [
"def",
"perm2tensor",
"(",
"t",
",",
"g",
",",
"canon_bp",
"=",
"False",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"t",
",",
"TensExpr",
")",
")",
":",
"return",
"t",
"new_tids",
"=",
"get_tids",
"(",
"t",
")",
".",
"perm2tensor",
"(",
"g",
... | returns the tensor corresponding to the permutation g for further details . | train | false |
55,347 | def get_build_results(build):
r_url = get_results_raw_url(build)
if (not r_url):
return
return convert_json_to_df(r_url)
| [
"def",
"get_build_results",
"(",
"build",
")",
":",
"r_url",
"=",
"get_results_raw_url",
"(",
"build",
")",
"if",
"(",
"not",
"r_url",
")",
":",
"return",
"return",
"convert_json_to_df",
"(",
"r_url",
")"
] | returns a df with the results of the vbench job associated with the travis build . | train | false |
55,348 | def get_hash(f):
import hashlib
m = hashlib.md5()
m.update(f)
return m.hexdigest()
| [
"def",
"get_hash",
"(",
"f",
")",
":",
"import",
"hashlib",
"m",
"=",
"hashlib",
".",
"md5",
"(",
")",
"m",
".",
"update",
"(",
"f",
")",
"return",
"m",
".",
"hexdigest",
"(",
")"
] | gets hexadmecimal md5 hash of a string . | train | false |
55,349 | def _parse_output(output, template):
ret = {}
index = 0
if (not (output and template)):
return ret
if ('translate' in template):
ret = _translate_output(output)
else:
output_list = output.strip().replace('\n', '').split(' ')
if (sum(template.values()) != len(output_list)):
raise ipmiexcept.IPMIException(_('ipmitool output length mismatch'))
for item in template.items():
index_end = (index + item[1])
update_value = output_list[index:index_end]
ret[item[0]] = update_value
index = index_end
return ret
| [
"def",
"_parse_output",
"(",
"output",
",",
"template",
")",
":",
"ret",
"=",
"{",
"}",
"index",
"=",
"0",
"if",
"(",
"not",
"(",
"output",
"and",
"template",
")",
")",
":",
"return",
"ret",
"if",
"(",
"'translate'",
"in",
"template",
")",
":",
"re... | parse the return value of ipmi command into dict . | train | false |
55,350 | def timestamp(x):
if (x.tzinfo is None):
x = x.replace(tzinfo=utc)
if hasattr(x, 'timestamp'):
return x.timestamp()
else:
return (x - datetime(1970, 1, 1, tzinfo=utc)).total_seconds()
| [
"def",
"timestamp",
"(",
"x",
")",
":",
"if",
"(",
"x",
".",
"tzinfo",
"is",
"None",
")",
":",
"x",
"=",
"x",
".",
"replace",
"(",
"tzinfo",
"=",
"utc",
")",
"if",
"hasattr",
"(",
"x",
",",
"'timestamp'",
")",
":",
"return",
"x",
".",
"timestam... | get a timestamp from a date in python 3 and python 2 . | train | true |
55,351 | def get_latest_repository_metadata(app, decoded_repository_id, downloadable=False):
sa_session = app.model.context.current
repository = sa_session.query(app.model.Repository).get(decoded_repository_id)
repo = hg_util.get_repo_for_repository(app, repository=repository, repo_path=None, create=False)
if downloadable:
changeset_revision = get_latest_downloadable_changeset_revision(app, repository, repo)
else:
changeset_revision = get_latest_changeset_revision(app, repository, repo)
return get_repository_metadata_by_changeset_revision(app, app.security.encode_id(repository.id), changeset_revision)
| [
"def",
"get_latest_repository_metadata",
"(",
"app",
",",
"decoded_repository_id",
",",
"downloadable",
"=",
"False",
")",
":",
"sa_session",
"=",
"app",
".",
"model",
".",
"context",
".",
"current",
"repository",
"=",
"sa_session",
".",
"query",
"(",
"app",
"... | get last metadata defined for a specified repository from the database . | train | false |
55,353 | @gen.engine
def ListRecursively(store, pattern, callback):
results = (yield gen.Task(ListAllKeys, store, prefix=PrefixFromPattern(pattern)))
callback(results)
| [
"@",
"gen",
".",
"engine",
"def",
"ListRecursively",
"(",
"store",
",",
"pattern",
",",
"callback",
")",
":",
"results",
"=",
"(",
"yield",
"gen",
".",
"Task",
"(",
"ListAllKeys",
",",
"store",
",",
"prefix",
"=",
"PrefixFromPattern",
"(",
"pattern",
")"... | recursively list all files matching pattern . | train | false |
55,354 | def _extract_doc_comment(content, line, column, markers):
if (markers[1] == ''):
return _extract_doc_comment_simple(content, line, column, markers)
elif (markers[1] == markers[2]):
return _extract_doc_comment_continuous(content, line, column, markers)
else:
return _extract_doc_comment_standard(content, line, column, markers)
| [
"def",
"_extract_doc_comment",
"(",
"content",
",",
"line",
",",
"column",
",",
"markers",
")",
":",
"if",
"(",
"markers",
"[",
"1",
"]",
"==",
"''",
")",
":",
"return",
"_extract_doc_comment_simple",
"(",
"content",
",",
"line",
",",
"column",
",",
"mar... | delegates depending on the given markers to the right extraction method . | train | false |
55,355 | def parseSdr(s):
assert isinstance(s, basestring)
sdr = [int(c) for c in s if (c in ('0', '1'))]
if (len(sdr) != len(s)):
raise ValueError("The provided string %s is malformed. The string should have only 0's and 1's.")
return sdr
| [
"def",
"parseSdr",
"(",
"s",
")",
":",
"assert",
"isinstance",
"(",
"s",
",",
"basestring",
")",
"sdr",
"=",
"[",
"int",
"(",
"c",
")",
"for",
"c",
"in",
"s",
"if",
"(",
"c",
"in",
"(",
"'0'",
",",
"'1'",
")",
")",
"]",
"if",
"(",
"len",
"(... | parses a string containing only 0s and 1s and return a python list object . | train | true |
55,357 | def floating_ip_create(kwargs, call=None):
if (call != 'function'):
raise SaltCloudSystemExit('The floating_ip_create action must be called with -f or --function')
if ('pool' not in kwargs):
log.error('pool is required')
return False
conn = get_conn()
return conn.floating_ip_create(kwargs['pool'])
| [
"def",
"floating_ip_create",
"(",
"kwargs",
",",
"call",
"=",
"None",
")",
":",
"if",
"(",
"call",
"!=",
"'function'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The floating_ip_create action must be called with -f or --function'",
")",
"if",
"(",
"'pool'",
"n... | allocate a floating ip . | train | true |
55,358 | def top_contributors_l10n(start=None, end=None, locale=None, product=None, count=10, page=1):
query = RevisionMetricsMappingType.search().facet('creator_id', filtered=True, size=BIG_NUMBER)
if (locale is None):
query = query.filter((~ F(locale=settings.WIKI_DEFAULT_LANGUAGE)))
query = _apply_filters(query, start, end, locale, product)
return _get_creator_counts(query, count, page)
| [
"def",
"top_contributors_l10n",
"(",
"start",
"=",
"None",
",",
"end",
"=",
"None",
",",
"locale",
"=",
"None",
",",
"product",
"=",
"None",
",",
"count",
"=",
"10",
",",
"page",
"=",
"1",
")",
":",
"query",
"=",
"RevisionMetricsMappingType",
".",
"sea... | get the top l10n contributors for the kb . | train | false |
55,359 | @login_required
@ensure_csrf_cookie
def assets_handler(request, course_key_string=None, asset_key_string=None):
course_key = CourseKey.from_string(course_key_string)
if (not has_course_author_access(request.user, course_key)):
raise PermissionDenied()
response_format = (request.GET.get('format') or request.POST.get('format') or 'html')
if ((response_format == 'json') or ('application/json' in request.META.get('HTTP_ACCEPT', 'application/json'))):
if (request.method == 'GET'):
return _assets_json(request, course_key)
else:
asset_key = (AssetKey.from_string(asset_key_string) if asset_key_string else None)
return _update_asset(request, course_key, asset_key)
elif (request.method == 'GET'):
return _asset_index(request, course_key)
else:
return HttpResponseNotFound()
| [
"@",
"login_required",
"@",
"ensure_csrf_cookie",
"def",
"assets_handler",
"(",
"request",
",",
"course_key_string",
"=",
"None",
",",
"asset_key_string",
"=",
"None",
")",
":",
"course_key",
"=",
"CourseKey",
".",
"from_string",
"(",
"course_key_string",
")",
"if... | the restful handler for assets . | train | false |
55,360 | def clamav(registry, xml_parent, data):
clamav = XML.SubElement(xml_parent, 'org.jenkinsci.plugins.clamav.ClamAvRecorder')
clamav.set('plugin', 'clamav')
mappings = [('includes', 'includes', ''), ('excludes', 'excludes', '')]
helpers.convert_mapping_to_xml(clamav, data, mappings, fail_required=True)
| [
"def",
"clamav",
"(",
"registry",
",",
"xml_parent",
",",
"data",
")",
":",
"clamav",
"=",
"XML",
".",
"SubElement",
"(",
"xml_parent",
",",
"'org.jenkinsci.plugins.clamav.ClamAvRecorder'",
")",
"clamav",
".",
"set",
"(",
"'plugin'",
",",
"'clamav'",
")",
"map... | yaml: clamav check files with clamav . | train | false |
55,361 | def tree_support(master, subsampled_tree):
master_tipnames = set(master.getTipNames())
subsampled_tree_trimmed = copy.deepcopy(subsampled_tree)
def delete_test(node):
if (not node.isTip()):
return False
else:
return (node.Name not in master_tipnames)
subsampled_tree_trimmed.removeDeleted(delete_test)
subsampled_tree_trimmed.prune()
subsampled_tree_nodes_names = []
for node in subsampled_tree_trimmed.iterNontips(include_self=True):
subsampled_tree_nodes_names.append(node.getTipNames())
subsampled_tree_nodes_names = map(set, subsampled_tree_nodes_names)
for master_node in master.iterNontips(include_self=True):
if (set(master_node.getTipNames()) in subsampled_tree_nodes_names):
try:
master_node.bootstrap_support += 1
except AttributeError:
master_node.bootstrap_support = 1
| [
"def",
"tree_support",
"(",
"master",
",",
"subsampled_tree",
")",
":",
"master_tipnames",
"=",
"set",
"(",
"master",
".",
"getTipNames",
"(",
")",
")",
"subsampled_tree_trimmed",
"=",
"copy",
".",
"deepcopy",
"(",
"subsampled_tree",
")",
"def",
"delete_test",
... | compares master tree to subsampled_tree . | train | false |
55,363 | def dlcs_api_request(path, params='', user='', passwd='', throttle=True):
if throttle:
Waiter()
if params:
url = ('%s/%s?%s' % (DLCS_API, path, urllib.urlencode(dict0(params))))
else:
url = ('%s/%s' % (DLCS_API, path))
if DEBUG:
print >>sys.stderr, ('dlcs_api_request: %s' % url)
try:
return http_auth_request(url, DLCS_API_HOST, user, passwd, USER_AGENT)
except DefaultErrorHandler as e:
print >>sys.stderr, ('%s' % e)
| [
"def",
"dlcs_api_request",
"(",
"path",
",",
"params",
"=",
"''",
",",
"user",
"=",
"''",
",",
"passwd",
"=",
"''",
",",
"throttle",
"=",
"True",
")",
":",
"if",
"throttle",
":",
"Waiter",
"(",
")",
"if",
"params",
":",
"url",
"=",
"(",
"'%s/%s?%s'... | retrieve/query a path within the del . | train | false |
55,364 | def conserve_mpmath_dps(func):
import functools
import mpmath
def func_wrapper(*args, **kwargs):
dps = mpmath.mp.dps
try:
return func(*args, **kwargs)
finally:
mpmath.mp.dps = dps
func_wrapper = functools.update_wrapper(func_wrapper, func)
return func_wrapper
| [
"def",
"conserve_mpmath_dps",
"(",
"func",
")",
":",
"import",
"functools",
"import",
"mpmath",
"def",
"func_wrapper",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"dps",
"=",
"mpmath",
".",
"mp",
".",
"dps",
"try",
":",
"return",
"func",
"(",
"*",... | after the function finishes . | train | false |
55,365 | def pathMatchPatterns(path, repos):
for repo in repos:
if fnmatch(path, repo):
return True
return False
| [
"def",
"pathMatchPatterns",
"(",
"path",
",",
"repos",
")",
":",
"for",
"repo",
"in",
"repos",
":",
"if",
"fnmatch",
"(",
"path",
",",
"repo",
")",
":",
"return",
"True",
"return",
"False"
] | check existence of given path against list of path patterns the pattern definition is the as fnmatch . | train | false |
55,366 | @domain_constructor(loss_target=(-2))
def distractor():
x = hp.uniform('x', (-15), 15)
f1 = old_div(1.0, (1.0 + scope.exp((- x))))
f2 = (2 * scope.exp((- ((x + 10) ** 2))))
return {'loss': ((- f1) - f2), 'status': base.STATUS_OK}
| [
"@",
"domain_constructor",
"(",
"loss_target",
"=",
"(",
"-",
"2",
")",
")",
"def",
"distractor",
"(",
")",
":",
"x",
"=",
"hp",
".",
"uniform",
"(",
"'x'",
",",
"(",
"-",
"15",
")",
",",
"15",
")",
"f1",
"=",
"old_div",
"(",
"1.0",
",",
"(",
... | this is a nasty function: it has a max in a spike near -10 . | train | false |
55,369 | def get_c_init(r, name, sub):
pre = ('\n py_%(name)s = Py_None;\n {Py_XINCREF(py_%(name)s);}\n ' % locals())
return (pre + r.type.c_init(name, sub))
| [
"def",
"get_c_init",
"(",
"r",
",",
"name",
",",
"sub",
")",
":",
"pre",
"=",
"(",
"'\\n py_%(name)s = Py_None;\\n {Py_XINCREF(py_%(name)s);}\\n '",
"%",
"locals",
"(",
")",
")",
"return",
"(",
"pre",
"+",
"r",
".",
"type",
".",
"c_init",
"(",
"name... | wrapper around c_init that initializes py_name to py_none . | train | false |
55,372 | def get_messages(request):
return getattr(request, '_messages', [])
| [
"def",
"get_messages",
"(",
"request",
")",
":",
"return",
"getattr",
"(",
"request",
",",
"'_messages'",
",",
"[",
"]",
")"
] | returns the message storage on the request if it exists . | train | false |
55,373 | def configure_paramiko_logging():
l = logging.getLogger('paramiko')
l.setLevel(logging.DEBUG)
static.create_sc_config_dirs()
lh = logging.handlers.RotatingFileHandler(static.SSH_DEBUG_FILE, maxBytes=1048576, backupCount=2)
lh.setLevel(logging.DEBUG)
format = (('PID: %s ' % str(static.PID)) + '%(levelname)-.3s [%(asctime)s.%(msecs)03d] thr=%(_threadid)-3d %(name)s: %(message)s')
date_format = '%Y%m%d-%H:%M:%S'
lh.setFormatter(logging.Formatter(format, date_format))
l.addHandler(lh)
| [
"def",
"configure_paramiko_logging",
"(",
")",
":",
"l",
"=",
"logging",
".",
"getLogger",
"(",
"'paramiko'",
")",
"l",
".",
"setLevel",
"(",
"logging",
".",
"DEBUG",
")",
"static",
".",
"create_sc_config_dirs",
"(",
")",
"lh",
"=",
"logging",
".",
"handle... | configure ssh to log to a file for debug . | train | false |
55,374 | def check_non_negative(X, whom):
X = (X.data if sp.issparse(X) else X)
if (X < 0).any():
raise ValueError(('Negative values in data passed to %s' % whom))
| [
"def",
"check_non_negative",
"(",
"X",
",",
"whom",
")",
":",
"X",
"=",
"(",
"X",
".",
"data",
"if",
"sp",
".",
"issparse",
"(",
"X",
")",
"else",
"X",
")",
"if",
"(",
"X",
"<",
"0",
")",
".",
"any",
"(",
")",
":",
"raise",
"ValueError",
"(",... | check if there is any negative value in an array . | train | false |
55,376 | def setvariable(cursor, mysqlvar, value):
query = ('SET GLOBAL %s = ' % mysql_quote_identifier(mysqlvar, 'vars'))
try:
cursor.execute((query + '%s'), (value,))
cursor.fetchall()
result = True
except Exception:
e = get_exception()
result = str(e)
return result
| [
"def",
"setvariable",
"(",
"cursor",
",",
"mysqlvar",
",",
"value",
")",
":",
"query",
"=",
"(",
"'SET GLOBAL %s = '",
"%",
"mysql_quote_identifier",
"(",
"mysqlvar",
",",
"'vars'",
")",
")",
"try",
":",
"cursor",
".",
"execute",
"(",
"(",
"query",
"+",
... | set a global mysql variable to a given value the db driver will handle quoting of the given value based on its type . | train | false |
55,378 | @login_required
@enforce_shopping_cart_enabled
def show_cart(request):
cart = Order.get_cart_for_user(request.user)
(is_any_course_expired, expired_cart_items, expired_cart_item_names, valid_cart_item_tuples) = verify_for_closed_enrollment(request.user, cart)
site_name = configuration_helpers.get_value('SITE_NAME', settings.SITE_NAME)
if is_any_course_expired:
for expired_item in expired_cart_items:
Order.remove_cart_item_from_order(expired_item, request.user)
cart.update_order_type()
callback_url = request.build_absolute_uri(reverse('shoppingcart.views.postpay_callback'))
form_html = render_purchase_form_html(cart, callback_url=callback_url)
context = {'order': cart, 'shoppingcart_items': valid_cart_item_tuples, 'amount': cart.total_cost, 'is_course_enrollment_closed': is_any_course_expired, 'expired_course_names': expired_cart_item_names, 'site_name': site_name, 'form_html': form_html, 'currency_symbol': settings.PAID_COURSE_REGISTRATION_CURRENCY[1], 'currency': settings.PAID_COURSE_REGISTRATION_CURRENCY[0], 'enable_bulk_purchase': configuration_helpers.get_value('ENABLE_SHOPPING_CART_BULK_PURCHASE', True)}
return render_to_response('shoppingcart/shopping_cart.html', context)
| [
"@",
"login_required",
"@",
"enforce_shopping_cart_enabled",
"def",
"show_cart",
"(",
"request",
")",
":",
"cart",
"=",
"Order",
".",
"get_cart_for_user",
"(",
"request",
".",
"user",
")",
"(",
"is_any_course_expired",
",",
"expired_cart_items",
",",
"expired_cart_i... | this view shows cart items . | train | false |
55,380 | def PostVimMessage(message, warning=True, truncate=False):
echo_command = (u'echom' if warning else u'echo')
vim.command(u'redraw')
if warning:
vim.command(u'echohl WarningMsg')
message = ToUnicode(message)
if truncate:
vim_width = GetIntValue(u'&columns')
message = message.replace(u'\n', u' ')
if (len(message) > vim_width):
message = (message[:(vim_width - 4)] + u'...')
old_ruler = GetIntValue(u'&ruler')
old_showcmd = GetIntValue(u'&showcmd')
vim.command(u'set noruler noshowcmd')
vim.command(u"{0} '{1}'".format(echo_command, EscapeForVim(message)))
SetVariableValue(u'&ruler', old_ruler)
SetVariableValue(u'&showcmd', old_showcmd)
else:
for line in message.split(u'\n'):
vim.command(u"{0} '{1}'".format(echo_command, EscapeForVim(line)))
if warning:
vim.command(u'echohl None')
| [
"def",
"PostVimMessage",
"(",
"message",
",",
"warning",
"=",
"True",
",",
"truncate",
"=",
"False",
")",
":",
"echo_command",
"=",
"(",
"u'echom'",
"if",
"warning",
"else",
"u'echo'",
")",
"vim",
".",
"command",
"(",
"u'redraw'",
")",
"if",
"warning",
"... | display a message on the vim status line . | train | false |
55,381 | def getRadiusByPrefix(prefix, sideLength, xmlElement):
radius = getFloatByPrefixSide((prefix + 'radius'), sideLength, xmlElement)
radius += (0.5 * getFloatByPrefixSide((prefix + 'diameter'), sideLength, xmlElement))
return (radius + (0.5 * getFloatByPrefixSide((prefix + 'size'), sideLength, xmlElement)))
| [
"def",
"getRadiusByPrefix",
"(",
"prefix",
",",
"sideLength",
",",
"xmlElement",
")",
":",
"radius",
"=",
"getFloatByPrefixSide",
"(",
"(",
"prefix",
"+",
"'radius'",
")",
",",
"sideLength",
",",
"xmlElement",
")",
"radius",
"+=",
"(",
"0.5",
"*",
"getFloatB... | get radius by prefix . | train | false |
55,385 | def _filter_doc_html(request, doc, doc_html, rendering_params):
if rendering_params['summary']:
return doc.get_summary_html()
if (not (rendering_params['section'] or rendering_params['raw'] or rendering_params['edit_links'] or rendering_params['include'])):
return doc_html
tool = kuma.wiki.content.parse(doc_html)
if rendering_params['raw']:
tool.injectSectionIDs()
tool.filterEditorSafety()
if rendering_params['section']:
tool.extractSection(rendering_params['section'])
if ((rendering_params['edit_links'] or (not rendering_params['raw'])) and request.user.is_authenticated() and doc.allows_revision_by(request.user)):
tool.injectSectionEditingLinks(doc.slug, doc.locale)
doc_html = tool.serialize()
if rendering_params['include']:
doc_html = kuma.wiki.content.filter_out_noinclude(doc_html)
return doc_html
| [
"def",
"_filter_doc_html",
"(",
"request",
",",
"doc",
",",
"doc_html",
",",
"rendering_params",
")",
":",
"if",
"rendering_params",
"[",
"'summary'",
"]",
":",
"return",
"doc",
".",
"get_summary_html",
"(",
")",
"if",
"(",
"not",
"(",
"rendering_params",
"[... | apply needed filtering/annotating operations to a documents html . | train | false |
55,386 | def create_minibatch_x(minibatches, minibatch_markers, epoch_axis):
if epoch_axis:
x = np.zeros((minibatches,))
last_e = 0
for (e_idx, e) in enumerate(minibatch_markers):
e_minibatches = (e - last_e)
x[last_e:e] = (e_idx + (np.arange(float(e_minibatches)) / e_minibatches))
last_e = e
else:
x = np.arange(minibatches)
return x
| [
"def",
"create_minibatch_x",
"(",
"minibatches",
",",
"minibatch_markers",
",",
"epoch_axis",
")",
":",
"if",
"epoch_axis",
":",
"x",
"=",
"np",
".",
"zeros",
"(",
"(",
"minibatches",
",",
")",
")",
"last_e",
"=",
"0",
"for",
"(",
"e_idx",
",",
"e",
")... | helper function to build x axis for data captured per minibatch . | train | false |
55,387 | def CDLMORNINGDOJISTAR(barDs, count, penetration=(-4e+37)):
return call_talib_with_ohlc(barDs, count, talib.CDLMORNINGDOJISTAR, penetration)
| [
"def",
"CDLMORNINGDOJISTAR",
"(",
"barDs",
",",
"count",
",",
"penetration",
"=",
"(",
"-",
"4e+37",
")",
")",
":",
"return",
"call_talib_with_ohlc",
"(",
"barDs",
",",
"count",
",",
"talib",
".",
"CDLMORNINGDOJISTAR",
",",
"penetration",
")"
] | morning doji star . | train | false |
55,388 | def gaussian_convolution(h, Xi, x):
return ((1.0 / np.sqrt((4 * np.pi))) * np.exp(((- ((Xi - x) ** 2)) / ((h ** 2) * 4.0))))
| [
"def",
"gaussian_convolution",
"(",
"h",
",",
"Xi",
",",
"x",
")",
":",
"return",
"(",
"(",
"1.0",
"/",
"np",
".",
"sqrt",
"(",
"(",
"4",
"*",
"np",
".",
"pi",
")",
")",
")",
"*",
"np",
".",
"exp",
"(",
"(",
"(",
"-",
"(",
"(",
"Xi",
"-",... | calculates the gaussian convolution kernel . | train | false |
55,389 | def _map_plays_to_roles(graph, dirs, git_dir, key, type_1, type_2):
Node = namedtuple('Node', ['name', 'type'])
for d in dirs:
d = pathlib2.Path(git_dir, d)
for item in d.iterdir():
if item.match('*.yml'):
yaml_file = _open_yaml_file(item)
if (yaml_file is not None):
for play in yaml_file:
if (key in play):
for role in play[key]:
name = _get_role_name(role)
node_1 = Node(item.stem, type_1)
node_2 = Node(name, type_2)
graph.add_edge(node_2, node_1)
| [
"def",
"_map_plays_to_roles",
"(",
"graph",
",",
"dirs",
",",
"git_dir",
",",
"key",
",",
"type_1",
",",
"type_2",
")",
":",
"Node",
"=",
"namedtuple",
"(",
"'Node'",
",",
"[",
"'name'",
",",
"'type'",
"]",
")",
"for",
"d",
"in",
"dirs",
":",
"d",
... | maps plays to the roles they use . | train | false |
55,391 | def handler_url(block, handler_name, suffix='', query='', thirdparty=False):
view_name = 'xblock_handler'
if handler_name:
func = getattr(block.__class__, handler_name, None)
if (not func):
raise ValueError('{!r} is not a function name'.format(handler_name))
if thirdparty:
view_name = 'xblock_handler_noauth'
url = reverse(view_name, kwargs={'course_id': unicode(block.location.course_key), 'usage_id': quote_slashes(unicode(block.scope_ids.usage_id).encode('utf-8')), 'handler': handler_name, 'suffix': suffix})
if (not suffix):
url = url.rstrip('/')
if query:
url += ('?' + query)
if thirdparty:
scheme = ('https' if (settings.HTTPS == 'on') else 'http')
url = '{scheme}://{host}{path}'.format(scheme=scheme, host=settings.SITE_NAME, path=url)
return url
| [
"def",
"handler_url",
"(",
"block",
",",
"handler_name",
",",
"suffix",
"=",
"''",
",",
"query",
"=",
"''",
",",
"thirdparty",
"=",
"False",
")",
":",
"view_name",
"=",
"'xblock_handler'",
"if",
"handler_name",
":",
"func",
"=",
"getattr",
"(",
"block",
... | this method matches the signature for xblock . | train | false |
55,392 | @protocol.commands.add(u'command_list_ok_begin', list_command=False)
def command_list_ok_begin(context):
context.dispatcher.command_list_receiving = True
context.dispatcher.command_list_ok = True
context.dispatcher.command_list = []
| [
"@",
"protocol",
".",
"commands",
".",
"add",
"(",
"u'command_list_ok_begin'",
",",
"list_command",
"=",
"False",
")",
"def",
"command_list_ok_begin",
"(",
"context",
")",
":",
"context",
".",
"dispatcher",
".",
"command_list_receiving",
"=",
"True",
"context",
... | see :meth:command_list_begin() . | train | false |
55,393 | def _ace_to_text(ace, objectType):
dc = daclConstants()
objectType = dc.getObjectTypeBit(objectType)
try:
userSid = win32security.LookupAccountSid('', ace[2])
if userSid[1]:
userSid = '{1}\\{0}'.format(userSid[0], userSid[1])
else:
userSid = '{0}'.format(userSid[0])
except Exception:
userSid = win32security.ConvertSidToStringSid(ace[2])
tPerm = ace[1]
tAceType = ace[0][0]
tProps = ace[0][1]
tInherited = ''
for x in dc.validAceTypes:
if (dc.validAceTypes[x]['BITS'] == tAceType):
tAceType = dc.validAceTypes[x]['TEXT']
break
for x in dc.rights[objectType]:
if (dc.rights[objectType][x]['BITS'] == tPerm):
tPerm = dc.rights[objectType][x]['TEXT']
break
if ((tProps & win32security.INHERITED_ACE) == win32security.INHERITED_ACE):
tInherited = '[Inherited]'
tProps = (tProps ^ win32security.INHERITED_ACE)
for x in dc.validPropagations[objectType]:
if (dc.validPropagations[objectType][x]['BITS'] == tProps):
tProps = dc.validPropagations[objectType][x]['TEXT']
break
return '{0} {1} {2} on {3} {4}'.format(userSid, tAceType, tPerm, tProps, tInherited)
| [
"def",
"_ace_to_text",
"(",
"ace",
",",
"objectType",
")",
":",
"dc",
"=",
"daclConstants",
"(",
")",
"objectType",
"=",
"dc",
".",
"getObjectTypeBit",
"(",
"objectType",
")",
"try",
":",
"userSid",
"=",
"win32security",
".",
"LookupAccountSid",
"(",
"''",
... | helper function to convert an ace to a textual representation . | train | true |
55,395 | def prime(nth):
n = as_int(nth)
if (n < 1):
raise ValueError('nth must be a positive integer; prime(1) == 2')
if (n <= len(sieve._list)):
return sieve[n]
from sympy.functions.special.error_functions import li
from sympy.functions.elementary.exponential import log
a = 2
b = int((n * (log(n) + log(log(n)))))
while (a < b):
mid = ((a + b) >> 1)
if (li(mid) > n):
b = mid
else:
a = (mid + 1)
n_primes = primepi((a - 1))
while (n_primes < n):
if isprime(a):
n_primes += 1
a += 1
return (a - 1)
| [
"def",
"prime",
"(",
"nth",
")",
":",
"n",
"=",
"as_int",
"(",
"nth",
")",
"if",
"(",
"n",
"<",
"1",
")",
":",
"raise",
"ValueError",
"(",
"'nth must be a positive integer; prime(1) == 2'",
")",
"if",
"(",
"n",
"<=",
"len",
"(",
"sieve",
".",
"_list",
... | return the nth prime . | train | false |
55,396 | @register.function
@jinja2.contextfunction
def favorites_widget(context, addon, condensed=False):
c = dict(context.items())
request = c['request']
if request.user.is_authenticated():
is_favorite = (addon.id in request.user.favorite_addons)
faved_class = ('faved' if is_favorite else '')
unfaved_text = ('' if condensed else _('Add to favorites'))
faved_text = (_('Favorite') if condensed else _('Remove from favorites'))
add_url = reverse('collections.alter', args=[request.user.username, 'favorites', 'add'])
remove_url = reverse('collections.alter', args=[request.user.username, 'favorites', 'remove'])
c.update(locals())
t = get_env().get_template('bandwagon/favorites_widget.html').render(c)
return jinja2.Markup(t)
| [
"@",
"register",
".",
"function",
"@",
"jinja2",
".",
"contextfunction",
"def",
"favorites_widget",
"(",
"context",
",",
"addon",
",",
"condensed",
"=",
"False",
")",
":",
"c",
"=",
"dict",
"(",
"context",
".",
"items",
"(",
")",
")",
"request",
"=",
"... | displays add to favorites widget . | train | false |
55,398 | def get_argnames(func):
if six.PY2:
if isinstance(func, functools.partial):
spec = inspect.getargspec(func.func)
elif inspect.isroutine(func):
spec = inspect.getargspec(func)
else:
spec = inspect.getargspec(func.__call__)
args = [arg for arg in spec.args if (arg != 'self')]
else:
sig = inspect.signature(func)
args = [param.name for param in sig.parameters.values() if (param.kind not in (inspect.Parameter.VAR_POSITIONAL, inspect.Parameter.VAR_KEYWORD))]
return args
| [
"def",
"get_argnames",
"(",
"func",
")",
":",
"if",
"six",
".",
"PY2",
":",
"if",
"isinstance",
"(",
"func",
",",
"functools",
".",
"partial",
")",
":",
"spec",
"=",
"inspect",
".",
"getargspec",
"(",
"func",
".",
"func",
")",
"elif",
"inspect",
".",... | introspecs the arguments of a callable . | train | false |
55,399 | def get_indices(client):
try:
indices = list(client.indices.get_settings(index='_all', params={'expand_wildcards': 'open,closed'}))
version_number = get_version(client)
logger.debug('Detected Elasticsearch version {0}'.format('.'.join(map(str, version_number))))
if ((version_number >= (2, 4, 2)) and (version_number < (5, 0, 0))):
logger.debug('Using Elasticsearch >= 2.4.2 < 5.0.0')
if client.indices.exists(index='.security'):
logger.debug('Found the ".security" index. Adding to list of all indices')
if (not ('.security' in indices)):
indices.append('.security')
logger.debug('All indices: {0}'.format(indices))
return indices
except Exception as e:
raise FailedExecution('Failed to get indices. Error: {0}'.format(e))
| [
"def",
"get_indices",
"(",
"client",
")",
":",
"try",
":",
"indices",
"=",
"list",
"(",
"client",
".",
"indices",
".",
"get_settings",
"(",
"index",
"=",
"'_all'",
",",
"params",
"=",
"{",
"'expand_wildcards'",
":",
"'open,closed'",
"}",
")",
")",
"versi... | get the current list of indices from the cluster . | train | false |
55,400 | def _initialize_headers(headers):
return ({} if (headers is None) else dict(headers))
| [
"def",
"_initialize_headers",
"(",
"headers",
")",
":",
"return",
"(",
"{",
"}",
"if",
"(",
"headers",
"is",
"None",
")",
"else",
"dict",
"(",
"headers",
")",
")"
] | creates a copy of the headers . | train | false |
55,401 | @contextmanager
def secret_model():
orig_model = None
try:
orig_model = loading.cache.app_models['tests']['secret']
del loading.cache.app_models['tests']['secret']
except KeyError:
pass
try:
attrs = {'name': EncryptedCharField('Name', max_length=Secret._meta.get_field('name').max_length), 'text': EncryptedTextField('Text'), '__module__': 'django_extensions.tests.models', 'Meta': type('Meta', (object,), {'managed': False, 'db_table': Secret._meta.db_table})}
(yield type('Secret', (models.Model,), attrs))
except:
raise
finally:
try:
loading.cache.app_models['tests']['secret'] = orig_model
except KeyError:
pass
| [
"@",
"contextmanager",
"def",
"secret_model",
"(",
")",
":",
"orig_model",
"=",
"None",
"try",
":",
"orig_model",
"=",
"loading",
".",
"cache",
".",
"app_models",
"[",
"'tests'",
"]",
"[",
"'secret'",
"]",
"del",
"loading",
".",
"cache",
".",
"app_models",... | a context manager that yields a secret model defined at runtime . | train | false |
55,402 | def mark_as_cover(container, name):
if (name not in container.mime_map):
raise ValueError((u'Cannot mark %s as cover as it does not exist' % name))
mt = container.mime_map[name]
if (not is_raster_image(mt)):
raise ValueError((u'Cannot mark %s as the cover image as it is not a raster image' % name))
if (container.book_type == u'azw3'):
mark_as_cover_azw3(container, name)
else:
mark_as_cover_epub(container, name)
| [
"def",
"mark_as_cover",
"(",
"container",
",",
"name",
")",
":",
"if",
"(",
"name",
"not",
"in",
"container",
".",
"mime_map",
")",
":",
"raise",
"ValueError",
"(",
"(",
"u'Cannot mark %s as cover as it does not exist'",
"%",
"name",
")",
")",
"mt",
"=",
"co... | mark the specified image as the cover image . | train | false |
55,403 | def is_unit(xblock, parent_xblock=None):
if (xblock.category == 'vertical'):
if (parent_xblock is None):
parent_xblock = get_parent_xblock(xblock)
parent_category = (parent_xblock.category if parent_xblock else None)
return (parent_category == 'sequential')
return False
| [
"def",
"is_unit",
"(",
"xblock",
",",
"parent_xblock",
"=",
"None",
")",
":",
"if",
"(",
"xblock",
".",
"category",
"==",
"'vertical'",
")",
":",
"if",
"(",
"parent_xblock",
"is",
"None",
")",
":",
"parent_xblock",
"=",
"get_parent_xblock",
"(",
"xblock",
... | returns true if the specified xblock is a vertical that is treated as a unit . | train | false |
55,405 | def assert_no_element_by_id(context, _id, wait_time=MAX_WAIT_FOR_UNEXPECTED_ELEMENT):
_assert_no_element_by(context, By.ID, _id, wait_time)
| [
"def",
"assert_no_element_by_id",
"(",
"context",
",",
"_id",
",",
"wait_time",
"=",
"MAX_WAIT_FOR_UNEXPECTED_ELEMENT",
")",
":",
"_assert_no_element_by",
"(",
"context",
",",
"By",
".",
"ID",
",",
"_id",
",",
"wait_time",
")"
] | assert that no element is found . | train | false |
55,406 | @flake8ext
def use_jsonutils(logical_line, filename):
msg = 'N321: jsonutils.%(fun)s must be used instead of json.%(fun)s'
json_check_skipped_patterns = ['neutron/plugins/ml2/drivers/openvswitch/agent/xenapi/etc/xapi.d/plugins/netwrap']
for pattern in json_check_skipped_patterns:
if (pattern in filename):
return
if ('json.' in logical_line):
json_funcs = ['dumps(', 'dump(', 'loads(', 'load(']
for f in json_funcs:
pos = logical_line.find(('json.%s' % f))
if (pos != (-1)):
(yield (pos, (msg % {'fun': f[:(-1)]})))
| [
"@",
"flake8ext",
"def",
"use_jsonutils",
"(",
"logical_line",
",",
"filename",
")",
":",
"msg",
"=",
"'N321: jsonutils.%(fun)s must be used instead of json.%(fun)s'",
"json_check_skipped_patterns",
"=",
"[",
"'neutron/plugins/ml2/drivers/openvswitch/agent/xenapi/etc/xapi.d/plugins/n... | n321 - use jsonutils instead of json . | train | false |
55,407 | def addAssemblyCage(derivation, negatives, positives):
addCageGroove(derivation, negatives, positives)
for pegCenterX in derivation.pegCenterXs:
addPositivePeg(derivation, positives, pegCenterX, (- derivation.pegY))
addPositivePeg(derivation, positives, pegCenterX, derivation.pegY)
translate.translateNegativesPositives(negatives, positives, Vector3(0.0, (- derivation.halfSeparationWidth)))
femaleNegatives = []
femalePositives = []
addCageGroove(derivation, femaleNegatives, femalePositives)
for pegCenterX in derivation.pegCenterXs:
addNegativePeg(derivation, femaleNegatives, pegCenterX, (- derivation.pegY))
addNegativePeg(derivation, femaleNegatives, pegCenterX, derivation.pegY)
translate.translateNegativesPositives(femaleNegatives, femalePositives, Vector3(0.0, derivation.halfSeparationWidth))
negatives += femaleNegatives
positives += femalePositives
| [
"def",
"addAssemblyCage",
"(",
"derivation",
",",
"negatives",
",",
"positives",
")",
":",
"addCageGroove",
"(",
"derivation",
",",
"negatives",
",",
"positives",
")",
"for",
"pegCenterX",
"in",
"derivation",
".",
"pegCenterXs",
":",
"addPositivePeg",
"(",
"deri... | add assembly linear bearing cage . | train | false |
55,408 | @treeio_login_required
@handle_response_format
def order_invoice_view(request, order_id, response_format='html'):
order = get_object_or_404(SaleOrder, pk=order_id)
if ((not request.user.profile.has_permission(order)) and (not request.user.profile.is_admin('treeio.sales'))):
return user_denied(request, message="You don't have access to this Sale")
ordered_products = order.orderedproduct_set.filter(trash=False)
try:
conf = ModuleSetting.get_for_module('treeio.finance', 'my_company')[0]
my_company = Contact.objects.get(pk=long(conf.value))
except:
my_company = None
return render_to_response('sales/order_invoice_view', {'order': order, 'ordered_products': ordered_products, 'my_company': my_company}, context_instance=RequestContext(request), response_format=response_format)
| [
"@",
"treeio_login_required",
"@",
"handle_response_format",
"def",
"order_invoice_view",
"(",
"request",
",",
"order_id",
",",
"response_format",
"=",
"'html'",
")",
":",
"order",
"=",
"get_object_or_404",
"(",
"SaleOrder",
",",
"pk",
"=",
"order_id",
")",
"if",
... | order view as invoice . | train | false |
55,409 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get the repository constructor . | train | false |
55,410 | def json_underscore(body, charset='utf-8', **kwargs):
return _underscore_dict(json(body, charset=charset))
| [
"def",
"json_underscore",
"(",
"body",
",",
"charset",
"=",
"'utf-8'",
",",
"**",
"kwargs",
")",
":",
"return",
"_underscore_dict",
"(",
"json",
"(",
"body",
",",
"charset",
"=",
"charset",
")",
")"
] | converts json formatted date to native python objects . | train | false |
55,411 | def git_hook():
(_, files_modified, _) = run('git diff-index --cached --name-only HEAD')
options = parse_options()
setup_logger(options)
candidates = list(map(str, files_modified))
if candidates:
process_paths(options, candidates=candidates)
| [
"def",
"git_hook",
"(",
")",
":",
"(",
"_",
",",
"files_modified",
",",
"_",
")",
"=",
"run",
"(",
"'git diff-index --cached --name-only HEAD'",
")",
"options",
"=",
"parse_options",
"(",
")",
"setup_logger",
"(",
"options",
")",
"candidates",
"=",
"list",
"... | run pylama after git commit . | train | false |
55,412 | def role_delete(role_id=None, name=None, profile=None, **connection_args):
kstone = auth(profile, **connection_args)
if name:
for role in kstone.roles.list():
if (role.name == name):
role_id = role.id
break
if (not role_id):
return {'Error': 'Unable to resolve role id'}
role = kstone.roles.get(role_id)
kstone.roles.delete(role)
ret = 'Role ID {0} deleted'.format(role_id)
if name:
ret += ' ({0})'.format(name)
return ret
| [
"def",
"role_delete",
"(",
"role_id",
"=",
"None",
",",
"name",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"**",
"connection_args",
")",
":",
"kstone",
"=",
"auth",
"(",
"profile",
",",
"**",
"connection_args",
")",
"if",
"name",
":",
"for",
"role"... | delete a role cli examples: . | train | true |
55,413 | def _needs_eeg_average_ref_proj(info):
eeg_sel = pick_types(info, meg=False, eeg=True, ref_meg=False, exclude='bads')
return ((len(eeg_sel) > 0) and (not info['custom_ref_applied']) and (not _has_eeg_average_ref_proj(info['projs'])))
| [
"def",
"_needs_eeg_average_ref_proj",
"(",
"info",
")",
":",
"eeg_sel",
"=",
"pick_types",
"(",
"info",
",",
"meg",
"=",
"False",
",",
"eeg",
"=",
"True",
",",
"ref_meg",
"=",
"False",
",",
"exclude",
"=",
"'bads'",
")",
"return",
"(",
"(",
"len",
"(",... | determine if the eeg needs an averge eeg reference . | train | false |
55,414 | @frappe.whitelist()
def add_tag(tag, dt, dn, color=None):
DocTags(dt).add(dn, tag)
return tag
| [
"@",
"frappe",
".",
"whitelist",
"(",
")",
"def",
"add_tag",
"(",
"tag",
",",
"dt",
",",
"dn",
",",
"color",
"=",
"None",
")",
":",
"DocTags",
"(",
"dt",
")",
".",
"add",
"(",
"dn",
",",
"tag",
")",
"return",
"tag"
] | adds a new tag to a record . | train | false |
55,415 | def CreateBudget(client):
budget_service = client.GetService('BudgetService', version='v201609')
budget = {'name': ('Interplanetary Cruise App Budget #%s' % uuid.uuid4()), 'amount': {'microAmount': '50000000'}, 'deliveryMethod': 'STANDARD', 'isExplicitlyShared': False}
budget_operations = [{'operator': 'ADD', 'operand': budget}]
budget_id = budget_service.mutate(budget_operations)['value'][0]['budgetId']
return budget_id
| [
"def",
"CreateBudget",
"(",
"client",
")",
":",
"budget_service",
"=",
"client",
".",
"GetService",
"(",
"'BudgetService'",
",",
"version",
"=",
"'v201609'",
")",
"budget",
"=",
"{",
"'name'",
":",
"(",
"'Interplanetary Cruise App Budget #%s'",
"%",
"uuid",
".",... | creates a budget and returns its budgetid . | train | true |
55,417 | def _nova_to_osvif_network(network):
netobj = objects.network.Network(id=network['id'], bridge_interface=network.get_meta('bridge_interface'), subnets=_nova_to_osvif_subnets(network['subnets']))
if (network['bridge'] is not None):
netobj.bridge = network['bridge']
if (network['label'] is not None):
netobj.label = network['label']
if (network.get_meta('mtu') is not None):
netobj.mtu = network.get_meta('mtu')
if (network.get_meta('multi_host') is not None):
netobj.multi_host = network.get_meta('multi_host')
if (network.get_meta('should_create_bridge') is not None):
netobj.should_provide_bridge = network.get_meta('should_create_bridge')
if (network.get_meta('should_create_vlan') is not None):
netobj.should_provide_vlan = network.get_meta('should_create_vlan')
if (network.get_meta('vlan') is None):
raise exception.NovaException((_('Missing vlan number in %s') % network))
netobj.vlan = network.get_meta('vlan')
return netobj
| [
"def",
"_nova_to_osvif_network",
"(",
"network",
")",
":",
"netobj",
"=",
"objects",
".",
"network",
".",
"Network",
"(",
"id",
"=",
"network",
"[",
"'id'",
"]",
",",
"bridge_interface",
"=",
"network",
".",
"get_meta",
"(",
"'bridge_interface'",
")",
",",
... | convert nova network object into os_vif object . | train | false |
55,418 | def delete_disk(kwargs=None, call=None):
if (call != 'function'):
raise SaltCloudSystemExit('The delete_disk function must be called with -f or --function.')
if ((not kwargs) or ('disk_name' not in kwargs)):
log.error('A disk_name must be specified when deleting a disk.')
return False
conn = get_conn()
disk = conn.ex_get_volume(kwargs.get('disk_name'))
__utils__['cloud.fire_event']('event', 'delete disk', 'salt/cloud/disk/deleting', args={'name': disk.name, 'location': disk.extra['zone'].name, 'size': disk.size}, sock_dir=__opts__['sock_dir'], transport=__opts__['transport'])
try:
result = conn.destroy_volume(disk)
except ResourceInUseError as exc:
log.error('Disk {0} is in use and must be detached before deleting.\nThe following exception was thrown by libcloud:\n{1}'.format(disk.name, exc), exc_info_on_loglevel=logging.DEBUG)
return False
__utils__['cloud.fire_event']('event', 'deleted disk', 'salt/cloud/disk/deleted', args={'name': disk.name, 'location': disk.extra['zone'].name, 'size': disk.size}, sock_dir=__opts__['sock_dir'], transport=__opts__['transport'])
return result
| [
"def",
"delete_disk",
"(",
"kwargs",
"=",
"None",
",",
"call",
"=",
"None",
")",
":",
"if",
"(",
"call",
"!=",
"'function'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The delete_disk function must be called with -f or --function.'",
")",
"if",
"(",
"(",
"... | permanently delete a persistent disk . | train | true |
55,420 | def externals_finder(dirname, filename):
found = False
f = open(filename, 'rt')
for line in iter(f.readline, ''):
parts = line.split()
if (len(parts) == 2):
(kind, length) = parts
data = f.read(int(length))
if ((kind == 'K') and (data == 'svn:externals')):
found = True
elif ((kind == 'V') and found):
f.close()
break
else:
f.close()
return
for line in data.splitlines():
parts = line.split()
if parts:
(yield joinpath(dirname, parts[0]))
| [
"def",
"externals_finder",
"(",
"dirname",
",",
"filename",
")",
":",
"found",
"=",
"False",
"f",
"=",
"open",
"(",
"filename",
",",
"'rt'",
")",
"for",
"line",
"in",
"iter",
"(",
"f",
".",
"readline",
",",
"''",
")",
":",
"parts",
"=",
"line",
"."... | find any svn:externals directories . | train | true |
55,421 | def quiet_close(closable):
try:
closable.close()
except Exception:
logger.debug(u'Exception while closing', exc_info=True)
| [
"def",
"quiet_close",
"(",
"closable",
")",
":",
"try",
":",
"closable",
".",
"close",
"(",
")",
"except",
"Exception",
":",
"logger",
".",
"debug",
"(",
"u'Exception while closing'",
",",
"exc_info",
"=",
"True",
")"
] | quietly closes a closable object without throwing an exception . | train | false |
55,423 | def contracted_edge(G, edge, self_loops=True):
if (not G.has_edge(*edge)):
raise ValueError('Edge {0} does not exist in graph G; cannot contract it'.format(edge))
return contracted_nodes(G, self_loops=self_loops, *edge)
| [
"def",
"contracted_edge",
"(",
"G",
",",
"edge",
",",
"self_loops",
"=",
"True",
")",
":",
"if",
"(",
"not",
"G",
".",
"has_edge",
"(",
"*",
"edge",
")",
")",
":",
"raise",
"ValueError",
"(",
"'Edge {0} does not exist in graph G; cannot contract it'",
".",
"... | returns the graph that results from contracting the specified edge . | train | false |
55,424 | def base_vectors(n):
n = (n / np.sqrt(np.square(n).sum(axis=(-1))))
if (abs(n[0]) == 1):
l = np.r_[(n[2], 0, (- n[0]))]
else:
l = np.r_[(0, n[2], (- n[1]))]
l = (l / np.sqrt(np.square(l).sum(axis=(-1))))
m = np.cross(n, l)
return (n, l, m)
| [
"def",
"base_vectors",
"(",
"n",
")",
":",
"n",
"=",
"(",
"n",
"/",
"np",
".",
"sqrt",
"(",
"np",
".",
"square",
"(",
"n",
")",
".",
"sum",
"(",
"axis",
"=",
"(",
"-",
"1",
")",
")",
")",
")",
"if",
"(",
"abs",
"(",
"n",
"[",
"0",
"]",
... | returns 3 orthognal base vectors . | train | false |
55,426 | def instance_tag_delete_all(context, instance_uuid):
return IMPL.instance_tag_delete_all(context, instance_uuid)
| [
"def",
"instance_tag_delete_all",
"(",
"context",
",",
"instance_uuid",
")",
":",
"return",
"IMPL",
".",
"instance_tag_delete_all",
"(",
"context",
",",
"instance_uuid",
")"
] | delete all tags from the instance . | train | false |
55,427 | def invalidate_star_import_cache(path):
try:
parser_cache_item = parser_cache[path]
except KeyError:
pass
else:
_invalidate_star_import_cache_module(parser_cache_item.parser.module)
| [
"def",
"invalidate_star_import_cache",
"(",
"path",
")",
":",
"try",
":",
"parser_cache_item",
"=",
"parser_cache",
"[",
"path",
"]",
"except",
"KeyError",
":",
"pass",
"else",
":",
"_invalidate_star_import_cache_module",
"(",
"parser_cache_item",
".",
"parser",
"."... | on success returns true . | train | false |
55,428 | def test_system_dynamic():
print 'TODO'
| [
"def",
"test_system_dynamic",
"(",
")",
":",
"print",
"'TODO'"
] | URL only a sanity check here . | train | false |
55,430 | def test_custom_model_subclass():
@custom_model
def model_a(x, a=1):
return (x * a)
class model_b(model_a, ):
@classmethod
def evaluate(cls, x, a):
return (- super(model_b, cls).evaluate(x, a))
b = model_b()
assert (b.param_names == (u'a',))
assert (b.a == 1)
assert (b(1) == (-1))
sig = signature(model_b.__init__)
assert (list(sig.parameters.keys()) == [u'self', u'a', u'kwargs'])
sig = signature(model_b.__call__)
assert (list(sig.parameters.keys()) == [u'self', u'x', u'model_set_axis'])
| [
"def",
"test_custom_model_subclass",
"(",
")",
":",
"@",
"custom_model",
"def",
"model_a",
"(",
"x",
",",
"a",
"=",
"1",
")",
":",
"return",
"(",
"x",
"*",
"a",
")",
"class",
"model_b",
"(",
"model_a",
",",
")",
":",
"@",
"classmethod",
"def",
"evalu... | test that custom models can be subclassed . | train | false |
55,431 | def get_scores(video):
if isinstance(video, Episode):
return episode_scores
elif isinstance(video, Movie):
return movie_scores
raise ValueError('video must be an instance of Episode or Movie')
| [
"def",
"get_scores",
"(",
"video",
")",
":",
"if",
"isinstance",
"(",
"video",
",",
"Episode",
")",
":",
"return",
"episode_scores",
"elif",
"isinstance",
"(",
"video",
",",
"Movie",
")",
":",
"return",
"movie_scores",
"raise",
"ValueError",
"(",
"'video mus... | get the scores dict for the given video . | train | true |
55,433 | def int_to_str(value, length=2):
try:
int(value)
except:
raise ValueError('expected an integer value')
content = str(value)
while (len(content) < length):
content = ('0' + content)
return content
| [
"def",
"int_to_str",
"(",
"value",
",",
"length",
"=",
"2",
")",
":",
"try",
":",
"int",
"(",
"value",
")",
"except",
":",
"raise",
"ValueError",
"(",
"'expected an integer value'",
")",
"content",
"=",
"str",
"(",
"value",
")",
"while",
"(",
"len",
"(... | converts integer to string eg 3 to "03" . | train | false |
55,434 | def validate_payload(payload, api_model, check_required=True):
if check_required:
for key in api_model:
if (api_model[key].required and (key not in payload)):
raise ValidationError(field=key, message="Required field '{}' missing".format(key))
for key in payload:
field = api_model[key]
if isinstance(field, fields.List):
field = field.container
data = payload[key]
elif isinstance(field, fields.Nested):
if payload[key]:
validate_payload(payload[key], field.model)
else:
data = [payload[key]]
if (isinstance(field, CustomField) and hasattr(field, 'validate')):
field.payload = payload
for i in data:
if (not field.validate(i)):
raise ValidationError(field=key, message=(field.validation_error % ("'%s'" % key)))
| [
"def",
"validate_payload",
"(",
"payload",
",",
"api_model",
",",
"check_required",
"=",
"True",
")",
":",
"if",
"check_required",
":",
"for",
"key",
"in",
"api_model",
":",
"if",
"(",
"api_model",
"[",
"key",
"]",
".",
"required",
"and",
"(",
"key",
"no... | validate payload against an api_model . | train | false |
55,435 | def serialize_item(collection, item):
if ((item.name is None) or (item.name == '')):
raise exceptions.RuntimeError('name unset for item!')
if (collection.collection_type() in ['mgmtclass']):
filename = ('/var/lib/cobbler/collections/%ses/%s' % (collection.collection_type(), item.name))
else:
filename = ('/var/lib/cobbler/collections/%ss/%s' % (collection.collection_type(), item.name))
_dict = item.to_dict()
if capi.CobblerAPI().settings().serializer_pretty_json:
sort_keys = True
indent = 4
else:
sort_keys = False
indent = None
filename += '.json'
_dict = item.to_dict()
fd = open(filename, 'w+')
data = simplejson.dumps(_dict, encoding='utf-8', sort_keys=sort_keys, indent=indent)
fd.write(data)
fd.close()
| [
"def",
"serialize_item",
"(",
"collection",
",",
"item",
")",
":",
"if",
"(",
"(",
"item",
".",
"name",
"is",
"None",
")",
"or",
"(",
"item",
".",
"name",
"==",
"''",
")",
")",
":",
"raise",
"exceptions",
".",
"RuntimeError",
"(",
"'name unset for item... | save a collection item to file system . | train | false |
55,436 | def get_raising_file_and_line(tb=None):
if (not tb):
tb = sys.exc_info()[2]
(filename, lineno, _context, _line) = traceback.extract_tb(tb)[(-1)]
return (filename, lineno)
| [
"def",
"get_raising_file_and_line",
"(",
"tb",
"=",
"None",
")",
":",
"if",
"(",
"not",
"tb",
")",
":",
"tb",
"=",
"sys",
".",
"exc_info",
"(",
")",
"[",
"2",
"]",
"(",
"filename",
",",
"lineno",
",",
"_context",
",",
"_line",
")",
"=",
"traceback"... | return the file and line number of the statement that raised the tb . | train | false |
55,437 | def quitWindows(event=None):
global globalRepositoryDialogListTable
globalRepositoryDialogValues = euclidean.getListTableElements(globalRepositoryDialogListTable)
for globalRepositoryDialogValue in globalRepositoryDialogValues:
quitWindow(globalRepositoryDialogValue.root)
| [
"def",
"quitWindows",
"(",
"event",
"=",
"None",
")",
":",
"global",
"globalRepositoryDialogListTable",
"globalRepositoryDialogValues",
"=",
"euclidean",
".",
"getListTableElements",
"(",
"globalRepositoryDialogListTable",
")",
"for",
"globalRepositoryDialogValue",
"in",
"g... | quit all windows . | train | false |
55,439 | @app.route('/scans/<int:scan_id>/exceptions/', methods=['GET'])
@requires_auth
def list_exceptions(scan_id):
scan_info = get_scan_info_from_id(scan_id)
if (scan_info is None):
abort(404, 'Scan not found')
data = []
all_exceptions = scan_info.w3af_core.exception_handler.get_all_exceptions()
for (exception_id, exception_data) in enumerate(all_exceptions):
data.append(exception_to_json(exception_data, scan_id, exception_id))
return jsonify({'items': data})
| [
"@",
"app",
".",
"route",
"(",
"'/scans/<int:scan_id>/exceptions/'",
",",
"methods",
"=",
"[",
"'GET'",
"]",
")",
"@",
"requires_auth",
"def",
"list_exceptions",
"(",
"scan_id",
")",
":",
"scan_info",
"=",
"get_scan_info_from_id",
"(",
"scan_id",
")",
"if",
"(... | list all exceptions found during a scan :return: a json containing a list of: - exception resource url - the exceptions id - exception string - exception file name - exception line number . | train | false |
55,440 | def _check_even_rewrite(func, arg):
return (func(arg).args[0] == (- arg))
| [
"def",
"_check_even_rewrite",
"(",
"func",
",",
"arg",
")",
":",
"return",
"(",
"func",
"(",
"arg",
")",
".",
"args",
"[",
"0",
"]",
"==",
"(",
"-",
"arg",
")",
")"
] | checks that the expr has been rewritten using f -> f(x) arg : -x . | train | false |
55,441 | def test_hsl_to_rgb_part_3():
assert (hsl_to_rgb(6120, 100, 50) == (255, 0, 0))
assert (hsl_to_rgb((-9660), 100, 50) == (255, 255, 0))
assert (hsl_to_rgb(99840, 100, 50) == (0, 255, 0))
assert (hsl_to_rgb((-900), 100, 50) == (0, 255, 255))
assert (hsl_to_rgb((-104880), 100, 50) == (0, 0, 255))
assert (hsl_to_rgb(2820, 100, 50) == (255, 0, 255))
| [
"def",
"test_hsl_to_rgb_part_3",
"(",
")",
":",
"assert",
"(",
"hsl_to_rgb",
"(",
"6120",
",",
"100",
",",
"50",
")",
"==",
"(",
"255",
",",
"0",
",",
"0",
")",
")",
"assert",
"(",
"hsl_to_rgb",
"(",
"(",
"-",
"9660",
")",
",",
"100",
",",
"50",
... | test hsl to rgb color function . | train | false |
55,442 | def test_cx_Oracle():
if (('ORACLE_HOME' not in os.environ) and ('ORACLE_INSTANTCLIENT_HOME' not in os.environ)):
raise SkipTest
try:
import cx_Oracle
return
except ImportError as ex:
if ('No module named' in ex.message):
assert_true(False, 'cx_Oracle skipped its build. This happens if env var ORACLE_HOME or ORACLE_INSTANTCLIENT_HOME is not defined. So ignore this test failure if your build does not need to work with an oracle backend.')
| [
"def",
"test_cx_Oracle",
"(",
")",
":",
"if",
"(",
"(",
"'ORACLE_HOME'",
"not",
"in",
"os",
".",
"environ",
")",
"and",
"(",
"'ORACLE_INSTANTCLIENT_HOME'",
"not",
"in",
"os",
".",
"environ",
")",
")",
":",
"raise",
"SkipTest",
"try",
":",
"import",
"cx_O... | tests that cx_oracle is built correctly . | train | false |
55,443 | def default_channel():
try:
chan = _open_session()
except ssh.SSHException as err:
if (str(err) == 'SSH session not active'):
connections[env.host_string].close()
del connections[env.host_string]
chan = _open_session()
else:
raise
chan.settimeout(0.1)
chan.input_enabled = True
return chan
| [
"def",
"default_channel",
"(",
")",
":",
"try",
":",
"chan",
"=",
"_open_session",
"(",
")",
"except",
"ssh",
".",
"SSHException",
"as",
"err",
":",
"if",
"(",
"str",
"(",
"err",
")",
"==",
"'SSH session not active'",
")",
":",
"connections",
"[",
"env",... | return a channel object based on env . | train | false |
55,444 | def global_fixes():
for function in list(globals().values()):
if inspect.isfunction(function):
arguments = inspect.getargspec(function)[0]
if (arguments[:1] != [u'source']):
continue
code = extract_code_from_function(function)
if code:
(yield (code, function))
| [
"def",
"global_fixes",
"(",
")",
":",
"for",
"function",
"in",
"list",
"(",
"globals",
"(",
")",
".",
"values",
"(",
")",
")",
":",
"if",
"inspect",
".",
"isfunction",
"(",
"function",
")",
":",
"arguments",
"=",
"inspect",
".",
"getargspec",
"(",
"f... | yield multiple tuples . | train | false |
55,445 | def _memoize(func, *args, **opts):
if opts:
key = (args, frozenset(opts.items()))
else:
key = args
cache = func.cache
try:
result = cache[key]
except KeyError:
result = cache[key] = func(*args, **opts)
return result
| [
"def",
"_memoize",
"(",
"func",
",",
"*",
"args",
",",
"**",
"opts",
")",
":",
"if",
"opts",
":",
"key",
"=",
"(",
"args",
",",
"frozenset",
"(",
"opts",
".",
"items",
"(",
")",
")",
")",
"else",
":",
"key",
"=",
"args",
"cache",
"=",
"func",
... | implements memoized cache lookups . | train | false |
55,446 | def _get_css_imports_cssutils(data, inline=False):
try:
import cssutils
except (ImportError, re.error):
return None
parser = cssutils.CSSParser(loglevel=100, fetcher=(lambda url: (None, '')), validate=False)
if (not inline):
sheet = parser.parseString(data)
return list(cssutils.getUrls(sheet))
else:
urls = []
declaration = parser.parseStyle(data)
for prop in declaration:
for value in prop.propertyValue:
if isinstance(value, cssutils.css.URIValue):
if value.uri:
urls.append(value.uri)
return urls
| [
"def",
"_get_css_imports_cssutils",
"(",
"data",
",",
"inline",
"=",
"False",
")",
":",
"try",
":",
"import",
"cssutils",
"except",
"(",
"ImportError",
",",
"re",
".",
"error",
")",
":",
"return",
"None",
"parser",
"=",
"cssutils",
".",
"CSSParser",
"(",
... | return all assets that are referenced in the given css document . | train | false |
55,447 | def make_script_path(script):
s_path = None
path = cfg.script_dir.get_path()
if (path and script):
if (script.lower() not in ('none', 'default')):
s_path = os.path.join(path, script)
if (not os.path.exists(s_path)):
s_path = None
return s_path
| [
"def",
"make_script_path",
"(",
"script",
")",
":",
"s_path",
"=",
"None",
"path",
"=",
"cfg",
".",
"script_dir",
".",
"get_path",
"(",
")",
"if",
"(",
"path",
"and",
"script",
")",
":",
"if",
"(",
"script",
".",
"lower",
"(",
")",
"not",
"in",
"("... | return full script path . | train | false |
55,451 | def InstallDriver(kext_path):
km = objc.KextManager()
cf_kext_path = km.PyStringToCFString(kext_path)
kext_url = km.dll.CFURLCreateWithFileSystemPath(objc.CF_DEFAULT_ALLOCATOR, cf_kext_path, objc.POSIX_PATH_STYLE, True)
status = km.iokit.KextManagerLoadKextWithURL(kext_url, None)
km.dll.CFRelease(kext_url)
km.dll.CFRelease(cf_kext_path)
if (status is not objc.OS_SUCCESS):
raise OSError('Failed to load kext at {0}: {1}'.format(kext_path, status))
| [
"def",
"InstallDriver",
"(",
"kext_path",
")",
":",
"km",
"=",
"objc",
".",
"KextManager",
"(",
")",
"cf_kext_path",
"=",
"km",
".",
"PyStringToCFString",
"(",
"kext_path",
")",
"kext_url",
"=",
"km",
".",
"dll",
".",
"CFURLCreateWithFileSystemPath",
"(",
"o... | calls into the iokit to load a kext by file-system path . | train | true |
55,453 | def setup_form_view(view, request, form, *args, **kwargs):
view.request = request
try:
view.request.user = request.user
except AttributeError:
view.request.user = UserFactory()
view.args = args
view.kwargs = kwargs
view.form = form
return view
| [
"def",
"setup_form_view",
"(",
"view",
",",
"request",
",",
"form",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"view",
".",
"request",
"=",
"request",
"try",
":",
"view",
".",
"request",
".",
"user",
"=",
"request",
".",
"user",
"except",
"Attr... | mimic as_view and with forms to skip some of the context . | train | false |
55,454 | def parse_bdist_wininst(name):
lower = name.lower()
(base, py_ver) = (None, None)
if lower.endswith('.exe'):
if lower.endswith('.win32.exe'):
base = name[:(-10)]
elif lower.startswith('.win32-py', (-16)):
py_ver = name[(-7):(-4)]
base = name[:(-16)]
return (base, py_ver)
| [
"def",
"parse_bdist_wininst",
"(",
"name",
")",
":",
"lower",
"=",
"name",
".",
"lower",
"(",
")",
"(",
"base",
",",
"py_ver",
")",
"=",
"(",
"None",
",",
"None",
")",
"if",
"lower",
".",
"endswith",
"(",
"'.exe'",
")",
":",
"if",
"lower",
".",
"... | return or for possible . | train | false |
55,456 | def scrub_text(text):
scrubbed_text = text.rstrip().replace('\n', '\\n').replace(' DCTB ', (' ' * 4))
return scrubbed_text
| [
"def",
"scrub_text",
"(",
"text",
")",
":",
"scrubbed_text",
"=",
"text",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"'\\n'",
",",
"'\\\\n'",
")",
".",
"replace",
"(",
"' DCTB '",
",",
"(",
"' '",
"*",
"4",
")",
")",
"return",
"scrubbed_text"
] | cleans up text . | train | false |
55,457 | def _get_immediate_cls_attr(cls, attrname, strict=False):
if (not issubclass(cls, object)):
return None
for base in cls.__mro__:
_is_declarative_inherits = hasattr(base, '_decl_class_registry')
if ((attrname in base.__dict__) and ((base is cls) or (((base in cls.__bases__) if strict else True) and (not _is_declarative_inherits)))):
return getattr(base, attrname)
else:
return None
| [
"def",
"_get_immediate_cls_attr",
"(",
"cls",
",",
"attrname",
",",
"strict",
"=",
"False",
")",
":",
"if",
"(",
"not",
"issubclass",
"(",
"cls",
",",
"object",
")",
")",
":",
"return",
"None",
"for",
"base",
"in",
"cls",
".",
"__mro__",
":",
"_is_decl... | return an attribute of the class that is either present directly on the class . | train | false |
55,458 | def wiki_escape(s):
ret = []
for word in s.split():
if re.match('[A-Z]+[a-z]+[A-Z]', word):
word = ('!%s' % word)
ret.append(word)
return ' '.join(ret)
| [
"def",
"wiki_escape",
"(",
"s",
")",
":",
"ret",
"=",
"[",
"]",
"for",
"word",
"in",
"s",
".",
"split",
"(",
")",
":",
"if",
"re",
".",
"match",
"(",
"'[A-Z]+[a-z]+[A-Z]'",
",",
"word",
")",
":",
"word",
"=",
"(",
"'!%s'",
"%",
"word",
")",
"re... | detect wikisyntax and escape it . | train | false |
55,459 | def getaddresses(fieldvalues):
all = COMMASPACE.join(fieldvalues)
a = _AddressList(all)
return a.addresslist
| [
"def",
"getaddresses",
"(",
"fieldvalues",
")",
":",
"all",
"=",
"COMMASPACE",
".",
"join",
"(",
"fieldvalues",
")",
"a",
"=",
"_AddressList",
"(",
"all",
")",
"return",
"a",
".",
"addresslist"
] | return a list of for each fieldvalue . | train | true |
55,461 | def _extend_external_network_default(core_plugin, net_res, net_db):
if (net_db.external is not None):
net_res[IS_DEFAULT] = net_db.external.is_default
return net_res
| [
"def",
"_extend_external_network_default",
"(",
"core_plugin",
",",
"net_res",
",",
"net_db",
")",
":",
"if",
"(",
"net_db",
".",
"external",
"is",
"not",
"None",
")",
":",
"net_res",
"[",
"IS_DEFAULT",
"]",
"=",
"net_db",
".",
"external",
".",
"is_default",... | add is_default field to show response . | train | false |
55,462 | def modify_node(hostname, username, password, name, connection_limit=None, description=None, dynamic_ratio=None, logging=None, monitor=None, rate_limit=None, ratio=None, session=None, state=None, trans_label=None):
params = {'connection-limit': connection_limit, 'description': description, 'dynamic-ratio': dynamic_ratio, 'logging': logging, 'monitor': monitor, 'rate-limit': rate_limit, 'ratio': ratio, 'session': session, 'state': state}
bigip_session = _build_session(username, password, trans_label)
payload = _loop_payload(params)
payload['name'] = name
try:
response = bigip_session.put((BIG_IP_URL_BASE.format(host=hostname) + '/ltm/node/{name}'.format(name=name)), data=json.dumps(payload))
except requests.exceptions.ConnectionError as e:
return _load_connection_error(hostname, e)
return _load_response(response)
| [
"def",
"modify_node",
"(",
"hostname",
",",
"username",
",",
"password",
",",
"name",
",",
"connection_limit",
"=",
"None",
",",
"description",
"=",
"None",
",",
"dynamic_ratio",
"=",
"None",
",",
"logging",
"=",
"None",
",",
"monitor",
"=",
"None",
",",
... | a function to connect to a bigip device and modify an existing node . | train | true |
55,463 | def set_edit_mode(request, flag):
if (flag and could_edit(request)):
request.session[EDIT_FLAG_NAME] = True
else:
request.session.pop(EDIT_FLAG_NAME, None)
| [
"def",
"set_edit_mode",
"(",
"request",
",",
"flag",
")",
":",
"if",
"(",
"flag",
"and",
"could_edit",
"(",
"request",
")",
")",
":",
"request",
".",
"session",
"[",
"EDIT_FLAG_NAME",
"]",
"=",
"True",
"else",
":",
"request",
".",
"session",
".",
"pop"... | enable or disable edit mode for the request . | train | false |
55,465 | def getVector3RemoveByPrefix(prefix, vector3, xmlElement):
vector3RemoveByPrefix = getVector3ByPrefix(vector3, prefix, xmlElement)
euclidean.removePrefixFromDictionary(xmlElement.attributeDictionary, prefix)
return vector3RemoveByPrefix
| [
"def",
"getVector3RemoveByPrefix",
"(",
"prefix",
",",
"vector3",
",",
"xmlElement",
")",
":",
"vector3RemoveByPrefix",
"=",
"getVector3ByPrefix",
"(",
"vector3",
",",
"prefix",
",",
"xmlElement",
")",
"euclidean",
".",
"removePrefixFromDictionary",
"(",
"xmlElement",... | get vector3 from prefix and xml element . | train | false |
55,466 | def _escape_jid(jid):
jid = str(jid)
jid = re.sub("'*", '', jid)
return jid
| [
"def",
"_escape_jid",
"(",
"jid",
")",
":",
"jid",
"=",
"str",
"(",
"jid",
")",
"jid",
"=",
"re",
".",
"sub",
"(",
"\"'*\"",
",",
"''",
",",
"jid",
")",
"return",
"jid"
] | do proper formatting of the jid . | train | false |
55,468 | def unparse_vs(tup):
return '.'.join(map(str, tup))
| [
"def",
"unparse_vs",
"(",
"tup",
")",
":",
"return",
"'.'",
".",
"join",
"(",
"map",
"(",
"str",
",",
"tup",
")",
")"
] | version list to string . | train | false |
55,470 | @apply_to_binary_file
def xmlminify(data):
parser = lxml.etree.XMLParser(remove_blank_text=True)
newdata = lxml.etree.XML(data, parser=parser)
return lxml.etree.tostring(newdata, encoding='utf-8', method='xml', xml_declaration=True)
| [
"@",
"apply_to_binary_file",
"def",
"xmlminify",
"(",
"data",
")",
":",
"parser",
"=",
"lxml",
".",
"etree",
".",
"XMLParser",
"(",
"remove_blank_text",
"=",
"True",
")",
"newdata",
"=",
"lxml",
".",
"etree",
".",
"XML",
"(",
"data",
",",
"parser",
"=",
... | minify xml files . | train | false |
55,471 | def qd(A, B, output='real', lwork=None, sort=None, overwrite_a=False, overwrite_b=False, check_finite=True):
(result, _) = _qd(A, B, output=output, lwork=lwork, sort=sort, overwrite_a=overwrite_a, overwrite_b=overwrite_b, check_finite=check_finite)
return (result[0], result[1], result[(-4)], result[(-3)])
| [
"def",
"qd",
"(",
"A",
",",
"B",
",",
"output",
"=",
"'real'",
",",
"lwork",
"=",
"None",
",",
"sort",
"=",
"None",
",",
"overwrite_a",
"=",
"False",
",",
"overwrite_b",
"=",
"False",
",",
"check_finite",
"=",
"True",
")",
":",
"(",
"result",
",",
... | qz decomposition for generalized eigenvalues of a pair of matrices . | train | false |
55,472 | @pytest.mark.parametrize('stream', ['stdout', 'stderr'])
def test_exit_unsuccessful_output(qtbot, proc, caplog, py_proc, stream):
with caplog.at_level(logging.ERROR):
with qtbot.waitSignal(proc.finished, timeout=10000):
proc.start(*py_proc('\n import sys\n print("test", file=sys.{})\n sys.exit(1)\n '.format(stream)))
assert (len(caplog.records) == 2)
assert (caplog.records[1].msg == 'Process {}:\ntest'.format(stream))
| [
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'stream'",
",",
"[",
"'stdout'",
",",
"'stderr'",
"]",
")",
"def",
"test_exit_unsuccessful_output",
"(",
"qtbot",
",",
"proc",
",",
"caplog",
",",
"py_proc",
",",
"stream",
")",
":",
"with",
"caplog",
... | when a process fails . | train | false |
55,473 | def generate_replace_result_xml(result_sourcedid, score):
elem = ElementMaker(nsmap={None: 'http://www.imsglobal.org/services/ltiv1p1/xsd/imsoms_v1p0'})
xml = elem.imsx_POXEnvelopeRequest(elem.imsx_POXHeader(elem.imsx_POXRequestHeaderInfo(elem.imsx_version('V1.0'), elem.imsx_messageIdentifier(str(uuid.uuid4())))), elem.imsx_POXBody(elem.replaceResultRequest(elem.resultRecord(elem.sourcedGUID(elem.sourcedId(result_sourcedid)), elem.result(elem.resultScore(elem.language('en'), elem.textString(str(score))))))))
return etree.tostring(xml, xml_declaration=True, encoding='UTF-8')
| [
"def",
"generate_replace_result_xml",
"(",
"result_sourcedid",
",",
"score",
")",
":",
"elem",
"=",
"ElementMaker",
"(",
"nsmap",
"=",
"{",
"None",
":",
"'http://www.imsglobal.org/services/ltiv1p1/xsd/imsoms_v1p0'",
"}",
")",
"xml",
"=",
"elem",
".",
"imsx_POXEnvelope... | create the xml document that contains the new score to be sent to the lti consumer . | train | false |
55,476 | def find_duplicative_certs(config, domains):
def update_certs_for_domain_matches(candidate_lineage, rv):
'Return cert as identical_names_cert if it matches,\n or subset_names_cert if it matches as subset\n '
(identical_names_cert, subset_names_cert) = rv
candidate_names = set(candidate_lineage.names())
if (candidate_names == set(domains)):
identical_names_cert = candidate_lineage
elif candidate_names.issubset(set(domains)):
if (subset_names_cert is None):
subset_names_cert = candidate_lineage
elif (len(candidate_names) > len(subset_names_cert.names())):
subset_names_cert = candidate_lineage
return (identical_names_cert, subset_names_cert)
return _search_lineages(config, update_certs_for_domain_matches, (None, None))
| [
"def",
"find_duplicative_certs",
"(",
"config",
",",
"domains",
")",
":",
"def",
"update_certs_for_domain_matches",
"(",
"candidate_lineage",
",",
"rv",
")",
":",
"(",
"identical_names_cert",
",",
"subset_names_cert",
")",
"=",
"rv",
"candidate_names",
"=",
"set",
... | find existing certs that duplicate the request . | train | false |
55,478 | def make_colorizer(color):
def inner(text):
return colorizer.colorize(color, text)
return inner
| [
"def",
"make_colorizer",
"(",
"color",
")",
":",
"def",
"inner",
"(",
"text",
")",
":",
"return",
"colorizer",
".",
"colorize",
"(",
"color",
",",
"text",
")",
"return",
"inner"
] | creates a function that colorizes text with the given color . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.