id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1
value | is_duplicated bool 2
classes |
|---|---|---|---|---|---|
10,587 | def get_case(d, t):
if (not d.has(t)):
if d.is_one:
return 'base'
return 'primitive'
if d.rem(Poly(t, t)).is_zero:
return 'exp'
if d.rem(Poly((1 + (t ** 2)), t)).is_zero:
return 'tan'
if (d.degree(t) > 1):
return 'other_nonlinear'
return 'other_linear'
| [
"def",
"get_case",
"(",
"d",
",",
"t",
")",
":",
"if",
"(",
"not",
"d",
".",
"has",
"(",
"t",
")",
")",
":",
"if",
"d",
".",
"is_one",
":",
"return",
"'base'",
"return",
"'primitive'",
"if",
"d",
".",
"rem",
"(",
"Poly",
"(",
"t",
",",
"t",
... | returns the type of the derivation d . | train | false |
10,588 | def newDerOctetString(binstring):
if isinstance(binstring, DerObject):
der = DerOctetString(binstring.encode())
else:
der = DerOctetString(binstring)
return der
| [
"def",
"newDerOctetString",
"(",
"binstring",
")",
":",
"if",
"isinstance",
"(",
"binstring",
",",
"DerObject",
")",
":",
"der",
"=",
"DerOctetString",
"(",
"binstring",
".",
"encode",
"(",
")",
")",
"else",
":",
"der",
"=",
"DerOctetString",
"(",
"binstri... | create a deroctetstring object . | train | false |
10,590 | def get_recently_published_exp_summaries(limit):
return _get_exploration_summaries_from_models(exp_models.ExpSummaryModel.get_recently_published(limit))
| [
"def",
"get_recently_published_exp_summaries",
"(",
"limit",
")",
":",
"return",
"_get_exploration_summaries_from_models",
"(",
"exp_models",
".",
"ExpSummaryModel",
".",
"get_recently_published",
"(",
"limit",
")",
")"
] | returns a dict with all featured exploration summary domain objects . | train | false |
10,591 | def _async_wait_until_running(reactor, instance):
action = start_action(action_type=u'flocker:provision:aws:wait_until_running', instance_id=instance.id)
def check_final_state(ignored):
if (instance.state != u'running'):
raise FailedToRun(instance.state_reason)
action.add_success_fields(instance_state=instance.state, instance_state_reason=instance.state_reason)
return instance
def finished_booting():
d = maybeDeferred(_node_is_booting, instance)
d.addCallback((lambda x: (not x)))
return d
with action.context():
d = loop_until(reactor, finished_booting, repeat(5, INSTANCE_TIMEOUT))
d = DeferredContext(d)
d.addCallback(check_final_state)
d.addActionFinish()
return d.result
| [
"def",
"_async_wait_until_running",
"(",
"reactor",
",",
"instance",
")",
":",
"action",
"=",
"start_action",
"(",
"action_type",
"=",
"u'flocker:provision:aws:wait_until_running'",
",",
"instance_id",
"=",
"instance",
".",
"id",
")",
"def",
"check_final_state",
"(",
... | wait until a instance is running . | train | false |
10,592 | def get_run_usr():
if ('SUDO_USER' in os.environ):
usr = os.environ['SUDO_USER']
else:
usr = init_app("who -m | awk '{print $1;}'")
try:
getpwnam(usr)
except:
usr = None
return usr
| [
"def",
"get_run_usr",
"(",
")",
":",
"if",
"(",
"'SUDO_USER'",
"in",
"os",
".",
"environ",
")",
":",
"usr",
"=",
"os",
".",
"environ",
"[",
"'SUDO_USER'",
"]",
"else",
":",
"usr",
"=",
"init_app",
"(",
"\"who -m | awk '{print $1;}'\"",
")",
"try",
":",
... | fetch the user that launched zarp . | train | false |
10,593 | def standard_normal_sample(shape):
return tf.random_normal(shape)
| [
"def",
"standard_normal_sample",
"(",
"shape",
")",
":",
"return",
"tf",
".",
"random_normal",
"(",
"shape",
")"
] | samples from standard gaussian distribution . | train | false |
10,594 | def _CopySortExpressionToProtocolBuffer(sort_expression, pb):
pb.set_sort_expression(sort_expression.expression.encode('utf-8'))
if (sort_expression.direction == SortExpression.ASCENDING):
pb.set_sort_descending(False)
if (sort_expression.default_value is not None):
if isinstance(sort_expression.default_value, basestring):
pb.set_default_value_text(sort_expression.default_value.encode('utf-8'))
elif (isinstance(sort_expression.default_value, datetime.datetime) or isinstance(sort_expression.default_value, datetime.date)):
pb.set_default_value_text(str(search_util.EpochTime(sort_expression.default_value)))
else:
pb.set_default_value_numeric(sort_expression.default_value)
return pb
| [
"def",
"_CopySortExpressionToProtocolBuffer",
"(",
"sort_expression",
",",
"pb",
")",
":",
"pb",
".",
"set_sort_expression",
"(",
"sort_expression",
".",
"expression",
".",
"encode",
"(",
"'utf-8'",
")",
")",
"if",
"(",
"sort_expression",
".",
"direction",
"==",
... | copies a sortexpression to a search_service_pb . | train | false |
10,595 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get new repository . | train | false |
10,597 | @rule(u'$nick(?i)(help|doc) +([A-Za-z]+)(?:\\?+)?$')
@example(u'.help tell')
@commands(u'help', u'commands')
@priority(u'low')
def help(bot, trigger):
if trigger.group(2):
name = trigger.group(2)
name = name.lower()
threshold = 3
if (name in bot.doc):
if ((len(bot.doc[name][0]) + (1 if bot.doc[name][1] else 0)) > threshold):
if (trigger.nick != trigger.sender):
bot.reply(u"The documentation for this command is too long; I'm sending it to you in a private message.")
msgfun = (lambda l: bot.msg(trigger.nick, l))
else:
msgfun = bot.reply
for line in bot.doc[name][0]:
msgfun(line)
if bot.doc[name][1]:
msgfun((u'e.g. ' + bot.doc[name][1]))
else:
if ((u'command-gist' in bot.memory) and (bot.memory[u'command-gist'][0] == len(bot.command_groups))):
url = bot.memory[u'command-gist'][1]
else:
bot.say(u"Hang on, I'm creating a list.")
msgs = []
name_length = max(6, max((len(k) for k in bot.command_groups.keys())))
for (category, cmds) in collections.OrderedDict(sorted(bot.command_groups.items())).items():
category = category.upper().ljust(name_length)
cmds = u' '.join(cmds)
msg = ((category + u' ') + cmds)
indent = (u' ' * (name_length + 2))
msgs.append(u'\n'.join(textwrap.wrap(msg, subsequent_indent=indent)))
url = create_gist(bot, u'\n\n'.join(msgs))
if (not url):
return
bot.memory[u'command-gist'] = (len(bot.command_groups), url)
bot.say(u"I've posted a list of my commands at {} - You can see more info about any of these commands by doing .help <command> (e.g. .help time)".format(url))
| [
"@",
"rule",
"(",
"u'$nick(?i)(help|doc) +([A-Za-z]+)(?:\\\\?+)?$'",
")",
"@",
"example",
"(",
"u'.help tell'",
")",
"@",
"commands",
"(",
"u'help'",
",",
"u'commands'",
")",
"@",
"priority",
"(",
"u'low'",
")",
"def",
"help",
"(",
"bot",
",",
"trigger",
")",
... | adds a charthelp object to the help attribute of the function . | train | false |
10,598 | def flash_size_bytes(size):
if ('MB' in size):
return ((int(size[:size.index('MB')]) * 1024) * 1024)
elif ('KB' in size):
return (int(size[:size.index('KB')]) * 1024)
else:
raise FatalError(('Unknown size %s' % size))
| [
"def",
"flash_size_bytes",
"(",
"size",
")",
":",
"if",
"(",
"'MB'",
"in",
"size",
")",
":",
"return",
"(",
"(",
"int",
"(",
"size",
"[",
":",
"size",
".",
"index",
"(",
"'MB'",
")",
"]",
")",
"*",
"1024",
")",
"*",
"1024",
")",
"elif",
"(",
... | given a flash size of the type passed in args . | train | true |
10,599 | def _patch():
import sys
from OpenGL import GL
if (sys.version_info > (3,)):
buffersubdatafunc = GL.glBufferSubData
if hasattr(buffersubdatafunc, 'wrapperFunction'):
buffersubdatafunc = buffersubdatafunc.wrapperFunction
_m = sys.modules[buffersubdatafunc.__module__]
_m.long = int
try:
from OpenGL.GL.VERSION import GL_2_0
GL_2_0.GL_OBJECT_SHADER_SOURCE_LENGTH = GL_2_0.GL_SHADER_SOURCE_LENGTH
except Exception:
pass
| [
"def",
"_patch",
"(",
")",
":",
"import",
"sys",
"from",
"OpenGL",
"import",
"GL",
"if",
"(",
"sys",
".",
"version_info",
">",
"(",
"3",
",",
")",
")",
":",
"buffersubdatafunc",
"=",
"GL",
".",
"glBufferSubData",
"if",
"hasattr",
"(",
"buffersubdatafunc"... | monkey-patch pyopengl to fix a bug in glbuffersubdata . | train | true |
10,601 | def cell_delete(context, cell_name):
return IMPL.cell_delete(context, cell_name)
| [
"def",
"cell_delete",
"(",
"context",
",",
"cell_name",
")",
":",
"return",
"IMPL",
".",
"cell_delete",
"(",
"context",
",",
"cell_name",
")"
] | delete a child cell . | train | false |
10,602 | def test_scamp_sip_distortion_parameters():
header = get_pkg_data_contents(u'data/validate.fits', encoding=u'binary')
w = wcs.WCS(header)
w.all_pix2world(0, 0, 0)
| [
"def",
"test_scamp_sip_distortion_parameters",
"(",
")",
":",
"header",
"=",
"get_pkg_data_contents",
"(",
"u'data/validate.fits'",
",",
"encoding",
"=",
"u'binary'",
")",
"w",
"=",
"wcs",
".",
"WCS",
"(",
"header",
")",
"w",
".",
"all_pix2world",
"(",
"0",
",... | test parsing of wcs parameters with redundant sip and scamp distortion parameters . | train | false |
10,604 | def test_install_from_wheel_with_legacy(script, data):
result = script.pip('install', 'script.wheel2a==0.1', '--no-index', ('--find-links=' + data.find_links), expect_error=False)
legacy_file1 = (script.bin / 'testscript1.bat')
legacy_file2 = (script.bin / 'testscript2')
assert (legacy_file1 in result.files_created)
assert (legacy_file2 in result.files_created)
| [
"def",
"test_install_from_wheel_with_legacy",
"(",
"script",
",",
"data",
")",
":",
"result",
"=",
"script",
".",
"pip",
"(",
"'install'",
",",
"'script.wheel2a==0.1'",
",",
"'--no-index'",
",",
"(",
"'--find-links='",
"+",
"data",
".",
"find_links",
")",
",",
... | test installing scripts . | train | false |
10,605 | def setup_services(hass, track_new_found_calendars, calendar_service):
def _found_calendar(call):
'Check if we know about a calendar and generate PLATFORM_DISCOVER.'
calendar = get_calendar_info(hass, call.data)
if (hass.data[DATA_INDEX].get(calendar[CONF_CAL_ID], None) is not None):
return
hass.data[DATA_INDEX].update({calendar[CONF_CAL_ID]: calendar})
update_config(hass.config.path(YAML_DEVICES), hass.data[DATA_INDEX][calendar[CONF_CAL_ID]])
discovery.load_platform(hass, 'calendar', DOMAIN, hass.data[DATA_INDEX][calendar[CONF_CAL_ID]])
hass.services.register(DOMAIN, SERVICE_FOUND_CALENDARS, _found_calendar, None, schema=None)
def _scan_for_calendars(service):
'Scan for new calendars.'
service = calendar_service.get()
cal_list = service.calendarList()
calendars = cal_list.list().execute()['items']
for calendar in calendars:
calendar['track'] = track_new_found_calendars
hass.services.call(DOMAIN, SERVICE_FOUND_CALENDARS, calendar)
hass.services.register(DOMAIN, SERVICE_SCAN_CALENDARS, _scan_for_calendars, None, schema=None)
return True
| [
"def",
"setup_services",
"(",
"hass",
",",
"track_new_found_calendars",
",",
"calendar_service",
")",
":",
"def",
"_found_calendar",
"(",
"call",
")",
":",
"calendar",
"=",
"get_calendar_info",
"(",
"hass",
",",
"call",
".",
"data",
")",
"if",
"(",
"hass",
"... | setup service listeners . | train | false |
10,606 | def LoadConfig(config_file_name, exit_fn=sys.exit):
if config_file_name:
config_file = open(config_file_name, 'r')
try:
bulkloader_config = imp.load_module('bulkloader_config', config_file, config_file_name, ('', 'r', imp.PY_SOURCE))
sys.modules['bulkloader_config'] = bulkloader_config
if hasattr(bulkloader_config, 'loaders'):
for cls in bulkloader_config.loaders:
Loader.RegisterLoader(cls())
if hasattr(bulkloader_config, 'exporters'):
for cls in bulkloader_config.exporters:
Exporter.RegisterExporter(cls())
if hasattr(bulkloader_config, 'mappers'):
for cls in bulkloader_config.mappers:
Mapper.RegisterMapper(cls())
except NameError as e:
m = re.search("[^']*'([^']*)'.*", str(e))
if (m.groups() and (m.group(1) == 'Loader')):
print >>sys.stderr, '\nThe config file format has changed and you appear to be using an old-style\nconfig file. Please make the following changes:\n\n1. At the top of the file, add this:\n\nfrom google.appengine.tools.bulkloader import Loader\n\n2. For each of your Loader subclasses add the following at the end of the\n __init__ definitioion:\n\nself.alias_old_names()\n\n3. At the bottom of the file, add this:\n\nloaders = [MyLoader1,...,MyLoaderN]\n\nWhere MyLoader1,...,MyLoaderN are the Loader subclasses you want the bulkloader\nto have access to.\n'
exit_fn(1)
else:
raise
except Exception as e:
if (isinstance(e, NameClashError) or (('bulkloader_config' in vars()) and (hasattr(bulkloader_config, 'bulkloader') and isinstance(e, bulkloader_config.bulkloader.NameClashError)))):
print >>sys.stderr, ('Found both %s and %s while aliasing old names on %s.' % (e.old_name, e.new_name, e.klass))
exit_fn(1)
else:
raise
| [
"def",
"LoadConfig",
"(",
"config_file_name",
",",
"exit_fn",
"=",
"sys",
".",
"exit",
")",
":",
"if",
"config_file_name",
":",
"config_file",
"=",
"open",
"(",
"config_file_name",
",",
"'r'",
")",
"try",
":",
"bulkloader_config",
"=",
"imp",
".",
"load_modu... | initialize a configmanager with the specified options . | train | false |
10,607 | def add_host(zone, name, ttl, ip, keyname, keyfile, nameserver, timeout, port=53, keyalgorithm='hmac-md5'):
res = []
if (zone in name):
name = name.replace(zone, '').rstrip('.')
fqdn = '{0}.{1}'.format(name, zone)
ret = create(zone, name, ttl, 'A', ip, keyname, keyfile, nameserver, timeout, port, keyalgorithm)
res.append(ret[fqdn])
parts = ip.split('.')[::(-1)]
i = len(parts)
popped = []
while (i > 1):
p = parts.pop(0)
i -= 1
popped.append(p)
zone = '{0}.{1}'.format('.'.join(parts), 'in-addr.arpa.')
name = '.'.join(popped)
rev_fqdn = '{0}.{1}'.format(name, zone)
ret = create(zone, name, ttl, 'PTR', '{0}.'.format(fqdn), keyname, keyfile, nameserver, timeout, port, keyalgorithm)
if ('Created' in ret[rev_fqdn]):
res.append(ret[rev_fqdn])
return {fqdn: res}
res.append(ret[rev_fqdn])
return {fqdn: res}
| [
"def",
"add_host",
"(",
"zone",
",",
"name",
",",
"ttl",
",",
"ip",
",",
"keyname",
",",
"keyfile",
",",
"nameserver",
",",
"timeout",
",",
"port",
"=",
"53",
",",
"keyalgorithm",
"=",
"'hmac-md5'",
")",
":",
"res",
"=",
"[",
"]",
"if",
"(",
"zone"... | add host . | train | true |
10,608 | def EvalNormalCdf(x, mu=0, sigma=1):
return stats.norm.cdf(x, loc=mu, scale=sigma)
| [
"def",
"EvalNormalCdf",
"(",
"x",
",",
"mu",
"=",
"0",
",",
"sigma",
"=",
"1",
")",
":",
"return",
"stats",
".",
"norm",
".",
"cdf",
"(",
"x",
",",
"loc",
"=",
"mu",
",",
"scale",
"=",
"sigma",
")"
] | evaluates the cdf of the normal distribution . | train | false |
10,609 | def NDP_Attack_DAD_DoS_via_NS(iface=None, mac_src_filter=None, tgt_filter=None, reply_mac=None):
def ns_reply_callback(req, reply_mac, iface):
'\n Callback that reply to a NS by sending a similar NS\n '
mac = req[Ether].src
dst = req[IPv6].dst
tgt = req[ICMPv6ND_NS].tgt
rep = ((Ether(src=reply_mac) / IPv6(src='::', dst=dst)) / ICMPv6ND_NS(tgt=tgt))
sendp(rep, iface=iface, verbose=0)
print ('Reply NS for target address %s (received from %s)' % (tgt, mac))
_NDP_Attack_DAD_DoS(ns_reply_callback, iface, mac_src_filter, tgt_filter, reply_mac)
| [
"def",
"NDP_Attack_DAD_DoS_via_NS",
"(",
"iface",
"=",
"None",
",",
"mac_src_filter",
"=",
"None",
",",
"tgt_filter",
"=",
"None",
",",
"reply_mac",
"=",
"None",
")",
":",
"def",
"ns_reply_callback",
"(",
"req",
",",
"reply_mac",
",",
"iface",
")",
":",
"m... | perform the dad dos attack using ns described in section 4 . | train | true |
10,610 | def path_isdir(path):
return os.path.isdir(path)
| [
"def",
"path_isdir",
"(",
"path",
")",
":",
"return",
"os",
".",
"path",
".",
"isdir",
"(",
"path",
")"
] | check to see whether the file already exists . | train | false |
10,611 | def obj_dict_dictize(obj_dict, context, sort_key=(lambda x: x)):
result_list = []
for (key, obj) in obj_dict.items():
result_list.append(table_dictize(obj, context))
return sorted(result_list, key=sort_key)
| [
"def",
"obj_dict_dictize",
"(",
"obj_dict",
",",
"context",
",",
"sort_key",
"=",
"(",
"lambda",
"x",
":",
"x",
")",
")",
":",
"result_list",
"=",
"[",
"]",
"for",
"(",
"key",
",",
"obj",
")",
"in",
"obj_dict",
".",
"items",
"(",
")",
":",
"result_... | get a dict whose values are model objects and represent it as a list of dicts . | train | false |
10,612 | def get_database_engine_options(kwargs):
conversions = {'convert_unicode': string_as_bool, 'pool_timeout': int, 'echo': string_as_bool, 'echo_pool': string_as_bool, 'pool_recycle': int, 'pool_size': int, 'max_overflow': int, 'pool_threadlocal': string_as_bool}
prefix = 'database_engine_option_'
prefix_len = len(prefix)
rval = {}
for (key, value) in kwargs.items():
if key.startswith(prefix):
key = key[prefix_len:]
if (key in conversions):
value = conversions[key](value)
rval[key] = value
return rval
| [
"def",
"get_database_engine_options",
"(",
"kwargs",
")",
":",
"conversions",
"=",
"{",
"'convert_unicode'",
":",
"string_as_bool",
",",
"'pool_timeout'",
":",
"int",
",",
"'echo'",
":",
"string_as_bool",
",",
"'echo_pool'",
":",
"string_as_bool",
",",
"'pool_recycl... | allow options for the sqlalchemy database engine to be passed by using the prefix "database_engine_option" . | train | false |
10,613 | def split_into_keywords(string, to_filter_stopwords=False):
if to_filter_stopwords:
return [kw for kw in RE_KEYWORD_SPLIT.split(string.lower()) if ((len(kw) > 0) and (kw not in DIALOG_STOPWORDS))]
else:
return [kw for kw in RE_KEYWORD_SPLIT.split(string.lower()) if (len(kw) > 0)]
| [
"def",
"split_into_keywords",
"(",
"string",
",",
"to_filter_stopwords",
"=",
"False",
")",
":",
"if",
"to_filter_stopwords",
":",
"return",
"[",
"kw",
"for",
"kw",
"in",
"RE_KEYWORD_SPLIT",
".",
"split",
"(",
"string",
".",
"lower",
"(",
")",
")",
"if",
"... | takes a string and returns a list of lowercase strings . | train | false |
10,614 | def missing(name, limit=''):
if (limit == 'upstart'):
return (not _service_is_upstart(name))
elif (limit == 'sysvinit'):
return (not _service_is_sysv(name))
elif (_service_is_upstart(name) or _service_is_sysv(name)):
return False
else:
return True
| [
"def",
"missing",
"(",
"name",
",",
"limit",
"=",
"''",
")",
":",
"if",
"(",
"limit",
"==",
"'upstart'",
")",
":",
"return",
"(",
"not",
"_service_is_upstart",
"(",
"name",
")",
")",
"elif",
"(",
"limit",
"==",
"'sysvinit'",
")",
":",
"return",
"(",
... | the inverse of service . | train | true |
10,615 | def globber(path, pattern=u'*'):
if os.path.exists(path):
return [f for f in os.listdir(path) if fnmatch.fnmatch(f, pattern)]
else:
return []
| [
"def",
"globber",
"(",
"path",
",",
"pattern",
"=",
"u'*'",
")",
":",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"path",
")",
":",
"return",
"[",
"f",
"for",
"f",
"in",
"os",
".",
"listdir",
"(",
"path",
")",
"if",
"fnmatch",
".",
"fnmatch",
... | return matching base file/folder names in folder path . | train | false |
10,616 | def central_server_only(handler):
def central_server_only_wrapper_fn(*args, **kwargs):
if (not settings.CENTRAL_SERVER):
raise Http404(_('This path is only available on the central server.'))
return handler(*args, **kwargs)
return central_server_only_wrapper_fn
| [
"def",
"central_server_only",
"(",
"handler",
")",
":",
"def",
"central_server_only_wrapper_fn",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"not",
"settings",
".",
"CENTRAL_SERVER",
")",
":",
"raise",
"Http404",
"(",
"_",
"(",
"'This path is... | assert-like decorator that marks a function for use only on the central server . | train | false |
10,617 | @docfiller
def mat_reader_factory(file_name, appendmat=True, **kwargs):
byte_stream = _open_file(file_name, appendmat)
(mjv, mnv) = get_matfile_version(byte_stream)
if (mjv == 0):
return MatFile4Reader(byte_stream, **kwargs)
elif (mjv == 1):
return MatFile5Reader(byte_stream, **kwargs)
elif (mjv == 2):
raise NotImplementedError('Please use HDF reader for matlab v7.3 files')
else:
raise TypeError(('Did not recognize version %s' % mjv))
| [
"@",
"docfiller",
"def",
"mat_reader_factory",
"(",
"file_name",
",",
"appendmat",
"=",
"True",
",",
"**",
"kwargs",
")",
":",
"byte_stream",
"=",
"_open_file",
"(",
"file_name",
",",
"appendmat",
")",
"(",
"mjv",
",",
"mnv",
")",
"=",
"get_matfile_version",... | create reader for matlab . | train | false |
10,618 | def determineDefaultFunctionName():
try:
(1 / 0)
except:
return traceback.extract_stack()[(-2)][2]
| [
"def",
"determineDefaultFunctionName",
"(",
")",
":",
"try",
":",
"(",
"1",
"/",
"0",
")",
"except",
":",
"return",
"traceback",
".",
"extract_stack",
"(",
")",
"[",
"(",
"-",
"2",
")",
"]",
"[",
"2",
"]"
] | return the string used by python as the name for code objects which are compiled from interactive input or at the top-level of modules . | train | false |
10,619 | def subreddit_messages_nocache(sr):
from r2.lib.db import queries
inbox = queries.get_subreddit_messages(sr)
messages = _load_messages(inbox)
return compute_message_trees(messages)
| [
"def",
"subreddit_messages_nocache",
"(",
"sr",
")",
":",
"from",
"r2",
".",
"lib",
".",
"db",
"import",
"queries",
"inbox",
"=",
"queries",
".",
"get_subreddit_messages",
"(",
"sr",
")",
"messages",
"=",
"_load_messages",
"(",
"inbox",
")",
"return",
"compu... | just like user_messages . | train | false |
10,620 | def _block_diag(A, n):
if sparse.issparse(A):
raise NotImplemented('sparse reversal not implemented yet')
(ma, na) = A.shape
bdn = (na // int(n))
if ((na % n) > 0):
raise ValueError('Width of matrix must be a multiple of n')
tmp = np.arange((ma * bdn), dtype=np.int).reshape(bdn, ma)
tmp = np.tile(tmp, (1, n))
ii = tmp.ravel()
jj = np.arange(na, dtype=np.int)[None, :]
jj = (jj * np.ones(ma, dtype=np.int)[:, None])
jj = jj.T.ravel()
bd = sparse.coo_matrix((A.T.ravel(), np.c_[(ii, jj)].T)).tocsc()
return bd
| [
"def",
"_block_diag",
"(",
"A",
",",
"n",
")",
":",
"if",
"sparse",
".",
"issparse",
"(",
"A",
")",
":",
"raise",
"NotImplemented",
"(",
"'sparse reversal not implemented yet'",
")",
"(",
"ma",
",",
"na",
")",
"=",
"A",
".",
"shape",
"bdn",
"=",
"(",
... | construct a block diagonal from a packed structure . | train | false |
10,621 | @require_POST
@login_required
@permitted
def vote_for_thread(request, course_id, thread_id, value):
thread = cc.Thread.find(thread_id)
result = _vote_or_unvote(request, course_id, thread, value)
thread_voted.send(sender=None, user=request.user, post=thread)
return result
| [
"@",
"require_POST",
"@",
"login_required",
"@",
"permitted",
"def",
"vote_for_thread",
"(",
"request",
",",
"course_id",
",",
"thread_id",
",",
"value",
")",
":",
"thread",
"=",
"cc",
".",
"Thread",
".",
"find",
"(",
"thread_id",
")",
"result",
"=",
"_vot... | given a course id and thread id vote for this thread ajax only . | train | false |
10,622 | def device_ctypes_pointer(obj):
if (obj is None):
return c_void_p(0)
require_device_memory(obj)
return obj.device_ctypes_pointer
| [
"def",
"device_ctypes_pointer",
"(",
"obj",
")",
":",
"if",
"(",
"obj",
"is",
"None",
")",
":",
"return",
"c_void_p",
"(",
"0",
")",
"require_device_memory",
"(",
"obj",
")",
"return",
"obj",
".",
"device_ctypes_pointer"
] | get the ctypes object for the device pointer . | train | false |
10,625 | def parsehttpdate(string_):
try:
t = time.strptime(string_, '%a, %d %b %Y %H:%M:%S %Z')
except ValueError:
return None
return datetime.datetime(*t[:6])
| [
"def",
"parsehttpdate",
"(",
"string_",
")",
":",
"try",
":",
"t",
"=",
"time",
".",
"strptime",
"(",
"string_",
",",
"'%a, %d %b %Y %H:%M:%S %Z'",
")",
"except",
"ValueError",
":",
"return",
"None",
"return",
"datetime",
".",
"datetime",
"(",
"*",
"t",
"[... | parses an http date into a datetime object . | train | true |
10,626 | def cache_headers_valid(policy_name, headers):
policy_headers = CACHE_POLICY_DIRECTIVES[policy_name]
for (header_name, expected_vals) in policy_headers.items():
found_vals = set(headers.get(header_name, []))
if (header_name == 'cache-control'):
parsed_cache_control = set()
for cache_header in found_vals:
for split_header in cache_header.split(','):
cache_directive = split_header.strip().lower()
parsed_cache_control.add(cache_directive)
if (parsed_cache_control != expected_vals):
return False
elif (found_vals != expected_vals):
return False
return True
| [
"def",
"cache_headers_valid",
"(",
"policy_name",
",",
"headers",
")",
":",
"policy_headers",
"=",
"CACHE_POLICY_DIRECTIVES",
"[",
"policy_name",
"]",
"for",
"(",
"header_name",
",",
"expected_vals",
")",
"in",
"policy_headers",
".",
"items",
"(",
")",
":",
"fou... | check if a responses headers make sense given a cache policy . | train | false |
10,628 | def test_iba_geo_mean_binary():
(y_true, y_pred, _) = make_prediction(binary=True)
iba_gmean = make_index_balanced_accuracy(alpha=0.5, squared=True)(geometric_mean_score)
iba = iba_gmean(y_true, y_pred)
assert_allclose(iba, 0.5948, rtol=R_TOL)
| [
"def",
"test_iba_geo_mean_binary",
"(",
")",
":",
"(",
"y_true",
",",
"y_pred",
",",
"_",
")",
"=",
"make_prediction",
"(",
"binary",
"=",
"True",
")",
"iba_gmean",
"=",
"make_index_balanced_accuracy",
"(",
"alpha",
"=",
"0.5",
",",
"squared",
"=",
"True",
... | test to test the iba using the geometric mean . | train | false |
10,629 | @public
def factor(f, *gens, **args):
f = sympify(f)
if args.pop('deep', False):
partials = {}
muladd = f.atoms(Mul, Add)
for p in muladd:
fac = factor(p, *gens, **args)
if ((fac.is_Mul or fac.is_Pow) and (fac != p)):
partials[p] = fac
return f.xreplace(partials)
try:
return _generic_factor(f, gens, args, method='factor')
except PolynomialError as msg:
if (not f.is_commutative):
from sympy.core.exprtools import factor_nc
return factor_nc(f)
else:
raise PolynomialError(msg)
| [
"@",
"public",
"def",
"factor",
"(",
"f",
",",
"*",
"gens",
",",
"**",
"args",
")",
":",
"f",
"=",
"sympify",
"(",
"f",
")",
"if",
"args",
".",
"pop",
"(",
"'deep'",
",",
"False",
")",
":",
"partials",
"=",
"{",
"}",
"muladd",
"=",
"f",
".",
... | compute the factorization of expression . | train | false |
10,630 | def is_command(text):
return text.startswith('/')
| [
"def",
"is_command",
"(",
"text",
")",
":",
"return",
"text",
".",
"startswith",
"(",
"'/'",
")"
] | checks if text is a command . | train | false |
10,631 | def _find_parametrized_scope(argnames, arg2fixturedefs, indirect):
from _pytest.fixtures import scopes
indirect_as_list = isinstance(indirect, (list, tuple))
all_arguments_are_fixtures = ((indirect is True) or (indirect_as_list and (len(indirect) == argnames)))
if all_arguments_are_fixtures:
fixturedefs = (arg2fixturedefs or {})
used_scopes = [fixturedef[0].scope for (name, fixturedef) in fixturedefs.items()]
if used_scopes:
for scope in reversed(scopes):
if (scope in used_scopes):
return scope
return 'function'
| [
"def",
"_find_parametrized_scope",
"(",
"argnames",
",",
"arg2fixturedefs",
",",
"indirect",
")",
":",
"from",
"_pytest",
".",
"fixtures",
"import",
"scopes",
"indirect_as_list",
"=",
"isinstance",
"(",
"indirect",
",",
"(",
"list",
",",
"tuple",
")",
")",
"al... | find the most appropriate scope for a parametrized call based on its arguments . | train | false |
10,632 | def _maybe_align_partitions(args):
dfs = [df for df in args if isinstance(df, _Frame)]
if (not dfs):
return args
divisions = dfs[0].divisions
if (not all(((df.divisions == divisions) for df in dfs))):
dfs2 = iter(align_partitions(*dfs)[0])
return [(a if (not isinstance(a, _Frame)) else next(dfs2)) for a in args]
return args
| [
"def",
"_maybe_align_partitions",
"(",
"args",
")",
":",
"dfs",
"=",
"[",
"df",
"for",
"df",
"in",
"args",
"if",
"isinstance",
"(",
"df",
",",
"_Frame",
")",
"]",
"if",
"(",
"not",
"dfs",
")",
":",
"return",
"args",
"divisions",
"=",
"dfs",
"[",
"0... | align dataframe blocks if divisions are different . | train | false |
10,633 | def get_start_date():
today = datetime.today()
date = datetime((today.year - 1), today.month, today.day, 12)
weekday = datetime.weekday(date)
while (weekday < 6):
date = (date + timedelta(1))
weekday = datetime.weekday(date)
return date
| [
"def",
"get_start_date",
"(",
")",
":",
"today",
"=",
"datetime",
".",
"today",
"(",
")",
"date",
"=",
"datetime",
"(",
"(",
"today",
".",
"year",
"-",
"1",
")",
",",
"today",
".",
"month",
",",
"today",
".",
"day",
",",
"12",
")",
"weekday",
"="... | returns a datetime object for the first sunday after one year ago today at 12:00 noon . | train | false |
10,635 | def ranking(R):
l = sorted(list(enumerate(R)), cmp=(lambda a, b: cmp(a[1], b[1])))
l = sorted(list(enumerate(l)), cmp=(lambda a, b: cmp(a[1], b[1])))
return array([kv[0] for kv in l])
| [
"def",
"ranking",
"(",
"R",
")",
":",
"l",
"=",
"sorted",
"(",
"list",
"(",
"enumerate",
"(",
"R",
")",
")",
",",
"cmp",
"=",
"(",
"lambda",
"a",
",",
"b",
":",
"cmp",
"(",
"a",
"[",
"1",
"]",
",",
"b",
"[",
"1",
"]",
")",
")",
")",
"l"... | produces a linear ranking of the values in r . | train | false |
10,636 | def dmp_compose(f, g, u, K):
if (not u):
return dup_compose(f, g, K)
if dmp_zero_p(f, u):
return f
h = [f[0]]
for c in f[1:]:
h = dmp_mul(h, g, u, K)
h = dmp_add_term(h, c, 0, u, K)
return h
| [
"def",
"dmp_compose",
"(",
"f",
",",
"g",
",",
"u",
",",
"K",
")",
":",
"if",
"(",
"not",
"u",
")",
":",
"return",
"dup_compose",
"(",
"f",
",",
"g",
",",
"K",
")",
"if",
"dmp_zero_p",
"(",
"f",
",",
"u",
")",
":",
"return",
"f",
"h",
"=",
... | evaluate functional composition f(g) in k[x] . | train | false |
10,637 | @pytest.mark.parametrize('error_message', ["error: pathspec 'unknown_branch' did not match any file(s) known to git.", "hg: abort: unknown revision 'unknown_branch'!"])
def test_clone_handles_branch_typo(mocker, clone_dir, error_message):
mocker.patch('cookiecutter.vcs.subprocess.check_output', autospec=True, side_effect=[subprocess.CalledProcessError((-1), 'cmd', output=error_message)])
repository_url = 'https://github.com/pytest-dev/cookiecutter-pytest-plugin'
with pytest.raises(exceptions.RepositoryCloneFailed) as err:
vcs.clone(repository_url, clone_to_dir=clone_dir, checkout='unknown_branch', no_input=True)
assert (str(err.value) == 'The unknown_branch branch of repository {} could not found, have you made a typo?'.format(repository_url))
| [
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'error_message'",
",",
"[",
"\"error: pathspec 'unknown_branch' did not match any file(s) known to git.\"",
",",
"\"hg: abort: unknown revision 'unknown_branch'!\"",
"]",
")",
"def",
"test_clone_handles_branch_typo",
"(",
"moc... | in clone() . | train | false |
10,638 | def result_to_country_region_city(result):
country = result_to_country(result)
if country:
region = result_to_region(result, country)
city = result_to_city(result, country, region)
else:
region = city = None
return (country, region, city)
| [
"def",
"result_to_country_region_city",
"(",
"result",
")",
":",
"country",
"=",
"result_to_country",
"(",
"result",
")",
"if",
"country",
":",
"region",
"=",
"result_to_region",
"(",
"result",
",",
"country",
")",
"city",
"=",
"result_to_city",
"(",
"result",
... | given one result from mapbox . | train | false |
10,639 | def get_default_model_permissions(model):
permissions = set()
for default in model._meta.default_permissions:
permissions.add(('%s.%s_%s' % (model._meta.app_label, default, model._meta.model_name)))
return permissions
| [
"def",
"get_default_model_permissions",
"(",
"model",
")",
":",
"permissions",
"=",
"set",
"(",
")",
"for",
"default",
"in",
"model",
".",
"_meta",
".",
"default_permissions",
":",
"permissions",
".",
"add",
"(",
"(",
"'%s.%s_%s'",
"%",
"(",
"model",
".",
... | return a set of all default permissions for a given model . | train | false |
10,642 | def extra_oauth2_scope(*scopes):
def extra_oauth2_wrapper(fn):
@functools.wraps(fn)
def wrapper_fn(*a, **kw):
if (not c.oauth_user):
return fn(*a, **kw)
elif c.oauth_scope.has_access(c.site.name, set(scopes)):
return fn(*a, **kw)
else:
raise OAuth2Scope.InsufficientScopeError(scopes)
return wrapper_fn
return extra_oauth2_wrapper
| [
"def",
"extra_oauth2_scope",
"(",
"*",
"scopes",
")",
":",
"def",
"extra_oauth2_wrapper",
"(",
"fn",
")",
":",
"@",
"functools",
".",
"wraps",
"(",
"fn",
")",
"def",
"wrapper_fn",
"(",
"*",
"a",
",",
"**",
"kw",
")",
":",
"if",
"(",
"not",
"c",
"."... | wrap a function so that it only returns data if user has all scopes when not in an oauth2 context . | train | false |
10,644 | def aggregationToMonthsSeconds(interval):
seconds = (interval.get('microseconds', 0) * 1e-06)
seconds += (interval.get('milliseconds', 0) * 0.001)
seconds += interval.get('seconds', 0)
seconds += (interval.get('minutes', 0) * 60)
seconds += ((interval.get('hours', 0) * 60) * 60)
seconds += (((interval.get('days', 0) * 24) * 60) * 60)
seconds += ((((interval.get('weeks', 0) * 7) * 24) * 60) * 60)
months = interval.get('months', 0)
months += (12 * interval.get('years', 0))
return {'months': months, 'seconds': seconds}
| [
"def",
"aggregationToMonthsSeconds",
"(",
"interval",
")",
":",
"seconds",
"=",
"(",
"interval",
".",
"get",
"(",
"'microseconds'",
",",
"0",
")",
"*",
"1e-06",
")",
"seconds",
"+=",
"(",
"interval",
".",
"get",
"(",
"'milliseconds'",
",",
"0",
")",
"*",... | return the number of months and seconds from an aggregation dict that represents a date and time . | train | true |
10,645 | def _AddPrefix(element, prefix):
if (element is None):
return element
if (isinstance(element, list) or isinstance(element, tuple)):
return [(prefix + e) for e in element]
else:
return (prefix + element)
| [
"def",
"_AddPrefix",
"(",
"element",
",",
"prefix",
")",
":",
"if",
"(",
"element",
"is",
"None",
")",
":",
"return",
"element",
"if",
"(",
"isinstance",
"(",
"element",
",",
"list",
")",
"or",
"isinstance",
"(",
"element",
",",
"tuple",
")",
")",
":... | add |prefix| to |element| or each subelement if element is iterable . | train | false |
10,646 | def update_context(name, ctx=None):
if (ctx is None):
ctx = builtins.__xonsh_ctx__
if (not hasattr(update_context, 'bad_imports')):
update_context.bad_imports = []
modctx = xontrib_context(name)
if (modctx is None):
update_context.bad_imports.append(name)
return ctx
return ctx.update(modctx)
| [
"def",
"update_context",
"(",
"name",
",",
"ctx",
"=",
"None",
")",
":",
"if",
"(",
"ctx",
"is",
"None",
")",
":",
"ctx",
"=",
"builtins",
".",
"__xonsh_ctx__",
"if",
"(",
"not",
"hasattr",
"(",
"update_context",
",",
"'bad_imports'",
")",
")",
":",
... | updates a context in place from a xontrib . | train | false |
10,648 | def S_nw_panel(xw, weights, groupidx):
nlags = (len(weights) - 1)
S = (weights[0] * np.dot(xw.T, xw))
for lag in range(1, (nlags + 1)):
(xw0, xwlag) = lagged_groups(xw, lag, groupidx)
s = np.dot(xw0.T, xwlag)
S += (weights[lag] * (s + s.T))
return S
| [
"def",
"S_nw_panel",
"(",
"xw",
",",
"weights",
",",
"groupidx",
")",
":",
"nlags",
"=",
"(",
"len",
"(",
"weights",
")",
"-",
"1",
")",
"S",
"=",
"(",
"weights",
"[",
"0",
"]",
"*",
"np",
".",
"dot",
"(",
"xw",
".",
"T",
",",
"xw",
")",
")... | inner covariance matrix for hac for panel data no denominator nobs used no reference for this . | train | false |
10,649 | def test_url_req_case_mismatch_no_index(script, data):
Upper = os.path.join(data.find_links, 'Upper-1.0.tar.gz')
result = script.pip('install', '--no-index', '-f', data.find_links, Upper, 'requiresupper')
egg_folder = ((script.site_packages / 'Upper-1.0-py%s.egg-info') % pyversion)
assert (egg_folder in result.files_created), str(result)
egg_folder = ((script.site_packages / 'Upper-2.0-py%s.egg-info') % pyversion)
assert (egg_folder not in result.files_created), str(result)
| [
"def",
"test_url_req_case_mismatch_no_index",
"(",
"script",
",",
"data",
")",
":",
"Upper",
"=",
"os",
".",
"path",
".",
"join",
"(",
"data",
".",
"find_links",
",",
"'Upper-1.0.tar.gz'",
")",
"result",
"=",
"script",
".",
"pip",
"(",
"'install'",
",",
"'... | tar ball url requirements . | train | false |
10,650 | def _ReconstructPath(initial_state, current):
path = collections.deque()
while current.previous:
path.appendleft(current)
current = current.previous
for node in path:
initial_state.AddTokenToState(newline=node.newline, dry_run=False)
| [
"def",
"_ReconstructPath",
"(",
"initial_state",
",",
"current",
")",
":",
"path",
"=",
"collections",
".",
"deque",
"(",
")",
"while",
"current",
".",
"previous",
":",
"path",
".",
"appendleft",
"(",
"current",
")",
"current",
"=",
"current",
".",
"previo... | reconstruct the path through the queue with lowest penalty . | train | false |
10,651 | def setup_transaction_hook(config):
def _notify_resource_events_before(request):
'Notify the accumulated resource events before end of transaction.\n '
for event in request.get_resource_events():
request.registry.notify(event)
def _notify_resource_events_after(success, request):
'Notify the accumulated resource events if transaction succeeds.\n '
if success:
for event in request.get_resource_events(after_commit=True):
try:
request.registry.notify(event)
except Exception:
logger.error('Unable to notify', exc_info=True)
def on_new_request(event):
'When a new request comes in, hook on transaction commit.\n '
if hasattr(event.request, 'parent'):
return
current = transaction.get()
current.addBeforeCommitHook(_notify_resource_events_before, args=(event.request,))
current.addAfterCommitHook(_notify_resource_events_after, args=(event.request,))
config.add_subscriber(on_new_request, NewRequest)
| [
"def",
"setup_transaction_hook",
"(",
"config",
")",
":",
"def",
"_notify_resource_events_before",
"(",
"request",
")",
":",
"for",
"event",
"in",
"request",
".",
"get_resource_events",
"(",
")",
":",
"request",
".",
"registry",
".",
"notify",
"(",
"event",
")... | resource events are plugged with the transactions of pyramid_tm . | train | false |
10,652 | @mock_streams('stdout')
def test_puts_with_encoding_type_none_output():
s = u'string!'
output.user = True
sys.stdout.encoding = None
puts(s, show_prefix=False)
eq_(sys.stdout.getvalue(), (s + '\n'))
| [
"@",
"mock_streams",
"(",
"'stdout'",
")",
"def",
"test_puts_with_encoding_type_none_output",
"(",
")",
":",
"s",
"=",
"u'string!'",
"output",
".",
"user",
"=",
"True",
"sys",
".",
"stdout",
".",
"encoding",
"=",
"None",
"puts",
"(",
"s",
",",
"show_prefix",... | puts() should print unicode output without a stream encoding . | train | false |
10,654 | def getSegmentPathDefault():
return [Vector3(), Vector3(0.0, 1.0)]
| [
"def",
"getSegmentPathDefault",
"(",
")",
":",
"return",
"[",
"Vector3",
"(",
")",
",",
"Vector3",
"(",
"0.0",
",",
"1.0",
")",
"]"
] | get segment path default . | train | false |
10,655 | def get_br_trunk_port_name(prefix, port_id):
return ('%st-%s' % (prefix, port_id))[:constants.DEVICE_NAME_MAX_LEN]
| [
"def",
"get_br_trunk_port_name",
"(",
"prefix",
",",
"port_id",
")",
":",
"return",
"(",
"'%st-%s'",
"%",
"(",
"prefix",
",",
"port_id",
")",
")",
"[",
":",
"constants",
".",
"DEVICE_NAME_MAX_LEN",
"]"
] | return the ovs port name for the given port id . | train | false |
10,656 | def load_diabetes(return_X_y=False):
base_dir = join(dirname(__file__), 'data')
data = np.loadtxt(join(base_dir, 'diabetes_data.csv.gz'))
target = np.loadtxt(join(base_dir, 'diabetes_target.csv.gz'))
if return_X_y:
return (data, target)
return Bunch(data=data, target=target, feature_names=['age', 'sex', 'bmi', 'bp', 's1', 's2', 's3', 's4', 's5', 's6'])
| [
"def",
"load_diabetes",
"(",
"return_X_y",
"=",
"False",
")",
":",
"base_dir",
"=",
"join",
"(",
"dirname",
"(",
"__file__",
")",
",",
"'data'",
")",
"data",
"=",
"np",
".",
"loadtxt",
"(",
"join",
"(",
"base_dir",
",",
"'diabetes_data.csv.gz'",
")",
")"... | load and return the diabetes dataset . | train | false |
10,657 | def init_bh_match():
global bh_match
bh_match = BhCore().match
debug('Match object loaded.')
| [
"def",
"init_bh_match",
"(",
")",
":",
"global",
"bh_match",
"bh_match",
"=",
"BhCore",
"(",
")",
".",
"match",
"debug",
"(",
"'Match object loaded.'",
")"
] | initialize the match object . | train | false |
10,658 | def _resource_path_list(resource, *elements):
path = [(loc.__name__ or '') for loc in lineage(resource)]
path.reverse()
path.extend(elements)
return path
| [
"def",
"_resource_path_list",
"(",
"resource",
",",
"*",
"elements",
")",
":",
"path",
"=",
"[",
"(",
"loc",
".",
"__name__",
"or",
"''",
")",
"for",
"loc",
"in",
"lineage",
"(",
"resource",
")",
"]",
"path",
".",
"reverse",
"(",
")",
"path",
".",
... | implementation detail shared by resource_path and resource_path_tuple . | train | false |
10,659 | def inf_loop_multiple_back_edge(rec):
while True:
rec.mark('yield')
(yield)
p = rec('p')
if p:
rec.mark('bra')
pass
| [
"def",
"inf_loop_multiple_back_edge",
"(",
"rec",
")",
":",
"while",
"True",
":",
"rec",
".",
"mark",
"(",
"'yield'",
")",
"(",
"yield",
")",
"p",
"=",
"rec",
"(",
"'p'",
")",
"if",
"p",
":",
"rec",
".",
"mark",
"(",
"'bra'",
")",
"pass"
] | test to reveal bug of invalid liveness when infinite loop has multiple backedge . | train | false |
10,660 | def substitute_variables(input_data, substitutions):
def f(value):
if (not isinstance(value, basestring)):
return value
new_value = Template(value).safe_substitute(substitutions)
if (new_value != value):
LOG.debug(('Substituted %s -> %s' % (repr(value), repr(new_value))))
return new_value
return recursive_walk(f, input_data)
| [
"def",
"substitute_variables",
"(",
"input_data",
",",
"substitutions",
")",
":",
"def",
"f",
"(",
"value",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"value",
",",
"basestring",
")",
")",
":",
"return",
"value",
"new_value",
"=",
"Template",
"(",
"v... | replaces variables with values from substitutions . | train | false |
10,661 | def get_comment_app():
comments_app = get_comment_app_name()
if (comments_app not in settings.INSTALLED_APPS):
raise ImproperlyConfigured(('The COMMENTS_APP (%r) must be in INSTALLED_APPS' % settings.COMMENTS_APP))
try:
package = import_module(comments_app)
except ImportError:
raise ImproperlyConfigured('The COMMENTS_APP setting refers to a non-existing package.')
return package
| [
"def",
"get_comment_app",
"(",
")",
":",
"comments_app",
"=",
"get_comment_app_name",
"(",
")",
"if",
"(",
"comments_app",
"not",
"in",
"settings",
".",
"INSTALLED_APPS",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"(",
"'The COMMENTS_APP (%r) must be in INSTALLED... | get the comment app as defined in the settings . | train | false |
10,662 | def tensor_product_simp(e, **hints):
if isinstance(e, Add):
return Add(*[tensor_product_simp(arg) for arg in e.args])
elif isinstance(e, Pow):
return (tensor_product_simp(e.base) ** e.exp)
elif isinstance(e, Mul):
return tensor_product_simp_Mul(e)
elif isinstance(e, Commutator):
return Commutator(*[tensor_product_simp(arg) for arg in e.args])
elif isinstance(e, AntiCommutator):
return AntiCommutator(*[tensor_product_simp(arg) for arg in e.args])
else:
return e
| [
"def",
"tensor_product_simp",
"(",
"e",
",",
"**",
"hints",
")",
":",
"if",
"isinstance",
"(",
"e",
",",
"Add",
")",
":",
"return",
"Add",
"(",
"*",
"[",
"tensor_product_simp",
"(",
"arg",
")",
"for",
"arg",
"in",
"e",
".",
"args",
"]",
")",
"elif"... | try to simplify and combine tensorproducts . | train | false |
10,663 | def get_credentials_file(*args):
if check_file_permissions():
ensure_local_plotly_files()
return utils.load_json_dict(CREDENTIALS_FILE, *args)
else:
return FILE_CONTENT[CREDENTIALS_FILE]
| [
"def",
"get_credentials_file",
"(",
"*",
"args",
")",
":",
"if",
"check_file_permissions",
"(",
")",
":",
"ensure_local_plotly_files",
"(",
")",
"return",
"utils",
".",
"load_json_dict",
"(",
"CREDENTIALS_FILE",
",",
"*",
"args",
")",
"else",
":",
"return",
"F... | return specified args from ~/ . | train | false |
10,664 | def unicode_join(seq):
return concat(imap(text_type, seq))
| [
"def",
"unicode_join",
"(",
"seq",
")",
":",
"return",
"concat",
"(",
"imap",
"(",
"text_type",
",",
"seq",
")",
")"
] | simple args to unicode conversion and concatenation . | train | false |
10,666 | def TSF(ds, count, timeperiod=(- (2 ** 31))):
return call_talib_with_ds(ds, count, talib.TSF, timeperiod)
| [
"def",
"TSF",
"(",
"ds",
",",
"count",
",",
"timeperiod",
"=",
"(",
"-",
"(",
"2",
"**",
"31",
")",
")",
")",
":",
"return",
"call_talib_with_ds",
"(",
"ds",
",",
"count",
",",
"talib",
".",
"TSF",
",",
"timeperiod",
")"
] | time series forecast . | train | false |
10,667 | def assert_opened_help_link_is_correct(test, url):
test.browser.switch_to_window(test.browser.window_handles[(-1)])
test.assertEqual(url, test.browser.current_url)
test.assertNotIn('Maze Found', test.browser.title)
| [
"def",
"assert_opened_help_link_is_correct",
"(",
"test",
",",
"url",
")",
":",
"test",
".",
"browser",
".",
"switch_to_window",
"(",
"test",
".",
"browser",
".",
"window_handles",
"[",
"(",
"-",
"1",
")",
"]",
")",
"test",
".",
"assertEqual",
"(",
"url",
... | asserts that url of browser when help link is clicked is correct . | train | false |
10,668 | def get_subnetid(vm_):
subnetid = config.get_cloud_config_value('subnetid', vm_, __opts__, search_global=False)
if subnetid:
return subnetid
subnetname = config.get_cloud_config_value('subnetname', vm_, __opts__, search_global=False)
if subnetname:
return _get_subnetname_id(subnetname)
return None
| [
"def",
"get_subnetid",
"(",
"vm_",
")",
":",
"subnetid",
"=",
"config",
".",
"get_cloud_config_value",
"(",
"'subnetid'",
",",
"vm_",
",",
"__opts__",
",",
"search_global",
"=",
"False",
")",
"if",
"subnetid",
":",
"return",
"subnetid",
"subnetname",
"=",
"c... | returns the subnetid to use . | train | true |
10,669 | def test_ast_bad_get():
cant_compile(u'(get)')
cant_compile(u'(get 1)')
| [
"def",
"test_ast_bad_get",
"(",
")",
":",
"cant_compile",
"(",
"u'(get)'",
")",
"cant_compile",
"(",
"u'(get 1)'",
")"
] | make sure ast cant compile invalid get . | train | false |
10,671 | def perform_all_actions(all_actions):
for action_group in all_actions:
to_destroy = getattr(action_group, 'destroy', [])
for resource in to_destroy:
destroy_resource(resource)
| [
"def",
"perform_all_actions",
"(",
"all_actions",
")",
":",
"for",
"action_group",
"in",
"all_actions",
":",
"to_destroy",
"=",
"getattr",
"(",
"action_group",
",",
"'destroy'",
",",
"[",
"]",
")",
"for",
"resource",
"in",
"to_destroy",
":",
"destroy_resource",
... | loop through all the actions and destroy the resources that need to be destroyed . | train | false |
10,672 | def test_parse_annotation():
annot = '+180\x14Lights off\x14Close door\x14\x00\x00\x00\x00\x00+180\x14Lights off\x14\x00\x00\x00\x00\x00\x00\x00\x00+180\x14Close door\x14\x00\x00\x00\x00\x00\x00\x00\x00+3.14\x1504.20\x14nothing\x14\x00\x00\x00\x00+1800.2\x1525.5\x14Apnea\x14\x00\x00\x00\x00\x00\x00\x00+123\x14\x14\x00\x00\x00\x00\x00\x00\x00'
annot = [a for a in iterbytes(annot)]
annot[1::2] = [(a * 256) for a in annot[1::2]]
tal_channel = map(sum, zip(annot[0::2], annot[1::2]))
assert_equal(_parse_tal_channel([tal_channel]), [[180.0, 0, 'Lights off'], [180.0, 0, 'Close door'], [180.0, 0, 'Lights off'], [180.0, 0, 'Close door'], [3.14, 4.2, 'nothing'], [1800.2, 25.5, 'Apnea']])
| [
"def",
"test_parse_annotation",
"(",
")",
":",
"annot",
"=",
"'+180\\x14Lights off\\x14Close door\\x14\\x00\\x00\\x00\\x00\\x00+180\\x14Lights off\\x14\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x00+180\\x14Close door\\x14\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x00+3.14\\x1504.20\\x14nothing\\x14\\x00\\x00\\x00... | test parsing the tal channel . | train | false |
10,673 | def splitdrive(p):
return ('', p)
| [
"def",
"splitdrive",
"(",
"p",
")",
":",
"return",
"(",
"''",
",",
"p",
")"
] | split a pathname into drive/unc sharepoint and relative path specifiers . | train | false |
10,674 | def CreateProcess(appName, cmdline, procSecurity, threadSecurity, inheritHandles, newEnvironment, env, workingDir, startupInfo):
(hProcess, hThread, dwPid, dwTid) = win32process.CreateProcess(appName, cmdline, procSecurity, threadSecurity, inheritHandles, newEnvironment, env, workingDir, startupInfo)
dwPid = 42
return (hProcess, hThread, dwPid, dwTid)
| [
"def",
"CreateProcess",
"(",
"appName",
",",
"cmdline",
",",
"procSecurity",
",",
"threadSecurity",
",",
"inheritHandles",
",",
"newEnvironment",
",",
"env",
",",
"workingDir",
",",
"startupInfo",
")",
":",
"(",
"hProcess",
",",
"hThread",
",",
"dwPid",
",",
... | this function mocks the generated pid aspect of the win32 . | train | false |
10,675 | def load_backends_and_plugins(plugins, working_set, backends, build_configuration=None):
build_configuration = (build_configuration or BuildConfiguration())
load_build_configuration_from_source(build_configuration, backends)
load_plugins(build_configuration, (plugins or []), working_set)
return build_configuration
| [
"def",
"load_backends_and_plugins",
"(",
"plugins",
",",
"working_set",
",",
"backends",
",",
"build_configuration",
"=",
"None",
")",
":",
"build_configuration",
"=",
"(",
"build_configuration",
"or",
"BuildConfiguration",
"(",
")",
")",
"load_build_configuration_from_... | load named plugins and source backends . | train | true |
10,676 | def get_public_key(key, passphrase=None, asObj=False):
if isinstance(key, M2Crypto.X509.X509):
rsa = key.get_pubkey().get_rsa()
text = ''
else:
text = _text_or_file(key)
text = get_pem_entry(text)
if text.startswith('-----BEGIN PUBLIC KEY-----'):
if (not asObj):
return text
bio = M2Crypto.BIO.MemoryBuffer()
bio.write(text)
rsa = M2Crypto.RSA.load_pub_key_bio(bio)
bio = M2Crypto.BIO.MemoryBuffer()
if text.startswith('-----BEGIN CERTIFICATE-----'):
cert = M2Crypto.X509.load_cert_string(text)
rsa = cert.get_pubkey().get_rsa()
if text.startswith('-----BEGIN CERTIFICATE REQUEST-----'):
csr = M2Crypto.X509.load_request_string(text)
rsa = csr.get_pubkey().get_rsa()
if (text.startswith('-----BEGIN PRIVATE KEY-----') or text.startswith('-----BEGIN RSA PRIVATE KEY-----')):
rsa = M2Crypto.RSA.load_key_string(text, callback=_passphrase_callback(passphrase))
if asObj:
evppubkey = M2Crypto.EVP.PKey()
evppubkey.assign_rsa(rsa)
return evppubkey
rsa.save_pub_key_bio(bio)
return bio.read_all()
| [
"def",
"get_public_key",
"(",
"key",
",",
"passphrase",
"=",
"None",
",",
"asObj",
"=",
"False",
")",
":",
"if",
"isinstance",
"(",
"key",
",",
"M2Crypto",
".",
"X509",
".",
"X509",
")",
":",
"rsa",
"=",
"key",
".",
"get_pubkey",
"(",
")",
".",
"ge... | create the public key object from a retrieved certificate . | train | true |
10,678 | def escape(text):
text = text.replace('\\', '\\\\')
text = text.replace('"""', '""\\"')
text = text.replace(' \n', ' \\n\\\n')
return text
| [
"def",
"escape",
"(",
"text",
")",
":",
"text",
"=",
"text",
".",
"replace",
"(",
"'\\\\'",
",",
"'\\\\\\\\'",
")",
"text",
"=",
"text",
".",
"replace",
"(",
"'\"\"\"'",
",",
"'\"\"\\\\\"'",
")",
"text",
"=",
"text",
".",
"replace",
"(",
"' \\n'",
",... | escape a string in an oauth-compatible fashion . | train | false |
10,679 | def smart_truncate(string, max_length=0, word_boundaries=False, separator=' '):
string = string.strip(separator)
if (not max_length):
return string
if (len(string) < max_length):
return string
if (not word_boundaries):
return string[:max_length].strip(separator)
if (separator not in string):
return string[:max_length]
truncated = ''
for word in string.split(separator):
if word:
next_len = ((len(truncated) + len(word)) + len(separator))
if (next_len <= max_length):
truncated += '{0}{1}'.format(word, separator)
if (not truncated):
truncated = string[:max_length]
return truncated.strip(separator)
| [
"def",
"smart_truncate",
"(",
"string",
",",
"max_length",
"=",
"0",
",",
"word_boundaries",
"=",
"False",
",",
"separator",
"=",
"' '",
")",
":",
"string",
"=",
"string",
".",
"strip",
"(",
"separator",
")",
"if",
"(",
"not",
"max_length",
")",
":",
"... | truncate a string . | train | true |
10,680 | def template_localtime(value, use_tz=None):
should_convert = (isinstance(value, datetime) and (settings.USE_TZ if (use_tz is None) else use_tz) and (not is_naive(value)) and getattr(value, 'convert_to_local_time', True))
return (localtime(value) if should_convert else value)
| [
"def",
"template_localtime",
"(",
"value",
",",
"use_tz",
"=",
"None",
")",
":",
"should_convert",
"=",
"(",
"isinstance",
"(",
"value",
",",
"datetime",
")",
"and",
"(",
"settings",
".",
"USE_TZ",
"if",
"(",
"use_tz",
"is",
"None",
")",
"else",
"use_tz"... | checks if value is a datetime and converts it to local time if necessary . | train | false |
10,681 | def _coerce_exceptions(function):
def wrapper(*args, **kwargs):
try:
return function(*args, **kwargs)
except IOException as e:
e.msg = force_unicode(e.msg, errors='replace')
e.stack = force_unicode(e.stack, errors='replace')
LOG.exception(('Exception in Hadoop FS call ' + function.__name__))
if (e.clazz == HADOOP_ACCESSCONTROLEXCEPTION):
raise PermissionDeniedException(e.msg, e)
else:
raise
return wrapper
| [
"def",
"_coerce_exceptions",
"(",
"function",
")",
":",
"def",
"wrapper",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"return",
"function",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"except",
"IOException",
"as",
"e",
":",
"e",
"."... | decorator that causes exceptions thrown by the decorated function to be coerced into generic exceptions from the hadoop . | train | false |
10,682 | @task()
def enable_self_generated_certs(course_key):
course_key = CourseKey.from_string(course_key)
course = CourseOverview.get_from_id(course_key)
is_enabled_for_course = CertificateGenerationCourseSetting.is_enabled_for_course(course_key)
if (course.self_paced and (not is_enabled_for_course)):
CertificateGenerationCourseSetting.set_enabled_for_course(course_key, True)
| [
"@",
"task",
"(",
")",
"def",
"enable_self_generated_certs",
"(",
"course_key",
")",
":",
"course_key",
"=",
"CourseKey",
".",
"from_string",
"(",
"course_key",
")",
"course",
"=",
"CourseOverview",
".",
"get_from_id",
"(",
"course_key",
")",
"is_enabled_for_cours... | enable the self-generated certificates by default for self-paced courses . | train | false |
10,685 | def _unusedTestDirectory(base):
counter = 0
while True:
if counter:
testdir = base.sibling(('%s-%d' % (base.basename(), counter)))
else:
testdir = base
testDirLock = FilesystemLock((testdir.path + '.lock'))
if testDirLock.lock():
if testdir.exists():
_removeSafely(testdir)
testdir.makedirs()
testdir.child('_trial_marker').setContent('')
return (testdir, testDirLock)
elif (base.basename() == '_trial_temp'):
counter += 1
else:
raise _WorkingDirectoryBusy()
| [
"def",
"_unusedTestDirectory",
"(",
"base",
")",
":",
"counter",
"=",
"0",
"while",
"True",
":",
"if",
"counter",
":",
"testdir",
"=",
"base",
".",
"sibling",
"(",
"(",
"'%s-%d'",
"%",
"(",
"base",
".",
"basename",
"(",
")",
",",
"counter",
")",
")",... | find an unused directory named similarly to c{base} . | train | false |
10,688 | def saveAnswers(questions, series_id, complete_id, rvars):
text = ''
table = current.s3db.survey_complete
for question in questions:
code = question['code']
if ((code in rvars) and (rvars[code] != '')):
line = ('"%s","%s"\n' % (code, rvars[code]))
text += line
if (complete_id is None):
record_id = table.insert(series_id=series_id, answer_list=text)
S3SurveyCompleteModel.completeOnAccept(record_id)
return record_id
else:
current.db((table.id == complete_id)).update(answer_list=text)
S3SurveyCompleteModel.completeOnAccept(complete_id)
return complete_id
| [
"def",
"saveAnswers",
"(",
"questions",
",",
"series_id",
",",
"complete_id",
",",
"rvars",
")",
":",
"text",
"=",
"''",
"table",
"=",
"current",
".",
"s3db",
".",
"survey_complete",
"for",
"question",
"in",
"questions",
":",
"code",
"=",
"question",
"[",
... | insert/update a record in survey_complete . | train | false |
10,689 | def setup_cluster(num_cpus, outdir, verbose, error_profile):
server_socket = setup_server()
(workers, client_socks_and_adrs) = setup_workers(num_cpus, outdir, server_socket, verbose=verbose, error_profile=error_profile)
client_sockets = [sock for (sock, addr) in client_socks_and_adrs]
return (client_sockets, workers, server_socket)
| [
"def",
"setup_cluster",
"(",
"num_cpus",
",",
"outdir",
",",
"verbose",
",",
"error_profile",
")",
":",
"server_socket",
"=",
"setup_server",
"(",
")",
"(",
"workers",
",",
"client_socks_and_adrs",
")",
"=",
"setup_workers",
"(",
"num_cpus",
",",
"outdir",
","... | setup server and clients . | train | false |
10,690 | def _subsets(n):
if (n == 1):
a = [[1]]
elif (n == 2):
a = [[1, 0], [0, 1], [1, 1]]
elif (n == 3):
a = [[1, 0, 0], [0, 1, 0], [1, 1, 0], [0, 0, 1], [1, 0, 1], [0, 1, 1], [1, 1, 1]]
else:
b = _subsets((n - 1))
a0 = [(x + [0]) for x in b]
a1 = [(x + [1]) for x in b]
a = ((a0 + [(([0] * (n - 1)) + [1])]) + a1)
return a
| [
"def",
"_subsets",
"(",
"n",
")",
":",
"if",
"(",
"n",
"==",
"1",
")",
":",
"a",
"=",
"[",
"[",
"1",
"]",
"]",
"elif",
"(",
"n",
"==",
"2",
")",
":",
"a",
"=",
"[",
"[",
"1",
",",
"0",
"]",
",",
"[",
"0",
",",
"1",
"]",
",",
"[",
... | returns all possible subsets of the set except the empty set . | train | false |
10,691 | def _adapt(iface, obj):
for (wrapper_test, wrapper) in IMOUNTABLE_FILESYSTEM_ADAPTERS.items():
if wrapper_test(obj):
return wrapper(obj)
| [
"def",
"_adapt",
"(",
"iface",
",",
"obj",
")",
":",
"for",
"(",
"wrapper_test",
",",
"wrapper",
")",
"in",
"IMOUNTABLE_FILESYSTEM_ADAPTERS",
".",
"items",
"(",
")",
":",
"if",
"wrapper_test",
"(",
"obj",
")",
":",
"return",
"wrapper",
"(",
"obj",
")"
] | adapt the filesystem argument of mount to imountablefilesystem . | train | false |
10,692 | def releaseNetToMs():
a = TpPd(pd=3)
b = MessageType(mesType=45)
c = CauseHdr(ieiC=8, eightBitC=0)
d = CauseHdr(ieiC=8, eightBitC=0)
e = FacilityHdr(ieiF=28, eightBitF=0)
f = UserUserHdr(ieiUU=126, eightBitUU=0)
packet = (((((a / b) / c) / d) / e) / f)
return packet
| [
"def",
"releaseNetToMs",
"(",
")",
":",
"a",
"=",
"TpPd",
"(",
"pd",
"=",
"3",
")",
"b",
"=",
"MessageType",
"(",
"mesType",
"=",
"45",
")",
"c",
"=",
"CauseHdr",
"(",
"ieiC",
"=",
"8",
",",
"eightBitC",
"=",
"0",
")",
"d",
"=",
"CauseHdr",
"("... | release section 9 . | train | true |
10,693 | def currentTag():
cmd = 'git describe --always --tag'.split()
tag = subprocess.check_output(cmd, cwd=VERSIONSDIR).split('-')[0]
return tag
| [
"def",
"currentTag",
"(",
")",
":",
"cmd",
"=",
"'git describe --always --tag'",
".",
"split",
"(",
")",
"tag",
"=",
"subprocess",
".",
"check_output",
"(",
"cmd",
",",
"cwd",
"=",
"VERSIONSDIR",
")",
".",
"split",
"(",
"'-'",
")",
"[",
"0",
"]",
"retu... | returns the current tag name from the version repository . | train | false |
10,694 | def get_ssh_key_fingerprint(ssh_key):
parts = ssh_key.strip().split()
if (len(parts) == 0):
return None
key_type = parts[0]
key = base64.b64decode(parts[1].encode('ascii'))
fp_plain = hashlib.md5(key).hexdigest()
key_fp = ':'.join(((a + b) for (a, b) in zip(fp_plain[::2], fp_plain[1::2]))).upper()
if (len(parts) < 3):
return ('%s (%s)' % (key_fp, key_type))
else:
user_host = parts[2]
return ('%s %s (%s)' % (key_fp, user_host, key_type))
| [
"def",
"get_ssh_key_fingerprint",
"(",
"ssh_key",
")",
":",
"parts",
"=",
"ssh_key",
".",
"strip",
"(",
")",
".",
"split",
"(",
")",
"if",
"(",
"len",
"(",
"parts",
")",
"==",
"0",
")",
":",
"return",
"None",
"key_type",
"=",
"parts",
"[",
"0",
"]"... | return the public key fingerprint of a given public ssh key in format "fb:0c:ac:0a:07:94:5b:ce:75:6e:63:32:13:ad:ad:d7 [user@host] " . | train | false |
10,696 | @deprecated('use cls.register(cli)')
def register_commands(commands):
for command_klass in commands:
_COMMANDS.register(command_klass)
| [
"@",
"deprecated",
"(",
"'use cls.register(cli)'",
")",
"def",
"register_commands",
"(",
"commands",
")",
":",
"for",
"command_klass",
"in",
"commands",
":",
"_COMMANDS",
".",
"register",
"(",
"command_klass",
")"
] | called when the emr command table is being built . | train | false |
10,699 | def caffe_load_image(filename, color=True, as_uint=False):
with WithTimer('imread', quiet=True):
if as_uint:
img = skimage.io.imread(filename)
else:
img = skimage.img_as_float(skimage.io.imread(filename)).astype(np.float32)
if (img.ndim == 2):
img = img[:, :, np.newaxis]
if color:
img = np.tile(img, (1, 1, 3))
elif (img.shape[2] == 4):
img = img[:, :, :3]
return img
| [
"def",
"caffe_load_image",
"(",
"filename",
",",
"color",
"=",
"True",
",",
"as_uint",
"=",
"False",
")",
":",
"with",
"WithTimer",
"(",
"'imread'",
",",
"quiet",
"=",
"True",
")",
":",
"if",
"as_uint",
":",
"img",
"=",
"skimage",
".",
"io",
".",
"im... | copied from caffe to simplify potential import problems . | train | false |
10,700 | def _finalize_fairy(connection, connection_record, pool, ref, echo, fairy=None):
_refs.discard(connection_record)
if ((ref is not None) and (connection_record.fairy_ref is not ref)):
return
if (connection is not None):
if (connection_record and echo):
pool.logger.debug('Connection %r being returned to pool', connection)
try:
fairy = (fairy or _ConnectionFairy(connection, connection_record))
assert (fairy.connection is connection)
fairy._reset(pool, echo)
if (not connection_record):
pool._close_connection(connection)
except Exception as e:
if connection_record:
connection_record.invalidate(e=e)
if isinstance(e, (SystemExit, KeyboardInterrupt)):
raise
if connection_record:
connection_record.checkin()
| [
"def",
"_finalize_fairy",
"(",
"connection",
",",
"connection_record",
",",
"pool",
",",
"ref",
",",
"echo",
",",
"fairy",
"=",
"None",
")",
":",
"_refs",
".",
"discard",
"(",
"connection_record",
")",
"if",
"(",
"(",
"ref",
"is",
"not",
"None",
")",
"... | cleanup for a :class: . | train | false |
10,702 | def _clean_probes(probes):
probes = _ordered_dict_to_dict(probes)
probes_copy = deepcopy(probes)
for (probe_name, probe_tests) in six.iteritems(probes_copy):
if (not probe_tests):
probes.pop(probe_name)
continue
for (test_name, test_params) in six.iteritems(probe_tests):
if (not test_params):
probes[probe_name].pop(test_name)
if (not probes.get(probe_name)):
probes.pop(probe_name)
return True
| [
"def",
"_clean_probes",
"(",
"probes",
")",
":",
"probes",
"=",
"_ordered_dict_to_dict",
"(",
"probes",
")",
"probes_copy",
"=",
"deepcopy",
"(",
"probes",
")",
"for",
"(",
"probe_name",
",",
"probe_tests",
")",
"in",
"six",
".",
"iteritems",
"(",
"probes_co... | will remove empty and useless values from the probes dictionary . | train | true |
10,703 | def clear_known_categories(x, cols=None, index=True):
if isinstance(x, (pd.Series, pd.DataFrame)):
x = x.copy()
if isinstance(x, pd.DataFrame):
mask = (x.dtypes == 'category')
if (cols is None):
cols = mask[mask].index
elif (not mask.loc[cols].all()):
raise ValueError('Not all columns are categoricals')
for c in cols:
x[c].cat.set_categories([UNKNOWN_CATEGORIES], inplace=True)
elif isinstance(x, pd.Series):
if is_categorical_dtype(x.dtype):
x.cat.set_categories([UNKNOWN_CATEGORIES], inplace=True)
if (index and isinstance(x.index, pd.CategoricalIndex)):
x.index = x.index.set_categories([UNKNOWN_CATEGORIES])
elif isinstance(x, pd.CategoricalIndex):
x = x.set_categories([UNKNOWN_CATEGORIES])
return x
| [
"def",
"clear_known_categories",
"(",
"x",
",",
"cols",
"=",
"None",
",",
"index",
"=",
"True",
")",
":",
"if",
"isinstance",
"(",
"x",
",",
"(",
"pd",
".",
"Series",
",",
"pd",
".",
"DataFrame",
")",
")",
":",
"x",
"=",
"x",
".",
"copy",
"(",
... | set categories to be unknown . | train | false |
10,704 | def getSPClass(spatialImp):
if (spatialImp == 'py'):
return PYSpatialPooler
elif (spatialImp == 'cpp'):
return CPPSpatialPooler
else:
raise RuntimeError(("Invalid spatialImp '%s'. Legal values are: 'py', 'cpp'" % spatialImp))
| [
"def",
"getSPClass",
"(",
"spatialImp",
")",
":",
"if",
"(",
"spatialImp",
"==",
"'py'",
")",
":",
"return",
"PYSpatialPooler",
"elif",
"(",
"spatialImp",
"==",
"'cpp'",
")",
":",
"return",
"CPPSpatialPooler",
"else",
":",
"raise",
"RuntimeError",
"(",
"(",
... | return the class corresponding to the given spatialimp string . | train | false |
10,705 | def escapeRegExpString(string):
toEscapeChars = ['\\', '(', ')', '.', '|', '^', '$', '*', '+', '?', '[', ']']
escapedValue = ''
for i in range(len(string)):
if (string[i] in toEscapeChars):
escapedValue += ('\\' + string[i])
else:
escapedValue += string[i]
return escapedValue
| [
"def",
"escapeRegExpString",
"(",
"string",
")",
":",
"toEscapeChars",
"=",
"[",
"'\\\\'",
",",
"'('",
",",
"')'",
",",
"'.'",
",",
"'|'",
",",
"'^'",
",",
"'$'",
",",
"'*'",
",",
"'+'",
",",
"'?'",
",",
"'['",
",",
"']'",
"]",
"escapedValue",
"=",... | escape the given string to include it as a regular expression . | train | false |
10,706 | def get_impression_pixel_url(codename):
mac = (codename + hashlib.sha1((codename + g.tracking_secret)).hexdigest())
v_param = ('?v=%s&' % _get_encrypted_user_slug())
hash_and_id_params = urllib.urlencode({'hash': mac, 'id': codename})
return ((g.adframetracker_url + v_param) + hash_and_id_params)
| [
"def",
"get_impression_pixel_url",
"(",
"codename",
")",
":",
"mac",
"=",
"(",
"codename",
"+",
"hashlib",
".",
"sha1",
"(",
"(",
"codename",
"+",
"g",
".",
"tracking_secret",
")",
")",
".",
"hexdigest",
"(",
")",
")",
"v_param",
"=",
"(",
"'?v=%s&'",
... | return a url to use for tracking impressions of the given advert . | train | false |
10,708 | def getComplexDefaultByDictionary(defaultComplex, dictionary, key):
if (key in dictionary):
return complex(dictionary[key].strip().replace('(', '').replace(')', ''))
return defaultComplex
| [
"def",
"getComplexDefaultByDictionary",
"(",
"defaultComplex",
",",
"dictionary",
",",
"key",
")",
":",
"if",
"(",
"key",
"in",
"dictionary",
")",
":",
"return",
"complex",
"(",
"dictionary",
"[",
"key",
"]",
".",
"strip",
"(",
")",
".",
"replace",
"(",
... | get the value as a complex . | train | false |
10,709 | def row_by_value(idl_, table, column, match, default=_NO_DEFAULT):
tab = idl_.tables[table]
for r in tab.rows.values():
if (getattr(r, column) == match):
return r
if (default is not _NO_DEFAULT):
return default
raise RowNotFound(table=table, col=column, match=match)
| [
"def",
"row_by_value",
"(",
"idl_",
",",
"table",
",",
"column",
",",
"match",
",",
"default",
"=",
"_NO_DEFAULT",
")",
":",
"tab",
"=",
"idl_",
".",
"tables",
"[",
"table",
"]",
"for",
"r",
"in",
"tab",
".",
"rows",
".",
"values",
"(",
")",
":",
... | lookup an idl row in a table by column/value . | train | false |
10,711 | def display_timestamps_pair(time_m_2):
if (len(time_m_2) == 0):
return '(empty)'
time_m_2 = np.array(time_m_2)
return '({}, {})'.format(display_timestamps(time_m_2[:, 0]), display_timestamps(time_m_2[:, 1]))
| [
"def",
"display_timestamps_pair",
"(",
"time_m_2",
")",
":",
"if",
"(",
"len",
"(",
"time_m_2",
")",
"==",
"0",
")",
":",
"return",
"'(empty)'",
"time_m_2",
"=",
"np",
".",
"array",
"(",
"time_m_2",
")",
"return",
"'({}, {})'",
".",
"format",
"(",
"displ... | takes a list of the following form: [ . | train | true |
10,712 | def get_fun(fun):
serv = _get_serv(ret=None)
ret = {}
for minion in serv.smembers('minions'):
ind_str = '{0}:{1}'.format(minion, fun)
try:
jid = serv.get(ind_str)
except Exception:
continue
if (not jid):
continue
data = serv.get('{0}:{1}'.format(minion, jid))
if data:
ret[minion] = json.loads(data)
return ret
| [
"def",
"get_fun",
"(",
"fun",
")",
":",
"serv",
"=",
"_get_serv",
"(",
"ret",
"=",
"None",
")",
"ret",
"=",
"{",
"}",
"for",
"minion",
"in",
"serv",
".",
"smembers",
"(",
"'minions'",
")",
":",
"ind_str",
"=",
"'{0}:{1}'",
".",
"format",
"(",
"mini... | return the most recent jobs that have executed the named function . | train | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.