id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
4,309 | def get_num_cat(sample_by_cat, samples_in_otus):
num_cat = defaultdict(int)
for (cat, samples) in sample_by_cat.items():
num_samples = len((set(samples_in_otus) & set(samples)))
num_cat[cat[0]] += ((num_samples * (num_samples - 1)) / 2)
return num_cat
| [
"def",
"get_num_cat",
"(",
"sample_by_cat",
",",
"samples_in_otus",
")",
":",
"num_cat",
"=",
"defaultdict",
"(",
"int",
")",
"for",
"(",
"cat",
",",
"samples",
")",
"in",
"sample_by_cat",
".",
"items",
"(",
")",
":",
"num_samples",
"=",
"len",
"(",
"(",... | builds a dictionary of numbers of samples keyed by metadata value . | train | false |
4,310 | @gof.local_optimizer([csm_properties])
def local_csm_properties_csm(node):
if (node.op == csm_properties):
(csm,) = node.inputs
if (csm.owner and ((csm.owner.op == CSC) or (csm.owner.op == CSR))):
ret_var = [theano.tensor.patternbroadcast(i, o.broadcastable) for (i, o) in izip(csm.owner.inputs, node.outputs)]
return ret_var
return False
| [
"@",
"gof",
".",
"local_optimizer",
"(",
"[",
"csm_properties",
"]",
")",
"def",
"local_csm_properties_csm",
"(",
"node",
")",
":",
"if",
"(",
"node",
".",
"op",
"==",
"csm_properties",
")",
":",
"(",
"csm",
",",
")",
"=",
"node",
".",
"inputs",
"if",
... | if we find csm_properties(csm) . | train | false |
4,311 | def getThreadPolicy(getDefault, flavour):
if importCtypesFailed:
return False
extendedPolicy = _timeConstraintThreadPolicy()
getDefault = ctypes.c_int(getDefault)
err = cocoa.thread_policy_get(cocoa.mach_thread_self(), THREAD_TIME_CONSTRAINT_POLICY, ctypes.byref(extendedPolicy), ctypes.byref(THREAD_TIME_CONSTRAINT_POLICY_COUNT), ctypes.byref(getDefault))
return extendedPolicy
| [
"def",
"getThreadPolicy",
"(",
"getDefault",
",",
"flavour",
")",
":",
"if",
"importCtypesFailed",
":",
"return",
"False",
"extendedPolicy",
"=",
"_timeConstraintThreadPolicy",
"(",
")",
"getDefault",
"=",
"ctypes",
".",
"c_int",
"(",
"getDefault",
")",
"err",
"... | retrieve the current thread policy . | train | false |
4,312 | @docstring.dedent_interpd
def phase_spectrum(x, Fs=None, window=None, pad_to=None, sides=None):
return _single_spectrum_helper(x=x, Fs=Fs, window=window, pad_to=pad_to, sides=sides, mode=u'phase')
| [
"@",
"docstring",
".",
"dedent_interpd",
"def",
"phase_spectrum",
"(",
"x",
",",
"Fs",
"=",
"None",
",",
"window",
"=",
"None",
",",
"pad_to",
"=",
"None",
",",
"sides",
"=",
"None",
")",
":",
"return",
"_single_spectrum_helper",
"(",
"x",
"=",
"x",
",... | compute the phase of the frequency spectrum of *x* . | train | false |
4,314 | def annotated(letter):
ucode_pics = {'F': (2, 0, 2, 0, u'\u250c\u2500\n\u251c\u2500\n\u2575'), 'G': (3, 0, 3, 1, u'\u256d\u2500\u256e\n\u2502\u2576\u2510\n\u2570\u2500\u256f')}
ascii_pics = {'F': (3, 0, 3, 0, ' _\n|_\n|\n'), 'G': (3, 0, 3, 1, ' __\n/__\n\\_|')}
if _use_unicode:
return ucode_pics[letter]
else:
return ascii_pics[letter]
| [
"def",
"annotated",
"(",
"letter",
")",
":",
"ucode_pics",
"=",
"{",
"'F'",
":",
"(",
"2",
",",
"0",
",",
"2",
",",
"0",
",",
"u'\\u250c\\u2500\\n\\u251c\\u2500\\n\\u2575'",
")",
",",
"'G'",
":",
"(",
"3",
",",
"0",
",",
"3",
",",
"1",
",",
"u'\\u2... | return a stylised drawing of the letter letter . | train | false |
4,315 | def features_contains(value):
return var_contains('FEATURES', value)
| [
"def",
"features_contains",
"(",
"value",
")",
":",
"return",
"var_contains",
"(",
"'FEATURES'",
",",
"value",
")"
] | verify if features variable contains a value in make . | train | false |
4,316 | def find_control_points(c1x, c1y, mmx, mmy, c2x, c2y):
cmx = (0.5 * ((4 * mmx) - (c1x + c2x)))
cmy = (0.5 * ((4 * mmy) - (c1y + c2y)))
return [(c1x, c1y), (cmx, cmy), (c2x, c2y)]
| [
"def",
"find_control_points",
"(",
"c1x",
",",
"c1y",
",",
"mmx",
",",
"mmy",
",",
"c2x",
",",
"c2y",
")",
":",
"cmx",
"=",
"(",
"0.5",
"*",
"(",
"(",
"4",
"*",
"mmx",
")",
"-",
"(",
"c1x",
"+",
"c2x",
")",
")",
")",
"cmy",
"=",
"(",
"0.5",... | find control points of the bezier line throught c1 . | train | false |
4,317 | def getRightStripAlphabetPercent(word):
word = word.strip()
for characterIndex in xrange((len(word) - 1), (-1), (-1)):
character = word[characterIndex]
if ((not character.isalpha()) and (not (character == '%'))):
return float(word[:(characterIndex + 1)])
return None
| [
"def",
"getRightStripAlphabetPercent",
"(",
"word",
")",
":",
"word",
"=",
"word",
".",
"strip",
"(",
")",
"for",
"characterIndex",
"in",
"xrange",
"(",
"(",
"len",
"(",
"word",
")",
"-",
"1",
")",
",",
"(",
"-",
"1",
")",
",",
"(",
"-",
"1",
")"... | get word with alphabet characters and the percent sign stripped from the right . | train | false |
4,318 | def _check_minions_directories_raetkey(pki_dir):
accepted = os.path.join(pki_dir, salt.key.RaetKey.ACC)
pre = os.path.join(pki_dir, salt.key.RaetKey.PEND)
rejected = os.path.join(pki_dir, salt.key.RaetKey.REJ)
return (accepted, pre, rejected)
| [
"def",
"_check_minions_directories_raetkey",
"(",
"pki_dir",
")",
":",
"accepted",
"=",
"os",
".",
"path",
".",
"join",
"(",
"pki_dir",
",",
"salt",
".",
"key",
".",
"RaetKey",
".",
"ACC",
")",
"pre",
"=",
"os",
".",
"path",
".",
"join",
"(",
"pki_dir"... | return the minion keys directory paths . | train | false |
4,319 | def assert_urls(host_blocker, blocked=BLOCKLIST_HOSTS, whitelisted=WHITELISTED_HOSTS, urls_to_check=URLS_TO_CHECK):
whitelisted = (list(whitelisted) + list(host_blocker.WHITELISTED))
for str_url in urls_to_check:
url = QUrl(str_url)
host = url.host()
if ((host in blocked) and (host not in whitelisted)):
assert host_blocker.is_blocked(url)
else:
assert (not host_blocker.is_blocked(url))
| [
"def",
"assert_urls",
"(",
"host_blocker",
",",
"blocked",
"=",
"BLOCKLIST_HOSTS",
",",
"whitelisted",
"=",
"WHITELISTED_HOSTS",
",",
"urls_to_check",
"=",
"URLS_TO_CHECK",
")",
":",
"whitelisted",
"=",
"(",
"list",
"(",
"whitelisted",
")",
"+",
"list",
"(",
"... | test if urls to check are blocked or not by hostblocker . | train | false |
4,320 | def register_scale(scale_class):
_scale_mapping[scale_class.name] = scale_class
| [
"def",
"register_scale",
"(",
"scale_class",
")",
":",
"_scale_mapping",
"[",
"scale_class",
".",
"name",
"]",
"=",
"scale_class"
] | register a new kind of scale . | train | false |
4,322 | def _compare_epochs_infos(info1, info2, ind):
info1._check_consistency()
info2._check_consistency()
if (info1['nchan'] != info2['nchan']):
raise ValueError(("epochs[%d]['info']['nchan'] must match" % ind))
if (info1['bads'] != info2['bads']):
raise ValueError(("epochs[%d]['info']['bads'] must match" % ind))
if (info1['sfreq'] != info2['sfreq']):
raise ValueError(("epochs[%d]['info']['sfreq'] must match" % ind))
if (set(info1['ch_names']) != set(info2['ch_names'])):
raise ValueError(("epochs[%d]['info']['ch_names'] must match" % ind))
if (len(info2['projs']) != len(info1['projs'])):
raise ValueError('SSP projectors in epochs files must be the same')
if any(((not _proj_equal(p1, p2)) for (p1, p2) in zip(info2['projs'], info1['projs']))):
raise ValueError('SSP projectors in epochs files must be the same')
if (((info1['dev_head_t'] is None) != (info2['dev_head_t'] is None)) or ((info1['dev_head_t'] is not None) and (not np.allclose(info1['dev_head_t']['trans'], info2['dev_head_t']['trans'], rtol=1e-06)))):
raise ValueError(("epochs[%d]['info']['dev_head_t'] must match. The epochs probably come from different runs, and are therefore associated with different head positions. Manually change info['dev_head_t'] to avoid this message but beware that this means the MEG sensors will not be properly spatially aligned. See mne.preprocessing.maxwell_filter to realign the runs to a common head position." % ind))
| [
"def",
"_compare_epochs_infos",
"(",
"info1",
",",
"info2",
",",
"ind",
")",
":",
"info1",
".",
"_check_consistency",
"(",
")",
"info2",
".",
"_check_consistency",
"(",
")",
"if",
"(",
"info1",
"[",
"'nchan'",
"]",
"!=",
"info2",
"[",
"'nchan'",
"]",
")"... | compare infos . | train | false |
4,323 | def dup_ff_lcm(f, g, K):
h = dup_quo(dup_mul(f, g, K), dup_gcd(f, g, K), K)
return dup_monic(h, K)
| [
"def",
"dup_ff_lcm",
"(",
"f",
",",
"g",
",",
"K",
")",
":",
"h",
"=",
"dup_quo",
"(",
"dup_mul",
"(",
"f",
",",
"g",
",",
"K",
")",
",",
"dup_gcd",
"(",
"f",
",",
"g",
",",
"K",
")",
",",
"K",
")",
"return",
"dup_monic",
"(",
"h",
",",
"... | computes polynomial lcm over a field in k[x] . | train | false |
4,325 | def regen_keys():
for fn_ in os.listdir(__opts__['pki_dir']):
path = os.path.join(__opts__['pki_dir'], fn_)
try:
os.remove(path)
except os.error:
pass
channel = salt.transport.Channel.factory(__opts__)
| [
"def",
"regen_keys",
"(",
")",
":",
"for",
"fn_",
"in",
"os",
".",
"listdir",
"(",
"__opts__",
"[",
"'pki_dir'",
"]",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"__opts__",
"[",
"'pki_dir'",
"]",
",",
"fn_",
")",
"try",
":",
"os"... | used to regenerate the minion keys . | train | false |
4,326 | def init(mpstate):
return SerialModule(mpstate)
| [
"def",
"init",
"(",
"mpstate",
")",
":",
"return",
"SerialModule",
"(",
"mpstate",
")"
] | initalize the platform with devices . | train | false |
4,329 | @contextmanager
def silence_stderr():
if DEBUG:
(yield)
else:
with threading.Lock():
stderr = sys.stderr
sys.stderr = StringIO()
(yield)
with threading.Lock():
sys.stderr = stderr
| [
"@",
"contextmanager",
"def",
"silence_stderr",
"(",
")",
":",
"if",
"DEBUG",
":",
"(",
"yield",
")",
"else",
":",
"with",
"threading",
".",
"Lock",
"(",
")",
":",
"stderr",
"=",
"sys",
".",
"stderr",
"sys",
".",
"stderr",
"=",
"StringIO",
"(",
")",
... | redirect stderr . | train | false |
4,330 | @_np.deprecate(message='scipy.constants.C2F is deprecated in scipy 0.18.0. Use scipy.constants.convert_temperature instead. Note that the new function has a different signature.')
def C2F(C):
return ((1.8 * _np.asanyarray(C)) + 32)
| [
"@",
"_np",
".",
"deprecate",
"(",
"message",
"=",
"'scipy.constants.C2F is deprecated in scipy 0.18.0. Use scipy.constants.convert_temperature instead. Note that the new function has a different signature.'",
")",
"def",
"C2F",
"(",
"C",
")",
":",
"return",
"(",
"(",
"1.8",
"*... | convert celsius to fahrenheit parameters c : array_like celsius temperature(s) to be converted . | train | false |
4,331 | def vserver_exists(v_name, v_ip=None, v_port=None, v_type=None, **connection_args):
vserver = _vserver_get(v_name, **connection_args)
if (vserver is None):
return False
if ((v_ip is not None) and (vserver.get_ipv46() != v_ip)):
return False
if ((v_port is not None) and (vserver.get_port() != v_port)):
return False
if ((v_type is not None) and (vserver.get_servicetype().upper() != v_type.upper())):
return False
return True
| [
"def",
"vserver_exists",
"(",
"v_name",
",",
"v_ip",
"=",
"None",
",",
"v_port",
"=",
"None",
",",
"v_type",
"=",
"None",
",",
"**",
"connection_args",
")",
":",
"vserver",
"=",
"_vserver_get",
"(",
"v_name",
",",
"**",
"connection_args",
")",
"if",
"(",... | checks if a vserver exists cli example: . | train | true |
4,333 | def task_coverage():
return {'task_dep': ['locale', 'doctest'], 'actions': ['py.test --cov nikola --cov-report term-missing tests/'], 'verbosity': 2}
| [
"def",
"task_coverage",
"(",
")",
":",
"return",
"{",
"'task_dep'",
":",
"[",
"'locale'",
",",
"'doctest'",
"]",
",",
"'actions'",
":",
"[",
"'py.test --cov nikola --cov-report term-missing tests/'",
"]",
",",
"'verbosity'",
":",
"2",
"}"
] | run unit-tests using py . | train | false |
4,334 | def prepare_request(uri, headers=None, data=None, method=None):
if (headers is None):
headers = {}
if (data and (not method)):
method = 'POST'
elif (not method):
method = 'GET'
if ((method == 'GET') and data):
uri = add_params_to_uri(uri, data)
data = None
return (uri, headers, data, method)
| [
"def",
"prepare_request",
"(",
"uri",
",",
"headers",
"=",
"None",
",",
"data",
"=",
"None",
",",
"method",
"=",
"None",
")",
":",
"if",
"(",
"headers",
"is",
"None",
")",
":",
"headers",
"=",
"{",
"}",
"if",
"(",
"data",
"and",
"(",
"not",
"meth... | make request parameters right . | train | true |
4,335 | def _sysfs_attr(name, value=None, log_lvl=None, log_msg=None):
if isinstance(name, six.string_types):
name = [name]
res = __salt__['sysfs.attr'](os.path.join(*name), value)
if ((not res) and (log_lvl is not None)):
log.log(LOG[log_lvl], log_msg)
return res
| [
"def",
"_sysfs_attr",
"(",
"name",
",",
"value",
"=",
"None",
",",
"log_lvl",
"=",
"None",
",",
"log_msg",
"=",
"None",
")",
":",
"if",
"isinstance",
"(",
"name",
",",
"six",
".",
"string_types",
")",
":",
"name",
"=",
"[",
"name",
"]",
"res",
"=",... | simple wrapper with logging around sysfs . | train | true |
4,338 | def platform_supported(rospack, pkg, os, version):
return _platform_supported(rospack.get_manifest(pkg), os, version)
| [
"def",
"platform_supported",
"(",
"rospack",
",",
"pkg",
",",
"os",
",",
"version",
")",
":",
"return",
"_platform_supported",
"(",
"rospack",
".",
"get_manifest",
"(",
"pkg",
")",
",",
"os",
",",
"version",
")"
] | return whether the platform defined by os and version is marked as supported in the package . | train | false |
4,339 | def _check_is_not_multitable(values, model):
used_models = set()
for field in values:
if isinstance(field, sqlalchemy.orm.attributes.InstrumentedAttribute):
used_models.add(field.class_)
elif isinstance(field, six.string_types):
used_models.add(model)
else:
raise exception.ProgrammingError(reason='DB Conditional update - Unknown field type, must be string or ORM field.')
if (len(used_models) > 1):
raise exception.ProgrammingError(reason='DB Conditional update - Error in query, multitable updates are not supported.')
| [
"def",
"_check_is_not_multitable",
"(",
"values",
",",
"model",
")",
":",
"used_models",
"=",
"set",
"(",
")",
"for",
"field",
"in",
"values",
":",
"if",
"isinstance",
"(",
"field",
",",
"sqlalchemy",
".",
"orm",
".",
"attributes",
".",
"InstrumentedAttribut... | check that we dont try to do multitable updates . | train | false |
4,340 | def captured_stderr():
return captured_output('stderr')
| [
"def",
"captured_stderr",
"(",
")",
":",
"return",
"captured_output",
"(",
"'stderr'",
")"
] | return a context manager used by captured_stdout/stdin/stderr that temporarily replaces the sys stream *stream_name* with a stringio . | train | false |
4,341 | def use_conn_pool(func):
def wrapper(self, *args, **kwargs):
with self._get_pool_connection() as conn:
self._apply_options(conn)
return func(self, conn, *args, **kwargs)
return wrapper
| [
"def",
"use_conn_pool",
"(",
"func",
")",
":",
"def",
"wrapper",
"(",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"with",
"self",
".",
"_get_pool_connection",
"(",
")",
"as",
"conn",
":",
"self",
".",
"_apply_options",
"(",
"conn",
")",
... | use this only for connection pool specific ldap api . | train | false |
4,343 | def block_diag(*arrs):
if (arrs == ()):
arrs = ([],)
arrs = [np.atleast_2d(a) for a in arrs]
bad_args = [k for k in range(len(arrs)) if (arrs[k].ndim > 2)]
if bad_args:
raise ValueError(('arguments in the following positions have dimension greater than 2: %s' % bad_args))
shapes = np.array([a.shape for a in arrs])
out_dtype = np.find_common_type([arr.dtype for arr in arrs], [])
out = np.zeros(np.sum(shapes, axis=0), dtype=out_dtype)
(r, c) = (0, 0)
for (i, (rr, cc)) in enumerate(shapes):
out[r:(r + rr), c:(c + cc)] = arrs[i]
r += rr
c += cc
return out
| [
"def",
"block_diag",
"(",
"*",
"arrs",
")",
":",
"if",
"(",
"arrs",
"==",
"(",
")",
")",
":",
"arrs",
"=",
"(",
"[",
"]",
",",
")",
"arrs",
"=",
"[",
"np",
".",
"atleast_2d",
"(",
"a",
")",
"for",
"a",
"in",
"arrs",
"]",
"bad_args",
"=",
"[... | build a block diagonal sparse matrix from provided matrices . | train | false |
4,345 | def discardLogs():
global logfile
logfile = NullFile()
| [
"def",
"discardLogs",
"(",
")",
":",
"global",
"logfile",
"logfile",
"=",
"NullFile",
"(",
")"
] | throw away all logs . | train | false |
4,348 | def commit_message_path():
path = git.git_path(u'GIT_COLA_MSG')
if core.exists(path):
return path
return None
| [
"def",
"commit_message_path",
"(",
")",
":",
"path",
"=",
"git",
".",
"git_path",
"(",
"u'GIT_COLA_MSG'",
")",
"if",
"core",
".",
"exists",
"(",
"path",
")",
":",
"return",
"path",
"return",
"None"
] | return the path to . | train | false |
4,349 | def Var(xs, mu=None, ddof=0):
xs = np.asarray(xs)
if (mu is None):
mu = xs.mean()
ds = (xs - mu)
return (np.dot(ds, ds) / (len(xs) - ddof))
| [
"def",
"Var",
"(",
"xs",
",",
"mu",
"=",
"None",
",",
"ddof",
"=",
"0",
")",
":",
"xs",
"=",
"np",
".",
"asarray",
"(",
"xs",
")",
"if",
"(",
"mu",
"is",
"None",
")",
":",
"mu",
"=",
"xs",
".",
"mean",
"(",
")",
"ds",
"=",
"(",
"xs",
"-... | computes variance . | train | false |
4,354 | def get_closest_dir(workdir):
closest_dir = ''
for wdi in path_split_all(workdir):
if os.path.isdir(os.path.join(closest_dir, wdi)):
closest_dir = os.path.join(closest_dir, wdi)
else:
break
assert (closest_dir != workdir)
return (closest_dir, wdi)
| [
"def",
"get_closest_dir",
"(",
"workdir",
")",
":",
"closest_dir",
"=",
"''",
"for",
"wdi",
"in",
"path_split_all",
"(",
"workdir",
")",
":",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"os",
".",
"path",
".",
"join",
"(",
"closest_dir",
",",
"wdi",
... | returns the topmost already-existing directory in the given path erasing work-dirs should never progress above this file . | train | false |
4,356 | @pytest.mark.parametrize(u'mode', modes)
def test_gaussian_eval_1D(mode):
model = Gaussian1D(1, 0, 20)
x = np.arange((-100), 101)
values = model(x)
disc_values = discretize_model(model, ((-100), 101), mode=mode)
assert_allclose(values, disc_values, atol=0.001)
| [
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"u'mode'",
",",
"modes",
")",
"def",
"test_gaussian_eval_1D",
"(",
"mode",
")",
":",
"model",
"=",
"Gaussian1D",
"(",
"1",
",",
"0",
",",
"20",
")",
"x",
"=",
"np",
".",
"arange",
"(",
"(",
"-",
... | discretize gaussian with different modes and check if result is at least similar to gaussian1d . | train | false |
4,357 | def _get_read_region(read_name):
return int(read_name[8])
| [
"def",
"_get_read_region",
"(",
"read_name",
")",
":",
"return",
"int",
"(",
"read_name",
"[",
"8",
"]",
")"
] | extract region from read name . | train | false |
4,358 | def zpad(x, l):
return (('\x00' * max(0, (l - len(x)))) + x)
| [
"def",
"zpad",
"(",
"x",
",",
"l",
")",
":",
"return",
"(",
"(",
"'\\x00'",
"*",
"max",
"(",
"0",
",",
"(",
"l",
"-",
"len",
"(",
"x",
")",
")",
")",
")",
"+",
"x",
")"
] | left zero pad value x at least to length l . | train | false |
4,359 | def set_parser(new_parser_fun=None):
global parser_fun
if (new_parser_fun is None):
new_parser_fun = mb_parser_xml
if (not callable(new_parser_fun)):
raise ValueError('new_parser_fun must be callable')
parser_fun = new_parser_fun
| [
"def",
"set_parser",
"(",
"new_parser_fun",
"=",
"None",
")",
":",
"global",
"parser_fun",
"if",
"(",
"new_parser_fun",
"is",
"None",
")",
":",
"new_parser_fun",
"=",
"mb_parser_xml",
"if",
"(",
"not",
"callable",
"(",
"new_parser_fun",
")",
")",
":",
"raise... | sets the function used to parse the response from the musicbrainz web service . | train | false |
4,360 | def select_row(view, row):
selmodel = view.selectionModel()
selmodel.select(view.model().index(row, 0), QItemSelectionModel.ClearAndSelect)
| [
"def",
"select_row",
"(",
"view",
",",
"row",
")",
":",
"selmodel",
"=",
"view",
".",
"selectionModel",
"(",
")",
"selmodel",
".",
"select",
"(",
"view",
".",
"model",
"(",
")",
".",
"index",
"(",
"row",
",",
"0",
")",
",",
"QItemSelectionModel",
"."... | select a row in an item view . | train | false |
4,361 | @login_required
def preview_handler(request, usage_key_string, handler, suffix=''):
usage_key = UsageKey.from_string(usage_key_string)
if isinstance(usage_key, (AsideUsageKeyV1, AsideUsageKeyV2)):
descriptor = modulestore().get_item(usage_key.usage_key)
for aside in descriptor.runtime.get_asides(descriptor):
if (aside.scope_ids.block_type == usage_key.aside_type):
asides = [aside]
instance = aside
break
else:
descriptor = modulestore().get_item(usage_key)
instance = _load_preview_module(request, descriptor)
asides = []
req = django_to_webob_request(request)
try:
resp = instance.handle(handler, req, suffix)
except NoSuchHandlerError:
log.exception('XBlock %s attempted to access missing handler %r', instance, handler)
raise Http404
except NotFoundError:
log.exception("Module indicating to user that request doesn't exist")
raise Http404
except ProcessingError:
log.warning('Module raised an error while processing AJAX request', exc_info=True)
return HttpResponseBadRequest()
except Exception:
log.exception('error processing ajax call')
raise
modulestore().update_item(descriptor, request.user.id, asides=asides)
return webob_to_django_response(resp)
| [
"@",
"login_required",
"def",
"preview_handler",
"(",
"request",
",",
"usage_key_string",
",",
"handler",
",",
"suffix",
"=",
"''",
")",
":",
"usage_key",
"=",
"UsageKey",
".",
"from_string",
"(",
"usage_key_string",
")",
"if",
"isinstance",
"(",
"usage_key",
... | dispatch an ajax action to an xblock usage_key_string: the usage_key_string-id of the block to dispatch to . | train | false |
4,362 | def _deserialize_spec_test(data, file_path):
context = data['data']
description = data['desc']
expected = unicode(data['expected'])
partials = ((data.has_key('partials') and data['partials']) or {})
template = data['template']
test_name = data['name']
_convert_children(context)
test_case = _make_spec_test(expected, template, context, partials, description, test_name, file_path)
return test_case
| [
"def",
"_deserialize_spec_test",
"(",
"data",
",",
"file_path",
")",
":",
"context",
"=",
"data",
"[",
"'data'",
"]",
"description",
"=",
"data",
"[",
"'desc'",
"]",
"expected",
"=",
"unicode",
"(",
"data",
"[",
"'expected'",
"]",
")",
"partials",
"=",
"... | return a unittest . | train | false |
4,363 | def getIndexOfStartingWithSecond(letter, splitLine):
for wordIndex in xrange(1, len(splitLine)):
word = splitLine[wordIndex]
firstLetter = word[0]
if (firstLetter == letter):
return wordIndex
return (-1)
| [
"def",
"getIndexOfStartingWithSecond",
"(",
"letter",
",",
"splitLine",
")",
":",
"for",
"wordIndex",
"in",
"xrange",
"(",
"1",
",",
"len",
"(",
"splitLine",
")",
")",
":",
"word",
"=",
"splitLine",
"[",
"wordIndex",
"]",
"firstLetter",
"=",
"word",
"[",
... | get index of the first occurence of the given letter in the split line . | train | false |
4,364 | def subscribe_to_exploration(user_id, exploration_id):
subscriptions_model = user_models.UserSubscriptionsModel.get(user_id, strict=False)
if (not subscriptions_model):
subscriptions_model = user_models.UserSubscriptionsModel(id=user_id)
if (exploration_id not in subscriptions_model.activity_ids):
subscriptions_model.activity_ids.append(exploration_id)
subscriptions_model.put()
| [
"def",
"subscribe_to_exploration",
"(",
"user_id",
",",
"exploration_id",
")",
":",
"subscriptions_model",
"=",
"user_models",
".",
"UserSubscriptionsModel",
".",
"get",
"(",
"user_id",
",",
"strict",
"=",
"False",
")",
"if",
"(",
"not",
"subscriptions_model",
")"... | subscribes a user to an exploration . | train | false |
4,366 | def _conv(obj, dtype=None):
if (obj is None):
return obj
else:
if (dtype is None):
obj = numpy.asarray(obj)
else:
obj = numpy.asarray(obj, dtype)
if (obj.shape == ()):
return obj.dtype.type(obj)
else:
return obj
| [
"def",
"_conv",
"(",
"obj",
",",
"dtype",
"=",
"None",
")",
":",
"if",
"(",
"obj",
"is",
"None",
")",
":",
"return",
"obj",
"else",
":",
"if",
"(",
"dtype",
"is",
"None",
")",
":",
"obj",
"=",
"numpy",
".",
"asarray",
"(",
"obj",
")",
"else",
... | convert an object to the preferred form for input to the odr routine . | train | false |
4,367 | def sysctlTestAndSet(name, limit):
if ('/' not in name):
name = ('/proc/sys/' + name.replace('.', '/'))
with open(name, 'r') as readFile:
oldLimit = readFile.readline()
if isinstance(limit, int):
if (int(oldLimit) < limit):
with open(name, 'w') as writeFile:
writeFile.write(('%d' % limit))
else:
with open(name, 'w') as writeFile:
writeFile.write(limit)
| [
"def",
"sysctlTestAndSet",
"(",
"name",
",",
"limit",
")",
":",
"if",
"(",
"'/'",
"not",
"in",
"name",
")",
":",
"name",
"=",
"(",
"'/proc/sys/'",
"+",
"name",
".",
"replace",
"(",
"'.'",
",",
"'/'",
")",
")",
"with",
"open",
"(",
"name",
",",
"'... | helper function to set sysctl limits . | train | false |
4,368 | def remove_borders_from_image(img, fuzz=None):
fuzz = (tweaks[u'cover_trim_fuzz_value'] if (fuzz is None) else fuzz)
img = image_from_data(img)
ans = imageops.remove_borders(img, max(0, fuzz))
return (ans if (ans.size() != img.size()) else img)
| [
"def",
"remove_borders_from_image",
"(",
"img",
",",
"fuzz",
"=",
"None",
")",
":",
"fuzz",
"=",
"(",
"tweaks",
"[",
"u'cover_trim_fuzz_value'",
"]",
"if",
"(",
"fuzz",
"is",
"None",
")",
"else",
"fuzz",
")",
"img",
"=",
"image_from_data",
"(",
"img",
")... | try to auto-detect and remove any borders from the image . | train | false |
4,369 | def remove_course_milestones(course_key, user, relationship):
if (not settings.FEATURES.get('MILESTONES_APP')):
return None
course_milestones = milestones_api.get_course_milestones(course_key=course_key, relationship=relationship)
for milestone in course_milestones:
milestones_api.remove_user_milestone({'id': user.id}, milestone)
| [
"def",
"remove_course_milestones",
"(",
"course_key",
",",
"user",
",",
"relationship",
")",
":",
"if",
"(",
"not",
"settings",
".",
"FEATURES",
".",
"get",
"(",
"'MILESTONES_APP'",
")",
")",
":",
"return",
"None",
"course_milestones",
"=",
"milestones_api",
"... | remove all user milestones for the course specified by course_key . | train | false |
4,372 | def load_processed_files(path):
processed_files = {}
with open(path) as input_file:
for line in input_file.readlines():
line = line.strip()
if (not line):
continue
if (' ' not in line):
raise TypeError(('Malformed line: %s' % line))
(path, timestamp) = line.rsplit(' ', 1)
if (not os.path.isabs(path)):
raise TypeError(("'%s' is not an absolute path" % path))
elif (not timestamp.isdigit()):
raise TypeError(("'%s' is not an integer timestamp" % timestamp))
processed_files[path] = int(timestamp)
return processed_files
| [
"def",
"load_processed_files",
"(",
"path",
")",
":",
"processed_files",
"=",
"{",
"}",
"with",
"open",
"(",
"path",
")",
"as",
"input_file",
":",
"for",
"line",
"in",
"input_file",
".",
"readlines",
"(",
")",
":",
"line",
"=",
"line",
".",
"strip",
"(... | loads a dictionary of path => last modified timestamp mappings . | train | false |
4,373 | def get_basename(fileName):
if fileName.endswith(os.path.sep):
fileName = fileName[:(-1)]
return os.path.basename(fileName)
| [
"def",
"get_basename",
"(",
"fileName",
")",
":",
"if",
"fileName",
".",
"endswith",
"(",
"os",
".",
"path",
".",
"sep",
")",
":",
"fileName",
"=",
"fileName",
"[",
":",
"(",
"-",
"1",
")",
"]",
"return",
"os",
".",
"path",
".",
"basename",
"(",
... | get the name of a file or folder specified in a path . | train | false |
4,374 | def get_industry_classified(standard='sina'):
if (standard == 'sw'):
df = _get_type_data((ct.SINA_INDUSTRY_INDEX_URL % (ct.P_TYPE['http'], ct.DOMAINS['vsf'], ct.PAGES['ids_sw'])))
else:
df = _get_type_data((ct.SINA_INDUSTRY_INDEX_URL % (ct.P_TYPE['http'], ct.DOMAINS['vsf'], ct.PAGES['ids'])))
data = []
ct._write_head()
for row in df.values:
rowDf = _get_detail(row[0], retry_count=10, pause=0.01)
rowDf['c_name'] = row[1]
data.append(rowDf)
data = pd.concat(data, ignore_index=True)
return data
| [
"def",
"get_industry_classified",
"(",
"standard",
"=",
"'sina'",
")",
":",
"if",
"(",
"standard",
"==",
"'sw'",
")",
":",
"df",
"=",
"_get_type_data",
"(",
"(",
"ct",
".",
"SINA_INDUSTRY_INDEX_URL",
"%",
"(",
"ct",
".",
"P_TYPE",
"[",
"'http'",
"]",
","... | parameters standard sina:新浪行业 sw:申万 行业 returns dataframe code :股票代码 name :股票名称 c_name :行业名称 . | train | false |
4,375 | def bins(data, values=None, column=None, bins=None, labels=None, **kwargs):
if isinstance(data, str):
column = data
values = None
else:
column = None
return Bins(values=values, column=column, bins=bins, **kwargs)
| [
"def",
"bins",
"(",
"data",
",",
"values",
"=",
"None",
",",
"column",
"=",
"None",
",",
"bins",
"=",
"None",
",",
"labels",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"if",
"isinstance",
"(",
"data",
",",
"str",
")",
":",
"column",
"=",
"data",... | specify binning or bins to be used for column or values . | train | false |
4,376 | def format_datetime(time_obj):
timestamp = time_obj.isoformat('T')
if (time_obj.tzinfo == tzutc()):
timestamp = timestamp[:(-6)]
return ('%sZ' % timestamp)
return timestamp
| [
"def",
"format_datetime",
"(",
"time_obj",
")",
":",
"timestamp",
"=",
"time_obj",
".",
"isoformat",
"(",
"'T'",
")",
"if",
"(",
"time_obj",
".",
"tzinfo",
"==",
"tzutc",
"(",
")",
")",
":",
"timestamp",
"=",
"timestamp",
"[",
":",
"(",
"-",
"6",
")"... | see :meth:i18n . | train | false |
4,377 | @memoize
def get_parent_until(path):
dirname = osp.dirname(path)
try:
mod = osp.basename(path)
mod = osp.splitext(mod)[0]
imp.find_module(mod, [dirname])
except ImportError:
return
items = [mod]
while 1:
items.append(osp.basename(dirname))
try:
dirname = osp.dirname(dirname)
imp.find_module('__init__', [(dirname + os.sep)])
except ImportError:
break
return '.'.join(reversed(items))
| [
"@",
"memoize",
"def",
"get_parent_until",
"(",
"path",
")",
":",
"dirname",
"=",
"osp",
".",
"dirname",
"(",
"path",
")",
"try",
":",
"mod",
"=",
"osp",
".",
"basename",
"(",
"path",
")",
"mod",
"=",
"osp",
".",
"splitext",
"(",
"mod",
")",
"[",
... | given a file path . | train | true |
4,378 | def volume_create(**kwargs):
return create_volume(kwargs, 'function')
| [
"def",
"volume_create",
"(",
"**",
"kwargs",
")",
":",
"return",
"create_volume",
"(",
"kwargs",
",",
"'function'",
")"
] | create a block storage volume name name of the new volume size volume size snapshot block storage snapshot id voltype type of storage profile profile to build on cli example: . | train | false |
4,379 | def window_none(x):
return x
| [
"def",
"window_none",
"(",
"x",
")",
":",
"return",
"x"
] | no window function; simply return x . | train | false |
4,380 | def date_from_str(date_str):
today = datetime.date.today()
if (date_str in (u'now', u'today')):
return today
if (date_str == u'yesterday'):
return (today - datetime.timedelta(days=1))
match = re.match(u'(now|today)(?P<sign>[+-])(?P<time>\\d+)(?P<unit>day|week|month|year)(s)?', date_str)
if (match is not None):
sign = match.group(u'sign')
time = int(match.group(u'time'))
if (sign == u'-'):
time = (- time)
unit = match.group(u'unit')
if (unit == u'month'):
unit = u'day'
time *= 30
elif (unit == u'year'):
unit = u'day'
time *= 365
unit += u's'
delta = datetime.timedelta(**{unit: time})
return (today + delta)
return datetime.datetime.strptime(date_str, u'%Y%m%d').date()
| [
"def",
"date_from_str",
"(",
"date_str",
")",
":",
"today",
"=",
"datetime",
".",
"date",
".",
"today",
"(",
")",
"if",
"(",
"date_str",
"in",
"(",
"u'now'",
",",
"u'today'",
")",
")",
":",
"return",
"today",
"if",
"(",
"date_str",
"==",
"u'yesterday'"... | return a datetime object from a string in the format yyyymmdd or [+-][0-9](s)? . | train | false |
4,384 | def facilityNetToMs():
a = TpPd(pd=3)
b = MessageType(mesType=58)
c = Facility()
packet = ((a / b) / c)
return packet
| [
"def",
"facilityNetToMs",
"(",
")",
":",
"a",
"=",
"TpPd",
"(",
"pd",
"=",
"3",
")",
"b",
"=",
"MessageType",
"(",
"mesType",
"=",
"58",
")",
"c",
"=",
"Facility",
"(",
")",
"packet",
"=",
"(",
"(",
"a",
"/",
"b",
")",
"/",
"c",
")",
"return"... | facility section 9 . | train | true |
4,387 | def load_images(img_names):
if (img_names[0] == ''):
return {}
for image_name in img_names:
img = open(image_name)
loaded_imgs = {}
img_list = ''
img_line = ' '
name = img.readline().replace('\n', '')
name = name[1:]
while True:
img_line = img.readline()
if (img_line == ''):
break
img_line.replace('\n', '')
if (img_line[0] == ':'):
loaded_imgs[name] = json.loads(img_list)
name = img_line[1:]
img_list = ''
else:
img_list += img_line
loaded_imgs[name] = json.loads(img_list)
return loaded_imgs
| [
"def",
"load_images",
"(",
"img_names",
")",
":",
"if",
"(",
"img_names",
"[",
"0",
"]",
"==",
"''",
")",
":",
"return",
"{",
"}",
"for",
"image_name",
"in",
"img_names",
":",
"img",
"=",
"open",
"(",
"image_name",
")",
"loaded_imgs",
"=",
"{",
"}",
... | loads user images from given file(s) . | train | false |
4,388 | def upgrade_to_float64(*types):
return (get_scalar_type('float64'),)
| [
"def",
"upgrade_to_float64",
"(",
"*",
"types",
")",
":",
"return",
"(",
"get_scalar_type",
"(",
"'float64'",
")",
",",
")"
] | upgrade any int and float32 to float64 to do as scipy . | train | false |
4,389 | def geocode():
vars = request.post_vars
street = vars.get('address', None)
postcode = vars.get('postcode', None)
L0 = vars.get('L0', None)
if L0:
L0 = int(L0)
L1 = vars.get('L1', None)
if L1:
L1 = int(L1)
L2 = vars.get('L2', None)
if L2:
L2 = int(L2)
L3 = vars.get('L3', None)
if L3:
L3 = int(L3)
L4 = vars.get('L4', None)
if L4:
L4 = int(L4)
L5 = vars.get('L5', None)
if L5:
L5 = int(L5)
if street:
Lx_ids = []
append = Lx_ids.append
for id in (L0, L1, L2, L3, L4, L5):
if id:
append(id)
gis.google_geocode_retry = False
results = gis.geocode(street, postcode, Lx_ids)
else:
results = 'NotImplementedError'
results = json.dumps(results, separators=SEPARATORS)
response.headers['Content-Type'] = 'application/json'
return results
| [
"def",
"geocode",
"(",
")",
":",
"vars",
"=",
"request",
".",
"post_vars",
"street",
"=",
"vars",
".",
"get",
"(",
"'address'",
",",
"None",
")",
"postcode",
"=",
"vars",
".",
"get",
"(",
"'postcode'",
",",
"None",
")",
"L0",
"=",
"vars",
".",
"get... | geocoding is the process of converting addresses into geographic coordinates . | train | false |
4,390 | def make_alert(warnfile, msg_type, msg_template, timestamp_format=None):
def t_format():
return int(time.time())
if (timestamp_format is None):
timestamp_format = t_format
def alert(*params):
formatted_msg = ((msg_type + ' DCTB ') + (msg_template % params))
timestamped_msg = prepend_timestamp(formatted_msg, timestamp_format)
print >>warnfile, timestamped_msg
return alert
| [
"def",
"make_alert",
"(",
"warnfile",
",",
"msg_type",
",",
"msg_template",
",",
"timestamp_format",
"=",
"None",
")",
":",
"def",
"t_format",
"(",
")",
":",
"return",
"int",
"(",
"time",
".",
"time",
"(",
")",
")",
"if",
"(",
"timestamp_format",
"is",
... | create an alert generation function that writes to warnfile . | train | false |
4,391 | def validate_labels(db_session, added_categories, removed_categories):
add = {c.name for c in added_categories if c.name}
add_all = ('all' in add)
add_trash = ('trash' in add)
add_spam = ('spam' in add)
if ((add_all and (add_trash or add_spam)) or (add_trash and add_spam)):
raise InputError('Only one of "all", "trash" or "spam" can be added')
remove = {c.name for c in removed_categories if c.name}
remove_all = ('all' in remove)
remove_trash = ('trash' in remove)
remove_spam = ('spam' in remove)
if (remove_all and remove_trash and remove_spam):
raise InputError('"all", "trash" and "spam" cannot all be removed')
| [
"def",
"validate_labels",
"(",
"db_session",
",",
"added_categories",
",",
"removed_categories",
")",
":",
"add",
"=",
"{",
"c",
".",
"name",
"for",
"c",
"in",
"added_categories",
"if",
"c",
".",
"name",
"}",
"add_all",
"=",
"(",
"'all'",
"in",
"add",
")... | validate that the labels added and removed obey gmails semantics -- gmail messages must belong to exactly one of the [gmail]all mail . | train | false |
4,397 | def removeElementFromDictionary(dictionary, key):
if (key in dictionary):
del dictionary[key]
| [
"def",
"removeElementFromDictionary",
"(",
"dictionary",
",",
"key",
")",
":",
"if",
"(",
"key",
"in",
"dictionary",
")",
":",
"del",
"dictionary",
"[",
"key",
"]"
] | remove element from the dictionary . | train | false |
4,398 | def has_progress(toppath):
state = progress_read()
return (toppath in state)
| [
"def",
"has_progress",
"(",
"toppath",
")",
":",
"state",
"=",
"progress_read",
"(",
")",
"return",
"(",
"toppath",
"in",
"state",
")"
] | return true if there exist paths that have already been imported under toppath . | train | false |
4,399 | @ownership_required
def page_create(request, slug, template_name='groups/pages/page_form.html'):
group = get_object_or_404(Group, slug=slug)
form = GroupPageForm(initial={'group': group})
if (request.method == 'POST'):
form = GroupPageForm(request.POST)
if form.is_valid():
page = form.save(commit=False)
page.group = group
page.save()
return redirect(request, page)
return render(request, template_name, {'group': group, 'form': form})
| [
"@",
"ownership_required",
"def",
"page_create",
"(",
"request",
",",
"slug",
",",
"template_name",
"=",
"'groups/pages/page_form.html'",
")",
":",
"group",
"=",
"get_object_or_404",
"(",
"Group",
",",
"slug",
"=",
"slug",
")",
"form",
"=",
"GroupPageForm",
"(",... | creates a group page . | train | false |
4,400 | def extract_backgrounds(archive_name):
os.mkdir('bgs')
t = tarfile.open(name=archive_name)
def members():
m = t.next()
while m:
(yield m)
m = t.next()
index = 0
for m in members():
if (not m.name.endswith('.jpg')):
continue
f = t.extractfile(m)
try:
im = im_from_file(f)
finally:
f.close()
if (im is None):
continue
if (im.shape[0] > im.shape[1]):
im = im[:im.shape[1], :]
else:
im = im[:, :im.shape[0]]
if (im.shape[0] > 256):
im = cv2.resize(im, (256, 256))
fname = 'bgs/{:08}.jpg'.format(index)
print fname
rc = cv2.imwrite(fname, im)
if (not rc):
raise Exception('Failed to write file {}'.format(fname))
index += 1
| [
"def",
"extract_backgrounds",
"(",
"archive_name",
")",
":",
"os",
".",
"mkdir",
"(",
"'bgs'",
")",
"t",
"=",
"tarfile",
".",
"open",
"(",
"name",
"=",
"archive_name",
")",
"def",
"members",
"(",
")",
":",
"m",
"=",
"t",
".",
"next",
"(",
")",
"whi... | extract backgrounds from provided tar archive . | train | false |
4,402 | @register.simple_tag
def get_advertisement_html_mail():
advertisement = Advertisement.objects.get_advertisement(Advertisement.PLACEMENT_MAIL_HTML)
if (advertisement is None):
return u''
return mark_safe(advertisement.text)
| [
"@",
"register",
".",
"simple_tag",
"def",
"get_advertisement_html_mail",
"(",
")",
":",
"advertisement",
"=",
"Advertisement",
".",
"objects",
".",
"get_advertisement",
"(",
"Advertisement",
".",
"PLACEMENT_MAIL_HTML",
")",
"if",
"(",
"advertisement",
"is",
"None",... | returns advertisement text . | train | false |
4,403 | def test_ast_good_try():
can_compile(u'(try)')
can_compile(u'(try 1)')
can_compile(u'(try 1 (except) (else 1))')
can_compile(u'(try 1 (else 1) (except))')
can_compile(u'(try 1 (finally 1) (except))')
can_compile(u'(try 1 (finally 1))')
can_compile(u'(try 1 (except) (finally 1))')
can_compile(u'(try 1 (except) (finally 1) (else 1))')
can_compile(u'(try 1 (except) (else 1) (finally 1))')
| [
"def",
"test_ast_good_try",
"(",
")",
":",
"can_compile",
"(",
"u'(try)'",
")",
"can_compile",
"(",
"u'(try 1)'",
")",
"can_compile",
"(",
"u'(try 1 (except) (else 1))'",
")",
"can_compile",
"(",
"u'(try 1 (else 1) (except))'",
")",
"can_compile",
"(",
"u'(try 1 (finall... | make sure ast can compile valid try . | train | false |
4,404 | def test_vstack_bytes():
t = table.Table([['a']], names=['a'])
assert (t['a'].itemsize == 1)
t2 = table.vstack([t, t])
assert (len(t2) == 2)
assert (t2['a'].itemsize == 1)
| [
"def",
"test_vstack_bytes",
"(",
")",
":",
"t",
"=",
"table",
".",
"Table",
"(",
"[",
"[",
"'a'",
"]",
"]",
",",
"names",
"=",
"[",
"'a'",
"]",
")",
"assert",
"(",
"t",
"[",
"'a'",
"]",
".",
"itemsize",
"==",
"1",
")",
"t2",
"=",
"table",
"."... | test for issue #5617 when vstacking bytes columns in py3 . | train | false |
4,405 | def _parse_list_rule(rule):
if (not rule):
return TrueCheck()
or_list = []
for inner_rule in rule:
if (not inner_rule):
continue
if isinstance(inner_rule, basestring):
inner_rule = [inner_rule]
and_list = [_parse_check(r) for r in inner_rule]
if (len(and_list) == 1):
or_list.append(and_list[0])
else:
or_list.append(AndCheck(and_list))
if (not or_list):
return FalseCheck()
elif (len(or_list) == 1):
return or_list[0]
return OrCheck(or_list)
| [
"def",
"_parse_list_rule",
"(",
"rule",
")",
":",
"if",
"(",
"not",
"rule",
")",
":",
"return",
"TrueCheck",
"(",
")",
"or_list",
"=",
"[",
"]",
"for",
"inner_rule",
"in",
"rule",
":",
"if",
"(",
"not",
"inner_rule",
")",
":",
"continue",
"if",
"isin... | provided for backwards compatibility . | train | false |
4,406 | def E_n(n, omega):
return ((hbar * omega) * (n + Rational(1, 2)))
| [
"def",
"E_n",
"(",
"n",
",",
"omega",
")",
":",
"return",
"(",
"(",
"hbar",
"*",
"omega",
")",
"*",
"(",
"n",
"+",
"Rational",
"(",
"1",
",",
"2",
")",
")",
")"
] | returns the energy psi_{n} for a 1d potential hole with infinity borders n the "principal" quantum number . | train | false |
4,407 | def randstr(length, alphabet='abcdefghijklmnopqrstuvwxyz0123456789'):
return ''.join((random.choice(alphabet) for _ in xrange(length)))
| [
"def",
"randstr",
"(",
"length",
",",
"alphabet",
"=",
"'abcdefghijklmnopqrstuvwxyz0123456789'",
")",
":",
"return",
"''",
".",
"join",
"(",
"(",
"random",
".",
"choice",
"(",
"alphabet",
")",
"for",
"_",
"in",
"xrange",
"(",
"length",
")",
")",
")"
] | return a string made up of random chars from alphabet . | train | false |
4,408 | def test_add_events():
raw = read_raw_fif(raw_fname)
events = np.array([[raw.first_samp, 0, 1]])
assert_raises(RuntimeError, raw.add_events, events, 'STI 014')
raw = read_raw_fif(raw_fname, preload=True)
orig_events = find_events(raw, 'STI 014')
events = np.array([raw.first_samp, 0, 1])
assert_raises(ValueError, raw.add_events, events, 'STI 014')
events[0] = ((raw.first_samp + raw.n_times) + 1)
events = events[np.newaxis, :]
assert_raises(ValueError, raw.add_events, events, 'STI 014')
events[(0, 0)] = (raw.first_samp - 1)
assert_raises(ValueError, raw.add_events, events, 'STI 014')
events[(0, 0)] = (raw.first_samp + 1)
assert_raises(ValueError, raw.add_events, events, 'STI FOO')
raw.add_events(events, 'STI 014')
new_events = find_events(raw, 'STI 014')
assert_array_equal(new_events, np.concatenate((events, orig_events)))
| [
"def",
"test_add_events",
"(",
")",
":",
"raw",
"=",
"read_raw_fif",
"(",
"raw_fname",
")",
"events",
"=",
"np",
".",
"array",
"(",
"[",
"[",
"raw",
".",
"first_samp",
",",
"0",
",",
"1",
"]",
"]",
")",
"assert_raises",
"(",
"RuntimeError",
",",
"raw... | test adding events to a raw file . | train | false |
4,409 | def read_float(fid):
return _unpack_simple(fid, '>f4', np.float32)
| [
"def",
"read_float",
"(",
"fid",
")",
":",
"return",
"_unpack_simple",
"(",
"fid",
",",
"'>f4'",
",",
"np",
".",
"float32",
")"
] | read 32bit float from bti file . | train | false |
4,410 | def hprModelSynth(hfreq, hmag, hphase, xr, N, H, fs):
yh = SM.sineModelSynth(hfreq, hmag, hphase, N, H, fs)
y = (yh[:min(yh.size, xr.size)] + xr[:min(yh.size, xr.size)])
return (y, yh)
| [
"def",
"hprModelSynth",
"(",
"hfreq",
",",
"hmag",
",",
"hphase",
",",
"xr",
",",
"N",
",",
"H",
",",
"fs",
")",
":",
"yh",
"=",
"SM",
".",
"sineModelSynth",
"(",
"hfreq",
",",
"hmag",
",",
"hphase",
",",
"N",
",",
"H",
",",
"fs",
")",
"y",
"... | synthesis of a sound using the sinusoidal plus residual model tfreq . | train | false |
4,412 | def absolute_path_link(path):
if os.path.islink(path):
link = os.readlink(path)
if (not os.path.isabs(link)):
link = os.path.join(os.path.dirname(path), link)
else:
link = os.path.abspath(path)
return link
| [
"def",
"absolute_path_link",
"(",
"path",
")",
":",
"if",
"os",
".",
"path",
".",
"islink",
"(",
"path",
")",
":",
"link",
"=",
"os",
".",
"readlink",
"(",
"path",
")",
"if",
"(",
"not",
"os",
".",
"path",
".",
"isabs",
"(",
"link",
")",
")",
"... | returns an absolute path for the destination of a symlink . | train | false |
4,413 | @require_context
def device_array(shape, dtype=np.float, strides=None, order='C', stream=0):
(shape, strides, dtype) = _prepare_shape_strides_dtype(shape, strides, dtype, order)
return devicearray.DeviceNDArray(shape=shape, strides=strides, dtype=dtype, stream=stream)
| [
"@",
"require_context",
"def",
"device_array",
"(",
"shape",
",",
"dtype",
"=",
"np",
".",
"float",
",",
"strides",
"=",
"None",
",",
"order",
"=",
"'C'",
",",
"stream",
"=",
"0",
")",
":",
"(",
"shape",
",",
"strides",
",",
"dtype",
")",
"=",
"_pr... | device_array allocate an empty device ndarray . | train | false |
4,414 | @command('history')
def view_history(duplicates=True):
history = g.userhist.get('history')
try:
hist_list = list(reversed(history.songs))
message = 'Viewing play history'
if (not duplicates):
seen = set()
seen_add = seen.add
hist_list = [x for x in hist_list if (not ((x in seen) or seen_add(x)))]
message = 'Viewing recent played songs'
paginatesongs(hist_list)
g.message = message
except AttributeError:
g.content = logo(c.r)
g.message = 'History empty'
| [
"@",
"command",
"(",
"'history'",
")",
"def",
"view_history",
"(",
"duplicates",
"=",
"True",
")",
":",
"history",
"=",
"g",
".",
"userhist",
".",
"get",
"(",
"'history'",
")",
"try",
":",
"hist_list",
"=",
"list",
"(",
"reversed",
"(",
"history",
".",... | display the users play history . | train | false |
4,415 | def detect_text(path):
vision_client = vision.Client()
with io.open(path, 'rb') as image_file:
content = image_file.read()
image = vision_client.image(content=content)
texts = image.detect_text()
print 'Texts:'
for text in texts:
print text.description
| [
"def",
"detect_text",
"(",
"path",
")",
":",
"vision_client",
"=",
"vision",
".",
"Client",
"(",
")",
"with",
"io",
".",
"open",
"(",
"path",
",",
"'rb'",
")",
"as",
"image_file",
":",
"content",
"=",
"image_file",
".",
"read",
"(",
")",
"image",
"="... | detects text in the file . | train | false |
4,417 | def places():
from datetime import datetime, timedelta
response.headers['Expires'] = (datetime.now() + timedelta(days=7)).strftime('%a, %d %b %Y %H:%M:%S GMT')
return response.stream(open(_map_plugin().place_data(), 'rb'), chunk_size=4096)
| [
"def",
"places",
"(",
")",
":",
"from",
"datetime",
"import",
"datetime",
",",
"timedelta",
"response",
".",
"headers",
"[",
"'Expires'",
"]",
"=",
"(",
"datetime",
".",
"now",
"(",
")",
"+",
"timedelta",
"(",
"days",
"=",
"7",
")",
")",
".",
"strfti... | places search . | train | false |
4,419 | def _traverse_results(value, fields, row, path):
for (f, v) in value.iteritems():
field_name = ('{path}.{name}'.format(path=path, name=f) if path else f)
if (not isinstance(v, (dict, list, tuple))):
if (field_name in fields):
row[fields.index(field_name)] = ensure_utf(v)
elif (isinstance(v, dict) and (f != 'attributes')):
_traverse_results(v, fields, row, field_name)
| [
"def",
"_traverse_results",
"(",
"value",
",",
"fields",
",",
"row",
",",
"path",
")",
":",
"for",
"(",
"f",
",",
"v",
")",
"in",
"value",
".",
"iteritems",
"(",
")",
":",
"field_name",
"=",
"(",
"'{path}.{name}'",
".",
"format",
"(",
"path",
"=",
... | helper method for parse_results() . | train | true |
4,420 | def find_empty_redis_database():
for dbnum in range(4, 17):
testconn = StrictRedis(db=dbnum)
empty = (len(testconn.keys(u'*')) == 0)
if empty:
return testconn
assert False, u'No empty Redis database found to run tests in.'
| [
"def",
"find_empty_redis_database",
"(",
")",
":",
"for",
"dbnum",
"in",
"range",
"(",
"4",
",",
"17",
")",
":",
"testconn",
"=",
"StrictRedis",
"(",
"db",
"=",
"dbnum",
")",
"empty",
"=",
"(",
"len",
"(",
"testconn",
".",
"keys",
"(",
"u'*'",
")",
... | tries to connect to a random redis database . | train | false |
4,421 | def XRI(xri):
if (not xri.startswith('xri://')):
xri = ('xri://' + xri)
return xri
| [
"def",
"XRI",
"(",
"xri",
")",
":",
"if",
"(",
"not",
"xri",
".",
"startswith",
"(",
"'xri://'",
")",
")",
":",
"xri",
"=",
"(",
"'xri://'",
"+",
"xri",
")",
"return",
"xri"
] | an xri object allowing comparison of xri . | train | false |
4,422 | @pytest.fixture
def english_tutorial(english, tutorial):
return _require_tp(english, tutorial)
| [
"@",
"pytest",
".",
"fixture",
"def",
"english_tutorial",
"(",
"english",
",",
"tutorial",
")",
":",
"return",
"_require_tp",
"(",
"english",
",",
"tutorial",
")"
] | require english tutorial . | train | false |
4,425 | def roundup(number, ndigits=0, return_type=None):
sign = (1 if (number >= 0) else (-1))
precision = (10 ** ((-1) * ndigits))
if (not return_type):
return_type = (float if (ndigits > 0) else int)
(quotient, remainder) = divmod(abs(number), precision)
if ((not (IRONPYTHON and (((quotient * precision) + remainder) > abs(number)))) and (remainder >= (precision / 2))):
quotient += 1
return (sign * return_type((quotient * precision)))
| [
"def",
"roundup",
"(",
"number",
",",
"ndigits",
"=",
"0",
",",
"return_type",
"=",
"None",
")",
":",
"sign",
"=",
"(",
"1",
"if",
"(",
"number",
">=",
"0",
")",
"else",
"(",
"-",
"1",
")",
")",
"precision",
"=",
"(",
"10",
"**",
"(",
"(",
"-... | rounds number to the given number of digits . | train | false |
4,427 | def default_resolver():
if DNSPYTHON_AVAILABLE:
return dns.resolver.get_default_resolver()
return None
| [
"def",
"default_resolver",
"(",
")",
":",
"if",
"DNSPYTHON_AVAILABLE",
":",
"return",
"dns",
".",
"resolver",
".",
"get_default_resolver",
"(",
")",
"return",
"None"
] | return a basic dns resolver object . | train | false |
4,428 | def test_disconnect_one_invalid(timer):
func1 = mock.Mock()
func2 = mock.Mock()
timer.timeout.connect(func1)
with pytest.raises(TypeError):
timer.timeout.disconnect(func2)
assert (not func1.called)
assert (not func2.called)
timer.timeout.emit()
func1.assert_called_once_with()
| [
"def",
"test_disconnect_one_invalid",
"(",
"timer",
")",
":",
"func1",
"=",
"mock",
".",
"Mock",
"(",
")",
"func2",
"=",
"mock",
".",
"Mock",
"(",
")",
"timer",
".",
"timeout",
".",
"connect",
"(",
"func1",
")",
"with",
"pytest",
".",
"raises",
"(",
... | test disconnecting with an invalid connection . | train | false |
4,429 | def _modify_tagids(source, add=True):
output = []
tagcount = 0
if (not isinstance(source, HtmlPage)):
source = HtmlPage(body=source)
for element in source.parsed_body:
if _must_add_tagid(element):
if add:
element.attributes[TAGID] = str(tagcount)
tagcount += 1
else:
element.attributes.pop(TAGID, None)
output.append(serialize_tag(element))
else:
output.append(source.body[element.start:element.end])
return u''.join(output)
| [
"def",
"_modify_tagids",
"(",
"source",
",",
"add",
"=",
"True",
")",
":",
"output",
"=",
"[",
"]",
"tagcount",
"=",
"0",
"if",
"(",
"not",
"isinstance",
"(",
"source",
",",
"HtmlPage",
")",
")",
":",
"source",
"=",
"HtmlPage",
"(",
"body",
"=",
"s... | add or remove tags ids to/from html document . | train | false |
4,430 | def imagej_description_dict(description):
def _bool(val):
return {'true': True, 'false': False}[val.lower()]
_str = (str if (sys.version_info[0] < 3) else (lambda x: str(x, 'cp1252')))
result = {}
for line in description.splitlines():
try:
(key, val) = line.split('=')
except Exception:
continue
key = key.strip()
val = val.strip()
for dtype in (int, float, _bool, _str):
try:
val = dtype(val)
break
except Exception:
pass
result[_str(key)] = val
if ('ImageJ' not in result):
raise ValueError('not a ImageJ image description')
return result
| [
"def",
"imagej_description_dict",
"(",
"description",
")",
":",
"def",
"_bool",
"(",
"val",
")",
":",
"return",
"{",
"'true'",
":",
"True",
",",
"'false'",
":",
"False",
"}",
"[",
"val",
".",
"lower",
"(",
")",
"]",
"_str",
"=",
"(",
"str",
"if",
"... | return dictionary from imagej image description byte string . | train | false |
4,431 | def get_doctype_module(doctype):
def make_modules_dict():
return dict(frappe.db.sql(u'select name, module from tabDocType'))
return frappe.cache().get_value(u'doctype_modules', make_modules_dict)[doctype]
| [
"def",
"get_doctype_module",
"(",
"doctype",
")",
":",
"def",
"make_modules_dict",
"(",
")",
":",
"return",
"dict",
"(",
"frappe",
".",
"db",
".",
"sql",
"(",
"u'select name, module from tabDocType'",
")",
")",
"return",
"frappe",
".",
"cache",
"(",
")",
"."... | returns **module def** name of given doctype . | train | false |
4,432 | def ndarray_device_allocate_data(ary):
datasize = driver.host_memory_size(ary)
gpu_data = devices.get_context().memalloc(datasize)
return gpu_data
| [
"def",
"ndarray_device_allocate_data",
"(",
"ary",
")",
":",
"datasize",
"=",
"driver",
".",
"host_memory_size",
"(",
"ary",
")",
"gpu_data",
"=",
"devices",
".",
"get_context",
"(",
")",
".",
"memalloc",
"(",
"datasize",
")",
"return",
"gpu_data"
] | allocate gpu data buffer . | train | false |
4,433 | def remove_if(predicate, lst):
return [elem for elem in lst if (not predicate(elem))]
| [
"def",
"remove_if",
"(",
"predicate",
",",
"lst",
")",
":",
"return",
"[",
"elem",
"for",
"elem",
"in",
"lst",
"if",
"(",
"not",
"predicate",
"(",
"elem",
")",
")",
"]"
] | returns a new list with elements of the iterable lst excepting those satisfying predicate . | train | false |
4,436 | def name_yaml_formatter(table_dict):
table_string = 'name: {\n'
namerecord_list = table_dict['names']
for record in namerecord_list:
if (record.__dict__['langID'] == 0):
record_name = str(record.__dict__['nameID'])
else:
record_name = (str(record.__dict__['nameID']) + 'u')
record_field = (((' ' * 4) + 'nameID') + record_name)
table_string = ((((table_string + record_field) + ': ') + str(record.__dict__)) + ',\n')
table_string = (table_string + '}\n\n')
return table_string
| [
"def",
"name_yaml_formatter",
"(",
"table_dict",
")",
":",
"table_string",
"=",
"'name: {\\n'",
"namerecord_list",
"=",
"table_dict",
"[",
"'names'",
"]",
"for",
"record",
"in",
"namerecord_list",
":",
"if",
"(",
"record",
".",
"__dict__",
"[",
"'langID'",
"]",
... | formats the yaml table string for opentype name tables . | train | false |
4,437 | def download_pricing_file(file_url=DEFAULT_FILE_URL, file_path=CUSTOM_PRICING_FILE_PATH):
dir_name = os.path.dirname(file_path)
if (not os.path.exists(dir_name)):
msg = ("Can't write to %s, directory %s, doesn't exist" % (file_path, dir_name))
raise ValueError(msg)
if (os.path.exists(file_path) and os.path.isdir(file_path)):
msg = ("Can't write to %s file path because it's a directory" % file_path)
raise ValueError(msg)
response = get_response_object(file_url)
body = response.body
try:
data = json.loads(body)
except JSONDecodeError:
msg = "Provided URL doesn't contain valid pricing data"
raise Exception(msg)
if (not data.get('updated', None)):
msg = "Provided URL doesn't contain valid pricing data"
raise Exception(msg)
with open(file_path, 'w') as file_handle:
file_handle.write(body)
| [
"def",
"download_pricing_file",
"(",
"file_url",
"=",
"DEFAULT_FILE_URL",
",",
"file_path",
"=",
"CUSTOM_PRICING_FILE_PATH",
")",
":",
"dir_name",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"file_path",
")",
"if",
"(",
"not",
"os",
".",
"path",
".",
"exist... | download pricing file from the file_url and save it to file_path . | train | false |
4,438 | def netapi(opts):
return LazyLoader(_module_dirs(opts, 'netapi'), opts, tag='netapi')
| [
"def",
"netapi",
"(",
"opts",
")",
":",
"return",
"LazyLoader",
"(",
"_module_dirs",
"(",
"opts",
",",
"'netapi'",
")",
",",
"opts",
",",
"tag",
"=",
"'netapi'",
")"
] | return the network api functions . | train | false |
4,439 | def project_time_day(row):
try:
thisdate = row['project_time.date']
except AttributeError:
return current.messages['NONE']
if (not thisdate):
return current.messages['NONE']
now = current.request.utcnow
week = datetime.timedelta(days=7)
return thisdate.date().strftime('%d %B %y')
| [
"def",
"project_time_day",
"(",
"row",
")",
":",
"try",
":",
"thisdate",
"=",
"row",
"[",
"'project_time.date'",
"]",
"except",
"AttributeError",
":",
"return",
"current",
".",
"messages",
"[",
"'NONE'",
"]",
"if",
"(",
"not",
"thisdate",
")",
":",
"return... | virtual field for project_time - abbreviated string format for date . | train | false |
4,442 | def cov_crosssection_0(results, group):
scale = S_crosssection(results.resid[:, None], group)
scale = np.squeeze(scale)
cov = _HCCM1(results, scale)
return cov
| [
"def",
"cov_crosssection_0",
"(",
"results",
",",
"group",
")",
":",
"scale",
"=",
"S_crosssection",
"(",
"results",
".",
"resid",
"[",
":",
",",
"None",
"]",
",",
"group",
")",
"scale",
"=",
"np",
".",
"squeeze",
"(",
"scale",
")",
"cov",
"=",
"_HCC... | this one is still wrong . | train | false |
4,443 | def testing_on(name, reload=False):
ret = {'name': 'testing mode', 'changes': {}, 'result': True, 'comment': 'Testing mode already ON.'}
result = {}
testing = __salt__['csf.get_testing_status']()
if (int(testing) == 1):
return ret
enable = __salt__['csf.enable_testing_mode']()
if enable:
comment = 'Csf testing mode enabled'
if reload:
if __salt__['csf.reload']():
comment += ' and csf reloaded.'
ret['changes']['Testing Mode'] = 'on'
ret['comment'] = result
return ret
| [
"def",
"testing_on",
"(",
"name",
",",
"reload",
"=",
"False",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"'testing mode'",
",",
"'changes'",
":",
"{",
"}",
",",
"'result'",
":",
"True",
",",
"'comment'",
":",
"'Testing mode already ON.'",
"}",
"result",
... | ensure testing mode is enabled in csf . | train | true |
4,444 | def update_user_contributions(user_id, created_exploration_ids, edited_exploration_ids):
user_contributions = get_user_contributions(user_id, strict=False)
if (not user_contributions):
raise Exception(('User contributions model for user %s does not exist.' % user_id))
user_contributions.created_exploration_ids = created_exploration_ids
user_contributions.edited_exploration_ids = edited_exploration_ids
_save_user_contributions(user_contributions)
| [
"def",
"update_user_contributions",
"(",
"user_id",
",",
"created_exploration_ids",
",",
"edited_exploration_ids",
")",
":",
"user_contributions",
"=",
"get_user_contributions",
"(",
"user_id",
",",
"strict",
"=",
"False",
")",
"if",
"(",
"not",
"user_contributions",
... | updates an existing usercontributionsmodel with new calculated contributions . | train | false |
4,445 | def extract_description(texts):
document = ''
for text in texts:
try:
document += text['description']
except KeyError as e:
print ('KeyError: %s\n%s' % (e, text))
return document
| [
"def",
"extract_description",
"(",
"texts",
")",
":",
"document",
"=",
"''",
"for",
"text",
"in",
"texts",
":",
"try",
":",
"document",
"+=",
"text",
"[",
"'description'",
"]",
"except",
"KeyError",
"as",
"e",
":",
"print",
"(",
"'KeyError: %s\\n%s'",
"%",... | returns text annotations as a single string . | train | false |
4,446 | def get_os_args():
if (PY2 and WIN and (_initial_argv_hash == _hash_py_argv())):
return _get_windows_argv()
return sys.argv[1:]
| [
"def",
"get_os_args",
"(",
")",
":",
"if",
"(",
"PY2",
"and",
"WIN",
"and",
"(",
"_initial_argv_hash",
"==",
"_hash_py_argv",
"(",
")",
")",
")",
":",
"return",
"_get_windows_argv",
"(",
")",
"return",
"sys",
".",
"argv",
"[",
"1",
":",
"]"
] | this returns the argument part of sys . | train | false |
4,447 | def _maybe_convert_scalar(values):
if is_scalar(values):
(dtype, values) = _infer_dtype_from_scalar(values)
try:
values = dtype(values)
except TypeError:
pass
return values
| [
"def",
"_maybe_convert_scalar",
"(",
"values",
")",
":",
"if",
"is_scalar",
"(",
"values",
")",
":",
"(",
"dtype",
",",
"values",
")",
"=",
"_infer_dtype_from_scalar",
"(",
"values",
")",
"try",
":",
"values",
"=",
"dtype",
"(",
"values",
")",
"except",
... | convert a python scalar to the appropriate numpy dtype if possible this avoids numpy directly converting according to platform preferences . | train | false |
4,448 | def tanh_cutoff(x, cutoff):
y = tf.tanh(x)
if (cutoff < 1.01):
return y
d = ((cutoff - 1.0) / 2.0)
return tf.minimum(1.0, tf.maximum((-1.0), ((1.0 + d) * y)))
| [
"def",
"tanh_cutoff",
"(",
"x",
",",
"cutoff",
")",
":",
"y",
"=",
"tf",
".",
"tanh",
"(",
"x",
")",
"if",
"(",
"cutoff",
"<",
"1.01",
")",
":",
"return",
"y",
"d",
"=",
"(",
"(",
"cutoff",
"-",
"1.0",
")",
"/",
"2.0",
")",
"return",
"tf",
... | tanh with cutoff . | train | false |
4,449 | def jobs():
job_names = _jobs().keys()
job_names.sort()
return {'jobs': job_names}
| [
"def",
"jobs",
"(",
")",
":",
"job_names",
"=",
"_jobs",
"(",
")",
".",
"keys",
"(",
")",
"job_names",
".",
"sort",
"(",
")",
"return",
"{",
"'jobs'",
":",
"job_names",
"}"
] | return a list of the currently installed job names . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.