id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1
value | is_duplicated bool 2
classes |
|---|---|---|---|---|---|
8,892 | def create_files_from_macho_zip(fileobj, project=None):
scratchpad = tempfile.mkdtemp()
try:
safe_extract_zip(fileobj, scratchpad)
to_create = []
for (dirpath, dirnames, filenames) in os.walk(scratchpad):
for fn in filenames:
fn = os.path.join(dirpath, fn)
try:
uuids = get_macho_uuids(fn)
ex... | [
"def",
"create_files_from_macho_zip",
"(",
"fileobj",
",",
"project",
"=",
"None",
")",
":",
"scratchpad",
"=",
"tempfile",
".",
"mkdtemp",
"(",
")",
"try",
":",
"safe_extract_zip",
"(",
"fileobj",
",",
"scratchpad",
")",
"to_create",
"=",
"[",
"]",
"for",
... | creates all missing dsym files from the given zip file . | train | false |
8,893 | def composite(image1, image2, mask):
return Image.composite(image1, image2, mask)
| [
"def",
"composite",
"(",
"image1",
",",
"image2",
",",
"mask",
")",
":",
"return",
"Image",
".",
"composite",
"(",
"image1",
",",
"image2",
",",
"mask",
")"
] | create composite image by blending images using a transparency mask . | train | false |
8,894 | def fetch_val_for_key(key, delete_key=False):
try:
if delete_key:
return THEANO_FLAGS_DICT.pop(key)
return THEANO_FLAGS_DICT[key]
except KeyError:
pass
key_tokens = key.rsplit('.', 1)
if (len(key_tokens) > 2):
raise KeyError(key)
if (len(key_tokens) == 2):
(section, option) = key_tokens
else:
(sect... | [
"def",
"fetch_val_for_key",
"(",
"key",
",",
"delete_key",
"=",
"False",
")",
":",
"try",
":",
"if",
"delete_key",
":",
"return",
"THEANO_FLAGS_DICT",
".",
"pop",
"(",
"key",
")",
"return",
"THEANO_FLAGS_DICT",
"[",
"key",
"]",
"except",
"KeyError",
":",
"... | return the overriding config value for a key . | train | false |
8,895 | def checkFieldName(field_name):
if (field_name not in data_fields):
raise ValueError(('%r is not a defined simple registration field' % (field_name,)))
| [
"def",
"checkFieldName",
"(",
"field_name",
")",
":",
"if",
"(",
"field_name",
"not",
"in",
"data_fields",
")",
":",
"raise",
"ValueError",
"(",
"(",
"'%r is not a defined simple registration field'",
"%",
"(",
"field_name",
",",
")",
")",
")"
] | check to see that the given value is a valid simple registration data field name . | train | false |
8,896 | def contact_autocreate_handler(sender, instance, created, **kwargs):
if created:
try:
contact_type = ContactType.objects.filter((models.Q(name='Person') | models.Q(slug='person')))[0]
contact = Contact(contact_type=contact_type, name=instance.name, related_user=instance)
contact.save()
except:
pass
| [
"def",
"contact_autocreate_handler",
"(",
"sender",
",",
"instance",
",",
"created",
",",
"**",
"kwargs",
")",
":",
"if",
"created",
":",
"try",
":",
"contact_type",
"=",
"ContactType",
".",
"objects",
".",
"filter",
"(",
"(",
"models",
".",
"Q",
"(",
"n... | when a user is created . | train | false |
8,897 | def _prepare_pre_auth_info_request(env, path, swift_source):
newenv = make_pre_authed_env(env, 'HEAD', path, agent='Swift', query_string='', swift_source=swift_source)
newenv.pop('HTTP_ORIGIN', None)
newenv['swift_owner'] = True
return Request.blank(quote(path), environ=newenv)
| [
"def",
"_prepare_pre_auth_info_request",
"(",
"env",
",",
"path",
",",
"swift_source",
")",
":",
"newenv",
"=",
"make_pre_authed_env",
"(",
"env",
",",
"'HEAD'",
",",
"path",
",",
"agent",
"=",
"'Swift'",
",",
"query_string",
"=",
"''",
",",
"swift_source",
... | prepares a pre authed request to obtain info using a head . | train | false |
8,898 | def fetch_hdu(filename, cache=True):
for retry in range(MAX_RETRIES):
try:
path = download_file((URL + filename), cache=cache, timeout=30)
except:
if (retry == (MAX_RETRIES - 1)):
raise
else:
time.sleep(TIME_BETWEEN_RETRIES)
else:
break
else:
raise Exception('Failed to download file {0}'.f... | [
"def",
"fetch_hdu",
"(",
"filename",
",",
"cache",
"=",
"True",
")",
":",
"for",
"retry",
"in",
"range",
"(",
"MAX_RETRIES",
")",
":",
"try",
":",
"path",
"=",
"download_file",
"(",
"(",
"URL",
"+",
"filename",
")",
",",
"cache",
"=",
"cache",
",",
... | download a fits file to the cache and open hdu 0 . | train | false |
8,899 | def stftSynth(mY, pY, M, H):
hM1 = int(math.floor(((M + 1) / 2)))
hM2 = int(math.floor((M / 2)))
nFrames = mY[:, 0].size
y = np.zeros((((nFrames * H) + hM1) + hM2))
pin = hM1
for i in range(nFrames):
y1 = DFT.dftSynth(mY[i, :], pY[i, :], M)
y[(pin - hM1):(pin + hM2)] += (H * y1)
pin += H
y = np.delete(y, r... | [
"def",
"stftSynth",
"(",
"mY",
",",
"pY",
",",
"M",
",",
"H",
")",
":",
"hM1",
"=",
"int",
"(",
"math",
".",
"floor",
"(",
"(",
"(",
"M",
"+",
"1",
")",
"/",
"2",
")",
")",
")",
"hM2",
"=",
"int",
"(",
"math",
".",
"floor",
"(",
"(",
"M... | synthesis of a sound using the short-time fourier transform my: magnitude spectra . | train | false |
8,900 | def addNameToCache(name, tvdb_id):
name = sanitizeSceneName(name)
if (not tvdb_id):
tvdb_id = 0
cacheDB = db.DBConnection('cache.db')
cacheDB.action('INSERT INTO scene_names (tvdb_id, name) VALUES (?, ?)', [tvdb_id, name])
| [
"def",
"addNameToCache",
"(",
"name",
",",
"tvdb_id",
")",
":",
"name",
"=",
"sanitizeSceneName",
"(",
"name",
")",
"if",
"(",
"not",
"tvdb_id",
")",
":",
"tvdb_id",
"=",
"0",
"cacheDB",
"=",
"db",
".",
"DBConnection",
"(",
"'cache.db'",
")",
"cacheDB",
... | adds the show & tvdb id to the scene_names table in cache . | train | false |
8,901 | def _bgp_capability_dispatcher(payload):
cls = _capabilities_registry['BGPCapGeneric']
if (payload is None):
cls = _capabilities_registry['BGPCapGeneric']
else:
length = len(payload)
if (length >= _BGP_CAPABILITY_MIN_SIZE):
code = struct.unpack('!B', payload[0])[0]
cls = _get_cls(_capabilities_objects.ge... | [
"def",
"_bgp_capability_dispatcher",
"(",
"payload",
")",
":",
"cls",
"=",
"_capabilities_registry",
"[",
"'BGPCapGeneric'",
"]",
"if",
"(",
"payload",
"is",
"None",
")",
":",
"cls",
"=",
"_capabilities_registry",
"[",
"'BGPCapGeneric'",
"]",
"else",
":",
"lengt... | returns the right class for a given bgp capability . | train | false |
8,902 | @memoize(prefix='get_excluded_in')
def get_excluded_in(region_id):
aers = list(AddonExcludedRegion.objects.filter(region=region_id).values_list('addon', flat=True))
geodata_qs = Q()
region = parse_region(region_id)
if (region in (mkt.regions.BRA, mkt.regions.DEU)):
geodata_qs |= Q(**{('region_%s_iarc_exclude' % r... | [
"@",
"memoize",
"(",
"prefix",
"=",
"'get_excluded_in'",
")",
"def",
"get_excluded_in",
"(",
"region_id",
")",
":",
"aers",
"=",
"list",
"(",
"AddonExcludedRegion",
".",
"objects",
".",
"filter",
"(",
"region",
"=",
"region_id",
")",
".",
"values_list",
"(",... | return ids of webapp objects excluded from a particular region or excluded due to geodata flags . | train | false |
8,903 | def dump_objects():
from meliae import scanner
scanner.dump_all_objects((PROFILING_OUTPUT_FMT % get_filename_fmt()))
| [
"def",
"dump_objects",
"(",
")",
":",
"from",
"meliae",
"import",
"scanner",
"scanner",
".",
"dump_all_objects",
"(",
"(",
"PROFILING_OUTPUT_FMT",
"%",
"get_filename_fmt",
"(",
")",
")",
")"
] | get all registered objects in all registries as a string . | train | false |
8,904 | def set_probes(probes, test=False, commit=True):
return __salt__['net.load_template']('set_probes', probes=probes, test=test, commit=commit)
| [
"def",
"set_probes",
"(",
"probes",
",",
"test",
"=",
"False",
",",
"commit",
"=",
"True",
")",
":",
"return",
"__salt__",
"[",
"'net.load_template'",
"]",
"(",
"'set_probes'",
",",
"probes",
"=",
"probes",
",",
"test",
"=",
"test",
",",
"commit",
"=",
... | configures rpm/sla probes on the device . | train | false |
8,905 | def make_api_project(project_data):
from readthedocs.projects.models import Project
for key in ['users', 'resource_uri', 'absolute_url', 'downloads', 'main_language_project', 'related_projects']:
if (key in project_data):
del project_data[key]
project = Project(**project_data)
project.save = _new_save
return ... | [
"def",
"make_api_project",
"(",
"project_data",
")",
":",
"from",
"readthedocs",
".",
"projects",
".",
"models",
"import",
"Project",
"for",
"key",
"in",
"[",
"'users'",
",",
"'resource_uri'",
",",
"'absolute_url'",
",",
"'downloads'",
",",
"'main_language_project... | make mock project instance from api return . | train | false |
8,906 | def createservicesid(svc):
uni = ''.join([(c + '\x00') for c in svc])
sha = hashlib.sha1(uni.upper()).digest()
dec = list()
for i in range(5):
dec.append(struct.unpack('<I', sha[(i * 4):((i * 4) + 4)])[0])
return ('S-1-5-80-' + '-'.join([str(n) for n in dec]))
| [
"def",
"createservicesid",
"(",
"svc",
")",
":",
"uni",
"=",
"''",
".",
"join",
"(",
"[",
"(",
"c",
"+",
"'\\x00'",
")",
"for",
"c",
"in",
"svc",
"]",
")",
"sha",
"=",
"hashlib",
".",
"sha1",
"(",
"uni",
".",
"upper",
"(",
")",
")",
".",
"dig... | calculate the service sid . | train | false |
8,907 | def ravel(a):
return a.ravel()
| [
"def",
"ravel",
"(",
"a",
")",
":",
"return",
"a",
".",
"ravel",
"(",
")"
] | returns a flattened array . | train | false |
8,908 | @cacheit
def _nT(n, k):
if (k == 0):
return (1 if (k == n) else 0)
return sum((_nT((n - k), j) for j in range((min(k, (n - k)) + 1))))
| [
"@",
"cacheit",
"def",
"_nT",
"(",
"n",
",",
"k",
")",
":",
"if",
"(",
"k",
"==",
"0",
")",
":",
"return",
"(",
"1",
"if",
"(",
"k",
"==",
"n",
")",
"else",
"0",
")",
"return",
"sum",
"(",
"(",
"_nT",
"(",
"(",
"n",
"-",
"k",
")",
",",
... | return the partitions of n items into k parts . | train | false |
8,909 | def uninstalled(name, password, keychain='/Library/Keychains/System.keychain', keychain_password=None):
ret = {'name': name, 'result': True, 'comment': '', 'changes': {}}
certs = __salt__['keychain.list_certs'](keychain)
if ('.p12' in name):
if (('http' in name) or ('salt' in name)):
name = __salt__['cp.cache_f... | [
"def",
"uninstalled",
"(",
"name",
",",
"password",
",",
"keychain",
"=",
"'/Library/Keychains/System.keychain'",
",",
"keychain_password",
"=",
"None",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'result'",
":",
"True",
",",
"'comment'",
":",
"'... | uninstalls a package name the name of the package to be uninstalled version uninstalls a specific version of the package . | train | true |
8,910 | def configure_proxy(proxyname, start=True):
changes_new = []
changes_old = []
status_file = True
test = __opts__['test']
proxyfile = '/etc/salt/proxy'
(status_file, msg_new, msg_old) = _proxy_conf_file(proxyfile, test)
changes_new.extend(msg_new)
changes_old.extend(msg_old)
status_proc = False
if start:
(st... | [
"def",
"configure_proxy",
"(",
"proxyname",
",",
"start",
"=",
"True",
")",
":",
"changes_new",
"=",
"[",
"]",
"changes_old",
"=",
"[",
"]",
"status_file",
"=",
"True",
"test",
"=",
"__opts__",
"[",
"'test'",
"]",
"proxyfile",
"=",
"'/etc/salt/proxy'",
"("... | create the salt proxy file and start the proxy process if required parameters: name: the name of this state proxyname: name to be used for this proxy start: boolean indicating if the process should be started example: . | train | true |
8,913 | def render_common_pairs(profile=None):
if (profile is None):
profile = snapshot_profile()
def seq():
for (_, ops, count) in common_pairs(profile):
(yield ('%s: %s\n' % (count, ops)))
return ''.join(seq())
| [
"def",
"render_common_pairs",
"(",
"profile",
"=",
"None",
")",
":",
"if",
"(",
"profile",
"is",
"None",
")",
":",
"profile",
"=",
"snapshot_profile",
"(",
")",
"def",
"seq",
"(",
")",
":",
"for",
"(",
"_",
",",
"ops",
",",
"count",
")",
"in",
"com... | renders the most common opcode pairs to a string in order of descending frequency . | train | false |
8,914 | def test_unset_text(qtbot):
label = TextBase()
qtbot.add_widget(label)
label.setText('foo')
label.setText('')
assert (not label._elided_text)
| [
"def",
"test_unset_text",
"(",
"qtbot",
")",
":",
"label",
"=",
"TextBase",
"(",
")",
"qtbot",
".",
"add_widget",
"(",
"label",
")",
"label",
".",
"setText",
"(",
"'foo'",
")",
"label",
".",
"setText",
"(",
"''",
")",
"assert",
"(",
"not",
"label",
"... | make sure the text is cleared properly . | train | false |
8,915 | def with_settings(*arg_settings, **kw_settings):
def outer(func):
@wraps(func)
def inner(*args, **kwargs):
with settings(*arg_settings, **kw_settings):
return func(*args, **kwargs)
return _wrap_as_new(func, inner)
return outer
| [
"def",
"with_settings",
"(",
"*",
"arg_settings",
",",
"**",
"kw_settings",
")",
":",
"def",
"outer",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"inner",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"with",
"settings",
"(",
"... | decorator equivalent of fabric . | train | false |
8,916 | def maybe_unwrap_results(results):
return getattr(results, '_results', results)
| [
"def",
"maybe_unwrap_results",
"(",
"results",
")",
":",
"return",
"getattr",
"(",
"results",
",",
"'_results'",
",",
"results",
")"
] | gets raw results back from wrapped results . | train | false |
8,917 | def is_base(text):
return min([(ord(char) in BASE_CHARS) for char in text])
| [
"def",
"is_base",
"(",
"text",
")",
":",
"return",
"min",
"(",
"[",
"(",
"ord",
"(",
"char",
")",
"in",
"BASE_CHARS",
")",
"for",
"char",
"in",
"text",
"]",
")"
] | checks whether text should use cjk fonts . | train | false |
8,918 | def testdoc(*arguments, **options):
TestDoc().execute(*arguments, **options)
| [
"def",
"testdoc",
"(",
"*",
"arguments",
",",
"**",
"options",
")",
":",
"TestDoc",
"(",
")",
".",
"execute",
"(",
"*",
"arguments",
",",
"**",
"options",
")"
] | executes testdoc programmatically . | train | false |
8,919 | def _roll_vectorized(M, roll_indices, axis):
assert (axis in [0, 1])
ndim = M.ndim
assert (ndim == 3)
ndim_roll = roll_indices.ndim
assert (ndim_roll == 1)
sh = M.shape
(r, c) = sh[(-2):]
assert (sh[0] == roll_indices.shape[0])
vec_indices = np.arange(sh[0], dtype=np.int32)
M_roll = np.empty_like(M)
if (axis... | [
"def",
"_roll_vectorized",
"(",
"M",
",",
"roll_indices",
",",
"axis",
")",
":",
"assert",
"(",
"axis",
"in",
"[",
"0",
",",
"1",
"]",
")",
"ndim",
"=",
"M",
".",
"ndim",
"assert",
"(",
"ndim",
"==",
"3",
")",
"ndim_roll",
"=",
"roll_indices",
".",... | rolls an array of matrices along an axis according to an array of indices *roll_indices* *axis* can be either 0 or 1 . | train | false |
8,921 | def SecurityCheck(func):
def Wrapper(request, *args, **kwargs):
'Wrapping function.'
if (WEBAUTH_MANAGER is None):
raise RuntimeError('Attempt to initialize before WEBAUTH_MANAGER set.')
return WEBAUTH_MANAGER.SecurityCheck(func, request, *args, **kwargs)
return Wrapper
| [
"def",
"SecurityCheck",
"(",
"func",
")",
":",
"def",
"Wrapper",
"(",
"request",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"WEBAUTH_MANAGER",
"is",
"None",
")",
":",
"raise",
"RuntimeError",
"(",
"'Attempt to initialize before WEBAUTH_MANAGER... | a decorator applied to protected web handlers . | train | true |
8,922 | def _build_id_tuple(params, spec):
if (spec is None):
return (None, None)
required_class = spec.class_
required_tag = spec.tag
tag_type = params.get(u'tag_type', spec.tag_type)
if (tag_type is not None):
required_class = 2
required_class = params.get(u'class_', required_class)
required_tag = params.get(u'tag... | [
"def",
"_build_id_tuple",
"(",
"params",
",",
"spec",
")",
":",
"if",
"(",
"spec",
"is",
"None",
")",
":",
"return",
"(",
"None",
",",
"None",
")",
"required_class",
"=",
"spec",
".",
"class_",
"required_tag",
"=",
"spec",
".",
"tag",
"tag_type",
"=",
... | builds a 2-element tuple used to identify fields by grabbing the class_ and tag from an asn1value class and the params dict being passed to it . | train | false |
8,925 | @calculator(272696320)
def calculate_perf_counter_counter(previous, current, property_name):
n0 = previous[property_name]
n1 = current[property_name]
d0 = previous['Timestamp_Sys100NS']
d1 = current['Timestamp_Sys100NS']
f = current['Frequency_Sys100NS']
if ((n0 is None) or (n1 is None)):
return
return ((n1 - ... | [
"@",
"calculator",
"(",
"272696320",
")",
"def",
"calculate_perf_counter_counter",
"(",
"previous",
",",
"current",
",",
"property_name",
")",
":",
"n0",
"=",
"previous",
"[",
"property_name",
"]",
"n1",
"=",
"current",
"[",
"property_name",
"]",
"d0",
"=",
... | perf_counter_counter URL . | train | true |
8,926 | def myTakeStep2(x):
s = 0.5
x += np.random.uniform((- s), s, np.shape(x))
return x
| [
"def",
"myTakeStep2",
"(",
"x",
")",
":",
"s",
"=",
"0.5",
"x",
"+=",
"np",
".",
"random",
".",
"uniform",
"(",
"(",
"-",
"s",
")",
",",
"s",
",",
"np",
".",
"shape",
"(",
"x",
")",
")",
"return",
"x"
] | redo randomdisplacement in function form without the attribute stepsize to make sure still everything works ok . | train | false |
8,927 | def tied_rank(x):
sorted_x = sorted(zip(x, range(len(x))))
r = [0 for k in x]
cur_val = sorted_x[0][0]
last_rank = 0
for i in range(len(sorted_x)):
if (cur_val != sorted_x[i][0]):
cur_val = sorted_x[i][0]
for j in range(last_rank, i):
r[sorted_x[j][1]] = (float(((last_rank + 1) + i)) / 2.0)
last_ran... | [
"def",
"tied_rank",
"(",
"x",
")",
":",
"sorted_x",
"=",
"sorted",
"(",
"zip",
"(",
"x",
",",
"range",
"(",
"len",
"(",
"x",
")",
")",
")",
")",
"r",
"=",
"[",
"0",
"for",
"k",
"in",
"x",
"]",
"cur_val",
"=",
"sorted_x",
"[",
"0",
"]",
"[",... | computes the tied rank of elements in x . | train | false |
8,929 | def get_link_name(connection, link):
if link:
return connection.follow_link(link).name
else:
return None
| [
"def",
"get_link_name",
"(",
"connection",
",",
"link",
")",
":",
"if",
"link",
":",
"return",
"connection",
".",
"follow_link",
"(",
"link",
")",
".",
"name",
"else",
":",
"return",
"None"
] | this method returns the name of the element which link points to . | train | false |
8,931 | def nthstr(n):
assert (n >= 0)
if ((n % 100) in [11, 12, 13]):
return ('%sth' % n)
return ({1: '%sst', 2: '%snd', 3: '%srd'}.get((n % 10), '%sth') % n)
| [
"def",
"nthstr",
"(",
"n",
")",
":",
"assert",
"(",
"n",
">=",
"0",
")",
"if",
"(",
"(",
"n",
"%",
"100",
")",
"in",
"[",
"11",
",",
"12",
",",
"13",
"]",
")",
":",
"return",
"(",
"'%sth'",
"%",
"n",
")",
"return",
"(",
"{",
"1",
":",
"... | formats an ordinal . | train | false |
8,932 | def _encode_objectid(name, value, dummy0, dummy1):
return (('\x07' + name) + value.binary)
| [
"def",
"_encode_objectid",
"(",
"name",
",",
"value",
",",
"dummy0",
",",
"dummy1",
")",
":",
"return",
"(",
"(",
"'\\x07'",
"+",
"name",
")",
"+",
"value",
".",
"binary",
")"
] | encode bson . | train | false |
8,934 | def build_nbt():
os.chdir('pymclevel')
os.system((sys.executable + ' setup.py build_ext --inplace --force'))
os.chdir('..')
| [
"def",
"build_nbt",
"(",
")",
":",
"os",
".",
"chdir",
"(",
"'pymclevel'",
")",
"os",
".",
"system",
"(",
"(",
"sys",
".",
"executable",
"+",
"' setup.py build_ext --inplace --force'",
")",
")",
"os",
".",
"chdir",
"(",
"'..'",
")"
] | builds _nbt . | train | false |
8,935 | def pluginPackagePaths(name):
package = name.split('.')
return [os.path.abspath(os.path.join(x, *package)) for x in sys.path if (not os.path.exists(os.path.join(x, *(package + ['__init__.py']))))]
| [
"def",
"pluginPackagePaths",
"(",
"name",
")",
":",
"package",
"=",
"name",
".",
"split",
"(",
"'.'",
")",
"return",
"[",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"join",
"(",
"x",
",",
"*",
"package",
")",
")",
"for",
"x",... | return a list of additional directories which should be searched for modules to be included as part of the named plugin package . | train | true |
8,936 | def test_docinfo():
exampledocinfo = env.metadata['metadata']
expecteddocinfo = {'author': u'David Goodger', 'authors': [u'Me', u'Myself', u'I'], 'address': u'123 Example Street\nExample, EX Canada\nA1B 2C3', 'field name': u'This is a generic bibliographic field.', 'field name 2': u'Generic bibliographic fields may ... | [
"def",
"test_docinfo",
"(",
")",
":",
"exampledocinfo",
"=",
"env",
".",
"metadata",
"[",
"'metadata'",
"]",
"expecteddocinfo",
"=",
"{",
"'author'",
":",
"u'David Goodger'",
",",
"'authors'",
":",
"[",
"u'Me'",
",",
"u'Myself'",
",",
"u'I'",
"]",
",",
"'a... | inspect the docinfo metadata stored in the first node of the document . | train | false |
8,937 | def raw_post_data(request):
return HttpResponse(request.raw_post_data)
| [
"def",
"raw_post_data",
"(",
"request",
")",
":",
"return",
"HttpResponse",
"(",
"request",
".",
"raw_post_data",
")"
] | a view that is requested with get and accesses request . | train | false |
8,938 | def p_function_type(p):
if (p[1] == 'void'):
p[0] = TType.VOID
else:
p[0] = p[1]
| [
"def",
"p_function_type",
"(",
"p",
")",
":",
"if",
"(",
"p",
"[",
"1",
"]",
"==",
"'void'",
")",
":",
"p",
"[",
"0",
"]",
"=",
"TType",
".",
"VOID",
"else",
":",
"p",
"[",
"0",
"]",
"=",
"p",
"[",
"1",
"]"
] | function_type : field_type | void . | train | false |
8,941 | def _compile_template_re(delimiters):
tag_types = '!>&/#^'
tag = ('\n (?P<whitespace>[\\ \\t]*)\n %(otag)s \\s*\n (?:\n (?P<change>=) \\s* (?P<delims>.+?) \\s* = |\n (?P<raw>{) \\s* (?P<raw_name>.+?) \\s* } |\n (?P<tag>[%(tag_types)s]?) \\s* (?P<tag_key>[\\s\\S]+?... | [
"def",
"_compile_template_re",
"(",
"delimiters",
")",
":",
"tag_types",
"=",
"'!>&/#^'",
"tag",
"=",
"(",
"'\\n (?P<whitespace>[\\\\ \\\\t]*)\\n %(otag)s \\\\s*\\n (?:\\n (?P<change>=) \\\\s* (?P<delims>.+?) \\\\s* = |\\n (?P<raw>{) \\\\s* (?P<raw... | return a regular expression object instance . | train | true |
8,942 | def describe_domain_brief(domain):
items = OrderedDict()
if (domain is None):
return items
items['Features'] = (len(domain.attributes) or 'None')
items['Meta attributes'] = (len(domain.metas) or 'None')
if domain.has_discrete_class:
items['Target'] = "Class '{}'".format(domain.class_var.name)
elif domain.has_... | [
"def",
"describe_domain_brief",
"(",
"domain",
")",
":",
"items",
"=",
"OrderedDict",
"(",
")",
"if",
"(",
"domain",
"is",
"None",
")",
":",
"return",
"items",
"items",
"[",
"'Features'",
"]",
"=",
"(",
"len",
"(",
"domain",
".",
"attributes",
")",
"or... | return an :obj:ordereddict with the number of features . | train | false |
8,944 | def dmp_gf_sqf_part(f, K):
raise NotImplementedError('multivariate polynomials over finite fields')
| [
"def",
"dmp_gf_sqf_part",
"(",
"f",
",",
"K",
")",
":",
"raise",
"NotImplementedError",
"(",
"'multivariate polynomials over finite fields'",
")"
] | compute square-free part of f in gf(p)[x] . | train | false |
8,948 | def clear_flows():
for c in core.openflow.connections:
d = of.ofp_flow_mod(command=of.OFPFC_DELETE)
c.send(d)
| [
"def",
"clear_flows",
"(",
")",
":",
"for",
"c",
"in",
"core",
".",
"openflow",
".",
"connections",
":",
"d",
"=",
"of",
".",
"ofp_flow_mod",
"(",
"command",
"=",
"of",
".",
"OFPFC_DELETE",
")",
"c",
".",
"send",
"(",
"d",
")"
] | clear flows on all switches . | train | false |
8,949 | def list_networks(kwargs=None, call=None):
if (call != 'function'):
raise SaltCloudSystemExit('The list_networks function must be called with -f or --function.')
return {'Networks': salt.utils.vmware.list_networks(_get_si())}
| [
"def",
"list_networks",
"(",
"kwargs",
"=",
"None",
",",
"call",
"=",
"None",
")",
":",
"if",
"(",
"call",
"!=",
"'function'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The list_networks function must be called with -f or --function.'",
")",
"return",
"{",
... | list virtual networks . | train | true |
8,950 | def SetIfNotNone(dict, attr_name, value):
if (value is not None):
dict[attr_name] = value
| [
"def",
"SetIfNotNone",
"(",
"dict",
",",
"attr_name",
",",
"value",
")",
":",
"if",
"(",
"value",
"is",
"not",
"None",
")",
":",
"dict",
"[",
"attr_name",
"]",
"=",
"value"
] | if "value" is not none . | train | false |
8,951 | def dead_items(vfs_top, paths):
dead_branches = {}
dead_saves = {}
for path in paths:
try:
n = vfs_top.lresolve(path)
except vfs.NodeError as e:
add_error(('unable to resolve %s: %s' % (path, e)))
else:
if isinstance(n, vfs.BranchList):
branchname = n.name
dead_branches[branchname] = n
dea... | [
"def",
"dead_items",
"(",
"vfs_top",
",",
"paths",
")",
":",
"dead_branches",
"=",
"{",
"}",
"dead_saves",
"=",
"{",
"}",
"for",
"path",
"in",
"paths",
":",
"try",
":",
"n",
"=",
"vfs_top",
".",
"lresolve",
"(",
"path",
")",
"except",
"vfs",
".",
"... | return an optimized set of removals . | train | false |
8,953 | def make_paginator(request, items, per_page=50):
try:
page = int(request.GET.get('page', '1'))
except ValueError:
page = 1
paginator = Paginator(items, per_page)
try:
items = paginator.page(page)
except (InvalidPage, EmptyPage):
items = paginator.page(paginator.num_pages)
return items
| [
"def",
"make_paginator",
"(",
"request",
",",
"items",
",",
"per_page",
"=",
"50",
")",
":",
"try",
":",
"page",
"=",
"int",
"(",
"request",
".",
"GET",
".",
"get",
"(",
"'page'",
",",
"'1'",
")",
")",
"except",
"ValueError",
":",
"page",
"=",
"1",... | make paginator . | train | false |
8,954 | def cxTwoPointCopy(ind1, ind2):
size = len(ind1)
cxpoint1 = random.randint(1, size)
cxpoint2 = random.randint(1, (size - 1))
if (cxpoint2 >= cxpoint1):
cxpoint2 += 1
else:
(cxpoint1, cxpoint2) = (cxpoint2, cxpoint1)
(ind1[cxpoint1:cxpoint2], ind2[cxpoint1:cxpoint2]) = (ind2[cxpoint1:cxpoint2].copy(), ind1[cxp... | [
"def",
"cxTwoPointCopy",
"(",
"ind1",
",",
"ind2",
")",
":",
"size",
"=",
"len",
"(",
"ind1",
")",
"cxpoint1",
"=",
"random",
".",
"randint",
"(",
"1",
",",
"size",
")",
"cxpoint2",
"=",
"random",
".",
"randint",
"(",
"1",
",",
"(",
"size",
"-",
... | execute a two points crossover with copy on the input individuals . | train | false |
8,956 | def server_started_with_option(client, cmdline_opt, config_opt):
command_line = get_command_line(client)
if ('parsed' in command_line):
parsed = command_line['parsed']
if (config_opt in parsed):
return parsed[config_opt]
argv = command_line['argv']
return (cmdline_opt in argv)
| [
"def",
"server_started_with_option",
"(",
"client",
",",
"cmdline_opt",
",",
"config_opt",
")",
":",
"command_line",
"=",
"get_command_line",
"(",
"client",
")",
"if",
"(",
"'parsed'",
"in",
"command_line",
")",
":",
"parsed",
"=",
"command_line",
"[",
"'parsed'... | check if the server was started with a particular option . | train | false |
8,957 | def rgba(s, *args):
if isinstance(s, string_types):
return get_color_from_hex(s)
elif isinstance(s, (list, tuple)):
s = map((lambda x: (x / 255.0)), s)
if (len(s) == 3):
return (list(s) + [1])
return s
elif isinstance(s, (int, float)):
s = map((lambda x: (x / 255.0)), ([s] + list(args)))
if (len(s) ==... | [
"def",
"rgba",
"(",
"s",
",",
"*",
"args",
")",
":",
"if",
"isinstance",
"(",
"s",
",",
"string_types",
")",
":",
"return",
"get_color_from_hex",
"(",
"s",
")",
"elif",
"isinstance",
"(",
"s",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"s",
"... | return a kivy color from either a hex string or a list of 0-255 values . | train | false |
8,958 | def button_action(button, action):
connect_button(button, action.trigger)
| [
"def",
"button_action",
"(",
"button",
",",
"action",
")",
":",
"connect_button",
"(",
"button",
",",
"action",
".",
"trigger",
")"
] | make a button trigger an action . | train | false |
8,959 | def humanize_key(key):
return u':'.join([(u'%02x' % ord(c)) for c in key.get_fingerprint()])
| [
"def",
"humanize_key",
"(",
"key",
")",
":",
"return",
"u':'",
".",
"join",
"(",
"[",
"(",
"u'%02x'",
"%",
"ord",
"(",
"c",
")",
")",
"for",
"c",
"in",
"key",
".",
"get_fingerprint",
"(",
")",
"]",
")"
] | returns a human-readable key as a series of hex characters . | train | false |
8,960 | def register_widget(widget):
WIDGETS[widget.name] = widget
return widget
| [
"def",
"register_widget",
"(",
"widget",
")",
":",
"WIDGETS",
"[",
"widget",
".",
"name",
"]",
"=",
"widget",
"return",
"widget"
] | registers widget in dictionary . | train | false |
8,962 | def install_pyenv(name, user=None):
ret = {'name': name, 'result': None, 'comment': '', 'changes': {}}
if __opts__['test']:
ret['comment'] = 'pyenv is set to be installed'
return ret
return _check_and_install_python(ret, user)
| [
"def",
"install_pyenv",
"(",
"name",
",",
"user",
"=",
"None",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'result'",
":",
"None",
",",
"'comment'",
":",
"''",
",",
"'changes'",
":",
"{",
"}",
"}",
"if",
"__opts__",
"[",
"'test'",
"]",
... | install pyenv if not installed . | train | true |
8,964 | @composite
def unique_name_strategy(draw):
return unicode(draw(st.uuids()))
| [
"@",
"composite",
"def",
"unique_name_strategy",
"(",
"draw",
")",
":",
"return",
"unicode",
"(",
"draw",
"(",
"st",
".",
"uuids",
"(",
")",
")",
")"
] | a hypothesis strategy to generate an always unique name . | train | false |
8,965 | def has_uniform_batch_size(mode):
return resolve_iterator_class(mode).uniform_batch_size
| [
"def",
"has_uniform_batch_size",
"(",
"mode",
")",
":",
"return",
"resolve_iterator_class",
"(",
"mode",
")",
".",
"uniform_batch_size"
] | returns true if the iteration scheme has uniform batch size . | train | false |
8,966 | def maybe_callable(thing, context=None):
if callable(thing):
return thing
if isinstance(thing, six.string_types):
thing = getattr(context, thing, None)
if callable(thing):
return thing
return None
| [
"def",
"maybe_callable",
"(",
"thing",
",",
"context",
"=",
"None",
")",
":",
"if",
"callable",
"(",
"thing",
")",
":",
"return",
"thing",
"if",
"isinstance",
"(",
"thing",
",",
"six",
".",
"string_types",
")",
":",
"thing",
"=",
"getattr",
"(",
"conte... | if thing is callable . | train | false |
8,967 | def _list_queues():
queue_dir = __opts__['sqlite_queue_dir']
files = os.path.join(queue_dir, '*.db')
paths = glob.glob(files)
queues = [os.path.splitext(os.path.basename(item))[0] for item in paths]
return queues
| [
"def",
"_list_queues",
"(",
")",
":",
"queue_dir",
"=",
"__opts__",
"[",
"'sqlite_queue_dir'",
"]",
"files",
"=",
"os",
".",
"path",
".",
"join",
"(",
"queue_dir",
",",
"'*.db'",
")",
"paths",
"=",
"glob",
".",
"glob",
"(",
"files",
")",
"queues",
"=",... | return a list of sqlite databases in the queue_dir . | train | true |
8,969 | def same_node(node1, node2):
return (node1.uuid == node2.uuid)
| [
"def",
"same_node",
"(",
"node1",
",",
"node2",
")",
":",
"return",
"(",
"node1",
".",
"uuid",
"==",
"node2",
".",
"uuid",
")"
] | return whether these two objects both refer to same cluster node . | train | false |
8,970 | def topic_name_from_path(path, project):
return _name_from_project_path(path, project, _TOPIC_TEMPLATE)
| [
"def",
"topic_name_from_path",
"(",
"path",
",",
"project",
")",
":",
"return",
"_name_from_project_path",
"(",
"path",
",",
"project",
",",
"_TOPIC_TEMPLATE",
")"
] | validate a topic uri path and get the topic name . | train | false |
8,971 | def total_sum_of_squares(y):
return sum(((v ** 2) for v in de_mean(y)))
| [
"def",
"total_sum_of_squares",
"(",
"y",
")",
":",
"return",
"sum",
"(",
"(",
"(",
"v",
"**",
"2",
")",
"for",
"v",
"in",
"de_mean",
"(",
"y",
")",
")",
")"
] | the total squared variation of y_is from their mean . | train | false |
8,972 | def detail(device='/dev/md0'):
ret = {}
ret['members'] = {}
if (not os.path.exists(device)):
msg = "Device {0} doesn't exist!"
raise CommandExecutionError(msg.format(device))
cmd = ['mdadm', '--detail', device]
for line in __salt__['cmd.run_stdout'](cmd, python_shell=False).splitlines():
if line.startswith(d... | [
"def",
"detail",
"(",
"device",
"=",
"'/dev/md0'",
")",
":",
"ret",
"=",
"{",
"}",
"ret",
"[",
"'members'",
"]",
"=",
"{",
"}",
"if",
"(",
"not",
"os",
".",
"path",
".",
"exists",
"(",
"device",
")",
")",
":",
"msg",
"=",
"\"Device {0} doesn't exis... | action detail view . | train | true |
8,973 | def is_float(input_string):
try:
float(input_string)
return True
except ValueError:
return False
| [
"def",
"is_float",
"(",
"input_string",
")",
":",
"try",
":",
"float",
"(",
"input_string",
")",
"return",
"True",
"except",
"ValueError",
":",
"return",
"False"
] | a check that tests that a given value is a float . | train | false |
8,974 | def disable_server(name, backend, socket='/var/run/haproxy.sock'):
if (backend == '*'):
backends = show_backends(socket=socket).split('\n')
else:
backends = [backend]
results = {}
for backend in backends:
ha_conn = _get_conn(socket)
ha_cmd = haproxy.cmds.disableServer(server=name, backend=backend)
ha_conn... | [
"def",
"disable_server",
"(",
"name",
",",
"backend",
",",
"socket",
"=",
"'/var/run/haproxy.sock'",
")",
":",
"if",
"(",
"backend",
"==",
"'*'",
")",
":",
"backends",
"=",
"show_backends",
"(",
"socket",
"=",
"socket",
")",
".",
"split",
"(",
"'\\n'",
"... | disable server . | train | false |
8,975 | @addons_reviewer_required
def beta_signed_log(request):
form = forms.BetaSignedLogForm(request.GET)
beta_signed_log = ActivityLog.objects.beta_signed_events()
motd_editable = acl.action_allowed(request, 'AddonReviewerMOTD', 'Edit')
if form.is_valid():
if form.cleaned_data['filter']:
beta_signed_log = beta_sign... | [
"@",
"addons_reviewer_required",
"def",
"beta_signed_log",
"(",
"request",
")",
":",
"form",
"=",
"forms",
".",
"BetaSignedLogForm",
"(",
"request",
".",
"GET",
")",
"beta_signed_log",
"=",
"ActivityLog",
".",
"objects",
".",
"beta_signed_events",
"(",
")",
"mot... | log of all the beta files that got signed . | train | false |
8,977 | def read_fiducials(fname):
(fid, tree, _) = fiff_open(fname)
with fid:
isotrak = dir_tree_find(tree, FIFF.FIFFB_ISOTRAK)
isotrak = isotrak[0]
pts = []
coord_frame = FIFF.FIFFV_COORD_UNKNOWN
for k in range(isotrak['nent']):
kind = isotrak['directory'][k].kind
pos = isotrak['directory'][k].pos
if (ki... | [
"def",
"read_fiducials",
"(",
"fname",
")",
":",
"(",
"fid",
",",
"tree",
",",
"_",
")",
"=",
"fiff_open",
"(",
"fname",
")",
"with",
"fid",
":",
"isotrak",
"=",
"dir_tree_find",
"(",
"tree",
",",
"FIFF",
".",
"FIFFB_ISOTRAK",
")",
"isotrak",
"=",
"i... | read fiducials from a fiff file . | train | false |
8,978 | def _loop_payload(params):
payload = {}
for (param, value) in six.iteritems(params):
if (value is not None):
payload[param] = value
return payload
| [
"def",
"_loop_payload",
"(",
"params",
")",
":",
"payload",
"=",
"{",
"}",
"for",
"(",
"param",
",",
"value",
")",
"in",
"six",
".",
"iteritems",
"(",
"params",
")",
":",
"if",
"(",
"value",
"is",
"not",
"None",
")",
":",
"payload",
"[",
"param",
... | pass in a dictionary of parameters . | train | true |
8,980 | def era_main():
return FlockerScriptRunner(script=EraScript(), options=EraOptions(), logging=False).main()
| [
"def",
"era_main",
"(",
")",
":",
"return",
"FlockerScriptRunner",
"(",
"script",
"=",
"EraScript",
"(",
")",
",",
"options",
"=",
"EraOptions",
"(",
")",
",",
"logging",
"=",
"False",
")",
".",
"main",
"(",
")"
] | entry point for flocker-node-era command-line tool . | train | false |
8,981 | def get_location_metres(original_location, dNorth, dEast):
earth_radius = 6378137.0
dLat = (dNorth / earth_radius)
dLon = (dEast / (earth_radius * math.cos(((math.pi * original_location.lat) / 180))))
newlat = (original_location.lat + ((dLat * 180) / math.pi))
newlon = (original_location.lon + ((dLon * 180) / math... | [
"def",
"get_location_metres",
"(",
"original_location",
",",
"dNorth",
",",
"dEast",
")",
":",
"earth_radius",
"=",
"6378137.0",
"dLat",
"=",
"(",
"dNorth",
"/",
"earth_radius",
")",
"dLon",
"=",
"(",
"dEast",
"/",
"(",
"earth_radius",
"*",
"math",
".",
"c... | returns a locationglobal object containing the latitude/longitude dnorth and deast metres from the specified original_location . | train | true |
8,982 | def islink(p):
return _false
| [
"def",
"islink",
"(",
"p",
")",
":",
"return",
"_false"
] | test for symbolic link . | train | false |
8,983 | def write_timeseries_value(client, project_resource, custom_metric_type, instance_id, metric_kind):
now = get_now_rfc3339()
timeseries_data = {'metric': {'type': custom_metric_type, 'labels': {'environment': 'STAGING'}}, 'resource': {'type': 'gce_instance', 'labels': {'instance_id': instance_id, 'zone': 'us-central1-... | [
"def",
"write_timeseries_value",
"(",
"client",
",",
"project_resource",
",",
"custom_metric_type",
",",
"instance_id",
",",
"metric_kind",
")",
":",
"now",
"=",
"get_now_rfc3339",
"(",
")",
"timeseries_data",
"=",
"{",
"'metric'",
":",
"{",
"'type'",
":",
"cust... | write the custom metric obtained by get_custom_data_point at a point in time . | train | false |
8,985 | def messy_split(long_line):
new_list = []
old_list = long_line.split('=')
for line in old_list:
if (len(line) != 0):
line += ('=' * ((4 - (len(line) % 4)) % 4))
new_list.append(line)
return new_list
| [
"def",
"messy_split",
"(",
"long_line",
")",
":",
"new_list",
"=",
"[",
"]",
"old_list",
"=",
"long_line",
".",
"split",
"(",
"'='",
")",
"for",
"line",
"in",
"old_list",
":",
"if",
"(",
"len",
"(",
"line",
")",
"!=",
"0",
")",
":",
"line",
"+=",
... | split on = gives me the right sections but deletes the b64 padding use modulo math to restore padding . | train | false |
8,986 | def lfsr_sequence(key, fill, n):
if (not isinstance(key, list)):
raise TypeError('key must be a list')
if (not isinstance(fill, list)):
raise TypeError('fill must be a list')
p = key[0].mod
F = FF(p)
s = fill
k = len(fill)
L = []
for i in range(n):
s0 = s[:]
L.append(s[0])
s = s[1:k]
x = sum([int((k... | [
"def",
"lfsr_sequence",
"(",
"key",
",",
"fill",
",",
"n",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"key",
",",
"list",
")",
")",
":",
"raise",
"TypeError",
"(",
"'key must be a list'",
")",
"if",
"(",
"not",
"isinstance",
"(",
"fill",
",",
"li... | this function creates an lfsr sequence . | train | false |
8,987 | def system_path():
ca_path = None
paths = [u'/usr/lib/ssl/certs/ca-certificates.crt', u'/etc/ssl/certs/ca-certificates.crt', u'/etc/ssl/certs/ca-bundle.crt', u'/etc/pki/tls/certs/ca-bundle.crt', u'/etc/ssl/ca-bundle.pem', u'/usr/local/share/certs/ca-root-nss.crt', u'/etc/ssl/cert.pem']
if (u'SSL_CERT_FILE' in os.env... | [
"def",
"system_path",
"(",
")",
":",
"ca_path",
"=",
"None",
"paths",
"=",
"[",
"u'/usr/lib/ssl/certs/ca-certificates.crt'",
",",
"u'/etc/ssl/certs/ca-certificates.crt'",
",",
"u'/etc/ssl/certs/ca-bundle.crt'",
",",
"u'/etc/pki/tls/certs/ca-bundle.crt'",
",",
"u'/etc/ssl/ca-bun... | tries to find a ca certs bundle in common locations :raises: oserror - when no valid ca certs bundle was found on the filesystem :return: the full filesystem path to a ca certs bundle file . | train | false |
8,988 | def show_all_prices(call=None, kwargs=None):
if (call == 'action'):
raise SaltCloudSystemExit('The show_all_prices function must be called with -f or --function.')
if (kwargs is None):
kwargs = {}
conn = get_conn(service='SoftLayer_Product_Package')
if ('code' not in kwargs):
return conn.getCategories(id=50)
... | [
"def",
"show_all_prices",
"(",
"call",
"=",
"None",
",",
"kwargs",
"=",
"None",
")",
":",
"if",
"(",
"call",
"==",
"'action'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The show_all_prices function must be called with -f or --function.'",
")",
"if",
"(",
"k... | return a dict of all prices on the cloud provider . | train | true |
8,990 | def pnio_get_config(pkt):
ether = _get_ethernet(pkt)
config = None
if ((ether is not None) and ((ether.src, ether.dst) in conf.contribs['PNIO_RTC'])):
config = conf.contribs['PNIO_RTC'][(ether.src, ether.dst)]
return config
| [
"def",
"pnio_get_config",
"(",
"pkt",
")",
":",
"ether",
"=",
"_get_ethernet",
"(",
"pkt",
")",
"config",
"=",
"None",
"if",
"(",
"(",
"ether",
"is",
"not",
"None",
")",
"and",
"(",
"(",
"ether",
".",
"src",
",",
"ether",
".",
"dst",
")",
"in",
"... | retrieve the config for a given communication . | train | false |
8,992 | def node_region(patched_ast_node):
return patched_ast_node.region
| [
"def",
"node_region",
"(",
"patched_ast_node",
")",
":",
"return",
"patched_ast_node",
".",
"region"
] | get the region of a patched ast node . | train | false |
8,993 | def floating_ip_get_all_by_host(context, host):
return IMPL.floating_ip_get_all_by_host(context, host)
| [
"def",
"floating_ip_get_all_by_host",
"(",
"context",
",",
"host",
")",
":",
"return",
"IMPL",
".",
"floating_ip_get_all_by_host",
"(",
"context",
",",
"host",
")"
] | get all floating ips by host . | train | false |
8,996 | def cpu_usage_for_process(results, process):
process_results = [r for r in results if ((r['metric']['type'] == 'cputime') and (r['process'] == process))]
cpu_values = sum((r['value'] for r in process_results))
wallclock_values = sum((r['wallclock'] for r in process_results))
if (wallclock_values > 0):
return (flo... | [
"def",
"cpu_usage_for_process",
"(",
"results",
",",
"process",
")",
":",
"process_results",
"=",
"[",
"r",
"for",
"r",
"in",
"results",
"if",
"(",
"(",
"r",
"[",
"'metric'",
"]",
"[",
"'type'",
"]",
"==",
"'cputime'",
")",
"and",
"(",
"r",
"[",
"'pr... | calculate the cpu percentage for a process running in a particular scenario . | train | false |
8,998 | def retrieveAcknowledge():
a = TpPd(pd=3)
b = MessageType(mesType=29)
packet = (a / b)
return packet
| [
"def",
"retrieveAcknowledge",
"(",
")",
":",
"a",
"=",
"TpPd",
"(",
"pd",
"=",
"3",
")",
"b",
"=",
"MessageType",
"(",
"mesType",
"=",
"29",
")",
"packet",
"=",
"(",
"a",
"/",
"b",
")",
"return",
"packet"
] | retrieve acknowledge section 9 . | train | true |
8,999 | def _adjust_start_date_for_beta_testers(user, descriptor, course_key):
return adjust_start_date(user, descriptor.days_early_for_beta, descriptor.start, course_key)
| [
"def",
"_adjust_start_date_for_beta_testers",
"(",
"user",
",",
"descriptor",
",",
"course_key",
")",
":",
"return",
"adjust_start_date",
"(",
"user",
",",
"descriptor",
".",
"days_early_for_beta",
",",
"descriptor",
".",
"start",
",",
"course_key",
")"
] | if user is in a beta test group . | train | false |
9,001 | def configure_debug_log_handlers(logger):
global printed_log_start_message
logger.setLevel(logging.DEBUG)
logging_to_file = True
try:
make_sure_path_exists(os.path.dirname(log_filepath), 448)
debug_handler = logging.FileHandler(log_filepath)
except (OSError, IOError):
logging_to_file = False
debug_handler ... | [
"def",
"configure_debug_log_handlers",
"(",
"logger",
")",
":",
"global",
"printed_log_start_message",
"logger",
".",
"setLevel",
"(",
"logging",
".",
"DEBUG",
")",
"logging_to_file",
"=",
"True",
"try",
":",
"make_sure_path_exists",
"(",
"os",
".",
"path",
".",
... | warnings and above to stderr . | train | false |
9,002 | def hrm_training_month(row):
if hasattr(row, 'hrm_training'):
row = row.hrm_training
try:
date = row.date
except AttributeError:
date = None
if date:
return ('%s/%02d' % (date.year, date.month))
else:
return current.messages['NONE']
| [
"def",
"hrm_training_month",
"(",
"row",
")",
":",
"if",
"hasattr",
"(",
"row",
",",
"'hrm_training'",
")",
":",
"row",
"=",
"row",
".",
"hrm_training",
"try",
":",
"date",
"=",
"row",
".",
"date",
"except",
"AttributeError",
":",
"date",
"=",
"None",
... | year/month of the start date of the training event . | train | false |
9,004 | def load_pycryptodome_raw_lib(name, cdecl):
split = name.split('.')
(dir_comps, basename) = (split[:(-1)], split[(-1)])
for (ext, mod, typ) in imp.get_suffixes():
if (typ == imp.C_EXTENSION):
try:
return load_lib(pycryptodome_filename(dir_comps, (basename + ext)), cdecl)
except OSError:
pass
raise O... | [
"def",
"load_pycryptodome_raw_lib",
"(",
"name",
",",
"cdecl",
")",
":",
"split",
"=",
"name",
".",
"split",
"(",
"'.'",
")",
"(",
"dir_comps",
",",
"basename",
")",
"=",
"(",
"split",
"[",
":",
"(",
"-",
"1",
")",
"]",
",",
"split",
"[",
"(",
"-... | load a shared library and return a handle to it . | train | false |
9,006 | def enable_insecure_serializers(choices=[u'pickle', u'yaml', u'msgpack']):
for choice in choices:
try:
registry.enable(choice)
except KeyError:
pass
| [
"def",
"enable_insecure_serializers",
"(",
"choices",
"=",
"[",
"u'pickle'",
",",
"u'yaml'",
",",
"u'msgpack'",
"]",
")",
":",
"for",
"choice",
"in",
"choices",
":",
"try",
":",
"registry",
".",
"enable",
"(",
"choice",
")",
"except",
"KeyError",
":",
"pas... | enable serializers that are considered to be unsafe . | train | false |
9,007 | def get_all_groups(path_prefix='/', region=None, key=None, keyid=None, profile=None):
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
if (not conn):
return None
_groups = conn.get_all_groups(path_prefix=path_prefix)
groups = _groups.list_groups_response.list_groups_result.groups
marker = g... | [
"def",
"get_all_groups",
"(",
"path_prefix",
"=",
"'/'",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key",
"=",
... | retrieve all ansible groups . | train | true |
9,008 | def nonthreadsafe(fn):
@wraps(fn)
def core(*args, **kwargs):
with _cuda_compiler_lock:
return fn(*args, **kwargs)
return core
| [
"def",
"nonthreadsafe",
"(",
"fn",
")",
":",
"@",
"wraps",
"(",
"fn",
")",
"def",
"core",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"with",
"_cuda_compiler_lock",
":",
"return",
"fn",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"return",
"co... | wraps a function to prevent multiple threads from executing it in parallel due to llvm is not threadsafe . | train | false |
9,009 | def get_master_uri(required=True, env=None, argv=None):
if (env is None):
env = os.environ
if (argv is None):
argv = sys.argv
try:
for arg in argv:
if arg.startswith('__master:='):
val = None
try:
(_, val) = arg.split(':=')
except:
pass
if (not val):
raise ROSEnvException(("__... | [
"def",
"get_master_uri",
"(",
"required",
"=",
"True",
",",
"env",
"=",
"None",
",",
"argv",
"=",
"None",
")",
":",
"if",
"(",
"env",
"is",
"None",
")",
":",
"env",
"=",
"os",
".",
"environ",
"if",
"(",
"argv",
"is",
"None",
")",
":",
"argv",
"... | get the ros_master_uri setting from the command-line args or environment . | train | false |
9,010 | def add_tagids(source):
return _modify_tagids(source)
| [
"def",
"add_tagids",
"(",
"source",
")",
":",
"return",
"_modify_tagids",
"(",
"source",
")"
] | applies a unique attribute code number for each tag element in order to be identified later in the process of apply annotation . | train | false |
9,012 | def get_localzone():
global _cache_tz
if (_cache_tz is None):
_cache_tz = pytz.timezone(get_localzone_name())
return _cache_tz
| [
"def",
"get_localzone",
"(",
")",
":",
"global",
"_cache_tz",
"if",
"(",
"_cache_tz",
"is",
"None",
")",
":",
"_cache_tz",
"=",
"pytz",
".",
"timezone",
"(",
"get_localzone_name",
"(",
")",
")",
"return",
"_cache_tz"
] | returns the current underlying local timezone object . | train | false |
9,013 | def autosummary_toc_visit_html(self, node):
raise nodes.SkipNode
| [
"def",
"autosummary_toc_visit_html",
"(",
"self",
",",
"node",
")",
":",
"raise",
"nodes",
".",
"SkipNode"
] | hide autosummary toctree list in html output . | train | false |
9,014 | def get_stock_basics(date=None):
wdate = (du.last_tddate() if (date is None) else date)
wdate = wdate.replace('-', '')
if (wdate < '20160809'):
return None
datepre = ('' if (date is None) else ((wdate[0:4] + wdate[4:6]) + '/'))
request = Request((ct.ALL_STOCK_BASICS_FILE % (datepre, ('' if (date is None) else wd... | [
"def",
"get_stock_basics",
"(",
"date",
"=",
"None",
")",
":",
"wdate",
"=",
"(",
"du",
".",
"last_tddate",
"(",
")",
"if",
"(",
"date",
"is",
"None",
")",
"else",
"date",
")",
"wdate",
"=",
"wdate",
".",
"replace",
"(",
"'-'",
",",
"''",
")",
"i... | parameters date:日期yyyy-mm-dd,默认为上一个交易日,目前只能提供2016-08-09之后的历史数据 return dataframe code . | train | false |
9,015 | @pytest.mark.parametrize((u'in_val', u'in_unit'), [([0.1, 5000.0, 10000.0], u.AA), ([100000.0, 2.0, 1.0], (u.micron ** (-1))), ([2.99792458e+19, 599584916000000.0, 299792458000000.0], u.Hz), ([1.98644568e-14, 3.97289137e-19, 1.98644568e-19], u.J)])
def test_spectral4(in_val, in_unit):
out_units = [(u.micron ** (-1)), ... | [
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"(",
"u'in_val'",
",",
"u'in_unit'",
")",
",",
"[",
"(",
"[",
"0.1",
",",
"5000.0",
",",
"10000.0",
"]",
",",
"u",
".",
"AA",
")",
",",
"(",
"[",
"100000.0",
",",
"2.0",
",",
"1.0",
"]",
",",... | wave number conversion w . | train | false |
9,017 | def atomic_group_add_labels(id, labels):
label_objs = models.Label.smart_get_bulk(labels)
models.AtomicGroup.smart_get(id).label_set.add(*label_objs)
| [
"def",
"atomic_group_add_labels",
"(",
"id",
",",
"labels",
")",
":",
"label_objs",
"=",
"models",
".",
"Label",
".",
"smart_get_bulk",
"(",
"labels",
")",
"models",
".",
"AtomicGroup",
".",
"smart_get",
"(",
"id",
")",
".",
"label_set",
".",
"add",
"(",
... | add labels to atomic group . | train | false |
9,018 | def get_pyos_inputhook():
return ctypes.c_void_p.in_dll(ctypes.pythonapi, 'PyOS_InputHook')
| [
"def",
"get_pyos_inputhook",
"(",
")",
":",
"return",
"ctypes",
".",
"c_void_p",
".",
"in_dll",
"(",
"ctypes",
".",
"pythonapi",
",",
"'PyOS_InputHook'",
")"
] | return the current pyos_inputhook as a ctypes . | train | false |
9,019 | def GenApiConfig(service_class_names, generator=None, hostname=None):
service_map = {}
generator = (generator or api_config.ApiConfigGenerator())
for service_class_name in service_class_names:
(module_name, base_service_class_name) = service_class_name.rsplit('.', 1)
module = __import__(module_name, fromlist=bas... | [
"def",
"GenApiConfig",
"(",
"service_class_names",
",",
"generator",
"=",
"None",
",",
"hostname",
"=",
"None",
")",
":",
"service_map",
"=",
"{",
"}",
"generator",
"=",
"(",
"generator",
"or",
"api_config",
".",
"ApiConfigGenerator",
"(",
")",
")",
"for",
... | write an api configuration for endpoints annotated protorpc services . | train | false |
9,024 | def get_logger_name_for_module(module):
module_file = module.__file__
base_dir = os.path.dirname(os.path.abspath(module_file))
module_name = os.path.basename(module_file)
module_name = module_name.replace('.pyc', '').replace('.py', '')
split = base_dir.split(os.path.sep)
split = [component for component in split ... | [
"def",
"get_logger_name_for_module",
"(",
"module",
")",
":",
"module_file",
"=",
"module",
".",
"__file__",
"base_dir",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"module_file",
")",
")",
"module_name",
"=",
"os",... | retrieve fully qualified logger name for current module :type: str . | train | false |
9,025 | def get_ami_metadata():
res = {}
base_url = 'http://169.254.169.254/latest/meta-data'
instance_id = run(('curl %s/instance-id' % base_url))
assert re.match('i-[0-9a-f]{8}', instance_id)
res['instance-id'] = instance_id
return res
| [
"def",
"get_ami_metadata",
"(",
")",
":",
"res",
"=",
"{",
"}",
"base_url",
"=",
"'http://169.254.169.254/latest/meta-data'",
"instance_id",
"=",
"run",
"(",
"(",
"'curl %s/instance-id'",
"%",
"base_url",
")",
")",
"assert",
"re",
".",
"match",
"(",
"'i-[0-9a-f]... | fetch ami metadata for the local instance . | train | false |
9,026 | def running_under_virtualenv():
return hasattr(sys, 'real_prefix')
| [
"def",
"running_under_virtualenv",
"(",
")",
":",
"return",
"hasattr",
"(",
"sys",
",",
"'real_prefix'",
")"
] | return true if were running inside a virtualenv . | train | false |
9,027 | def _build_element_wise_ufunc_wrapper(cres, signature):
ctx = cres.target_context
library = cres.library
fname = cres.fndesc.llvm_func_name
env = None
if cres.objectmode:
env = cres.environment
assert (env is not None)
ll_intp = cres.target_context.get_value_type(types.intp)
ll_pyobj = cres.target_context.... | [
"def",
"_build_element_wise_ufunc_wrapper",
"(",
"cres",
",",
"signature",
")",
":",
"ctx",
"=",
"cres",
".",
"target_context",
"library",
"=",
"cres",
".",
"library",
"fname",
"=",
"cres",
".",
"fndesc",
".",
"llvm_func_name",
"env",
"=",
"None",
"if",
"cre... | build a wrapper for the ufunc loop entry point given by the compilation result object . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.