id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1
value | is_duplicated bool 2
classes |
|---|---|---|---|---|---|
25,142 | def on_valid(valid_content_type, on_invalid=json):
invalid_kwargs = introspect.generate_accepted_kwargs(on_invalid, 'request', 'response')
invalid_takes_response = introspect.takes_all_arguments(on_invalid, 'response')
def wrapper(function):
valid_kwargs = introspect.generate_accepted_kwargs(function, 'request', 'response')
valid_takes_response = introspect.takes_all_arguments(function, 'response')
@content_type(valid_content_type)
@wraps(function)
def output_content(content, response, **kwargs):
if ((type(content) == dict) and ('errors' in content)):
response.content_type = on_invalid.content_type
if invalid_takes_response:
kwargs['response'] = response
return on_invalid(content, **invalid_kwargs(kwargs))
if valid_takes_response:
kwargs['response'] = response
return function(content, **valid_kwargs(kwargs))
return output_content
return wrapper
| [
"def",
"on_valid",
"(",
"valid_content_type",
",",
"on_invalid",
"=",
"json",
")",
":",
"invalid_kwargs",
"=",
"introspect",
".",
"generate_accepted_kwargs",
"(",
"on_invalid",
",",
"'request'",
",",
"'response'",
")",
"invalid_takes_response",
"=",
"introspect",
".... | renders as the specified content type only if no errors are found in the provided data object . | train | true |
25,143 | def getManipulatedPaths(close, elementNode, loop, prefix, sideLength):
if (len(loop) < 1):
return [[]]
derivation = BottomDerivation(elementNode, prefix)
targetMatrix = matrix.getBranchMatrixSetElementNode(elementNode)
transformedLoop = matrix.getTransformedVector3s(matrix.getIdentityTetragrid(targetMatrix.tetragrid), loop)
lift = ((derivation.altitude + derivation.getAdditionalPathLift()) - euclidean.getBottomByPath(transformedLoop))
for point in loop:
point.z += lift
return [loop]
| [
"def",
"getManipulatedPaths",
"(",
"close",
",",
"elementNode",
",",
"loop",
",",
"prefix",
",",
"sideLength",
")",
":",
"if",
"(",
"len",
"(",
"loop",
")",
"<",
"1",
")",
":",
"return",
"[",
"[",
"]",
"]",
"derivation",
"=",
"BottomDerivation",
"(",
... | get round loop . | train | false |
25,144 | def create_comm_note(obj, version, author, body, note_type=comm.NO_ACTION, perms=None, attachments=None):
perms = (perms or {})
if (('developer' not in perms) and (note_type in comm.REVIEWER_NOTE_TYPES)):
perms['developer'] = False
create_perms = dict(((('read_permission_%s' % key), has_perm) for (key, has_perm) in perms.iteritems()))
version_param = {}
if (obj.__class__ == Webapp):
version_param['_version'] = version
elif (obj.__class__ == Extension):
version_param['_extension_version'] = version
(thread, created_thread) = obj.threads.safer_get_or_create(defaults=create_perms, **version_param)
note = thread.notes.create(note_type=note_type, body=body, author=author, **create_perms)
if attachments:
create_attachments(note, attachments)
post_create_comm_note(note)
return (thread, note)
| [
"def",
"create_comm_note",
"(",
"obj",
",",
"version",
",",
"author",
",",
"body",
",",
"note_type",
"=",
"comm",
".",
"NO_ACTION",
",",
"perms",
"=",
"None",
",",
"attachments",
"=",
"None",
")",
":",
"perms",
"=",
"(",
"perms",
"or",
"{",
"}",
")",... | creates a note on an obj versions thread . | train | false |
25,145 | def parse_partial_identifier(word):
p = sqlparse.parse(word)[0]
n_tok = len(p.tokens)
if ((n_tok == 1) and isinstance(p.tokens[0], Identifier)):
return p.tokens[0]
elif p.token_next_by(m=(Error, '"'))[1]:
return parse_partial_identifier((word + '"'))
else:
return None
| [
"def",
"parse_partial_identifier",
"(",
"word",
")",
":",
"p",
"=",
"sqlparse",
".",
"parse",
"(",
"word",
")",
"[",
"0",
"]",
"n_tok",
"=",
"len",
"(",
"p",
".",
"tokens",
")",
"if",
"(",
"(",
"n_tok",
"==",
"1",
")",
"and",
"isinstance",
"(",
"... | attempt to parse a word as an identifier word may include a schema qualification . | train | false |
25,146 | def stub_out_image_service(test):
image_service = FakeImageService()
test.stub_out('nova.image.glance.get_remote_image_service', (lambda x, y: (image_service, y)))
test.stub_out('nova.image.glance.get_default_image_service', (lambda : image_service))
test.useFixture(nova_fixtures.ConfPatcher(group='glance', api_servers=['http://localhost:9292']))
return image_service
| [
"def",
"stub_out_image_service",
"(",
"test",
")",
":",
"image_service",
"=",
"FakeImageService",
"(",
")",
"test",
".",
"stub_out",
"(",
"'nova.image.glance.get_remote_image_service'",
",",
"(",
"lambda",
"x",
",",
"y",
":",
"(",
"image_service",
",",
"y",
")",... | stubs out the image service for the test with the fakeimageservice . | train | false |
25,148 | def reorder_title(title, articles=('the',), separators=(',', ', ')):
ltitle = title.lower()
for article in articles:
for separator in separators:
suffix = (separator + article)
if (ltitle[(- len(suffix)):] == suffix):
return ((title[((- len(suffix)) + len(separator)):] + ' ') + title[:(- len(suffix))])
return title
| [
"def",
"reorder_title",
"(",
"title",
",",
"articles",
"=",
"(",
"'the'",
",",
")",
",",
"separators",
"=",
"(",
"','",
",",
"', '",
")",
")",
":",
"ltitle",
"=",
"title",
".",
"lower",
"(",
")",
"for",
"article",
"in",
"articles",
":",
"for",
"sep... | reorder the title . | train | false |
25,149 | def get_descrs(arrays, col_name_map):
out_descrs = []
for (out_name, in_names) in six.iteritems(col_name_map):
in_cols = [arr[name] for (arr, name) in zip(arrays, in_names) if (name is not None)]
names = [name for name in in_names if (name is not None)]
try:
dtype = common_dtype(in_cols)
except TableMergeError as tme:
raise TableMergeError(u"The '{0}' columns have incompatible types: {1}".format(names[0], tme._incompat_types))
uniq_shapes = set((col.shape[1:] for col in in_cols))
if (len(uniq_shapes) != 1):
raise TableMergeError(u'Key columns {0!r} have different shape'.format(name))
shape = uniq_shapes.pop()
out_descrs.append((fix_column_name(out_name), dtype, shape))
return out_descrs
| [
"def",
"get_descrs",
"(",
"arrays",
",",
"col_name_map",
")",
":",
"out_descrs",
"=",
"[",
"]",
"for",
"(",
"out_name",
",",
"in_names",
")",
"in",
"six",
".",
"iteritems",
"(",
"col_name_map",
")",
":",
"in_cols",
"=",
"[",
"arr",
"[",
"name",
"]",
... | find the dtypes descrs resulting from merging the list of arrays dtypes . | train | false |
25,150 | def generate_config(hass, passwd):
from homeassistant.components.mqtt import PROTOCOL_311
config = {'listeners': {'default': {'max-connections': 50000, 'bind': '0.0.0.0:1883', 'type': 'tcp'}, 'ws-1': {'bind': '0.0.0.0:8080', 'type': 'ws'}}, 'auth': {'allow-anonymous': (hass.config.api.api_password is None)}, 'plugins': ['auth_anonymous']}
if hass.config.api.api_password:
username = 'homeassistant'
password = hass.config.api.api_password
from passlib.apps import custom_app_context
passwd.write('homeassistant:{}\n'.format(custom_app_context.encrypt(hass.config.api.api_password)).encode('utf-8'))
passwd.flush()
config['auth']['password-file'] = passwd.name
config['plugins'].append('auth_file')
else:
username = None
password = None
client_config = ('localhost', 1883, username, password, None, PROTOCOL_311)
return (config, client_config)
| [
"def",
"generate_config",
"(",
"hass",
",",
"passwd",
")",
":",
"from",
"homeassistant",
".",
"components",
".",
"mqtt",
"import",
"PROTOCOL_311",
"config",
"=",
"{",
"'listeners'",
":",
"{",
"'default'",
":",
"{",
"'max-connections'",
":",
"50000",
",",
"'b... | generate a configuration based on current home assistant instance . | train | false |
25,151 | def upload_imageattachment(request, obj):
return upload_media(request, ImageAttachmentUploadForm, create_imageattachment, obj=obj)
| [
"def",
"upload_imageattachment",
"(",
"request",
",",
"obj",
")",
":",
"return",
"upload_media",
"(",
"request",
",",
"ImageAttachmentUploadForm",
",",
"create_imageattachment",
",",
"obj",
"=",
"obj",
")"
] | uploads image attachments . | train | false |
25,152 | def simple_rnn(incoming, n_units, activation='sigmoid', dropout=None, bias=True, weights_init=None, return_seq=False, return_state=False, initial_state=None, dynamic=False, trainable=True, restore=True, reuse=False, scope=None, name='SimpleRNN'):
cell = BasicRNNCell(n_units, activation=activation, bias=bias, weights_init=weights_init, trainable=trainable, restore=restore, reuse=reuse)
x = _rnn_template(incoming, cell=cell, dropout=dropout, return_seq=return_seq, return_state=return_state, initial_state=initial_state, dynamic=dynamic, scope=scope, name=name)
return x
| [
"def",
"simple_rnn",
"(",
"incoming",
",",
"n_units",
",",
"activation",
"=",
"'sigmoid'",
",",
"dropout",
"=",
"None",
",",
"bias",
"=",
"True",
",",
"weights_init",
"=",
"None",
",",
"return_seq",
"=",
"False",
",",
"return_state",
"=",
"False",
",",
"... | simple rnn . | train | false |
25,157 | def is_git_dir(d):
if osp.isdir(d):
if (osp.isdir(osp.join(d, 'objects')) and osp.isdir(osp.join(d, 'refs'))):
headref = osp.join(d, 'HEAD')
return (osp.isfile(headref) or (osp.islink(headref) and os.readlink(headref).startswith('refs')))
elif (osp.isfile(osp.join(d, 'gitdir')) and osp.isfile(osp.join(d, 'commondir')) and osp.isfile(osp.join(d, 'gitfile'))):
raise WorkTreeRepositoryUnsupported(d)
return False
| [
"def",
"is_git_dir",
"(",
"d",
")",
":",
"if",
"osp",
".",
"isdir",
"(",
"d",
")",
":",
"if",
"(",
"osp",
".",
"isdir",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'objects'",
")",
")",
"and",
"osp",
".",
"isdir",
"(",
"osp",
".",
"join",
"(",
... | from gits setup . | train | true |
25,158 | def advantage(cpu_dt, gpu_dt):
assert ((gpu_dt >= 0) and (cpu_dt >= 0))
if ((gpu_dt == 0) and (cpu_dt == 0)):
return numpy.nan
elif (gpu_dt == 0):
return numpy.inf
else:
return (cpu_dt / gpu_dt)
| [
"def",
"advantage",
"(",
"cpu_dt",
",",
"gpu_dt",
")",
":",
"assert",
"(",
"(",
"gpu_dt",
">=",
"0",
")",
"and",
"(",
"cpu_dt",
">=",
"0",
")",
")",
"if",
"(",
"(",
"gpu_dt",
"==",
"0",
")",
"and",
"(",
"cpu_dt",
"==",
"0",
")",
")",
":",
"re... | return ratio of cpu_dt / gpu_dt . | train | false |
25,161 | def _validate_connectivity(image_dim, connectivity, offset):
if (connectivity is None):
connectivity = 1
if np.isscalar(connectivity):
c_connectivity = ndi.generate_binary_structure(image_dim, connectivity)
else:
c_connectivity = np.array(connectivity, bool)
if (c_connectivity.ndim != image_dim):
raise ValueError('Connectivity dimension must be same as image')
if (offset is None):
if any([((x % 2) == 0) for x in c_connectivity.shape]):
raise ValueError('Connectivity array must have an unambiguous center')
offset = (np.array(c_connectivity.shape) // 2)
return (c_connectivity, offset)
| [
"def",
"_validate_connectivity",
"(",
"image_dim",
",",
"connectivity",
",",
"offset",
")",
":",
"if",
"(",
"connectivity",
"is",
"None",
")",
":",
"connectivity",
"=",
"1",
"if",
"np",
".",
"isscalar",
"(",
"connectivity",
")",
":",
"c_connectivity",
"=",
... | convert any valid connectivity to a structuring element and offset . | train | false |
25,164 | def unlock_file(filename):
log.trace('Removing lock for {0}'.format(filename))
lock = (filename + '.lock')
try:
os.remove(lock)
except OSError as exc:
log.trace('Unable to remove lock for {0}: {1}'.format(filename, exc))
| [
"def",
"unlock_file",
"(",
"filename",
")",
":",
"log",
".",
"trace",
"(",
"'Removing lock for {0}'",
".",
"format",
"(",
"filename",
")",
")",
"lock",
"=",
"(",
"filename",
"+",
"'.lock'",
")",
"try",
":",
"os",
".",
"remove",
"(",
"lock",
")",
"excep... | unlock file . | train | true |
25,165 | def _course_outline_json(request, course_module):
return create_xblock_info(course_module, include_child_info=True, course_outline=True, include_children_predicate=(lambda xblock: (not (xblock.category == 'vertical'))), user=request.user)
| [
"def",
"_course_outline_json",
"(",
"request",
",",
"course_module",
")",
":",
"return",
"create_xblock_info",
"(",
"course_module",
",",
"include_child_info",
"=",
"True",
",",
"course_outline",
"=",
"True",
",",
"include_children_predicate",
"=",
"(",
"lambda",
"x... | returns a json representation of the course module and recursively all of its children . | train | false |
25,166 | def ip_to_int(ip):
ret = 0
for octet in ip.split('.'):
ret = ((ret * 256) + int(octet))
return ret
| [
"def",
"ip_to_int",
"(",
"ip",
")",
":",
"ret",
"=",
"0",
"for",
"octet",
"in",
"ip",
".",
"split",
"(",
"'.'",
")",
":",
"ret",
"=",
"(",
"(",
"ret",
"*",
"256",
")",
"+",
"int",
"(",
"octet",
")",
")",
"return",
"ret"
] | converts an ip address to an integer . | train | true |
25,169 | @testing.requires_testing_data
@requires_mayavi
@requires_mne
@requires_freesurfer
def test_subject_source_with_fsaverage():
from mne.gui._file_traits import MRISubjectSource
tempdir = _TempDir()
mri = MRISubjectSource()
assert_false(mri.can_create_fsaverage)
assert_raises(RuntimeError, mri.create_fsaverage)
mri.subjects_dir = tempdir
assert_true(mri.can_create_fsaverage)
mri.create_fsaverage()
| [
"@",
"testing",
".",
"requires_testing_data",
"@",
"requires_mayavi",
"@",
"requires_mne",
"@",
"requires_freesurfer",
"def",
"test_subject_source_with_fsaverage",
"(",
")",
":",
"from",
"mne",
".",
"gui",
".",
"_file_traits",
"import",
"MRISubjectSource",
"tempdir",
... | test subjectselector . | train | false |
25,170 | def build_versions_form(project):
attrs = {'project': project}
versions_qs = project.versions.all()
active = versions_qs.filter(active=True)
if active.exists():
choices = [(version.slug, version.verbose_name) for version in active]
attrs['default-version'] = forms.ChoiceField(label=_('Default Version'), choices=choices, initial=project.get_default_version())
for version in versions_qs:
field_name = ('version-%s' % version.slug)
privacy_name = ('privacy-%s' % version.slug)
if (version.type == TAG):
label = ('%s (%s)' % (version.verbose_name, version.identifier[:8]))
else:
label = version.verbose_name
attrs[field_name] = forms.BooleanField(label=label, widget=DualCheckboxWidget(version), initial=version.active, required=False)
attrs[privacy_name] = forms.ChoiceField(label='privacy', choices=constants.PRIVACY_CHOICES, initial=version.privacy_level)
return type('VersionsForm', (BaseVersionsForm,), attrs)
| [
"def",
"build_versions_form",
"(",
"project",
")",
":",
"attrs",
"=",
"{",
"'project'",
":",
"project",
"}",
"versions_qs",
"=",
"project",
".",
"versions",
".",
"all",
"(",
")",
"active",
"=",
"versions_qs",
".",
"filter",
"(",
"active",
"=",
"True",
")... | versions form with a list of versions and version privacy levels . | train | false |
25,172 | def award():
mode = session.s3.hrm.mode
def prep(r):
if (mode is not None):
auth.permission.fail()
return True
s3.prep = prep
output = s3_rest_controller()
return output
| [
"def",
"award",
"(",
")",
":",
"mode",
"=",
"session",
".",
"s3",
".",
"hrm",
".",
"mode",
"def",
"prep",
"(",
"r",
")",
":",
"if",
"(",
"mode",
"is",
"not",
"None",
")",
":",
"auth",
".",
"permission",
".",
"fail",
"(",
")",
"return",
"True",
... | volunteer awards controller . | train | false |
25,173 | def is_iterable(x):
try:
iter(x)
except TypeError:
return False
else:
return True
| [
"def",
"is_iterable",
"(",
"x",
")",
":",
"try",
":",
"iter",
"(",
"x",
")",
"except",
"TypeError",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | are we being asked to look up a list of things . | train | false |
25,174 | def os_like_constants():
os_like = util.get_systemd_os_like()
if os_like:
for os_name in os_like:
if (os_name in CLI_DEFAULTS.keys()):
return CLI_DEFAULTS[os_name]
return {}
| [
"def",
"os_like_constants",
"(",
")",
":",
"os_like",
"=",
"util",
".",
"get_systemd_os_like",
"(",
")",
"if",
"os_like",
":",
"for",
"os_name",
"in",
"os_like",
":",
"if",
"(",
"os_name",
"in",
"CLI_DEFAULTS",
".",
"keys",
"(",
")",
")",
":",
"return",
... | try to get constants for distribution with similar layout and configuration . | train | false |
25,176 | def is_standard_module(modname, std_path=None):
modname = modname.split('.')[0]
try:
filename = file_from_modpath([modname])
except ImportError:
return False
if (filename is None):
return True
filename = _normalize_path(filename)
if filename.startswith(_cache_normalize_path(EXT_LIB_DIR)):
return False
if (std_path is None):
std_path = STD_LIB_DIRS
for path in std_path:
if filename.startswith(_cache_normalize_path(path)):
return True
return False
| [
"def",
"is_standard_module",
"(",
"modname",
",",
"std_path",
"=",
"None",
")",
":",
"modname",
"=",
"modname",
".",
"split",
"(",
"'.'",
")",
"[",
"0",
"]",
"try",
":",
"filename",
"=",
"file_from_modpath",
"(",
"[",
"modname",
"]",
")",
"except",
"Im... | try to guess if a module is a standard python module :type modname: str . | train | false |
25,177 | def test_bootstrap_ols():
ols_fit = (lambda X, y: np.dot(np.dot(np.linalg.inv(np.dot(X.T, X)), X.T), y))
X = np.column_stack((rs.randn(50, 4), np.ones(50)))
w = [2, 4, 0, 3, 5]
y_noisy = (np.dot(X, w) + (rs.randn(50) * 20))
y_lownoise = (np.dot(X, w) + rs.randn(50))
n_boot = 500
w_boot_noisy = algo.bootstrap(X, y_noisy, n_boot=n_boot, func=ols_fit)
w_boot_lownoise = algo.bootstrap(X, y_lownoise, n_boot=n_boot, func=ols_fit)
assert_equal(w_boot_noisy.shape, (n_boot, 5))
assert_equal(w_boot_lownoise.shape, (n_boot, 5))
nose.tools.assert_greater(w_boot_noisy.std(), w_boot_lownoise.std())
| [
"def",
"test_bootstrap_ols",
"(",
")",
":",
"ols_fit",
"=",
"(",
"lambda",
"X",
",",
"y",
":",
"np",
".",
"dot",
"(",
"np",
".",
"dot",
"(",
"np",
".",
"linalg",
".",
"inv",
"(",
"np",
".",
"dot",
"(",
"X",
".",
"T",
",",
"X",
")",
")",
","... | test bootstrap of ols model fit . | train | false |
25,179 | @contextlib.contextmanager
def report_scope(observation):
current = _reporters[(-1)]
old = current.observation
current.observation = observation
(yield)
current.observation = old
| [
"@",
"contextlib",
".",
"contextmanager",
"def",
"report_scope",
"(",
"observation",
")",
":",
"current",
"=",
"_reporters",
"[",
"(",
"-",
"1",
")",
"]",
"old",
"=",
"current",
".",
"observation",
"current",
".",
"observation",
"=",
"observation",
"(",
"y... | returns a report scope with the current reporter . | train | false |
25,180 | def test_masked_row_with_object_col():
t = table.Table([[1]], dtype=['O'], masked=True)
if NUMPY_LT_1_8:
with pytest.raises(ValueError):
t['col0'].mask = False
t[0].as_void()
with pytest.raises(ValueError):
t['col0'].mask = True
t[0].as_void()
else:
t['col0'].mask = False
assert (t[0]['col0'] == 1)
t['col0'].mask = True
assert (t[0]['col0'] is np.ma.masked)
| [
"def",
"test_masked_row_with_object_col",
"(",
")",
":",
"t",
"=",
"table",
".",
"Table",
"(",
"[",
"[",
"1",
"]",
"]",
",",
"dtype",
"=",
"[",
"'O'",
"]",
",",
"masked",
"=",
"True",
")",
"if",
"NUMPY_LT_1_8",
":",
"with",
"pytest",
".",
"raises",
... | numpy < 1 . | train | false |
25,182 | def get_organization_courses(organization_id):
if (not organizations_enabled()):
return []
from organizations import api as organizations_api
return organizations_api.get_organization_courses(organization_id)
| [
"def",
"get_organization_courses",
"(",
"organization_id",
")",
":",
"if",
"(",
"not",
"organizations_enabled",
"(",
")",
")",
":",
"return",
"[",
"]",
"from",
"organizations",
"import",
"api",
"as",
"organizations_api",
"return",
"organizations_api",
".",
"get_or... | client api operation adapter/wrapper . | train | false |
25,185 | def supportsTeams(endpoint):
return endpoint.usesExtension(ns_uri)
| [
"def",
"supportsTeams",
"(",
"endpoint",
")",
":",
"return",
"endpoint",
".",
"usesExtension",
"(",
"ns_uri",
")"
] | does the given endpoint advertise support for launchpad teams? . | train | false |
25,186 | @pytest.mark.django_db
@pytest.mark.parametrize('view,model,get_object,service_provider_attr', [(PaymentMethodEditView, PaymentMethod, get_default_payment_method, 'payment_processor'), (ShippingMethodEditView, ShippingMethod, get_default_shipping_method, 'carrier')])
def test_behavior_add_save(rf, admin_user, view, model, get_object, service_provider_attr):
get_default_shop()
with override_settings(LANGUAGES=[('en', 'en')]):
object = get_object()
view = view.as_view()
service_provider_attr_field = ('base-%s' % service_provider_attr)
data = get_default_data(object, service_provider_attr, service_provider_attr_field)
components_before = object.behavior_components.count()
assert (not components_before)
request = apply_request_middleware(rf.post('/', data=data, user=admin_user))
view(request, pk=object.pk)
components_after = object.behavior_components.count()
assert (components_after == len(get_default_behavior_settings()))
| [
"@",
"pytest",
".",
"mark",
".",
"django_db",
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'view,model,get_object,service_provider_attr'",
",",
"[",
"(",
"PaymentMethodEditView",
",",
"PaymentMethod",
",",
"get_default_payment_method",
",",
"'payment_processor'... | to make things little bit more simple lets use only english as a language . | train | false |
25,187 | def preprocess_file(filename, cpp_path='cpp', cpp_args=''):
path_list = [cpp_path]
if isinstance(cpp_args, list):
path_list += cpp_args
elif (cpp_args != ''):
path_list += [cpp_args]
path_list += [filename]
try:
pipe = Popen(path_list, stdout=PIPE, universal_newlines=True)
text = pipe.communicate()[0]
except OSError as e:
raise RuntimeError((("Unable to invoke 'cpp'. " + 'Make sure its path was passed correctly\n') + ('Original error: %s' % e)))
return text
| [
"def",
"preprocess_file",
"(",
"filename",
",",
"cpp_path",
"=",
"'cpp'",
",",
"cpp_args",
"=",
"''",
")",
":",
"path_list",
"=",
"[",
"cpp_path",
"]",
"if",
"isinstance",
"(",
"cpp_args",
",",
"list",
")",
":",
"path_list",
"+=",
"cpp_args",
"elif",
"("... | preprocess a file using cpp . | train | false |
25,188 | @cli.command()
@click.option('--username', prompt=True, help="The developer's shown username.")
@click.option('--email', prompt='E-Mail', help="The developer's email address")
@click.password_option(help='The login password.')
@pass_repo
def setuser(repo, username, email, password):
repo.set_config('username', username)
repo.set_config('email', email)
repo.set_config('password', ('*' * len(password)))
click.echo('Changed credentials.')
| [
"@",
"cli",
".",
"command",
"(",
")",
"@",
"click",
".",
"option",
"(",
"'--username'",
",",
"prompt",
"=",
"True",
",",
"help",
"=",
"\"The developer's shown username.\"",
")",
"@",
"click",
".",
"option",
"(",
"'--email'",
",",
"prompt",
"=",
"'E-Mail'",... | sets the user credentials . | train | false |
25,189 | def load_pytest_conf(path, parser):
namespace = {}
exec open(path, 'rb').read() in namespace
if ('update_parser' in namespace):
namespace['update_parser'](parser)
return namespace.get('CustomPyTester', PyTester)
| [
"def",
"load_pytest_conf",
"(",
"path",
",",
"parser",
")",
":",
"namespace",
"=",
"{",
"}",
"exec",
"open",
"(",
"path",
",",
"'rb'",
")",
".",
"read",
"(",
")",
"in",
"namespace",
"if",
"(",
"'update_parser'",
"in",
"namespace",
")",
":",
"namespace"... | loads a pytestconf . | train | false |
25,190 | def gen_lib_options(compiler, library_dirs, runtime_library_dirs, libraries):
lib_opts = []
for dir in library_dirs:
lib_opts.append(compiler.library_dir_option(dir))
for dir in runtime_library_dirs:
opt = compiler.runtime_library_dir_option(dir)
if isinstance(opt, list):
lib_opts = (lib_opts + opt)
else:
lib_opts.append(opt)
for lib in libraries:
(lib_dir, lib_name) = os.path.split(lib)
if lib_dir:
lib_file = compiler.find_library_file([lib_dir], lib_name)
if lib_file:
lib_opts.append(lib_file)
else:
compiler.warn(("no library file corresponding to '%s' found (skipping)" % lib))
else:
lib_opts.append(compiler.library_option(lib))
return lib_opts
| [
"def",
"gen_lib_options",
"(",
"compiler",
",",
"library_dirs",
",",
"runtime_library_dirs",
",",
"libraries",
")",
":",
"lib_opts",
"=",
"[",
"]",
"for",
"dir",
"in",
"library_dirs",
":",
"lib_opts",
".",
"append",
"(",
"compiler",
".",
"library_dir_option",
... | generate linker options for searching library directories and linking with specific libraries . | train | false |
25,191 | def py_encode(symb2freq):
heap = [[wt, [sym, u'']] for (sym, wt) in symb2freq.items()]
heapify(heap)
while (len(heap) > 1):
lo = heappop(heap)
hi = heappop(heap)
for pair in lo[1:]:
pair[1] = (u'0' + pair[1])
for pair in hi[1:]:
pair[1] = (u'1' + pair[1])
heappush(heap, (([(lo[0] + hi[0])] + lo[1:]) + hi[1:]))
return dict(heappop(heap)[1:])
| [
"def",
"py_encode",
"(",
"symb2freq",
")",
":",
"heap",
"=",
"[",
"[",
"wt",
",",
"[",
"sym",
",",
"u''",
"]",
"]",
"for",
"(",
"sym",
",",
"wt",
")",
"in",
"symb2freq",
".",
"items",
"(",
")",
"]",
"heapify",
"(",
"heap",
")",
"while",
"(",
... | huffman encode the given dict mapping symbols to weights from rosetta code . | train | false |
25,193 | def sign_certificate_request(keypair, dn, request, serial, validity_period, digest, start=None, additional_extensions=()):
if (start is None):
start = datetime.datetime.utcnow()
expire = (start + datetime.timedelta(seconds=validity_period))
start = start.strftime('%Y%m%d%H%M%SZ')
expire = expire.strftime('%Y%m%d%H%M%SZ')
req = request.original
cert = crypto.X509()
cert.set_issuer(dn)
cert.set_subject(req.get_subject())
cert.set_pubkey(req.get_pubkey())
cert.set_notBefore(start)
cert.set_notAfter(expire)
cert.set_serial_number(serial)
cert.add_extensions(additional_extensions)
cert.sign(keypair.original, digest)
return Certificate(cert)
| [
"def",
"sign_certificate_request",
"(",
"keypair",
",",
"dn",
",",
"request",
",",
"serial",
",",
"validity_period",
",",
"digest",
",",
"start",
"=",
"None",
",",
"additional_extensions",
"=",
"(",
")",
")",
":",
"if",
"(",
"start",
"is",
"None",
")",
"... | sign a certificaterequest and return a certificate . | train | false |
25,194 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get the repository constructor . | train | false |
25,195 | def remove_profile_images(profile_image_names):
storage = get_profile_image_storage()
for name in profile_image_names.values():
storage.delete(name)
| [
"def",
"remove_profile_images",
"(",
"profile_image_names",
")",
":",
"storage",
"=",
"get_profile_image_storage",
"(",
")",
"for",
"name",
"in",
"profile_image_names",
".",
"values",
"(",
")",
":",
"storage",
".",
"delete",
"(",
"name",
")"
] | physically remove the image files specified in profile_image_names . | train | false |
25,196 | def profile_start(name):
_profiles_running[name] = time.time()
_profiles_stack.append(name)
| [
"def",
"profile_start",
"(",
"name",
")",
":",
"_profiles_running",
"[",
"name",
"]",
"=",
"time",
".",
"time",
"(",
")",
"_profiles_stack",
".",
"append",
"(",
"name",
")"
] | starts a profiling interval with specific name profiling data is sent to the client with next data batch . | train | false |
25,197 | def getLargestCenterOutsetLoopFromLoop(loop, radius, thresholdRatio=0.9):
if (radius == 0.0):
return loop
radius = abs(radius)
points = getPointsFromLoop(loop, radius, thresholdRatio)
centers = getCentersFromPoints(points, (globalIntercircleMultiplier * radius))
largestCenterOutset = None
largestOutsetArea = (-987654321.0)
for center in centers:
outset = getSimplifiedInsetFromClockwiseLoop(center, radius)
if isLargeSameDirection(outset, center, radius):
if (euclidean.isPathInsideLoop(loop, outset) != euclidean.isWiddershins(loop)):
centerOutset = CenterOutset(center, outset)
outsetArea = abs(euclidean.getAreaLoop(outset))
if (outsetArea > largestOutsetArea):
largestOutsetArea = outsetArea
largestCenterOutset = centerOutset
if (largestCenterOutset == None):
return None
largestCenterOutset.center = euclidean.getSimplifiedLoop(largestCenterOutset.center, radius)
return largestCenterOutset
| [
"def",
"getLargestCenterOutsetLoopFromLoop",
"(",
"loop",
",",
"radius",
",",
"thresholdRatio",
"=",
"0.9",
")",
":",
"if",
"(",
"radius",
"==",
"0.0",
")",
":",
"return",
"loop",
"radius",
"=",
"abs",
"(",
"radius",
")",
"points",
"=",
"getPointsFromLoop",
... | get the largest circle outset loop from the loop . | train | false |
25,198 | def copy_data(data, dest, header=None, use_put=None):
ret = None
if use_put:
udata = data
else:
udata = urllib.urlencode(data)
if utils.is_url(dest):
ret = copy_remote(udata, dest, use_put)
if header:
return ret[header]
else:
if header:
ret = (dest + str(time.time()))
dest = (ret + '/_task_result')
copy_local(udata, dest, use_put)
return ret
| [
"def",
"copy_data",
"(",
"data",
",",
"dest",
",",
"header",
"=",
"None",
",",
"use_put",
"=",
"None",
")",
":",
"ret",
"=",
"None",
"if",
"use_put",
":",
"udata",
"=",
"data",
"else",
":",
"udata",
"=",
"urllib",
".",
"urlencode",
"(",
"data",
")"... | copy data to a destination to aid in debugging . | train | false |
25,199 | def _any_pandas_objects(terms):
return any((isinstance(term.value, pd.core.generic.PandasObject) for term in terms))
| [
"def",
"_any_pandas_objects",
"(",
"terms",
")",
":",
"return",
"any",
"(",
"(",
"isinstance",
"(",
"term",
".",
"value",
",",
"pd",
".",
"core",
".",
"generic",
".",
"PandasObject",
")",
"for",
"term",
"in",
"terms",
")",
")"
] | check a sequence of terms for instances of pandasobject . | train | true |
25,200 | def make_dataset(X, y, sample_weight, random_state=None):
rng = check_random_state(random_state)
seed = rng.randint(1, np.iinfo(np.int32).max)
if sp.issparse(X):
dataset = CSRDataset(X.data, X.indptr, X.indices, y, sample_weight, seed=seed)
intercept_decay = SPARSE_INTERCEPT_DECAY
else:
dataset = ArrayDataset(X, y, sample_weight, seed=seed)
intercept_decay = 1.0
return (dataset, intercept_decay)
| [
"def",
"make_dataset",
"(",
"X",
",",
"y",
",",
"sample_weight",
",",
"random_state",
"=",
"None",
")",
":",
"rng",
"=",
"check_random_state",
"(",
"random_state",
")",
"seed",
"=",
"rng",
".",
"randint",
"(",
"1",
",",
"np",
".",
"iinfo",
"(",
"np",
... | create dataset abstraction for sparse and dense inputs . | train | false |
25,201 | def askUser(text, parent=None, help='', defaultno=False, msgfunc=None, title='Anki'):
if (not parent):
parent = aqt.mw.app.activeWindow()
if (not msgfunc):
msgfunc = QMessageBox.question
sb = (QMessageBox.Yes | QMessageBox.No)
if help:
sb |= QMessageBox.Help
while 1:
if defaultno:
default = QMessageBox.No
else:
default = QMessageBox.Yes
r = msgfunc(parent, title, text, sb, default)
if (r == QMessageBox.Help):
openHelp(help)
else:
break
return (r == QMessageBox.Yes)
| [
"def",
"askUser",
"(",
"text",
",",
"parent",
"=",
"None",
",",
"help",
"=",
"''",
",",
"defaultno",
"=",
"False",
",",
"msgfunc",
"=",
"None",
",",
"title",
"=",
"'Anki'",
")",
":",
"if",
"(",
"not",
"parent",
")",
":",
"parent",
"=",
"aqt",
"."... | show a yes/no question . | train | false |
25,203 | def broadcast_to_sharejs(action, sharejs_uuid, node=None, wiki_name='home', data=None):
url = 'http://{host}:{port}/{action}/{id}/'.format(host=wiki_settings.SHAREJS_HOST, port=wiki_settings.SHAREJS_PORT, action=action, id=sharejs_uuid)
if ((action == 'redirect') or (action == 'delete')):
redirect_url = urllib.quote(node.web_url_for('project_wiki_view', wname=wiki_name, _guid=True), safe='')
url = os.path.join(url, redirect_url)
try:
requests.post(url, json=data)
except requests.ConnectionError:
pass
| [
"def",
"broadcast_to_sharejs",
"(",
"action",
",",
"sharejs_uuid",
",",
"node",
"=",
"None",
",",
"wiki_name",
"=",
"'home'",
",",
"data",
"=",
"None",
")",
":",
"url",
"=",
"'http://{host}:{port}/{action}/{id}/'",
".",
"format",
"(",
"host",
"=",
"wiki_settin... | broadcast an action to all documents connected to a wiki . | train | false |
25,204 | def dict_formatter(view, value):
return json.dumps(value, ensure_ascii=False)
| [
"def",
"dict_formatter",
"(",
"view",
",",
"value",
")",
":",
"return",
"json",
".",
"dumps",
"(",
"value",
",",
"ensure_ascii",
"=",
"False",
")"
] | removes unicode entities when displaying dict as string . | train | false |
25,205 | def test_eppstein_matching():
G = nx.Graph()
G.add_nodes_from(['a', 2, 3, 4], bipartite=0)
G.add_nodes_from([1, 'b', 'c'], bipartite=1)
G.add_edges_from([('a', 1), ('a', 'b'), (2, 'b'), (2, 'c'), (3, 'c'), (4, 1)])
matching = eppstein_matching(G)
assert_true((len(matching) == len(maximum_matching(G))))
assert all(((x in set(matching.keys())) for x in set(matching.values())))
| [
"def",
"test_eppstein_matching",
"(",
")",
":",
"G",
"=",
"nx",
".",
"Graph",
"(",
")",
"G",
".",
"add_nodes_from",
"(",
"[",
"'a'",
",",
"2",
",",
"3",
",",
"4",
"]",
",",
"bipartite",
"=",
"0",
")",
"G",
".",
"add_nodes_from",
"(",
"[",
"1",
... | test in accordance to issue #1927 . | train | false |
25,206 | def clashing():
ns = {}
exec_('from sympy import *', ns)
clash1 = {}
clash2 = {}
while ns:
(k, _) = ns.popitem()
if (k in _greek):
clash2[k] = Symbol(k)
_greek.remove(k)
elif (k in _latin):
clash1[k] = Symbol(k)
_latin.remove(k)
clash = {}
clash.update(clash1)
clash.update(clash2)
return (clash1, clash2, clash)
| [
"def",
"clashing",
"(",
")",
":",
"ns",
"=",
"{",
"}",
"exec_",
"(",
"'from sympy import *'",
",",
"ns",
")",
"clash1",
"=",
"{",
"}",
"clash2",
"=",
"{",
"}",
"while",
"ns",
":",
"(",
"k",
",",
"_",
")",
"=",
"ns",
".",
"popitem",
"(",
")",
... | return the clashing-symbols dictionaries . | train | false |
25,207 | @handle_response_format
@treeio_login_required
def item_view(request, folderPath, itemPath, response_format='html'):
try:
item = KnowledgeItem.by_path(folderPath, itemPath)
except KnowledgeItem.DoesNotExist:
raise Http404
if (not item):
raise Http404
items = Object.filter_permitted(manager=KnowledgeItem.objects, user=request.user.profile, mode='r')
if (not request.user.profile.has_permission(item)):
return user_denied(request, message="You don't have access to this Knowledge Item")
context = _get_default_context(request)
context.update({'items': items, 'item': item})
return render_to_response('knowledge/item_view', context, context_instance=RequestContext(request), response_format=response_format)
| [
"@",
"handle_response_format",
"@",
"treeio_login_required",
"def",
"item_view",
"(",
"request",
",",
"folderPath",
",",
"itemPath",
",",
"response_format",
"=",
"'html'",
")",
":",
"try",
":",
"item",
"=",
"KnowledgeItem",
".",
"by_path",
"(",
"folderPath",
","... | item view . | train | false |
25,208 | @before.each_scenario
def process_requires_tags(scenario):
tag_re = re.compile('requires_stub_(?P<server>[^_]+)')
for tag in scenario.tags:
requires = tag_re.match(tag)
if requires:
if (requires.group('server') == 'youtube'):
if (not is_youtube_available(YOUTUBE_API_URLS)):
scenario.steps = []
return
start_stub(requires.group('server'))
| [
"@",
"before",
".",
"each_scenario",
"def",
"process_requires_tags",
"(",
"scenario",
")",
":",
"tag_re",
"=",
"re",
".",
"compile",
"(",
"'requires_stub_(?P<server>[^_]+)'",
")",
"for",
"tag",
"in",
"scenario",
".",
"tags",
":",
"requires",
"=",
"tag_re",
"."... | process the scenario tags to make sure that any requirements are met prior to that scenario being executed . | train | false |
25,209 | def api_prefix(url_prefix=None, api_version=None):
if (url_prefix is None):
url_prefix = config.URL_PREFIX
if (api_version is None):
api_version = config.API_VERSION
prefix = (('/%s' % url_prefix) if url_prefix else '')
version = (('/%s' % api_version) if api_version else '')
return (prefix + version)
| [
"def",
"api_prefix",
"(",
"url_prefix",
"=",
"None",
",",
"api_version",
"=",
"None",
")",
":",
"if",
"(",
"url_prefix",
"is",
"None",
")",
":",
"url_prefix",
"=",
"config",
".",
"URL_PREFIX",
"if",
"(",
"api_version",
"is",
"None",
")",
":",
"api_versio... | returns the prefix to api endpoints . | train | false |
25,212 | def test_human_readable_custom():
f = formatters.HumanReadable()
assert (f(None) == u('\xe2\x88\x85'))
f = formatters.HumanReadable(none_char='/')
assert (f(None) == '/')
| [
"def",
"test_human_readable_custom",
"(",
")",
":",
"f",
"=",
"formatters",
".",
"HumanReadable",
"(",
")",
"assert",
"(",
"f",
"(",
"None",
")",
"==",
"u",
"(",
"'\\xe2\\x88\\x85'",
")",
")",
"f",
"=",
"formatters",
".",
"HumanReadable",
"(",
"none_char",... | test human_readable formatter option . | train | false |
25,213 | def gather_bootstrap_script(bootstrap=None):
if (not HAS_CLOUD):
return (False, 'config.gather_bootstrap_script is unavailable')
ret = salt.utils.cloud.update_bootstrap(__opts__, url=bootstrap)
if (('Success' in ret) and (len(ret['Success']['Files updated']) > 0)):
return ret['Success']['Files updated'][0]
| [
"def",
"gather_bootstrap_script",
"(",
"bootstrap",
"=",
"None",
")",
":",
"if",
"(",
"not",
"HAS_CLOUD",
")",
":",
"return",
"(",
"False",
",",
"'config.gather_bootstrap_script is unavailable'",
")",
"ret",
"=",
"salt",
".",
"utils",
".",
"cloud",
".",
"updat... | download the salt-bootstrap script . | train | true |
25,214 | def float_to_byte(value, mantissabits=5, zeroexp=2):
fzero = ((63 - zeroexp) << mantissabits)
bits = unpack('i', pack('f', value))[0]
smallfloat = (bits >> (24 - mantissabits))
if (smallfloat < fzero):
if (bits <= 0):
result = chr(0)
else:
result = chr(1)
elif (smallfloat >= (fzero + 256)):
result = chr(255)
else:
result = chr((smallfloat - fzero))
return b(result)
| [
"def",
"float_to_byte",
"(",
"value",
",",
"mantissabits",
"=",
"5",
",",
"zeroexp",
"=",
"2",
")",
":",
"fzero",
"=",
"(",
"(",
"63",
"-",
"zeroexp",
")",
"<<",
"mantissabits",
")",
"bits",
"=",
"unpack",
"(",
"'i'",
",",
"pack",
"(",
"'f'",
",",
... | encodes a floating point number in a single byte . | train | false |
25,216 | def test_sigmoid():
def ref_sigmoid(x):
if (x >= 0):
return (1 / (1 + np.exp((- x))))
else:
z = np.exp(x)
return (z / (1 + z))
sigmoid = np.vectorize(ref_sigmoid)
x = K.placeholder(ndim=2)
f = K.function([x], [activations.sigmoid(x)])
test_values = get_standard_values()
result = f([test_values])[0]
expected = sigmoid(test_values)
assert_allclose(result, expected, rtol=1e-05)
| [
"def",
"test_sigmoid",
"(",
")",
":",
"def",
"ref_sigmoid",
"(",
"x",
")",
":",
"if",
"(",
"x",
">=",
"0",
")",
":",
"return",
"(",
"1",
"/",
"(",
"1",
"+",
"np",
".",
"exp",
"(",
"(",
"-",
"x",
")",
")",
")",
")",
"else",
":",
"z",
"=",
... | test using a numerically stable reference sigmoid implementation . | train | false |
25,217 | def generate_tasks(**kwargs):
tasks = _local_tasks()
for (fun, args) in tasks:
fun(*args)
tasks.clear()
| [
"def",
"generate_tasks",
"(",
"**",
"kwargs",
")",
":",
"tasks",
"=",
"_local_tasks",
"(",
")",
"for",
"(",
"fun",
",",
"args",
")",
"in",
"tasks",
":",
"fun",
"(",
"*",
"args",
")",
"tasks",
".",
"clear",
"(",
")"
] | goes through thread local index update tasks set and generates celery tasks for all tasks in the set . | train | false |
25,220 | def get_cursor_position(fd=1):
csbi = get_console_screen_buffer_info(fd=fd)
coord = csbi.dwCursorPosition
return (coord.X, coord.Y)
| [
"def",
"get_cursor_position",
"(",
"fd",
"=",
"1",
")",
":",
"csbi",
"=",
"get_console_screen_buffer_info",
"(",
"fd",
"=",
"fd",
")",
"coord",
"=",
"csbi",
".",
"dwCursorPosition",
"return",
"(",
"coord",
".",
"X",
",",
"coord",
".",
"Y",
")"
] | gets the current cursor position as an tuple . | train | false |
25,222 | def get_local_facts_from_file(filename):
local_facts = dict()
try:
ini_facts = configparser.SafeConfigParser()
ini_facts.read(filename)
for section in ini_facts.sections():
local_facts[section] = dict()
for (key, value) in ini_facts.items(section):
local_facts[section][key] = value
except (configparser.MissingSectionHeaderError, configparser.ParsingError):
try:
with open(filename, 'r') as facts_file:
local_facts = json.load(facts_file)
except (ValueError, IOError):
pass
return local_facts
| [
"def",
"get_local_facts_from_file",
"(",
"filename",
")",
":",
"local_facts",
"=",
"dict",
"(",
")",
"try",
":",
"ini_facts",
"=",
"configparser",
".",
"SafeConfigParser",
"(",
")",
"ini_facts",
".",
"read",
"(",
"filename",
")",
"for",
"section",
"in",
"ini... | retrieve local facts from fact file args: filename : local facts file returns: dict: the retrieved facts . | train | false |
25,223 | def miniEditImages():
return {'Select': BitmapImage(file='/usr/include/X11/bitmaps/left_ptr'), 'Switch': PhotoImage(data='\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay0gBlmtnZ2snJ\nyr+2tAuMu6rY6D6kyfHx8XO/2Uqszjmly6DU5uXz+JLN4uz3+kSrzlKx0ZeZm2K21BuYw67a6QB9\nr+Xl5rW2uHW61On1+UGpzbrf6xiXwny9166vsMLCwgBdlAmHt8TFxgBwpNTs9C2hyO7t7ZnR5L/B\nw0yv0NXV1gBimKGjpABtoQBuoqKkpiaUvqWmqHbB2/j4+Pf39729vgB/sN7w9obH3hSMugCAsonJ\n4M/q8wBglgB6rCCaxLO0tX7C2wBqniGMuABzpuPl5f3+/v39/fr6+r7i7vP6/ABonV621LLc6zWk\nyrq6uq6wskGlyUaszp6gohmYw8HDxKaoqn3E3LGztWGuzcnLzKmrrOnp6gB1qCaex1q001ewz+Dg\n4QB3qrCxstHS09LR0dHR0s7Oz8zNzsfIyQaJuQB0pozL4YzI3re4uAGFtYDG3hOUwb+/wQB5rOvr\n6wB2qdju9TWfxgBpniOcxeLj48vn8dvc3VKuzwB2qp6fos/Q0aXV6D+jxwB7rsXHyLu8vb27vCSc\nxSGZwxyZxH3A2RuUv0+uzz+ozCedxgCDtABnnABroKutr/7+/n2/2LTd6wBvo9bX2OLo6lGv0C6d\nxS6avjmmzLTR2uzr6m651RuXw4jF3CqfxySaxSadyAuRv9bd4cPExRiMuDKjyUWevNPS0sXl8BeY\nxKytr8G/wABypXvC23vD3O73+3vE3cvU2PH5+7S1t7q7vCGVwO/v8JfM3zymyyyZwrWys+Hy90Ki\nxK6qqg+TwBKXxMvMzaWtsK7U4jemzLXEygBxpW++2aCho97Z18bP0/T09fX29vb19ViuzdDR0crf\n51qd01y00ujo6Onq6hCDs2Gpw3i71CqWv3S71nO92M/h52m207bJ0AN6rPPz9Nrh5Nvo7K/b6oTI\n37Td7ABqneHi4yScxo/M4RiWwRqVwcro8n3B2lGoylStzszMzAAAACH5BAEAAP8ALAAAAAAuACAA\nBwj/AP8JHEjw3wEkEY74WOjrQhUNBSNKnCjRSoYKCOwJcKWpEAACBFBRGEKxZMkDjRAg2OBlQyYL\nWhDEcOWxDwofv0zqHIhhDYIFC2p4MYFMS62ZaiYVWlJJAYIqO00KMlEjABYOQokaRbp0CYBKffpE\niDpxSKYC1gqswToUmYVaCFyp6QrgwwcCscaSJZhgQYBeAdRyqFBhgwWkGyct8WoXRZ8Ph/YOxMOB\nCIUAHsBxwGQBAII1YwpMI5Brcd0PKFA4Q2ZFMgYteZqkwxyu1KQNJzQc+CdFCrxypyqdRoEPX6x7\nki/n2TfbAxtNRHYTVCWpWTRbuRoX7yMgZ9QSFQa0/7LU/BXygjIWXVOBTR2sxp7BxGpENgKbY+PR\nreqyIOKnOh0M445AjTjDCgrPSBNFKt9w8wMVU5g0Bg8kDAAKOutQAkNEQNBwDRAEeVEcAV6w84Ay\nKowQSRhmzNGAASIAYow2IP6DySPk8ANKCv1wINE2cpjxCUEgOIOPAKicQMMbKnhyhhg97HDNF4vs\nIEYkNkzwjwSP/PHIE2VIgIdEnxjAiBwNGIKGDKS8I0sw2VAzApNOQimGLlyMAIkDw2yhZTF/KKGE\nlxCEMtEPBtDhACQurLDCLkFIsoUeZLyRpx8OmEGHN3AEcU0HkFAhUDFulDroJvOU5M44iDjgDTQO\n1P/hzRw2IFJPGw3AAY0LI/SAwxc7jEKQI2mkEUipRoxp0g821AMIGlG0McockMzihx5c1LkDDmSg\nUVAiafACRbGPVKDTFG3MYUYdLoThRxDE6DEMGUww8eQONGwTER9piFINFOPasaFJVIjTwC1xzOGP\nA3HUKoIMDTwJR4QRgdBOJzq8UM0Lj5QihU5ZdGMOCSSYUwYzAwwkDhNtUKTBOZ10koMOoohihDwm\nHZKPEDwb4fMe9An0g5Yl+SDKFTHnkMMLLQAjXUTxUCLEIyH0bIQAwuxVQhEMcEIIIUmHUEsWGCQg\nxQEaIFGAHV0+QnUIIWwyg2T/3MPLDQwwcAUhTjiswYsQl1SAxQKmbBJCIMe6ISjVmXwsWQKJEJJE\n3l1/TY8O4wZyh8ZQ3IF4qX9cggTdAmEwCAMs3IB311fsDfbMGv97BxSBQBAP6QMN0QUhLCSRhOp5\ne923zDpk/EIaRdyO+0C/eHBHEiz0vjrrfMfciSKD4LJ8RBEk88IN0ff+O/CEVEPLGK1tH1ECM7Dx\nRDWdcMLJFTpUQ44jfCyjvlShZNDE/0QAgT6ypr6AAAA7\n '), 'LegacySwitch': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXmDjbe4uAE5cjF7xwFWq2Sa0S9biSlrrdTW1k2Ly02a5xUvSQFHjmep\n6bfI2Q5SlQIYLwFfvj6M3Jaan8fHyDuFzwFp0Vah60uU3AEiRhFgrgFRogFr10N9uTFrpytHYQFM\nmGWt9wIwX+bm5kaT4gtFgR1cnJPF9yt80CF0yAIMGHmp2c/P0AEoUb/P4Fei7qK4zgpLjgFkyQlf\nt1mf5jKD1WWJrQ86ZwFAgBhYmVOa4MPV52uv8y+A0iR3ywFbtUyX5ECI0Q1UmwIcOUGQ3RBXoQI0\naRJbpr3BxVeJvQUJDafH5wIlS2aq7xBmv52lr7fH12el5Wml3097ph1ru7vM3HCz91Ke6lid40KQ\n4GSQvgQGClFnfwVJjszMzVCX3hljrdPT1AFLlBRnutPf6yd5zjeI2QE9eRBdrBNVl+3v70mV4ydf\nlwMVKwErVlul8AFChTGB1QE3bsTFxQImTVmAp0FjiUSM1k+b6QQvWQ1SlxMgLgFixEqU3xJhsgFT\npn2Xs5OluZ+1yz1Xb6HN+Td9wy1zuYClykV5r0x2oeDh4qmvt8LDwxhuxRlLfyRioo2124mft9bi\n71mDr7fT79nl8Z2hpQs9b7vN4QMQIOPj5XOPrU2Jx32z6xtvwzeBywFFikFnjwcPFa29yxJjuFmP\nxQFv3qGxwRc/Z8vb6wsRGBNqwqmpqTdvqQIbNQFPngMzZAEfP0mQ13mHlQFYsAFnznOXu2mPtQxj\nvQ1Vn4Ot1+/x8my0/CJgnxNNh8DT5CdJaWyx+AELFWmt8QxPkxBZpwMFB015pgFduGCNuyx7zdnZ\n2WKm6h1xyOPp8aW70QtPkUmM0LrCyr/FyztljwFPm0OJzwFny7/L1xFjswE/e12i50iR2VR8o2Gf\n3xszS2eTvz2BxSlloQdJiwMHDzF3u7bJ3T2I1WCp8+Xt80FokQFJklef6mORw2ap7SJ1y77Q47nN\n3wFfu1Kb5cXJyxdhrdDR0wlNkTSF11Oa4yp4yQEuW0WQ3QIDBQI7dSH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHDjKF6SDvhImPMHwhA6HOiLqUENRDYSLEIplxBcNHz4Z5GTI8BLKS5OBA1Ply2fDhxwf\nPlLITGFmmRkzP+DlVKHCmU9nnz45csSqKKsn9gileZKrVC4aRFACOGZu5UobNuRohRkzhc2b+36o\nqCaqrFmzZEV1ERBg3BOmMl5JZTBhwhm7ZyycYZnvJdeuNl21qkCHTiPDhxspTtKoQgUKCJ6wehMV\n5QctWupeo6TkjOd8e1lmdQkTGbTTMaDFiDGINeskX6YhEicUiQa5A/kUKaFFwQ0oXzjZ8Tbcm3Hj\nirwpMtTSgg9QMJf5WEZ9375AiED19ImpSQSUB4Kw/8HFSMyiRWJaqG/xhf2X91+oCbmq1e/MFD/2\nEcApVkWVJhp8J9AqsywQxDfAbLJJPAy+kMkL8shjxTkUnhOJZ5+JVp8cKfhwxwdf4fQLgG4MFAwW\nKOZRAxM81EAPPQvoE0QQfrDhx4399OMBMjz2yCMVivCoCAWXKLKMTPvoUYcsKwi0RCcwYCAlFjU0\nA6OBM4pXAhsl8FYELYWFWZhiZCbRQgIC2AGTLy408coxAoEDx5wwtGPALTVg0E4NKC7gp4FsBKoA\nKi8U+oIVmVih6DnZPMBMAlGwIARWOLiggSYC+ZNIOulwY4AkSZCyxaikbqHMqaeaIp4+rAaxQxBg\n2P+IozuRzvLZIS4syYVAfMAhwhSC1EPCGoskIIYY9yS7Hny75OFnEIAGyiVvWkjjRxF11fXIG3WU\nKNA6wghDTCW88PKMJZOkm24Z7LarSjPtoIjFn1lKyyVmmBVhwRtvaDDMgFL0Eu4VhaiDwhXCXNFD\nD8QQw7ATEDsBw8RSxotFHs7CKJ60XWrRBj91EOGPQCA48c7J7zTjSTPctOzynjVkkYU+O9S8Axg4\nZ6BzBt30003Ps+AhNB5C4PCGC5gKJMMTZJBRytOl/CH1HxvQkMbVVxujtdZGGKGL17rsEfYQe+xR\nzNnFcGQCv7LsKlAtp8R9Sgd0032BLXjPoPcMffTd3YcEgAMOxOBA1GJ4AYgXAMjiHDTgggveCgRI\n3RfcnffefgcOeDKEG3444osDwgEspMNiTQhx5FoOShxcrrfff0uQjOycD+554qFzMHrpp4cwBju/\n5+CmVNbArnntndeCO+O689777+w0IH0o1P/TRJMohRA4EJwn47nyiocOSOmkn/57COxE3wD11Mfh\nfg45zCGyVF4Ufvvyze8ewv5jQK9++6FwXxzglwM0GPAfR8AeSo4gwAHCbxsQNCAa/kHBAVhwAHPI\n4BE2eIRYeHAEIBwBP0Y4Qn41YWRSCQgAOw==\n '), 'LegacyRouter': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXZ8gQNAgL29vQNctjl/xVSa4j1dfCF+3QFq1DmL3wJMmAMzZZW11dnZ\n2SFrtyNdmTSO6gIZMUKa8gJVqEOHzR9Pf5W74wFjxgFx4jltn+np6Eyi+DuT6qKiohdtwwUPGWiq\n6ymF4LHH3Rh11CV81kKT5AMoUA9dq1ap/mV0gxdXlytRdR1ptRNPjTt9vwNgvwJZsX+69gsXJQFH\njTtjizF0tvHx8VOm9z2V736Dhz2N3QM2acPZ70qe8gFo0HS19wVRnTiR6hMpP0eP1i6J5iNlqAtg\ntktjfQFu3TNxryx4xAMTIzOE1XqAh1uf5SWC4AcfNy1XgQJny93n8a2trRh312Gt+VGm/AQIDTmB\nyAF37QJasydzvxM/ayF3zhdLf8zLywFdu4i56gFlyi2J4yV/1w8wUo2/8j+X8D2Q5Eee9jeR7Uia\n7DpeggFt2QNPm97e3jRong9bpziH2DuT7aipqQoVICmG45vI9R5720eT4Q1hs1er/yVVhwJJktPh\n70tfdbHP7Xev5xs5V7W1sz9jhz11rUVZcQ9WoCVVhQk7cRdtwWuw9QYOFyFHbSBnr0dznxtWkS18\nzKfP9wwcLAMHCwFFiS5UeqGtuRNNiwMfPS1hlQMtWRE5XzGM5yhxusLCwCljnwMdOFWh7cve8pG/\n7Tlxp+Tr8g9bpXF3f0lheStrrYu13QEXLS1ppTV3uUuR1RMjNTF3vU2X4TZupwRSolNne4nB+T+L\n2YGz4zJ/zYe99YGHjRdDcT95sx09XQldsgMLEwMrVc/X3yN3yQ1JhTRbggsdMQNfu9HPz6WlpW2t\n7RctQ0GFyeHh4dvl8SBZklCb5kOO2kWR3Vmt/zdjkQIQHi90uvPz8wIVKBp42SV5zbfT7wtXpStV\nfwFWrBVvyTt3swFz5kGBv2+1/QlbrVFjdQM7d1+j54i67UmX51qn9i1vsy+D2TuR5zddhQsjOR1t\nu0GV6ghbsDVZf4+76RRisent8Xd9hQFBgwFNmwJLlcPDwwFr1z2T5yH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHEiQYJY7Qwg9UsTplRIbENuxEiXJgpcz8e5YKsixY8Essh7JcbbOBwcOa1JOmJAmTY4c\nHeoIabJrCShI0XyB8YRso0eOjoAdWpciBZajJ1GuWcnSZY46Ed5N8hPATqEBoRB9gVJsxRlhPwHI\n0kDkVywcRpGe9LF0adOnMpt8CxDnxg1o9lphKoEACoIvmlxxvHOKVg0n/Tzku2WoVoU2J1P6WNkS\nrtwADuxCG/MOjwgRUEIjGG3FhaOBzaThiDSCil27G8Isc3LLjZwXsA6YYJmDjhTMmseoKQIFDx7R\noxHo2abnwygAlUj1mV6tWjlelEpRwfd6gzI7VeJQ/2vZoVaDUqigqftXpH0R46H9Kl++zUo4JnKq\n9dGvv09RHFhcIUMe0NiFDyql0OJUHWywMc87TXRhhCRGiHAccvNZUR8JxpDTH38p9HEUFhxgMSAv\njbBjQge8PSXEC6uo0IsHA6gAAShmgCbffNtsQwIJifhRHX/TpUUiSijlUk8AqgQixSwdNBjCa7CF\noVggmEgCyRf01WcFCYvYUgB104k4YlK5HONEXXfpokYdMrXRAzMhmNINNNzB9p0T57AgyZckpKKP\nGFNgw06ZWKR10jTw6MAmFWj4AJcQQkQQwSefvFeGCemMIQggeaJywSQ/wgHOAmJskQEfWqBlFBEH\n1P/QaGY3QOpDZXA2+A6m7hl3IRQKGDCIAj6iwE8yGKC6xbJv8IHNHgACQQybN2QiTi5NwdlBpZdi\nisd7vyanByOJ7CMGGRhgwE+qyy47DhnBPLDLEzLIAEQjBtChRmVPNWgpr+Be+Nc9icARww9TkIEu\nDAsQ0O7DzGIQzD2QdDEJHTsIAROc3F7qWQncyHPPHN5QQAAG/vjzw8oKp8sPPxDH3O44/kwBQzLB\nxBCMOTzzHEMMBMBARgJvZJBBEm/4k0ACKydMBgwYoKNNEjJXbTXE42Q9jtFIp8z0Dy1jQMA1AGzi\nz9VoW7310V0znYDTGMQgwUDXLDBO2nhvoTXbbyRk/XXL+pxWkAT8UJ331WsbnbTSK8MggDZhCTOM\nLQkcjvXeSPedAAw0nABWWARZIgEDfyTzxt15Z53BG1PEcEknrvgEelhZMDHKCTwI8EcQFHBBAAFc\ngGPLHwLwcMIo12Qxu0ABAQA7\n '), 'Controller': PhotoImage(data='\n R0lGODlhMAAwAPcAAAEBAWfNAYWFhcfHx+3t6/f390lJUaWlpfPz8/Hx72lpaZGRke/v77m5uc0B\n AeHh4e/v7WNjY3t7e5eXlyMjI4mJidPT0+3t7f///09PT7Ozs/X19fHx8ZWTk8HBwX9/fwAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACH5BAEAAAAALAAAAAAwADAA\n Bwj/AAEIHEiwoMGDCBMqXMiwocOHECNKnEixosWLGAEIeMCxo8ePHwVkBGABg8mTKFOmtDByAIYN\n MGPCRCCzQIENNzEMGOkBAwIKQIMKpYCgKAIHCDB4GNkAA4OnUJ9++CDhQ1QGFzA0GKkBA4GvYMOK\n BYtBA1cNaNOqXcuWq8q3b81m7Cqdbk2bMMu6/Tl0qFEEAZLKxdj1KlSqVA3rnet1rOOwiwmznUzZ\n LdzLJgdfpIv3pmebN2Pm1GyRbocNp1PLNMDaAM3Im1/alQk4gO28pCt2RdCBt+/eRg8IP1AUdmmf\n f5MrL56bYlcOvaP7Xo6Ag3HdGDho3869u/YE1507t+3AgLz58ujPMwg/sTBUCAzgy49PH0LW5u0x\n XFiwvz////5dcJ9bjxVIAHsSdUXAAgs2yOCDDn6FYEQaFGDgYxNCpEFfHHKIX4IDhCjiiCSS+CGF\n FlCmogYpcnVABTDGKGOMAlRQYwUHnKjhAjX2aOOPN8LImgAL6PiQBhLMqCSNAThQgQRGOqRBBD1W\n aaOVAggnQARRNqRBBxmEKeaYZIrZQZcMKbDiigqM5OabcMYp55x01ilnQAA7\n '), 'Host': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAiNAAH8G0iwoMGDCAcKTMiw4UBw\n BPXVm0ixosWLFvVBHFjPoUeC9Tb+6/jRY0iQ/8iVbHiS40CVKxG2\n HEkQZsyCM0mmvGkw50uePUV2tEnOZkyfQA8iTYpTKNOgKJ+C3AhO\n p9SWVaVOfWj1KdauTL9q5UgVbFKsEjGqXVtP40NwcBnCjXtw7tx/\n C8cSBBAQADs=\n '), 'OldSwitch': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAhwAAEIHEiwoMGDCBMqXMiwocOH\n ECNKnEixosWB3zJq3Mixo0eNAL7xG0mypMmTKPl9Cznyn8uWL/m5\n /AeTpsyYI1eKlBnO5r+eLYHy9Ck0J8ubPmPOrMmUpM6UUKMa/Ui1\n 6saLWLNq3cq1q9evYB0GBAA7\n '), 'NetLink': PhotoImage(data='\n R0lGODlhFgAWAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAWABYAAAhIAAEIHEiwoEGBrhIeXEgwoUKG\n Cx0+hGhQoiuKBy1irChxY0GNHgeCDAlgZEiTHlFuVImRJUWXEGEy\n lBmxI8mSNknm1Dnx5sCAADs=\n ')}
| [
"def",
"miniEditImages",
"(",
")",
":",
"return",
"{",
"'Select'",
":",
"BitmapImage",
"(",
"file",
"=",
"'/usr/include/X11/bitmaps/left_ptr'",
")",
",",
"'Switch'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay... | create and return images for miniedit . | train | false |
25,224 | def unpack_ext(ext_path):
modcache = os.path.join(OPTIONS.saltdir, 'running_data', 'var', 'cache', 'salt', 'minion', 'extmods')
tfile = tarfile.TarFile.gzopen(ext_path)
old_umask = os.umask(63)
tfile.extractall(path=modcache)
tfile.close()
os.umask(old_umask)
os.unlink(ext_path)
ver_path = os.path.join(modcache, 'ext_version')
ver_dst = os.path.join(OPTIONS.saltdir, 'ext_version')
shutil.move(ver_path, ver_dst)
| [
"def",
"unpack_ext",
"(",
"ext_path",
")",
":",
"modcache",
"=",
"os",
".",
"path",
".",
"join",
"(",
"OPTIONS",
".",
"saltdir",
",",
"'running_data'",
",",
"'var'",
",",
"'cache'",
",",
"'salt'",
",",
"'minion'",
",",
"'extmods'",
")",
"tfile",
"=",
"... | unpack the external modules . | train | true |
25,225 | def _stringsToObjects(strings, arglist, proto):
objects = {}
myStrings = strings.copy()
for (argname, argparser) in arglist:
argparser.fromBox(argname, myStrings, objects, proto)
return objects
| [
"def",
"_stringsToObjects",
"(",
"strings",
",",
"arglist",
",",
"proto",
")",
":",
"objects",
"=",
"{",
"}",
"myStrings",
"=",
"strings",
".",
"copy",
"(",
")",
"for",
"(",
"argname",
",",
"argparser",
")",
"in",
"arglist",
":",
"argparser",
".",
"fro... | convert an ampbox to a dictionary of python objects . | train | true |
25,226 | def _find_closest_point_on_path(lc, point):
ds = np.sum(((lc - point[None, :]) ** 2), 1)
imin = np.argmin(ds)
dmin = np.inf
xcmin = None
legmin = (None, None)
closed = mlab.is_closed_polygon(lc)
legs = []
if ((imin > 0) or closed):
legs.append((((imin - 1) % len(lc)), imin))
if ((imin < (len(lc) - 1)) or closed):
legs.append((imin, ((imin + 1) % len(lc))))
for leg in legs:
(d, xc) = _find_closest_point_on_leg(lc[leg[0]], lc[leg[1]], point)
if (d < dmin):
dmin = d
xcmin = xc
legmin = leg
return (dmin, xcmin, legmin)
| [
"def",
"_find_closest_point_on_path",
"(",
"lc",
",",
"point",
")",
":",
"ds",
"=",
"np",
".",
"sum",
"(",
"(",
"(",
"lc",
"-",
"point",
"[",
"None",
",",
":",
"]",
")",
"**",
"2",
")",
",",
"1",
")",
"imin",
"=",
"np",
".",
"argmin",
"(",
"d... | lc: coordinates of vertices point: coordinates of test point . | train | false |
25,227 | def flushdb(host=None, port=None, db=None, password=None):
server = _connect(host, port, db, password)
return server.flushdb()
| [
"def",
"flushdb",
"(",
"host",
"=",
"None",
",",
"port",
"=",
"None",
",",
"db",
"=",
"None",
",",
"password",
"=",
"None",
")",
":",
"server",
"=",
"_connect",
"(",
"host",
",",
"port",
",",
"db",
",",
"password",
")",
"return",
"server",
".",
"... | remove all keys from the selected database cli example: . | train | true |
25,228 | def _addHeaderToRequest(request, header):
requestHeaders = request.requestHeaders
(name, value) = header
values = requestHeaders.getRawHeaders(name)
if (values is not None):
values.append(value)
else:
requestHeaders.setRawHeaders(name, [value])
if (name == 'content-length'):
request.gotLength(int(value))
return True
return False
| [
"def",
"_addHeaderToRequest",
"(",
"request",
",",
"header",
")",
":",
"requestHeaders",
"=",
"request",
".",
"requestHeaders",
"(",
"name",
",",
"value",
")",
"=",
"header",
"values",
"=",
"requestHeaders",
".",
"getRawHeaders",
"(",
"name",
")",
"if",
"(",... | add a header tuple to a request header object . | train | false |
25,229 | def project_update(project_id=None, name=None, description=None, enabled=None, profile=None, **connection_args):
auth(profile, **connection_args)
if (_OS_IDENTITY_API_VERSION > 2):
return tenant_update(tenant_id=project_id, name=name, description=description, enabled=enabled, profile=profile, **connection_args)
else:
return False
| [
"def",
"project_update",
"(",
"project_id",
"=",
"None",
",",
"name",
"=",
"None",
",",
"description",
"=",
"None",
",",
"enabled",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"**",
"connection_args",
")",
":",
"auth",
"(",
"profile",
",",
"**",
"co... | update a tenants information the following fields may be updated: name . | train | true |
25,230 | def v7(method_v7):
frame = currentframe().f_back
return frame.f_locals.get(method_v7.__name__, method_v7)
| [
"def",
"v7",
"(",
"method_v7",
")",
":",
"frame",
"=",
"currentframe",
"(",
")",
".",
"f_back",
"return",
"frame",
".",
"f_locals",
".",
"get",
"(",
"method_v7",
".",
"__name__",
",",
"method_v7",
")"
] | decorate a method that supports the old-style api only . | train | false |
25,232 | def format_jnlp_file_lines(web_flag, url, tep_fp):
lines = [jnlp_top_block]
if web_flag:
lines += ['http://topiaryexplorer.sourceforge.net/app/']
else:
topiaryexplorer_project_dir = load_qiime_config()['topiaryexplorer_project_dir']
if topiaryexplorer_project_dir:
lines += [('file:' + topiaryexplorer_project_dir)]
else:
print "WARNING: Couldn't create jnlp file - topiaryexplorer_project_dir is not defined in your qiime_config. tep file was created sucessfully."
lines += [jnlp_middle_block]
if url:
lines += [url]
else:
lines += [abspath(tep_fp)]
lines += [jnlp_bottom_block]
return lines
| [
"def",
"format_jnlp_file_lines",
"(",
"web_flag",
",",
"url",
",",
"tep_fp",
")",
":",
"lines",
"=",
"[",
"jnlp_top_block",
"]",
"if",
"web_flag",
":",
"lines",
"+=",
"[",
"'http://topiaryexplorer.sourceforge.net/app/'",
"]",
"else",
":",
"topiaryexplorer_project_di... | format the jnlp file for topiaryexplorer . | train | false |
25,233 | @core_helper
def sorted_extras(package_extras, auto_clean=False, subs=None, exclude=None):
if (not exclude):
exclude = config.get('package_hide_extras', [])
output = []
for extra in sorted(package_extras, key=(lambda x: x['key'])):
if (extra.get('state') == 'deleted'):
continue
(k, v) = (extra['key'], extra['value'])
if (k in exclude):
continue
if (subs and (k in subs)):
k = subs[k]
elif auto_clean:
k = k.replace('_', ' ').replace('-', ' ').title()
if isinstance(v, (list, tuple)):
v = ', '.join(map(unicode, v))
output.append((k, v))
return output
| [
"@",
"core_helper",
"def",
"sorted_extras",
"(",
"package_extras",
",",
"auto_clean",
"=",
"False",
",",
"subs",
"=",
"None",
",",
"exclude",
"=",
"None",
")",
":",
"if",
"(",
"not",
"exclude",
")",
":",
"exclude",
"=",
"config",
".",
"get",
"(",
"'pac... | used for outputting package extras . | train | false |
25,236 | def getitem(a, b):
return a[b]
| [
"def",
"getitem",
"(",
"a",
",",
"b",
")",
":",
"return",
"a",
"[",
"b",
"]"
] | get item from an iterable at index . | train | false |
25,237 | def normalize_spaces(s):
return ' '.join(s.split())
| [
"def",
"normalize_spaces",
"(",
"s",
")",
":",
"return",
"' '",
".",
"join",
"(",
"s",
".",
"split",
"(",
")",
")"
] | replace any sequence of whitespace characters with a single space . | train | false |
25,238 | def test_solve_rational():
assert (solve(((x - (y ** 3)) / ((y ** 2) * sqrt((1 - (y ** 2))))), x) == [(y ** 3)])
| [
"def",
"test_solve_rational",
"(",
")",
":",
"assert",
"(",
"solve",
"(",
"(",
"(",
"x",
"-",
"(",
"y",
"**",
"3",
")",
")",
"/",
"(",
"(",
"y",
"**",
"2",
")",
"*",
"sqrt",
"(",
"(",
"1",
"-",
"(",
"y",
"**",
"2",
")",
")",
")",
")",
"... | test solve for rational functions . | train | false |
25,239 | def backup_create(context, values):
return IMPL.backup_create(context, values)
| [
"def",
"backup_create",
"(",
"context",
",",
"values",
")",
":",
"return",
"IMPL",
".",
"backup_create",
"(",
"context",
",",
"values",
")"
] | create a backup from the values dictionary . | train | false |
25,240 | def _delete_asset(course_key, asset_key_string):
if asset_key_string:
try:
asset_key = AssetKey.from_string(asset_key_string)
except InvalidKeyError:
if ('/' == asset_key_string[0]):
asset_key_string = asset_key_string[1:]
try:
asset_key = AssetKey.from_string(asset_key_string)
except InvalidKeyError:
LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string)
return
else:
LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string)
return
try:
delete_asset(course_key, asset_key)
except AssetNotFoundException:
pass
| [
"def",
"_delete_asset",
"(",
"course_key",
",",
"asset_key_string",
")",
":",
"if",
"asset_key_string",
":",
"try",
":",
"asset_key",
"=",
"AssetKey",
".",
"from_string",
"(",
"asset_key_string",
")",
"except",
"InvalidKeyError",
":",
"if",
"(",
"'/'",
"==",
"... | internal method used to create asset key from string and remove asset by calling delete_asset method of assets module . | train | false |
25,241 | def no_os_popen(logical_line):
if ('os.popen(' in logical_line):
(yield (0, 'N348 Deprecated library function os.popen(). Replace it using subprocess module. '))
| [
"def",
"no_os_popen",
"(",
"logical_line",
")",
":",
"if",
"(",
"'os.popen('",
"in",
"logical_line",
")",
":",
"(",
"yield",
"(",
"0",
",",
"'N348 Deprecated library function os.popen(). Replace it using subprocess module. '",
")",
")"
] | disallow os . | train | false |
25,243 | def instance_metadata_update(context, instance_uuid, metadata, delete):
return IMPL.instance_metadata_update(context, instance_uuid, metadata, delete)
| [
"def",
"instance_metadata_update",
"(",
"context",
",",
"instance_uuid",
",",
"metadata",
",",
"delete",
")",
":",
"return",
"IMPL",
".",
"instance_metadata_update",
"(",
"context",
",",
"instance_uuid",
",",
"metadata",
",",
"delete",
")"
] | update metadata if it exists . | train | false |
25,244 | @utils.arg('fixed_ip', metavar='<fixed_ip>', help=_('Fixed IP Address.'))
@deprecated_network
def do_fixed_ip_reserve(cs, args):
cs.fixed_ips.reserve(args.fixed_ip)
| [
"@",
"utils",
".",
"arg",
"(",
"'fixed_ip'",
",",
"metavar",
"=",
"'<fixed_ip>'",
",",
"help",
"=",
"_",
"(",
"'Fixed IP Address.'",
")",
")",
"@",
"deprecated_network",
"def",
"do_fixed_ip_reserve",
"(",
"cs",
",",
"args",
")",
":",
"cs",
".",
"fixed_ips"... | reserve a fixed ip . | train | false |
25,245 | def getWrappedInteger(integer, modulo):
if (integer >= modulo):
return modulo
if (integer >= 0):
return integer
return (integer % modulo)
| [
"def",
"getWrappedInteger",
"(",
"integer",
",",
"modulo",
")",
":",
"if",
"(",
"integer",
">=",
"modulo",
")",
":",
"return",
"modulo",
"if",
"(",
"integer",
">=",
"0",
")",
":",
"return",
"integer",
"return",
"(",
"integer",
"%",
"modulo",
")"
] | get wrapped integer . | train | false |
25,246 | def Std(xs, mu=None, ddof=0):
var = Var(xs, mu, ddof)
return math.sqrt(var)
| [
"def",
"Std",
"(",
"xs",
",",
"mu",
"=",
"None",
",",
"ddof",
"=",
"0",
")",
":",
"var",
"=",
"Var",
"(",
"xs",
",",
"mu",
",",
"ddof",
")",
"return",
"math",
".",
"sqrt",
"(",
"var",
")"
] | computes standard deviation . | train | false |
25,247 | @hook.command()
def shorten(text):
args = text.split()
url = args[0]
custom = (args[1] if (len(args) > 1) else None)
try:
return web.shorten(url, custom=custom)
except web.ServiceError as e:
return e.message
| [
"@",
"hook",
".",
"command",
"(",
")",
"def",
"shorten",
"(",
"text",
")",
":",
"args",
"=",
"text",
".",
"split",
"(",
")",
"url",
"=",
"args",
"[",
"0",
"]",
"custom",
"=",
"(",
"args",
"[",
"1",
"]",
"if",
"(",
"len",
"(",
"args",
")",
"... | smartly shorten a given string . | train | false |
25,248 | def clientresp(data, **kwargs):
creds = config.get_credentials()
cfg = config.get_config()
dumps_kwargs = {'sort_keys': True, 'cls': utils.PlotlyJSONEncoder}
payload = {'platform': 'python', 'version': version.__version__, 'args': _json.dumps(data, **dumps_kwargs), 'un': creds['username'], 'key': creds['api_key'], 'origin': 'plot', 'kwargs': _json.dumps(kwargs, **dumps_kwargs)}
url = '{plotly_domain}/clientresp'.format(**cfg)
response = request('post', url, data=payload)
parsed_content = response.json()
if parsed_content.get('warning'):
warnings.warn(parsed_content['warning'])
if parsed_content.get('message'):
print parsed_content['message']
return response
| [
"def",
"clientresp",
"(",
"data",
",",
"**",
"kwargs",
")",
":",
"creds",
"=",
"config",
".",
"get_credentials",
"(",
")",
"cfg",
"=",
"config",
".",
"get_config",
"(",
")",
"dumps_kwargs",
"=",
"{",
"'sort_keys'",
":",
"True",
",",
"'cls'",
":",
"util... | deprecated endpoint . | train | false |
25,250 | def _check_header_magic_bytes(file_obj):
file_obj.seek(0, 0)
magic = file_obj.read(4)
return (magic == 'PAR1')
| [
"def",
"_check_header_magic_bytes",
"(",
"file_obj",
")",
":",
"file_obj",
".",
"seek",
"(",
"0",
",",
"0",
")",
"magic",
"=",
"file_obj",
".",
"read",
"(",
"4",
")",
"return",
"(",
"magic",
"==",
"'PAR1'",
")"
] | check if the file-like obj has the par1 magic bytes at the header . | train | false |
25,252 | def cliques_containing_node(G, nodes=None, cliques=None):
if (cliques is None):
cliques = list(find_cliques(G))
if (nodes is None):
nodes = list(G.nodes())
if (not isinstance(nodes, list)):
v = nodes
vcliques = [c for c in cliques if (v in c)]
else:
vcliques = {}
for v in nodes:
vcliques[v] = [c for c in cliques if (v in c)]
return vcliques
| [
"def",
"cliques_containing_node",
"(",
"G",
",",
"nodes",
"=",
"None",
",",
"cliques",
"=",
"None",
")",
":",
"if",
"(",
"cliques",
"is",
"None",
")",
":",
"cliques",
"=",
"list",
"(",
"find_cliques",
"(",
"G",
")",
")",
"if",
"(",
"nodes",
"is",
"... | returns a list of cliques containing the given node . | train | false |
25,253 | def _add_retry_host(filter_properties, host, node):
retry = filter_properties.get('retry', None)
if (not retry):
return
hosts = retry['hosts']
hosts.append([host, node])
| [
"def",
"_add_retry_host",
"(",
"filter_properties",
",",
"host",
",",
"node",
")",
":",
"retry",
"=",
"filter_properties",
".",
"get",
"(",
"'retry'",
",",
"None",
")",
"if",
"(",
"not",
"retry",
")",
":",
"return",
"hosts",
"=",
"retry",
"[",
"'hosts'",... | add a retry entry for the selected compute node . | train | false |
25,254 | def mirror_y(clip, apply_to='mask'):
return clip.fl_image((lambda f: f[::(-1)]), apply_to=apply_to)
| [
"def",
"mirror_y",
"(",
"clip",
",",
"apply_to",
"=",
"'mask'",
")",
":",
"return",
"clip",
".",
"fl_image",
"(",
"(",
"lambda",
"f",
":",
"f",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
")",
",",
"apply_to",
"=",
"apply_to",
")"
] | flips the clip vertically . | train | false |
25,256 | def make_derivative_operator(M, z):
def doit(C):
r = ((z * C.diff(z)) + (C * M))
r = r.applyfunc(make_simp(z))
return r
return doit
| [
"def",
"make_derivative_operator",
"(",
"M",
",",
"z",
")",
":",
"def",
"doit",
"(",
"C",
")",
":",
"r",
"=",
"(",
"(",
"z",
"*",
"C",
".",
"diff",
"(",
"z",
")",
")",
"+",
"(",
"C",
"*",
"M",
")",
")",
"r",
"=",
"r",
".",
"applyfunc",
"(... | create a derivative operator . | train | false |
25,257 | def hmc_move(s_rng, positions, energy_fn, stepsize, n_steps):
initial_vel = s_rng.normal(size=positions.shape)
(final_pos, final_vel) = simulate_dynamics(initial_pos=positions, initial_vel=initial_vel, stepsize=stepsize, n_steps=n_steps, energy_fn=energy_fn)
accept = metropolis_hastings_accept(energy_prev=hamiltonian(positions, initial_vel, energy_fn), energy_next=hamiltonian(final_pos, final_vel, energy_fn), s_rng=s_rng)
return (accept, final_pos)
| [
"def",
"hmc_move",
"(",
"s_rng",
",",
"positions",
",",
"energy_fn",
",",
"stepsize",
",",
"n_steps",
")",
":",
"initial_vel",
"=",
"s_rng",
".",
"normal",
"(",
"size",
"=",
"positions",
".",
"shape",
")",
"(",
"final_pos",
",",
"final_vel",
")",
"=",
... | this function performs one-step of hybrid monte-carlo sampling . | train | false |
25,258 | def _zpklp2lp(z, p, k, wo=1):
degree = _relative_degree(z, p)
z_lp = [(wo * z1) for z1 in z]
p_lp = [(wo * p1) for p1 in p]
k_lp = (k * (wo ** degree))
return (z_lp, p_lp, k_lp)
| [
"def",
"_zpklp2lp",
"(",
"z",
",",
"p",
",",
"k",
",",
"wo",
"=",
"1",
")",
":",
"degree",
"=",
"_relative_degree",
"(",
"z",
",",
"p",
")",
"z_lp",
"=",
"[",
"(",
"wo",
"*",
"z1",
")",
"for",
"z1",
"in",
"z",
"]",
"p_lp",
"=",
"[",
"(",
... | transform a lowpass filter prototype to a different frequency . | train | false |
25,260 | def __get_tags(vm_):
t = config.get_cloud_config_value('tags', vm_, __opts__, default='[]', search_global=False)
try:
tags = literal_eval(t)
except Exception:
tags = None
if ((not tags) or (not isinstance(tags, list))):
tags = None
return tags
| [
"def",
"__get_tags",
"(",
"vm_",
")",
":",
"t",
"=",
"config",
".",
"get_cloud_config_value",
"(",
"'tags'",
",",
"vm_",
",",
"__opts__",
",",
"default",
"=",
"'[]'",
",",
"search_global",
"=",
"False",
")",
"try",
":",
"tags",
"=",
"literal_eval",
"(",
... | get configured tags . | train | true |
25,261 | def reset():
_runtime.reset()
| [
"def",
"reset",
"(",
")",
":",
"_runtime",
".",
"reset",
"(",
")"
] | interface to git-reset(1)_ . | train | false |
25,262 | def test_allknn_sk_estimator():
check_estimator(AllKNN)
| [
"def",
"test_allknn_sk_estimator",
"(",
")",
":",
"check_estimator",
"(",
"AllKNN",
")"
] | test the sklearn estimator compatibility . | train | false |
25,263 | def compareSegmentLength(endpoint, otherEndpoint):
if (endpoint.segmentLength > otherEndpoint.segmentLength):
return 1
if (endpoint.segmentLength < otherEndpoint.segmentLength):
return (-1)
return 0
| [
"def",
"compareSegmentLength",
"(",
"endpoint",
",",
"otherEndpoint",
")",
":",
"if",
"(",
"endpoint",
".",
"segmentLength",
">",
"otherEndpoint",
".",
"segmentLength",
")",
":",
"return",
"1",
"if",
"(",
"endpoint",
".",
"segmentLength",
"<",
"otherEndpoint",
... | get comparison in order to sort endpoints in ascending order of segment length . | train | false |
25,265 | def _stdin_ready_other():
return True
| [
"def",
"_stdin_ready_other",
"(",
")",
":",
"return",
"True"
] | return true . | train | false |
25,267 | def get_bind_data(zone_id, profile):
conn = _get_driver(profile=profile)
zone = conn.get_zone(zone_id)
return conn.export_zone_to_bind_format(zone)
| [
"def",
"get_bind_data",
"(",
"zone_id",
",",
"profile",
")",
":",
"conn",
"=",
"_get_driver",
"(",
"profile",
"=",
"profile",
")",
"zone",
"=",
"conn",
".",
"get_zone",
"(",
"zone_id",
")",
"return",
"conn",
".",
"export_zone_to_bind_format",
"(",
"zone",
... | export zone to the bind compatible format . | train | true |
25,269 | def manage_entries(options, selection, approved):
approved_text = (u'approved' if approved else u'pending')
with Session() as session:
if (selection == u'all'):
entries = list_pending_entries(session=session, approved=(not approved))
else:
try:
entry = get_entry_by_id(session, selection)
if (entry.approved is approved):
console((colorize(u'red', u'ERROR: ') + (u'Entry with ID %s is already %s' % (entry.id, approved_text))))
sys.exit(1)
except NoResultFound:
console((u'Pending entry with ID %s does not exist' % selection))
sys.exit(1)
else:
entries = [entry]
if (not entries):
console((u'All entries are already %s' % approved_text))
return
for entry in entries:
if (entry.approved is not approved):
console((u'Setting pending entry with ID %s status to %s' % (entry.id, approved_text)))
entry.approved = approved
| [
"def",
"manage_entries",
"(",
"options",
",",
"selection",
",",
"approved",
")",
":",
"approved_text",
"=",
"(",
"u'approved'",
"if",
"approved",
"else",
"u'pending'",
")",
"with",
"Session",
"(",
")",
"as",
"session",
":",
"if",
"(",
"selection",
"==",
"u... | manage pending entries . | train | false |
25,270 | def inthread(func):
@wraps(func)
def wrapped(*a, **kw):
return threads.deferToThread(func, *a, **kw)
return wrapped
| [
"def",
"inthread",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"wrapped",
"(",
"*",
"a",
",",
"**",
"kw",
")",
":",
"return",
"threads",
".",
"deferToThread",
"(",
"func",
",",
"*",
"a",
",",
"**",
"kw",
")",
"return",
"wrapped"... | decorator to call a function in a thread and return a deferred with the result . | train | false |
25,272 | def validateMasterOption(master):
try:
(hostname, port) = master.split(':')
port = int(port)
except (TypeError, ValueError):
raise usage.UsageError("master must have the form 'hostname:port'")
| [
"def",
"validateMasterOption",
"(",
"master",
")",
":",
"try",
":",
"(",
"hostname",
",",
"port",
")",
"=",
"master",
".",
"split",
"(",
"':'",
")",
"port",
"=",
"int",
"(",
"port",
")",
"except",
"(",
"TypeError",
",",
"ValueError",
")",
":",
"raise... | validate master command line option . | train | true |
25,273 | def is_cnf(expr):
return _is_form(expr, And, Or)
| [
"def",
"is_cnf",
"(",
"expr",
")",
":",
"return",
"_is_form",
"(",
"expr",
",",
"And",
",",
"Or",
")"
] | test whether or not an expression is in conjunctive normal form . | train | false |
25,275 | def cachedir_index_del(minion_id, base=None):
base = init_cachedir(base)
index_file = os.path.join(base, 'index.p')
lock_file(index_file)
if os.path.exists(index_file):
with salt.utils.fopen(index_file, 'r') as fh_:
index = msgpack.load(fh_)
else:
return
if (minion_id in index):
del index[minion_id]
with salt.utils.fopen(index_file, 'w') as fh_:
msgpack.dump(index, fh_)
unlock_file(index_file)
| [
"def",
"cachedir_index_del",
"(",
"minion_id",
",",
"base",
"=",
"None",
")",
":",
"base",
"=",
"init_cachedir",
"(",
"base",
")",
"index_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"base",
",",
"'index.p'",
")",
"lock_file",
"(",
"index_file",
")",... | delete an entry from the cachedir index . | train | false |
25,276 | def config(conf_path='/etc/rsyncd.conf'):
ret = ''
try:
with salt.utils.fopen(conf_path, 'r') as fp_:
for line in fp_:
ret += line
except IOError as exc:
if (exc.errno == errno.ENOENT):
raise CommandExecutionError('{0} does not exist'.format(conf_path))
elif (exc.errno == errno.EACCES):
raise CommandExecutionError('Unable to read {0}, access denied'.format(conf_path))
elif (exc.errno == errno.EISDIR):
raise CommandExecutionError('Unable to read {0}, path is a directory'.format(conf_path))
else:
raise CommandExecutionError('Error {0}: {1}'.format(exc.errno, exc.strerror))
else:
return ret
| [
"def",
"config",
"(",
"conf_path",
"=",
"'/etc/rsyncd.conf'",
")",
":",
"ret",
"=",
"''",
"try",
":",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"conf_path",
",",
"'r'",
")",
"as",
"fp_",
":",
"for",
"line",
"in",
"fp_",
":",
"ret",
"+=",
"li... | set traffic server configuration variable values . | train | true |
25,277 | def _createPluginDummy(entrypath, pluginContent, real, pluginModule):
entrypath.createDirectory()
pkg = entrypath.child('plugindummy')
pkg.createDirectory()
if real:
pkg.child('__init__.py').setContent('')
plugs = pkg.child('plugins')
plugs.createDirectory()
if real:
plugs.child('__init__.py').setContent(pluginInitFile)
plugs.child((pluginModule + '.py')).setContent(pluginContent)
return plugs
| [
"def",
"_createPluginDummy",
"(",
"entrypath",
",",
"pluginContent",
",",
"real",
",",
"pluginModule",
")",
":",
"entrypath",
".",
"createDirectory",
"(",
")",
"pkg",
"=",
"entrypath",
".",
"child",
"(",
"'plugindummy'",
")",
"pkg",
".",
"createDirectory",
"("... | create a plugindummy package . | train | false |
25,278 | def one(method):
def loop(method, self, *args, **kwargs):
result = [method(rec, *args, **kwargs) for rec in self]
return aggregate(method, result, self)
wrapper = decorator(loop, method)
wrapper._api = 'one'
return wrapper
| [
"def",
"one",
"(",
"method",
")",
":",
"def",
"loop",
"(",
"method",
",",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"result",
"=",
"[",
"method",
"(",
"rec",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"for",
"rec",
"in",
"self",... | one returns a single game matching the given criteria . | train | false |
25,279 | def contains_exact_files(directory, expected_files, ignore_links=False):
return (sorted(expected_files) == sorted(exact_files(directory, ignore_links=ignore_links)))
| [
"def",
"contains_exact_files",
"(",
"directory",
",",
"expected_files",
",",
"ignore_links",
"=",
"False",
")",
":",
"return",
"(",
"sorted",
"(",
"expected_files",
")",
"==",
"sorted",
"(",
"exact_files",
"(",
"directory",
",",
"ignore_links",
"=",
"ignore_link... | check if the only files which directory contains are expected_files . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.