id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
16,432
def notepadplusplus(exe=u'notepad++'): install_editor((exe + u' -n{line} {filename}'))
[ "def", "notepadplusplus", "(", "exe", "=", "u'notepad++'", ")", ":", "install_editor", "(", "(", "exe", "+", "u' -n{line} {filename}'", ")", ")" ]
notepad++ URL .
train
false
16,434
def _divide_side(lobe, x): lobe = np.asarray(lobe) median = np.median(x[lobe]) left = lobe[np.where((x[lobe] < median))[0]] right = lobe[np.where((x[lobe] > median))[0]] medians = np.where((x[lobe] == median))[0] left = np.sort(np.concatenate([left, lobe[medians[1::2]]])) right = np.sort(np.concatenate([right, lobe[medians[::2]]])) return (list(left), list(right))
[ "def", "_divide_side", "(", "lobe", ",", "x", ")", ":", "lobe", "=", "np", ".", "asarray", "(", "lobe", ")", "median", "=", "np", ".", "median", "(", "x", "[", "lobe", "]", ")", "left", "=", "lobe", "[", "np", ".", "where", "(", "(", "x", "["...
helper for making a separation between left and right lobe evenly .
train
false
16,436
def resolve_cert_reqs(candidate): if (candidate is None): return CERT_NONE if isinstance(candidate, str): res = getattr(ssl, candidate, None) if (res is None): res = getattr(ssl, ('CERT_' + candidate)) return res return candidate
[ "def", "resolve_cert_reqs", "(", "candidate", ")", ":", "if", "(", "candidate", "is", "None", ")", ":", "return", "CERT_NONE", "if", "isinstance", "(", "candidate", ",", "str", ")", ":", "res", "=", "getattr", "(", "ssl", ",", "candidate", ",", "None", ...
resolves the argument to a numeric constant .
train
true
16,437
def UploadUnconvertedFileSample(): client = CreateClient() doc = gdata.docs.data.Resource(type='document', title='My Sample Raw Doc') path = _GetDataFilePath('test.0.doc') media = gdata.data.MediaSource() media.SetFileHandle(path, 'application/msword') create_uri = (gdata.docs.client.RESOURCE_UPLOAD_URI + '?convert=false') doc = client.CreateResource(doc, create_uri=create_uri, media=media) print 'Created, and uploaded:', doc.title.text, doc.resource_id.text
[ "def", "UploadUnconvertedFileSample", "(", ")", ":", "client", "=", "CreateClient", "(", ")", "doc", "=", "gdata", ".", "docs", ".", "data", ".", "Resource", "(", "type", "=", "'document'", ",", "title", "=", "'My Sample Raw Doc'", ")", "path", "=", "_GetD...
upload a document .
train
false
16,440
def sync_languages(): with open(frappe.get_app_path(u'frappe', u'geo', u'languages.json'), u'r') as f: data = json.loads(f.read()) for l in data: if (not frappe.db.exists(u'Language', l[u'code'])): frappe.get_doc({u'doctype': u'Language', u'language_code': l[u'code'], u'language_name': l[u'name']}).insert()
[ "def", "sync_languages", "(", ")", ":", "with", "open", "(", "frappe", ".", "get_app_path", "(", "u'frappe'", ",", "u'geo'", ",", "u'languages.json'", ")", ",", "u'r'", ")", "as", "f", ":", "data", "=", "json", ".", "loads", "(", "f", ".", "read", "(...
sync frappe/geo/languages .
train
false
16,443
def _AddIOSDeviceConfigurations(targets): for target_dict in targets.itervalues(): toolset = target_dict['toolset'] configs = target_dict['configurations'] for (config_name, config_dict) in dict(configs).iteritems(): iphoneos_config_dict = copy.deepcopy(config_dict) configs[(config_name + '-iphoneos')] = iphoneos_config_dict configs[(config_name + '-iphonesimulator')] = config_dict if (toolset == 'target'): iphoneos_config_dict['xcode_settings']['SDKROOT'] = 'iphoneos' return targets
[ "def", "_AddIOSDeviceConfigurations", "(", "targets", ")", ":", "for", "target_dict", "in", "targets", ".", "itervalues", "(", ")", ":", "toolset", "=", "target_dict", "[", "'toolset'", "]", "configs", "=", "target_dict", "[", "'configurations'", "]", "for", "...
clone all targets and append -iphoneos to the name .
train
false
16,444
def test_error_handling(): class TestException(Exception, ): pass try: with cd('somewhere'): raise TestException('Houston, we have a problem.') except TestException: pass finally: with cd('else'): eq_(env.cwd, 'else')
[ "def", "test_error_handling", "(", ")", ":", "class", "TestException", "(", "Exception", ",", ")", ":", "pass", "try", ":", "with", "cd", "(", "'somewhere'", ")", ":", "raise", "TestException", "(", "'Houston, we have a problem.'", ")", "except", "TestException"...
cd cleans up after itself even in case of an exception .
train
false
16,447
def GetIndexedTimeZoneNames(index_key='Index'): for timeZoneName in GetTimeZoneNames(): tzRegKeyPath = os.path.join(TimeZoneInfo.tzRegKey, timeZoneName) key = _winreg.OpenKeyEx(_winreg.HKEY_LOCAL_MACHINE, tzRegKeyPath) (tzIndex, type) = _winreg.QueryValueEx(key, index_key) (yield (tzIndex, timeZoneName))
[ "def", "GetIndexedTimeZoneNames", "(", "index_key", "=", "'Index'", ")", ":", "for", "timeZoneName", "in", "GetTimeZoneNames", "(", ")", ":", "tzRegKeyPath", "=", "os", ".", "path", ".", "join", "(", "TimeZoneInfo", ".", "tzRegKey", ",", "timeZoneName", ")", ...
returns the names of the time zones as defined in the registry .
train
false
16,448
def styleof(expr, styles=default_styles): style = dict() for (typ, sty) in styles: if isinstance(expr, typ): style.update(sty) return style
[ "def", "styleof", "(", "expr", ",", "styles", "=", "default_styles", ")", ":", "style", "=", "dict", "(", ")", "for", "(", "typ", ",", "sty", ")", "in", "styles", ":", "if", "isinstance", "(", "expr", ",", "typ", ")", ":", "style", ".", "update", ...
merge style dictionaries in order .
train
false
16,449
def setup_students_and_grades(context): if context.course: context.student = student = UserFactory.create() CourseEnrollmentFactory.create(user=student, course_id=context.course.id) context.student2 = student2 = UserFactory.create() CourseEnrollmentFactory.create(user=student2, course_id=context.course.id) for chapter in context.course.get_children(): for (i, section) in enumerate(chapter.get_children()): for (j, problem) in enumerate(section.get_children()): StudentModuleFactory.create(grade=(1 if (i < j) else 0), max_grade=1, student=context.student, course_id=context.course.id, module_state_key=problem.location) StudentModuleFactory.create(grade=(1 if (i > j) else 0), max_grade=1, student=context.student2, course_id=context.course.id, module_state_key=problem.location)
[ "def", "setup_students_and_grades", "(", "context", ")", ":", "if", "context", ".", "course", ":", "context", ".", "student", "=", "student", "=", "UserFactory", ".", "create", "(", ")", "CourseEnrollmentFactory", ".", "create", "(", "user", "=", "student", ...
create students and set their grades .
train
false
16,450
@db_api.api_context_manager.reader def _refresh_from_db(ctx, cache): with db_api.api_context_manager.reader.connection.using(ctx) as conn: sel = sa.select([_RC_TBL.c.id, _RC_TBL.c.name]) res = conn.execute(sel).fetchall() cache.id_cache = {r[1]: r[0] for r in res} cache.str_cache = {r[0]: r[1] for r in res}
[ "@", "db_api", ".", "api_context_manager", ".", "reader", "def", "_refresh_from_db", "(", "ctx", ",", "cache", ")", ":", "with", "db_api", ".", "api_context_manager", ".", "reader", ".", "connection", ".", "using", "(", "ctx", ")", "as", "conn", ":", "sel"...
grabs all custom resource classes from the db table and populates the supplied cache objects internal integer and string identifier dicts .
train
false
16,451
def get_image_dimensions(file_or_path, close=False): from PIL import ImageFile as PillowImageFile p = PillowImageFile.Parser() if hasattr(file_or_path, 'read'): file = file_or_path file_pos = file.tell() file.seek(0) else: file = open(file_or_path, 'rb') close = True try: chunk_size = 1024 while 1: data = file.read(chunk_size) if (not data): break try: p.feed(data) except zlib.error as e: if e.args[0].startswith('Error -5'): pass else: raise except struct.error: pass if p.image: return p.image.size chunk_size *= 2 return (None, None) finally: if close: file.close() else: file.seek(file_pos)
[ "def", "get_image_dimensions", "(", "file_or_path", ",", "close", "=", "False", ")", ":", "from", "PIL", "import", "ImageFile", "as", "PillowImageFile", "p", "=", "PillowImageFile", ".", "Parser", "(", ")", "if", "hasattr", "(", "file_or_path", ",", "'read'", ...
returns the of an image .
train
false
16,452
@task @needs(['reset_test_database', 'clear_mongo', 'load_bok_choy_data', 'load_courses']) @might_call('start_servers') @cmdopts([BOKCHOY_FASTTEST], share_with=['start_servers']) @timed def prepare_bokchoy_run(options): if (not options.get('fasttest', False)): print colorize('green', 'Generating optimized static assets...') if (options.get('log_dir') is None): call_task('update_assets', args=['--settings', 'test_static_optimized']) else: call_task('update_assets', args=['--settings', 'test_static_optimized', '--collect-log', options.log_dir]) msg = colorize('green', 'Confirming servers are running...') print msg start_servers()
[ "@", "task", "@", "needs", "(", "[", "'reset_test_database'", ",", "'clear_mongo'", ",", "'load_bok_choy_data'", ",", "'load_courses'", "]", ")", "@", "might_call", "(", "'start_servers'", ")", "@", "cmdopts", "(", "[", "BOKCHOY_FASTTEST", "]", ",", "share_with"...
sets up and starts servers for a bok choy run .
train
false
16,453
def current_env(): if (config.default_prefix == config.root_dir): name = config.root_env_name else: name = basename(config.default_prefix) return name
[ "def", "current_env", "(", ")", ":", "if", "(", "config", ".", "default_prefix", "==", "config", ".", "root_dir", ")", ":", "name", "=", "config", ".", "root_env_name", "else", ":", "name", "=", "basename", "(", "config", ".", "default_prefix", ")", "ret...
retrieves dictionary with current environments name and prefix .
train
false
16,454
def read_fixture_lines(filename): lines = [] for line in codecs.open(filename, u'rb', encoding=u'utf-8'): lines.append(line.strip()) return lines
[ "def", "read_fixture_lines", "(", "filename", ")", ":", "lines", "=", "[", "]", "for", "line", "in", "codecs", ".", "open", "(", "filename", ",", "u'rb'", ",", "encoding", "=", "u'utf-8'", ")", ":", "lines", ".", "append", "(", "line", ".", "strip", ...
read lines of text from file .
train
false
16,455
def safe_concurrent_rename(src, dst): if os.path.isdir(src): safe_rmtree(dst) else: safe_delete(dst) try: shutil.move(src, dst) except IOError as e: if (e.errno != errno.EEXIST): raise
[ "def", "safe_concurrent_rename", "(", "src", ",", "dst", ")", ":", "if", "os", ".", "path", ".", "isdir", "(", "src", ")", ":", "safe_rmtree", "(", "dst", ")", "else", ":", "safe_delete", "(", "dst", ")", "try", ":", "shutil", ".", "move", "(", "sr...
rename src to dst .
train
true
16,459
def cnv_date(attribute, arg, element): return str(arg)
[ "def", "cnv_date", "(", "attribute", ",", "arg", ",", "element", ")", ":", "return", "str", "(", "arg", ")" ]
a dateordatetime value is either an [xmlschema-2] date value or an [xmlschema-2] datetime value .
train
false
16,460
def get_active_streams(realm): return Stream.objects.filter(realm=realm, deactivated=False)
[ "def", "get_active_streams", "(", "realm", ")", ":", "return", "Stream", ".", "objects", ".", "filter", "(", "realm", "=", "realm", ",", "deactivated", "=", "False", ")" ]
return all streams that have not been deactivated .
train
false
16,461
def _get_pretty_string(obj): sio = StringIO() pprint.pprint(obj, stream=sio) return sio.getvalue()
[ "def", "_get_pretty_string", "(", "obj", ")", ":", "sio", "=", "StringIO", "(", ")", "pprint", ".", "pprint", "(", "obj", ",", "stream", "=", "sio", ")", "return", "sio", ".", "getvalue", "(", ")" ]
return a prettier version of obj parameters obj : object object to pretty print returns s : str pretty print object repr .
train
true
16,462
def imview(*args, **kwargs): if ('figure' not in kwargs): f = plt.figure() else: f = kwargs['figure'] new_ax = matplotlib.axes.Axes(f, [0, 0, 1, 1], xticks=[], yticks=[], frame_on=False) f.delaxes(f.gca()) f.add_axes(new_ax) if ((len(args) < 5) and ('interpolation' not in kwargs)): kwargs['interpolation'] = 'nearest' plt.imshow(*args, **kwargs)
[ "def", "imview", "(", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "'figure'", "not", "in", "kwargs", ")", ":", "f", "=", "plt", ".", "figure", "(", ")", "else", ":", "f", "=", "kwargs", "[", "'figure'", "]", "new_ax", "=", "matplotlib"...
a matplotlib-based image viewer command .
train
false
16,464
def _build_tag_param_list(params, tags): keys = sorted(tags.keys()) i = 1 for key in keys: value = tags[key] params['Tags.member.{0}.Key'.format(i)] = key if (value is not None): params['Tags.member.{0}.Value'.format(i)] = value i += 1
[ "def", "_build_tag_param_list", "(", "params", ",", "tags", ")", ":", "keys", "=", "sorted", "(", "tags", ".", "keys", "(", ")", ")", "i", "=", "1", "for", "key", "in", "keys", ":", "value", "=", "tags", "[", "key", "]", "params", "[", "'Tags.membe...
helper function to build a tag parameter list to send .
train
true
16,465
def call_xenhost(session, method, arg_dict): try: result = session.call_plugin('xenhost.py', method, args=arg_dict) if (not result): return '' return jsonutils.loads(result) except ValueError: LOG.exception(_LE('Unable to get updated status')) return None except session.XenAPI.Failure as e: LOG.error(_LE('The call to %(method)s returned an error: %(e)s.'), {'method': method, 'e': e}) return e.details[1]
[ "def", "call_xenhost", "(", "session", ",", "method", ",", "arg_dict", ")", ":", "try", ":", "result", "=", "session", ".", "call_plugin", "(", "'xenhost.py'", ",", "method", ",", "args", "=", "arg_dict", ")", "if", "(", "not", "result", ")", ":", "ret...
there will be several methods that will need this general handling for interacting with the xenhost plugin .
train
false
16,466
def bump_shop_product_signal_handler(sender, instance, **kwargs): bump_cache_for_shop_product(instance)
[ "def", "bump_shop_product_signal_handler", "(", "sender", ",", "instance", ",", "**", "kwargs", ")", ":", "bump_cache_for_shop_product", "(", "instance", ")" ]
signal handler for clearing shop product cache .
train
false
16,467
def merge_geometries(geometries_str, sep='$'): geometries = geometries_str.split(sep) if (len(geometries) == 1): return geometries_str else: pool = OGRGeometry(geometries[0]) for geom in geometries: pool = pool.union(OGRGeometry(geom)) return pool.wkt
[ "def", "merge_geometries", "(", "geometries_str", ",", "sep", "=", "'$'", ")", ":", "geometries", "=", "geometries_str", ".", "split", "(", "sep", ")", "if", "(", "len", "(", "geometries", ")", "==", "1", ")", ":", "return", "geometries_str", "else", ":"...
take a list of geometries in a string .
train
false
16,468
@validator def app(environ, start_response): if (environ['REQUEST_METHOD'].upper() != 'POST'): data = 'Hello, World!\n' else: data = environ['wsgi.input'].read() status = '200 OK' response_headers = [('Content-type', 'text/plain'), ('Content-Length', str(len(data))), ('X-Gunicorn-Version', __version__), ('Test', 'test \xd1\x82\xd0\xb5\xd1\x81\xd1\x82')] start_response(status, response_headers) return iter([data])
[ "@", "validator", "def", "app", "(", "environ", ",", "start_response", ")", ":", "if", "(", "environ", "[", "'REQUEST_METHOD'", "]", ".", "upper", "(", ")", "!=", "'POST'", ")", ":", "data", "=", "'Hello, World!\\n'", "else", ":", "data", "=", "environ",...
simplest possible application object .
train
false
16,469
def safe_izip(*args): assert all([(len(arg) == len(args[0])) for arg in args]) return izip(*args)
[ "def", "safe_izip", "(", "*", "args", ")", ":", "assert", "all", "(", "[", "(", "len", "(", "arg", ")", "==", "len", "(", "args", "[", "0", "]", ")", ")", "for", "arg", "in", "args", "]", ")", "return", "izip", "(", "*", "args", ")" ]
like izip .
train
false
16,470
def _pragma_foreign_keys(connection, on): connection.execute(('PRAGMA foreign_keys=%s' % ('ON' if on else 'OFF')))
[ "def", "_pragma_foreign_keys", "(", "connection", ",", "on", ")", ":", "connection", ".", "execute", "(", "(", "'PRAGMA foreign_keys=%s'", "%", "(", "'ON'", "if", "on", "else", "'OFF'", ")", ")", ")" ]
sets the pragma foreign_keys state of the sqlite database .
train
false
16,471
def hyperfocal_distance(f, N, c): f = sympify(f) N = sympify(N) c = sympify(c) return ((1 / (N * c)) * (f ** 2))
[ "def", "hyperfocal_distance", "(", "f", ",", "N", ",", "c", ")", ":", "f", "=", "sympify", "(", "f", ")", "N", "=", "sympify", "(", "N", ")", "c", "=", "sympify", "(", "c", ")", "return", "(", "(", "1", "/", "(", "N", "*", "c", ")", ")", ...
parameters f: sympifiable focal length of a given lens n: sympifiable f-number of a given lens c: sympifiable circle of confusion of a given image format example .
train
false
16,473
def cycle_colors(chunk, palette=DEFAULT_PALETTE): colors = [] g = itertools.cycle(palette) for i in range(len(chunk)): colors.append(next(g)) return colors
[ "def", "cycle_colors", "(", "chunk", ",", "palette", "=", "DEFAULT_PALETTE", ")", ":", "colors", "=", "[", "]", "g", "=", "itertools", ".", "cycle", "(", "palette", ")", "for", "i", "in", "range", "(", "len", "(", "chunk", ")", ")", ":", "colors", ...
build a color list just cycling through a given palette .
train
false
16,474
def fetch_snippets_from_dir(path): rv = [] for filename in glob.glob(os.path.join(path, '*.tmSnippet')): print ('Reading file %s' % filename) f = open(filename) content = f.read() cont = parse_content(content) if cont: name = os.path.splitext(os.path.basename(filename))[0] rv.append((name, cont)) return rv
[ "def", "fetch_snippets_from_dir", "(", "path", ")", ":", "rv", "=", "[", "]", "for", "filename", "in", "glob", ".", "glob", "(", "os", ".", "path", ".", "join", "(", "path", ",", "'*.tmSnippet'", ")", ")", ":", "print", "(", "'Reading file %s'", "%", ...
fetch snippets from a given path .
train
false
16,475
def _oldGiInit(): _glibbase.ensureNotImported(_PYGTK_MODULES, "Introspected and static glib/gtk bindings must not be mixed; can't import gireactor since pygtk2 module is already imported.") global GLib from gi.repository import GLib if (getattr(GLib, 'threads_init', None) is not None): GLib.threads_init() _glibbase.ensureNotImported([], '', preventImports=_PYGTK_MODULES)
[ "def", "_oldGiInit", "(", ")", ":", "_glibbase", ".", "ensureNotImported", "(", "_PYGTK_MODULES", ",", "\"Introspected and static glib/gtk bindings must not be mixed; can't import gireactor since pygtk2 module is already imported.\"", ")", "global", "GLib", "from", "gi", ".", "re...
make sure pygtk and gi arent loaded at the same time .
train
false
16,476
def get_model_name(model): return model._meta.model_name
[ "def", "get_model_name", "(", "model", ")", ":", "return", "model", ".", "_meta", ".", "model_name" ]
returns the name of the model .
train
false
16,477
def parse_CPS_file(lines): chimeras = [] for line in lines: record = line.split() try: id = record[1] parent1 = record[2] parent2 = record[3] verdict = record[10] except IndexError: raise ValueError('Error parsing ChimeraSlayer CPS file.') if (verdict == 'YES'): chimeras.append((id, [parent1, parent2])) return chimeras
[ "def", "parse_CPS_file", "(", "lines", ")", ":", "chimeras", "=", "[", "]", "for", "line", "in", "lines", ":", "record", "=", "line", ".", "split", "(", ")", "try", ":", "id", "=", "record", "[", "1", "]", "parent1", "=", "record", "[", "2", "]",...
parse the cps file from chimeraslayer .
train
false
16,478
def mime_to_document_iters(input_file, boundary, read_chunk_size=4096): doc_files = iter_multipart_mime_documents(input_file, boundary, read_chunk_size) for (i, doc_file) in enumerate(doc_files): headers = parse_mime_headers(doc_file) (yield (headers, doc_file))
[ "def", "mime_to_document_iters", "(", "input_file", ",", "boundary", ",", "read_chunk_size", "=", "4096", ")", ":", "doc_files", "=", "iter_multipart_mime_documents", "(", "input_file", ",", "boundary", ",", "read_chunk_size", ")", "for", "(", "i", ",", "doc_file"...
takes a file-like object containing a multipart mime document and returns an iterator of tuples .
train
false
16,479
def get_field_data_type(field): return (field.description % field.__dict__)
[ "def", "get_field_data_type", "(", "field", ")", ":", "return", "(", "field", ".", "description", "%", "field", ".", "__dict__", ")" ]
returns the description for a given field type .
train
false
16,480
def flex_loader(alias): if (not alias.startswith('flex.')): return try: if alias.startswith('flex.messaging.messages'): import pyamf.flex.messaging elif alias.startswith('flex.messaging.io'): import pyamf.flex elif alias.startswith('flex.data.messages'): import pyamf.flex.data return CLASS_CACHE[alias] except KeyError: raise UnknownClassAlias(alias)
[ "def", "flex_loader", "(", "alias", ")", ":", "if", "(", "not", "alias", ".", "startswith", "(", "'flex.'", ")", ")", ":", "return", "try", ":", "if", "alias", ".", "startswith", "(", "'flex.messaging.messages'", ")", ":", "import", "pyamf", ".", "flex",...
loader for l{flex<pyamf .
train
true
16,481
@receiver(models.signals.post_save, sender=CreditCourse) @receiver(models.signals.post_delete, sender=CreditCourse) def invalidate_credit_courses_cache(sender, **kwargs): cache.delete(CreditCourse.CREDIT_COURSES_CACHE_KEY)
[ "@", "receiver", "(", "models", ".", "signals", ".", "post_save", ",", "sender", "=", "CreditCourse", ")", "@", "receiver", "(", "models", ".", "signals", ".", "post_delete", ",", "sender", "=", "CreditCourse", ")", "def", "invalidate_credit_courses_cache", "(...
invalidate the cache of credit courses .
train
false
16,482
def make_otu_labels(otu_ids, lineages, n_levels=1): if (len(lineages[0]) > 0): otu_labels = [] for (i, lineage) in enumerate(lineages): if (n_levels > len(lineage)): otu_label = ('%s (%s)' % (';'.join(lineage), otu_ids[i])) else: otu_label = ('%s (%s)' % (';'.join(lineage[(- n_levels):]), otu_ids[i])) otu_labels.append(otu_label) otu_labels = [lab.replace('"', '') for lab in otu_labels] else: otu_labels = otu_ids return otu_labels
[ "def", "make_otu_labels", "(", "otu_ids", ",", "lineages", ",", "n_levels", "=", "1", ")", ":", "if", "(", "len", "(", "lineages", "[", "0", "]", ")", ">", "0", ")", ":", "otu_labels", "=", "[", "]", "for", "(", "i", ",", "lineage", ")", "in", ...
returns pretty otu labels: lineage substring lineage substring includes the last n_levels lineage levels .
train
false
16,483
def delete_quantum_ports(ports, root_helper): for port in ports: if ip_lib.device_exists(port): device = ip_lib.IPDevice(port, root_helper) device.link.delete() LOG.info(_('Delete %s'), port)
[ "def", "delete_quantum_ports", "(", "ports", ",", "root_helper", ")", ":", "for", "port", "in", "ports", ":", "if", "ip_lib", ".", "device_exists", "(", "port", ")", ":", "device", "=", "ip_lib", ".", "IPDevice", "(", "port", ",", "root_helper", ")", "de...
delete non-internal ports created by quantum non-internal ovs ports need to be removed manually .
train
false
16,486
@requires_application() def test_capability(): non_default_vals = dict(title='foo', size=[100, 100], position=[0, 0], show=True, decorate=False, resizable=False, vsync=True) good_kwargs = dict() bad_kwargs = dict() with Canvas() as c: for (key, val) in c.app.backend_module.capability.items(): if (key in non_default_vals): if val: good_kwargs[key] = non_default_vals[key] else: bad_kwargs[key] = non_default_vals[key] with Canvas(**good_kwargs): pass for (key, val) in bad_kwargs.items(): assert_raises(RuntimeError, Canvas, **{key: val})
[ "@", "requires_application", "(", ")", "def", "test_capability", "(", ")", ":", "non_default_vals", "=", "dict", "(", "title", "=", "'foo'", ",", "size", "=", "[", "100", ",", "100", "]", ",", "position", "=", "[", "0", ",", "0", "]", ",", "show", ...
test application capability enumeration .
train
false
16,487
def service_update(context, service_id, values): return IMPL.service_update(context, service_id, values)
[ "def", "service_update", "(", "context", ",", "service_id", ",", "values", ")", ":", "return", "IMPL", ".", "service_update", "(", "context", ",", "service_id", ",", "values", ")" ]
set the given properties on an service and update it .
train
false
16,488
def is_rel(s): if (len(s) == 0): return True elif (all((isinstance(el, tuple) for el in s)) and (len(max(s)) == len(min(s)))): return True else: raise ValueError((u'Set %r contains sequences of different lengths' % s))
[ "def", "is_rel", "(", "s", ")", ":", "if", "(", "len", "(", "s", ")", "==", "0", ")", ":", "return", "True", "elif", "(", "all", "(", "(", "isinstance", "(", "el", ",", "tuple", ")", "for", "el", "in", "s", ")", ")", "and", "(", "len", "(",...
check whether a set represents a relation .
train
false
16,490
def sameopenfile(fp1, fp2): s1 = os.fstat(fp1) s2 = os.fstat(fp2) return samestat(s1, s2)
[ "def", "sameopenfile", "(", "fp1", ",", "fp2", ")", ":", "s1", "=", "os", ".", "fstat", "(", "fp1", ")", "s2", "=", "os", ".", "fstat", "(", "fp2", ")", "return", "samestat", "(", "s1", ",", "s2", ")" ]
test whether two open file objects reference the same file .
train
false
16,491
def make_decreasing_ohlc(open, high, low, close, dates, **kwargs): (flat_decrease_x, flat_decrease_y, text_decrease) = _OHLC(open, high, low, close, dates).get_decrease() kwargs.setdefault('line', dict(color=_DEFAULT_DECREASING_COLOR, width=1)) kwargs.setdefault('text', text_decrease) kwargs.setdefault('showlegend', False) kwargs.setdefault('name', 'Decreasing') ohlc_decr = dict(type='scatter', x=flat_decrease_x, y=flat_decrease_y, mode='lines', **kwargs) return ohlc_decr
[ "def", "make_decreasing_ohlc", "(", "open", ",", "high", ",", "low", ",", "close", ",", "dates", ",", "**", "kwargs", ")", ":", "(", "flat_decrease_x", ",", "flat_decrease_y", ",", "text_decrease", ")", "=", "_OHLC", "(", "open", ",", "high", ",", "low",...
makes decreasing ohlc sticks .
train
false
16,492
def unsubscribe_from_basket_action(newsletter): def unsubscribe_from_basket(modeladmin, request, queryset): 'Unsubscribe from Basket.' ts = [unsubscribe_from_basket_task.subtask(args=[userprofile.user.email, [newsletter]]) for userprofile in queryset] TaskSet(ts).apply_async() messages.success(request, 'Basket update started.') unsubscribe_from_basket.short_description = 'Unsubscribe from {0}'.format(newsletter) func_name = 'unsubscribe_from_basket_{0}'.format(newsletter.replace('-', '_')) unsubscribe_from_basket.__name__ = func_name return unsubscribe_from_basket
[ "def", "unsubscribe_from_basket_action", "(", "newsletter", ")", ":", "def", "unsubscribe_from_basket", "(", "modeladmin", ",", "request", ",", "queryset", ")", ":", "ts", "=", "[", "unsubscribe_from_basket_task", ".", "subtask", "(", "args", "=", "[", "userprofil...
unsubscribe from basket action .
train
false
16,493
def get_response_stream(response): try: getheader = response.headers.getheader except AttributeError: getheader = response.getheader if (getheader('content-encoding') == 'gzip'): return GzipDecodedResponse(response) return response
[ "def", "get_response_stream", "(", "response", ")", ":", "try", ":", "getheader", "=", "response", ".", "headers", ".", "getheader", "except", "AttributeError", ":", "getheader", "=", "response", ".", "getheader", "if", "(", "getheader", "(", "'content-encoding'...
helper function to return either a gzip reader if content-encoding is gzip otherwise the response itself .
train
false
16,494
def deltaE_ciede94(lab1, lab2, kH=1, kC=1, kL=1, k1=0.045, k2=0.015): (L1, C1) = np.rollaxis(lab2lch(lab1), (-1))[:2] (L2, C2) = np.rollaxis(lab2lch(lab2), (-1))[:2] dL = (L1 - L2) dC = (C1 - C2) dH2 = get_dH2(lab1, lab2) SL = 1 SC = (1 + (k1 * C1)) SH = (1 + (k2 * C1)) dE2 = ((dL / (kL * SL)) ** 2) dE2 += ((dC / (kC * SC)) ** 2) dE2 += (dH2 / ((kH * SH) ** 2)) return np.sqrt(dE2)
[ "def", "deltaE_ciede94", "(", "lab1", ",", "lab2", ",", "kH", "=", "1", ",", "kC", "=", "1", ",", "kL", "=", "1", ",", "k1", "=", "0.045", ",", "k2", "=", "0.015", ")", ":", "(", "L1", ",", "C1", ")", "=", "np", ".", "rollaxis", "(", "lab2l...
color difference according to ciede 94 standard accommodates perceptual non-uniformities through the use of application specific scale factors .
train
false
16,496
def create_access_port(network_switch, port_name, vlan): debug = False new_port = SwitchPort.objects.get_or_create(port_name=port_name, mode='access', access_vlan=vlan, network_switch=network_switch) if debug: print new_port
[ "def", "create_access_port", "(", "network_switch", ",", "port_name", ",", "vlan", ")", ":", "debug", "=", "False", "new_port", "=", "SwitchPort", ".", "objects", ".", "get_or_create", "(", "port_name", "=", "port_name", ",", "mode", "=", "'access'", ",", "a...
create a switchport object - access port .
train
false
16,497
def run_scripts(package_location, scripts): path = os.path.join(package_location, 'scripts/') cwd = os.getcwd() os.chdir(path) for script in scripts: if os.path.exists(script): try: subprocess.check_call(script, stdout=sys.stdout, stderr=sys.stderr) except Exception: os.chdir(cwd) raise os.chdir(cwd)
[ "def", "run_scripts", "(", "package_location", ",", "scripts", ")", ":", "path", "=", "os", ".", "path", ".", "join", "(", "package_location", ",", "'scripts/'", ")", "cwd", "=", "os", ".", "getcwd", "(", ")", "os", ".", "chdir", "(", "path", ")", "f...
search for installation scripts speficied by the scripts list .
train
false
16,498
def delete_branch(pr, session): refs_url = pr['head']['repo']['git_refs_url'] branch_url = refs_url.replace('{/sha}', ('/heads/' + pr['head']['ref'])) return session.delete(branch_url)
[ "def", "delete_branch", "(", "pr", ",", "session", ")", ":", "refs_url", "=", "pr", "[", "'head'", "]", "[", "'repo'", "]", "[", "'git_refs_url'", "]", "branch_url", "=", "refs_url", ".", "replace", "(", "'{/sha}'", ",", "(", "'/heads/'", "+", "pr", "[...
launch the delete branch dialog .
train
false
16,500
def find_hessian(point, vars=None, model=None): model = modelcontext(model) H = model.fastd2logp(vars) return H(Point(point, model=model))
[ "def", "find_hessian", "(", "point", ",", "vars", "=", "None", ",", "model", "=", "None", ")", ":", "model", "=", "modelcontext", "(", "model", ")", "H", "=", "model", ".", "fastd2logp", "(", "vars", ")", "return", "H", "(", "Point", "(", "point", ...
returns hessian of logp at the point passed .
train
false
16,501
def project_activity_postprocess(form): form_vars = form.vars project_id = form_vars.get('project_id', None) if project_id: s3db = current.s3db db = current.db activity_id = form_vars.get('id', None) ltable = s3db.project_activity_organisation org = db((ltable.activity_id == activity_id)).select(ltable.organisation_id, limitby=(0, 1)).first() if org: return ptable = s3db.project_project project = db((ptable.id == project_id)).select(ptable.organisation_id, limitby=(0, 1)).first() try: organisation_id = project.organisation_id except: return ltable.insert(activity_id=activity_id, organisation_id=organisation_id)
[ "def", "project_activity_postprocess", "(", "form", ")", ":", "form_vars", "=", "form", ".", "vars", "project_id", "=", "form_vars", ".", "get", "(", "'project_id'", ",", "None", ")", "if", "project_id", ":", "s3db", "=", "current", ".", "s3db", "db", "=",...
default the activitys organisation to that of the project .
train
false
16,502
@given(u'a new working directory') def step_a_new_working_directory(context): command_util.ensure_context_attribute_exists(context, 'workdir', None) command_util.ensure_workdir_exists(context) shutil.rmtree(context.workdir, ignore_errors=True) command_util.ensure_workdir_exists(context)
[ "@", "given", "(", "u'a new working directory'", ")", "def", "step_a_new_working_directory", "(", "context", ")", ":", "command_util", ".", "ensure_context_attribute_exists", "(", "context", ",", "'workdir'", ",", "None", ")", "command_util", ".", "ensure_workdir_exists...
creates a new .
train
true
16,503
def copy_file_from_manifest(repo, ctx, filename, dir): for changeset in reversed_upper_bounded_changelog(repo, ctx): changeset_ctx = repo.changectx(changeset) fctx = get_file_context_from_ctx(changeset_ctx, filename) if (fctx and (fctx not in ['DELETED'])): file_path = os.path.join(dir, filename) fh = open(file_path, 'wb') fh.write(fctx.data()) fh.close() return file_path return None
[ "def", "copy_file_from_manifest", "(", "repo", ",", "ctx", ",", "filename", ",", "dir", ")", ":", "for", "changeset", "in", "reversed_upper_bounded_changelog", "(", "repo", ",", "ctx", ")", ":", "changeset_ctx", "=", "repo", ".", "changectx", "(", "changeset",...
copy the latest version of the file named filename from the repository manifest to the directory to which dir refers .
train
false
16,504
def deeper2net_conv2d(teacher_w): (nb_filter, nb_channel, kh, kw) = teacher_w.shape student_w = np.zeros((nb_filter, nb_filter, kh, kw)) for i in xrange(nb_filter): student_w[(i, i, ((kh - 1) / 2), ((kw - 1) / 2))] = 1.0 student_b = np.zeros(nb_filter) return (student_w, student_b)
[ "def", "deeper2net_conv2d", "(", "teacher_w", ")", ":", "(", "nb_filter", ",", "nb_channel", ",", "kh", ",", "kw", ")", "=", "teacher_w", ".", "shape", "student_w", "=", "np", ".", "zeros", "(", "(", "nb_filter", ",", "nb_filter", ",", "kh", ",", "kw",...
get initial weights for a deeper conv2d layer by net2deeper .
train
false
16,505
def get_alembic_version(meta): try: a_ver = sa.Table('alembic_version', meta, autoload=True) return sa.select([a_ver.c.version_num]).scalar() except sa.exc.NoSuchTableError: return None
[ "def", "get_alembic_version", "(", "meta", ")", ":", "try", ":", "a_ver", "=", "sa", ".", "Table", "(", "'alembic_version'", ",", "meta", ",", "autoload", "=", "True", ")", "return", "sa", ".", "select", "(", "[", "a_ver", ".", "c", ".", "version_num",...
return alembic version or none if no alembic table exists .
train
false
16,506
def _parse_global_variables(user_cidr, inventory, user_defined_config): if ('all' not in inventory): inventory['all'] = {} if ('vars' not in inventory['all']): inventory['all']['vars'] = {} inventory['all']['vars']['container_cidr'] = user_cidr if ('global_overrides' in user_defined_config): if isinstance(user_defined_config['global_overrides'], dict): inventory['all']['vars'].update(user_defined_config['global_overrides']) logger.debug('Applied global_overrides') kept_vars = user_defined_config['global_overrides'].keys() kept_vars.append('container_cidr') for key in inventory['all']['vars'].keys(): if (key not in kept_vars): logger.debug('Deleting key %s from inventory', key) del inventory['all']['vars'][key]
[ "def", "_parse_global_variables", "(", "user_cidr", ",", "inventory", ",", "user_defined_config", ")", ":", "if", "(", "'all'", "not", "in", "inventory", ")", ":", "inventory", "[", "'all'", "]", "=", "{", "}", "if", "(", "'vars'", "not", "in", "inventory"...
add any extra variables that may have been set in config .
train
false
16,507
def topic_rule_exists(ruleName, region=None, key=None, keyid=None, profile=None): try: conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) rule = conn.get_topic_rule(ruleName=ruleName) return {'exists': True} except ClientError as e: err = salt.utils.boto3.get_error(e) if (e.response.get('Error', {}).get('Code') == 'UnauthorizedException'): return {'exists': False} return {'error': salt.utils.boto3.get_error(e)}
[ "def", "topic_rule_exists", "(", "ruleName", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "try", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "key", "=", ...
given a rule name .
train
false
16,508
def bind_floating_ip(floating_ip, device): _execute('ip', 'addr', 'add', (str(floating_ip) + '/32'), 'dev', device, run_as_root=True, check_exit_code=[0, 2, 254]) if (CONF.send_arp_for_ha and (CONF.send_arp_for_ha_count > 0)): send_arp_for_ip(floating_ip, device, CONF.send_arp_for_ha_count)
[ "def", "bind_floating_ip", "(", "floating_ip", ",", "device", ")", ":", "_execute", "(", "'ip'", ",", "'addr'", ",", "'add'", ",", "(", "str", "(", "floating_ip", ")", "+", "'/32'", ")", ",", "'dev'", ",", "device", ",", "run_as_root", "=", "True", ","...
bind ip to public interface .
train
false
16,509
def splrep(x, y, w=None, xb=None, xe=None, k=3, task=0, s=None, t=None, full_output=0, per=0, quiet=1): if (task <= 0): _curfit_cache = {} (x, y) = map(atleast_1d, [x, y]) m = len(x) if (w is None): w = ones(m, float) if (s is None): s = 0.0 else: w = atleast_1d(w) if (s is None): s = (m - sqrt((2 * m))) if (not (len(w) == m)): raise TypeError(('len(w)=%d is not equal to m=%d' % (len(w), m))) if ((m != len(y)) or (m != len(w))): raise TypeError('Lengths of the first three arguments (x,y,w) must be equal') if (not (1 <= k <= 5)): raise TypeError(('Given degree of the spline (k=%d) is not supported. (1<=k<=5)' % k)) if (m <= k): raise TypeError('m > k must hold') if (xb is None): xb = x[0] if (xe is None): xe = x[(-1)] if (not ((-1) <= task <= 1)): raise TypeError('task must be -1, 0 or 1') if (t is not None): task = (-1) if (task == (-1)): if (t is None): raise TypeError('Knots must be given for task=-1') numknots = len(t) _curfit_cache['t'] = empty((((numknots + (2 * k)) + 2),), float) _curfit_cache['t'][(k + 1):((- k) - 1)] = t nest = len(_curfit_cache['t']) elif (task == 0): if per: nest = max((m + (2 * k)), ((2 * k) + 3)) else: nest = max(((m + k) + 1), ((2 * k) + 3)) t = empty((nest,), float) _curfit_cache['t'] = t if (task <= 0): if per: _curfit_cache['wrk'] = empty((((m * (k + 1)) + (nest * (8 + (5 * k)))),), float) else: _curfit_cache['wrk'] = empty((((m * (k + 1)) + (nest * (7 + (3 * k)))),), float) _curfit_cache['iwrk'] = empty((nest,), intc) try: t = _curfit_cache['t'] wrk = _curfit_cache['wrk'] iwrk = _curfit_cache['iwrk'] except KeyError: raise TypeError('must call with task=1 only after call with task=0,-1') if (not per): (n, c, fp, ier) = dfitpack.curfit(task, x, y, w, t, wrk, iwrk, xb, xe, k, s) else: (n, c, fp, ier) = dfitpack.percur(task, x, y, w, t, wrk, iwrk, k, s) tck = (t[:n], c[:n], k) if ((ier <= 0) and (not quiet)): _mess = (_iermess[ier][0] + (' DCTB k=%d n=%d m=%d fp=%f s=%f' % (k, len(t), m, fp, s))) warnings.warn(RuntimeWarning(_mess)) if ((ier > 0) and (not full_output)): if (ier in [1, 2, 3]): warnings.warn(RuntimeWarning(_iermess[ier][0])) else: try: raise _iermess[ier][1](_iermess[ier][0]) except KeyError: raise _iermess['unknown'][1](_iermess['unknown'][0]) if full_output: try: return (tck, fp, ier, _iermess[ier][0]) except KeyError: return (tck, fp, ier, _iermess['unknown'][0]) else: return tck
[ "def", "splrep", "(", "x", ",", "y", ",", "w", "=", "None", ",", "xb", "=", "None", ",", "xe", "=", "None", ",", "k", "=", "3", ",", "task", "=", "0", ",", "s", "=", "None", ",", "t", "=", "None", ",", "full_output", "=", "0", ",", "per",...
find the b-spline representation of 1-d curve .
train
false
16,510
def getAbridgedSettings(gcodeText): abridgedSettings = [] lines = archive.getTextLines(gcodeText) settingsStart = False for line in lines: splitLine = gcodec.getSplitLineBeforeBracketSemicolon(line) firstWord = gcodec.getFirstWord(splitLine) if ((firstWord == '(<setting>') and settingsStart): if (len(splitLine) > 4): abridgedSettings.append(AbridgedSetting(splitLine)) elif (firstWord == '(<settings>)'): settingsStart = True elif (firstWord == '(</settings>)'): return abridgedSettings return []
[ "def", "getAbridgedSettings", "(", "gcodeText", ")", ":", "abridgedSettings", "=", "[", "]", "lines", "=", "archive", ".", "getTextLines", "(", "gcodeText", ")", "settingsStart", "=", "False", "for", "line", "in", "lines", ":", "splitLine", "=", "gcodec", "....
get the abridged settings from the gcode text .
train
false
16,511
def tag_string_convert(key, data, errors, context): if isinstance(data[key], basestring): tags = [tag.strip() for tag in data[key].split(',') if tag.strip()] else: tags = data[key] current_index = max(([int(k[1]) for k in data.keys() if ((len(k) == 3) and (k[0] == 'tags'))] + [(-1)])) for (num, tag) in zip(count((current_index + 1)), tags): data[('tags', num, 'name')] = tag for tag in tags: tag_length_validator(tag, context) tag_name_validator(tag, context)
[ "def", "tag_string_convert", "(", "key", ",", "data", ",", "errors", ",", "context", ")", ":", "if", "isinstance", "(", "data", "[", "key", "]", ",", "basestring", ")", ":", "tags", "=", "[", "tag", ".", "strip", "(", ")", "for", "tag", "in", "data...
takes a list of tags that is a comma-separated string and parses tag names .
train
false
16,512
def list_cidr_ips_ipv6(cidr): ips = netaddr.IPNetwork(cidr) return [str(ip.ipv6()) for ip in list(ips)]
[ "def", "list_cidr_ips_ipv6", "(", "cidr", ")", ":", "ips", "=", "netaddr", ".", "IPNetwork", "(", "cidr", ")", "return", "[", "str", "(", "ip", ".", "ipv6", "(", ")", ")", "for", "ip", "in", "list", "(", "ips", ")", "]" ]
get a list of ipv6 addresses from a cidr .
train
false
16,515
def _render_configuration(): text_repr = _current_statement.build() _INDENT = '' return text_repr
[ "def", "_render_configuration", "(", ")", ":", "text_repr", "=", "_current_statement", ".", "build", "(", ")", "_INDENT", "=", "''", "return", "text_repr" ]
renders the configuration tree into syslog-ngs configuration syntax .
train
false
16,516
def get_secret(): return file_io.read(constants.SECRET_LOC).rstrip()
[ "def", "get_secret", "(", ")", ":", "return", "file_io", ".", "read", "(", "constants", ".", "SECRET_LOC", ")", ".", "rstrip", "(", ")" ]
reads a secret key string from the specified file and returns it .
train
false
16,517
@cwd_at('test/test_evaluate/not_in_sys_path/pkg') def test_import_not_in_sys_path(): a = jedi.Script(path='module.py', line=5).goto_definitions() assert (a[0].name == 'int') a = jedi.Script(path='module.py', line=6).goto_definitions() assert (a[0].name == 'str') a = jedi.Script(path='module.py', line=7).goto_definitions() assert (a[0].name == 'str')
[ "@", "cwd_at", "(", "'test/test_evaluate/not_in_sys_path/pkg'", ")", "def", "test_import_not_in_sys_path", "(", ")", ":", "a", "=", "jedi", ".", "Script", "(", "path", "=", "'module.py'", ",", "line", "=", "5", ")", ".", "goto_definitions", "(", ")", "assert",...
non-direct imports .
train
false
16,520
def publish_msgstr(app, source, source_path, source_line, config, settings): from sphinx.io import SphinxI18nReader reader = SphinxI18nReader(app=app, parsers=config.source_parsers, parser_name='restructuredtext') reader.set_lineno_for_reporter(source_line) doc = reader.read(source=StringInput(source=source, source_path=source_path), parser=reader.parser, settings=settings) try: doc = doc[0] except IndexError: pass return doc
[ "def", "publish_msgstr", "(", "app", ",", "source", ",", "source_path", ",", "source_line", ",", "config", ",", "settings", ")", ":", "from", "sphinx", ".", "io", "import", "SphinxI18nReader", "reader", "=", "SphinxI18nReader", "(", "app", "=", "app", ",", ...
publish msgstr into docutils document .
train
false
16,521
def ensure_bytes(s): if isinstance(s, bytes): return s if hasattr(s, 'encode'): return s.encode() msg = 'Object %s is neither a bytes object nor has an encode method' raise TypeError((msg % s))
[ "def", "ensure_bytes", "(", "s", ")", ":", "if", "isinstance", "(", "s", ",", "bytes", ")", ":", "return", "s", "if", "hasattr", "(", "s", ",", "'encode'", ")", ":", "return", "s", ".", "encode", "(", ")", "msg", "=", "'Object %s is neither a bytes obj...
turn string or bytes to bytes .
train
false
16,525
def b1282int(st): e = 1 i = 0 for char in iterbytes(st): n = ord(char) i += (n * e) e <<= 7 return i
[ "def", "b1282int", "(", "st", ")", ":", "e", "=", "1", "i", "=", "0", "for", "char", "in", "iterbytes", "(", "st", ")", ":", "n", "=", "ord", "(", "char", ")", "i", "+=", "(", "n", "*", "e", ")", "e", "<<=", "7", "return", "i" ]
convert an integer represented as a base 128 string into an c{int} or c{long} .
train
false
16,526
@profiler.trace def server_console_output(request, instance_id, tail_length=None): return novaclient(request).servers.get_console_output(instance_id, length=tail_length)
[ "@", "profiler", ".", "trace", "def", "server_console_output", "(", "request", ",", "instance_id", ",", "tail_length", "=", "None", ")", ":", "return", "novaclient", "(", "request", ")", ".", "servers", ".", "get_console_output", "(", "instance_id", ",", "leng...
gets console output of an instance .
train
false
16,527
def user_config_dir(appname, roaming=True): if WINDOWS: path = user_data_dir(appname, roaming=roaming) elif (sys.platform == 'darwin'): path = user_data_dir(appname) else: path = os.getenv('XDG_CONFIG_HOME', expanduser('~/.config')) path = os.path.join(path, appname) return path
[ "def", "user_config_dir", "(", "appname", ",", "roaming", "=", "True", ")", ":", "if", "WINDOWS", ":", "path", "=", "user_data_dir", "(", "appname", ",", "roaming", "=", "roaming", ")", "elif", "(", "sys", ".", "platform", "==", "'darwin'", ")", ":", "...
return full path to the user-specific config dir for this application .
train
true
16,531
def _serialize_inventories(inventories, generation): inventories_by_class = {inventory.resource_class: inventory for inventory in inventories} inventories_dict = {} for (resource_class, inventory) in inventories_by_class.items(): inventories_dict[resource_class] = _serialize_inventory(inventory, generation=None) return {'resource_provider_generation': generation, 'inventories': inventories_dict}
[ "def", "_serialize_inventories", "(", "inventories", ",", "generation", ")", ":", "inventories_by_class", "=", "{", "inventory", ".", "resource_class", ":", "inventory", "for", "inventory", "in", "inventories", "}", "inventories_dict", "=", "{", "}", "for", "(", ...
turn a list of inventories in a dict by resource class .
train
false
16,532
def render_to_response(renderer_name, value, request=None, package=None, response=None): try: registry = request.registry except AttributeError: registry = None if (package is None): package = caller_package() helper = RendererHelper(name=renderer_name, package=package, registry=registry) with hide_attrs(request, 'response'): if (response is not None): request.response = response result = helper.render_to_response(value, None, request=request) return result
[ "def", "render_to_response", "(", "renderer_name", ",", "value", ",", "request", "=", "None", ",", "package", "=", "None", ",", "response", "=", "None", ")", ":", "try", ":", "registry", "=", "request", ".", "registry", "except", "AttributeError", ":", "re...
returns a httpresponse whose content is filled with the result of calling lookup .
train
false
16,533
def local_random(): global _local_random if (_local_random is None): _local_random = random.Random() return _local_random
[ "def", "local_random", "(", ")", ":", "global", "_local_random", "if", "(", "_local_random", "is", "None", ")", ":", "_local_random", "=", "random", ".", "Random", "(", ")", "return", "_local_random" ]
get the local random number generator .
train
false
16,534
def get_provider_metadata(metadata_url, supports_recursive=False, headers=None, expect_json=False): try: if supports_recursive: metadata = query_metadata(metadata_url, headers, expect_json) else: metadata = walk_metadata(metadata_url, headers, expect_json) except OpenShiftFactsMetadataUnavailableError: metadata = None return metadata
[ "def", "get_provider_metadata", "(", "metadata_url", ",", "supports_recursive", "=", "False", ",", "headers", "=", "None", ",", "expect_json", "=", "False", ")", ":", "try", ":", "if", "supports_recursive", ":", "metadata", "=", "query_metadata", "(", "metadata_...
retrieve the provider metadata args: metadata_url : metadata url supports_recursive : does the provider metadata api support recursion headers : headers to set for metadata request expect_json : does the metadata_url return json returns: dict: the provider metadata .
train
false
16,535
def MakeCdfFromHist(hist, label=None): if (label is None): label = hist.label return Cdf(hist, label=label)
[ "def", "MakeCdfFromHist", "(", "hist", ",", "label", "=", "None", ")", ":", "if", "(", "label", "is", "None", ")", ":", "label", "=", "hist", ".", "label", "return", "Cdf", "(", "hist", ",", "label", "=", "label", ")" ]
makes a cdf from a hist object .
train
false
16,536
def systemInformationType2ter(): a = L2PseudoLength(l2pLength=18) b = TpPd(pd=6) c = MessageType(mesType=3) d = NeighbourCellsDescription2() e = Si2terRestOctets() packet = ((((a / b) / c) / d) / e) return packet
[ "def", "systemInformationType2ter", "(", ")", ":", "a", "=", "L2PseudoLength", "(", "l2pLength", "=", "18", ")", "b", "=", "TpPd", "(", "pd", "=", "6", ")", "c", "=", "MessageType", "(", "mesType", "=", "3", ")", "d", "=", "NeighbourCellsDescription2", ...
system information type 2ter section 9 .
train
true
16,537
def int_divmod(context, builder, ty, x, y): if ty.signed: return int_divmod_signed(context, builder, ty, x, y) else: return (builder.udiv(x, y), builder.urem(x, y))
[ "def", "int_divmod", "(", "context", ",", "builder", ",", "ty", ",", "x", ",", "y", ")", ":", "if", "ty", ".", "signed", ":", "return", "int_divmod_signed", "(", "context", ",", "builder", ",", "ty", ",", "x", ",", "y", ")", "else", ":", "return", ...
integer divmod .
train
false
16,540
def qnwbeta(n, a=1.0, b=1.0): return _make_multidim_func(_qnwbeta1, n, a, b)
[ "def", "qnwbeta", "(", "n", ",", "a", "=", "1.0", ",", "b", "=", "1.0", ")", ":", "return", "_make_multidim_func", "(", "_qnwbeta1", ",", "n", ",", "a", ",", "b", ")" ]
computes nodes and weights for beta distribution parameters n : int or array_like a length-d iterable of the number of nodes in each dimension a : scalar or array_like .
train
false
16,541
def test_text(): test_data = BytesIO('{"a": "b"}') assert (hug.input_format.text(test_data) == '{"a": "b"}')
[ "def", "test_text", "(", ")", ":", "test_data", "=", "BytesIO", "(", "'{\"a\": \"b\"}'", ")", "assert", "(", "hug", ".", "input_format", ".", "text", "(", "test_data", ")", "==", "'{\"a\": \"b\"}'", ")" ]
test that "text" image can be loaded .
train
false
16,542
def test_iszero_substitution(): m = Matrix([[0.9, (-0.1), (-0.2), 0], [(-0.8), 0.9, (-0.4), 0], [(-0.1), (-0.8), 0.6, 0]]) m_rref = m.rref(iszerofunc=(lambda x: (abs(x) < 6e-15)))[0] m_correct = Matrix([[1.0, 0, (-0.301369863013699), 0], [0, 1.0, (-0.712328767123288), 0], [0, 0, 0, 0]]) m_diff = (m_rref - m_correct) assert (m_diff.norm() < 1e-15) assert (m_rref[(2, 2)] == 0)
[ "def", "test_iszero_substitution", "(", ")", ":", "m", "=", "Matrix", "(", "[", "[", "0.9", ",", "(", "-", "0.1", ")", ",", "(", "-", "0.2", ")", ",", "0", "]", ",", "[", "(", "-", "0.8", ")", ",", "0.9", ",", "(", "-", "0.4", ")", ",", "...
when doing numerical computations .
train
false
16,543
def net_send_object(sock, obj): data = pickle.dumps(obj, pickle.HIGHEST_PROTOCOL) sock.sendall(('%10d' % len(data))) sock.sendall(data)
[ "def", "net_send_object", "(", "sock", ",", "obj", ")", ":", "data", "=", "pickle", ".", "dumps", "(", "obj", ",", "pickle", ".", "HIGHEST_PROTOCOL", ")", "sock", ".", "sendall", "(", "(", "'%10d'", "%", "len", "(", "data", ")", ")", ")", "sock", "...
send python object over network .
train
false
16,546
def _define_nrt_meminfo_data(module): fn = module.get_or_insert_function(meminfo_data_ty, name='NRT_MemInfo_data_fast') builder = ir.IRBuilder(fn.append_basic_block()) [ptr] = fn.args struct_ptr = builder.bitcast(ptr, _meminfo_struct_type.as_pointer()) data_ptr = builder.load(cgutils.gep(builder, struct_ptr, 0, 3)) builder.ret(data_ptr)
[ "def", "_define_nrt_meminfo_data", "(", "module", ")", ":", "fn", "=", "module", ".", "get_or_insert_function", "(", "meminfo_data_ty", ",", "name", "=", "'NRT_MemInfo_data_fast'", ")", "builder", "=", "ir", ".", "IRBuilder", "(", "fn", ".", "append_basic_block", ...
implement nrt_meminfo_data_fast in the module .
train
false
16,547
@task @write def calc_checksum(theme_id, **kw): lfs = LocalFileStorage() theme = Persona.objects.get(id=theme_id) header = theme.header_path footer = theme.footer_path try: Image.open(header) Image.open(footer) except IOError: log.info(('Deleting invalid theme [%s] (header: %s) (footer: %s)' % (theme.addon.id, header, footer))) theme.addon.delete() theme.delete() rm_stored_dir(header.replace('header.png', ''), storage=lfs) return try: theme.checksum = make_checksum(header, footer) theme.save() except IOError as e: log.error(str(e))
[ "@", "task", "@", "write", "def", "calc_checksum", "(", "theme_id", ",", "**", "kw", ")", ":", "lfs", "=", "LocalFileStorage", "(", ")", "theme", "=", "Persona", ".", "objects", ".", "get", "(", "id", "=", "theme_id", ")", "header", "=", "theme", "."...
for migration 596 .
train
false
16,549
def test_get_syslog_facility_empty(monkeypatch): assert (os.getenv('WALE_SYSLOG_FACILITY') is None) (out, valid_facility) = log_help.get_syslog_facility() assert (valid_facility is True) assert (out == handlers.SysLogHandler.LOG_USER)
[ "def", "test_get_syslog_facility_empty", "(", "monkeypatch", ")", ":", "assert", "(", "os", ".", "getenv", "(", "'WALE_SYSLOG_FACILITY'", ")", "is", "None", ")", "(", "out", ",", "valid_facility", ")", "=", "log_help", ".", "get_syslog_facility", "(", ")", "as...
wale_syslog_facility is not set .
train
false
16,551
def predict_help_ver(args): (ns, _) = HELP_VER_PREDICTOR_PARSER.parse_known_args(args) pred = ((ns.help is not None) or (ns.version is not None)) return pred
[ "def", "predict_help_ver", "(", "args", ")", ":", "(", "ns", ",", "_", ")", "=", "HELP_VER_PREDICTOR_PARSER", ".", "parse_known_args", "(", "args", ")", "pred", "=", "(", "(", "ns", ".", "help", "is", "not", "None", ")", "or", "(", "ns", ".", "versio...
precict the backgroundability of commands that have help & version switches: -h .
train
false
16,552
def ensure_arg(args, arg, param=None): found = False for (idx, found_arg) in enumerate(args): if (found_arg == arg): if (param is not None): args[(idx + 1)] = param return args if (not found): args += [arg] if (param is not None): args += [param] return args
[ "def", "ensure_arg", "(", "args", ",", "arg", ",", "param", "=", "None", ")", ":", "found", "=", "False", "for", "(", "idx", ",", "found_arg", ")", "in", "enumerate", "(", "args", ")", ":", "if", "(", "found_arg", "==", "arg", ")", ":", "if", "("...
make sure the arg is present in the list of args .
train
false
16,553
def check_resolver(resolver): check_method = getattr(resolver, 'check', None) if (check_method is not None): return check_method() elif (not hasattr(resolver, 'resolve')): return get_warning_for_invalid_pattern(resolver) else: return []
[ "def", "check_resolver", "(", "resolver", ")", ":", "check_method", "=", "getattr", "(", "resolver", ",", "'check'", ",", "None", ")", "if", "(", "check_method", "is", "not", "None", ")", ":", "return", "check_method", "(", ")", "elif", "(", "not", "hasa...
recursively check the resolver .
train
false
16,554
def _cannotInstallHandler(fd): raise RuntimeError('Cannot install a SIGCHLD handler')
[ "def", "_cannotInstallHandler", "(", "fd", ")", ":", "raise", "RuntimeError", "(", "'Cannot install a SIGCHLD handler'", ")" ]
fail to install a signal handler for i{sigchld} .
train
false
16,555
def is_dictlist(data): if isinstance(data, list): for element in data: if isinstance(element, dict): if (len(element) != 1): return False else: return False return True return False
[ "def", "is_dictlist", "(", "data", ")", ":", "if", "isinstance", "(", "data", ",", "list", ")", ":", "for", "element", "in", "data", ":", "if", "isinstance", "(", "element", ",", "dict", ")", ":", "if", "(", "len", "(", "element", ")", "!=", "1", ...
returns true if data is a list of one-element dicts .
train
true
16,557
@requires_segment_info def capslock_indicator(pl, segment_info, text=u'CAPS'): if (not vim_func_exists(u'CapsLockStatusline')): return None return (text if vim.eval(u'CapsLockStatusline()') else None)
[ "@", "requires_segment_info", "def", "capslock_indicator", "(", "pl", ",", "segment_info", ",", "text", "=", "u'CAPS'", ")", ":", "if", "(", "not", "vim_func_exists", "(", "u'CapsLockStatusline'", ")", ")", ":", "return", "None", "return", "(", "text", "if", ...
shows the indicator if tpope/vim-capslock plugin is enabled .
train
false
16,558
def reverse_cuthill_mckee_ordering(G, heuristic=None): return reversed(list(cuthill_mckee_ordering(G, heuristic=heuristic)))
[ "def", "reverse_cuthill_mckee_ordering", "(", "G", ",", "heuristic", "=", "None", ")", ":", "return", "reversed", "(", "list", "(", "cuthill_mckee_ordering", "(", "G", ",", "heuristic", "=", "heuristic", ")", ")", ")" ]
generate an ordering of the graph nodes to make a sparse matrix .
train
false
16,559
def get_flavor_access_by_flavor_id(flavorid, ctxt=None): if (ctxt is None): ctxt = context.get_admin_context() flavor = objects.Flavor.get_by_flavor_id(ctxt, flavorid) return flavor.projects
[ "def", "get_flavor_access_by_flavor_id", "(", "flavorid", ",", "ctxt", "=", "None", ")", ":", "if", "(", "ctxt", "is", "None", ")", ":", "ctxt", "=", "context", ".", "get_admin_context", "(", ")", "flavor", "=", "objects", ".", "Flavor", ".", "get_by_flavo...
retrieve flavor access list by flavor id .
train
false
16,562
def getFaceGivenLines(triangleMesh, vertexStartIndex, vertexIndexTable, vertexes): faceGivenLines = face.Face() faceGivenLines.index = len(triangleMesh.faces) for vertexIndex in xrange(vertexStartIndex, (vertexStartIndex + 3)): vertex = vertexes[vertexIndex] vertexUniqueIndex = len(vertexIndexTable) if (str(vertex) in vertexIndexTable): vertexUniqueIndex = vertexIndexTable[str(vertex)] else: vertexIndexTable[str(vertex)] = vertexUniqueIndex triangleMesh.vertexes.append(vertex) faceGivenLines.vertexIndexes.append(vertexUniqueIndex) return faceGivenLines
[ "def", "getFaceGivenLines", "(", "triangleMesh", ",", "vertexStartIndex", ",", "vertexIndexTable", ",", "vertexes", ")", ":", "faceGivenLines", "=", "face", ".", "Face", "(", ")", "faceGivenLines", ".", "index", "=", "len", "(", "triangleMesh", ".", "faces", "...
add face given line index and lines .
train
false
16,563
def jsmin(js): if (not is_3): if (cStringIO and (not isinstance(js, unicode))): klass = cStringIO.StringIO else: klass = StringIO.StringIO else: klass = io.StringIO ins = klass(js) outs = klass() JavascriptMinify(ins, outs).minify() return outs.getvalue()
[ "def", "jsmin", "(", "js", ")", ":", "if", "(", "not", "is_3", ")", ":", "if", "(", "cStringIO", "and", "(", "not", "isinstance", "(", "js", ",", "unicode", ")", ")", ")", ":", "klass", "=", "cStringIO", ".", "StringIO", "else", ":", "klass", "="...
returns a minified version of the javascript string .
train
true
16,564
def libvlc_media_list_player_previous(p_mlp): f = (_Cfunctions.get('libvlc_media_list_player_previous', None) or _Cfunction('libvlc_media_list_player_previous', ((1,),), None, ctypes.c_int, MediaListPlayer)) return f(p_mlp)
[ "def", "libvlc_media_list_player_previous", "(", "p_mlp", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_media_list_player_previous'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_media_list_player_previous'", ",", "(", "(", "1", ",", ")"...
play previous item from media list .
train
true
16,565
def is_path_within_repo(app, path, repository_id): repo_path = os.path.abspath(repository_util.get_repository_by_id(app, repository_id).repo_path(app)) resolved_path = os.path.realpath(path) return (os.path.commonprefix([repo_path, resolved_path]) == repo_path)
[ "def", "is_path_within_repo", "(", "app", ",", "path", ",", "repository_id", ")", ":", "repo_path", "=", "os", ".", "path", ".", "abspath", "(", "repository_util", ".", "get_repository_by_id", "(", "app", ",", "repository_id", ")", ".", "repo_path", "(", "ap...
detect whether the given path is within the repository folder on the disk .
train
false
16,566
def _strip_schema(url): result = parse_url(url) return (result.netloc + result.path)
[ "def", "_strip_schema", "(", "url", ")", ":", "result", "=", "parse_url", "(", "url", ")", "return", "(", "result", ".", "netloc", "+", "result", ".", "path", ")" ]
returns the url without the s3:// part .
train
false
16,567
def test_settings_with_multiple_kwargs(): env.testval1 = 'outer 1' env.testval2 = 'outer 2' with settings(testval1='inner 1', testval2='inner 2'): eq_(env.testval1, 'inner 1') eq_(env.testval2, 'inner 2') eq_(env.testval1, 'outer 1') eq_(env.testval2, 'outer 2')
[ "def", "test_settings_with_multiple_kwargs", "(", ")", ":", "env", ".", "testval1", "=", "'outer 1'", "env", ".", "testval2", "=", "'outer 2'", "with", "settings", "(", "testval1", "=", "'inner 1'", ",", "testval2", "=", "'inner 2'", ")", ":", "eq_", "(", "e...
settings() should temporarily override env dict with given key/value pairs .
train
false
16,568
def test_topic_tracker_needs_update_cleared(database, user, topic): forumsread = ForumsRead.query.filter((ForumsRead.user_id == user.id), (ForumsRead.forum_id == topic.forum_id)).first() topicsread = TopicsRead.query.filter((TopicsRead.user_id == user.id), (TopicsRead.topic_id == topic.id)).first() with current_app.test_request_context(): assert topic.tracker_needs_update(forumsread, topicsread) forumsread = ForumsRead() forumsread.user_id = user.id forumsread.forum_id = topic.forum_id forumsread.last_read = datetime.utcnow() forumsread.cleared = datetime.utcnow() forumsread.save() assert (not topic.tracker_needs_update(forumsread, topicsread))
[ "def", "test_topic_tracker_needs_update_cleared", "(", "database", ",", "user", ",", "topic", ")", ":", "forumsread", "=", "ForumsRead", ".", "query", ".", "filter", "(", "(", "ForumsRead", ".", "user_id", "==", "user", ".", "id", ")", ",", "(", "ForumsRead"...
tests if the topicsread needs an update if the forum has been marked as cleared .
train
false