id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
1,737
def is_sqrt(expr): return (expr.is_Pow and expr.exp.is_Rational and (abs(expr.exp) is S.Half))
[ "def", "is_sqrt", "(", "expr", ")", ":", "return", "(", "expr", ".", "is_Pow", "and", "expr", ".", "exp", ".", "is_Rational", "and", "(", "abs", "(", "expr", ".", "exp", ")", "is", "S", ".", "Half", ")", ")" ]
return true if expr is a sqrt .
train
false
1,738
def create_textfile_with_contents(filename, contents, encoding='utf-8'): ensure_directory_exists(os.path.dirname(filename)) if os.path.exists(filename): os.remove(filename) outstream = codecs.open(filename, 'w', encoding) outstream.write(contents) if (contents and (not contents.endswith('\n'))): outstream.write('\n') outstream.flush() outstream.close() assert os.path.exists(filename), ('ENSURE file exists: %s' % filename)
[ "def", "create_textfile_with_contents", "(", "filename", ",", "contents", ",", "encoding", "=", "'utf-8'", ")", ":", "ensure_directory_exists", "(", "os", ".", "path", ".", "dirname", "(", "filename", ")", ")", "if", "os", ".", "path", ".", "exists", "(", ...
creates a textual file with the provided contents in the workdir .
train
true
1,739
def _fixed_cbLogin(self, xxx_todo_changeme): (interface, avatar, logout) = xxx_todo_changeme if (not IJellyable.providedBy(avatar)): avatar = pb.AsReferenceable(avatar, 'perspective') puid = avatar.processUniqueID() logout = [logout] def maybeLogout(): if (not logout): return fn = logout[0] del logout[0] fn() self.broker._localCleanup[puid] = maybeLogout self.broker.notifyOnDisconnect(maybeLogout) return avatar
[ "def", "_fixed_cbLogin", "(", "self", ",", "xxx_todo_changeme", ")", ":", "(", "interface", ",", "avatar", ",", "logout", ")", "=", "xxx_todo_changeme", "if", "(", "not", "IJellyable", ".", "providedBy", "(", "avatar", ")", ")", ":", "avatar", "=", "pb", ...
ensure that the avatar to be returned to the client is jellyable and set up disconnection notification to call the realms logout object .
train
false
1,740
def getSupportedKeyExchanges(): from cryptography.hazmat.backends import default_backend from cryptography.hazmat.primitives.asymmetric import ec from twisted.conch.ssh.keys import _curveTable backend = default_backend() kexAlgorithms = _kexAlgorithms.copy() for keyAlgorithm in list(kexAlgorithms): if keyAlgorithm.startswith('ecdh'): keyAlgorithmDsa = keyAlgorithm.replace('ecdh', 'ecdsa') supported = backend.elliptic_curve_exchange_algorithm_supported(ec.ECDH(), _curveTable[keyAlgorithmDsa]) if (not supported): kexAlgorithms.pop(keyAlgorithm) return sorted(kexAlgorithms, key=(lambda kexAlgorithm: kexAlgorithms[kexAlgorithm].preference))
[ "def", "getSupportedKeyExchanges", "(", ")", ":", "from", "cryptography", ".", "hazmat", ".", "backends", "import", "default_backend", "from", "cryptography", ".", "hazmat", ".", "primitives", ".", "asymmetric", "import", "ec", "from", "twisted", ".", "conch", "...
get a list of supported key exchange algorithm names in order of preference .
train
false
1,741
def _format_range_unified(start, stop): beginning = (start + 1) length = (stop - start) if (length == 1): return '{}'.format(beginning) if (not length): beginning -= 1 return '{},{}'.format(beginning, length)
[ "def", "_format_range_unified", "(", "start", ",", "stop", ")", ":", "beginning", "=", "(", "start", "+", "1", ")", "length", "=", "(", "stop", "-", "start", ")", "if", "(", "length", "==", "1", ")", ":", "return", "'{}'", ".", "format", "(", "begi...
convert range to the "ed" format .
train
true
1,742
def handle_register(request, text_template=None, html_template=None, subject=None, email_data=None, *args, **kwargs): if (request.method == 'POST'): form = RegisterForm(request.POST) if form.is_valid(): form = try_send_email_with_form(RegistrationProfile.objects.create_inactive_user, form, 'email', form.cleaned_data['username'], form.cleaned_data['password'], form.cleaned_data['email'], locale=request.LANGUAGE_CODE, text_template=text_template, html_template=html_template, subject=subject, email_data=email_data, volunteer_interest=form.cleaned_data['interested'], *args, **kwargs) if (not form.is_valid()): User.objects.filter(email=form.instance.email).delete() else: statsd.incr('user.register') return form return RegisterForm()
[ "def", "handle_register", "(", "request", ",", "text_template", "=", "None", ",", "html_template", "=", "None", ",", "subject", "=", "None", ",", "email_data", "=", "None", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "request", ".", "me...
handle to help registration .
train
false
1,744
@handle_response_format @treeio_login_required @module_admin_required() def pagefolder_add(request, response_format='html'): if request.POST: form = PageFolderForm(request.POST) if form.is_valid(): folder = form.save() return HttpResponseRedirect(reverse('core_admin_pagefolder_view', args=[folder.id])) else: form = PageFolderForm() return render_to_response('core/administration/pagefolder_add', {'form': form}, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "@", "module_admin_required", "(", ")", "def", "pagefolder_add", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "if", "request", ".", "POST", ":", "form", "=", "PageFolderForm", "(", ...
folder for static pages add .
train
false
1,745
def check_fixed_len_bcs_dups(header, mapping_data, errors): header_field_to_check = 'BarcodeSequence' try: check_ix = header.index(header_field_to_check) except ValueError: return errors barcodes = [] correction = 1 for curr_data in mapping_data: barcodes.append(upper(curr_data[check_ix])) dups = duplicates_indices(barcodes) for curr_dup in dups: for curr_loc in dups[curr_dup]: errors.append(('Duplicate barcode %s found. DCTB %d,%d' % (curr_dup, (curr_loc + correction), check_ix))) return errors
[ "def", "check_fixed_len_bcs_dups", "(", "header", ",", "mapping_data", ",", "errors", ")", ":", "header_field_to_check", "=", "'BarcodeSequence'", "try", ":", "check_ix", "=", "header", ".", "index", "(", "header_field_to_check", ")", "except", "ValueError", ":", ...
checks barcodes of same length for duplicates .
train
false
1,746
def all_index_generator(k=10): all_make_index_funcs = [makeIntIndex, makeFloatIndex, makeStringIndex, makeUnicodeIndex, makeDateIndex, makePeriodIndex, makeTimedeltaIndex, makeBoolIndex, makeCategoricalIndex] for make_index_func in all_make_index_funcs: (yield make_index_func(k=k))
[ "def", "all_index_generator", "(", "k", "=", "10", ")", ":", "all_make_index_funcs", "=", "[", "makeIntIndex", ",", "makeFloatIndex", ",", "makeStringIndex", ",", "makeUnicodeIndex", ",", "makeDateIndex", ",", "makePeriodIndex", ",", "makeTimedeltaIndex", ",", "make...
generator which can be iterated over to get instances of all the various index classes .
train
false
1,748
def structured(inputSchema, outputSchema, schema_store=None, ignore_body=False): if (schema_store is None): schema_store = {} inputValidator = getValidator(inputSchema, schema_store) outputValidator = getValidator(outputSchema, schema_store) def deco(original): @wraps(original) @_remote_logging @_logging @_serialize(outputValidator) def loadAndDispatch(self, request, **routeArguments): if ((request.method in ('GET', 'DELETE')) or ignore_body): objects = {} else: body = request.content.read() try: objects = loads(body) except ValueError: raise DECODING_ERROR errors = [] for error in inputValidator.iter_errors(objects): errors.append(error.message) if errors: raise InvalidRequestJSON(errors=errors, schema=inputSchema) objects.update(routeArguments) return maybeDeferred(original, self, **objects) loadAndDispatch.inputSchema = inputSchema loadAndDispatch.outputSchema = outputSchema return loadAndDispatch return deco
[ "def", "structured", "(", "inputSchema", ",", "outputSchema", ",", "schema_store", "=", "None", ",", "ignore_body", "=", "False", ")", ":", "if", "(", "schema_store", "is", "None", ")", ":", "schema_store", "=", "{", "}", "inputValidator", "=", "getValidator...
decorate a klein-style endpoint method so that the request body is automatically decoded and the response body is automatically encoded .
train
false
1,749
def interleave(inter, f, seq): seq = iter(seq) try: f(seq.next()) except StopIteration: pass else: for x in seq: inter() f(x)
[ "def", "interleave", "(", "inter", ",", "f", ",", "seq", ")", ":", "seq", "=", "iter", "(", "seq", ")", "try", ":", "f", "(", "seq", ".", "next", "(", ")", ")", "except", "StopIteration", ":", "pass", "else", ":", "for", "x", "in", "seq", ":", ...
call f on each item in seq .
train
true
1,750
def compress(body, compress_level): import zlib (yield '\x1f\x8b') (yield '\x08') (yield '\x00') (yield struct.pack('<L', (int(time.time()) & 4294967295L))) (yield '\x02') (yield '\xff') crc = zlib.crc32('') size = 0 zobj = zlib.compressobj(compress_level, zlib.DEFLATED, (- zlib.MAX_WBITS), zlib.DEF_MEM_LEVEL, 0) for line in body: size += len(line) crc = zlib.crc32(line, crc) (yield zobj.compress(line)) (yield zobj.flush()) (yield struct.pack('<L', (crc & 4294967295L))) (yield struct.pack('<L', (size & 4294967295L)))
[ "def", "compress", "(", "body", ",", "compress_level", ")", ":", "import", "zlib", "(", "yield", "'\\x1f\\x8b'", ")", "(", "yield", "'\\x08'", ")", "(", "yield", "'\\x00'", ")", "(", "yield", "struct", ".", "pack", "(", "'<L'", ",", "(", "int", "(", ...
compress a block of data .
train
false
1,751
@require_GET def contributors(request): product = _get_product(request) category = _get_category(request) return render_readouts(request, CONTRIBUTOR_READOUTS, 'contributors.html', locale=settings.WIKI_DEFAULT_LANGUAGE, product=product, extra_data={'overview_rows': kb_overview_rows(locale=request.LANGUAGE_CODE, product=product, mode=smart_int(request.GET.get('mode'), None), max=smart_int(request.GET.get('max'), 10), category=category), 'overview_modes': PERIODS, 'category': category, 'categories': CATEGORIES})
[ "@", "require_GET", "def", "contributors", "(", "request", ")", ":", "product", "=", "_get_product", "(", "request", ")", "category", "=", "_get_category", "(", "request", ")", "return", "render_readouts", "(", "request", ",", "CONTRIBUTOR_READOUTS", ",", "'cont...
render aggregate data about the articles in the default locale .
train
false
1,752
def qt5_qml_plugins_datas(directory): datas = [] qmldir = qt5_qml_dir() qt5_qml_plugin_dir = os.path.join(qmldir, directory) files = [] for (root, _dirs, _files) in os.walk(qt5_qml_plugin_dir): files.extend(misc.files_in_dir(root, ['qmldir', '*.qmltypes'])) for f in files: relpath = os.path.relpath(f, qmldir) (instdir, file) = os.path.split(relpath) instdir = os.path.join('qml', instdir) logger.debug(('qt5_qml_plugins_datas installing %s in %s' % (f, instdir))) datas.append((f, instdir)) return datas
[ "def", "qt5_qml_plugins_datas", "(", "directory", ")", ":", "datas", "=", "[", "]", "qmldir", "=", "qt5_qml_dir", "(", ")", "qt5_qml_plugin_dir", "=", "os", ".", "path", ".", "join", "(", "qmldir", ",", "directory", ")", "files", "=", "[", "]", "for", ...
return list of data files for mod .
train
false
1,753
def readpipe(argv, preexec_fn=None, shell=False): p = subprocess.Popen(argv, stdout=subprocess.PIPE, preexec_fn=preexec_fn, shell=shell) (out, err) = p.communicate() if (p.returncode != 0): raise Exception(('subprocess %r failed with status %d' % (' '.join(argv), p.returncode))) return out
[ "def", "readpipe", "(", "argv", ",", "preexec_fn", "=", "None", ",", "shell", "=", "False", ")", ":", "p", "=", "subprocess", ".", "Popen", "(", "argv", ",", "stdout", "=", "subprocess", ".", "PIPE", ",", "preexec_fn", "=", "preexec_fn", ",", "shell", ...
run a subprocess and return its output .
train
false
1,756
def _make_update_dict(update): return {'id': update['id'], 'date': update['date'], 'content': update['content']}
[ "def", "_make_update_dict", "(", "update", ")", ":", "return", "{", "'id'", ":", "update", "[", "'id'", "]", ",", "'date'", ":", "update", "[", "'date'", "]", ",", "'content'", ":", "update", "[", "'content'", "]", "}" ]
return course update item as a dictionary with required keys .
train
false
1,757
@intercept_errors(UserAPIInternalError, ignore_errors=[UserAPIRequestError]) def update_email_opt_in(user, org, opt_in): (preference, _) = UserOrgTag.objects.get_or_create(user=user, org=org, key='email-optin') try: user_profile = UserProfile.objects.get(user=user) except ObjectDoesNotExist: raise UserNotFound() if user_profile.requires_parental_consent(age_limit=getattr(settings, 'EMAIL_OPTIN_MINIMUM_AGE', 13), default_requires_consent=False): opt_in = False preference.value = str(opt_in) try: preference.save() if (hasattr(settings, 'LMS_SEGMENT_KEY') and settings.LMS_SEGMENT_KEY): _track_update_email_opt_in(user.id, org, opt_in) except IntegrityError as err: log.warn(u'Could not update organization wide preference due to IntegrityError: {}'.format(err.message))
[ "@", "intercept_errors", "(", "UserAPIInternalError", ",", "ignore_errors", "=", "[", "UserAPIRequestError", "]", ")", "def", "update_email_opt_in", "(", "user", ",", "org", ",", "opt_in", ")", ":", "(", "preference", ",", "_", ")", "=", "UserOrgTag", ".", "...
updates a users preference for receiving org-wide emails .
train
false
1,758
def load_backend(build_configuration, backend_package): backend_module = (backend_package + u'.register') try: module = importlib.import_module(backend_module) except ImportError as e: traceback.print_exc() raise BackendConfigurationError(u'Failed to load the {backend} backend: {error}'.format(backend=backend_module, error=e)) def invoke_entrypoint(name): entrypoint = getattr(module, name, (lambda : None)) try: return entrypoint() except TypeError as e: traceback.print_exc() raise BackendConfigurationError(u'Entrypoint {entrypoint} in {backend} must be a zero-arg callable: {error}'.format(entrypoint=name, backend=backend_module, error=e)) build_file_aliases = invoke_entrypoint(u'build_file_aliases') if build_file_aliases: build_configuration.register_aliases(build_file_aliases) subsystems = invoke_entrypoint(u'global_subsystems') if subsystems: build_configuration.register_subsystems(subsystems) invoke_entrypoint(u'register_goals')
[ "def", "load_backend", "(", "build_configuration", ",", "backend_package", ")", ":", "backend_module", "=", "(", "backend_package", "+", "u'.register'", ")", "try", ":", "module", "=", "importlib", ".", "import_module", "(", "backend_module", ")", "except", "Impor...
loads a backend for interacting with the search engine .
train
false
1,759
def get_umc_admin_objects(): import univention.admin return univention.admin.objects
[ "def", "get_umc_admin_objects", "(", ")", ":", "import", "univention", ".", "admin", "return", "univention", ".", "admin", ".", "objects" ]
convenience accessor for getting univention .
train
false
1,760
def obtain_serial_number(show_ver): match = re.search('Processor board ID (.+)', show_ver) if match: return match.group(1).strip() else: return None
[ "def", "obtain_serial_number", "(", "show_ver", ")", ":", "match", "=", "re", ".", "search", "(", "'Processor board ID (.+)'", ",", "show_ver", ")", "if", "match", ":", "return", "match", ".", "group", "(", "1", ")", ".", "strip", "(", ")", "else", ":", ...
example string from cisco ios: processor board id ftx1000008x return the serial_number .
train
false
1,763
def convert_serializable(records): for r in records: r.msg = r.getMessage() r.args = ()
[ "def", "convert_serializable", "(", "records", ")", ":", "for", "r", "in", "records", ":", "r", ".", "msg", "=", "r", ".", "getMessage", "(", ")", "r", ".", "args", "=", "(", ")" ]
convert logrecord serializable .
train
false
1,764
def check_valid_naming(pattern=None, multi=None, anime_type=None): if (pattern is None): pattern = sickbeard.NAMING_PATTERN if (anime_type is None): anime_type = sickbeard.NAMING_ANIME logger.log(((u'Checking whether the pattern ' + pattern) + u' is valid for a single episode'), logger.DEBUG) valid = validate_name(pattern, None, anime_type) if (multi is not None): logger.log(((u'Checking whether the pattern ' + pattern) + u' is valid for a multi episode'), logger.DEBUG) valid = (valid and validate_name(pattern, multi, anime_type)) return valid
[ "def", "check_valid_naming", "(", "pattern", "=", "None", ",", "multi", "=", "None", ",", "anime_type", "=", "None", ")", ":", "if", "(", "pattern", "is", "None", ")", ":", "pattern", "=", "sickbeard", ".", "NAMING_PATTERN", "if", "(", "anime_type", "is"...
checks if the name is can be parsed back to its original form for both single and multi episodes .
train
false
1,765
def volume_attach(name, server_name, device='/dev/xvdb', profile=None, timeout=300): conn = _auth(profile) return conn.volume_attach(name, server_name, device, timeout)
[ "def", "volume_attach", "(", "name", ",", "server_name", ",", "device", "=", "'/dev/xvdb'", ",", "profile", "=", "None", ",", "timeout", "=", "300", ")", ":", "conn", "=", "_auth", "(", "profile", ")", "return", "conn", ".", "volume_attach", "(", "name",...
attach a volume .
train
true
1,766
@not_implemented_for('undirected') def out_degree_centrality(G): centrality = {} s = (1.0 / (len(G) - 1.0)) centrality = {n: (d * s) for (n, d) in G.out_degree()} return centrality
[ "@", "not_implemented_for", "(", "'undirected'", ")", "def", "out_degree_centrality", "(", "G", ")", ":", "centrality", "=", "{", "}", "s", "=", "(", "1.0", "/", "(", "len", "(", "G", ")", "-", "1.0", ")", ")", "centrality", "=", "{", "n", ":", "("...
compute the out-degree centrality for nodes .
train
false
1,767
def set_sharing_strategy(new_strategy): global _sharing_strategy assert (new_strategy in _all_sharing_strategies) _sharing_strategy = new_strategy
[ "def", "set_sharing_strategy", "(", "new_strategy", ")", ":", "global", "_sharing_strategy", "assert", "(", "new_strategy", "in", "_all_sharing_strategies", ")", "_sharing_strategy", "=", "new_strategy" ]
sets the strategy for sharing cpu tensors .
train
false
1,769
def get_dup_labels_perc(fasta_labels): fasta_labels_count = float(len(fasta_labels)) fasta_labels_derep = float(len(set(fasta_labels))) perc_dup = ('%1.3f' % ((fasta_labels_count - fasta_labels_derep) / fasta_labels_count)) label_counts = defaultdict(int) for curr_label in fasta_labels: label_counts[curr_label] += 1 labels_from_dups = [] for label in label_counts: if (label_counts[label] > 1): labels_from_dups.append(label) return (perc_dup, labels_from_dups)
[ "def", "get_dup_labels_perc", "(", "fasta_labels", ")", ":", "fasta_labels_count", "=", "float", "(", "len", "(", "fasta_labels", ")", ")", "fasta_labels_derep", "=", "float", "(", "len", "(", "set", "(", "fasta_labels", ")", ")", ")", "perc_dup", "=", "(", ...
calculates percentage of sequences with duplicate labels fasta_labels: list of fasta labels .
train
false
1,770
def _write_batch_lmdb(db, batch, image_count): try: with db.begin(write=True) as lmdb_txn: for (i, datum) in enumerate(batch): key = ('%08d_%d' % ((image_count + i), datum.label)) lmdb_txn.put(key, datum.SerializeToString()) except lmdb.MapFullError: curr_limit = db.info()['map_size'] new_limit = (curr_limit * 2) try: db.set_mapsize(new_limit) except AttributeError as e: version = tuple((int(x) for x in lmdb.__version__.split('.'))) if (version < (0, 87)): raise Error(('py-lmdb is out of date (%s vs 0.87)' % lmdb.__version__)) else: raise e _write_batch_lmdb(db, batch, image_count)
[ "def", "_write_batch_lmdb", "(", "db", ",", "batch", ",", "image_count", ")", ":", "try", ":", "with", "db", ".", "begin", "(", "write", "=", "True", ")", "as", "lmdb_txn", ":", "for", "(", "i", ",", "datum", ")", "in", "enumerate", "(", "batch", "...
write a batch to an lmdb database .
train
false
1,772
def tag_list(context, data_dict): model = context['model'] vocab_id_or_name = data_dict.get('vocabulary_id') query = (data_dict.get('query') or data_dict.get('q')) if query: query = query.strip() all_fields = data_dict.get('all_fields', None) _check_access('tag_list', context, data_dict) if query: (tags, count) = _tag_search(context, data_dict) else: tags = model.Tag.all(vocab_id_or_name) if tags: if all_fields: tag_list = model_dictize.tag_list_dictize(tags, context) else: tag_list = [tag.name for tag in tags] else: tag_list = [] return tag_list
[ "def", "tag_list", "(", "context", ",", "data_dict", ")", ":", "model", "=", "context", "[", "'model'", "]", "vocab_id_or_name", "=", "data_dict", ".", "get", "(", "'vocabulary_id'", ")", "query", "=", "(", "data_dict", ".", "get", "(", "'query'", ")", "...
display list of tags .
train
false
1,773
def get_trending(obj, region=None): return _property_value_by_region(obj, region=region, property='trending')
[ "def", "get_trending", "(", "obj", ",", "region", "=", "None", ")", ":", "return", "_property_value_by_region", "(", "obj", ",", "region", "=", "region", ",", "property", "=", "'trending'", ")" ]
returns trending value for the given obj to use in elasticsearch .
train
false
1,775
def getElementsPath(subName=''): return getJoinedPath(getGeometryUtilitiesPath('evaluate_elements'), subName)
[ "def", "getElementsPath", "(", "subName", "=", "''", ")", ":", "return", "getJoinedPath", "(", "getGeometryUtilitiesPath", "(", "'evaluate_elements'", ")", ",", "subName", ")" ]
get the evaluate_elements directory path .
train
false
1,776
def rewrite_links(root, link_repl_func, resolve_base_href=False): from cssutils import replaceUrls, log, CSSParser log.setLevel(logging.WARN) log.raiseExceptions = False if resolve_base_href: resolve_base_href(root) for (el, attrib, link, pos) in iterlinks(root, find_links_in_css=False): new_link = link_repl_func(link.strip()) if (new_link == link): continue if (new_link is None): if (attrib is None): el.text = '' else: del el.attrib[attrib] continue if (attrib is None): new = ((el.text[:pos] + new_link) + el.text[(pos + len(link)):]) el.text = new else: cur = el.attrib[attrib] if ((not pos) and (len(cur) == len(link))): el.attrib[attrib] = new_link else: new = ((cur[:pos] + new_link) + cur[(pos + len(link)):]) el.attrib[attrib] = new parser = CSSParser(raiseExceptions=False, log=_css_logger, fetcher=(lambda x: (None, None))) for el in root.iter(etree.Element): try: tag = el.tag except UnicodeDecodeError: continue if ((tag == XHTML('style')) and el.text and ((_css_url_re.search(el.text) is not None) or ('@import' in el.text))): stylesheet = parser.parseString(el.text, validate=False) replaceUrls(stylesheet, link_repl_func) repl = stylesheet.cssText if isbytestring(repl): repl = repl.decode('utf-8') el.text = (('\n' + repl) + '\n') if ('style' in el.attrib): text = el.attrib['style'] if (_css_url_re.search(text) is not None): try: stext = parser.parseStyle(text, validate=False) except: continue replaceUrls(stext, link_repl_func) repl = stext.cssText.replace('\n', ' ').replace('\r', ' ') if isbytestring(repl): repl = repl.decode('utf-8') el.attrib['style'] = repl
[ "def", "rewrite_links", "(", "root", ",", "link_repl_func", ",", "resolve_base_href", "=", "False", ")", ":", "from", "cssutils", "import", "replaceUrls", ",", "log", ",", "CSSParser", "log", ".", "setLevel", "(", "logging", ".", "WARN", ")", "log", ".", "...
thin wrapper around lxmls rewrite_links() that prevents extra html markup from being produced when theres no root tag present in the input html .
train
false
1,777
def version_requirement(version): return (version >= REQUIRED_VERSION)
[ "def", "version_requirement", "(", "version", ")", ":", "return", "(", "version", ">=", "REQUIRED_VERSION", ")" ]
test for minimum required version .
train
false
1,778
def to_one_hot(y, nb_class, dtype=None): ret = theano.tensor.zeros((y.shape[0], nb_class), dtype=dtype) ret = theano.tensor.set_subtensor(ret[(theano.tensor.arange(y.shape[0]), y)], 1) return ret
[ "def", "to_one_hot", "(", "y", ",", "nb_class", ",", "dtype", "=", "None", ")", ":", "ret", "=", "theano", ".", "tensor", ".", "zeros", "(", "(", "y", ".", "shape", "[", "0", "]", ",", "nb_class", ")", ",", "dtype", "=", "dtype", ")", "ret", "=...
return a matrix where each row correspond to the one hot encoding of each element in y .
train
false
1,780
def get_object_jsonschema(question, required_fields, is_reviewer): object_jsonschema = {'type': 'object', 'additionalProperties': False, 'properties': {}} required = [] properties = question.get('properties') if properties: for property in properties: if (property.get('required', False) and required_fields): required.append(property['id']) values = extract_question_values(property, required_fields, is_reviewer) object_jsonschema['properties'][property['id']] = {'type': 'object', 'additionalProperties': False, 'properties': values} if required_fields: object_jsonschema['properties'][property['id']]['required'] = ['value'] if (required_fields and is_required(question)): object_jsonschema['required'] = required return object_jsonschema
[ "def", "get_object_jsonschema", "(", "question", ",", "required_fields", ",", "is_reviewer", ")", ":", "object_jsonschema", "=", "{", "'type'", ":", "'object'", ",", "'additionalProperties'", ":", "False", ",", "'properties'", ":", "{", "}", "}", "required", "="...
returns jsonschema for nested objects within schema .
train
false
1,781
def save_traceback(app): import platform exc = traceback.format_exc() (fd, path) = tempfile.mkstemp('.log', 'sphinx-err-') os.write(fd, (_DEBUG_HEADER % (sphinx.__version__, platform.python_version(), docutils.__version__, docutils.__version_details__, jinja2.__version__)).encode('utf-8')) if (app is not None): for (extname, extmod) in app._extensions.iteritems(): os.write(fd, ('# %s from %s\n' % (extname, getattr(extmod, '__file__', 'unknown'))).encode('utf-8')) os.write(fd, exc.encode('utf-8')) os.close(fd) return path
[ "def", "save_traceback", "(", "app", ")", ":", "import", "platform", "exc", "=", "traceback", ".", "format_exc", "(", ")", "(", "fd", ",", "path", ")", "=", "tempfile", ".", "mkstemp", "(", "'.log'", ",", "'sphinx-err-'", ")", "os", ".", "write", "(", ...
save the current exceptions traceback in a temporary file .
train
false
1,782
def vn_free_ar(call=None, kwargs=None): if (call != 'function'): raise SaltCloudSystemExit('The vn_free_ar function must be called with -f or --function.') if (kwargs is None): kwargs = {} vn_id = kwargs.get('vn_id', None) vn_name = kwargs.get('vn_name', None) ar_id = kwargs.get('ar_id', None) if (ar_id is None): raise SaltCloudSystemExit("The vn_free_ar function requires an 'rn_id' to be provided.") if vn_id: if vn_name: log.warning("Both the 'vn_id' and 'vn_name' arguments were provided. 'vn_id' will take precedence.") elif vn_name: vn_id = get_vn_id(kwargs={'name': vn_name}) else: raise SaltCloudSystemExit("The vn_free_ar function requires a 'vn_id' or a 'vn_name' to be provided.") (server, user, password) = _get_xml_rpc() auth = ':'.join([user, password]) response = server.one.vn.free_ar(auth, int(vn_id), int(ar_id)) data = {'action': 'vn.free_ar', 'ar_freed': response[0], 'resource_id': response[1], 'error_code': response[2]} return data
[ "def", "vn_free_ar", "(", "call", "=", "None", ",", "kwargs", "=", "None", ")", ":", "if", "(", "call", "!=", "'function'", ")", ":", "raise", "SaltCloudSystemExit", "(", "'The vn_free_ar function must be called with -f or --function.'", ")", "if", "(", "kwargs", ...
frees a reserved address range from a virtual network .
train
true
1,785
def unpickleStringI(val, sek): x = _cStringIO(val) x.seek(sek) return x
[ "def", "unpickleStringI", "(", "val", ",", "sek", ")", ":", "x", "=", "_cStringIO", "(", "val", ")", "x", ".", "seek", "(", "sek", ")", "return", "x" ]
convert the output of l{picklestringi} into an appropriate type for the current python version .
train
false
1,786
def CallSetAllowedModule(name, desired): if (USING_SDK and (name == 'django')): sys.path[:] = [dirname for dirname in sys.path if (not dirname.startswith(os.path.join(PYTHON_LIB, 'lib', 'django')))] if (desired in ('0.96', '1.2', '1.3')): sys.path.insert(1, os.path.join(PYTHON_LIB, 'lib', ('django-' + desired))) SetAllowedModule(name)
[ "def", "CallSetAllowedModule", "(", "name", ",", "desired", ")", ":", "if", "(", "USING_SDK", "and", "(", "name", "==", "'django'", ")", ")", ":", "sys", ".", "path", "[", ":", "]", "=", "[", "dirname", "for", "dirname", "in", "sys", ".", "path", "...
helper to call setallowedmodule .
train
false
1,788
def _assure_identity(fnc): def _wrapped(*args, **kwargs): if (identity is None): _create_identity() return fnc(*args, **kwargs) return _wrapped
[ "def", "_assure_identity", "(", "fnc", ")", ":", "def", "_wrapped", "(", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "identity", "is", "None", ")", ":", "_create_identity", "(", ")", "return", "fnc", "(", "*", "args", ",", "**", "kwargs", ...
ensures that the identity attribute is not none .
train
true
1,789
def add_error_patch(result): return make_instancemethod(TextTestResult.addError, result)
[ "def", "add_error_patch", "(", "result", ")", ":", "return", "make_instancemethod", "(", "TextTestResult", ".", "addError", ",", "result", ")" ]
create a new adderror method to patch into a result instance that recognizes the errorclasses attribute and deals with errorclasses correctly .
train
false
1,791
def setup_fog(): glEnable(GL_FOG) glFogfv(GL_FOG_COLOR, (GLfloat * 4)(0.5, 0.69, 1.0, 1)) glHint(GL_FOG_HINT, GL_DONT_CARE) glFogi(GL_FOG_MODE, GL_LINEAR) glFogf(GL_FOG_START, 20.0) glFogf(GL_FOG_END, 60.0)
[ "def", "setup_fog", "(", ")", ":", "glEnable", "(", "GL_FOG", ")", "glFogfv", "(", "GL_FOG_COLOR", ",", "(", "GLfloat", "*", "4", ")", "(", "0.5", ",", "0.69", ",", "1.0", ",", "1", ")", ")", "glHint", "(", "GL_FOG_HINT", ",", "GL_DONT_CARE", ")", ...
configure the opengl fog properties .
train
false
1,792
def _log_backport(ovo, target_version): if (target_version and (target_version != ovo.VERSION)): LOG.debug('Backporting %(obj_name)s from version %(src_vers)s to version %(dst_vers)s', {'obj_name': ovo.obj_name(), 'src_vers': ovo.VERSION, 'dst_vers': target_version})
[ "def", "_log_backport", "(", "ovo", ",", "target_version", ")", ":", "if", "(", "target_version", "and", "(", "target_version", "!=", "ovo", ".", "VERSION", ")", ")", ":", "LOG", ".", "debug", "(", "'Backporting %(obj_name)s from version %(src_vers)s to version %(ds...
log backported versioned objects .
train
false
1,793
def _fwd_bem_linear_collocation_solution(m): for surf in m['surfs']: complete_surface_info(surf, copy=False, verbose=False) logger.info('Computing the linear collocation solution...') logger.info(' Matrix coefficients...') coeff = _fwd_bem_lin_pot_coeff(m['surfs']) m['nsol'] = len(coeff) logger.info(' Inverting the coefficient matrix...') nps = [surf['np'] for surf in m['surfs']] m['solution'] = _fwd_bem_multi_solution(coeff, m['gamma'], nps) if (len(m['surfs']) == 3): ip_mult = (m['sigma'][1] / m['sigma'][2]) if (ip_mult <= FIFF.FWD_BEM_IP_APPROACH_LIMIT): logger.info('IP approach required...') logger.info(' Matrix coefficients (homog)...') coeff = _fwd_bem_lin_pot_coeff([m['surfs'][(-1)]]) logger.info(' Inverting the coefficient matrix (homog)...') ip_solution = _fwd_bem_homog_solution(coeff, [m['surfs'][(-1)]['np']]) logger.info(' Modify the original solution to incorporate IP approach...') _fwd_bem_ip_modify_solution(m['solution'], ip_solution, ip_mult, nps) m['bem_method'] = FIFF.FWD_BEM_LINEAR_COLL logger.info('Solution ready.')
[ "def", "_fwd_bem_linear_collocation_solution", "(", "m", ")", ":", "for", "surf", "in", "m", "[", "'surfs'", "]", ":", "complete_surface_info", "(", "surf", ",", "copy", "=", "False", ",", "verbose", "=", "False", ")", "logger", ".", "info", "(", "'Computi...
compute the linear collocation potential solution .
train
false
1,794
def polynomial_multiply_mod(m1, m2, polymod, p): prod = (((len(m1) + len(m2)) - 1) * [0]) for i in range(len(m1)): for j in range(len(m2)): prod[(i + j)] = ((prod[(i + j)] + (m1[i] * m2[j])) % p) return polynomial_reduce_mod(prod, polymod, p)
[ "def", "polynomial_multiply_mod", "(", "m1", ",", "m2", ",", "polymod", ",", "p", ")", ":", "prod", "=", "(", "(", "(", "len", "(", "m1", ")", "+", "len", "(", "m2", ")", ")", "-", "1", ")", "*", "[", "0", "]", ")", "for", "i", "in", "range...
polynomial multiplication modulo a polynomial over ints mod p .
train
true
1,796
def buildSlicedNetwork(): N = FeedForwardNetwork('sliced') a = LinearLayer(2, name='a') b = LinearLayer(2, name='b') N.addInputModule(a) N.addOutputModule(b) N.addConnection(FullConnection(a, b, inSliceTo=1, outSliceFrom=1)) N.addConnection(FullConnection(a, b, inSliceFrom=1, outSliceTo=1)) N.sortModules() return N
[ "def", "buildSlicedNetwork", "(", ")", ":", "N", "=", "FeedForwardNetwork", "(", "'sliced'", ")", "a", "=", "LinearLayer", "(", "2", ",", "name", "=", "'a'", ")", "b", "=", "LinearLayer", "(", "2", ",", "name", "=", "'b'", ")", "N", ".", "addInputMod...
build a network with shared connections .
train
false
1,797
def lookupZone(name, timeout=None): return getResolver().lookupZone(name, timeout)
[ "def", "lookupZone", "(", "name", ",", "timeout", "=", "None", ")", ":", "return", "getResolver", "(", ")", ".", "lookupZone", "(", "name", ",", "timeout", ")" ]
perform an axfr record lookup .
train
false
1,798
def _find_function(name, region=None, key=None, keyid=None, profile=None): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) for funcs in salt.utils.boto3.paged_call(conn.list_functions): for func in funcs['Functions']: if (func['FunctionName'] == name): return func return None
[ "def", "_find_function", "(", "name", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "key", "=", "key", ",", "keyid...
given function name .
train
false
1,800
@blueprint.route('/jobs/<job_id>', methods=['GET']) def show_job(job_id): job = scheduler.get_job(job_id) if (job is None): raise werkzeug.exceptions.NotFound('Job not found') if isinstance(job, dataset.DatasetJob): return flask.redirect(flask.url_for('digits.dataset.views.show', job_id=job_id)) if isinstance(job, model.ModelJob): return flask.redirect(flask.url_for('digits.model.views.show', job_id=job_id)) if isinstance(job, pretrained_model.PretrainedModelJob): return flask.redirect(flask.url_for('digits.pretrained_model.views.show', job_id=job_id)) else: raise werkzeug.exceptions.BadRequest('Invalid job type')
[ "@", "blueprint", ".", "route", "(", "'/jobs/<job_id>'", ",", "methods", "=", "[", "'GET'", "]", ")", "def", "show_job", "(", "job_id", ")", ":", "job", "=", "scheduler", ".", "get_job", "(", "job_id", ")", "if", "(", "job", "is", "None", ")", ":", ...
redirects to the appropriate /datasets/ or /models/ page .
train
false
1,802
def _setup_selinux_for_keys(fs, sshdir): if (not fs.has_file(os.path.join('etc', 'selinux'))): return rclocal = os.path.join('etc', 'rc.local') rc_d = os.path.join('etc', 'rc.d') if ((not fs.has_file(rclocal)) and fs.has_file(rc_d)): rclocal = os.path.join(rc_d, 'rc.local') restorecon = ['\n', '# Added by Nova to ensure injected ssh keys have the right context\n', ('restorecon -RF %s 2>/dev/null || :\n' % sshdir)] if (not fs.has_file(rclocal)): restorecon.insert(0, '#!/bin/sh') _inject_file_into_fs(fs, rclocal, ''.join(restorecon), append=True) fs.set_permissions(rclocal, 448)
[ "def", "_setup_selinux_for_keys", "(", "fs", ",", "sshdir", ")", ":", "if", "(", "not", "fs", ".", "has_file", "(", "os", ".", "path", ".", "join", "(", "'etc'", ",", "'selinux'", ")", ")", ")", ":", "return", "rclocal", "=", "os", ".", "path", "."...
get selinux guests to ensure correct context on injected keys .
train
false
1,804
def firebase_delete(path): (response, content) = _get_http().request(path, method='DELETE')
[ "def", "firebase_delete", "(", "path", ")", ":", "(", "response", ",", "content", ")", "=", "_get_http", "(", ")", ".", "request", "(", "path", ",", "method", "=", "'DELETE'", ")" ]
removes the data at a particular path .
train
false
1,806
def create_image_metadata(data): disk_format = data['disk_format'] if (disk_format in ('ami', 'aki', 'ari')): container_format = disk_format elif (disk_format == 'docker'): disk_format = 'raw' container_format = 'docker' else: container_format = 'bare' meta = {'protected': data['protected'], 'disk_format': disk_format, 'container_format': container_format, 'min_disk': (data['minimum_disk'] or 0), 'min_ram': (data['minimum_ram'] or 0), 'name': data['name']} is_public = data.get('is_public', data.get('public', False)) properties = {} if data.get('description'): properties['description'] = data['description'] if data.get('kernel'): properties['kernel_id'] = data['kernel'] if data.get('ramdisk'): properties['ramdisk_id'] = data['ramdisk'] if data.get('architecture'): properties['architecture'] = data['architecture'] if (api.glance.VERSIONS.active < 2): meta.update({'is_public': is_public, 'properties': properties}) else: meta['visibility'] = ('public' if is_public else 'private') meta.update(properties) return meta
[ "def", "create_image_metadata", "(", "data", ")", ":", "disk_format", "=", "data", "[", "'disk_format'", "]", "if", "(", "disk_format", "in", "(", "'ami'", ",", "'aki'", ",", "'ari'", ")", ")", ":", "container_format", "=", "disk_format", "elif", "(", "dis...
use the given dict of image form data to generate the metadata used for creating the image in glance .
train
false
1,807
def _get_recurse_set(recurse): if (not recurse): return set() if (not isinstance(recurse, list)): raise TypeError('"recurse" must be formed as a list of strings') try: recurse_set = set(recurse) except TypeError: recurse_set = None if ((recurse_set is None) or (not (set(_RECURSE_TYPES) >= recurse_set))): raise ValueError('Types for "recurse" limited to {0}.'.format(', '.join(('"{0}"'.format(rtype) for rtype in _RECURSE_TYPES)))) if (('ignore_files' in recurse_set) and ('ignore_dirs' in recurse_set)): raise ValueError('Must not specify "recurse" options "ignore_files" and "ignore_dirs" at the same time.') return recurse_set
[ "def", "_get_recurse_set", "(", "recurse", ")", ":", "if", "(", "not", "recurse", ")", ":", "return", "set", "(", ")", "if", "(", "not", "isinstance", "(", "recurse", ",", "list", ")", ")", ":", "raise", "TypeError", "(", "'\"recurse\" must be formed as a ...
converse *recurse* definition to a set of strings .
train
true
1,808
def get_service_module_name(service_model): name = service_model.metadata.get('serviceAbbreviation', service_model.metadata.get('serviceFullName', service_model.service_name)) name = name.replace('Amazon', '') name = name.replace('AWS', '') name = re.sub('\\W+', '', name) return name
[ "def", "get_service_module_name", "(", "service_model", ")", ":", "name", "=", "service_model", ".", "metadata", ".", "get", "(", "'serviceAbbreviation'", ",", "service_model", ".", "metadata", ".", "get", "(", "'serviceFullName'", ",", "service_model", ".", "serv...
returns the module name for a service this is the value used in both the documentation and client class name .
train
false
1,810
def s_one_set(topics): s_one_set = [] for top_words in topics: s_one_set_t = [] for w_prime in top_words: s_one_set_t.append((w_prime, top_words)) s_one_set.append(s_one_set_t) return s_one_set
[ "def", "s_one_set", "(", "topics", ")", ":", "s_one_set", "=", "[", "]", "for", "top_words", "in", "topics", ":", "s_one_set_t", "=", "[", "]", "for", "w_prime", "in", "top_words", ":", "s_one_set_t", ".", "append", "(", "(", "w_prime", ",", "top_words",...
this function performs s_one_set segmentation on a list of topics .
train
false
1,811
def reversed_arguments(func): def wrapped(*args): return func(*reversed(args)) return wrapped
[ "def", "reversed_arguments", "(", "func", ")", ":", "def", "wrapped", "(", "*", "args", ")", ":", "return", "func", "(", "*", "reversed", "(", "args", ")", ")", "return", "wrapped" ]
return a function with reversed argument order .
train
false
1,812
@require_http_methods('GET') @login_required @expect_json def xblock_container_handler(request, usage_key_string): usage_key = usage_key_with_run(usage_key_string) if (not has_studio_read_access(request.user, usage_key.course_key)): raise PermissionDenied() response_format = request.GET.get('format', 'html') if ((response_format == 'json') or ('application/json' in request.META.get('HTTP_ACCEPT', 'application/json'))): with modulestore().bulk_operations(usage_key.course_key): response = _get_module_info(_get_xblock(usage_key, request.user), include_ancestor_info=True, include_publishing_info=True) return JsonResponse(response) else: return Http404
[ "@", "require_http_methods", "(", "'GET'", ")", "@", "login_required", "@", "expect_json", "def", "xblock_container_handler", "(", "request", ",", "usage_key_string", ")", ":", "usage_key", "=", "usage_key_with_run", "(", "usage_key_string", ")", "if", "(", "not", ...
the restful handler for requests for xblock information about the block and its children .
train
false
1,813
def elink(**keywds): cgi = 'https://eutils.ncbi.nlm.nih.gov/entrez/eutils/elink.fcgi' variables = {} variables.update(keywds) return _open(cgi, variables)
[ "def", "elink", "(", "**", "keywds", ")", ":", "cgi", "=", "'https://eutils.ncbi.nlm.nih.gov/entrez/eutils/elink.fcgi'", "variables", "=", "{", "}", "variables", ".", "update", "(", "keywds", ")", "return", "_open", "(", "cgi", ",", "variables", ")" ]
elink checks for linked external articles and returns a handle .
train
false
1,816
def assertRaises(_exception, _callable=None, *args, **kwargs): manager = _AssertRaisesContextmanager(exception=_exception) if (_callable is not None): with manager: _callable(*args, **kwargs) else: return manager
[ "def", "assertRaises", "(", "_exception", ",", "_callable", "=", "None", ",", "*", "args", ",", "**", "kwargs", ")", ":", "manager", "=", "_AssertRaisesContextmanager", "(", "exception", "=", "_exception", ")", "if", "(", "_callable", "is", "not", "None", ...
assertraises that is usable as context manager or in a with statement exceptions that dont match the given exception type fall through:: .
train
false
1,818
def renew_hook(config, domains, lineage_path): if config.renew_hook: if (not config.dry_run): os.environ['RENEWED_DOMAINS'] = ' '.join(domains) os.environ['RENEWED_LINEAGE'] = lineage_path logger.info('Running renew-hook command: %s', config.renew_hook) _run_hook(config.renew_hook) else: logger.warning('Dry run: skipping renewal hook command: %s', config.renew_hook)
[ "def", "renew_hook", "(", "config", ",", "domains", ",", "lineage_path", ")", ":", "if", "config", ".", "renew_hook", ":", "if", "(", "not", "config", ".", "dry_run", ")", ":", "os", ".", "environ", "[", "'RENEWED_DOMAINS'", "]", "=", "' '", ".", "join...
run post-renewal hook if defined .
train
false
1,820
def example_data(infile='functional.nii'): filepath = os.path.abspath(__file__) basedir = os.path.dirname(filepath) outfile = os.path.join(basedir, 'data', infile) if (not os.path.exists(outfile)): raise IOError(('%s empty data file does NOT exist' % outfile)) return outfile
[ "def", "example_data", "(", "infile", "=", "'functional.nii'", ")", ":", "filepath", "=", "os", ".", "path", ".", "abspath", "(", "__file__", ")", "basedir", "=", "os", ".", "path", ".", "dirname", "(", "filepath", ")", "outfile", "=", "os", ".", "path...
returns path to empty example data files for doc tests it will raise an exception if filename is not in the directory .
train
false
1,821
@gen.engine def ListClientLogUsers(logs_store, callback): user_dir_re = re.compile('^([0-9]+)/$') (subdirs, _) = (yield gen.Task(store_utils.ListFilesAndDirs, logs_store, '/')) filtered = [] for s in subdirs: res = user_dir_re.match(s) if (res is not None): filtered.append(res.group(1)) callback(sorted(filtered))
[ "@", "gen", ".", "engine", "def", "ListClientLogUsers", "(", "logs_store", ",", "callback", ")", ":", "user_dir_re", "=", "re", ".", "compile", "(", "'^([0-9]+)/$'", ")", "(", "subdirs", ",", "_", ")", "=", "(", "yield", "gen", ".", "Task", "(", "store...
return the list of all users with data in the client log repository .
train
false
1,822
def durable_hash(item, ephemeral_paths): durable_item = deepcopy(item) for path in ephemeral_paths: try: dpath.util.delete(durable_item, path, separator='$') except PathNotFound: pass return hash_config(durable_item)
[ "def", "durable_hash", "(", "item", ",", "ephemeral_paths", ")", ":", "durable_item", "=", "deepcopy", "(", "item", ")", "for", "path", "in", "ephemeral_paths", ":", "try", ":", "dpath", ".", "util", ".", "delete", "(", "durable_item", ",", "path", ",", ...
remove all ephemeral paths from the item and return the hash of the new structure .
train
false
1,823
def registerDateHandler(func): _date_handlers.insert(0, func)
[ "def", "registerDateHandler", "(", "func", ")", ":", "_date_handlers", ".", "insert", "(", "0", ",", "func", ")" ]
register a date handler function .
train
false
1,824
@bdd.when(bdd.parsers.re('I press the keys? "(?P<keys>[^"]*)"')) def press_keys(quteproc, keys): quteproc.press_keys(keys)
[ "@", "bdd", ".", "when", "(", "bdd", ".", "parsers", ".", "re", "(", "'I press the keys? \"(?P<keys>[^\"]*)\"'", ")", ")", "def", "press_keys", "(", "quteproc", ",", "keys", ")", ":", "quteproc", ".", "press_keys", "(", "keys", ")" ]
send the given fake keys to qutebrowser .
train
false
1,827
def _process_tools_arg(plot, tools): tool_objs = [] tool_map = {} temp_tool_str = '' repeated_tools = [] if isinstance(tools, (list, tuple)): for tool in tools: if isinstance(tool, Tool): tool_objs.append(tool) elif isinstance(tool, string_types): temp_tool_str += (tool + ',') else: raise ValueError('tool should be a string or an instance of Tool class') tools = temp_tool_str for tool in re.split('\\s*,\\s*', tools.strip()): if (tool == ''): continue tool_obj = _tool_from_string(tool) tool_objs.append(tool_obj) tool_map[tool] = tool_obj for (typename, group) in itertools.groupby(sorted([tool.__class__.__name__ for tool in tool_objs])): if (len(list(group)) > 1): repeated_tools.append(typename) if repeated_tools: warnings.warn(('%s are being repeated' % ','.join(repeated_tools))) return (tool_objs, tool_map)
[ "def", "_process_tools_arg", "(", "plot", ",", "tools", ")", ":", "tool_objs", "=", "[", "]", "tool_map", "=", "{", "}", "temp_tool_str", "=", "''", "repeated_tools", "=", "[", "]", "if", "isinstance", "(", "tools", ",", "(", "list", ",", "tuple", ")",...
adds tools to the plot object args: plot : instance of a plot object tools : list of tool types or string listing the tool names .
train
true
1,829
@ioflo.base.deeding.deedify('SaltRaetEventReturnFork', ioinits={'opts': '.salt.opts', 'proc_mgr': '.salt.usr.proc_mgr'}) def event_return_fork(self): self.proc_mgr.value.add_process(salt.utils.event.EventReturn, args=(self.opts.value,))
[ "@", "ioflo", ".", "base", ".", "deeding", ".", "deedify", "(", "'SaltRaetEventReturnFork'", ",", "ioinits", "=", "{", "'opts'", ":", "'.salt.opts'", ",", "'proc_mgr'", ":", "'.salt.usr.proc_mgr'", "}", ")", "def", "event_return_fork", "(", "self", ")", ":", ...
add a reactor object to the process manager .
train
false
1,830
def matchSetStrengthNoNoise(match_set, target_set, noise): sum = 0.0 for t in target_set: sum += max((matchStrengthNoNoise(m, t, noise) for m in match_set)) return ((sum / len(target_set)),)
[ "def", "matchSetStrengthNoNoise", "(", "match_set", ",", "target_set", ",", "noise", ")", ":", "sum", "=", "0.0", "for", "t", "in", "target_set", ":", "sum", "+=", "max", "(", "(", "matchStrengthNoNoise", "(", "m", ",", "t", ",", "noise", ")", "for", "...
compute the match strength of a set of strings on the target set of strings .
train
false
1,831
@queued_task @app.task @transaction() def unsubscribe_mailchimp_async(list_name, user_id, username=None, send_goodbye=True): unsubscribe_mailchimp(list_name=list_name, user_id=user_id, username=username, send_goodbye=send_goodbye)
[ "@", "queued_task", "@", "app", ".", "task", "@", "transaction", "(", ")", "def", "unsubscribe_mailchimp_async", "(", "list_name", ",", "user_id", ",", "username", "=", "None", ",", "send_goodbye", "=", "True", ")", ":", "unsubscribe_mailchimp", "(", "list_nam...
same args as unsubscribe_mailchimp .
train
false
1,832
def get_breadcrumbs(url, request=None): from rest_framework.reverse import preserve_builtin_query_params from rest_framework.views import APIView def breadcrumbs_recursive(url, breadcrumbs_list, prefix, seen): u'\n Add tuples of (name, url) to the breadcrumbs list,\n progressively chomping off parts of the url.\n ' try: (view, unused_args, unused_kwargs) = resolve(url) except Exception: pass else: cls = getattr(view, u'cls', None) if ((cls is not None) and issubclass(cls, APIView)): if ((not seen) or (seen[(-1)] != view)): c = cls() c.suffix = getattr(view, u'suffix', None) name = c.get_view_name() insert_url = preserve_builtin_query_params((prefix + url), request) breadcrumbs_list.insert(0, (name, insert_url)) seen.append(view) if (url == u''): return breadcrumbs_list elif url.endswith(u'/'): url = url.rstrip(u'/') return breadcrumbs_recursive(url, breadcrumbs_list, prefix, seen) url = url[:(url.rfind(u'/') + 1)] return breadcrumbs_recursive(url, breadcrumbs_list, prefix, seen) prefix = get_script_prefix().rstrip(u'/') url = url[len(prefix):] return breadcrumbs_recursive(url, [], prefix, [])
[ "def", "get_breadcrumbs", "(", "url", ",", "request", "=", "None", ")", ":", "from", "rest_framework", ".", "reverse", "import", "preserve_builtin_query_params", "from", "rest_framework", ".", "views", "import", "APIView", "def", "breadcrumbs_recursive", "(", "url",...
get breadcrumbs .
train
false
1,834
def _asfarray(x): if (hasattr(x, 'dtype') and (x.dtype.char in numpy.typecodes['AllFloat'])): if (x.dtype == numpy.half): return numpy.asarray(x, dtype=numpy.float32) return numpy.asarray(x, dtype=x.dtype) else: ret = numpy.asarray(x) if (ret.dtype == numpy.half): return numpy.asarray(ret, dtype=numpy.float32) elif (ret.dtype.char not in numpy.typecodes['AllFloat']): return numpy.asfarray(x) return ret
[ "def", "_asfarray", "(", "x", ")", ":", "if", "(", "hasattr", "(", "x", ",", "'dtype'", ")", "and", "(", "x", ".", "dtype", ".", "char", "in", "numpy", ".", "typecodes", "[", "'AllFloat'", "]", ")", ")", ":", "if", "(", "x", ".", "dtype", "==",...
like numpy asfarray .
train
false
1,835
def general_context(request): return {'game_name': GAME_NAME, 'game_slogan': GAME_SLOGAN, 'evennia_userapps': PLAYER_RELATED, 'evennia_entityapps': GAME_ENTITIES, 'evennia_setupapps': GAME_SETUP, 'evennia_connectapps': CONNECTIONS, 'evennia_websiteapps': WEBSITE, 'webclient_enabled': WEBCLIENT_ENABLED, 'websocket_enabled': WEBSOCKET_CLIENT_ENABLED, 'websocket_port': WEBSOCKET_PORT, 'websocket_url': WEBSOCKET_URL}
[ "def", "general_context", "(", "request", ")", ":", "return", "{", "'game_name'", ":", "GAME_NAME", ",", "'game_slogan'", ":", "GAME_SLOGAN", ",", "'evennia_userapps'", ":", "PLAYER_RELATED", ",", "'evennia_entityapps'", ":", "GAME_ENTITIES", ",", "'evennia_setupapps'...
returns common evennia-related context stuff .
train
false
1,836
def lowercase_value(value): if isinstance(value, six.string_types): result = value.lower() elif isinstance(value, (list, tuple)): result = [str(item).lower() for item in value] elif isinstance(value, dict): result = {} for (key, value) in six.iteritems(value): result[key.lower()] = str(value).lower() else: result = value return result
[ "def", "lowercase_value", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "six", ".", "string_types", ")", ":", "result", "=", "value", ".", "lower", "(", ")", "elif", "isinstance", "(", "value", ",", "(", "list", ",", "tuple", ")", ")...
lowercase the provided value .
train
false
1,838
def connect_awslambda(aws_access_key_id=None, aws_secret_access_key=None, **kwargs): from boto.awslambda.layer1 import AWSLambdaConnection return AWSLambdaConnection(aws_access_key_id=aws_access_key_id, aws_secret_access_key=aws_secret_access_key, **kwargs)
[ "def", "connect_awslambda", "(", "aws_access_key_id", "=", "None", ",", "aws_secret_access_key", "=", "None", ",", "**", "kwargs", ")", ":", "from", "boto", ".", "awslambda", ".", "layer1", "import", "AWSLambdaConnection", "return", "AWSLambdaConnection", "(", "aw...
connect to aws lambda :type aws_access_key_id: string .
train
false
1,841
def update_activity_first_published_msec(activity_type, activity_id, first_published_msec): activity_rights = _get_activity_rights(activity_type, activity_id) commit_cmds = [{'cmd': CMD_UPDATE_FIRST_PUBLISHED_MSEC, 'old_first_published_msec': activity_rights.first_published_msec, 'new_first_published_msec': first_published_msec}] activity_rights.first_published_msec = first_published_msec _save_activity_rights(feconf.SYSTEM_COMMITTER_ID, activity_rights, activity_type, 'set first published time in msec', commit_cmds)
[ "def", "update_activity_first_published_msec", "(", "activity_type", ",", "activity_id", ",", "first_published_msec", ")", ":", "activity_rights", "=", "_get_activity_rights", "(", "activity_type", ",", "activity_id", ")", "commit_cmds", "=", "[", "{", "'cmd'", ":", "...
updates the first_published_msec field for an activity .
train
false
1,842
def identify_hasher(encoded): if (((len(encoded) == 32) and (u'$' not in encoded)) or ((len(encoded) == 37) and encoded.startswith(u'md5$$'))): algorithm = u'unsalted_md5' elif ((len(encoded) == 46) and encoded.startswith(u'sha1$$')): algorithm = u'unsalted_sha1' else: algorithm = encoded.split(u'$', 1)[0] return get_hasher(algorithm)
[ "def", "identify_hasher", "(", "encoded", ")", ":", "if", "(", "(", "(", "len", "(", "encoded", ")", "==", "32", ")", "and", "(", "u'$'", "not", "in", "encoded", ")", ")", "or", "(", "(", "len", "(", "encoded", ")", "==", "37", ")", "and", "enc...
returns an instance of a loaded password hasher .
train
false
1,843
def configure_codelab_igor_jenkins(): YamlBindings.update_yml_source('/opt/spinnaker/config/spinnaker-local.yml', {'jenkins': {'defaultMaster': {'name': 'CodelabJenkins', 'baseUrl': 'http://localhost:9090', 'username': 'admin', 'password': 'admin'}}, 'igor': {'enabled': 'true'}})
[ "def", "configure_codelab_igor_jenkins", "(", ")", ":", "YamlBindings", ".", "update_yml_source", "(", "'/opt/spinnaker/config/spinnaker-local.yml'", ",", "{", "'jenkins'", ":", "{", "'defaultMaster'", ":", "{", "'name'", ":", "'CodelabJenkins'", ",", "'baseUrl'", ":", ...
configures igor to be enabled and to point to the codelab jenkins instance .
train
false
1,844
def simulate_user_history(session, user): total_sessions = random.randrange(10) previous_user_session = None for _ in range(total_sessions): user_session = simulate_user_session(session, user, previous_user_session) previous_user_session = user_session
[ "def", "simulate_user_history", "(", "session", ",", "user", ")", ":", "total_sessions", "=", "random", ".", "randrange", "(", "10", ")", "previous_user_session", "=", "None", "for", "_", "in", "range", "(", "total_sessions", ")", ":", "user_session", "=", "...
simulates the entire history of activity for a single user .
train
false
1,845
def conv_output_length(input_length, filter_size, border_mode, stride, dilation=1): if (input_length is None): return None assert (border_mode in {'same', 'valid', 'full'}) dilated_filter_size = (filter_size + ((filter_size - 1) * (dilation - 1))) if (border_mode == 'same'): output_length = input_length elif (border_mode == 'valid'): output_length = ((input_length - dilated_filter_size) + 1) elif (border_mode == 'full'): output_length = ((input_length + dilated_filter_size) - 1) return (((output_length + stride) - 1) // stride)
[ "def", "conv_output_length", "(", "input_length", ",", "filter_size", ",", "border_mode", ",", "stride", ",", "dilation", "=", "1", ")", ":", "if", "(", "input_length", "is", "None", ")", ":", "return", "None", "assert", "(", "border_mode", "in", "{", "'sa...
helper function to compute the output size of a convolution operation this function computes the length along a single axis .
train
true
1,846
def CreateCGate(name, latexname=None): if (not latexname): latexname = name onequbitgate = CreateOneQubitGate(name, latexname) def ControlledGate(ctrls, target): return CGate(tuple(ctrls), onequbitgate(target)) return ControlledGate
[ "def", "CreateCGate", "(", "name", ",", "latexname", "=", "None", ")", ":", "if", "(", "not", "latexname", ")", ":", "latexname", "=", "name", "onequbitgate", "=", "CreateOneQubitGate", "(", "name", ",", "latexname", ")", "def", "ControlledGate", "(", "ctr...
use a lexical closure to make a controlled gate .
train
false
1,847
def get_materialized_checkpoints(checkpoints, colnames, lower_dt, odo_kwargs): if (checkpoints is not None): ts = checkpoints[TS_FIELD_NAME] checkpoints_ts = odo(ts[(ts <= lower_dt)].max(), pd.Timestamp, **odo_kwargs) if pd.isnull(checkpoints_ts): materialized_checkpoints = pd.DataFrame(columns=colnames) lower = None else: materialized_checkpoints = odo(checkpoints[(ts == checkpoints_ts)][colnames], pd.DataFrame, **odo_kwargs) lower = checkpoints_ts else: materialized_checkpoints = pd.DataFrame(columns=colnames) lower = None return (lower, materialized_checkpoints)
[ "def", "get_materialized_checkpoints", "(", "checkpoints", ",", "colnames", ",", "lower_dt", ",", "odo_kwargs", ")", ":", "if", "(", "checkpoints", "is", "not", "None", ")", ":", "ts", "=", "checkpoints", "[", "TS_FIELD_NAME", "]", "checkpoints_ts", "=", "odo"...
computes a lower bound and a dataframe checkpoints .
train
true
1,848
def convert(t, convstr=True): for (i, op) in enumerate(t.op): context.clear_status() if (op in RoundModes): t.cop.append(op) t.pop.append(op) elif (((not t.contextfunc) and (i == 0)) or (convstr and isinstance(op, str))): try: c = C.Decimal(op) cex = None except (TypeError, ValueError, OverflowError) as e: c = None cex = e.__class__ try: p = RestrictedDecimal(op) pex = None except (TypeError, ValueError, OverflowError) as e: p = None pex = e.__class__ t.cop.append(c) t.cex.append(cex) t.pop.append(p) t.pex.append(pex) if (cex is pex): if ((str(c) != str(p)) or (not context.assert_eq_status())): raise_error(t) if (cex and pex): return 0 else: raise_error(t) elif isinstance(op, Context): t.context = op t.cop.append(op.c) t.pop.append(op.p) else: t.cop.append(op) t.pop.append(op) return 1
[ "def", "convert", "(", "t", ",", "convstr", "=", "True", ")", ":", "for", "(", "i", ",", "op", ")", "in", "enumerate", "(", "t", ".", "op", ")", ":", "context", ".", "clear_status", "(", ")", "if", "(", "op", "in", "RoundModes", ")", ":", "t", ...
convert the input to a properly typed numpy value according to the current casting policy .
train
false
1,850
def skipWithClientIf(condition, reason): def client_pass(*args, **kwargs): pass def skipdec(obj): retval = unittest.skip(reason)(obj) if (not isinstance(obj, type)): retval.client_skip = (lambda f: client_pass) return retval def noskipdec(obj): if (not (isinstance(obj, type) or hasattr(obj, 'client_skip'))): obj.client_skip = (lambda f: f) return obj return (skipdec if condition else noskipdec)
[ "def", "skipWithClientIf", "(", "condition", ",", "reason", ")", ":", "def", "client_pass", "(", "*", "args", ",", "**", "kwargs", ")", ":", "pass", "def", "skipdec", "(", "obj", ")", ":", "retval", "=", "unittest", ".", "skip", "(", "reason", ")", "...
skip decorated test if condition is true .
train
false
1,852
def remove_series(name, forget=False): downloaded_releases = [] with Session() as session: series = session.query(Series).filter((Series.name == name)).all() if series: for s in series: if forget: for episode in s.episodes: for release in episode.downloaded_releases: downloaded_releases.append(release.title) session.delete(s) session.commit() log.debug(u'Removed series %s from database.', name) else: raise ValueError((u'Unknown series %s' % name)) for downloaded_release in downloaded_releases: fire_event(u'forget', downloaded_release)
[ "def", "remove_series", "(", "name", ",", "forget", "=", "False", ")", ":", "downloaded_releases", "=", "[", "]", "with", "Session", "(", ")", "as", "session", ":", "series", "=", "session", ".", "query", "(", "Series", ")", ".", "filter", "(", "(", ...
remove a whole series name from database .
train
false
1,854
def ipaddress_from_string(ip_address_string): parts = ip_address_string.rsplit('%', 1) ip_address_string = parts[0] return IPAddress(ip_address_string)
[ "def", "ipaddress_from_string", "(", "ip_address_string", ")", ":", "parts", "=", "ip_address_string", ".", "rsplit", "(", "'%'", ",", "1", ")", "ip_address_string", "=", "parts", "[", "0", "]", "return", "IPAddress", "(", "ip_address_string", ")" ]
parse an ipv4 or ipv6 address string and return an ipaddress instance .
train
false
1,855
def deploy_member_filters(status=False): T = current.T widgets = [S3TextFilter(['person_id$first_name', 'person_id$middle_name', 'person_id$last_name'], label=T('Name')), S3OptionsFilter('organisation_id', filter=True, hidden=True), S3OptionsFilter('credential.job_title_id', label=T('Sector'), hidden=True)] settings = current.deployment_settings if settings.get_hrm_teams(): widgets.append(S3OptionsFilter('group_membership.group_id', label=T('Teams'), hidden=True)) if settings.get_org_regions(): if settings.get_org_regions_hierarchical(): widgets.insert(1, S3HierarchyFilter('organisation_id$region_id', lookup='org_region', hidden=True, none=T('No Region'))) else: widgets.insert(1, S3OptionsFilter('organisation_id$region_id', widget='multiselect', filter=True)) if status: widgets.insert(1, S3OptionsFilter('application.active', cols=2, default=True, label=T('Status'), options={'True': T('active'), 'False': T('inactive')})) return widgets
[ "def", "deploy_member_filters", "(", "status", "=", "False", ")", ":", "T", "=", "current", ".", "T", "widgets", "=", "[", "S3TextFilter", "(", "[", "'person_id$first_name'", ",", "'person_id$middle_name'", ",", "'person_id$last_name'", "]", ",", "label", "=", ...
filter widgets for members .
train
false
1,856
@treeio_login_required def widget_index(request, response_format='html'): tickets = Object.filter_by_request(request, Ticket.objects.filter(status__hidden=False)) context = _get_default_context(request) context.update({'tickets': tickets}) return render_to_response('services/widgets/index', context, context_instance=RequestContext(request), response_format=response_format)
[ "@", "treeio_login_required", "def", "widget_index", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "tickets", "=", "Object", ".", "filter_by_request", "(", "request", ",", "Ticket", ".", "objects", ".", "filter", "(", "status__hidden", "=", ...
all active tickets .
train
false
1,858
def _search_coding_line(txt): coding_pattern = u'coding[:=]\\s*([-\\w.]+)' pat_coding = re.search(coding_pattern, txt) if (pat_coding and (pat_coding.groups()[0] != u'None')): return pat_coding.groups()[0] return None
[ "def", "_search_coding_line", "(", "txt", ")", ":", "coding_pattern", "=", "u'coding[:=]\\\\s*([-\\\\w.]+)'", "pat_coding", "=", "re", ".", "search", "(", "coding_pattern", ",", "txt", ")", "if", "(", "pat_coding", "and", "(", "pat_coding", ".", "groups", "(", ...
search a pattern like this: # -*- coding: utf-8 -*- .
train
false
1,860
def req_factory_factory(url='', user=None, post=False, data=None, **kwargs): req = RequestFactory() if post: req = req.post(url, (data or {})) else: req = req.get(url, (data or {})) if user: req.user = UserProfile.objects.get(id=user.id) req.groups = user.groups.all() else: req.user = AnonymousUser() req.check_ownership = partial(check_ownership, req) req.REGION = kwargs.pop('region', mkt.regions.REGIONS_CHOICES[0][1]) req.API_VERSION = 2 for key in kwargs: setattr(req, key, kwargs[key]) return req
[ "def", "req_factory_factory", "(", "url", "=", "''", ",", "user", "=", "None", ",", "post", "=", "False", ",", "data", "=", "None", ",", "**", "kwargs", ")", ":", "req", "=", "RequestFactory", "(", ")", "if", "post", ":", "req", "=", "req", ".", ...
creates a request factory .
train
false
1,862
def add_problem_to_course(course, problem_type, extra_meta=None): assert (problem_type in PROBLEM_DICT) factory_dict = PROBLEM_DICT[problem_type] problem_xml = factory_dict['factory'].build_xml(**factory_dict['kwargs']) metadata = ({'rerandomize': 'always'} if ('metadata' not in factory_dict) else factory_dict['metadata']) if extra_meta: metadata = dict(metadata, **extra_meta) category_name = 'problem' return world.ItemFactory.create(parent_location=section_location(course), category=category_name, display_name=str(problem_type), data=problem_xml, metadata=metadata)
[ "def", "add_problem_to_course", "(", "course", ",", "problem_type", ",", "extra_meta", "=", "None", ")", ":", "assert", "(", "problem_type", "in", "PROBLEM_DICT", ")", "factory_dict", "=", "PROBLEM_DICT", "[", "problem_type", "]", "problem_xml", "=", "factory_dict...
add a problem to the course we have created using factories .
train
false
1,865
def slstm(c_prev1, c_prev2, x1, x2): return SLSTM()(c_prev1, c_prev2, x1, x2)
[ "def", "slstm", "(", "c_prev1", ",", "c_prev2", ",", "x1", ",", "x2", ")", ":", "return", "SLSTM", "(", ")", "(", "c_prev1", ",", "c_prev2", ",", "x1", ",", "x2", ")" ]
s-lstm units as an activation function .
train
false
1,866
def reduceby(key, binop, seq, init=no_default): is_no_default = (init == no_default) if ((not is_no_default) and (not callable(init))): _init = init init = (lambda : _init) if (not callable(key)): key = getter(key) d = {} for item in seq: k = key(item) if (k not in d): if is_no_default: d[k] = item continue else: d[k] = init() d[k] = binop(d[k], item) return d
[ "def", "reduceby", "(", "key", ",", "binop", ",", "seq", ",", "init", "=", "no_default", ")", ":", "is_no_default", "=", "(", "init", "==", "no_default", ")", "if", "(", "(", "not", "is_no_default", ")", "and", "(", "not", "callable", "(", "init", ")...
perform a simultaneous groupby and reduction the computation: .
train
false
1,867
def getRightStripMinusSplit(lineString): oldLineStringLength = (-1) while (oldLineStringLength < len(lineString)): oldLineStringLength = len(lineString) lineString = lineString.replace('- ', '-') return lineString.split()
[ "def", "getRightStripMinusSplit", "(", "lineString", ")", ":", "oldLineStringLength", "=", "(", "-", "1", ")", "while", "(", "oldLineStringLength", "<", "len", "(", "lineString", ")", ")", ":", "oldLineStringLength", "=", "len", "(", "lineString", ")", "lineSt...
get string with spaces after the minus sign stripped .
train
false
1,868
def _gl_score(estimators, scoring, X, y): n_iter = X.shape[(-1)] n_est = len(estimators) for (ii, est) in enumerate(estimators): for jj in range(X.shape[(-1)]): if (scoring is not None): _score = scoring(est, X[..., jj], y) else: _score = est.score(X[..., jj], y) if ((ii == 0) & (jj == 0)): if isinstance(_score, np.ndarray): dtype = _score.dtype shape = np.r_[(n_est, n_iter, _score.shape)] else: dtype = type(_score) shape = [n_est, n_iter] score = np.zeros(shape, dtype) score[ii, jj, ...] = _score return score
[ "def", "_gl_score", "(", "estimators", ",", "scoring", ",", "X", ",", "y", ")", ":", "n_iter", "=", "X", ".", "shape", "[", "(", "-", "1", ")", "]", "n_est", "=", "len", "(", "estimators", ")", "for", "(", "ii", ",", "est", ")", "in", "enumerat...
score _generalizationlight in parallel .
train
false
1,872
def tag_as_test_install(flocker_version, distribution, package_name): repository_url = get_repository_url(distribution=distribution, flocker_version=flocker_version) repository_host = urlparse(repository_url).hostname tag_url = bytes('https://{host}/clusterhq-internal-acceptance-test/{distribution}/{package}/{version}'.format(host=repository_host, distribution=distribution, package=package_name, version=flocker_version)) return http_get(tag_url)
[ "def", "tag_as_test_install", "(", "flocker_version", ",", "distribution", ",", "package_name", ")", ":", "repository_url", "=", "get_repository_url", "(", "distribution", "=", "distribution", ",", "flocker_version", "=", "flocker_version", ")", "repository_host", "=", ...
creates an effect of making an http get to a specific url in an s3 bucket that has logging enabled .
train
false
1,875
def scalar_symbols(exprs): new_names = (('_%d' % i) for i in itertools.count(1)) scalars = [] names = set() for expr in exprs: if (expr._name and (expr._name not in names)): name = expr._name names.add(name) else: name = next(new_names) s = symbol(name, expr.schema) scalars.append(s) return scalars
[ "def", "scalar_symbols", "(", "exprs", ")", ":", "new_names", "=", "(", "(", "'_%d'", "%", "i", ")", "for", "i", "in", "itertools", ".", "count", "(", "1", ")", ")", "scalars", "=", "[", "]", "names", "=", "set", "(", ")", "for", "expr", "in", ...
gives a sequence of scalar symbols to mirror these expressions examples .
train
false
1,876
def connect_to_region(region_name, **kw_params): for region in regions(): if (region.name == region_name): return region.connect(**kw_params) return None
[ "def", "connect_to_region", "(", "region_name", ",", "**", "kw_params", ")", ":", "for", "region", "in", "regions", "(", ")", ":", "if", "(", "region", ".", "name", "==", "region_name", ")", ":", "return", "region", ".", "connect", "(", "**", "kw_params"...
given a valid region name .
train
true
1,877
def github(registry, xml_parent, data): ghtrig = XML.SubElement(xml_parent, 'com.cloudbees.jenkins.GitHubPushTrigger') XML.SubElement(ghtrig, 'spec').text = ''
[ "def", "github", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "ghtrig", "=", "XML", ".", "SubElement", "(", "xml_parent", ",", "'com.cloudbees.jenkins.GitHubPushTrigger'", ")", "XML", ".", "SubElement", "(", "ghtrig", ",", "'spec'", ")", ".", "...
yaml: github sets the github url for the project .
train
false
1,878
def _matches(event, token_values): if (event.user_id is not None): for attribute_name in ['user_id', 'trustor_id', 'trustee_id']: if (event.user_id == token_values[attribute_name]): break else: return False if (event.domain_id is not None): for attribute_name in ['identity_domain_id', 'assignment_domain_id']: if (event.domain_id == token_values[attribute_name]): break else: return False if (event.domain_scope_id is not None): if (event.domain_scope_id != token_values['assignment_domain_id']): return False attribute_names = ['project_id', 'expires_at', 'trust_id', 'consumer_id', 'access_token_id', 'audit_id', 'audit_chain_id'] for attribute_name in attribute_names: if (getattr(event, attribute_name) is not None): if (getattr(event, attribute_name) != token_values[attribute_name]): return False if (event.role_id is not None): roles = token_values['roles'] for role in roles: if (event.role_id == role): break else: return False if (token_values['issued_at'] > event.issued_before): return False return True
[ "def", "_matches", "(", "event", ",", "token_values", ")", ":", "if", "(", "event", ".", "user_id", "is", "not", "None", ")", ":", "for", "attribute_name", "in", "[", "'user_id'", ",", "'trustor_id'", ",", "'trustee_id'", "]", ":", "if", "(", "event", ...
see if the token matches the revocation event .
train
false
1,879
def sanitize_address(addr, encoding): if (not isinstance(addr, tuple)): addr = parseaddr(force_text(addr)) (nm, addr) = addr (localpart, domain) = (None, None) nm = Header(nm, encoding).encode() try: addr.encode('ascii') except UnicodeEncodeError: (localpart, domain) = split_addr(addr, encoding) if (localpart and domain): address = Address(nm, username=localpart, domain=domain) return str(address) try: address = Address(nm, addr_spec=addr) except (InvalidHeaderDefect, NonASCIILocalPartDefect): (localpart, domain) = split_addr(addr, encoding) address = Address(nm, username=localpart, domain=domain) return str(address)
[ "def", "sanitize_address", "(", "addr", ",", "encoding", ")", ":", "if", "(", "not", "isinstance", "(", "addr", ",", "tuple", ")", ")", ":", "addr", "=", "parseaddr", "(", "force_text", "(", "addr", ")", ")", "(", "nm", ",", "addr", ")", "=", "addr...
format a pair of or an email address string .
train
false
1,880
def is_valid_closer(string): openers = ['[', '(', '{'] closers = [']', ')', '}'] stack = [] for ch in string: if ((not stack) and (ch in closers)): return False elif (ch in openers): stack.append(ch) elif (ch == ']'): if (stack.pop() != '['): return False elif (ch == ')'): if (stack.pop() != '('): return False elif (ch == '}'): if (stack.pop() != '{'): return False return (not stack)
[ "def", "is_valid_closer", "(", "string", ")", ":", "openers", "=", "[", "'['", ",", "'('", ",", "'{'", "]", "closers", "=", "[", "']'", ",", "')'", ",", "'}'", "]", "stack", "=", "[", "]", "for", "ch", "in", "string", ":", "if", "(", "(", "not"...
returns true if every opener has a valid closer .
train
false