id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
6,217
def _get_capa_types(): capa_types = {tag: _get_human_name(registry.get_class_for_tag(tag)) for tag in registry.registered_tags()} return ([{'value': ANY_CAPA_TYPE_VALUE, 'display_name': _('Any Type')}] + sorted([{'value': capa_type, 'display_name': caption} for (capa_type, caption) in capa_types.items()], key=(lambda item: item.get('display_name'))))
[ "def", "_get_capa_types", "(", ")", ":", "capa_types", "=", "{", "tag", ":", "_get_human_name", "(", "registry", ".", "get_class_for_tag", "(", "tag", ")", ")", "for", "tag", "in", "registry", ".", "registered_tags", "(", ")", "}", "return", "(", "[", "{...
gets capa types tags and labels .
train
false
6,218
def _best_mime(): supported = [] renders = {} for mime in _MIME_TYPES: if app.config.get(mime['tag'], True): for mime_type in mime['mime']: supported.append(mime_type) renders[mime_type] = mime['renderer'] if (len(supported) == 0): abort(500, description=debug_error_message('Configuration error: no supported mime types')) best_match = (request.accept_mimetypes.best_match(supported) or supported[0]) return (best_match, renders[best_match])
[ "def", "_best_mime", "(", ")", ":", "supported", "=", "[", "]", "renders", "=", "{", "}", "for", "mime", "in", "_MIME_TYPES", ":", "if", "app", ".", "config", ".", "get", "(", "mime", "[", "'tag'", "]", ",", "True", ")", ":", "for", "mime_type", ...
returns the best match between the requested mime type and the ones supported by eve .
train
false
6,219
def test_refresh_with_callbacks(refresher): callbacks = [Mock()] pgexecute_class = Mock() pgexecute = Mock() special = Mock() with patch('pgcli.completion_refresher.PGExecute', pgexecute_class): refresher.refreshers = {} refresher.refresh(pgexecute, special, callbacks) time.sleep(1) assert (callbacks[0].call_count == 1)
[ "def", "test_refresh_with_callbacks", "(", "refresher", ")", ":", "callbacks", "=", "[", "Mock", "(", ")", "]", "pgexecute_class", "=", "Mock", "(", ")", "pgexecute", "=", "Mock", "(", ")", "special", "=", "Mock", "(", ")", "with", "patch", "(", "'pgcli....
callbacks must be called .
train
false
6,220
def make_naive(value, timezone): value = value.astimezone(timezone) if hasattr(timezone, 'normalize'): value = timezone.normalize(value) return value.replace(tzinfo=None)
[ "def", "make_naive", "(", "value", ",", "timezone", ")", ":", "value", "=", "value", ".", "astimezone", "(", "timezone", ")", "if", "hasattr", "(", "timezone", ",", "'normalize'", ")", ":", "value", "=", "timezone", ".", "normalize", "(", "value", ")", ...
makes an aware datetime .
train
true
6,221
@image_comparison(baseline_images=[u'test_eventplot_defaults'], extensions=[u'png'], remove_text=True) def test_eventplot_defaults(): np.random.seed(0) data1 = np.random.random([32, 20]).tolist() data2 = np.random.random([6, 20]).tolist() data = (data1 + data2) fig = plt.figure() axobj = fig.add_subplot(111) colls = axobj.eventplot(data)
[ "@", "image_comparison", "(", "baseline_images", "=", "[", "u'test_eventplot_defaults'", "]", ",", "extensions", "=", "[", "u'png'", "]", ",", "remove_text", "=", "True", ")", "def", "test_eventplot_defaults", "(", ")", ":", "np", ".", "random", ".", "seed", ...
test that eventplot produces the correct output given the default params .
train
false
6,224
def format_group(group, show_url=True): out = '\x02{}\x02'.format(group['name']) if group['description']: out += ': "{}"'.format(formatting.truncate(group['description'])) out += ' - Owned by \x02{}\x02.'.format(group['creator']['username']) if show_url: out += ' - {}'.format(web.try_shorten(group['permalink_url'])) return out
[ "def", "format_group", "(", "group", ",", "show_url", "=", "True", ")", ":", "out", "=", "'\\x02{}\\x02'", ".", "format", "(", "group", "[", "'name'", "]", ")", "if", "group", "[", "'description'", "]", ":", "out", "+=", "': \"{}\"'", ".", "format", "(...
takes a soundcloud group and returns a formatting string .
train
false
6,225
def salt_spm(): import salt.cli.spm spm = salt.cli.spm.SPM() spm.run()
[ "def", "salt_spm", "(", ")", ":", "import", "salt", ".", "cli", ".", "spm", "spm", "=", "salt", ".", "cli", ".", "spm", ".", "SPM", "(", ")", "spm", ".", "run", "(", ")" ]
the main function for spm .
train
true
6,226
def _paramsFileHead(): str = (getCopyrightHead() + '\n\n## This file defines parameters for a prediction experiment.\n\n###############################################################################\n# IMPORTANT!!!\n# This params file is dynamically generated by the RunExperimentPermutations\n# script. Any changes made manually will be over-written the next time\n# RunExperimentPermutations is run!!!\n###############################################################################\n\n\nfrom nupic.frameworks.opf.expdescriptionhelpers import importBaseDescription\n\n# the sub-experiment configuration\nconfig ={\n') return str
[ "def", "_paramsFileHead", "(", ")", ":", "str", "=", "(", "getCopyrightHead", "(", ")", "+", "'\\n\\n## This file defines parameters for a prediction experiment.\\n\\n###############################################################################\\n# IMPORTANT!...
this is the first portion of every sub-experiment params file we generate .
train
false
6,227
def get_profit_data(year, quarter): if (ct._check_input(year, quarter) is True): ct._write_head() data = _get_profit_data(year, quarter, 1, pd.DataFrame()) if (data is not None): data['code'] = data['code'].map((lambda x: str(x).zfill(6))) return data
[ "def", "get_profit_data", "(", "year", ",", "quarter", ")", ":", "if", "(", "ct", ".", "_check_input", "(", "year", ",", "quarter", ")", "is", "True", ")", ":", "ct", ".", "_write_head", "(", ")", "data", "=", "_get_profit_data", "(", "year", ",", "q...
parameters year:int 年度 e .
train
false
6,228
def minor_version(best=False): return _distro.minor_version(best)
[ "def", "minor_version", "(", "best", "=", "False", ")", ":", "return", "_distro", ".", "minor_version", "(", "best", ")" ]
return the minor version of the current linux distribution .
train
false
6,230
def getChanges(request, options=None): payload = json.loads(request.args['payload'][0]) repo_url = ('%s%s' % (payload['canon_url'], payload['repository']['absolute_url'])) project = request.args.get('project', [''])[0] changes = [] for commit in payload['commits']: changes.append({'author': commit['raw_author'], 'files': [f['file'] for f in commit['files']], 'comments': commit['message'], 'revision': commit['raw_node'], 'when_timestamp': dateparse(commit['utctimestamp']), 'branch': commit['branch'], 'revlink': ('%scommits/%s' % (repo_url, commit['raw_node'])), 'repository': repo_url, 'project': project}) log.msg(('New revision: %s' % (commit['node'],))) log.msg(('Received %s changes from bitbucket' % (len(changes),))) return (changes, payload['repository']['scm'])
[ "def", "getChanges", "(", "request", ",", "options", "=", "None", ")", ":", "payload", "=", "json", ".", "loads", "(", "request", ".", "args", "[", "'payload'", "]", "[", "0", "]", ")", "repo_url", "=", "(", "'%s%s'", "%", "(", "payload", "[", "'ca...
reponds only to post events and starts the build process :arguments: request the http request object .
train
false
6,234
def package_absent(m, name): retvals = {'rc': 0, 'stdout': '', 'stderr': ''} (name_install, name_remove, urls) = get_want_state(m, name, remove=True) if name_install: m.fail_json(msg="Can not combine '+' prefix with state=remove/absent.") if urls: m.fail_json(msg='Can not remove via URL.') if (m.params['type'] == 'patch'): m.fail_json(msg='Can not remove patches.') prerun_state = get_installed_state(m, name_remove) remove_version = [(p + name_remove[p]) for p in name_remove if name_remove[p]] name_remove = [p for p in name_remove if (p in prerun_state)] if ((not name_remove) and (not remove_version)): return (None, retvals) cmd = get_cmd(m, 'remove') cmd.extend(name_remove) cmd.extend(remove_version) retvals['cmd'] = cmd (result, retvals['rc'], retvals['stdout'], retvals['stderr']) = parse_zypper_xml(m, cmd) return (result, retvals)
[ "def", "package_absent", "(", "m", ",", "name", ")", ":", "retvals", "=", "{", "'rc'", ":", "0", ",", "'stdout'", ":", "''", ",", "'stderr'", ":", "''", "}", "(", "name_install", ",", "name_remove", ",", "urls", ")", "=", "get_want_state", "(", "m", ...
remove the packages in name .
train
false
6,235
def test_autocall_binops(): ip.magic('autocall 2') f = (lambda x: x) ip.user_ns['f'] = f try: nt.assert_equal(ip.prefilter('f 1'), 'f(1)') for t in ['f +1', 'f -1']: nt.assert_equal(ip.prefilter(t), t) pm = ip.prefilter_manager ac = AutocallChecker(shell=pm.shell, prefilter_manager=pm, config=pm.config) try: ac.priority = 1 ac.exclude_regexp = '^[,&^\\|\\*/]|^is |^not |^in |^and |^or ' pm.sort_checkers() nt.assert_equal(ip.prefilter('f -1'), 'f(-1)') nt.assert_equal(ip.prefilter('f +1'), 'f(+1)') finally: pm.unregister_checker(ac) finally: ip.magic('autocall 0') del ip.user_ns['f']
[ "def", "test_autocall_binops", "(", ")", ":", "ip", ".", "magic", "(", "'autocall 2'", ")", "f", "=", "(", "lambda", "x", ":", "x", ")", "ip", ".", "user_ns", "[", "'f'", "]", "=", "f", "try", ":", "nt", ".", "assert_equal", "(", "ip", ".", "pref...
see URL .
train
false
6,237
def set_guess(guess): global _GUESS _GUESS = guess
[ "def", "set_guess", "(", "guess", ")", ":", "global", "_GUESS", "_GUESS", "=", "guess" ]
set the default value of the guess parameter for read() parameters guess : bool new default guess value .
train
false
6,238
def resolve_stream_name(streams, stream_name): if ((stream_name in STREAM_SYNONYMS) and (stream_name in streams)): for (name, stream) in streams.items(): if ((stream is streams[stream_name]) and (name not in STREAM_SYNONYMS)): return name return stream_name
[ "def", "resolve_stream_name", "(", "streams", ",", "stream_name", ")", ":", "if", "(", "(", "stream_name", "in", "STREAM_SYNONYMS", ")", "and", "(", "stream_name", "in", "streams", ")", ")", ":", "for", "(", "name", ",", "stream", ")", "in", "streams", "...
returns the real stream name of a synonym .
train
true
6,239
def find_all_tests(suite): suites = [suite] while suites: s = suites.pop() try: suites.extend(s) except TypeError: (yield (s, ('%s.%s.%s' % (s.__class__.__module__, s.__class__.__name__, s._testMethodName))))
[ "def", "find_all_tests", "(", "suite", ")", ":", "suites", "=", "[", "suite", "]", "while", "suites", ":", "s", "=", "suites", ".", "pop", "(", ")", "try", ":", "suites", ".", "extend", "(", "s", ")", "except", "TypeError", ":", "(", "yield", "(", ...
yields all the tests and their names from a given suite .
train
false
6,240
def get_outbound_pdus(text, recipient): encoding = 'ucs2' try: encoded_text = text.encode('gsm') encoding = 'gsm' except: encoded_text = text csm_max = MSG_LIMITS[encoding][1] if (len(encoded_text) > (MAX_CSM_SEGMENTS * csm_max)): raise ValueError('Message text too long') if (len(encoded_text) <= MSG_LIMITS[encoding][0]): return [OutboundGsmPdu(text, recipient)] with __ref_lock: if (recipient not in __csm_refs): __csm_refs[recipient] = 0 csm_ref = (__csm_refs[recipient] % 256) __csm_refs[recipient] += 1 num = int(math.ceil((len(encoded_text) / float(MSG_LIMITS[encoding][0])))) pdus = [] for seq in range(num): i = (seq * csm_max) seg_txt = encoded_text[i:(i + csm_max)] if (encoding == 'gsm'): seg_txt = seg_txt.decode('gsm') pdus.append(OutboundGsmPdu(seg_txt, recipient, csm_ref=csm_ref, csm_seq=(seq + 1), csm_total=num)) return pdus
[ "def", "get_outbound_pdus", "(", "text", ",", "recipient", ")", ":", "encoding", "=", "'ucs2'", "try", ":", "encoded_text", "=", "text", ".", "encode", "(", "'gsm'", ")", "encoding", "=", "'gsm'", "except", ":", "encoded_text", "=", "text", "csm_max", "=",...
returns a list of pdus to send the provided text to the given recipient .
train
false
6,242
def __converting_factory(specimen_cls, original_factory): instrumented_cls = __canned_instrumentation[specimen_cls] def wrapper(): collection = original_factory() return instrumented_cls(collection) wrapper.__name__ = ('%sWrapper' % original_factory.__name__) wrapper.__doc__ = original_factory.__doc__ return wrapper
[ "def", "__converting_factory", "(", "specimen_cls", ",", "original_factory", ")", ":", "instrumented_cls", "=", "__canned_instrumentation", "[", "specimen_cls", "]", "def", "wrapper", "(", ")", ":", "collection", "=", "original_factory", "(", ")", "return", "instrum...
return a wrapper that converts a "canned" collection like set .
train
false
6,243
def run_path(path_name, init_globals=None, run_name=None): if (run_name is None): run_name = '<run_path>' importer = _get_importer(path_name) if isinstance(importer, imp.NullImporter): code = _get_code_from_file(path_name) return _run_module_code(code, init_globals, run_name, path_name) else: sys.path.insert(0, path_name) try: main_name = '__main__' saved_main = sys.modules[main_name] del sys.modules[main_name] try: (mod_name, loader, code, fname) = _get_main_module_details() finally: sys.modules[main_name] = saved_main pkg_name = '' with _TempModule(run_name) as temp_module: with _ModifiedArgv0(path_name): mod_globals = temp_module.module.__dict__ return _run_code(code, mod_globals, init_globals, run_name, fname, loader, pkg_name).copy() finally: try: sys.path.remove(path_name) except ValueError: pass
[ "def", "run_path", "(", "path_name", ",", "init_globals", "=", "None", ",", "run_name", "=", "None", ")", ":", "if", "(", "run_name", "is", "None", ")", ":", "run_name", "=", "'<run_path>'", "importer", "=", "_get_importer", "(", "path_name", ")", "if", ...
execute code located at the specified filesystem location returns the resulting top level namespace dictionary the file path may refer directly to a python script or else it may refer to a zipfile or directory containing a top level __main__ .
train
true
6,244
def rank_order(image): flat_image = image.ravel() sort_order = flat_image.argsort().astype(np.uint32) flat_image = flat_image[sort_order] sort_rank = np.zeros_like(sort_order) is_different = (flat_image[:(-1)] != flat_image[1:]) np.cumsum(is_different, out=sort_rank[1:]) original_values = np.zeros(((sort_rank[(-1)] + 1),), image.dtype) original_values[0] = flat_image[0] original_values[1:] = flat_image[1:][is_different] int_image = np.zeros_like(sort_order) int_image[sort_order] = sort_rank return (int_image.reshape(image.shape), original_values)
[ "def", "rank_order", "(", "image", ")", ":", "flat_image", "=", "image", ".", "ravel", "(", ")", "sort_order", "=", "flat_image", ".", "argsort", "(", ")", ".", "astype", "(", "np", ".", "uint32", ")", "flat_image", "=", "flat_image", "[", "sort_order", ...
return an image of the same shape where each pixel is the index of the pixel value in the ascending order of the unique values of image .
train
false
6,245
def get_plus_sign_symbol(locale=LC_NUMERIC): return Locale.parse(locale).number_symbols.get('plusSign', u'+')
[ "def", "get_plus_sign_symbol", "(", "locale", "=", "LC_NUMERIC", ")", ":", "return", "Locale", ".", "parse", "(", "locale", ")", ".", "number_symbols", ".", "get", "(", "'plusSign'", ",", "u'+'", ")" ]
return the plus sign symbol used by the current locale .
train
false
6,247
def ports_open(name, ports, proto='tcp', direction='in'): ports = list(map(str, ports)) diff = False ret = {'name': ','.join(ports), 'changes': {}, 'result': True, 'comment': 'Ports open.'} current_ports = __salt__['csf.get_ports'](proto=proto, direction=direction) direction = direction.upper() directions = __salt__['csf.build_directions'](direction) for direction in directions: print current_ports[direction] print ports if (current_ports[direction] != ports): diff = True if diff: result = __salt__['csf.allow_ports'](ports, proto=proto, direction=direction) ret['changes']['Ports'] = 'Changed' ret['comment'] = result return ret
[ "def", "ports_open", "(", "name", ",", "ports", ",", "proto", "=", "'tcp'", ",", "direction", "=", "'in'", ")", ":", "ports", "=", "list", "(", "map", "(", "str", ",", "ports", ")", ")", "diff", "=", "False", "ret", "=", "{", "'name'", ":", "','"...
ensure ports are open for a protocol .
train
false
6,250
def _package(package, subdirs=None): dirs = package.split(u'.') app_dir = os.path.join(u'share', u'git-cola', u'lib', *dirs) if subdirs: dirs = (list(subdirs) + dirs) src_dir = os.path.join(*dirs) return (app_dir, glob(os.path.join(src_dir, u'*.py')))
[ "def", "_package", "(", "package", ",", "subdirs", "=", "None", ")", ":", "dirs", "=", "package", ".", "split", "(", "u'.'", ")", "app_dir", "=", "os", ".", "path", ".", "join", "(", "u'share'", ",", "u'git-cola'", ",", "u'lib'", ",", "*", "dirs", ...
collect python files for a given python "package" name .
train
false
6,252
@pytest.mark.xfail(reason='pickle does not produce same error') @pytest.mark.parametrize('serial', all_formats) def test_builtin_403_exception(iris_server, serial): t = symbol('t', discover(iris)) for name in ('map', 'apply'): func = getattr(t.species, name) expr = func(eval, 'int') query = {'expr': to_tree(expr)} response = iris_server.post('/compute', data=serial.dumps(query), headers=mimetype(serial)) assert ('403 FORBIDDEN'.lower() in response.status.lower())
[ "@", "pytest", ".", "mark", ".", "xfail", "(", "reason", "=", "'pickle does not produce same error'", ")", "@", "pytest", ".", "mark", ".", "parametrize", "(", "'serial'", ",", "all_formats", ")", "def", "test_builtin_403_exception", "(", "iris_server", ",", "se...
ensure exception is raised when both map and apply are invoked .
train
false
6,253
def test_mixed_mro_respected(): class OC: abc = 3 class OC2(OC, ): pass class NC(object, ): pass class NC2(object, ): abc = 5 class MC(NC, OC2, NC2, OC, ): pass AreEqual(MC.abc, 5)
[ "def", "test_mixed_mro_respected", "(", ")", ":", "class", "OC", ":", "abc", "=", "3", "class", "OC2", "(", "OC", ",", ")", ":", "pass", "class", "NC", "(", "object", ",", ")", ":", "pass", "class", "NC2", "(", "object", ",", ")", ":", "abc", "="...
creates a class with an mro of "mc .
train
false
6,254
def copytree_hardlink(source, dest): copy2 = shutil.copy2 try: shutil.copy2 = os.link shutil.copytree(source, dest) finally: shutil.copy2 = copy2
[ "def", "copytree_hardlink", "(", "source", ",", "dest", ")", ":", "copy2", "=", "shutil", ".", "copy2", "try", ":", "shutil", ".", "copy2", "=", "os", ".", "link", "shutil", ".", "copytree", "(", "source", ",", "dest", ")", "finally", ":", "shutil", ...
recursively copy a directory ala shutils .
train
true
6,255
def assert_identical(a, b): assert_equal(a, b) if (type(b) is str): assert_equal(type(a), type(b)) else: assert_equal(np.asarray(a).dtype.type, np.asarray(b).dtype.type)
[ "def", "assert_identical", "(", "a", ",", "b", ")", ":", "assert_equal", "(", "a", ",", "b", ")", "if", "(", "type", "(", "b", ")", "is", "str", ")", ":", "assert_equal", "(", "type", "(", "a", ")", ",", "type", "(", "b", ")", ")", "else", ":...
assert whether value and type are the same .
train
false
6,258
def _sanitize_params(prefix, suffix, dir): output_type = _infer_return_type(prefix, suffix, dir) if (suffix is None): suffix = output_type() if (prefix is None): if (output_type is str): prefix = template else: prefix = _os.fsencode(template) if (dir is None): if (output_type is str): dir = gettempdir() else: dir = gettempdirb() return (prefix, suffix, dir, output_type)
[ "def", "_sanitize_params", "(", "prefix", ",", "suffix", ",", "dir", ")", ":", "output_type", "=", "_infer_return_type", "(", "prefix", ",", "suffix", ",", "dir", ")", "if", "(", "suffix", "is", "None", ")", ":", "suffix", "=", "output_type", "(", ")", ...
common parameter processing for most apis in this module .
train
false
6,259
def _flag_default_rules(rules): if (len(rules) >= len(DEFAULT_POLICY_RULES)): rules_suffix = tuple(rules[(- len(DEFAULT_POLICY_RULES)):]) if (rules_suffix == DEFAULT_POLICY_RULES): for rule in rules_suffix: rule._is_default_suffix = True
[ "def", "_flag_default_rules", "(", "rules", ")", ":", "if", "(", "len", "(", "rules", ")", ">=", "len", "(", "DEFAULT_POLICY_RULES", ")", ")", ":", "rules_suffix", "=", "tuple", "(", "rules", "[", "(", "-", "len", "(", "DEFAULT_POLICY_RULES", ")", ")", ...
determine if part of our policy ends with the defaultly appended suffix .
train
false
6,260
def _tconfint_generic(mean, std_mean, dof, alpha, alternative): if (alternative in ['two-sided', '2-sided', '2s']): tcrit = stats.t.ppf((1 - (alpha / 2.0)), dof) lower = (mean - (tcrit * std_mean)) upper = (mean + (tcrit * std_mean)) elif (alternative in ['larger', 'l']): tcrit = stats.t.ppf(alpha, dof) lower = (mean + (tcrit * std_mean)) upper = np.inf elif (alternative in ['smaller', 's']): tcrit = stats.t.ppf((1 - alpha), dof) lower = (- np.inf) upper = (mean + (tcrit * std_mean)) else: raise ValueError('invalid alternative') return (lower, upper)
[ "def", "_tconfint_generic", "(", "mean", ",", "std_mean", ",", "dof", ",", "alpha", ",", "alternative", ")", ":", "if", "(", "alternative", "in", "[", "'two-sided'", ",", "'2-sided'", ",", "'2s'", "]", ")", ":", "tcrit", "=", "stats", ".", "t", ".", ...
generic t-confint to save typing .
train
false
6,261
def count(session, query): counts = query.selectable.with_only_columns([func.count()]) num_results = session.execute(counts.order_by(None)).scalar() if ((num_results is None) or (query._limit is not None)): return query.order_by(None).count() return num_results
[ "def", "count", "(", "session", ",", "query", ")", ":", "counts", "=", "query", ".", "selectable", ".", "with_only_columns", "(", "[", "func", ".", "count", "(", ")", "]", ")", "num_results", "=", "session", ".", "execute", "(", "counts", ".", "order_b...
get the count of properties for a namespace .
train
false
6,262
def get_fallback_languages(language, site_id=None): try: language = get_language_object(language, site_id) except LanguageError: language = get_languages(site_id)[0] return language.get('fallbacks', [])
[ "def", "get_fallback_languages", "(", "language", ",", "site_id", "=", "None", ")", ":", "try", ":", "language", "=", "get_language_object", "(", "language", ",", "site_id", ")", "except", "LanguageError", ":", "language", "=", "get_languages", "(", "site_id", ...
returns a list of fallback languages for the given language .
train
false
6,263
def _build_gecos(gecos_dict): return '{0},{1},{2},{3}'.format(gecos_dict.get('fullname', ''), gecos_dict.get('roomnumber', ''), gecos_dict.get('workphone', ''), gecos_dict.get('homephone', ''))
[ "def", "_build_gecos", "(", "gecos_dict", ")", ":", "return", "'{0},{1},{2},{3}'", ".", "format", "(", "gecos_dict", ".", "get", "(", "'fullname'", ",", "''", ")", ",", "gecos_dict", ".", "get", "(", "'roomnumber'", ",", "''", ")", ",", "gecos_dict", ".", ...
accepts a dictionary entry containing gecos field names and their values .
train
true
6,264
def parse_snmp(src_ip_port, dst_ip_port, snmp_layer): if (type(snmp_layer.community.val) == str): ver = snmp_layer.version.val msg = ('SNMPv%d community string: %s' % (ver, snmp_layer.community.val)) printer(src_ip_port, dst_ip_port, msg) return True
[ "def", "parse_snmp", "(", "src_ip_port", ",", "dst_ip_port", ",", "snmp_layer", ")", ":", "if", "(", "type", "(", "snmp_layer", ".", "community", ".", "val", ")", "==", "str", ")", ":", "ver", "=", "snmp_layer", ".", "version", ".", "val", "msg", "=", ...
parse out the snmp version and community string .
train
false
6,265
def getNewRepository(): return ExportRepository()
[ "def", "getNewRepository", "(", ")", ":", "return", "ExportRepository", "(", ")" ]
get new repository .
train
false
6,266
def test_pypi_xml_transformation(): pypi_hits = [{'name': 'foo', 'summary': 'foo summary', 'version': '1.0'}, {'name': 'foo', 'summary': 'foo summary v2', 'version': '2.0'}, {'_pypi_ordering': 50, 'name': 'bar', 'summary': 'bar summary', 'version': '1.0'}] expected = [{'versions': ['1.0', '2.0'], 'name': 'foo', 'summary': 'foo summary v2'}, {'versions': ['1.0'], 'name': 'bar', 'summary': 'bar summary'}] assert (transform_hits(pypi_hits) == expected)
[ "def", "test_pypi_xml_transformation", "(", ")", ":", "pypi_hits", "=", "[", "{", "'name'", ":", "'foo'", ",", "'summary'", ":", "'foo summary'", ",", "'version'", ":", "'1.0'", "}", ",", "{", "'name'", ":", "'foo'", ",", "'summary'", ":", "'foo summary v2'"...
test transformation of data structures .
train
false
6,267
def get_role_assignments_for_user(user_db): result = UserRoleAssignment.query(user=user_db.name) return result
[ "def", "get_role_assignments_for_user", "(", "user_db", ")", ":", "result", "=", "UserRoleAssignment", ".", "query", "(", "user", "=", "user_db", ".", "name", ")", "return", "result" ]
retrieve all the userroleassignmentdb objects for a particular user .
train
false
6,268
def _cmp_by_reachable_nh(path1, path2): return None
[ "def", "_cmp_by_reachable_nh", "(", "path1", ",", "path2", ")", ":", "return", "None" ]
compares given paths and selects best path based on reachable next-hop .
train
false
6,269
def read_csv_file(path): from csv import reader with codecs.open(path, u'r', u'utf-8') as msgfile: data = msgfile.read() data = data.replace(chr(28), u'').replace(chr(29), u'') data = reader([r.encode(u'utf-8') for r in data.splitlines()]) newdata = [[unicode(val, u'utf-8') for val in row] for row in data] return newdata
[ "def", "read_csv_file", "(", "path", ")", ":", "from", "csv", "import", "reader", "with", "codecs", ".", "open", "(", "path", ",", "u'r'", ",", "u'utf-8'", ")", "as", "msgfile", ":", "data", "=", "msgfile", ".", "read", "(", ")", "data", "=", "data",...
read csv file and return as list of list .
train
false
6,270
def getMaximumByComplexPaths(paths): maximum = complex((-9.876543219876543e+17), (-9.876543219876543e+17)) for path in paths: for point in path: maximum = getMaximum(maximum, point) return maximum
[ "def", "getMaximumByComplexPaths", "(", "paths", ")", ":", "maximum", "=", "complex", "(", "(", "-", "9.876543219876543e+17", ")", ",", "(", "-", "9.876543219876543e+17", ")", ")", "for", "path", "in", "paths", ":", "for", "point", "in", "path", ":", "maxi...
get a complex with each component the maximum of the respective components of complex paths .
train
false
6,271
def make_transient(instance): state = attributes.instance_state(instance) s = _state_session(state) if s: s._expunge_states([state]) state.expired_attributes.clear() if state.callables: del state.callables if state.key: del state.key if state._deleted: del state._deleted
[ "def", "make_transient", "(", "instance", ")", ":", "state", "=", "attributes", ".", "instance_state", "(", "instance", ")", "s", "=", "_state_session", "(", "state", ")", "if", "s", ":", "s", ".", "_expunge_states", "(", "[", "state", "]", ")", "state",...
make the given instance transient .
train
false
6,272
def _path_for_test_id(test_id, max_segment_length=32): if (test_id.count('.') < 2): raise ValueError(('Must have at least three components (e.g. foo.bar.baz), got: %r' % (test_id,))) return '/'.join((segment[:max_segment_length] for segment in test_id.rsplit('.', 2)))
[ "def", "_path_for_test_id", "(", "test_id", ",", "max_segment_length", "=", "32", ")", ":", "if", "(", "test_id", ".", "count", "(", "'.'", ")", "<", "2", ")", ":", "raise", "ValueError", "(", "(", "'Must have at least three components (e.g. foo.bar.baz), got: %r'...
get the temporary directory path for a test id .
train
false
6,273
def addSphere(elementNode, faces, radius, vertexes): bottom = (- radius.z) sides = evaluate.getSidesMinimumThreeBasedOnPrecision(elementNode, max(radius.x, radius.y, radius.z)) sphereSlices = max((sides / 2), 2) equator = euclidean.getComplexPolygonByComplexRadius(complex(radius.x, radius.y), sides) polygons = [triangle_mesh.getAddIndexedLoop([complex()], vertexes, bottom)] zIncrement = ((radius.z + radius.z) / float(sphereSlices)) z = bottom for sphereSlice in xrange(1, sphereSlices): z += zIncrement zPortion = (abs(z) / radius.z) multipliedPath = euclidean.getComplexPathByMultiplier(math.sqrt((1.0 - (zPortion * zPortion))), equator) polygons.append(triangle_mesh.getAddIndexedLoop(multipliedPath, vertexes, z)) polygons.append(triangle_mesh.getAddIndexedLoop([complex()], vertexes, radius.z)) triangle_mesh.addPillarByLoops(faces, polygons)
[ "def", "addSphere", "(", "elementNode", ",", "faces", ",", "radius", ",", "vertexes", ")", ":", "bottom", "=", "(", "-", "radius", ".", "z", ")", "sides", "=", "evaluate", ".", "getSidesMinimumThreeBasedOnPrecision", "(", "elementNode", ",", "max", "(", "r...
add sphere by radius .
train
false
6,274
def maskMatrix(matrix, shape='circle', radius=1.0, center=(0.0, 0.0)): alphaMask = makeMask(matrix.shape[0], shape, radius, center=(0.0, 0.0), range=[0, 1]) return (matrix * alphaMask)
[ "def", "maskMatrix", "(", "matrix", ",", "shape", "=", "'circle'", ",", "radius", "=", "1.0", ",", "center", "=", "(", "0.0", ",", "0.0", ")", ")", ":", "alphaMask", "=", "makeMask", "(", "matrix", ".", "shape", "[", "0", "]", ",", "shape", ",", ...
make and apply a mask to an input matrix .
train
false
6,275
def get_str_resource_ref_from_model(model): return get_resource_ref_from_model(model).ref
[ "def", "get_str_resource_ref_from_model", "(", "model", ")", ":", "return", "get_resource_ref_from_model", "(", "model", ")", ".", "ref" ]
return a resource reference as string given db_model .
train
false
6,276
def patch_sys(stdin=True, stdout=True, stderr=True): if PY3: return if stdin: _patch_sys_std('stdin') if stdout: _patch_sys_std('stdout') if stderr: _patch_sys_std('stderr')
[ "def", "patch_sys", "(", "stdin", "=", "True", ",", "stdout", "=", "True", ",", "stderr", "=", "True", ")", ":", "if", "PY3", ":", "return", "if", "stdin", ":", "_patch_sys_std", "(", "'stdin'", ")", "if", "stdout", ":", "_patch_sys_std", "(", "'stdout...
patch sys .
train
false
6,277
def local_url(port): return 'http://localhost:{}'.format(str(port))
[ "def", "local_url", "(", "port", ")", ":", "return", "'http://localhost:{}'", ".", "format", "(", "str", "(", "port", ")", ")" ]
generates url for a service when running locally .
train
false
6,278
def encoding_exists(encoding, _aliases=_ENCODING_ALIASES): try: codecs.lookup(resolve_encoding(encoding, _aliases)) except LookupError: return False return True
[ "def", "encoding_exists", "(", "encoding", ",", "_aliases", "=", "_ENCODING_ALIASES", ")", ":", "try", ":", "codecs", ".", "lookup", "(", "resolve_encoding", "(", "encoding", ",", "_aliases", ")", ")", "except", "LookupError", ":", "return", "False", "return",...
returns true if encoding is valid .
train
false
6,280
def colormaps(): return sorted(cm.cmap_d)
[ "def", "colormaps", "(", ")", ":", "return", "sorted", "(", "cm", ".", "cmap_d", ")" ]
matplotlib provides a number of colormaps .
train
false
6,281
def dict_to_one(dp_dict={}): return {x: 1 for x in dp_dict}
[ "def", "dict_to_one", "(", "dp_dict", "=", "{", "}", ")", ":", "return", "{", "x", ":", "1", "for", "x", "in", "dp_dict", "}" ]
input a dictionary .
train
false
6,282
def pts_to_midstep(x, *args): steps = np.zeros(((1 + len(args)), (2 * len(x)))) x = np.asanyarray(x) steps[0, 1:(-1):2] = steps[0, 2::2] = ((x[:(-1)] + x[1:]) / 2) (steps[(0, 0)], steps[(0, (-1))]) = (x[0], x[(-1)]) steps[1:, 0::2] = args steps[1:, 1::2] = steps[1:, 0::2] return steps
[ "def", "pts_to_midstep", "(", "x", ",", "*", "args", ")", ":", "steps", "=", "np", ".", "zeros", "(", "(", "(", "1", "+", "len", "(", "args", ")", ")", ",", "(", "2", "*", "len", "(", "x", ")", ")", ")", ")", "x", "=", "np", ".", "asanyar...
convert continuous line to mid-steps .
train
false
6,284
def removeIdentifiersFromDictionary(dictionary): euclidean.removeElementsFromDictionary(dictionary, ['id', 'name', 'tags']) return dictionary
[ "def", "removeIdentifiersFromDictionary", "(", "dictionary", ")", ":", "euclidean", ".", "removeElementsFromDictionary", "(", "dictionary", ",", "[", "'id'", ",", "'name'", ",", "'tags'", "]", ")", "return", "dictionary" ]
remove the identifier elements from a dictionary .
train
false
6,286
def test_compute_debiasing(): rng = np.random.RandomState(42) G = rng.randn(10, 4) X = rng.randn(4, 20) debias_true = np.arange(1, 5, dtype=np.float) M = np.dot(G, (X * debias_true[:, np.newaxis])) debias = compute_bias(M, G, X, max_iter=10000, n_orient=1, tol=1e-07) assert_almost_equal(debias, debias_true, decimal=5) debias = compute_bias(M, G, X, max_iter=10000, n_orient=2, tol=1e-05) assert_almost_equal(debias, [1.8, 1.8, 3.72, 3.72], decimal=2)
[ "def", "test_compute_debiasing", "(", ")", ":", "rng", "=", "np", ".", "random", ".", "RandomState", "(", "42", ")", "G", "=", "rng", ".", "randn", "(", "10", ",", "4", ")", "X", "=", "rng", ".", "randn", "(", "4", ",", "20", ")", "debias_true", ...
test source amplitude debiasing .
train
false
6,287
def set_nodename(facts): if (('node' in facts) and ('common' in facts)): if (('cloudprovider' in facts) and (facts['cloudprovider']['kind'] == 'openstack')): facts['node']['nodename'] = facts['provider']['metadata']['hostname'].replace('.novalocal', '') else: facts['node']['nodename'] = facts['common']['hostname'].lower() return facts
[ "def", "set_nodename", "(", "facts", ")", ":", "if", "(", "(", "'node'", "in", "facts", ")", "and", "(", "'common'", "in", "facts", ")", ")", ":", "if", "(", "(", "'cloudprovider'", "in", "facts", ")", "and", "(", "facts", "[", "'cloudprovider'", "]"...
set nodename .
train
false
6,290
@requires_pyopengl() def test_pyopengl(): from vispy.gloo.gl import pyopengl2 _test_function_names(pyopengl2) _test_constant_names(pyopengl2)
[ "@", "requires_pyopengl", "(", ")", "def", "test_pyopengl", "(", ")", ":", "from", "vispy", ".", "gloo", ".", "gl", "import", "pyopengl2", "_test_function_names", "(", "pyopengl2", ")", "_test_constant_names", "(", "pyopengl2", ")" ]
pyopengl backend should have all es 2 .
train
false
6,291
def get_trigger(trigger): if callable(trigger): return trigger elif (trigger is None): return _never_fire_trigger else: return interval.IntervalTrigger(*trigger)
[ "def", "get_trigger", "(", "trigger", ")", ":", "if", "callable", "(", "trigger", ")", ":", "return", "trigger", "elif", "(", "trigger", "is", "None", ")", ":", "return", "_never_fire_trigger", "else", ":", "return", "interval", ".", "IntervalTrigger", "(", ...
gets a trigger object .
train
false
6,292
def get_repository_type_from_tool_shed(app, tool_shed_url, name, owner): tool_shed_url = common_util.get_tool_shed_url_from_tool_shed_registry(app, tool_shed_url) params = dict(name=name, owner=owner) pathspec = ['repository', 'get_repository_type'] repository_type = util.url_get(tool_shed_url, password_mgr=app.tool_shed_registry.url_auth(tool_shed_url), pathspec=pathspec, params=params) return repository_type
[ "def", "get_repository_type_from_tool_shed", "(", "app", ",", "tool_shed_url", ",", "name", ",", "owner", ")", ":", "tool_shed_url", "=", "common_util", ".", "get_tool_shed_url_from_tool_shed_registry", "(", "app", ",", "tool_shed_url", ")", "params", "=", "dict", "...
send a request to the tool shed to retrieve the type for a repository defined by the combination of a name and owner .
train
false
6,293
def apply_received_command(event): device_id = slugify(event.device.id_string.lower()) if (device_id not in RFX_DEVICES): return _LOGGER.debug('Device_id: %s device_update. Command: %s', device_id, event.values['Command']) if ((event.values['Command'] == 'On') or (event.values['Command'] == 'Off')): is_on = (event.values['Command'] == 'On') RFX_DEVICES[device_id].update_state(is_on) elif (hasattr(RFX_DEVICES[device_id], 'brightness') and (event.values['Command'] == 'Set level')): _brightness = ((event.values['Dim level'] * 255) // 100) is_on = (_brightness > 0) RFX_DEVICES[device_id].update_state(is_on, _brightness) if RFX_DEVICES[device_id].should_fire_event: RFX_DEVICES[device_id].hass.bus.fire(EVENT_BUTTON_PRESSED, {ATTR_ENTITY_ID: RFX_DEVICES[device_id].entity_id, ATTR_STATE: event.values['Command'].lower()}) _LOGGER.info('Rfxtrx fired event: (event_type: %s, %s: %s, %s: %s)', EVENT_BUTTON_PRESSED, ATTR_ENTITY_ID, RFX_DEVICES[device_id].entity_id, ATTR_STATE, event.values['Command'].lower())
[ "def", "apply_received_command", "(", "event", ")", ":", "device_id", "=", "slugify", "(", "event", ".", "device", ".", "id_string", ".", "lower", "(", ")", ")", "if", "(", "device_id", "not", "in", "RFX_DEVICES", ")", ":", "return", "_LOGGER", ".", "deb...
apply command from rfxtrx .
train
false
6,294
def timing(function): @wraps(function) def wrapped(*args, **kwargs): start_time = time.time() ret = function(*args, **salt.utils.clean_kwargs(**kwargs)) end_time = time.time() if function.__module__.startswith('salt.loaded.int.'): mod_name = function.__module__[16:] else: mod_name = function.__module__ log.profile('Function {0}.{1} took {2:.20f} seconds to execute'.format(mod_name, function.__name__, (end_time - start_time))) return ret return wrapped
[ "def", "timing", "(", "function", ")", ":", "@", "wraps", "(", "function", ")", "def", "wrapped", "(", "*", "args", ",", "**", "kwargs", ")", ":", "start_time", "=", "time", ".", "time", "(", ")", "ret", "=", "function", "(", "*", "args", ",", "*...
timing -> t_total execute a function once .
train
true
6,295
@with_setup(prepare_stdout) def test_output_outlines_success_colorless(): runner = Runner(join_path('ru', 'success', 'outlines.feature'), verbosity=3, no_color=True) runner.run() assert_stdout_lines(u'\n\u0424\u0443\u043d\u043a\u0446\u0438\u043e\u043d\u0430\u043b: \u041f\u0440\u043e\u0432\u0435\u0440\u0438\u0442\u044c \u0432\u044b\u0432\u043e\u0434 \u0441\u0442\u0440\u0443\u043a\u0442\u0443\u0440\u043d\u043e\u0433\u043e \u0441\u0446\u0435\u043d\u0430\u0440\u0438\u044f # tests/functional/language_specific_features/ru/success/outlines.feature:3\n \u041a\u0430\u043a \u043f\u0440\u043e\u0433\u0440\u0430\u043c\u043c\u0438\u0441\u0442 # tests/functional/language_specific_features/ru/success/outlines.feature:4\n \u0414\u043b\u044f \u0442\u043e\u0433\u043e \u0447\u043e\u0431\u044b lettuce \u0431\u044b\u043b \u043d\u0430\u0434\u0435\u0436\u043d\u044b\u043c # tests/functional/language_specific_features/ru/success/outlines.feature:5\n \u042f \u0445\u043e\u0447\u0443, \u0447\u0442\u043e \u0431\u044b \u0441\u0446\u0435\u043d\u0430\u0440\u0438\u0438 \u0441\u043e \u0441\u0442\u0440\u0443\u043a\u0442\u0443\u0440\u043e\u0439 \u0440\u0430\u0431\u043e\u0442\u0430\u043b\u0438 \u043d\u0430 \u0440\u0443\u0441\u0441\u043a\u043e\u043c # tests/functional/language_specific_features/ru/success/outlines.feature:6\n\n \u0421\u0442\u0440\u0443\u043a\u0442\u0443\u0440\u0430 \u0441\u0446\u0435\u043d\u0430\u0440\u0438\u044f: \u0417\u0430\u043f\u043e\u043b\u043d\u0438\u0442\u044c \u0444\u043e\u0440\u043c\u0443 # tests/functional/language_specific_features/ru/success/outlines.feature:8\n \u041f\u0443\u0441\u043a\u0430\u0439 \u044f \u043e\u0442\u043a\u0440\u044b\u0432\u0430\u044e \u0432 \u0431\u0440\u0430\u0443\u0437\u0435\u0440\u0435 "http://sona-studio.com/contacts/" # tests/functional/language_specific_features/ru/success/outlines_steps.py:12\n \u041a\u043e\u0433\u0434\u0430 \u044f \u0437\u0430\u043f\u043e\u043b\u043d\u044f\u044e \u0432 \u043f\u043e\u043b\u0435 "\u0418\u043c\u044f" "<\u0438\u043c\u044f>" # tests/functional/language_specific_features/ru/success/outlines_steps.py:16\n \u0418 \u044f \u0437\u0430\u043f\u043e\u043b\u043d\u044f\u044e \u0432 \u043f\u043e\u043b\u0435 "Email" "<email>" # tests/functional/language_specific_features/ru/success/outlines_steps.py:24\n \u0418 \u044f \u0437\u0430\u043f\u043e\u043b\u043d\u044f\u044e \u0432 \u043f\u043e\u043b\u0435 "\u0421\u043e\u043e\u0431\u0449\u0435\u043d\u0438\u0435" "<\u0441\u043e\u043e\u0431\u0449\u0435\u043d\u0438\u0435>" # tests/functional/language_specific_features/ru/success/outlines_steps.py:32\n \u0418 \u044f \u043d\u0430\u0436\u0438\u043c\u0430\u044e "\u041e\u0442\u043f\u0440\u0430\u0432\u0438\u0442\u044c" # tests/functional/language_specific_features/ru/success/outlines_steps.py:40\n \u0422\u043e\u0433\u0434\u0430 \u044f \u043f\u043e\u043b\u0443\u0447\u0430\u044e \u0441\u043e\u043e\u0431\u0449\u0435\u043d\u0438\u0435 "\u0421\u043f\u0430\u0441\u0438\u0431\u043e \u0437\u0430 \u0432\u0430\u0448\u0435 \u0441\u043e\u043e\u0431\u0449\u0435\u043d\u0438\u0435" # tests/functional/language_specific_features/ru/success/outlines_steps.py:43\n\n \u041f\u0440\u0438\u043c\u0435\u0440\u044b:\n | \u0438\u043c\u044f | email | \u0441\u043e\u043e\u0431\u0449\u0435\u043d\u0438\u0435 |\n | \u0412\u0438\u0442\u0430\u043b\u0438\u0439 \u0418\u0433\u043e\u0440\u0435\u0432\u0438\u0447 | john@gmail.org | \u0415\u0441\u0442\u044c \u0438\u043d\u0442\u0435\u0440\u0435\u0441\u043d\u044b\u0439 \u043f\u0440\u043e\u0435\u043a\u0442, \u043d\u0443\u0436\u043d\u043e \u043e\u0431\u0441\u0443\u0434\u0438\u0442\u044c |\n | \u041c\u0430\u0440\u0438\u043d\u0430 \u0411\u0430\u043d\u0440\u0430\u0443\u043b | mary@email.com | \u041c\u043d\u0435 \u043d\u0440\u0430\u0432\u044f\u0442\u0441\u044f \u0432\u0430\u0448\u0438 \u0434\u0438\u0437\u0430\u0439\u043d\u044b, \u0445\u043e\u0447\u0443 \u0441\u0430\u0439\u0442 |\n\n1 feature (1 passed)\n2 scenarios (2 passed)\n12 steps (12 passed)\n')
[ "@", "with_setup", "(", "prepare_stdout", ")", "def", "test_output_outlines_success_colorless", "(", ")", ":", "runner", "=", "Runner", "(", "join_path", "(", "'ru'", ",", "'success'", ",", "'outlines.feature'", ")", ",", "verbosity", "=", "3", ",", "no_color", ...
language: fr -> sucess outlines colorless .
train
false
6,296
def parse_mtestfile(fname): with open(fname) as fp: for line in fp: if ('--' in line): line = line[:line.index('--')] if (not line.strip()): continue (lhs, rhs) = line.split('->') (id, fn, arg) = lhs.split() rhs_pieces = rhs.split() exp = rhs_pieces[0] flags = rhs_pieces[1:] (yield (id, fn, float(arg), float(exp), flags))
[ "def", "parse_mtestfile", "(", "fname", ")", ":", "with", "open", "(", "fname", ")", "as", "fp", ":", "for", "line", "in", "fp", ":", "if", "(", "'--'", "in", "line", ")", ":", "line", "=", "line", "[", ":", "line", ".", "index", "(", "'--'", "...
parse a file with test values -- starts a comment blank lines .
train
false
6,297
def GetServiceVersions(namespace): def compare(a, b): if (a == b): return 0 if (b in parentMap[a]): return (-1) if (a in parentMap[b]): return 1 return ((a > b) - (a < b)) if PY3: return sorted([v for (v, n) in iteritems(serviceNsMap) if (n == namespace)], key=cmp_to_key(compare)) else: return sorted([v for (v, n) in iteritems(serviceNsMap) if (n == namespace)], compare)
[ "def", "GetServiceVersions", "(", "namespace", ")", ":", "def", "compare", "(", "a", ",", "b", ")", ":", "if", "(", "a", "==", "b", ")", ":", "return", "0", "if", "(", "b", "in", "parentMap", "[", "a", "]", ")", ":", "return", "(", "-", "1", ...
get all the versions for the service with specified namespace ordered by compatibility .
train
true
6,298
def get_suggestion(exploration_id, thread_id): model = feedback_models.SuggestionModel.get_by_exploration_and_thread_id(exploration_id, thread_id) return (_get_suggestion_from_model(model) if model else None)
[ "def", "get_suggestion", "(", "exploration_id", ",", "thread_id", ")", ":", "model", "=", "feedback_models", ".", "SuggestionModel", ".", "get_by_exploration_and_thread_id", "(", "exploration_id", ",", "thread_id", ")", "return", "(", "_get_suggestion_from_model", "(", ...
fetches the suggestion for the given thread .
train
false
6,300
def push_notebook(document=None, state=None, handle=None): if (state is None): state = _state if state.server_enabled: raise RuntimeError('output_server() has been called, which is incompatible with push_notebook') if (not document): document = state.document if (not document): warnings.warn('No document to push') return if (handle is None): handle = state.last_comms_handle if (not handle): warnings.warn('Cannot find a last shown plot to update. Call output_notebook() and show(..., notebook_handle=True) before push_notebook()') return to_json = document.to_json() if (handle.doc is not document): msg = dict(doc=to_json) else: msg = Document._compute_patch_between_json(handle.json, to_json) handle.comms.send(json.dumps(msg)) handle.update(document, to_json)
[ "def", "push_notebook", "(", "document", "=", "None", ",", "state", "=", "None", ",", "handle", "=", "None", ")", ":", "if", "(", "state", "is", "None", ")", ":", "state", "=", "_state", "if", "state", ".", "server_enabled", ":", "raise", "RuntimeError...
update bokeh plots in a jupyter notebook output cells with new data or property values .
train
false
6,301
def _image_present(client, image_uuid): headers = client.get_image_meta(image_uuid) return ('status' in headers)
[ "def", "_image_present", "(", "client", ",", "image_uuid", ")", ":", "headers", "=", "client", ".", "get_image_meta", "(", "image_uuid", ")", "return", "(", "'status'", "in", "headers", ")" ]
check if an image is present in glance .
train
false
6,302
def build_request_with_data(url, data, api_key, method): http_redirect_with_data_handler = HTTPRedirectWithDataHandler(method=method) opener = urllib2.build_opener(http_redirect_with_data_handler) urllib2.install_opener(opener) url = make_url(url, api_key=api_key, args=None) request = urllib2.Request(url, headers={'Content-Type': 'application/json'}, data=json.dumps(data)) request_method = request.get_method() if (request_method != method): request.get_method = (lambda : method) return (opener, request)
[ "def", "build_request_with_data", "(", "url", ",", "data", ",", "api_key", ",", "method", ")", ":", "http_redirect_with_data_handler", "=", "HTTPRedirectWithDataHandler", "(", "method", "=", "method", ")", "opener", "=", "urllib2", ".", "build_opener", "(", "http_...
build a request with the received method .
train
false
6,304
def spArticlesForLang(lang): if (lang in _SP_ART_CACHE): return _SP_ART_CACHE[lang] spArticles = addTrailingSpace(LANG_ARTICLESget(lang, GENERIC_ARTICLES)) _SP_ART_CACHE[lang] = spArticles return spArticles
[ "def", "spArticlesForLang", "(", "lang", ")", ":", "if", "(", "lang", "in", "_SP_ART_CACHE", ")", ":", "return", "_SP_ART_CACHE", "[", "lang", "]", "spArticles", "=", "addTrailingSpace", "(", "LANG_ARTICLESget", "(", "lang", ",", "GENERIC_ARTICLES", ")", ")", ...
return lists of articles specific for the given language .
train
false
6,306
def createFontList(fontfiles, fontext=u'ttf'): fontlist = [] seen = {} for fpath in fontfiles: verbose.report((u'createFontDict: %s' % fpath), u'debug') fname = os.path.split(fpath)[1] if (fname in seen): continue else: seen[fname] = 1 if (fontext == u'afm'): try: fh = open(fpath, u'rb') except EnvironmentError: verbose.report((u'Could not open font file %s' % fpath)) continue try: font = afm.AFM(fh) except RuntimeError: verbose.report((u'Could not parse font file %s' % fpath)) continue finally: fh.close() try: prop = afmFontProperty(fpath, font) except KeyError: continue else: try: font = ft2font.FT2Font(fpath) except RuntimeError: verbose.report((u'Could not open font file %s' % fpath)) continue except UnicodeError: verbose.report(u'Cannot handle unicode filenames') continue except IOError: verbose.report((u'IO error - cannot open font file %s' % fpath)) continue try: prop = ttfFontProperty(font) except (KeyError, RuntimeError, ValueError): continue fontlist.append(prop) return fontlist
[ "def", "createFontList", "(", "fontfiles", ",", "fontext", "=", "u'ttf'", ")", ":", "fontlist", "=", "[", "]", "seen", "=", "{", "}", "for", "fpath", "in", "fontfiles", ":", "verbose", ".", "report", "(", "(", "u'createFontDict: %s'", "%", "fpath", ")", ...
a function to create a font lookup list .
train
false
6,307
@gof.local_optimizer([sparse.AddSD]) def local_inplace_addsd_ccode(node): if (isinstance(node.op, sparse.AddSD) and theano.config.cxx): out_dtype = scalar.upcast(*node.inputs) if (out_dtype != node.inputs[1].dtype): return new_node = AddSD_ccode(format=node.inputs[0].type.format, inplace=True)(*node.inputs) return [new_node] return False
[ "@", "gof", ".", "local_optimizer", "(", "[", "sparse", ".", "AddSD", "]", ")", "def", "local_inplace_addsd_ccode", "(", "node", ")", ":", "if", "(", "isinstance", "(", "node", ".", "op", ",", "sparse", ".", "AddSD", ")", "and", "theano", ".", "config"...
optimization to insert inplace versions of addsd .
train
false
6,308
def _item_to_project(iterator, resource): return Project.from_api_repr(resource)
[ "def", "_item_to_project", "(", "iterator", ",", "resource", ")", ":", "return", "Project", ".", "from_api_repr", "(", "resource", ")" ]
convert a json project to the native object .
train
false
6,309
@api_versions.wraps('2.35') @utils.arg('--user', metavar='<user-id>', default=None, help=_('List key-pairs of specified user ID (Admin only).')) @utils.arg('--marker', dest='marker', metavar='<marker>', default=None, help=_('The last keypair of the previous page; displays list of keypairs after "marker".')) @utils.arg('--limit', dest='limit', metavar='<limit>', type=int, default=None, help=_("Maximum number of keypairs to display. If limit is bigger than 'CONF.api.max_limit' option of Nova API, limit 'CONF.api.max_limit' will be used instead.")) def do_keypair_list(cs, args): keypairs = cs.keypairs.list(args.user, args.marker, args.limit) columns = _get_keypairs_list_columns(cs, args) utils.print_list(keypairs, columns)
[ "@", "api_versions", ".", "wraps", "(", "'2.35'", ")", "@", "utils", ".", "arg", "(", "'--user'", ",", "metavar", "=", "'<user-id>'", ",", "default", "=", "None", ",", "help", "=", "_", "(", "'List key-pairs of specified user ID (Admin only).'", ")", ")", "@...
print a list of keypairs for a user .
train
false
6,310
def _BuildArgList(fdesc, names): numArgs = max(fdesc[6], len(fdesc[2])) names = list(names) while (None in names): i = names.index(None) names[i] = ('arg%d' % (i,)) names = list(map(MakePublicAttributeName, names[1:])) name_num = 0 while (len(names) < numArgs): names.append(('arg%d' % (len(names),))) for i in range(0, len(names), 5): names[i] = (names[i] + '\n DCTB DCTB DCTB ') return (',' + ', '.join(names))
[ "def", "_BuildArgList", "(", "fdesc", ",", "names", ")", ":", "numArgs", "=", "max", "(", "fdesc", "[", "6", "]", ",", "len", "(", "fdesc", "[", "2", "]", ")", ")", "names", "=", "list", "(", "names", ")", "while", "(", "None", "in", "names", "...
builds list of args to the underlying invoke method .
train
false
6,314
def shells(): shells_fn = '/etc/shells' ret = [] if os.path.exists(shells_fn): try: with salt.utils.fopen(shells_fn, 'r') as shell_fp: lines = shell_fp.read().splitlines() for line in lines: line = line.strip() if line.startswith('#'): continue elif (not line): continue else: ret.append(line) except OSError: log.error("File '{0}' was not found".format(shells_fn)) return ret
[ "def", "shells", "(", ")", ":", "shells_fn", "=", "'/etc/shells'", "ret", "=", "[", "]", "if", "os", ".", "path", ".", "exists", "(", "shells_fn", ")", ":", "try", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "shells_fn", ",", "'r'", ")", ...
lists the valid shells on this system via the /etc/shells file .
train
false
6,315
def _urlmatcher_for_gcs_stub(url): (_, host, _, _, _) = urlparse.urlsplit(url) return (host == common.LOCAL_API_HOST)
[ "def", "_urlmatcher_for_gcs_stub", "(", "url", ")", ":", "(", "_", ",", "host", ",", "_", ",", "_", ",", "_", ")", "=", "urlparse", ".", "urlsplit", "(", "url", ")", "return", "(", "host", "==", "common", ".", "LOCAL_API_HOST", ")" ]
determines whether a url should be handled by gcs stub .
train
false
6,318
def get_power(**kwargs): with _IpmiCommand(**kwargs) as s: return s.get_power()['powerstate']
[ "def", "get_power", "(", "**", "kwargs", ")", ":", "with", "_IpmiCommand", "(", "**", "kwargs", ")", "as", "s", ":", "return", "s", ".", "get_power", "(", ")", "[", "'powerstate'", "]" ]
get current power state the response .
train
false
6,321
def _url_as_string(url): if isinstance(url, Request_): return url.get_full_url() elif isinstance(url, str): return url else: raise TypeError(('Expected type %r or %r' % (str, Request_)))
[ "def", "_url_as_string", "(", "url", ")", ":", "if", "isinstance", "(", "url", ",", "Request_", ")", ":", "return", "url", ".", "get_full_url", "(", ")", "elif", "isinstance", "(", "url", ",", "str", ")", ":", "return", "url", "else", ":", "raise", "...
returns the url string from a url value that is either a string or urllib2 .
train
false
6,322
def retrieve_seq_length_op2(data): return tf.reduce_sum(tf.cast(tf.greater(data, tf.zeros_like(data)), tf.int32), 1)
[ "def", "retrieve_seq_length_op2", "(", "data", ")", ":", "return", "tf", ".", "reduce_sum", "(", "tf", ".", "cast", "(", "tf", ".", "greater", "(", "data", ",", "tf", ".", "zeros_like", "(", "data", ")", ")", ",", "tf", ".", "int32", ")", ",", "1",...
an op to compute the length of a sequence .
train
true
6,324
def _create_ansi_color_dict(color_cls): return {u'ansidefault': color_cls.BLACK, u'ansiblack': color_cls.BLACK, u'ansidarkgray': (color_cls.BLACK | color_cls.INTENSITY), u'ansilightgray': color_cls.GRAY, u'ansiwhite': (color_cls.GRAY | color_cls.INTENSITY), u'ansidarkred': color_cls.RED, u'ansidarkgreen': color_cls.GREEN, u'ansibrown': color_cls.YELLOW, u'ansidarkblue': color_cls.BLUE, u'ansipurple': color_cls.MAGENTA, u'ansiteal': color_cls.CYAN, u'ansired': (color_cls.RED | color_cls.INTENSITY), u'ansigreen': (color_cls.GREEN | color_cls.INTENSITY), u'ansiyellow': (color_cls.YELLOW | color_cls.INTENSITY), u'ansiblue': (color_cls.BLUE | color_cls.INTENSITY), u'ansifuchsia': (color_cls.MAGENTA | color_cls.INTENSITY), u'ansiturquoise': (color_cls.CYAN | color_cls.INTENSITY)}
[ "def", "_create_ansi_color_dict", "(", "color_cls", ")", ":", "return", "{", "u'ansidefault'", ":", "color_cls", ".", "BLACK", ",", "u'ansiblack'", ":", "color_cls", ".", "BLACK", ",", "u'ansidarkgray'", ":", "(", "color_cls", ".", "BLACK", "|", "color_cls", "...
create a table that maps the 16 named ansi colors to their windows code .
train
false
6,325
def getPointsFromSegmentTable(segmentTable): points = [] endpoints = euclidean.getEndpointsFromSegmentTable(segmentTable) for endpoint in endpoints: points.append(endpoint.point) return points
[ "def", "getPointsFromSegmentTable", "(", "segmentTable", ")", ":", "points", "=", "[", "]", "endpoints", "=", "euclidean", ".", "getEndpointsFromSegmentTable", "(", "segmentTable", ")", "for", "endpoint", "in", "endpoints", ":", "points", ".", "append", "(", "en...
get the points from the segment table .
train
false
6,326
def dict_to_xml(metadata_dict): build = ET.Element('build') for (k, v) in metadata_dict.iteritems(): node = ET.SubElement(build, k) node.text = v return ET.tostring(build)
[ "def", "dict_to_xml", "(", "metadata_dict", ")", ":", "build", "=", "ET", ".", "Element", "(", "'build'", ")", "for", "(", "k", ",", "v", ")", "in", "metadata_dict", ".", "iteritems", "(", ")", ":", "node", "=", "ET", ".", "SubElement", "(", "build",...
turn a simple dict of key/value pairs into xml .
train
false
6,327
def enable_ssh(): _current = global_settings() if (_current['Global Settings']['SSH_STATUS']['VALUE'] == 'Y'): return True _xml = '<RIBCL VERSION="2.0">\n <LOGIN USER_LOGIN="adminname" PASSWORD="password">\n <RIB_INFO MODE="write">\n <MOD_GLOBAL_SETTINGS>\n <SSH_STATUS value="Yes"/>\n </MOD_GLOBAL_SETTINGS>\n </RIB_INFO>\n </LOGIN>\n </RIBCL>' return __execute_cmd('Enable_SSH', _xml)
[ "def", "enable_ssh", "(", ")", ":", "_current", "=", "global_settings", "(", ")", "if", "(", "_current", "[", "'Global Settings'", "]", "[", "'SSH_STATUS'", "]", "[", "'VALUE'", "]", "==", "'Y'", ")", ":", "return", "True", "_xml", "=", "'<RIBCL VERSION=\"...
enable the ssh daemon cli example: .
train
false
6,328
def relpath_to_config_or_make(filename): prefix = _find_prefix(filename) return os.path.relpath(os.path.dirname(filename), prefix)
[ "def", "relpath_to_config_or_make", "(", "filename", ")", ":", "prefix", "=", "_find_prefix", "(", "filename", ")", "return", "os", ".", "path", ".", "relpath", "(", "os", ".", "path", ".", "dirname", "(", "filename", ")", ",", "prefix", ")" ]
the following is refactored out of hook-sysconfig and hook-distutils .
train
false
6,329
def send_notification(device_name): current_time = datetime.now() sender = 'sender@domain.com' recipient = 'recipient@domain.com' subject = 'Device {0} was modified'.format(device_name) message = '\nThe running configuration of {0} was modified. \n\nThis change was detected at: {1}\n\n'.format(device_name, current_time) if send_mail(recipient, subject, message, sender): print 'Email notification sent to {}'.format(recipient) return True
[ "def", "send_notification", "(", "device_name", ")", ":", "current_time", "=", "datetime", ".", "now", "(", ")", "sender", "=", "'sender@domain.com'", "recipient", "=", "'recipient@domain.com'", "subject", "=", "'Device {0} was modified'", ".", "format", "(", "devic...
notify concerned persons about recurring document generation .
train
false
6,330
@register.filter def can_write(obj, user): return obj.can_write(user)
[ "@", "register", ".", "filter", "def", "can_write", "(", "obj", ",", "user", ")", ":", "return", "obj", ".", "can_write", "(", "user", ")" ]
takes article or related to article model .
train
false
6,331
def getProfileBaseNameSynonym(repository): if (repository.getProfileDirectory == None): return repository.baseNameSynonym return os.path.join(repository.getProfileDirectory(), repository.baseNameSynonym)
[ "def", "getProfileBaseNameSynonym", "(", "repository", ")", ":", "if", "(", "repository", ".", "getProfileDirectory", "==", "None", ")", ":", "return", "repository", ".", "baseNameSynonym", "return", "os", ".", "path", ".", "join", "(", "repository", ".", "get...
get the profile base file name synonym .
train
false
6,332
def delete_(*keyname): mdata = _check_mdata_delete() valid_keynames = list_() ret = {} for k in keyname: if (mdata and (k in valid_keynames)): cmd = '{0} {1}'.format(mdata, k) ret[k] = (__salt__['cmd.run_all'](cmd)['retcode'] == 0) else: ret[k] = True return ret
[ "def", "delete_", "(", "*", "keyname", ")", ":", "mdata", "=", "_check_mdata_delete", "(", ")", "valid_keynames", "=", "list_", "(", ")", "ret", "=", "{", "}", "for", "k", "in", "keyname", ":", "if", "(", "mdata", "and", "(", "k", "in", "valid_keynam...
delete metadata prop : string name of property cli example: .
train
true
6,333
def organisation(): def prep(r): query = (FS('organisation_id:po_referral_organisation.id') != None) r.resource.add_filter(query) onaccept = s3db.get_config('org_organisation', 'onaccept') s3db.configure('org_organisation', onaccept=(onaccept, s3db.po_organisation_onaccept)) if (r.record and (r.component_name == 'organisation_household')): atable = s3db.po_organisation_area query = ((atable.organisation_id == r.id) & (atable.deleted != True)) rows = db(query).select(atable.area_id) if rows: area_ids = [row.area_id for row in rows] area_ids.append(None) table = r.component.table table.household_id.requires.set_filter(filterby='area_id', filter_opts=area_ids) elif (not r.component): list_fields = ['name', 'name', (T('Type'), 'organisation_organisation_type.organisation_type_id'), (T('Areas'), 'organisation_area.area_id'), 'website'] s3db.configure('org_organisation', list_fields=list_fields) if r.interactive: s3.crud_strings['org_organisation'].update({'label_create': T('Create Agency'), 'title_list': T('Referral Agencies'), 'title_display': T('Agency Details'), 'title_update': T('Edit Agency Details'), 'label_delete_button': T('Delete Agency')}) if (r.component_name == 'area'): s3.crud_strings['po_organisation_area'].update({'label_create': T('Add Area')}) return True s3.prep = prep return s3_rest_controller('org', 'organisation', rheader=s3db.po_rheader)
[ "def", "organisation", "(", ")", ":", "def", "prep", "(", "r", ")", ":", "query", "=", "(", "FS", "(", "'organisation_id:po_referral_organisation.id'", ")", "!=", "None", ")", "r", ".", "resource", ".", "add_filter", "(", "query", ")", "onaccept", "=", "...
restful crud controller .
train
false
6,334
def getBranchMatrixSetElementNode(elementNode): branchMatrix = getBranchMatrix(elementNode) setElementNodeDictionaryMatrix(elementNode, branchMatrix) return branchMatrix
[ "def", "getBranchMatrixSetElementNode", "(", "elementNode", ")", ":", "branchMatrix", "=", "getBranchMatrix", "(", "elementNode", ")", "setElementNodeDictionaryMatrix", "(", "elementNode", ",", "branchMatrix", ")", "return", "branchMatrix" ]
get matrix starting from the object if it exists .
train
false
6,335
def add_ops(op_classes): def f(cls): for (op_attr_name, op_class) in compat.iteritems(op_classes): ops = getattr(cls, '{0}_ops'.format(op_attr_name)) ops_map = getattr(cls, '{0}_op_nodes_map'.format(op_attr_name)) for op in ops: op_node = ops_map[op] if (op_node is not None): made_op = _op_maker(op_class, op) setattr(cls, 'visit_{0}'.format(op_node), made_op) return cls return f
[ "def", "add_ops", "(", "op_classes", ")", ":", "def", "f", "(", "cls", ")", ":", "for", "(", "op_attr_name", ",", "op_class", ")", "in", "compat", ".", "iteritems", "(", "op_classes", ")", ":", "ops", "=", "getattr", "(", "cls", ",", "'{0}_ops'", "."...
decorator to add default implementation of ops .
train
false
6,336
def _ipv4_to_bits(ipaddr): return ''.join([bin(int(x))[2:].rjust(8, '0') for x in ipaddr.split('.')])
[ "def", "_ipv4_to_bits", "(", "ipaddr", ")", ":", "return", "''", ".", "join", "(", "[", "bin", "(", "int", "(", "x", ")", ")", "[", "2", ":", "]", ".", "rjust", "(", "8", ",", "'0'", ")", "for", "x", "in", "ipaddr", ".", "split", "(", "'.'", ...
accepts an ipv4 dotted quad and returns a string representing its binary counterpart .
train
true
6,337
def succeed_with_changes(name): ret = {'name': name, 'changes': {}, 'result': True, 'comment': 'Success!'} ret['changes'] = {'testing': {'old': 'Unchanged', 'new': 'Something pretended to change'}} if __opts__['test']: ret['result'] = None ret['comment'] = "If we weren't testing, this would be successful with changes" return ret
[ "def", "succeed_with_changes", "(", "name", ")", ":", "ret", "=", "{", "'name'", ":", "name", ",", "'changes'", ":", "{", "}", ",", "'result'", ":", "True", ",", "'comment'", ":", "'Success!'", "}", "ret", "[", "'changes'", "]", "=", "{", "'testing'", ...
returns successful and changes is not empty .
train
false
6,338
def get_generator_names_descriptions(): descs = [] for language in registered_languages: for generator in language.html_generators: description = getattr(generator, 'description', None) if (description is None): description = generator.name descs.append((generator.name, description)) return descs
[ "def", "get_generator_names_descriptions", "(", ")", ":", "descs", "=", "[", "]", "for", "language", "in", "registered_languages", ":", "for", "generator", "in", "language", ".", "html_generators", ":", "description", "=", "getattr", "(", "generator", ",", "'des...
return a tuple of the name and description .
train
false
6,340
def get_flowgram_ali_exe(): return 'FlowgramAli_4frame'
[ "def", "get_flowgram_ali_exe", "(", ")", ":", "return", "'FlowgramAli_4frame'" ]
return the executable name of the flowgram alignment prog .
train
false
6,341
@pytest.mark.django_db def test_verify_user(member_with_email): with pytest.raises(EmailAddress.DoesNotExist): EmailAddress.objects.get(user=member_with_email, verified=True) accounts.utils.verify_user(member_with_email) EmailAddress.objects.get(user=member_with_email, email='member_with_email@this.test', primary=True, verified=True)
[ "@", "pytest", ".", "mark", ".", "django_db", "def", "test_verify_user", "(", "member_with_email", ")", ":", "with", "pytest", ".", "raises", "(", "EmailAddress", ".", "DoesNotExist", ")", ":", "EmailAddress", ".", "objects", ".", "get", "(", "user", "=", ...
test verifying user using verify_user function .
train
false
6,342
def pretty_atom(atom_name, default=None): if _use_unicode: return atoms_table[atom_name] else: if (default is not None): return default raise KeyError('only unicode')
[ "def", "pretty_atom", "(", "atom_name", ",", "default", "=", "None", ")", ":", "if", "_use_unicode", ":", "return", "atoms_table", "[", "atom_name", "]", "else", ":", "if", "(", "default", "is", "not", "None", ")", ":", "return", "default", "raise", "Key...
return pretty representation of an atom .
train
false
6,344
def PRGA(S): i = 0 j = 0 while True: i = ((i + 1) % 256) j = ((j + S[i]) % 256) (S[i], S[j]) = (S[j], S[i]) K = S[((S[i] + S[j]) % 256)] (yield K)
[ "def", "PRGA", "(", "S", ")", ":", "i", "=", "0", "j", "=", "0", "while", "True", ":", "i", "=", "(", "(", "i", "+", "1", ")", "%", "256", ")", "j", "=", "(", "(", "j", "+", "S", "[", "i", "]", ")", "%", "256", ")", "(", "S", "[", ...
run pseudo-random generation algorithm .
train
true
6,345
def _rewrite_sin(m_n, s, a, b): from sympy import expand_mul, pi, ceiling, gamma (m, n) = m_n m = expand_mul((m / pi)) n = expand_mul((n / pi)) r = ceiling((((- m) * a) - n.as_real_imag()[0])) return (gamma((((m * s) + n) + r)), gamma((((1 - n) - r) - (m * s))), (((-1) ** r) * pi))
[ "def", "_rewrite_sin", "(", "m_n", ",", "s", ",", "a", ",", "b", ")", ":", "from", "sympy", "import", "expand_mul", ",", "pi", ",", "ceiling", ",", "gamma", "(", "m", ",", "n", ")", "=", "m_n", "m", "=", "expand_mul", "(", "(", "m", "/", "pi", ...
re-write the sine function sin as gamma functions .
train
false
6,348
def isunauthenticated(func): return getattr(func, 'unauthenticated', False)
[ "def", "isunauthenticated", "(", "func", ")", ":", "return", "getattr", "(", "func", ",", "'unauthenticated'", ",", "False", ")" ]
checks to see if the function is marked as not requiring authentication with the @unauthenticated decorator .
train
false
6,349
def convert_db_torrent_to_json(torrent, include_rel_score=False): torrent_name = torrent[2] if ((torrent_name is None) or (len(torrent_name.strip()) == 0)): torrent_name = 'Unnamed torrent' res_json = {'id': torrent[0], 'infohash': torrent[1].encode('hex'), 'name': torrent_name, 'size': torrent[3], 'category': torrent[4], 'num_seeders': (torrent[5] or 0), 'num_leechers': (torrent[6] or 0), 'last_tracker_check': (torrent[7] or 0)} if include_rel_score: res_json['relevance_score'] = torrent[9] return res_json
[ "def", "convert_db_torrent_to_json", "(", "torrent", ",", "include_rel_score", "=", "False", ")", ":", "torrent_name", "=", "torrent", "[", "2", "]", "if", "(", "(", "torrent_name", "is", "None", ")", "or", "(", "len", "(", "torrent_name", ".", "strip", "(...
this method converts a torrent in the database to a json dictionary .
train
false
6,350
def taggedsent_to_conll(sentence): for (i, (word, tag)) in enumerate(sentence, start=1): input_str = [str(i), word, '_', tag, tag, '_', '0', 'a', '_', '_'] input_str = (' DCTB '.join(input_str) + '\n') (yield input_str)
[ "def", "taggedsent_to_conll", "(", "sentence", ")", ":", "for", "(", "i", ",", "(", "word", ",", "tag", ")", ")", "in", "enumerate", "(", "sentence", ",", "start", "=", "1", ")", ":", "input_str", "=", "[", "str", "(", "i", ")", ",", "word", ",",...
a module to convert a single pos tagged sentence into conll format .
train
false