id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
677
def ensure_point(geom): ensure_geometry(geom) if (geom.geom_type != u'Point'): raise SpatialError((u"Provided geometry '%s' is not a 'Point'." % geom)) return geom
[ "def", "ensure_point", "(", "geom", ")", ":", "ensure_geometry", "(", "geom", ")", "if", "(", "geom", ".", "geom_type", "!=", "u'Point'", ")", ":", "raise", "SpatialError", "(", "(", "u\"Provided geometry '%s' is not a 'Point'.\"", "%", "geom", ")", ")", "retu...
makes sure the parameter passed in looks like a geos point .
train
false
678
@positional(1) def get_package_for_module(module): if isinstance(module, basestring): try: module = sys.modules[module] except KeyError: return None try: return unicode(module.package) except AttributeError: if (module.__name__ == '__main__'): try: file_name = module.__file__ except AttributeError: pass else: base_name = os.path.basename(file_name) split_name = os.path.splitext(base_name) if (len(split_name) == 1): return unicode(base_name) else: return u'.'.join(split_name[:(-1)]) return unicode(module.__name__)
[ "@", "positional", "(", "1", ")", "def", "get_package_for_module", "(", "module", ")", ":", "if", "isinstance", "(", "module", ",", "basestring", ")", ":", "try", ":", "module", "=", "sys", ".", "modules", "[", "module", "]", "except", "KeyError", ":", ...
get package name for a module .
train
false
679
def GetAllPosts(): posts = [{'content': str(row[1]), 'time': str(row[0])} for row in DB] posts.sort(key=(lambda row: row['time']), reverse=True) return posts
[ "def", "GetAllPosts", "(", ")", ":", "posts", "=", "[", "{", "'content'", ":", "str", "(", "row", "[", "1", "]", ")", ",", "'time'", ":", "str", "(", "row", "[", "0", "]", ")", "}", "for", "row", "in", "DB", "]", "posts", ".", "sort", "(", ...
get all the posts from the database .
train
false
680
def split_stem(sentence): sentence = re.sub('([a-z])([A-Z])', u'\\1 \\2', sentence) return sentence.split()
[ "def", "split_stem", "(", "sentence", ")", ":", "sentence", "=", "re", ".", "sub", "(", "'([a-z])([A-Z])'", ",", "u'\\\\1 \\\\2'", ",", "sentence", ")", "return", "sentence", ".", "split", "(", ")" ]
splits camel cased sentence into words .
train
false
682
def cookie_app(environ, start_response): response = Response(environ.get('HTTP_COOKIE', 'No Cookie'), mimetype='text/plain') response.set_cookie('test', 'test') return response(environ, start_response)
[ "def", "cookie_app", "(", "environ", ",", "start_response", ")", ":", "response", "=", "Response", "(", "environ", ".", "get", "(", "'HTTP_COOKIE'", ",", "'No Cookie'", ")", ",", "mimetype", "=", "'text/plain'", ")", "response", ".", "set_cookie", "(", "'tes...
a wsgi application which sets a cookie .
train
false
683
def team_status(): return s3_rest_controller()
[ "def", "team_status", "(", ")", ":", "return", "s3_rest_controller", "(", ")" ]
team statuses .
train
false
684
def asciiHexDecode(stream): eod = '>' decodedStream = '' char = '' index = 0 while (index < len(stream)): c = stream[index] if (c == eod): if ((len(decodedStream) % 2) != 0): char += '0' try: decodedStream += chr(int(char, base=16)) except: return ((-1), 'Error in hexadecimal conversion') break elif c.isspace(): index += 1 continue char += c if (len(char) == 2): try: decodedStream += chr(int(char, base=16)) except: return ((-1), 'Error in hexadecimal conversion') char = '' index += 1 return (0, decodedStream)
[ "def", "asciiHexDecode", "(", "stream", ")", ":", "eod", "=", "'>'", "decodedStream", "=", "''", "char", "=", "''", "index", "=", "0", "while", "(", "index", "<", "len", "(", "stream", ")", ")", ":", "c", "=", "stream", "[", "index", "]", "if", "...
method to decode streams using hexadecimal encoding .
train
false
685
def init(mpstate): return SerialModule(mpstate)
[ "def", "init", "(", "mpstate", ")", ":", "return", "SerialModule", "(", "mpstate", ")" ]
initialize all qwebsettings .
train
false
686
def list_stringify(inlist): outlist = [] for item in inlist: if (not isinstance(item, (tuple, list))): if (not isinstance(item, basestring)): item = str(item) else: item = list_stringify(item) outlist.append(item) return outlist
[ "def", "list_stringify", "(", "inlist", ")", ":", "outlist", "=", "[", "]", "for", "item", "in", "inlist", ":", "if", "(", "not", "isinstance", "(", "item", ",", "(", "tuple", ",", "list", ")", ")", ")", ":", "if", "(", "not", "isinstance", "(", ...
recursively rebuilds a list - making sure all the members are strings .
train
false
688
def distro_release_info(): return _distro.distro_release_info()
[ "def", "distro_release_info", "(", ")", ":", "return", "_distro", ".", "distro_release_info", "(", ")" ]
return a dictionary containing key-value pairs for the information items from the distro release file data source of the current linux distribution .
train
false
689
def set_vif_host_backend_802qbh_config(conf, devname, profileid, tapname=None): conf.net_type = 'direct' conf.source_dev = devname conf.source_mode = 'vepa' conf.vporttype = '802.1Qbh' conf.add_vport_param('profileid', profileid) if tapname: conf.target_dev = tapname
[ "def", "set_vif_host_backend_802qbh_config", "(", "conf", ",", "devname", ",", "profileid", ",", "tapname", "=", "None", ")", ":", "conf", ".", "net_type", "=", "'direct'", "conf", ".", "source_dev", "=", "devname", "conf", ".", "source_mode", "=", "'vepa'", ...
populate a libvirtconfigguestinterface instance with host backend details for an 802 .
train
false
690
@treeio_login_required @handle_response_format def source_add(request, response_format='html'): if (not request.user.profile.is_admin('treeio.sales')): return user_denied(request, message="You don't have administrator access to the Sales module") if request.POST: if ('cancel' not in request.POST): source = SaleSource() form = SaleSourceForm(request.user.profile, request.POST, instance=source) if form.is_valid(): source = form.save() source.set_user_from_request(request) return HttpResponseRedirect(reverse('sales_source_view', args=[source.id])) else: return HttpResponseRedirect(reverse('sales_settings_view')) else: form = SaleSourceForm(request.user.profile) all_products = Object.filter_by_request(request, Product.objects.filter(parent__isnull=True)) all_sources = Object.filter_by_request(request, SaleSource.objects) return render_to_response('sales/source_add', {'form': form, 'sources': all_sources, 'products': all_products}, context_instance=RequestContext(request), response_format=response_format)
[ "@", "treeio_login_required", "@", "handle_response_format", "def", "source_add", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "if", "(", "not", "request", ".", "user", ".", "profile", ".", "is_admin", "(", "'treeio.sales'", ")", ")", ":", ...
ticketstatus add .
train
false
691
def _single_spectrum_helper(x, mode, Fs=None, window=None, pad_to=None, sides=None): if ((mode is None) or (mode == u'psd') or (mode == u'default')): raise ValueError((u'_single_spectrum_helper does not work with %s mode' % mode)) if (pad_to is None): pad_to = len(x) (spec, freqs, _) = _spectral_helper(x=x, y=None, NFFT=len(x), Fs=Fs, detrend_func=detrend_none, window=window, noverlap=0, pad_to=pad_to, sides=sides, scale_by_freq=False, mode=mode) if (mode != u'complex'): spec = spec.real if ((spec.ndim == 2) and (spec.shape[1] == 1)): spec = spec[:, 0] return (spec, freqs)
[ "def", "_single_spectrum_helper", "(", "x", ",", "mode", ",", "Fs", "=", "None", ",", "window", "=", "None", ",", "pad_to", "=", "None", ",", "sides", "=", "None", ")", ":", "if", "(", "(", "mode", "is", "None", ")", "or", "(", "mode", "==", "u'p...
this is a helper function that implements the commonality between the complex .
train
false
692
def start(name, call=None): if (call != 'action'): raise SaltCloudSystemExit('The start action must be called with -a or --action.') data = show_instance(name, call='action') if (data.get('status') == 'active'): return {'success': True, 'action': 'start', 'status': 'active', 'msg': 'Machine is already running.'} ret = query(droplet_id=data['id'], command='actions', args={'type': 'power_on'}, http_method='post') return {'success': True, 'action': ret['action']['type'], 'state': ret['action']['status']}
[ "def", "start", "(", "name", ",", "call", "=", "None", ")", ":", "if", "(", "call", "!=", "'action'", ")", ":", "raise", "SaltCloudSystemExit", "(", "'The start action must be called with -a or --action.'", ")", "data", "=", "show_instance", "(", "name", ",", ...
start a machine by name cli example: .
train
true
694
def django_to_webob_request(django_request): return DjangoWebobRequest(django_request)
[ "def", "django_to_webob_request", "(", "django_request", ")", ":", "return", "DjangoWebobRequest", "(", "django_request", ")" ]
returns a webob request to the django_request .
train
false
695
def end_recording(status, firepython_set_extension_data=None): if (firepython_set_extension_data is not None): warnings.warn('Firepython is no longer supported') rec = recorder_proxy.get_for_current_request() recorder_proxy.clear_for_current_request() if config.DEBUG: logging.debug('Cleared recorder') if (rec is not None): try: rec.record_http_status(status) rec.save() finally: memcache.delete(lock_key(), namespace=config.KEY_NAMESPACE)
[ "def", "end_recording", "(", "status", ",", "firepython_set_extension_data", "=", "None", ")", ":", "if", "(", "firepython_set_extension_data", "is", "not", "None", ")", ":", "warnings", ".", "warn", "(", "'Firepython is no longer supported'", ")", "rec", "=", "re...
stop recording rpc traces and save all traces to memcache .
train
false
696
def _getitem_from_frame(f_locals, key, default=None): try: return f_locals[key] except Exception: return default
[ "def", "_getitem_from_frame", "(", "f_locals", ",", "key", ",", "default", "=", "None", ")", ":", "try", ":", "return", "f_locals", "[", "key", "]", "except", "Exception", ":", "return", "default" ]
f_locals is not guaranteed to have .
train
false
698
def coerce_core(result, dshape, odo_kwargs=None): if iscoretype(result): return result elif isscalar(dshape): result = coerce_scalar(result, dshape, odo_kwargs=odo_kwargs) elif (istabular(dshape) and isrecord(dshape.measure)): result = into(DataFrame, result, **(odo_kwargs or {})) elif iscollection(dshape): dim = _dimensions(dshape) if (dim == 1): result = into(Series, result, **(odo_kwargs or {})) elif (dim > 1): result = into(np.ndarray, result, **(odo_kwargs or {})) else: msg = 'Expr with dshape dimensions < 1 should have been handled earlier: dim={}' raise ValueError(msg.format(str(dim))) else: msg = 'Expr does not evaluate to a core return type' raise ValueError(msg) return result
[ "def", "coerce_core", "(", "result", ",", "dshape", ",", "odo_kwargs", "=", "None", ")", ":", "if", "iscoretype", "(", "result", ")", ":", "return", "result", "elif", "isscalar", "(", "dshape", ")", ":", "result", "=", "coerce_scalar", "(", "result", ","...
coerce data to a core data type .
train
false
700
def forget_sr(session, sr_ref): LOG.debug(_('Forgetting SR...')) unplug_pbds(session, sr_ref) session.call_xenapi('SR.forget', sr_ref)
[ "def", "forget_sr", "(", "session", ",", "sr_ref", ")", ":", "LOG", ".", "debug", "(", "_", "(", "'Forgetting SR...'", ")", ")", "unplug_pbds", "(", "session", ",", "sr_ref", ")", "session", ".", "call_xenapi", "(", "'SR.forget'", ",", "sr_ref", ")" ]
forgets the storage repository without destroying the vdis within .
train
false
701
def _get_cycles(graph_dict, path, visited, result, vertice): if (vertice in path): cycle = [vertice] for node in path[::(-1)]: if (node == vertice): break cycle.insert(0, node) start_from = min(cycle) index = cycle.index(start_from) cycle = (cycle[index:] + cycle[0:index]) if (not (cycle in result)): result.append(cycle) return path.append(vertice) try: for node in graph_dict[vertice]: if (node not in visited): _get_cycles(graph_dict, path, visited, result, node) visited.add(node) except KeyError: pass path.pop()
[ "def", "_get_cycles", "(", "graph_dict", ",", "path", ",", "visited", ",", "result", ",", "vertice", ")", ":", "if", "(", "vertice", "in", "path", ")", ":", "cycle", "=", "[", "vertice", "]", "for", "node", "in", "path", "[", ":", ":", "(", "-", ...
recursive function doing the real work for get_cycles .
train
true
702
def get_new_otu_id(old_otu_id, tree, dissim): create_tip_index(tree) cache_tip_names(tree) node = tree._tip_index[old_otu_id] distance_up_tree = 0 while ((not node.isRoot()) and ((distance_up_tree + node.Length) < dissim)): distance_up_tree += node.Length node = node.Parent if node.isTip(): return node.Name else: return choice(node._tip_names)
[ "def", "get_new_otu_id", "(", "old_otu_id", ",", "tree", ",", "dissim", ")", ":", "create_tip_index", "(", "tree", ")", "cache_tip_names", "(", "tree", ")", "node", "=", "tree", ".", "_tip_index", "[", "old_otu_id", "]", "distance_up_tree", "=", "0", "while"...
simulates an otu switching to related one input a tipname .
train
false
703
def copy_config(server_root, temp_dir): (copied_files, copied_dirs) = ([], []) dir_len = len(os.path.dirname(server_root)) for (config_path, config_dirs, config_files) in os.walk(server_root): temp_path = os.path.join(temp_dir, config_path[(dir_len + 1):]) os.mkdir(temp_path) copied_all = True copied_files_in_current_dir = [] for config_file in config_files: config_file_path = os.path.join(config_path, config_file) temp_file_path = os.path.join(temp_path, config_file) if os.path.islink(config_file_path): os.symlink(os.readlink(config_file_path), temp_file_path) elif safe_config_file(config_file_path): copy_file_without_comments(config_file_path, temp_file_path) copied_files_in_current_dir.append(config_file_path) else: copied_all = False if (copied_all and (not config_dirs)): copied_dirs.append(config_path) else: copied_files += copied_files_in_current_dir return (copied_files, copied_dirs)
[ "def", "copy_config", "(", "server_root", ",", "temp_dir", ")", ":", "(", "copied_files", ",", "copied_dirs", ")", "=", "(", "[", "]", ",", "[", "]", ")", "dir_len", "=", "len", "(", "os", ".", "path", ".", "dirname", "(", "server_root", ")", ")", ...
safely copies server_root to temp_dir and returns copied files .
train
false
704
def convert_filename(value): if NORMALIZE_FILENAME: chunks = value.split(os.extsep) normalized = [] for v in chunks: v = unicodedata.normalize('NFKD', six.text_type(v)).encode('ascii', 'ignore').decode('ascii') v = re.sub('[^\\w\\s-]', '', v).strip() normalized.append(v) if (len(normalized) > 1): value = '.'.join(normalized) else: value = normalized[0] if CONVERT_FILENAME: value = value.replace(' ', '_').lower() return value
[ "def", "convert_filename", "(", "value", ")", ":", "if", "NORMALIZE_FILENAME", ":", "chunks", "=", "value", ".", "split", "(", "os", ".", "extsep", ")", "normalized", "=", "[", "]", "for", "v", "in", "chunks", ":", "v", "=", "unicodedata", ".", "normal...
convert filename .
train
false
705
def plugin_order(): p = {} for func in plugin_store: p[func] = [plugin_name for (plugin_name, f) in plugin_store[func]] return p
[ "def", "plugin_order", "(", ")", ":", "p", "=", "{", "}", "for", "func", "in", "plugin_store", ":", "p", "[", "func", "]", "=", "[", "plugin_name", "for", "(", "plugin_name", ",", "f", ")", "in", "plugin_store", "[", "func", "]", "]", "return", "p"...
return the currently preferred plugin order .
train
false
706
def rehashPassword(password, hashParameters): try: (digestname, iterations, salt) = hashParameters.split(Delimiter) except ValueError: raise ValueError(u"Expected hash parameters string in format 'digestmod{0}iterations{0}salt".format(Delimiter)) if (digestname not in Hashes.keys()): raise ValueError(u"Unsupported hash algorithm '{0}' for hash parameters '{1}'.".format(digestname, hash)) iterations = int(iterations) salt = base64.b64decode(salt.encode(u'ascii')) password = password.encode(u'utf-8') return pbkdf2(password, salt, iterations, Hashes[digestname])
[ "def", "rehashPassword", "(", "password", ",", "hashParameters", ")", ":", "try", ":", "(", "digestname", ",", "iterations", ",", "salt", ")", "=", "hashParameters", ".", "split", "(", "Delimiter", ")", "except", "ValueError", ":", "raise", "ValueError", "("...
module function to recreate a password hash given the hash parameters .
train
false
707
def unique_values(func): @wraps(func) def wrapper(*args, **kwargs): return unique_everseen(func(*args, **kwargs)) return wrapper
[ "def", "unique_values", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "wrapper", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "unique_everseen", "(", "func", "(", "*", "args", ",", "**", "kwargs", ")", ")", "return", "...
wrap a function returning an iterable such that the resulting iterable only ever yields unique items .
train
true
708
def pages(post_objects, request): paginator = Paginator(post_objects, 20) try: current_page = int(request.GET.get('page', '1')) except ValueError: current_page = 1 page_range = page_list_return(len(paginator.page_range), current_page) try: page_objects = paginator.page(current_page) except (EmptyPage, InvalidPage): page_objects = paginator.page(paginator.num_pages) if (current_page >= 5): show_first = 1 else: show_first = 0 if (current_page <= (len(paginator.page_range) - 3)): show_end = 1 else: show_end = 0 return (post_objects, paginator, page_objects, page_range, current_page, show_first, show_end)
[ "def", "pages", "(", "post_objects", ",", "request", ")", ":", "paginator", "=", "Paginator", "(", "post_objects", ",", "20", ")", "try", ":", "current_page", "=", "int", "(", "request", ".", "GET", ".", "get", "(", "'page'", ",", "'1'", ")", ")", "e...
page public function .
train
false
709
def verbose_print(arg): if support.verbose: with _print_mutex: print arg
[ "def", "verbose_print", "(", "arg", ")", ":", "if", "support", ".", "verbose", ":", "with", "_print_mutex", ":", "print", "arg" ]
only prints s if verbose is true .
train
false
711
def exc_message(exc_info): exc = exc_info[1] if (exc is None): result = exc_info[0] else: try: result = str(exc) except UnicodeEncodeError: try: result = unicode(exc) except UnicodeError: result = exc.args[0] result = force_unicode(result, 'UTF-8') return xml_safe(result)
[ "def", "exc_message", "(", "exc_info", ")", ":", "exc", "=", "exc_info", "[", "1", "]", "if", "(", "exc", "is", "None", ")", ":", "result", "=", "exc_info", "[", "0", "]", "else", ":", "try", ":", "result", "=", "str", "(", "exc", ")", "except", ...
return the exceptions message .
train
true
712
def validate_target(func, *args, **kwargs): def inner(self, *args, **kwargs): target_id = None if (('target_id' in kwargs) and (kwargs['target_id'] != None)): target_id = kwargs['target_id'] else: target_id = 0 if (not self.target_is_valid(target_id)): raise NoSuchTargetException() return func(self, *args, **kwargs) return inner
[ "def", "validate_target", "(", "func", ",", "*", "args", ",", "**", "kwargs", ")", ":", "def", "inner", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "target_id", "=", "None", "if", "(", "(", "'target_id'", "in", "kwargs", ")", "and...
a decorator that ensures that the specified target_id exists and is valid .
train
true
713
@handle_response_format @treeio_login_required @module_admin_required() def pagefolder_delete(request, folder_id, response_format='html'): folder = get_object_or_404(PageFolder, pk=folder_id) pages = Page.objects.filter(folder=folder).order_by('name') if request.POST: if ('delete' in request.POST): folder.delete() return HttpResponseRedirect(reverse('core_admin_index_pages')) elif ('cancel' in request.POST): return HttpResponseRedirect(reverse('core_admin_pagefolder_view', args=[folder.id])) return render_to_response('core/administration/pagefolder_delete', {'folder': folder, 'pages': pages}, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "@", "module_admin_required", "(", ")", "def", "pagefolder_delete", "(", "request", ",", "folder_id", ",", "response_format", "=", "'html'", ")", ":", "folder", "=", "get_object_or_404", "(", "PageFolder", ...
folder for static pages delete .
train
false
714
def action_events_get(context, action_id): return IMPL.action_events_get(context, action_id)
[ "def", "action_events_get", "(", "context", ",", "action_id", ")", ":", "return", "IMPL", ".", "action_events_get", "(", "context", ",", "action_id", ")" ]
get the events by action id .
train
false
715
def test_double_start_finished(qtbot, proc, py_proc): with qtbot.waitSignals([proc.started, proc.finished], timeout=10000, order='strict'): argv = py_proc('import sys; sys.exit(0)') proc.start(*argv) with qtbot.waitSignals([proc.started, proc.finished], timeout=10000, order='strict'): argv = py_proc('import sys; sys.exit(0)') proc.start(*argv)
[ "def", "test_double_start_finished", "(", "qtbot", ",", "proc", ",", "py_proc", ")", ":", "with", "qtbot", ".", "waitSignals", "(", "[", "proc", ".", "started", ",", "proc", ".", "finished", "]", ",", "timeout", "=", "10000", ",", "order", "=", "'strict'...
test starting a guiprocess twice .
train
false
716
def StrJoin(delim, data): return delim.join((str(x) for x in data))
[ "def", "StrJoin", "(", "delim", ",", "data", ")", ":", "return", "delim", ".", "join", "(", "(", "str", "(", "x", ")", "for", "x", "in", "data", ")", ")" ]
string-ize & join data .
train
false
717
def _parse_date_columns(data_frame, parse_dates): if ((parse_dates is True) or (parse_dates is None) or (parse_dates is False)): parse_dates = [] if (not hasattr(parse_dates, '__iter__')): parse_dates = [parse_dates] for col_name in parse_dates: df_col = data_frame[col_name] try: fmt = parse_dates[col_name] except TypeError: fmt = None data_frame[col_name] = _handle_date_column(df_col, format=fmt) for (col_name, df_col) in data_frame.iteritems(): if is_datetime64tz_dtype(df_col): data_frame[col_name] = _handle_date_column(df_col) return data_frame
[ "def", "_parse_date_columns", "(", "data_frame", ",", "parse_dates", ")", ":", "if", "(", "(", "parse_dates", "is", "True", ")", "or", "(", "parse_dates", "is", "None", ")", "or", "(", "parse_dates", "is", "False", ")", ")", ":", "parse_dates", "=", "[",...
force non-datetime columns to be read as such .
train
false
718
def _ptransform(p): return ((-1.0) / (1.0 + (1.5 * _phi(((1.0 + p) / 2.0)))))
[ "def", "_ptransform", "(", "p", ")", ":", "return", "(", "(", "-", "1.0", ")", "/", "(", "1.0", "+", "(", "1.5", "*", "_phi", "(", "(", "(", "1.0", "+", "p", ")", "/", "2.0", ")", ")", ")", ")", ")" ]
function for p-value abcissa transformation .
train
false
720
def get_price_list(): rate = {} price_list = frappe.db.sql(u'select ip.item_code, ip.buying, ip.selling,\n DCTB DCTB concat(ifnull(cu.symbol,ip.currency), " ", round(ip.price_list_rate,2), " - ", ip.price_list) as price\n DCTB DCTB from `tabItem Price` ip, `tabPrice List` pl, `tabCurrency` cu\n DCTB DCTB where ip.price_list=pl.name and pl.currency=cu.name and pl.enabled=1', as_dict=1) for j in price_list: if j.price: rate.setdefault(j.item_code, {}).setdefault((u'Buying' if j.buying else u'Selling'), []).append(j.price) item_rate_map = {} for item in rate: for buying_or_selling in rate[item]: item_rate_map.setdefault(item, {}).setdefault(buying_or_selling, u', '.join(rate[item].get(buying_or_selling, []))) return item_rate_map
[ "def", "get_price_list", "(", ")", ":", "rate", "=", "{", "}", "price_list", "=", "frappe", ".", "db", ".", "sql", "(", "u'select ip.item_code, ip.buying, ip.selling,\\n DCTB DCTB concat(ifnull(cu.symbol,ip.currency), \" \", round(ip.price_list_rate,2), \" - \", ip.price_list) as ...
get selling & buying price list of every item .
train
false
722
def matrix_tensor_product(*product): if isinstance(product[0], Matrix): return _sympy_tensor_product(*product) elif isinstance(product[0], numpy_ndarray): return _numpy_tensor_product(*product) elif isinstance(product[0], scipy_sparse_matrix): return _scipy_sparse_tensor_product(*product)
[ "def", "matrix_tensor_product", "(", "*", "product", ")", ":", "if", "isinstance", "(", "product", "[", "0", "]", ",", "Matrix", ")", ":", "return", "_sympy_tensor_product", "(", "*", "product", ")", "elif", "isinstance", "(", "product", "[", "0", "]", "...
compute the matrix tensor product of sympy/numpy/scipy .
train
false
723
@hug.post() def post_here(body): return body
[ "@", "hug", ".", "post", "(", ")", "def", "post_here", "(", "body", ")", ":", "return", "body" ]
this example shows how to read in post data w/ hug outside of its automatic param parsing .
train
false
724
def _add_metaclass(metaclass): def wrapper(cls): orig_vars = cls.__dict__.copy() slots = orig_vars.get('__slots__') if (slots is not None): if isinstance(slots, str): slots = [slots] for slots_var in slots: orig_vars.pop(slots_var) orig_vars.pop('__dict__', None) orig_vars.pop('__weakref__', None) return metaclass(cls.__name__, cls.__bases__, orig_vars) return wrapper
[ "def", "_add_metaclass", "(", "metaclass", ")", ":", "def", "wrapper", "(", "cls", ")", ":", "orig_vars", "=", "cls", ".", "__dict__", ".", "copy", "(", ")", "slots", "=", "orig_vars", ".", "get", "(", "'__slots__'", ")", "if", "(", "slots", "is", "n...
class decorator for creating a class with a metaclass .
train
false
725
@handle_response_format @treeio_login_required def mlist_add(request, response_format='html'): user = request.user.profile if request.POST: mlist = MailingList() form = MailingListForm(user, request.POST, instance=mlist) if form.is_valid(): mlist = form.save() mlist.set_user_from_request(request) return HttpResponseRedirect('/messaging/') else: form = MailingListForm(user) context = _get_default_context(request) context.update({'form': form}) return render_to_response('messaging/mlist_add', context, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "def", "mlist_add", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "user", "=", "request", ".", "user", ".", "profile", "if", "request", ".", "POST", ":", "mlist", "=", "MailingList...
new message mlist .
train
false
726
def Confirm(message): return bool((PresentDialog(message, [u'Ok', u'Cancel']) == 0))
[ "def", "Confirm", "(", "message", ")", ":", "return", "bool", "(", "(", "PresentDialog", "(", "message", ",", "[", "u'Ok'", ",", "u'Cancel'", "]", ")", "==", "0", ")", ")" ]
display |message| with ok/cancel operations .
train
false
727
def fileOpenDlg(tryFilePath='', tryFileName='', prompt=_translate('Select file to open'), allowed=None): global qtapp qtapp = ensureQtApp() if (allowed is None): allowed = 'All files (*.*);;PsychoPy Data (*.psydat);;txt (*.txt *.dlm *.csv);;pickled files (*.pickle *.pkl);;shelved files (*.shelf)' fdir = os.path.join(tryFilePath, tryFileName) filesToOpen = QtWidgets.QFileDialog.getOpenFileNames(parent=None, caption=prompt, directory=fdir, filter=allowed) if (type(filesToOpen) == tuple): filesToOpen = filesToOpen[0] filesToOpen = [unicode(fpath) for fpath in filesToOpen if os.path.exists(fpath)] if (len(filesToOpen) == 0): return None return filesToOpen
[ "def", "fileOpenDlg", "(", "tryFilePath", "=", "''", ",", "tryFileName", "=", "''", ",", "prompt", "=", "_translate", "(", "'Select file to open'", ")", ",", "allowed", "=", "None", ")", ":", "global", "qtapp", "qtapp", "=", "ensureQtApp", "(", ")", "if", ...
a simple dialogue allowing read access to the file system .
train
false
728
@task @timed def install_node_prereqs(): if no_prereq_install(): print NO_PREREQ_MESSAGE return prereq_cache('Node prereqs', ['package.json'], node_prereqs_installation)
[ "@", "task", "@", "timed", "def", "install_node_prereqs", "(", ")", ":", "if", "no_prereq_install", "(", ")", ":", "print", "NO_PREREQ_MESSAGE", "return", "prereq_cache", "(", "'Node prereqs'", ",", "[", "'package.json'", "]", ",", "node_prereqs_installation", ")"...
installs node prerequisites .
train
false
729
def generate_task_families(task_class, n): ret = {} for i in range(n): class_name = '{}_{}'.format(task_class.task_family, i) ret[class_name] = type(class_name, (task_class,), {}) return ret
[ "def", "generate_task_families", "(", "task_class", ",", "n", ")", ":", "ret", "=", "{", "}", "for", "i", "in", "range", "(", "n", ")", ":", "class_name", "=", "'{}_{}'", ".", "format", "(", "task_class", ".", "task_family", ",", "i", ")", "ret", "["...
generate n copies of a task with different task_family names .
train
false
730
def _get_corrected_exif(image, original): if ('exif' in original.info): image_exif = image.info.get('exif', piexif.dump({})) original_exif = original.info['exif'] image_exif = _update_exif_orientation(image_exif, _get_exif_orientation(original_exif)) return image_exif
[ "def", "_get_corrected_exif", "(", "image", ",", "original", ")", ":", "if", "(", "'exif'", "in", "original", ".", "info", ")", ":", "image_exif", "=", "image", ".", "info", ".", "get", "(", "'exif'", ",", "piexif", ".", "dump", "(", "{", "}", ")", ...
if the original image contains exif data .
train
false
731
def save_to_disk(db, ids, root, opts=None, callback=None): (root, opts, length) = sanitize_args(root, opts) failures = [] for x in ids: tb = '' try: (failed, id, title) = save_book_to_disk(x, db, root, opts, length) tb = _('Requested formats not available') except: (failed, id, title) = (True, x, db.title(x, index_is_id=True)) tb = traceback.format_exc() if failed: failures.append((id, title, tb)) if callable(callback): if (not callback(int(id), title, failed, tb)): break return failures
[ "def", "save_to_disk", "(", "db", ",", "ids", ",", "root", ",", "opts", "=", "None", ",", "callback", "=", "None", ")", ":", "(", "root", ",", "opts", ",", "length", ")", "=", "sanitize_args", "(", "root", ",", "opts", ")", "failures", "=", "[", ...
save books from the database db to the path specified by root .
train
false
732
def crc_update(crc, data): if ((type(data) != array.array) or (data.itemsize != 1)): buf = array.array('B', data) else: buf = data crc = (crc ^ _MASK) for b in buf: table_index = ((crc ^ b) & 255) crc = ((CRC_TABLE[table_index] ^ (crc >> 8)) & _MASK) return (crc ^ _MASK)
[ "def", "crc_update", "(", "crc", ",", "data", ")", ":", "if", "(", "(", "type", "(", "data", ")", "!=", "array", ".", "array", ")", "or", "(", "data", ".", "itemsize", "!=", "1", ")", ")", ":", "buf", "=", "array", ".", "array", "(", "'B'", "...
update crc-32c checksum with data .
train
true
733
def asmodule(module): if isinstance(module, str): module = __import__(module, fromlist=[]) return module
[ "def", "asmodule", "(", "module", ")", ":", "if", "isinstance", "(", "module", ",", "str", ")", ":", "module", "=", "__import__", "(", "module", ",", "fromlist", "=", "[", "]", ")", "return", "module" ]
return the module references by module name .
train
false
734
def test_suffix(): transformer = hug.transform.suffix({'.js': int, '.txt': str}) class FakeRequest(object, ): path = 'hey.js' request = FakeRequest() assert (transformer('1', request) == 1) request.path = 'hey.txt' assert (transformer(2, request) == '2') request.path = 'hey.undefined' (transformer({'data': 'value'}, request) == {'data': 'value'})
[ "def", "test_suffix", "(", ")", ":", "transformer", "=", "hug", ".", "transform", ".", "suffix", "(", "{", "'.js'", ":", "int", ",", "'.txt'", ":", "str", "}", ")", "class", "FakeRequest", "(", "object", ",", ")", ":", "path", "=", "'hey.js'", "reque...
ensure that its possible to route the output type format by the suffix of the requested url .
train
false
735
def transcribe(dna): if isinstance(dna, Seq): return dna.transcribe() elif isinstance(dna, MutableSeq): return dna.toseq().transcribe() else: return dna.replace('T', 'U').replace('t', 'u')
[ "def", "transcribe", "(", "dna", ")", ":", "if", "isinstance", "(", "dna", ",", "Seq", ")", ":", "return", "dna", ".", "transcribe", "(", ")", "elif", "isinstance", "(", "dna", ",", "MutableSeq", ")", ":", "return", "dna", ".", "toseq", "(", ")", "...
transcribes a dna sequence into rna .
train
false
736
def test_prewitt_zeros(): result = filters.prewitt(np.zeros((10, 10)), np.ones((10, 10), bool)) assert_allclose(result, 0)
[ "def", "test_prewitt_zeros", "(", ")", ":", "result", "=", "filters", ".", "prewitt", "(", "np", ".", "zeros", "(", "(", "10", ",", "10", ")", ")", ",", "np", ".", "ones", "(", "(", "10", ",", "10", ")", ",", "bool", ")", ")", "assert_allclose", ...
prewitt on an array of all zeros .
train
false
737
def get_requests(name): return reduce((lambda memo, obj: (memo + get_rate(('%srequests_%s_count' % (NAME_PREFIX, obj))))), ['DELETE', 'GET', 'POST', 'PUT'], 0)
[ "def", "get_requests", "(", "name", ")", ":", "return", "reduce", "(", "(", "lambda", "memo", ",", "obj", ":", "(", "memo", "+", "get_rate", "(", "(", "'%srequests_%s_count'", "%", "(", "NAME_PREFIX", ",", "obj", ")", ")", ")", ")", ")", ",", "[", ...
return requests per second .
train
false
738
def get_latest_repository_metadata_if_it_includes_invalid_tools(trans, repository): repository_metadata = get_latest_repository_metadata(trans, repository) if (repository_metadata is not None): metadata = repository_metadata.metadata if ((metadata is not None) and ('invalid_tools' in metadata)): return repository_metadata return None
[ "def", "get_latest_repository_metadata_if_it_includes_invalid_tools", "(", "trans", ",", "repository", ")", ":", "repository_metadata", "=", "get_latest_repository_metadata", "(", "trans", ",", "repository", ")", "if", "(", "repository_metadata", "is", "not", "None", ")",...
return the latest repository_metadata record for the received repository that contains invalid tools if one exists .
train
false
740
def decode_int(v): if ((len(v) > 0) and ((v[0] == '\x00') or (v[0] == 0))): raise Exception('No leading zero bytes allowed for integers') return big_endian_to_int(v)
[ "def", "decode_int", "(", "v", ")", ":", "if", "(", "(", "len", "(", "v", ")", ">", "0", ")", "and", "(", "(", "v", "[", "0", "]", "==", "'\\x00'", ")", "or", "(", "v", "[", "0", "]", "==", "0", ")", ")", ")", ":", "raise", "Exception", ...
decode c{int} .
train
true
742
def detect_encoding(readline): bom_found = False encoding = None default = 'utf-8' def read_or_stop(): try: return readline() except StopIteration: return bytes() def find_cookie(line): try: line_string = line.decode('ascii') except UnicodeDecodeError: return None match = cookie_re.match(line_string) if (not match): return None encoding = _get_normal_name(match.group(1)) try: codec = lookup(encoding) except LookupError: raise SyntaxError(('unknown encoding: ' + encoding)) if bom_found: if (codec.name != 'utf-8'): raise SyntaxError('encoding problem: utf-8') encoding += '-sig' return encoding first = read_or_stop() if first.startswith(BOM_UTF8): bom_found = True first = first[3:] default = 'utf-8-sig' if (not first): return (default, []) encoding = find_cookie(first) if encoding: return (encoding, [first]) second = read_or_stop() if (not second): return (default, [first]) encoding = find_cookie(second) if encoding: return (encoding, [first, second]) return (default, [first, second])
[ "def", "detect_encoding", "(", "readline", ")", ":", "bom_found", "=", "False", "encoding", "=", "None", "default", "=", "'utf-8'", "def", "read_or_stop", "(", ")", ":", "try", ":", "return", "readline", "(", ")", "except", "StopIteration", ":", "return", ...
return file encoding .
train
true
744
def inverse_dicts(*dicts): inverse = {} for dictionary in dicts: for (key, value) in dictionary.items(): if isinstance(value, Iterable): for item in value: add_pair_to_dict(item, key, inverse) else: add_pair_to_dict(value, key, inverse) return inverse
[ "def", "inverse_dicts", "(", "*", "dicts", ")", ":", "inverse", "=", "{", "}", "for", "dictionary", "in", "dicts", ":", "for", "(", "key", ",", "value", ")", "in", "dictionary", ".", "items", "(", ")", ":", "if", "isinstance", "(", "value", ",", "I...
inverts the dicts .
train
false
745
def is_datetimelike(data): try: maybe_to_datetimelike(data) return True except Exception: pass return False
[ "def", "is_datetimelike", "(", "data", ")", ":", "try", ":", "maybe_to_datetimelike", "(", "data", ")", "return", "True", "except", "Exception", ":", "pass", "return", "False" ]
return a boolean if we can be successfully converted to a datetimelike .
train
false
746
def delete_usage_plan(plan_id, region=None, key=None, keyid=None, profile=None): try: existing = describe_usage_plans(plan_id=plan_id, region=region, key=key, keyid=keyid, profile=profile) if ('error' in existing): return {'error': existing['error']} if (('plans' in existing) and existing['plans']): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) res = conn.delete_usage_plan(usagePlanId=plan_id) return {'deleted': True, 'usagePlanId': plan_id} except ClientError as e: return {'error': salt.utils.boto3.get_error(e)}
[ "def", "delete_usage_plan", "(", "plan_id", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "try", ":", "existing", "=", "describe_usage_plans", "(", "plan_id", "=", "plan_id", ",", ...
deletes usage plan identified by plan_id .
train
false
747
def add_qiime_labels(mapping_f, fasta_dir, filename_column, output_dir='.', count_start=0): (headers, mapping_data, run_description, errors, warnings) = process_id_map(mapping_f, has_barcodes=False, disable_primer_check=True, added_demultiplex_field=None, variable_len_barcodes=False) fasta_name_to_sample_id = check_mapping_data(mapping_data, headers, filename_column) fasta_files = get_fasta_fps(fasta_dir, fasta_name_to_sample_id.keys()) write_combined_fasta(fasta_name_to_sample_id, fasta_files, output_dir, counter=count_start)
[ "def", "add_qiime_labels", "(", "mapping_f", ",", "fasta_dir", ",", "filename_column", ",", "output_dir", "=", "'.'", ",", "count_start", "=", "0", ")", ":", "(", "headers", ",", "mapping_data", ",", "run_description", ",", "errors", ",", "warnings", ")", "=...
main function for combining fasta files .
train
false
748
def qflags_key(base, value, add_base=False, klass=None): if (klass is None): klass = value.__class__ if (klass == int): raise TypeError("Can't guess enum class of an int!") bits = [] names = [] mask = 1 value = int(value) while (mask <= value): if (value & mask): bits.append(mask) mask <<= 1 for bit in bits: names.append(qenum_key(base, klass(bit), add_base)) return '|'.join(names)
[ "def", "qflags_key", "(", "base", ",", "value", ",", "add_base", "=", "False", ",", "klass", "=", "None", ")", ":", "if", "(", "klass", "is", "None", ")", ":", "klass", "=", "value", ".", "__class__", "if", "(", "klass", "==", "int", ")", ":", "r...
convert a qt qflags value to its keys as string .
train
false
750
def split_title_year(title): if (not title): return if (not re.search(u'\\d{4}', title)): return (title, None) match = re.search(u'(.*?)\\(?(\\d{4})?\\)?$', title) title = match.group(1).strip() if match.group(2): year = int(match.group(2)) else: year = None return (title, year)
[ "def", "split_title_year", "(", "title", ")", ":", "if", "(", "not", "title", ")", ":", "return", "if", "(", "not", "re", ".", "search", "(", "u'\\\\d{4}'", ",", "title", ")", ")", ":", "return", "(", "title", ",", "None", ")", "match", "=", "re", ...
splits title containing a year into a title .
train
false
751
def InstallManagementConsole(namespace, users={'admin': 'admin'}, port=503): from twisted.internet import reactor def build_protocol(): p = insults.ServerProtocol(manhole.ColoredManhole, namespace) return p r = manhole_ssh.TerminalRealm() r.chainedProtocolFactory = build_protocol c = checkers.InMemoryUsernamePasswordDatabaseDontUse(**users) p = portal.Portal(r, [c]) factory = manhole_ssh.ConchFactory(p) reactor.listenTCP(port, factory)
[ "def", "InstallManagementConsole", "(", "namespace", ",", "users", "=", "{", "'admin'", ":", "'admin'", "}", ",", "port", "=", "503", ")", ":", "from", "twisted", ".", "internet", "import", "reactor", "def", "build_protocol", "(", ")", ":", "p", "=", "in...
helper method to start an ssh management console for the modbus server .
train
false
752
def ip_addrs6(interface=None, include_loopback=False): return salt.utils.network.ip_addrs6(interface=interface, include_loopback=include_loopback)
[ "def", "ip_addrs6", "(", "interface", "=", "None", ",", "include_loopback", "=", "False", ")", ":", "return", "salt", ".", "utils", ".", "network", ".", "ip_addrs6", "(", "interface", "=", "interface", ",", "include_loopback", "=", "include_loopback", ")" ]
returns a list of ipv6 addresses assigned to the host .
train
false
753
def _logged_get(log, *args, **kwargs): req_kwargs = kwargs send_kwargs = {} for arg in ('stream', 'verify', 'proxies', 'cert', 'timeout'): if (arg in kwargs): send_kwargs[arg] = req_kwargs.pop(arg) if ('message' in kwargs): message = kwargs.pop('message') else: message = 'getting URL' req = requests.Request('GET', *args, **req_kwargs) with requests.Session() as s: s.headers = {'User-Agent': 'beets'} prepped = s.prepare_request(req) log.debug('{}: {}', message, prepped.url) return s.send(prepped, **send_kwargs)
[ "def", "_logged_get", "(", "log", ",", "*", "args", ",", "**", "kwargs", ")", ":", "req_kwargs", "=", "kwargs", "send_kwargs", "=", "{", "}", "for", "arg", "in", "(", "'stream'", ",", "'verify'", ",", "'proxies'", ",", "'cert'", ",", "'timeout'", ")", ...
like requests .
train
false
756
def _in_gce_environment(): if (SETTINGS.env_name is not None): return (SETTINGS.env_name == 'GCE_PRODUCTION') if ((NO_GCE_CHECK != 'True') and _detect_gce_environment()): SETTINGS.env_name = 'GCE_PRODUCTION' return True return False
[ "def", "_in_gce_environment", "(", ")", ":", "if", "(", "SETTINGS", ".", "env_name", "is", "not", "None", ")", ":", "return", "(", "SETTINGS", ".", "env_name", "==", "'GCE_PRODUCTION'", ")", "if", "(", "(", "NO_GCE_CHECK", "!=", "'True'", ")", "and", "_d...
detect if the code is running in the compute engine environment .
train
true
757
def rotatePoints(points, prefix, xmlElement): rotateMatrixTetragrid = matrix.getRotateMatrixTetragrid(prefix, xmlElement) if (rotateMatrixTetragrid == None): print 'Warning, rotateMatrixTetragrid was None in rotate so nothing will be done for:' print xmlElement return for point in points: matrix.transformVector3ByMatrix(rotateMatrixTetragrid, point)
[ "def", "rotatePoints", "(", "points", ",", "prefix", ",", "xmlElement", ")", ":", "rotateMatrixTetragrid", "=", "matrix", ".", "getRotateMatrixTetragrid", "(", "prefix", ",", "xmlElement", ")", "if", "(", "rotateMatrixTetragrid", "==", "None", ")", ":", "print",...
rotate the points .
train
false
758
def RotL_64(x, N): return (np.left_shift(x, (N & 63), dtype=np.uint64) | np.right_shift(x, ((64 - N) & 63), dtype=np.uint64))
[ "def", "RotL_64", "(", "x", ",", "N", ")", ":", "return", "(", "np", ".", "left_shift", "(", "x", ",", "(", "N", "&", "63", ")", ",", "dtype", "=", "np", ".", "uint64", ")", "|", "np", ".", "right_shift", "(", "x", ",", "(", "(", "64", "-",...
return x rotated left by n .
train
false
759
def align_left(text, length, left_edge='|', right_edge='|', text_length=None, left_padding=2): if (text_length is None): text_length = get_text_length(text) computed_length = (((text_length + left_padding) + get_text_length(left_edge)) + get_text_length(right_edge)) if ((length - computed_length) >= 0): padding = left_padding else: padding = 0 output = [] length_so_far = 0 output.append(left_edge) length_so_far += len(left_edge) output.append((' ' * padding)) length_so_far += padding output.append(text) length_so_far += text_length output.append((' ' * ((length - length_so_far) - len(right_edge)))) output.append(right_edge) return ''.join(output)
[ "def", "align_left", "(", "text", ",", "length", ",", "left_edge", "=", "'|'", ",", "right_edge", "=", "'|'", ",", "text_length", "=", "None", ",", "left_padding", "=", "2", ")", ":", "if", "(", "text_length", "is", "None", ")", ":", "text_length", "="...
left align text .
train
false
760
@docfiller def generic_filter1d(input, function, filter_size, axis=(-1), output=None, mode='reflect', cval=0.0, origin=0, extra_arguments=(), extra_keywords=None): if (extra_keywords is None): extra_keywords = {} input = numpy.asarray(input) if numpy.iscomplexobj(input): raise TypeError('Complex type not supported') (output, return_value) = _ni_support._get_output(output, input) if (filter_size < 1): raise RuntimeError('invalid filter size') axis = _ni_support._check_axis(axis, input.ndim) if ((((filter_size // 2) + origin) < 0) or (((filter_size // 2) + origin) >= filter_size)): raise ValueError('invalid origin') mode = _ni_support._extend_mode_to_code(mode) _nd_image.generic_filter1d(input, function, filter_size, axis, output, mode, cval, origin, extra_arguments, extra_keywords) return return_value
[ "@", "docfiller", "def", "generic_filter1d", "(", "input", ",", "function", ",", "filter_size", ",", "axis", "=", "(", "-", "1", ")", ",", "output", "=", "None", ",", "mode", "=", "'reflect'", ",", "cval", "=", "0.0", ",", "origin", "=", "0", ",", ...
calculate a one-dimensional filter along the given axis .
train
false
761
def linkcheck(): os.system('sphinx-build -b linkcheck -d build/doctrees . build/linkcheck')
[ "def", "linkcheck", "(", ")", ":", "os", ".", "system", "(", "'sphinx-build -b linkcheck -d build/doctrees . build/linkcheck'", ")" ]
check if all links are corect .
train
false
763
def format_slice(key_val, dim): if isinstance(key_val, slice): key_val = slice(to_int(key_val.start), to_int(key_val.stop), to_int(key_val.step)) return key_val else: key_val = to_int(key_val) key_val = wrap_neg_index(key_val, dim) if (0 <= key_val < dim): return slice(key_val, (key_val + 1), 1) else: raise IndexError('Index/slice out of bounds.')
[ "def", "format_slice", "(", "key_val", ",", "dim", ")", ":", "if", "isinstance", "(", "key_val", ",", "slice", ")", ":", "key_val", "=", "slice", "(", "to_int", "(", "key_val", ".", "start", ")", ",", "to_int", "(", "key_val", ".", "stop", ")", ",", ...
converts part of a key into a slice with a start and step .
train
false
764
def delocalize(string): conv = localeconv() ts = conv['thousands_sep'] if ts: string = string.replace(ts, '') dd = conv['decimal_point'] if dd: string = string.replace(dd, '.') return string
[ "def", "delocalize", "(", "string", ")", ":", "conv", "=", "localeconv", "(", ")", "ts", "=", "conv", "[", "'thousands_sep'", "]", "if", "ts", ":", "string", "=", "string", ".", "replace", "(", "ts", ",", "''", ")", "dd", "=", "conv", "[", "'decima...
parses a string as a normalized number according to the locale settings .
train
false
765
def local_git_clone(repo_url): with lcd(LOGDIR): local('if [ -d letsencrypt ]; then rm -rf letsencrypt; fi') local(('git clone %s letsencrypt' % repo_url)) local('tar czf le.tar.gz letsencrypt')
[ "def", "local_git_clone", "(", "repo_url", ")", ":", "with", "lcd", "(", "LOGDIR", ")", ":", "local", "(", "'if [ -d letsencrypt ]; then rm -rf letsencrypt; fi'", ")", "local", "(", "(", "'git clone %s letsencrypt'", "%", "repo_url", ")", ")", "local", "(", "'tar ...
clones master of repo_url .
train
false
766
def oo_collect(data, attribute=None, filters=None): if (not isinstance(data, list)): raise errors.AnsibleFilterError('|failed expects to filter on a List') if (not attribute): raise errors.AnsibleFilterError('|failed expects attribute to be set') if (filters is not None): if (not isinstance(filters, dict)): raise errors.AnsibleFilterError('|failed expects filter to be a dict') retval = [get_attr(d, attribute) for d in data if all([(d.get(key, None) == filters[key]) for key in filters])] else: retval = [get_attr(d, attribute) for d in data] retval = [val for val in retval if (val is not None)] return retval
[ "def", "oo_collect", "(", "data", ",", "attribute", "=", "None", ",", "filters", "=", "None", ")", ":", "if", "(", "not", "isinstance", "(", "data", ",", "list", ")", ")", ":", "raise", "errors", ".", "AnsibleFilterError", "(", "'|failed expects to filter ...
this takes a list of dict and collects all attributes specified into a list .
train
false
767
def prefixed_userid(request): authn_type = getattr(request, 'authn_type', None) if (authn_type is not None): return ((authn_type + ':') + request.selected_userid)
[ "def", "prefixed_userid", "(", "request", ")", ":", "authn_type", "=", "getattr", "(", "request", ",", "'authn_type'", ",", "None", ")", "if", "(", "authn_type", "is", "not", "None", ")", ":", "return", "(", "(", "authn_type", "+", "':'", ")", "+", "re...
in kinto users ids are prefixed with the policy name that is contained in pyramid multiauth .
train
false
769
def function_simple(a, b, c): return (a, b, c)
[ "def", "function_simple", "(", "a", ",", "b", ",", "c", ")", ":", "return", "(", "a", ",", "b", ",", "c", ")" ]
a function which accepts several arguments .
train
false
770
def set_vm_state_and_notify(context, instance_uuid, service, method, updates, ex, request_spec): LOG.warning(_LW('Failed to %(service)s_%(method)s: %(ex)s'), {'service': service, 'method': method, 'ex': ex}) vm_state = updates['vm_state'] properties = request_spec.get('instance_properties', {}) notifier = rpc.get_notifier(service) state = vm_state.upper() LOG.warning(_LW('Setting instance to %s state.'), state, instance_uuid=instance_uuid) instance = objects.Instance(context=context, uuid=instance_uuid, **updates) instance.obj_reset_changes(['uuid']) instance.save() compute_utils.add_instance_fault_from_exc(context, instance, ex, sys.exc_info()) payload = dict(request_spec=request_spec, instance_properties=properties, instance_id=instance_uuid, state=vm_state, method=method, reason=ex) event_type = ('%s.%s' % (service, method)) notifier.error(context, event_type, payload)
[ "def", "set_vm_state_and_notify", "(", "context", ",", "instance_uuid", ",", "service", ",", "method", ",", "updates", ",", "ex", ",", "request_spec", ")", ":", "LOG", ".", "warning", "(", "_LW", "(", "'Failed to %(service)s_%(method)s: %(ex)s'", ")", ",", "{", ...
changes vm state and notifies .
train
false
771
def register_deep_copy_op_c_code(typ, code, version=()): DeepCopyOp.c_code_and_version[typ] = (code, version)
[ "def", "register_deep_copy_op_c_code", "(", "typ", ",", "code", ",", "version", "=", "(", ")", ")", ":", "DeepCopyOp", ".", "c_code_and_version", "[", "typ", "]", "=", "(", "code", ",", "version", ")" ]
tell deepcopyop how to generate c code for a theano type .
train
false
772
def _length_hint(obj): try: return len(obj) except (AttributeError, TypeError): try: get_hint = type(obj).__length_hint__ except AttributeError: return None try: hint = get_hint(obj) except TypeError: return None if ((hint is NotImplemented) or (not isinstance(hint, int_types)) or (hint < 0)): return None return hint
[ "def", "_length_hint", "(", "obj", ")", ":", "try", ":", "return", "len", "(", "obj", ")", "except", "(", "AttributeError", ",", "TypeError", ")", ":", "try", ":", "get_hint", "=", "type", "(", "obj", ")", ".", "__length_hint__", "except", "AttributeErro...
returns the length hint of an object .
train
true
773
def policy_version_exists(policy_name, version_id, region=None, key=None, keyid=None, profile=None): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) policy_arn = _get_policy_arn(policy_name, region, key, keyid, profile) try: conn.get_policy_version(policy_arn, version_id) return True except boto.exception.BotoServerError: return False
[ "def", "policy_version_exists", "(", "policy_name", ",", "version_id", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "...
given a policy name and version id .
train
true
774
def test_cnot_commutators(): assert (Commutator(CNOT(0, 1), Z(0)).doit() == 0) assert (Commutator(CNOT(0, 1), T(0)).doit() == 0) assert (Commutator(CNOT(0, 1), S(0)).doit() == 0) assert (Commutator(CNOT(0, 1), X(1)).doit() == 0) assert (Commutator(CNOT(0, 1), CNOT(0, 1)).doit() == 0) assert (Commutator(CNOT(0, 1), CNOT(0, 2)).doit() == 0) assert (Commutator(CNOT(0, 2), CNOT(0, 1)).doit() == 0) assert (Commutator(CNOT(1, 2), CNOT(1, 0)).doit() == 0)
[ "def", "test_cnot_commutators", "(", ")", ":", "assert", "(", "Commutator", "(", "CNOT", "(", "0", ",", "1", ")", ",", "Z", "(", "0", ")", ")", ".", "doit", "(", ")", "==", "0", ")", "assert", "(", "Commutator", "(", "CNOT", "(", "0", ",", "1",...
test commutators of involving cnot gates .
train
false
775
def check_header_match_180_or_later(header1, header2): header1 = header1.split(':') header2 = header2.split(':') for (e1, e2) in zip(header1, header2): if (e1.split(' ')[0] != e2.split(' ')[0]): return False return True
[ "def", "check_header_match_180_or_later", "(", "header1", ",", "header2", ")", ":", "header1", "=", "header1", ".", "split", "(", "':'", ")", "header2", "=", "header2", ".", "split", "(", "':'", ")", "for", "(", "e1", ",", "e2", ")", "in", "zip", "(", ...
confirm headers are compatible in casava 1 .
train
false
777
@register.tag def get_unread_message_count_for(parser, token): try: (tag_name, arg) = token.contents.split(None, 1) except ValueError: raise template.TemplateSyntaxError(('%s tag requires arguments' % token.contents.split()[0])) m = re.search('(.*?) as (\\w+)', arg) if (not m): raise template.TemplateSyntaxError(('%s tag had invalid arguments' % tag_name)) (user, var_name) = m.groups() return MessageCount(user, var_name)
[ "@", "register", ".", "tag", "def", "get_unread_message_count_for", "(", "parser", ",", "token", ")", ":", "try", ":", "(", "tag_name", ",", "arg", ")", "=", "token", ".", "contents", ".", "split", "(", "None", ",", "1", ")", "except", "ValueError", ":...
returns the unread message count for a user .
train
true
778
def grubbs(timeseries): series = scipy.array([x[1] for x in timeseries]) stdDev = scipy.std(series) mean = np.mean(series) tail_average = tail_avg(timeseries) z_score = ((tail_average - mean) / stdDev) len_series = len(series) threshold = scipy.stats.t.isf((0.05 / (2 * len_series)), (len_series - 2)) threshold_squared = (threshold * threshold) grubbs_score = (((len_series - 1) / np.sqrt(len_series)) * np.sqrt((threshold_squared / ((len_series - 2) + threshold_squared)))) return (z_score > grubbs_score)
[ "def", "grubbs", "(", "timeseries", ")", ":", "series", "=", "scipy", ".", "array", "(", "[", "x", "[", "1", "]", "for", "x", "in", "timeseries", "]", ")", "stdDev", "=", "scipy", ".", "std", "(", "series", ")", "mean", "=", "np", ".", "mean", ...
a timeseries is anomalous if the z score is greater than the grubbs score .
train
false
779
def headers_url_generator(resp, fuzzable_req): resp_headers = resp.get_headers() for (parser, header_names) in URL_HEADERS.iteritems(): for header_name in header_names: (header_value, _) = resp_headers.iget(header_name, None) if (header_value is not None): header_value = smart_unicode(header_value, encoding=resp.charset) for ref in parser(resp, header_name, header_value): (yield (ref, fuzzable_req, resp, False))
[ "def", "headers_url_generator", "(", "resp", ",", "fuzzable_req", ")", ":", "resp_headers", "=", "resp", ".", "get_headers", "(", ")", "for", "(", "parser", ",", "header_names", ")", "in", "URL_HEADERS", ".", "iteritems", "(", ")", ":", "for", "header_name",...
yields tuples containing: * newly found url * the fuzzablerequest instance passed as parameter * the httpresponse generated by the fuzzablerequest * boolean indicating if we trust this reference or not the newly found urls are extracted from the http response headers such as "location" .
train
false
780
def ttest_ind_from_stats(mean1, std1, nobs1, mean2, std2, nobs2, equal_var=True): if equal_var: (df, denom) = _equal_var_ttest_denom((std1 ** 2), nobs1, (std2 ** 2), nobs2) else: (df, denom) = _unequal_var_ttest_denom((std1 ** 2), nobs1, (std2 ** 2), nobs2) res = _ttest_ind_from_stats(mean1, mean2, denom, df) return Ttest_indResult(*res)
[ "def", "ttest_ind_from_stats", "(", "mean1", ",", "std1", ",", "nobs1", ",", "mean2", ",", "std2", ",", "nobs2", ",", "equal_var", "=", "True", ")", ":", "if", "equal_var", ":", "(", "df", ",", "denom", ")", "=", "_equal_var_ttest_denom", "(", "(", "st...
t-test for means of two independent samples from descriptive statistics .
train
false
781
def dump_js_escaped_json(obj, cls=EdxJSONEncoder): json_string = json.dumps(obj, ensure_ascii=True, cls=cls) json_string = _escape_json_for_js(json_string) return json_string
[ "def", "dump_js_escaped_json", "(", "obj", ",", "cls", "=", "EdxJSONEncoder", ")", ":", "json_string", "=", "json", ".", "dumps", "(", "obj", ",", "ensure_ascii", "=", "True", ",", "cls", "=", "cls", ")", "json_string", "=", "_escape_json_for_js", "(", "js...
json dumps and escapes objects that are safe to be embedded in javascript .
train
false
782
@receiver(post_migrate) def sync_create_groups(sender, **kwargs): if (sender.label == u'accounts'): create_groups(False)
[ "@", "receiver", "(", "post_migrate", ")", "def", "sync_create_groups", "(", "sender", ",", "**", "kwargs", ")", ":", "if", "(", "sender", ".", "label", "==", "u'accounts'", ")", ":", "create_groups", "(", "False", ")" ]
create groups on syncdb .
train
false
783
def _make_logpt(global_RVs, local_RVs, observed_RVs, potentials): factors = ((([(c * v.logpt) for (v, c) in observed_RVs.items()] + [(c * v.logpt) for (v, c) in global_RVs.items()]) + [(c * v.logpt) for (v, (_, c)) in local_RVs.items()]) + potentials) logpt = tt.add(*map(tt.sum, factors)) return logpt
[ "def", "_make_logpt", "(", "global_RVs", ",", "local_RVs", ",", "observed_RVs", ",", "potentials", ")", ":", "factors", "=", "(", "(", "(", "[", "(", "c", "*", "v", ".", "logpt", ")", "for", "(", "v", ",", "c", ")", "in", "observed_RVs", ".", "item...
return expression of log probability .
train
false
784
def generate_prototypes_from_src(src_text): src_text = collapse_braces(strip(src_text)) prototype_pattern = re.compile(prototype, (re.M | re.S)) function_pattern = re.compile(function, (re.M | re.S)) prototypes = prototype_pattern.findall(src_text) functions = function_pattern.findall(src_text) declared_prototypes = list(map(sanitize_prototype, prototypes)) all_prototypes = list(map(sanitize_prototype, functions)) for declared_prototype in declared_prototypes: if (declared_prototype in all_prototypes): all_prototypes.remove(declared_prototype) all_prototypes = sorted(set(all_prototypes), key=all_prototypes.index) return all_prototypes
[ "def", "generate_prototypes_from_src", "(", "src_text", ")", ":", "src_text", "=", "collapse_braces", "(", "strip", "(", "src_text", ")", ")", "prototype_pattern", "=", "re", ".", "compile", "(", "prototype", ",", "(", "re", ".", "M", "|", "re", ".", "S", ...
generate prototypes for all functions of a given source code .
train
false
786
def intrinsic(func): name = getattr(func, '__name__', str(func)) llc = _Intrinsic(name, func) llc._register() return llc
[ "def", "intrinsic", "(", "func", ")", ":", "name", "=", "getattr", "(", "func", ",", "'__name__'", ",", "str", "(", "func", ")", ")", "llc", "=", "_Intrinsic", "(", "name", ",", "func", ")", "llc", ".", "_register", "(", ")", "return", "llc" ]
a decorator marking the decorated function as typing and implementing *func* in nopython mode using the llvmlite irbuilder api .
train
false
787
def __clean_tmp(sfn): if sfn.startswith(os.path.join(tempfile.gettempdir(), salt.utils.files.TEMPFILE_PREFIX)): all_roots = itertools.chain.from_iterable(six.itervalues(__opts__['file_roots'])) in_roots = any((sfn.startswith(root) for root in all_roots)) if (os.path.exists(sfn) and (not in_roots)): os.remove(sfn)
[ "def", "__clean_tmp", "(", "sfn", ")", ":", "if", "sfn", ".", "startswith", "(", "os", ".", "path", ".", "join", "(", "tempfile", ".", "gettempdir", "(", ")", ",", "salt", ".", "utils", ".", "files", ".", "TEMPFILE_PREFIX", ")", ")", ":", "all_roots"...
clean out a template temp file .
train
true
788
def from_tree(expr, namespace=None): if isinstance(expr, dict): (op, args) = (expr[u'op'], expr[u'args']) if (op == u'slice'): return expr_utils._slice(*[from_tree(arg, namespace) for arg in args]) if hasattr(blaze.expr, op): cls = getattr(blaze.expr, op) else: cls = expression_from_name(op) if (cls is Symbol): cls = symbol children = [from_tree(arg, namespace) for arg in args] return cls(*children) elif isinstance(expr, (list, tuple)): return tuple((from_tree(arg, namespace) for arg in expr)) if (namespace and (expr in namespace)): return namespace[expr] else: return expr
[ "def", "from_tree", "(", "expr", ",", "namespace", "=", "None", ")", ":", "if", "isinstance", "(", "expr", ",", "dict", ")", ":", "(", "op", ",", "args", ")", "=", "(", "expr", "[", "u'op'", "]", ",", "expr", "[", "u'args'", "]", ")", "if", "("...
convert core data structures to blaze expression core data structure representations created by to_tree are converted back into blaze expressions .
train
false
790
@datastore_rpc._positional(1) def NonTransactional(_func=None, allow_existing=True): if (_func is not None): return NonTransactional()(_func) def outer_wrapper(func): def inner_wrapper(*args, **kwds): if (not IsInTransaction()): return func(*args, **kwds) if (not allow_existing): raise datastore_errors.BadRequestError('Function cannot be called from within a transaction.') txn_connection = _GetConnection() _SetConnection(_thread_local.old_connection) try: return func(*args, **kwds) finally: _SetConnection(txn_connection) return inner_wrapper return outer_wrapper
[ "@", "datastore_rpc", ".", "_positional", "(", "1", ")", "def", "NonTransactional", "(", "_func", "=", "None", ",", "allow_existing", "=", "True", ")", ":", "if", "(", "_func", "is", "not", "None", ")", ":", "return", "NonTransactional", "(", ")", "(", ...
a decorator that insures a function is run outside a transaction .
train
false
793
def getStreamLabel(sampleRate, channels, blockSize): return '{}_{}_{}'.format(sampleRate, channels, blockSize)
[ "def", "getStreamLabel", "(", "sampleRate", ",", "channels", ",", "blockSize", ")", ":", "return", "'{}_{}_{}'", ".", "format", "(", "sampleRate", ",", "channels", ",", "blockSize", ")" ]
returns the string repr of the stream label .
train
false
796
def mkdtemp(suffix='', prefix=template, dir=None): if (dir is None): dir = gettempdir() names = _get_candidate_names() for seq in xrange(TMP_MAX): name = names.next() file = _os.path.join(dir, ((prefix + name) + suffix)) try: _os.mkdir(file, 448) return file except OSError as e: if (e.errno == _errno.EEXIST): continue raise raise IOError, (_errno.EEXIST, 'No usable temporary directory name found')
[ "def", "mkdtemp", "(", "suffix", "=", "''", ",", "prefix", "=", "template", ",", "dir", "=", "None", ")", ":", "if", "(", "dir", "is", "None", ")", ":", "dir", "=", "gettempdir", "(", ")", "names", "=", "_get_candidate_names", "(", ")", "for", "seq...
user-callable function to create and return a unique temporary directory .
train
false
797
def test_es2(): try: from vispy.gloo.gl import es2 except Exception: import ctypes ctypes.TEST_DLL = _DummyObject() from vispy.gloo.gl import es2 _test_function_names(es2) _test_constant_names(es2)
[ "def", "test_es2", "(", ")", ":", "try", ":", "from", "vispy", ".", "gloo", ".", "gl", "import", "es2", "except", "Exception", ":", "import", "ctypes", "ctypes", ".", "TEST_DLL", "=", "_DummyObject", "(", ")", "from", "vispy", ".", "gloo", ".", "gl", ...
es2 backend should have all es 2 .
train
false
798
def atoms(cls, instance_or_dict): field_getter = serializable_getter = instance_or_dict.get try: field_getter = instance_or_dict._data.get except AttributeError: pass sequences = ((cls._field_list, field_getter), (cls._serializables.items(), serializable_getter)) for (sequence, get) in sequences: for (field_name, field) in sequence: (yield (field_name, field, get(field_name, Undefined)))
[ "def", "atoms", "(", "cls", ",", "instance_or_dict", ")", ":", "field_getter", "=", "serializable_getter", "=", "instance_or_dict", ".", "get", "try", ":", "field_getter", "=", "instance_or_dict", ".", "_data", ".", "get", "except", "AttributeError", ":", "pass"...
iterator for the atomic components of a model definition and relevant data that creates a 3-tuple of the fields name .
train
false
799
def CDLEVENINGSTAR(barDs, count, penetration=(-4e+37)): return call_talib_with_ohlc(barDs, count, talib.CDLEVENINGSTAR, penetration)
[ "def", "CDLEVENINGSTAR", "(", "barDs", ",", "count", ",", "penetration", "=", "(", "-", "4e+37", ")", ")", ":", "return", "call_talib_with_ohlc", "(", "barDs", ",", "count", ",", "talib", ".", "CDLEVENINGSTAR", ",", "penetration", ")" ]
evening star .
train
false