rem
stringlengths
0
322k
add
stringlengths
0
2.05M
context
stringlengths
8
228k
if self.pointer == i: clr.append('selected')
y = i + 1 clr = deque(base_clr)
def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list()
descr = obj.get_description() self.addstr(y, 0, descr, self.wid) self.color_at(y, 0, self.wid, clr)
if self.pointer == i: clr.append('selected')
def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list()
else: if self.hei > 1: self.addstr(1, 0, "No task in the queue.") self.color_at(1, 0, self.wid, base_clr, 'error')
descr = obj.get_description() self.addstr(y, 0, descr, self.wid) self.color_at(y, 0, self.wid, clr)
def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list()
self.color_reset()
else: if self.hei > 1: self.addstr(1, 0, "No task in the queue.") self.color_at(1, 0, self.wid, base_clr, 'error') self.color_reset()
def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list()
alias(e=quit)
alias(e=edit)
def alias(**kw): for key, value in kw.items(): by_name[key] = value
@depends_on(app_editor, Applications.app_self)
def app_editor(self, c): try: default_editor = os.environ['EDITOR'] except KeyError: pass else: parts = default_editor.split() exe_name = os.path.basename(parts[0]) if exe_name in get_executables(): return tuple(parts) + tuple(c)
sys.stdout.write("\033]2;" + hostname + \ self.fm.env.pwd.path + "\007")
sys.stdout.write("\033]2;" + hostname + cwd + "\007")
def draw(self): """Erase the window, then draw all objects in the container""" self.win.touchwin() DisplayableContainer.draw(self) if self.settings.update_title: hostname = str(socket.gethostname()) try: cwd = self.fm.env.pwd.path except: cwd = ' - ranger' sys.stdout.write("\033]2;" + hostname + \ self.fm.env.pwd.path ...
import os
def load_content_if_outdated(self, *a, **k): """ Load the contents of the directory if it's outdated or not done yet """
cached_mtime = self.stat.st_mtime
cached_mtime = self.load_content_mtime
def load_content_if_outdated(self, *a, **k): """ Load the contents of the directory if it's outdated or not done yet """
return True
return False
def has_preview(self): if not self.fm.settings.preview_files: return False if self.is_socket or self.is_fifo or self.is_device: return False if not self.accessible: return False if self.image or self.container: return True if PREVIEW_WHITELIST.search(self.basename): return True if PREVIEW_BLACKLIST.search(self.basename...
if self.fm.settings.preview_script: return True
def has_preview(self): if self.fm.settings.preview_script: return True if not self.fm.settings.preview_files: return False if self.is_socket or self.is_fifo or self.is_device: return False if not self.accessible: return False if self.image or self.container: return False if PREVIEW_WHITELIST.search(self.basename): retu...
self.notify("%s complete" % path)
def on_after(signal): self.notify("%s complete" % path) exit = signal.process.poll() content = signal.loader.stdout_buffer content += signal.process.stdout.read() if exit == 0: data[(width, height)] = content elif exit == 3: data[(-1, height)] = content elif exit == 4: data[(width, -1)] = content elif exit == 5: data[(...
result = self.settings.collapse_preview and self.preview and \ not self.columns[-1].has_preview() and self.stretch_ratios if result: return True if self.columns[-1].target: target = self.columns[-1].target
if not self.settings.collapse_preview or not self.preview \ or not self.stretch_ratios: return False result = not self.columns[-1].has_preview() target = self.columns[-1].target if not result and target and target.is_file:
def _collapse(self): # Should the last column be cut off? (Because there is no preview) result = self.settings.collapse_preview and self.preview and \ not self.columns[-1].has_preview() and self.stretch_ratios if result: return True if self.columns[-1].target: target = self.columns[-1].target try: result = not self.fm....
for child in reversed(rows): if not child.has_preview(): right_end = child.x - 1 else: break if right_end < left_start: right_end = self.wid - 1
if not self.pager.visible: for child in reversed(rows): if not child.has_preview(): right_end = child.x - 1 else: break if right_end < left_start: right_end = self.wid - 1
def _draw_borders(self): win = self.win self.color('in_browser', 'border')
max(1, self.wid - left))
max(1, self.wid - left - pad))
def resize(self, y, x, hei, wid): """Resize all the columns according to the given ratio""" DisplayableContainer.resize(self, y, x, hei, wid) borders = self.settings.draw_borders pad = 1 if borders else 0 left = pad
if self.end_hook(process):
if self.end_hook:
def generate(self): self.process = process = Popen(self.args, stdout=open(os.devnull, 'w'), stderr=PIPE) if self.begin_hook: self.begin_hook(process) while process.poll() is None: rd, _, __ = select.select( [process.stderr], [], [], 0.05) if rd: self.notify(process.stderr.readline(), bad=True) sleep(0.02) yield if self...
and cls.allow_abbrev]
and cls.allow_abbrev \ or cmd == name]
def get_command(name, abbrev=True): if abbrev: lst = [cls for cmd, cls in by_name.items() \ if cmd.startswith(name) \ and cls.allow_abbrev] if len(lst) == 0: raise KeyError if len(lst) == 1: return lst[0] raise ValueError("Ambiguous command") else: try: return by_name[name] except KeyError: return None
i = absolute
if absolute < 0: i = absolute + len(self.lines) else: i = absolute
def move(self, relative=0, absolute=None, pages=False): i = self.scroll_begin if isinstance(absolute, int): i = absolute
if chr(key) in digits and ANYKEY not in tree:
if key in digitlist and ANYKEY not in tree:
def _do_eval_quantifier(self, key): if self.eval_command: tree = self.tree_pointer else: tree = self.dir_tree_pointer if chr(key) in digits and ANYKEY not in tree: attr = self.eval_command and 'quant' or 'direction_quant' if getattr(self, attr) is None: setattr(self, attr, 0) setattr(self, attr, getattr(self, attr) * 1...
chr(key) in digits or self.direction_keys._tree[key]
key in digitlist or self.direction_keys._tree[key]
def _do_eval_command(self, key): assert isinstance(self.tree_pointer, dict), self.tree_pointer try: self.tree_pointer = self.tree_pointer[key] except TypeError: self.failure = True return None except KeyError: try: chr(key) in digits or self.direction_keys._tree[key] self.tree_pointer = self.tree_pointer[DIRKEY] except...
kw['stdout'] = sys.stderr
kw['stdout'] = sys.stdout
def run(self): """ Run the application in the way specified by the options.
log("freeing {0}".format(self))
def __del__(self): log("freeing {0}".format(self)) for handler in self.handlers: self.settings.signal_unbind(handler)
for key, val in special_keys.items():
for key, val in tuple(special_keys.items()):
def parse_keybinding(obj): """ Translate a keybinding to a sequence of integers Example: lol<CR> => (ord('l'), ord('o'), ord('l'), ord('\n')) => (108, 111, 108, 10) x<A-Left> => (120, (27, curses.KEY_LEFT)) """ assert isinstance(obj, (tuple, int, str)) if isinstance(obj, tuple): for char in obj: yield char eli...
default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0])
try: default_editor = os.environ['EDITOR'] except KeyError: pass else: parts = default_editor.split() exe_name = os.path.basename(parts[0]) if exe_name in self.fm.executables: return tuple(parts) + tuple(c)
def app_editor(self, c): default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0])
if exe_name in self.fm.executables: return tuple(parts) + tuple(c) else: return self.either(c, 'vim', 'emacs', 'nano')
return self.either(c, 'vim', 'emacs', 'nano')
def app_editor(self, c): default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0])
pos = self.wid - 3
pos = self.wid - 1
def click(self, event): """Handle a MouseEvent""" direction = event.mouse_wheel_direction() if direction: self.fm.tab_move(direction) self.need_redraw = True return True
def open_console(self, mode=':', string='', prompt=None):
def open_console(self, mode=cmode.COMMAND, string='', prompt=None):
def open_console(self, mode=':', string='', prompt=None): """Open the console if the current UI supports that""" if hasattr(self.ui, 'open_console'): self.ui.open_console(mode, string, prompt=prompt)
map('zd', fm.toggle_boolean_option('directories_first'))
map('zd', fm.toggle_boolean_option('sort_directories_first'))
def move(arg): arg.wdg.move(narg=arg.n, **arg.direction)
map('<home>', wdg.move(right=0, absolute=True)) map('<end>', wdg.move(right=-1, absolute=True))
map('<home>', '<C-A>', wdg.move(right=0, absolute=True)) map('<end>', '<C-E>', wdg.move(right=-1, absolute=True))
def ctrl_c(arg): try: item = arg.fm.loader.queue[0] except: arg.fm.notify("Type Q or :quit<Enter> to exit Ranger") else: arg.fm.notify("Aborting: " + item.get_description()) arg.fm.loader.remove(index=0)
map('<backspace>', wdg.delete(-1)) map('<delete>', wdg.delete(0))
map('<backspace>', '<C-H>', wdg.delete(-1)) map('<delete>', '<C-D>', wdg.delete(0))
def ctrl_c(arg): try: item = arg.fm.loader.queue[0] except: arg.fm.notify("Type Q or :quit<Enter> to exit Ranger") else: arg.fm.notify("Aborting: " + item.get_description()) arg.fm.loader.remove(index=0)
obj = CommandLoader(args=['mv'] + mv_flags + + [f.path for f in copied_files]
obj = CommandLoader(args=['mv'] + mv_flags \ + [f.path for f in copied_files] \
def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content()
obj = CommandLoader(args=['cp', '-f'] + cp_flags
obj = CommandLoader(args=['cp', '-f'] + cp_flags \
def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content()
obj = CommandLoader(args=['cp'] + cp_flags + [f.path for f in copied_files]
obj = CommandLoader(args=['cp'] + cp_flags \ + [f.path for f in copied_files] \
def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content()
if ranger.debug:
if ranger.arg.debug:
def notify(self, text, duration=4, bad=False): if isinstance(text, Exception): if ranger.debug: raise bad = True text = str(text) self.log.appendleft(text) if hasattr(self.ui, 'notify'): self.ui.notify(text, duration=duration, bad=bad)
minY = goodY.min() yRange = maxY - minY
def plotHistogram(coaddName, chiSqOrder): coadd = pyfits.open(coaddName) coaddData = coadd[0].data # undo normalization coaddData *= float(chiSqOrder) # get rid of nans and infs goodData = numpy.extract(numpy.isfinite(coaddData.flat), coaddData.flat) goodData = numpy.extract(goodData < 50, goodData) hist, binEdges = n...
yEndVal = minY + (yRange * 0.01) smallYIndices = numpy.where(goodY < yEndVal)[0] endInd = numpy.extract(smallYIndices > maxYInd, smallYIndices)[0]
tailMinY = goodY[maxYInd:].min() yRange = maxY - tailMinY yEndVal = tailMinY + (yRange * 0.01) endInd = numpy.where(goodY[maxYInd:] <= yEndVal)[0][0] + maxYInd endInd = len(goodY)-1
def plotHistogram(coaddName, chiSqOrder): coadd = pyfits.open(coaddName) coaddData = coadd[0].data # undo normalization coaddData *= float(chiSqOrder) # get rid of nans and infs goodData = numpy.extract(numpy.isfinite(coaddData.flat), coaddData.flat) goodData = numpy.extract(goodData < 50, goodData) hist, binEdges = n...
BaseHTTPServer.HTTPServer.__init__(self, (host, port), RecordHandler)
try: BaseHTTPServer.HTTPServer.__init__(self, (host, port), RecordHandler) except Exception, e: logging.critical('Could not start HTTPServer on port %d: %s', port, e) return
def __init__( self, http_archive_filename, use_deterministic_script, real_dns_lookup, host='localhost', port=80): self.use_deterministic_script = use_deterministic_script self.archive_filename = http_archive_filename self.real_http_request = RealHttpRequest(real_dns_lookup)
BaseHTTPServer.HTTPServer.__init__(self, (host, port), ReplayHandler)
try: BaseHTTPServer.HTTPServer.__init__(self, (host, port), ReplayHandler) except Exception, e: logging.critical('Could not start HTTPServer on port %d: %s', port, e) return
def __init__(self, http_archive_filename, use_deterministic_script, host='localhost', port=80): self.use_deterministic_script = use_deterministic_script self.http_archive = httparchive.HttpArchive.Create(http_archive_filename) logging.info('Loaded %d responses from %s', len(self.http_archive), http_archive_filename) Ba...
query.filter("platform =", int(self.request.get("platform_filter"))) if self.request.get("version_vilter"): query.filter("version =", int(self.request.get("version_vilter")))
query.filter("platform =", self.request.get("platform_filter")) if self.request.get("version_filter"): query.filter("version =", self.request.get("version_filter"))
def do_set_search(self): query = models.TestSet.all(); query.order("-date")
self._ipfw([ 'pipe', dns_pipe, 'config', 'bw', '0', 'delay', delay_ms, 'plr', packet_loss_rate ]) self._ipfw(['add', self.pipe_set, 'pipe', dns_pipe, 'udp', 'from', 'any', 'to', '127.0.0.1', 'dst-port', '53'])
def set_traffic_shaping(self, up_bandwidth = '0', down_bandwidth = '0', delay_ms = '0', packet_loss_rate = '0'): """Start shaping traffic.
def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()):
def __init__(self, host='127.0.0.1', port=53, platform_settings=None): if not platformsettings: platformsettings.get_platform_settings()
def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()): self.host = host self.platform_settings = platform_settings try: SocketServer.ThreadingUDPServer.__init__( self, (host, port), UdpDnsHandler) except socket.error, (error_number, msg): if error_number == errno.EACC...
self.platform_settings = platform_settings
self.restore_primary_dns = platform_settings.restore_primary_dns
def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()): self.host = host self.platform_settings = platform_settings try: SocketServer.ThreadingUDPServer.__init__( self, (host, port), UdpDnsHandler) except socket.error, (error_number, msg): if error_number == errno.EACC...
self.connection.close()
self.close_connection = 1
def send_archived_http_response(self, response): try: # We need to set the server name before we start the response. # Take a scan through the response headers here use_chunked = False has_content_length = False server_name = "WebPageReplay" for header, value in response.headers: if header == "server": server_name = va...
logging.critical("Traffic Shaping Exception ", e)
logging.critical("Traffic Shaping Exception: %s ", e)
def set_traffic_shaping(self, up_bandwidth = '0', down_bandwidth = '0', delay_ms = '0', packet_loss_rate = '0'): """Start shaping traffic.
logging.basicConfig(level=log_level)
logging.basicConfig(level=log_level, format='%(asctime)s %(levelname)s %(message)s')
def format_description(self, description): if description: return description + '\n' else: return ''
def __call_(self, request, headers):
def __call__(self, request, headers):
def __call_(self, request, headers): logging.debug('RealHttpRequest: %s %s', request.host, request.path) host_ip = self._real_dns_lookup(request.host) connection = httplib.HTTPConnection(host_ip) connection.request( request.command, request.path, request.request_body, headers) response = connection.getresponse() connec...
'queue', queue_size
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'dst-port', '53'
'dst-port', '53',
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'mask', 'src-port', '0xffff'
'mask', 'src-port', '0xffff',
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'pipe', self._UPLOAD_PIPE,
'queue', self._UPLOAD_QUEUE,
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'dst-port', '80'
'dst-port', '80',
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'mask', 'dst-port', '0xffff'
'mask', 'dst-port', '0xffff',
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'pipe', self._DOWNLOAD_PIPE,
'queue', self._DOWNLOAD_QUEUE,
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
'src-port', '80'
'src-port', '80',
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return
"-p", str(self.config["packet_loss_rate"] / 100),
"-p", str(self.config["packet_loss_rate"] / 100.0),
def StartProxy(self): logging.debug("Starting Web-Page-Replay") log_level = "info" if self.log_level: log_level = self.log_level cmdline = [ replay_path, "-x", # Disables DNS intercepting "-d", str(self.config["download_bandwidth_kbps"]) + "KBit/s", "-u", str(self.config["upload_bandwidth_kbps"]) + "KBit/s", "-m", str...
json_error.error = format % args self.response.out.write(json.dumps(json_error))
json_error['error'] = format % args self.response.out.write(json.encode(json_error))
def send_json_error(self, format, *args): """Send a fatal request error to the error log and json output.""" logging.error(format, *args) json_error = {} json_error.error = format % args self.response.out.write(json.dumps(json_error))
json_output.obj = test_set json_output.summaries = test_set.summaries self.response.out.write(json.dumps(json_output))
json_output['obj'] = test_set json_output['summaries'] = [s for s in test_set.summaries] self.response.out.write(json.encode(json_output))
def do_set(self): """Lookup a specific TestSet.""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_set = models.TestSet.get(db.Key(set_id)) if not test_set: self.send_json_error("Could not find id: ", id) return
json_output.obj = test_summary
json_output['obj'] = test_summary
def do_summary(self): """ Lookup a specific TestSummary""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_summary = models.TestSummary.get(db.Key(set_id)) if not test_summary: self.send_json_error("Could not find id: ", id) return
json_output.results = test_results self.response.out.write(json.dumps(json_output))
json_output['results'] = [r for r in test_results] self.response.out.write(json.encode(json_output))
def do_summary(self): """ Lookup a specific TestSummary""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_summary = models.TestSummary.get(db.Key(set_id)) if not test_summary: self.send_json_error("Could not find id: ", id) return
if headers.get('content-type', '').startswith('text/'):
content_type = headers.get('content-type', '') if (content_type.startswith('text/') or content_type == 'application/x-javascript'):
def remove_header(self, key): for i, (k, v) in enumerate(self.headers): if key == k: self.headers.pop(i) return
'plr', self.packet_loss_rate
'plr', self.packet_loss_rate, 'queue', queue_size
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The...
'queue', '4000000'
'queue', queue_size
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The...
self._ipfw([
self.platformsettings.ipfw([
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The...
option_parser.add_option('-n', '--no-deterministic_script', default=True,
harness_group.add_option('-n', '--no-deterministic_script', default=True,
def format_description(self, description): if description: return description + '\n' else: return ''
option_parser.add_option('-P', '--no-dns_private_passthrough', default=True,
harness_group.add_option('-P', '--no-dns_private_passthrough', default=True,
def format_description(self, description): if description: return description + '\n' else: return ''
option_parser.add_option('-x', '--no-dns_forwarding', default=True,
harness_group.add_option('-x', '--no-dns_forwarding', default=True,
def format_description(self, description): if description: return description + '\n' else: return ''
delay_ms = str(int(delay_ms) / 2)
self.delay_ms = str(int(self.delay_ms) / 2)
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The...
if shape_dns:
if self.shape_dns:
def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The...
obj.read_bytes_kb = int(request.get('read_bytes_kb')) obj.write_bytes_kb = int(request.get('write_bytes_kb'))
obj.read_bytes_kb = int(float(request.get('read_bytes_kb'))) obj.write_bytes_kb = int(float(request.get('write_bytes_kb')))
def ApplyStatisticsData(request, obj): """Applies statistics uploaded via the request to the object.""" obj.using_spdy = bool(request.get('using_spdy')=="CHECKED") obj.start_load_time = int(request.get('start_load_time')) obj.commit_load_time = int(request.get('commit_load_time')) obj.doc_load_time = int(request.get('d...
d = self.process_rs(owner, content, rs)
d = self.process_rs(rs)
def collect(self, vs=None, rs=None): """ Collect data.
d.addCallback(lambda x: len(x) > 0 and x[0][0] or None)
d.addCallback(lambda x: len(x) > 0 and (x[0][0] + 1) or None)
def age(self, ctx, lb=None, vs=None, rs=None, sorry=False): """ Return the age of the given resource.
query that should return URL (without prefix) of matching results. The URL will be prefixed by API version.
query that should return URL of matching results.
def childFactory(self, ctx, name): """ Dispatch the search to the generic search handler. """ return SearchGenericResource(self.dbpool, name)
Run the search query and prefix the results with API URL.
Run the search query.
def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d
api = IApiVersion(ctx)
def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d
self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results])
self.query(), {'term': self.term}) d.addCallback(lambda results: [y[0] for y in results])
def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d
SELECT 'loadbalancer/' || name || '/'
SELECT '/loadbalancer/' || name || '/'
def query(self): return """
SELECT 'loadbalancer/' || lb || '/virtualserver/' || vs || '/'
SELECT '/loadbalancer/' || lb || '/virtualserver/' || vs || '/'
def query(self): return """
SELECT 'loadbalancer/' || lb || '/virtualserver/' || vs || '/realserver/' || rs || '/'
SELECT '/loadbalancer/' || lb || '/virtualserver/' || vs || '/realserver/' || rs || '/'
def query(self): return """
mo = re.match(r"[rb](\d+)", rs)
mo = re.match(r"([rb])(\d+)", rs)
def parse(self, vs=None, rs=None): """ Parse vs and rs into v, s, g, r. """ if vs is not None: mo = re.match(r"v(\d+)s(\d+)g(\d+)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) v, s, g = int(mo.group(1)), int(mo.group(2)), int(mo.group(3)) if rs is not None: mo = re.match(r"[rb](\d+)", rs) i...
r = int(mo.group(1)) return v, s, g, r return v, s, g, None return None, None, None, None
r = int(mo.group(2)) return v, s, g, r, mo.group(1) == "b" return v, s, g, None, None return None, None, None, None, None
def parse(self, vs=None, rs=None): """ Parse vs and rs into v, s, g, r. """ if vs is not None: mo = re.match(r"v(\d+)s(\d+)g(\d+)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) v, s, g = int(mo.group(1)), int(mo.group(2)), int(mo.group(3)) if rs is not None: mo = re.match(r"[rb](\d+)", rs) i...
v, s, g, r = self.parse(vs, rs)
v, s, g, r, backup = self.parse(vs, rs)
def collect(self, vs=None, rs=None): """ Collect data for an Alteon """ v, s, g, r = self.parse(vs, rs) if v is not None: if r is not None: # Collect data to refresh a specific real server d = self.process_rs(v, s, g, r) else: # Collect data to refresh a virtual server d = self.process_vs(v, s, g) else: # Otherwise, co...
d = self.process_rs(v, s, g, r)
d = self.process_rs(v, s, g, r, backup)
def collect(self, vs=None, rs=None): """ Collect data for an Alteon """ v, s, g, r = self.parse(vs, rs) if v is not None: if r is not None: # Collect data to refresh a specific real server d = self.process_rs(v, s, g, r) else: # Collect data to refresh a virtual server d = self.process_vs(v, s, g) else: # Otherwise, co...
v, s, g, r = self.parse(vs, rs)
v, s, g, r, backup = self.parse(vs, rs) if backup is True: yield {} return
def actions(self, vs=None, rs=None): """ List possible actions.
v, s, g, r = self.parse(vs, rs)
v, s, g, r, backup = self.parse(vs, rs) if backup is True: yield None return
def execute(self, action, actionargs=None, vs=None, rs=None): """ Execute an action.
g = defer.waitForDeferred(
gr = defer.waitForDeferred(
def process_vs(self, v, s, g): """ Process data for a given virtual server when no real server is provided
yield g g.getResult()
yield gr gr.getResult()
def process_vs(self, v, s, g): """ Process data for a given virtual server when no real server is provided
mo = re.match(r"(.*)|(.*)", vs)
mo = re.match(r"(.*)\|(.*)", vs)
def parse(self, vs=None, rs=None): """ Parse vs and rs into owner, content, rs """ if vs is not None: mo = re.match(r"(.*)|(.*)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) owner, content = mo.groups() return owner, content, rs return None, None, None
d = self.process_rs(rs)
d = self.process_rs(owner, content, rs, None)
def collect(self, vs=None, rs=None): """ Collect data.
rs = defer.waitForDeferred(self.process_rs(service))
rs = defer.waitForDeferred(self.process_rs(owner, content, service))
def process_vs(self, owner, content): """ Process data for a given virtual server when no real server is provided
rs = defer.waitForDeferred(self.process_rs(service, backup))
rs = defer.waitForDeferred(self.process_rs(owner, content, service, backup))
def process_vs(self, owner, content): """ Process data for a given virtual server when no real server is provided
def process_rs(self, service, backup=False):
def process_rs(self, owner, content, service, backup=False):
def process_rs(self, service, backup=False): """ Process data for a given virtual server and real server.
vip = "mark%d:%d" % tuple(self.cache(('virtualServerFwMark', v), ('virtualServerPort', v)))
vip = "mark%d:0" % self.cache(('virtualServerFwMark', v))
def process_vs(self, v): """ Process data for a given virtual server when no real server is provided
def nogateway(x):
def nogateway():
def nogateway(x): request.setResponseCode(504) # Gateway timeout return "No gateway available"
d.addBoth(lambda x: x is not None and collector.collect(vs, rs) and x or x)
d.addBoth(lambda x: refresh(x))
def execute_and_refresh(collector): d = collector.execute(action, actionargs, vs, rs) # We refresh only if the result is not None (action has # been executed) We don't alter the original result. Don't # refresh a whole load balancer. if vs is not None or rs is not None: d.addBoth(lambda x: x is not None and collector.c...
for rstate in state[1:]:
for rstate in states[1:]:
def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if c...
if cur == "disabled":
if state == "disabled":
def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if c...
if cur == "down":
if state == "down":
def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if c...
if cur == "ok":
if state == "ok":
def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if c...
fc_name, fc_date = get_flowcell_info(fc_dir)
def main(config_file, fc_dir, run_info_yaml=None): work_dir = os.getcwd() with open(config_file) as in_handle: config = yaml.load(in_handle) if run_info_yaml: with open(run_info_yaml) as in_handle: run_details = yaml.load(in_handle) run_info = dict(details=run_details, run_id="") else: galaxy_api = GalaxyApiAccess(conf...
os.makedirs(align_dir)
try: os.makedirs(align_dir) except OSError: pass assert os.path.exists(align_dir)
def do_alignment(fastq1, fastq2, align_ref, sam_ref, lane_name, sample_name, align_dir, config, config_file): """Align to the provided reference genome, returning an aligned SAM file. """ aligner_to_use = config["algorithm"]["aligner"] if not os.path.exists(align_dir): os.makedirs(align_dir) print lane_name, "Aligning ...