rem
stringlengths
0
322k
add
stringlengths
0
2.05M
context
stringlengths
8
228k
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").read().strip('\n').split() if getRecipeStores :
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in \"${getRecipeStores[@]}\"; do echo $i; done'").read().strip('\n').split('\n') if getRecipeStores and getRecipeStores[0] :
def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo...
return (v+'-'+r).strip('-')
if r : return v+'-'+r else : return v
def Join_Version_Revision(v, r) : return (v+'-'+r).strip('-')
os.environ[name] = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; echo -n $"+name+"'").read()
os.environ[name] = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${"+name+"[@]}; do echo $i; done'").read().strip('\n')
def getGoboVariable(name, filename = 'GoboPath', isList = False) : if filename != 'GoboPath' : goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboScriptsDefaults = goboPrograms+'/Scripts/Current/Resources/Defaults/Settings/' goboManagerDefaults = goboPrograms+'/Manager/Cu...
compileRecipeDirs = os.environ['compileRecipeDirs'].split()
compileRecipeDirs = os.environ['compileRecipeDirs'].strip('\n').split()
def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo...
compileRecipeDirs = os.popen("/bin/sh -c '. " +file+" 2> /dev/null; for i in ${compileRecipeDirs[@]}; do echo $i; done'").readlines()
compileRecipeDirs = os.popen("/bin/sh -c '. " +file+" 2> /dev/null; for i in ${compileRecipeDirs[@]}; do echo $i; done'").read().strip('\n').split()
def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo...
getRecipeStores = os.environ['getRecipeStores'].split()
getRecipeStores = os.environ['getRecipeStores'].strip('\n').split('\n')
def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo...
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").readlines()
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").read().strip('\n').split()
def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo...
def controller_scan(directory):
def controller_scan(directory=None):
def controller_scan(directory): """ Scan a directory for python files and use them as controllers """ def find_controllers(dirname, prefix=''): controllers = [] for fname in os.listdir(dirname): filename = os.path.join(dirname, fname) if os.path.isfile(filename) and re.match('^[^_]{1,1}.*\.py$', fname): controllers.app...
(reg, gaps) = self.buildnextreg(self.routelist, clist) print reg
(reg, noreqs, allblank) = self.buildnextreg(self.routelist, clist)
def makeregexp(self, clist): """ Create a regular expression for matching purposes, this MUST be called before match can function properly. clist should be a list of valid controller strings that can be matched, for this reason makeregexp should be called by the web framework after it knows all available controllers t...
default = '[^/]+'
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
(rest, moregaps) = ('', False)
(rest, noreqs, allblank) = ('', True, True)
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
(rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False
(rest, noreqs, allblank) = self.buildnextreg(path[1:], clist)
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
if self.reqs.has_key(var) and self.defaults.has_key(var) and not moregaps: gaps = True reg = '(/(?P<' + var + '>' + self.reqs[var] + ')?' reg = reg + rest elif self.reqs.has_key(var) and self.defaults.has_key(var) and moregaps: gaps = True reg = '(/(?P<' + var + '>' + self.reqs[var] + ')' reg = reg + rest elif self.req...
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
gaps = True
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
partreg = '(?P<' + var + '>' + default + ')' if self.defaults.has_key(var) and not moregaps: reg = '(/' + partreg + ')?'
partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if noreqs: if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(/' + partreg + rest + ')?' elif self.reqs.has_key(var): allblank = False reg = '(/' + partreg + ')' + rest elif self.defaults.has_key(var): reg = '(/' + part...
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
reg = '/' + partreg
if allblank and self.defaults.has_key(var): reg = '(/' + partreg + rest + ')?' else: reg = '/' + partreg + rest
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
reg = '(/' + '(?P<' + var + '>.*))*'
reg = '(/' + '(?P<' + var + '>.*))*' + rest
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
gaps = True reg = '/' + part if moregaps and gaps: reg = '(' + reg + rest + ')?' else: reg = reg + rest return (reg, gaps)
noreqs = False allblank = False reg = '/' + part + rest return (reg, noreqs, allblank)
def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor...
if environ.get('SERVER_PORT', 80) != 80: self.__shared_state.host += ':' + environ['SERVER_PORT']
self.__shared_state.host += port_info
def load_wsgi_environ(self, environ): """ Load the protocol/server info from the environ and store it. Also, match the incoming URL if there's already a mapper, and store the resulting match dict in mapper_dict. """ if environ.get('HTTPS'): self.__shared_state.protocol = 'https' else: self.__shared_state.protocol = 'ht...
primary = (method != 'GET' and lst.pop(0)) or None
if method not in ['POST', 'GET', 'any']: primary = lst.pop(0) else: primary = None
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
return self.not_found(environ, start_response)
match = {}
def __call__(self, environ, start_response): config = request_config() config.mapper = self.mapper old_method = None if self.use_method_override: req = WSGIRequest(environ) if '_method' in environ.get('QUERY_STRING', '') and '_method' in req.GET: old_method = environ['REQUEST_METHOD'] environ['REQUEST_METHOD'] = req.G...
def not_found(self, environ, start_response): start_response('404 Not Found', [('Content-type', 'text/plain')]) return ['Not found']
def __call__(self, environ, start_response): config = request_config() config.mapper = self.mapper old_method = None if self.use_method_override: req = WSGIRequest(environ) if '_method' in environ.get('QUERY_STRING', '') and '_method' in req.GET: old_method = environ['REQUEST_METHOD'] environ['REQUEST_METHOD'] = req.G...
self.conditions['method'] != environ.get('HTTP_METHOD'):
self.conditions['method'] != environ.get('REQUEST_METHOD'):
def match(self, url, environ=None): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, requirement regexp doesn't pass, etc.). Therefore the callin...
for key in self.reqs.keys(): if key not in matchdict.keys() or matchdict[key] is None: try: result[key] = self.defaults[key] except: return False else: value = matchdict[key] or (self.defaults.has_key(key) and self.defaults[key]) or '' if not re.compile('^' + self.reqs[key] + '$').match(value): return False
def match(self, url): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, requirement regexp doesn't pass, etc.). Therefore the calling function sho...
if match: return (match, route) return None
if self.debug: matchlog.append(dict(route=route, regexp=bool(match))) if match: return (match, route, matchlog) return (None, None, matchlog)
def _match(self, url): """Internal Route matcher Matches a URL against a route, and returns a tuple of the match dict and the route object if a match is successfull, otherwise it returns empty. For internal use only. """ if not self._created_regs and self.controller_scan: self.create_regs() elif not self._created_re...
if result:
if self.debug: return result[0], result[1], result[2] if result[0]:
def match(self, url): """Match a URL against against one of the routes contained. Will return None if no valid match is found. resultdict = m.match('/joe/sixpack') """ result = self._match(url) if result: return result[0] return None
if result:
if self.debug: return result[0], result[1], result[2] if result[0]:
def routematch(self, url): """Match a URL against against one of the routes contained. Will return None if no valid match is found, otherwise a result dict and a route object is returned. resultdict, route_obj = m.match('/joe/sixpack') """ result = self._match(url) if result: return result[0], result[1] return None ...
self.connect("formatted_" + route_name, "%s.:(format);%s")
self.connect("formatted_" + route_name, "%s.:(format);%s" % (collection_path, action))
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth.upper()]} return opts
controllers.sort(cmp=longest_first)
controllers.sort(longest_first)
def longest_first(a, b): return cmp(len(b), len(a))
if environ.get('SERVER_PORT') != '80':
if environ.get('SERVER_PORT', '80') != '80':
def load_wsgi_environ(self, environ): """ Load the protocol/server info from the environ and store it. Also, match the incoming URL if there's already a mapper, and store the resulting match dict in mapper_dict. """ port_info = '' if environ.get('HTTPS'): self.__shared_state.protocol = 'https' if environ.get('SERVER_PO...
if meth: kargs['_method'] = meth if meth not in self.conditions.get('method', [meth]): return False
if meth: kargs['_method'] = meth
def generate(self,**kargs): """Generate a URL from ourself given a set of keyword arguments Toss an exception if this set of keywords would cause a gap in the url. """ # Verify that our args pass any regexp requirements for key in self.reqs.keys(): val = kargs.get(key) if val and not self.req_regs[key].match(str(val)...
new_path = path_prefix + "/new" member_path = path_prefix + "/:id"
new_path = path + "/new" member_path = path + "/:(id)"
def swap(dct, newdct): for key, val in dct.iteritems(): newdct.setdefault(val, []).append(key) return newdct
opts['conditions'] = {'method':meth}
opts['conditions'] = {'method':[meth]}
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
self.connect("formatted_" + route_name, "%s.:(format);%s")
self.connect("formatted_" + route_name, "%s\.:(format);%s")
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
self.connect("%s.:(format)" % collection_path, **route_options)
self.connect("%s\.:(format)" % collection_path, **route_options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
conditions={'method':'GET'}, **options)
conditions={'method':['GET']}, **options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
collection_path + ".:(format)", action='index', conditions={'method':'GET'}, **options)
collection_path + "\.:(format)", action='index', conditions={'method':['GET']}, **options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
path = (action == 'new' and action + '.:(format)') or \ "%s.:(format);%s" % (new_path, action)
path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
"%s.:(format);%s" % (member_path, action), **route_options)
"%s\.:(format);%s" % (member_path, action), **route_options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
hostmatch = config.environ['HTTP_HOST'].split(':')
fullhost = config.environ.get('HTTP_HOST') or config.environ.get('SERVER_NAME') hostmatch = fullhost.split(':')
def _subdomain_check(config, kargs): if config.mapper.sub_domains: subdomain = kargs.pop('sub_domain', None) hostmatch = config.environ['HTTP_HOST'].split(':') host = hostmatch[0] port = '' if len(hostmatch) > 1: port += ':' + hostmatch[1] sub_match = re.compile('^.+?\.(%s)$' % config.mapper.domain_match) domain = re.s...
controllers.extend(find_controllers(filename, prefix=fname+'/'))
controllers.extend(find_controllers(filename, prefix=prefix+fname+'/'))
def find_controllers(dirname, prefix=''): controllers = [] for fname in os.listdir(dirname): filename = dirname + '/' + fname if os.path.isfile(filename) and re.match('^[^_]{1,1}.*\.py$', fname): controllers.append(prefix + fname[:-3]) elif os.path.isdir(filename): controllers.extend(find_controllers(filename, prefix=f...
class RouteConfig(threadinglocal.local): pass
def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 ...
_shared_state = RouteConfig()
__shared_state = threadinglocal.local()
def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 ...
return self._shared_state.__getattr__(name)
return self.__shared_state.__getattr__(name)
def __getattr__(self, name): return self._shared_state.__getattr__(name)
return self._shared_state.__setattr__(name, value)
return self.__shared_state.__setattr__(name, value)
def __setattr__(self, name, value): return self._shared_state.__setattr__(name, value)
newargs = route.defaults.copy() host = newargs.pop('host', '') protocol = newargs.pop('protocol', '') del newargs['_static'] newargs.update(kargs) url = route.generate(_ignore_req_list=True, **newargs) if not url: url = '' kargs = None
url = route.routepath
def url_for(*args, **kargs): """Generates a URL All keys given to url_for are sent to the Routes Mapper instance for generation except for:: anchor specified the anchor name to be appened to the path host overrides the default (current) host if provided protocol overrides the default (curre...
if isinsance(use_sd, list) and sub_domain not in use_sd:
if isinstance(use_sd, list) and sub_domain not in use_sd:
def match(self, url, environ=None, sub_domains=False, sub_domains_ignore=None, domain_match=''): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, ...
var = part['name']
var = re.escape(part['name'])
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the...
partreg = '(?P<' + var + '>' + '|'.join(clist) + ')'
partreg = '(?P<' + var + '>' + '|'.join(map(re.escape, clist)) + ')'
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the...
reg = part[:-1] + '(/' + rest + ')?'
reg = re.escape(part[:-1]) + '(/' + rest + ')?'
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the...
reg = part + rest
reg = re.escape(part) + rest
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the...
self.connect("formatted_" + route_name, "%s\.:(format);%s")
self.connect("formatted_" + route_name, "%s.:(format);%s")
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
self.connect("%s\.:(format)" % collection_path, **route_options)
self.connect("%s.:(format)" % collection_path, **route_options)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
collection_path + "\.:(format)", action='index',
collection_path + ".:(format)", action='index',
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action)
path = (action == 'new' and new_path + '.:(format)') or \ "%s.:(format);%s" % (new_path, action)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
"%s\.:(format);%s" % (member_path, action), **route_options)
"%s.:(format);%s" % (member_path, action), **route_options)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
for extrastyle in list(layer.wmsextrastyles):
for extrastyle in [layer.wmsdefaultstyle] + list(layer.wmsextrastyles):
def __init__(self, conf, mapfactory, opsonlineresource): self.conf = conf self.mapfactory = mapfactory if self.conf.has_option('service', 'epsg'): self.crs = CRS('EPSG', self.conf.get('service', 'epsg')) else: ServerConfigurationError('EPSG code not properly configured.')
self._set_servers()
def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman")
return def _set_servers(self): conf_file = open("/etc/pacman.conf", "r").read() self.servers = {} for repo in self.repos: if repo == "foreigners": continue self.servers[repo] = [] self._get_repo_servers(repo, conf_file) continue return def _get_repo_servers(self, repo, conf_file): try: begin = conf_file.index("[%s]" ...
def _set_filelist(self, pac, path): """Set installed files list for the given pac""" if not pac.installed: return _("%s is not installed") %pac.name files = open("%s/files" %path).read() begin = files.index("%FILES%") + len("%FILES%") end = files.find("%", begin) - len("%") filelist = files[begin:end].strip() pac.file...
self.show_all()
self.show()
def run(self): self.show_all() self.terminal.fork_command() self.terminal.feed_child("pacman -Sy;exit\n")
self.vbox.show_all() self.close_button.hide()
def _setup_layout(self):
def close(self, widget): self.destroy()
def run(self):
self.show_all()
self.show()
def run(self): self.show_all() self.terminal.do_upgrade()
if self.database["foreigners"]:
try:
def _setup_pacs_models(self): self.models = {}
self.models["foreigners"] = installed_list(self.database["foreigners"])
if self.database["foreigners"]: self.models["foreigners"] = installed_list(self.database["foreigners"])
def _setup_pacs_models(self): self.models = {}
elif keyword.count("+"): keys = keyword.split("+")
elif keywords.count("+"): keys = keywords.split("+")
def get_by_keywords(self, keywords): """Return pacs which have keywords as name or in description""" keys = [] #Split keywords by '+' or spaces if keywords.count("+") and keywords.count(" "): keys_1 = keywords.split("+") for key in keys_1: keys = key.split(" ") elif keywords.count(" "): keys = keywords.split() elif key...
self.pages["all"].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
self.pages["local"]["node"].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
self.pages["remote"][repo].toggle_renderer.connect("toggled", self.add_install )
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
self.pages["local"][repo].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
def _get_selected_page(self, repo_tree):
def _get_selected_page(self, repo_tree, toggle=None):
def _get_selected_page(self, repo_tree): model, tree_iter = repo_tree.get_selection().get_selected() if tree_iter is None: tree_iter = model.get_iter_from_string("0")
def add_install(self, widget, data=None):
def add_install(self, widget=None, data=None):
def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
return def remove_install(self, widget, data=None):
installed = model.get_value(l_iter, 0) if installed == "yellow" or installed == "green": model.set_value(l_iter, 1, STOCK_GO_UP) else: model.set_value(l_iter, 1, STOCK_ADD) self.queued[name] = (model, l_iter) return def remove_install(self, widget=None, data=None):
def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
return def add_remove(self, widget, data=None):
model.set_value(l_iter, 1, "") self.queued.pop(name) return def add_remove(self, widget=None, data=None):
def remove_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
return def remove_remove(self, widget, data=None):
model.set_value(l_iter, 1, STOCK_REMOVE) self.queued[name] = (model, l_iter) return def remove_remove(self, widget=None, data=None):
def add_remove(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
self.toggle_renderer = CellRendererToggle() self.toggle_renderer.activatable = True
action_renderer = CellRendererPixbuf()
def _setup_columns(self):
first_col.pack_start(self.toggle_renderer)
first_col.pack_start(action_renderer)
def _setup_columns(self):
first_col.set_attributes(self.toggle_renderer, active=1)
first_col.set_attributes(action_renderer, stock_id=1)
def _setup_columns(self):
ListStore.__init__(self, str, bool, str, str, str)
ListStore.__init__(self, str, str, str, str, str)
def __init__(self, packages, repo, node):
self.gld.get_widget("main_win").show_all()
self._setup_prog_bar()
def __init__(self, fname, icons): """Create the XML Object and connect handlers, then setup notebook""" self.gld = XML(fname, "main_win")
self.repos = os.listdir("/var/lib/pacman") self.repos.remove("local") try: self.repos.remove("wget-log") except ValueError: pass
self.repos = self._get_repos()
def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman")
print "Found a target: " + str(target) + ", for pattern: " + str(pattern)
def matchTargets(self): rePatternedTargets = {} for pattern in self.patternedTargets.keys(): rePatternedTargets[pattern] = re.compile(pattern)
print "Processing: " + str(path)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
if allInBetween: for i in range(len(path)): if i < (len(path) - 1): nodes.setdefault(path[i], []) if path[i + 1] not in nodes[path[i]]: nodes[path[i]].append(path[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep)
print "Matching path: " + str(newpath) if allInBetween: for i in range(len(newpath)): nodes.setdefault(newpath[i], []) if i < (len(newpath) - 1): if newpath[i + 1] not in nodes[newpath[i]]: nodes[newpath[i]].append(newpath[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
paths.append(path + [dep])
paths.append(newpath)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
def DoxyfileParse(file_contents, dir, data = None):
def DoxyfileParse(file): data = DoxyfileParse_(file,{}) for (k,v) in data.items(): if not v : del data[k] elif k in ("INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES") : continue elif len(v)==1 : data[k] = v[0] return data def DoxyfileParse_(file, data):
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac...
if data is None : data = {}
dir = os.path.dirname(file)
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac...
lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:"
lex = shlex.shlex(instream=open(file), posix=True) lex.wordchars += "*+./-:@~"
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac...
if token in ['\n']: if last_token not in ['\\']:
if token=='\n': if last_token!='\\':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
elif token in ['\\']:
elif token=='\\':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
if key == '@' : key += token else : key = token if token != '@' : key_token = False
key = token key_token = False
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
data[key] = list()
data[key] = []
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
DoxyfileParse(open(inc).read(),dir,data)
DoxyfileParse_(inc,data)
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
append_data( data, key, new_data, '\\' ) for (k, v) in data.items(): if len(v) == 0: data.pop(k) if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0]
append_data(data, key, new_data, '\\')
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
dep_add_keys = [
dep_add_keys = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++...
]
)
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++...
default_file_patterns = [
default_file_patterns = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++...
] default_exclude_patterns = [
) default_exclude_patterns = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++...
basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" )
basedir = node.dir.abspath data = DoxyfileParse(node.abspath) recursive = data.get("RECURSIVE", "NO").upper()=="YES"
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++...
data = DoxyfileParse(source[0].get_contents(), str(source[0].dir))
data = DoxyfileParse(source[0].abspath)
def DoxyEmitter(source, target, env): """Doxygen Doxyfile emitter""" # possible output formats and their default values and output locations output_formats = { "HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"), } data = DoxyfileParse(source[0].get_con...