rem stringlengths 0 322k | add stringlengths 0 2.05M | context stringlengths 8 228k |
|---|---|---|
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").read().strip('\n').split() if getRecipeStores : | getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in \"${getRecipeStores[@]}\"; do echo $i; done'").read().strip('\n').split('\n') if getRecipeStores and getRecipeStores[0] : | def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo... |
return (v+'-'+r).strip('-') | if r : return v+'-'+r else : return v | def Join_Version_Revision(v, r) : return (v+'-'+r).strip('-') |
os.environ[name] = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; echo -n $"+name+"'").read() | os.environ[name] = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${"+name+"[@]}; do echo $i; done'").read().strip('\n') | def getGoboVariable(name, filename = 'GoboPath', isList = False) : if filename != 'GoboPath' : goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboScriptsDefaults = goboPrograms+'/Scripts/Current/Resources/Defaults/Settings/' goboManagerDefaults = goboPrograms+'/Manager/Cu... |
compileRecipeDirs = os.environ['compileRecipeDirs'].split() | compileRecipeDirs = os.environ['compileRecipeDirs'].strip('\n').split() | def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo... |
compileRecipeDirs = os.popen("/bin/sh -c '. " +file+" 2> /dev/null; for i in ${compileRecipeDirs[@]}; do echo $i; done'").readlines() | compileRecipeDirs = os.popen("/bin/sh -c '. " +file+" 2> /dev/null; for i in ${compileRecipeDirs[@]}; do echo $i; done'").read().strip('\n').split() | def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo... |
getRecipeStores = os.environ['getRecipeStores'].split() | getRecipeStores = os.environ['getRecipeStores'].strip('\n').split('\n') | def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo... |
getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").readlines() | getRecipeStores = os.popen("/bin/sh -c '. " + file +" 2> /dev/null; for i in ${getRecipeStores[@]}; do echo $i; done'").read().strip('\n').split() | def getCompileOptions() : import os goboSettings = getGoboVariable('goboSettings') goboPrograms = getGoboVariable('goboPrograms') goboCompileDefaults = goboPrograms+'/Compile/Current/Resources/Defaults/Settings/' compileSettingsFiles = [ "~/.Settings/Compile/Compile.conf", goboSettings + "/Compile/Compile.conf", gobo... |
def controller_scan(directory): | def controller_scan(directory=None): | def controller_scan(directory): """ Scan a directory for python files and use them as controllers """ def find_controllers(dirname, prefix=''): controllers = [] for fname in os.listdir(dirname): filename = os.path.join(dirname, fname) if os.path.isfile(filename) and re.match('^[^_]{1,1}.*\.py$', fname): controllers.app... |
(reg, gaps) = self.buildnextreg(self.routelist, clist) print reg | (reg, noreqs, allblank) = self.buildnextreg(self.routelist, clist) | def makeregexp(self, clist): """ Create a regular expression for matching purposes, this MUST be called before match can function properly. clist should be a list of valid controller strings that can be matched, for this reason makeregexp should be called by the web framework after it knows all available controllers t... |
default = '[^/]+' | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... | |
(rest, moregaps) = ('', False) | (rest, noreqs, allblank) = ('', True, True) | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
(rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False | (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
if self.reqs.has_key(var) and self.defaults.has_key(var) and not moregaps: gaps = True reg = '(/(?P<' + var + '>' + self.reqs[var] + ')?' reg = reg + rest elif self.reqs.has_key(var) and self.defaults.has_key(var) and moregaps: gaps = True reg = '(/(?P<' + var + '>' + self.reqs[var] + ')' reg = reg + rest elif self.req... | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... | |
gaps = True | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... | |
partreg = '(?P<' + var + '>' + default + ')' if self.defaults.has_key(var) and not moregaps: reg = '(/' + partreg + ')?' | partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if noreqs: if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(/' + partreg + rest + ')?' elif self.reqs.has_key(var): allblank = False reg = '(/' + partreg + ')' + rest elif self.defaults.has_key(var): reg = '(/' + part... | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
reg = '/' + partreg | if allblank and self.defaults.has_key(var): reg = '(/' + partreg + rest + ')?' else: reg = '/' + partreg + rest | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
reg = '(/' + '(?P<' + var + '>.*))*' | reg = '(/' + '(?P<' + var + '>.*))*' + rest | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
gaps = True reg = '/' + part if moregaps and gaps: reg = '(' + reg + rest + ')?' else: reg = reg + rest return (reg, gaps) | noreqs = False allblank = False reg = '/' + part + rest return (reg, noreqs, allblank) | def buildnextreg(self, path, clist): default = '[^/]+' part = path[0] reg = '' (rest, moregaps) = ('', False) if len(path[1:]) > 0: (rest, moregaps) = self.buildnextreg(path[1:], clist) gaps = False if part.startswith(':'): var = part[1:] partreg = '' if self.reqs.has_key(var) and self.defaults.has_key(var) and not mor... |
if environ.get('SERVER_PORT', 80) != 80: self.__shared_state.host += ':' + environ['SERVER_PORT'] | self.__shared_state.host += port_info | def load_wsgi_environ(self, environ): """ Load the protocol/server info from the environ and store it. Also, match the incoming URL if there's already a mapper, and store the resulting match dict in mapper_dict. """ if environ.get('HTTPS'): self.__shared_state.protocol = 'https' else: self.__shared_state.protocol = 'ht... |
primary = (method != 'GET' and lst.pop(0)) or None | if method not in ['POST', 'GET', 'any']: primary = lst.pop(0) else: primary = None | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
return self.not_found(environ, start_response) | match = {} | def __call__(self, environ, start_response): config = request_config() config.mapper = self.mapper old_method = None if self.use_method_override: req = WSGIRequest(environ) if '_method' in environ.get('QUERY_STRING', '') and '_method' in req.GET: old_method = environ['REQUEST_METHOD'] environ['REQUEST_METHOD'] = req.G... |
def not_found(self, environ, start_response): start_response('404 Not Found', [('Content-type', 'text/plain')]) return ['Not found'] | def __call__(self, environ, start_response): config = request_config() config.mapper = self.mapper old_method = None if self.use_method_override: req = WSGIRequest(environ) if '_method' in environ.get('QUERY_STRING', '') and '_method' in req.GET: old_method = environ['REQUEST_METHOD'] environ['REQUEST_METHOD'] = req.G... | |
self.conditions['method'] != environ.get('HTTP_METHOD'): | self.conditions['method'] != environ.get('REQUEST_METHOD'): | def match(self, url, environ=None): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, requirement regexp doesn't pass, etc.). Therefore the callin... |
for key in self.reqs.keys(): if key not in matchdict.keys() or matchdict[key] is None: try: result[key] = self.defaults[key] except: return False else: value = matchdict[key] or (self.defaults.has_key(key) and self.defaults[key]) or '' if not re.compile('^' + self.reqs[key] + '$').match(value): return False | def match(self, url): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, requirement regexp doesn't pass, etc.). Therefore the calling function sho... | |
if match: return (match, route) return None | if self.debug: matchlog.append(dict(route=route, regexp=bool(match))) if match: return (match, route, matchlog) return (None, None, matchlog) | def _match(self, url): """Internal Route matcher Matches a URL against a route, and returns a tuple of the match dict and the route object if a match is successfull, otherwise it returns empty. For internal use only. """ if not self._created_regs and self.controller_scan: self.create_regs() elif not self._created_re... |
if result: | if self.debug: return result[0], result[1], result[2] if result[0]: | def match(self, url): """Match a URL against against one of the routes contained. Will return None if no valid match is found. resultdict = m.match('/joe/sixpack') """ result = self._match(url) if result: return result[0] return None |
if result: | if self.debug: return result[0], result[1], result[2] if result[0]: | def routematch(self, url): """Match a URL against against one of the routes contained. Will return None if no valid match is found, otherwise a result dict and a route object is returned. resultdict, route_obj = m.match('/joe/sixpack') """ result = self._match(url) if result: return result[0], result[1] return None ... |
self.connect("formatted_" + route_name, "%s.:(format);%s") | self.connect("formatted_" + route_name, "%s.:(format);%s" % (collection_path, action)) | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth.upper()]} return opts |
controllers.sort(cmp=longest_first) | controllers.sort(longest_first) | def longest_first(a, b): return cmp(len(b), len(a)) |
if environ.get('SERVER_PORT') != '80': | if environ.get('SERVER_PORT', '80') != '80': | def load_wsgi_environ(self, environ): """ Load the protocol/server info from the environ and store it. Also, match the incoming URL if there's already a mapper, and store the resulting match dict in mapper_dict. """ port_info = '' if environ.get('HTTPS'): self.__shared_state.protocol = 'https' if environ.get('SERVER_PO... |
if meth: kargs['_method'] = meth if meth not in self.conditions.get('method', [meth]): return False | if meth: kargs['_method'] = meth | def generate(self,**kargs): """Generate a URL from ourself given a set of keyword arguments Toss an exception if this set of keywords would cause a gap in the url. """ # Verify that our args pass any regexp requirements for key in self.reqs.keys(): val = kargs.get(key) if val and not self.req_regs[key].match(str(val)... |
new_path = path_prefix + "/new" member_path = path_prefix + "/:id" | new_path = path + "/new" member_path = path + "/:(id)" | def swap(dct, newdct): for key, val in dct.iteritems(): newdct.setdefault(val, []).append(key) return newdct |
opts['conditions'] = {'method':meth} | opts['conditions'] = {'method':[meth]} | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
self.connect("formatted_" + route_name, "%s.:(format);%s") | self.connect("formatted_" + route_name, "%s\.:(format);%s") | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
self.connect("%s.:(format)" % collection_path, **route_options) | self.connect("%s\.:(format)" % collection_path, **route_options) | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
conditions={'method':'GET'}, **options) | conditions={'method':['GET']}, **options) | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
collection_path + ".:(format)", action='index', conditions={'method':'GET'}, **options) | collection_path + "\.:(format)", action='index', conditions={'method':['GET']}, **options) | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
path = (action == 'new' and action + '.:(format)') or \ "%s.:(format);%s" % (new_path, action) | path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action) | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
"%s.:(format);%s" % (member_path, action), **route_options) | "%s\.:(format);%s" % (member_path, action), **route_options) | def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts |
hostmatch = config.environ['HTTP_HOST'].split(':') | fullhost = config.environ.get('HTTP_HOST') or config.environ.get('SERVER_NAME') hostmatch = fullhost.split(':') | def _subdomain_check(config, kargs): if config.mapper.sub_domains: subdomain = kargs.pop('sub_domain', None) hostmatch = config.environ['HTTP_HOST'].split(':') host = hostmatch[0] port = '' if len(hostmatch) > 1: port += ':' + hostmatch[1] sub_match = re.compile('^.+?\.(%s)$' % config.mapper.domain_match) domain = re.s... |
controllers.extend(find_controllers(filename, prefix=fname+'/')) | controllers.extend(find_controllers(filename, prefix=prefix+fname+'/')) | def find_controllers(dirname, prefix=''): controllers = [] for fname in os.listdir(dirname): filename = dirname + '/' + fname if os.path.isfile(filename) and re.match('^[^_]{1,1}.*\.py$', fname): controllers.append(prefix + fname[:-3]) elif os.path.isdir(filename): controllers.extend(find_controllers(filename, prefix=f... |
class RouteConfig(threadinglocal.local): pass | def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 ... | |
_shared_state = RouteConfig() | __shared_state = threadinglocal.local() | def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 ... |
return self._shared_state.__getattr__(name) | return self.__shared_state.__getattr__(name) | def __getattr__(self, name): return self._shared_state.__getattr__(name) |
return self._shared_state.__setattr__(name, value) | return self.__shared_state.__setattr__(name, value) | def __setattr__(self, name, value): return self._shared_state.__setattr__(name, value) |
newargs = route.defaults.copy() host = newargs.pop('host', '') protocol = newargs.pop('protocol', '') del newargs['_static'] newargs.update(kargs) url = route.generate(_ignore_req_list=True, **newargs) if not url: url = '' kargs = None | url = route.routepath | def url_for(*args, **kargs): """Generates a URL All keys given to url_for are sent to the Routes Mapper instance for generation except for:: anchor specified the anchor name to be appened to the path host overrides the default (current) host if provided protocol overrides the default (curre... |
if isinsance(use_sd, list) and sub_domain not in use_sd: | if isinstance(use_sd, list) and sub_domain not in use_sd: | def match(self, url, environ=None, sub_domains=False, sub_domains_ignore=None, domain_match=''): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, ... |
var = part['name'] | var = re.escape(part['name']) | def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the... |
partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' | partreg = '(?P<' + var + '>' + '|'.join(map(re.escape, clist)) + ')' | def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the... |
reg = part[:-1] + '(/' + rest + ')?' | reg = re.escape(part[:-1]) + '(/' + rest + ')?' | def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the... |
reg = part + rest | reg = re.escape(part) + rest | def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the... |
self.connect("formatted_" + route_name, "%s\.:(format);%s") | self.connect("formatted_" + route_name, "%s.:(format);%s") | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
self.connect("%s\.:(format)" % collection_path, **route_options) | self.connect("%s.:(format)" % collection_path, **route_options) | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
collection_path + "\.:(format)", action='index', | collection_path + ".:(format)", action='index', | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action) | path = (action == 'new' and new_path + '.:(format)') or \ "%s.:(format);%s" % (new_path, action) | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
"%s\.:(format);%s" % (member_path, action), **route_options) | "%s.:(format);%s" % (member_path, action), **route_options) | def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts |
for extrastyle in list(layer.wmsextrastyles): | for extrastyle in [layer.wmsdefaultstyle] + list(layer.wmsextrastyles): | def __init__(self, conf, mapfactory, opsonlineresource): self.conf = conf self.mapfactory = mapfactory if self.conf.has_option('service', 'epsg'): self.crs = CRS('EPSG', self.conf.get('service', 'epsg')) else: ServerConfigurationError('EPSG code not properly configured.') |
self._set_servers() | def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman") | |
return def _set_servers(self): conf_file = open("/etc/pacman.conf", "r").read() self.servers = {} for repo in self.repos: if repo == "foreigners": continue self.servers[repo] = [] self._get_repo_servers(repo, conf_file) continue return def _get_repo_servers(self, repo, conf_file): try: begin = conf_file.index("[%s]" ... | def _set_filelist(self, pac, path): """Set installed files list for the given pac""" if not pac.installed: return _("%s is not installed") %pac.name files = open("%s/files" %path).read() begin = files.index("%FILES%") + len("%FILES%") end = files.find("%", begin) - len("%") filelist = files[begin:end].strip() pac.file... | |
self.show_all() | self.show() | def run(self): self.show_all() self.terminal.fork_command() self.terminal.feed_child("pacman -Sy;exit\n") |
self.vbox.show_all() self.close_button.hide() | def _setup_layout(self): | |
def close(self, widget): self.destroy() | def run(self): | |
self.show_all() | self.show() | def run(self): self.show_all() self.terminal.do_upgrade() |
if self.database["foreigners"]: | try: | def _setup_pacs_models(self): self.models = {} |
self.models["foreigners"] = installed_list(self.database["foreigners"]) | if self.database["foreigners"]: self.models["foreigners"] = installed_list(self.database["foreigners"]) | def _setup_pacs_models(self): self.models = {} |
elif keyword.count("+"): keys = keyword.split("+") | elif keywords.count("+"): keys = keywords.split("+") | def get_by_keywords(self, keywords): """Return pacs which have keywords as name or in description""" keys = [] #Split keywords by '+' or spaces if keywords.count("+") and keywords.count(" "): keys_1 = keywords.split("+") for key in keys_1: keys = key.split(" ") elif keywords.count(" "): keys = keywords.split() elif key... |
self.pages["all"].toggle_renderer.connect("toggled", self.add_install) | def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook") | |
self.pages["local"]["node"].toggle_renderer.connect("toggled", self.add_install) | def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook") | |
self.pages["remote"][repo].toggle_renderer.connect("toggled", self.add_install ) | def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook") | |
self.pages["local"][repo].toggle_renderer.connect("toggled", self.add_install) | def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook") | |
def _get_selected_page(self, repo_tree): | def _get_selected_page(self, repo_tree, toggle=None): | def _get_selected_page(self, repo_tree): model, tree_iter = repo_tree.get_selection().get_selected() if tree_iter is None: tree_iter = model.get_iter_from_string("0") |
def add_install(self, widget, data=None): | def add_install(self, widget=None, data=None): | def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree) |
return def remove_install(self, widget, data=None): | installed = model.get_value(l_iter, 0) if installed == "yellow" or installed == "green": model.set_value(l_iter, 1, STOCK_GO_UP) else: model.set_value(l_iter, 1, STOCK_ADD) self.queued[name] = (model, l_iter) return def remove_install(self, widget=None, data=None): | def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree) |
return def add_remove(self, widget, data=None): | model.set_value(l_iter, 1, "") self.queued.pop(name) return def add_remove(self, widget=None, data=None): | def remove_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree) |
return def remove_remove(self, widget, data=None): | model.set_value(l_iter, 1, STOCK_REMOVE) self.queued[name] = (model, l_iter) return def remove_remove(self, widget=None, data=None): | def add_remove(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree) |
self.toggle_renderer = CellRendererToggle() self.toggle_renderer.activatable = True | action_renderer = CellRendererPixbuf() | def _setup_columns(self): |
first_col.pack_start(self.toggle_renderer) | first_col.pack_start(action_renderer) | def _setup_columns(self): |
first_col.set_attributes(self.toggle_renderer, active=1) | first_col.set_attributes(action_renderer, stock_id=1) | def _setup_columns(self): |
ListStore.__init__(self, str, bool, str, str, str) | ListStore.__init__(self, str, str, str, str, str) | def __init__(self, packages, repo, node): |
self.gld.get_widget("main_win").show_all() | self._setup_prog_bar() | def __init__(self, fname, icons): """Create the XML Object and connect handlers, then setup notebook""" self.gld = XML(fname, "main_win") |
self.repos = os.listdir("/var/lib/pacman") self.repos.remove("local") try: self.repos.remove("wget-log") except ValueError: pass | self.repos = self._get_repos() | def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman") |
print "Found a target: " + str(target) + ", for pattern: " + str(pattern) | def matchTargets(self): rePatternedTargets = {} for pattern in self.patternedTargets.keys(): rePatternedTargets[pattern] = re.compile(pattern) | |
print "Processing: " + str(path) | def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False): | |
if allInBetween: for i in range(len(path)): if i < (len(path) - 1): nodes.setdefault(path[i], []) if path[i + 1] not in nodes[path[i]]: nodes[path[i]].append(path[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep) | print "Matching path: " + str(newpath) if allInBetween: for i in range(len(newpath)): nodes.setdefault(newpath[i], []) if i < (len(newpath) - 1): if newpath[i + 1] not in nodes[newpath[i]]: nodes[newpath[i]].append(newpath[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep) | def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False): |
paths.append(path + [dep]) | paths.append(newpath) | def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False): |
def DoxyfileParse(file_contents, dir, data = None): | def DoxyfileParse(file): data = DoxyfileParse_(file,{}) for (k,v) in data.items(): if not v : del data[k] elif k in ("INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES") : continue elif len(v)==1 : data[k] = v[0] return data def DoxyfileParse_(file, data): | def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac... |
if data is None : data = {} | dir = os.path.dirname(file) | def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac... |
lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" | lex = shlex.shlex(instream=open(file), posix=True) lex.wordchars += "*+./-:@~" | def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespac... |
if token in ['\n']: if last_token not in ['\\']: | if token=='\n': if last_token!='\\': | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
elif token in ['\\']: | elif token=='\\': | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
if key == '@' : key += token else : key = token if token != '@' : key_token = False | key = token key_token = False | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
data[key] = list() | data[key] = [] | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
DoxyfileParse(open(inc).read(),dir,data) | DoxyfileParse_(inc,data) | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
append_data( data, key, new_data, '\\' ) for (k, v) in data.items(): if len(v) == 0: data.pop(k) if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0] | append_data(data, key, new_data, '\\') | def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token |
dep_add_keys = [ | dep_add_keys = ( | def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++... |
] | ) | def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++... |
default_file_patterns = [ | default_file_patterns = ( | def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++... |
] default_exclude_patterns = [ | ) default_exclude_patterns = ( | def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++... |
basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) | basedir = node.dir.abspath data = DoxyfileParse(node.abspath) recursive = data.get("RECURSIVE", "NO").upper()=="YES" | def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++... |
data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) | data = DoxyfileParse(source[0].abspath) | def DoxyEmitter(source, target, env): """Doxygen Doxyfile emitter""" # possible output formats and their default values and output locations output_formats = { "HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"), } data = DoxyfileParse(source[0].get_con... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.