rem
stringlengths
0
322k
add
stringlengths
0
2.05M
context
stringlengths
8
228k
_form_delnofilter(form['nofilterhosts'])
_form_delnofilter(form)
def exec_form (form): # reset info/error del info[:] del error[:] # proxy port if form.has_key('port'): _form_proxyport(form['port'].value) # proxy user if form.has_key('proxyuser'): _form_proxyuser(form['proxyuser'].value.strip()) # proxy pass if form.has_key('proxypass'): _form_proxypass(base64.encodestring(form['proxypass'].value.strip())) # parent proxy host if form.has_key('parentproxy'): _form_parentproxy(form['parentproxy'].value.strip()) # parent proxy port if form.has_key('parentproxyport'): _form_parentproxyport(form['parentproxyport'].value) # parent proxy user if form.has_key('parentproxyuser'): _form_parentproxyuser(form['parentproxyuser'].value.strip()) # parent proxy pass if form.has_key('parentproxypass'): _form_parentproxypass( base64.encodestring(form['parentproxypass'].value)) # timeout if form.has_key('timeout'): _form_timeout(form['timeout'].value) # filter modules _form_filtermodules(form) # allowed hosts if form.has_key('addallowed') and form.has_key('newallowed'): _form_addallowed(form['newallowed'].value.strip()) elif form.has_key('delallowed') and form.has_key('allowedhosts'): _form_delallowed(form['allowedhosts']) # no filter hosts if form.has_key('addnofilter') and form.has_key('newnofilter'): _form_addnofilter(form['newnofilter'].value.strip()) elif form.has_key('delnofilter') and form.has_key('nofilterhosts'): _form_delnofilter(form['nofilterhosts']) if info: # write changed config config.write_proxyconf()
def _form_delallowed (hosts): if hasattr(hosts, "value"): delhosts = [hosts.value.strip()] else: delhosts = [ host.value.strip() for host in hosts ] hosts = ip.map2hosts(config['allowedhosts'])
def _form_removehosts (form, key): toremove = getlist(form, key) hosts = ip.map2hosts(config[key])
def _form_delallowed (hosts): if hasattr(hosts, "value"): # single host delhosts = [hosts.value.strip()] else: # multiple delhosts = [ host.value.strip() for host in hosts ] hosts = ip.map2hosts(config['allowedhosts']) removed = 0 for host in delhosts: if host in hosts: hosts.remove(host) removed += 1 if removed > 0: config['allowedhosts'] = ip.hosts2map(hosts) config['allowedhostlist'] = sort_seq(hosts) if removed == 1: info.append(i18n._("Allowed host successfully removed")) else: info.append(i18n._("%d allowed hosts successfully removed") % \ removed)
for host in delhosts:
for host in toremove:
def _form_delallowed (hosts): if hasattr(hosts, "value"): # single host delhosts = [hosts.value.strip()] else: # multiple delhosts = [ host.value.strip() for host in hosts ] hosts = ip.map2hosts(config['allowedhosts']) removed = 0 for host in delhosts: if host in hosts: hosts.remove(host) removed += 1 if removed > 0: config['allowedhosts'] = ip.hosts2map(hosts) config['allowedhostlist'] = sort_seq(hosts) if removed == 1: info.append(i18n._("Allowed host successfully removed")) else: info.append(i18n._("%d allowed hosts successfully removed") % \ removed)
p = wc.ZapperParser(fullname, wconfig, compile_data=False)
p = wc.configuration.ZapperParser(fullname, wconfig, compile_data=False)
def update_filter (wconfig, dryrun=False, log=None): """Update the given configuration object with .zap files found at baseurl. If dryrun is True, only print out the changes but do nothing throws IOError on error """ print >> log, _("updating filters"), "..." chg = False baseurl = wconfig['baseurl']+"filter/" url = baseurl+"filter-md5sums.txt" try: page = open_url(url) except IOError, msg: print >> log, _("error fetching %s") % url, msg print >> log, "...", _("done") return chg # remember all local config files filemap = {} for filename in wc.configuration.filterconf_files(wconfig.filterdir): filemap[os.path.basename(filename)] = filename # read md5sums for line in page.read().splitlines(): if "<" in line: print >> log, _("error fetching %s") % url print >> log, "...", _("done") return chg if not line: continue md5sum, filename = line.split() assert filename.endswith('.zap') fullname = os.path.join(wc.ConfigDir, filename) # compare checksums if filemap.has_key(filename): f = file(fullname) data = f.read() digest = list(md5.new(data).digest()) f.close() digest = "".join([ "%0.2x"%ord(c) for c in digest ]) if digest == md5sum: print >> log, \ _("filter %s not changed, ignoring") % filename continue print >> log, _("updating filter %s") % filename else: print >> log, _("adding new filter %s") % filename # parse new filter url = baseurl+filename page = open_url(url) p = wc.ZapperParser(fullname, wconfig, compile_data=False) p.parse(fp=page) page.close() if wconfig.merge_folder(p.folder, dryrun=dryrun, log=log): chg = True url = baseurl+"extern-md5sums.txt" try: page = open_url(url) except IOError, msg: print >> log, _("error fetching %s:") % url, msg print >> log, "...", _("done") return chg lines = page.read().splitlines() page.close() for line in lines: if "<" in line: print >> log, _("error fetching %s:") % url, \ _("invalid content") print >> log, "...", _("done") return chg if not line: continue md5sum, filename = line.split() # XXX UNIX-generated md5sum filenames with subdirs are not portable fullname = os.path.join(wc.ConfigDir, filename) # compare checksums if os.path.exists(fullname): f = file(fullname) data = f.read() digest = list(md5.new(data).digest()) f.close() digest = "".join([ "%0.2x"%ord(c) for c in digest ]) if digest == md5sum: print >> log, \ _("extern filter %s not changed, ignoring")%filename continue print >> log, _("updating extern filter %s") % filename else: print >> log, _("adding new extern filter %s") % filename chg = True if not dryrun: url = baseurl+filename try: page = open_url(url) except IOError, msg: print >> log, _("error fetching %s:") % url, msg continue data = page.read() if not data: print >> log, _("error fetching %s:") % url, \ _("got no data") continue f = file(fullname, 'wb') f.write(data) f.close() print >> log, "...", _("done") return chg
os.umask(0277)
os.umask(0177)
def start (parent_exit=1): """start a daemon using the appropriate pidfile""" # already running? if os.path.exists(pidfile): return _("""WebCleaner already started (lock file found).
debug(NIGHTMARE, "matched")
debug(NIGHTMARE, "matched, replacing", `repl`)
def replace_one(self, ro, repl, data): debug(NIGHTMARE, "checking replacer rule", ro.pattern, "...") offset = 0 mo = ro.search(data, offset) while mo: debug(NIGHTMARE, "matched") data = data[:mo.start()] + repl + data[mo.end():] offset = mo.start()+len(repl) mo = ro.search(data, offset) return data
offset = mo.start()+len(repl)
offset = mo.start()+len(repl)+1
def replace_one(self, ro, repl, data): debug(NIGHTMARE, "checking replacer rule", ro.pattern, "...") offset = 0 mo = ro.search(data, offset) while mo: debug(NIGHTMARE, "matched") data = data[:mo.start()] + repl + data[mo.end():] offset = mo.start()+len(repl) mo = ro.search(data, offset) return data
else: ratings[category] = value if category not in ["generic", "modified"]: values[category] = {value: True}
values[key] = value
def _form_ratings (form): for key, value in _get_prefix_vals(form, 'category_'): category = _get_category(key) if not category.is_valid_value(value): error['categoryvalue'] = True return False else: ratings[category] = value if category not in ["generic", "modified"]: values[category] = {value: True} return True
global url rating = {} rating.update(ratings) if generic: rating['generic'] = u"true" rating['modified'] = u"%d" % int(_time.time()) if _rating_add(url, rating):
if url in rating_store: rating = rating_store[url] else: rating = _Rating(url, generic) rating.remove_categories() for catname, value in values.items(): category = _get_category(catname) rating.add_category_value(category, value) rating_store[url] = rating try: rating_store.write()
def _form_apply (): global url rating = {} rating.update(ratings) if generic: rating['generic'] = u"true" rating['modified'] = u"%d" % int(_time.time()) if _rating_add(url, rating): info['ratingupdated'] = True else: error['ratingupdated'] = True
else:
except:
def _form_apply (): global url rating = {} rating.update(ratings) if generic: rating['generic'] = u"true" rating['modified'] = u"%d" % int(_time.time()) if _rating_add(url, rating): info['ratingupdated'] = True else: error['ratingupdated'] = True
global config config = Configuration() config.init_filter_modules()
def startfunc (): if os.name=='posix': import signal signal.signal(signal.SIGHUP, reload_config) global config config = Configuration() config.init_filter_modules() if os.geteuid()==0: # drop root privileges import pwd, grp try: pentry = pwd.getpwnam("nobody") pw_uid = 2 nobody = pentry[pw_uid] gentry = grp.getgrnam("nogroup") gr_gid = 2 nogroup = gentry[gr_gid] os.setgid(nogroup) os.setuid(nobody) except KeyError: print >>sys.stderr, "warning: could not drop root privileges, user nobody and/or group nogroup not found" pass import wc.proxy wc.proxy.mainloop()
self['strict_whitelist'] = None
self['strict_whitelist'] = 0
def reset (self): """Reset to default values""" self['port'] = 8080 self['proxyuser'] = "" self['proxypass'] = "" self['parentproxy'] = "" self['parentproxyport'] = 3128 self['parentproxyuser'] = "" self['parentproxypass'] = "" self['logfile'] = "" self['strict_whitelist'] = None self['debuglevel'] = 0 self['rules'] = [] self['filters'] = [] self['filterlist'] = [[],[],[],[],[],[],[],[],[],[]] self['errorlen'] = ErrorLen self['errortext'] = ErrorText self['colorize'] = None self['noproxyfor'] = {} self['allowedhosts'] = {} self['starttime'] = time.time() self['requests'] = {'valid':0, 'error':0, 'blocked':0} self['local_sockets_only'] = None self['localip'] = socket.gethostbyname(socket.gethostname()) self['mime_content_rewriting'] = [] self['headersave'] = 100 self['showerrors'] = None
self['colorize'] = None
self['colorize'] = 0
def reset (self): """Reset to default values""" self['port'] = 8080 self['proxyuser'] = "" self['proxypass'] = "" self['parentproxy'] = "" self['parentproxyport'] = 3128 self['parentproxyuser'] = "" self['parentproxypass'] = "" self['logfile'] = "" self['strict_whitelist'] = None self['debuglevel'] = 0 self['rules'] = [] self['filters'] = [] self['filterlist'] = [[],[],[],[],[],[],[],[],[],[]] self['errorlen'] = ErrorLen self['errortext'] = ErrorText self['colorize'] = None self['noproxyfor'] = {} self['allowedhosts'] = {} self['starttime'] = time.time() self['requests'] = {'valid':0, 'error':0, 'blocked':0} self['local_sockets_only'] = None self['localip'] = socket.gethostbyname(socket.gethostname()) self['mime_content_rewriting'] = [] self['headersave'] = 100 self['showerrors'] = None
self['local_sockets_only'] = None
self['local_sockets_only'] = 0
def reset (self): """Reset to default values""" self['port'] = 8080 self['proxyuser'] = "" self['proxypass'] = "" self['parentproxy'] = "" self['parentproxyport'] = 3128 self['parentproxyuser'] = "" self['parentproxypass'] = "" self['logfile'] = "" self['strict_whitelist'] = None self['debuglevel'] = 0 self['rules'] = [] self['filters'] = [] self['filterlist'] = [[],[],[],[],[],[],[],[],[],[]] self['errorlen'] = ErrorLen self['errortext'] = ErrorText self['colorize'] = None self['noproxyfor'] = {} self['allowedhosts'] = {} self['starttime'] = time.time() self['requests'] = {'valid':0, 'error':0, 'blocked':0} self['local_sockets_only'] = None self['localip'] = socket.gethostbyname(socket.gethostname()) self['mime_content_rewriting'] = [] self['headersave'] = 100 self['showerrors'] = None
if self.defer_data and data: self.defer_data = False self.client.server_response(self, self.response, self.statuscode, self.headers) if not self.client: return
if data: if self.defer_data: self.defer_data = False self.client.server_response(self, self.response, self.statuscode, self.headers) if not self.client: return
def process_content (self): """ Process server data: filter it and write it to client. """ data = self.read(self.bytes_remaining) wc.log.debug(wc.LOG_PROXY, "%s process %d bytes", self, len(data)) if self.bytes_remaining is not None: # If we do know how many bytes we're dealing with, # we'll close the connection when we're done self.bytes_remaining -= len(data) wc.log.debug(wc.LOG_PROXY, "%s %d bytes remaining", self, self.bytes_remaining) is_closed = False for decoder in self.decoders: data = decoder.decode(data) wc.log.debug(wc.LOG_PROXY, "%s have run decoder %s", self, decoder) if not is_closed and decoder.closed: is_closed = True try: for stage in FilterStages: data = wc.filter.applyfilter(stage, data, "filter", self.attrs) except wc.filter.FilterWait, msg: wc.log.debug(wc.LOG_PROXY, "%s FilterWait %s", self, msg) except wc.filter.FilterRating, msg: wc.log.debug(wc.LOG_PROXY, "%s FilterRating from content %s", self, msg) self._show_rating_deny(str(msg)) return except wc.filter.FilterProxyError, e: self.client.error(e.status, e.msg, txt=e.text) self.handle_error("filter proxy error") return underflow = self.bytes_remaining is not None and \ self.bytes_remaining < 0 if underflow: wc.log.warn(wc.LOG_PROXY, _("server received %d bytes more than content-length"), (-self.bytes_remaining)) if self.statuscode != 407: if self.defer_data and data: # defer until data is non-empty, which ensures that # every filter above has seen at least some data self.defer_data = False self.client.server_response(self, self.response, self.statuscode, self.headers) if not self.client: return self.client.server_content(data) if is_closed or self.bytes_remaining == 0: # either we ran out of bytes, or the decoder says we're done self.state = 'recycle'
self['filterlist'] = [[], [], [], [], [], [], [], [], [], []]
self['filterlist'] = {}
def reset (self): """Reset to default values""" # The bind address specifies on which address the socket should # listen. # The default empty string represents INADDR_ANY which means to # accept incoming connections from all hosts. self['bindaddress'] = '' self['port'] = 8080 self['sslport'] = 8443 self['sslgateway'] = 0 self['adminuser'] = "" self['adminpass'] = "" self['proxyuser'] = "" self['proxypass'] = "" self['parentproxy'] = "" self['parentproxyport'] = 3128 self['parentproxyuser'] = "" self['parentproxypass'] = "" # dynamically stored parent proxy authorization credentials self['parentproxycreds'] = None self['folderrules'] = [] self['filters'] = [] self['filterlist'] = [[], [], [], [], [], [], [], [], [], []] self['colorize'] = 0 # DNS resolved nofilterhosts self['nofilterhosts'] = None self['allowedhosts'] = None self['starttime'] = time.time() self['mime_content_rewriting'] = sets.Set() self['gui_theme'] = "classic" self['timeout'] = 10 self['auth_ntlm'] = 0 if os.name == 'posix': self['clamavconf'] = "/etc/clamav/clamav.conf" elif os.name == 'nt': self['clamavconf'] = r"c:\clamav-devel\etc\clamav.conf" else: self['clamavconf'] = os.path.join(os.getcwd(), "clamav.conf") # in development mode some values have different defaults self['development'] = os.environ.get("WC_DEVELOPMENT", 0) self['baseurl'] = wc.Url self['try_google'] = 0 # delete all registered sids from wc.filter.rules import delete_registered_sids delete_registered_sids()
FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DEBUGLEVEL,ConfWindow.onCmdDebuglevel)
def eventMap (self): """attach all events to (member) functions""" FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ACCEPT,ConfWindow.onCmdAccept) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_CANCEL,ConfWindow.onCmdCancel) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_APPLY,ConfWindow.onCmdApply) FXMAPFUNC(self,SEL_UPDATE,ConfWindow.ID_APPLY,ConfWindow.onUpdApply) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ABOUT,ConfWindow.onCmdAbout) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PORT,ConfWindow.onCmdPort) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DEBUGLEVEL,ConfWindow.onCmdDebuglevel) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXY,ConfWindow.onCmdParentProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYPORT,ConfWindow.onCmdParentProxyPort) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_LOGFILE,ConfWindow.onCmdLogfile) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_FILTERMODULE,ConfWindow.onCmdFilterModule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_FILTER,ConfWindow.onCmdFilter) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NEWFOLDER,ConfWindow.onCmdNewFolder) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NEWRULE,ConfWindow.onCmdNewRule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_REMOVE,ConfWindow.onCmdRemove) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_TITLE,ConfWindow.onCmdTitle) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTART,ConfWindow.onCmdProxyStart) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTOP,ConfWindow.onCmdProxyStop) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYRESTART,ConfWindow.onCmdProxyRestart) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYRELOAD,ConfWindow.onCmdProxyReload) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTATUS,ConfWindow.onCmdProxyStatus) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_CONFUPDATE,ConfWindow.onCmdConfUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DISABLERULE,ConfWindow.onCmdDisableRule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_ADD,ConfWindow.onCmdNoProxyForAdd) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_EDIT,ConfWindow.onCmdNoProxyForEdit) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_NOPROXYFOR_EDIT,ConfWindow.onUpdNoProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_REMOVE,ConfWindow.onCmdNoProxyForRemove) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_NOPROXYFOR_REMOVE,ConfWindow.onUpdNoProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_ADD,ConfWindow.onCmdAllowedHostsAdd) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_EDIT,ConfWindow.onCmdAllowedHostsEdit) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_ALLOWEDHOSTS_EDIT,ConfWindow.onUpdAllowedHosts) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_REMOVE,ConfWindow.onCmdAllowedHostsRemove) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_ALLOWEDHOSTS_REMOVE,ConfWindow.onUpdAllowedHosts) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_UP,ConfWindow.onCmdUp) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_UP,ConfWindow.onCmdUpUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DOWN,ConfWindow.onCmdDown) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_DOWN,ConfWindow.onCmdDownUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYUSER,ConfWindow.onCmdProxyUser) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYPASS,ConfWindow.onCmdProxyPass) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYUSER,ConfWindow.onCmdParentProxyUser) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYPASS,ConfWindow.onCmdParentProxyPass)
FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_LOGFILE,ConfWindow.onCmdLogfile)
def eventMap (self): """attach all events to (member) functions""" FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ACCEPT,ConfWindow.onCmdAccept) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_CANCEL,ConfWindow.onCmdCancel) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_APPLY,ConfWindow.onCmdApply) FXMAPFUNC(self,SEL_UPDATE,ConfWindow.ID_APPLY,ConfWindow.onUpdApply) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ABOUT,ConfWindow.onCmdAbout) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PORT,ConfWindow.onCmdPort) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DEBUGLEVEL,ConfWindow.onCmdDebuglevel) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXY,ConfWindow.onCmdParentProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYPORT,ConfWindow.onCmdParentProxyPort) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_LOGFILE,ConfWindow.onCmdLogfile) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_FILTERMODULE,ConfWindow.onCmdFilterModule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_FILTER,ConfWindow.onCmdFilter) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NEWFOLDER,ConfWindow.onCmdNewFolder) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NEWRULE,ConfWindow.onCmdNewRule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_REMOVE,ConfWindow.onCmdRemove) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_TITLE,ConfWindow.onCmdTitle) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTART,ConfWindow.onCmdProxyStart) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTOP,ConfWindow.onCmdProxyStop) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYRESTART,ConfWindow.onCmdProxyRestart) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYRELOAD,ConfWindow.onCmdProxyReload) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYSTATUS,ConfWindow.onCmdProxyStatus) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_CONFUPDATE,ConfWindow.onCmdConfUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DISABLERULE,ConfWindow.onCmdDisableRule) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_ADD,ConfWindow.onCmdNoProxyForAdd) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_EDIT,ConfWindow.onCmdNoProxyForEdit) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_NOPROXYFOR_EDIT,ConfWindow.onUpdNoProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_NOPROXYFOR_REMOVE,ConfWindow.onCmdNoProxyForRemove) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_NOPROXYFOR_REMOVE,ConfWindow.onUpdNoProxy) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_ADD,ConfWindow.onCmdAllowedHostsAdd) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_EDIT,ConfWindow.onCmdAllowedHostsEdit) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_ALLOWEDHOSTS_EDIT,ConfWindow.onUpdAllowedHosts) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_ALLOWEDHOSTS_REMOVE,ConfWindow.onCmdAllowedHostsRemove) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_ALLOWEDHOSTS_REMOVE,ConfWindow.onUpdAllowedHosts) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_UP,ConfWindow.onCmdUp) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_UP,ConfWindow.onCmdUpUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_DOWN,ConfWindow.onCmdDown) FXMAPFUNC(self,SEL_UPDATE, ConfWindow.ID_DOWN,ConfWindow.onCmdDownUpdate) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYUSER,ConfWindow.onCmdProxyUser) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PROXYPASS,ConfWindow.onCmdProxyPass) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYUSER,ConfWindow.onCmdParentProxyUser) FXMAPFUNC(self,SEL_COMMAND,ConfWindow.ID_PARENTPROXYPASS,ConfWindow.onCmdParentProxyPass)
FXLabel(matrix, i18n._("Logfile\tThe name for the logfile can be empty (no logging), '<stdout>'\n(standard out) or a filename (relative or absolute)."), opts=LAYOUT_CENTER_Y|LAYOUT_RIGHT) FXTextField(matrix, 10, self, self.ID_LOGFILE).setText(self.logfile) FXLabel(matrix, i18n._("Debug level"), opts=LAYOUT_CENTER_Y|LAYOUT_RIGHT) cols=0 d = FXComboBox(matrix,0,4,self, self.ID_DEBUGLEVEL,opts=COMBOBOX_INSERT_LAST|FRAME_SUNKEN|FRAME_THICK|LAYOUT_SIDE_TOP) levels = [ i18n._("No debugging"), i18n._("Bring it on"), i18n._("Hurt me plenty"), i18n._("Nightmare"), ] for text in levels: cols = max(len(text), cols) d.appendItem(text) d.setEditable(0) d.setNumColumns(cols) d.setCurrentItem(self.debuglevel)
def proxySettings (self, tabbook): """generate the proxy setting tab""" FXTabItem(tabbook, i18n._("P&roxy Settings"), None) proxy = FXVerticalFrame(tabbook, FRAME_THICK|FRAME_RAISED) proxy_top = FXHorizontalFrame(proxy, LAYOUT_FILL_X|LAYOUT_FILL_Y|LAYOUT_SIDE_TOP)
return 1 def onCmdLogfile (self, sender, sel, ptr): self.logfile = sender.getText() self.getApp().dirty = 1 debug(GUI, "Logfile=%s", self.logfile)
def onCmdParentProxyPass (self, sender, sel, ptr): self.parentproxypass = base64.encodestring(sender.getText()).strip() self.getApp().dirty = 1 debug(GUI, "Parentproxypass was changed") return 1
'debuglevel','logfile',
def readconfig (self): """read the configuration from disc""" debug(GUI, "reading config") self.config = Configuration() for key in ('version','port','parentproxy','parentproxyport', 'debuglevel','logfile',
if self.logfile: self.logfile = self.logfile.name
def readconfig (self): """read the configuration from disc""" debug(GUI, "reading config") self.config = Configuration() for key in ('version','port','parentproxy','parentproxyport', 'debuglevel','logfile',
' debuglevel="%d"\n' % self.debuglevel +\
def toxml (self): s = """<?xml version="1.0"?>
if self.logfile: s += ' logfile="%s"\n' % xmlify(self.logfile)
def toxml (self): s = """<?xml version="1.0"?>
error(GUI, "Authentication with wrong status %d", status)
wc.log.error(wc.LOG_GUI, "Authentication with wrong status %d", status)
def __init__ (self, client, url, form, protocol, clientheaders, status=200, msg=wc.i18n._('Ok'), localcontext=None, auth=''): """load a web configuration template and return response""" wc.log.debug(wc.LOG_GUI, "WebConfig %s %s", url, form) self.client = client # we pretend to be the server self.connected = True headers = wc.proxy.Headers.WcMessage() headers['Server'] = 'Proxy\r' if auth: if status==407: headers['Proxy-Authenticate'] = "%s\r"%auth elif status==401: headers['WWW-Authenticate'] = "%s\r"%auth else: error(GUI, "Authentication with wrong status %d", status) if status in [301,302]: headers['Location'] = clientheaders['Location'] gm = mimetypes.guess_type(url, None) if gm[0] is not None: ctype = gm[0] else: # note: index.html is appended to directories ctype = 'text/html' if ctype=='text/html': ctype += "; charset=iso-8859-1" headers['Content-Type'] = "%s\r"%ctype try: lang = wc.i18n.get_headers_lang(clientheaders) # get the template filename path, dirs, lang = get_template_url(url, lang) if path.endswith('.html'): fp = file(path) # get TAL context context, newstatus = \ get_context(dirs, form, localcontext, lang) if newstatus==401 and status!=newstatus: client.error(401, wc.i18n._("Authentication Required"), auth=wc.proxy.auth.get_challenges()) return # get translator translator = gettext.translation(wc.Name, wc.LocaleDir, [lang], fallback=True) #wc.log.debug(wc.LOG_GUI, "Using translator %s", translator.info()) # expand template data = expand_template(fp, context, translator=translator) else: fp = file(path, 'rb') data = fp.read() fp.close() except IOError: wc.log.exception(GUI, "Wrong path %r", url) # XXX this can actually lead to a maximum recursion # error when client.error caused the exception client.error(404, wc.i18n._("Not Found")) return except StandardError: # catch standard exceptions and report internal error wc.log.exception(GUI, "Template error") client.error(500, wc.i18n._("Internal Error")) return # not catched builtin exceptions are: # SystemExit, StopIteration and all warnings
wc.log.exception(GUI, "Wrong path %r", url)
wc.log.exception(wc.LOG_GUI, "Wrong path %r", url)
def __init__ (self, client, url, form, protocol, clientheaders, status=200, msg=wc.i18n._('Ok'), localcontext=None, auth=''): """load a web configuration template and return response""" wc.log.debug(wc.LOG_GUI, "WebConfig %s %s", url, form) self.client = client # we pretend to be the server self.connected = True headers = wc.proxy.Headers.WcMessage() headers['Server'] = 'Proxy\r' if auth: if status==407: headers['Proxy-Authenticate'] = "%s\r"%auth elif status==401: headers['WWW-Authenticate'] = "%s\r"%auth else: error(GUI, "Authentication with wrong status %d", status) if status in [301,302]: headers['Location'] = clientheaders['Location'] gm = mimetypes.guess_type(url, None) if gm[0] is not None: ctype = gm[0] else: # note: index.html is appended to directories ctype = 'text/html' if ctype=='text/html': ctype += "; charset=iso-8859-1" headers['Content-Type'] = "%s\r"%ctype try: lang = wc.i18n.get_headers_lang(clientheaders) # get the template filename path, dirs, lang = get_template_url(url, lang) if path.endswith('.html'): fp = file(path) # get TAL context context, newstatus = \ get_context(dirs, form, localcontext, lang) if newstatus==401 and status!=newstatus: client.error(401, wc.i18n._("Authentication Required"), auth=wc.proxy.auth.get_challenges()) return # get translator translator = gettext.translation(wc.Name, wc.LocaleDir, [lang], fallback=True) #wc.log.debug(wc.LOG_GUI, "Using translator %s", translator.info()) # expand template data = expand_template(fp, context, translator=translator) else: fp = file(path, 'rb') data = fp.read() fp.close() except IOError: wc.log.exception(GUI, "Wrong path %r", url) # XXX this can actually lead to a maximum recursion # error when client.error caused the exception client.error(404, wc.i18n._("Not Found")) return except StandardError: # catch standard exceptions and report internal error wc.log.exception(GUI, "Template error") client.error(500, wc.i18n._("Internal Error")) return # not catched builtin exceptions are: # SystemExit, StopIteration and all warnings
wc.log.exception(GUI, "Template error")
wc.log.exception(wc.LOG_GUI, "Template error")
def __init__ (self, client, url, form, protocol, clientheaders, status=200, msg=wc.i18n._('Ok'), localcontext=None, auth=''): """load a web configuration template and return response""" wc.log.debug(wc.LOG_GUI, "WebConfig %s %s", url, form) self.client = client # we pretend to be the server self.connected = True headers = wc.proxy.Headers.WcMessage() headers['Server'] = 'Proxy\r' if auth: if status==407: headers['Proxy-Authenticate'] = "%s\r"%auth elif status==401: headers['WWW-Authenticate'] = "%s\r"%auth else: error(GUI, "Authentication with wrong status %d", status) if status in [301,302]: headers['Location'] = clientheaders['Location'] gm = mimetypes.guess_type(url, None) if gm[0] is not None: ctype = gm[0] else: # note: index.html is appended to directories ctype = 'text/html' if ctype=='text/html': ctype += "; charset=iso-8859-1" headers['Content-Type'] = "%s\r"%ctype try: lang = wc.i18n.get_headers_lang(clientheaders) # get the template filename path, dirs, lang = get_template_url(url, lang) if path.endswith('.html'): fp = file(path) # get TAL context context, newstatus = \ get_context(dirs, form, localcontext, lang) if newstatus==401 and status!=newstatus: client.error(401, wc.i18n._("Authentication Required"), auth=wc.proxy.auth.get_challenges()) return # get translator translator = gettext.translation(wc.Name, wc.LocaleDir, [lang], fallback=True) #wc.log.debug(wc.LOG_GUI, "Using translator %s", translator.info()) # expand template data = expand_template(fp, context, translator=translator) else: fp = file(path, 'rb') data = fp.read() fp.close() except IOError: wc.log.exception(GUI, "Wrong path %r", url) # XXX this can actually lead to a maximum recursion # error when client.error caused the exception client.error(404, wc.i18n._("Not Found")) return except StandardError: # catch standard exceptions and report internal error wc.log.exception(GUI, "Template error") client.error(500, wc.i18n._("Internal Error")) return # not catched builtin exceptions are: # SystemExit, StopIteration and all warnings
if data: p.feed(data)
p.feed(data)
def finish (self, data, **attrs): if not attrs.has_key('filter'): return data p = attrs['filter'] if data: p.feed(data) p.flush() p.buf2data() return p.flushbuf()
def _buf2data (buf, outbuf): """Append all tags of the buffer to the data""" for item in buf: if item[0]==DATA: outbuf.write(item[1]) elif item[0]==STARTTAG: s = "<"+item[1] for name,val in item[2].items(): s += ' %s'%name if val: s += "=%s"%val outbuf.write(s+">") elif item[0]==ENDTAG: outbuf.write("</%s>"%item[1]) elif item[0]==COMMENT: outbuf.write("<!--%s-->"%item[1]) else: error("unknown buffer element %s" % item[0])
def _buf2data (buf, outbuf): """Append all tags of the buffer to the data""" for item in buf: if item[0]==DATA: outbuf.write(item[1]) elif item[0]==STARTTAG: s = "<"+item[1] for name,val in item[2].items(): s += ' %s'%name if val: s += "=%s"%val outbuf.write(s+">") elif item[0]==ENDTAG: outbuf.write("</%s>"%item[1]) elif item[0]==COMMENT: outbuf.write("<!--%s-->"%item[1]) else: error("unknown buffer element %s" % item[0])
self.waitbuf = StringIO()
self.waitbuf = []
def __init__ (self, rules, url, **opts): if wc.config['showerrors']: self.error = self._error self.warning = self._warning self.fatalError = self._fatalError HtmlParser.__init__(self) self.rules = rules self.comments = opts['comments'] self.javascript = opts['javascript'] self.outbuf = StringIO() self.inbuf = StringIO() self.waitbuf = StringIO() self.state = 'parse' self.script = '' self.waited = 0 self.rulestack = [] self.buf = [] self.url = url or "unknown" if self.javascript: self.jsEnv = jslib.new_jsenv() self.output_counter = 0 self.popup_counter = 0
data = "%s%s" % (self.waitbuf.getvalue(), self.inbuf.getvalue(), )
self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': return data = self.inbuf.getvalue()
def feed (self, data): if self.state=='parse': if self.waited: data = "%s%s" % (self.waitbuf.getvalue(), self.inbuf.getvalue(), ) self.inbuf.close() self.waitbuf.close() self.inbuf = StringIO() self.waitbuf = StringIO() self.waited = 0 #debug(NIGHTMARE, "HtmlFilter: feed", `data`) HtmlParser.feed(self, data) #debug(NIGHTMARE, "HtmlFilter: feed finished") else: self.inbuf.write(data)
self.waitbuf.close()
def feed (self, data): if self.state=='parse': if self.waited: data = "%s%s" % (self.waitbuf.getvalue(), self.inbuf.getvalue(), ) self.inbuf.close() self.waitbuf.close() self.inbuf = StringIO() self.waitbuf = StringIO() self.waited = 0 #debug(NIGHTMARE, "HtmlFilter: feed", `data`) HtmlParser.feed(self, data) #debug(NIGHTMARE, "HtmlFilter: feed finished") else: self.inbuf.write(data)
self.waitbuf = StringIO() self.waited = 0 HtmlParser.feed(self, data)
if data: debug(NIGHTMARE, "HtmlFilter: feed", `data`) HtmlParser.feed(self, data) debug(NIGHTMARE, "HtmlFilter: feed finished")
def feed (self, data): if self.state=='parse': if self.waited: data = "%s%s" % (self.waitbuf.getvalue(), self.inbuf.getvalue(), ) self.inbuf.close() self.waitbuf.close() self.inbuf = StringIO() self.waitbuf = StringIO() self.waited = 0 #debug(NIGHTMARE, "HtmlFilter: feed", `data`) HtmlParser.feed(self, data) #debug(NIGHTMARE, "HtmlFilter: feed finished") else: self.inbuf.write(data)
_buf2data(self.buf, self.outbuf)
for item in self.buf: if item[0]==DATA: self.outbuf.write(item[1]) elif item[0]==STARTTAG: s = "<"+item[1] for name,val in item[2].items(): s += ' %s'%name if val: s += "=%s"%val self.outbuf.write(s+">") elif item[0]==ENDTAG: self.outbuf.write("</%s>"%item[1]) elif item[0]==COMMENT: self.outbuf.write("<!--%s-->"%item[1]) else: error("unknown buffer element %s" % item[0])
def buf2data (self): _buf2data(self.buf, self.outbuf) self.buf = []
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def cdata (self, d): """handler for data""" item = [DATA, d] if self.state=='wait': return _buf2data([item], self.waitbuf) self.buf_append_data(item)
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def comment (self, data): """a comment; accept only non-empty comments""" item = [COMMENT, data] if self.state=='wait': return _buf2data([item], self.waitbuf) if self.comments and data: self.buf.append(item)
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def characters (self, s): """handler for characters""" item = [DATA, s] if self.state=='wait': return _buf2data([item], self.waitbuf) self.buf_append_data(item)
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def doctype (self, data): item = [DATA, "<!DOCTYPE%s>"%data] if self.state=='wait': return _buf2data([item], self.waitbuf) self.buf_append_data(item)
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def pi (self, data): item = [DATA, "<?%s?>"%data] if self.state=='wait': return _buf2data([item], self.waitbuf) self.buf_append_data(item)
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data item = [STARTTAG, tag, attrs] if self.state=='wait': return _buf2data([item], self.waitbuf) rulelist = [] filtered = 0 # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): debug(NIGHTMARE, "HtmlFilter: matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" # give'em a chance to replace more than one attribute if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item continue else: break else: debug(NIGHTMARE, "HtmlFilter: put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.javascript: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.javascript: self.buf2data()
return _buf2data([item], self.waitbuf)
return self.waitbuf.append(item)
def endElement (self, tag): """We know the following: if a rule matches, it must be the one on the top of the stack. So we look only at the top rule.
if not filtered and self.javascript and tag=='script' and \ self.jsEndElement(tag): del self.buf[-1] del self.buf[-1] else: self.buf.append(item)
if not filtered: if self.javascript and tag=='script' and \ self.jsEndElement(tag): del self.buf[-1] del self.buf[-1] else: self.buf.append(item)
def endElement (self, tag): """We know the following: if a rule matches, it must be the one on the top of the stack. So we look only at the top rule.
elif self.jsScript(self.script, ver): print >> sys.stderr, "JS popup src", url
def jsScriptData (self, data, url, ver): assert self.state=='wait' if data is None: if not self.script: print >> sys.stderr, "empty JS src", url elif self.jsScript(self.script, ver): print >> sys.stderr, "JS popup src", url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) self.buf.append([DATA, "<!--\n%s\n//-->"%self.script]) self.state = 'parse' self.script = '' debug(NIGHTMARE, "XXX switching back to parse with") debug(NIGHTMARE, "self.buf", `self.buf`) #debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) debug(NIGHTMARE, "self.waitbuf", `self.waitbuf.getvalue()`) else: debug(HURT_ME_PLENTY, "JS: read", len(data)) self.script += data
debug(NIGHTMARE, "XXX switching back to parse with") debug(NIGHTMARE, "self.buf", `self.buf`) debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) debug(NIGHTMARE, "self.waitbuf", `self.waitbuf.getvalue()`)
debug(NIGHTMARE, "Filter: switching back to parse with") debug(NIGHTMARE, "Filter: self.buf", `self.buf`) debug(NIGHTMARE, "Filter: self.waitbuf", `self.waitbuf`) debug(NIGHTMARE, "Filter: self.inbuf", `self.inbuf.getvalue()`)
def jsScriptData (self, data, url, ver): assert self.state=='wait' if data is None: if not self.script: print >> sys.stderr, "empty JS src", url elif self.jsScript(self.script, ver): print >> sys.stderr, "JS popup src", url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) self.buf.append([DATA, "<!--\n%s\n//-->"%self.script]) self.state = 'parse' self.script = '' debug(NIGHTMARE, "XXX switching back to parse with") debug(NIGHTMARE, "self.buf", `self.buf`) #debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) debug(NIGHTMARE, "self.waitbuf", `self.waitbuf.getvalue()`) else: debug(HURT_ME_PLENTY, "JS: read", len(data)) self.script += data
debug(HURT_ME_PLENTY, "JS: read", len(data))
debug(HURT_ME_PLENTY, "JS: read", len(data), "<=", url)
def jsScriptData (self, data, url, ver): assert self.state=='wait' if data is None: if not self.script: print >> sys.stderr, "empty JS src", url elif self.jsScript(self.script, ver): print >> sys.stderr, "JS popup src", url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) self.buf.append([DATA, "<!--\n%s\n//-->"%self.script]) self.state = 'parse' self.script = '' debug(NIGHTMARE, "XXX switching back to parse with") debug(NIGHTMARE, "self.buf", `self.buf`) #debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) debug(NIGHTMARE, "self.waitbuf", `self.waitbuf.getvalue()`) else: debug(HURT_ME_PLENTY, "JS: read", len(data)) self.script += data
debug(HURT_ME_PLENTY, "JS: self.url", self.url) debug(HURT_ME_PLENTY, "JS: url", url)
def jsScriptSrc (self, url, language): assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) debug(HURT_ME_PLENTY, "JS: self.url", self.url) debug(HURT_ME_PLENTY, "JS: url", url) url = urlparse.urljoin(self.url, url) debug(HURT_ME_PLENTY, "JS: jsScriptSrc", url, ver) self.state = 'wait' client = HttpProxyClient(self.jsScriptData, (url, ver)) ClientServerMatchmaker(client, "GET %s HTTP/1.1" % url, #request {}, #headers '', #content {'nofilter': None}, 'identity', # compress ) self.waited = 1
self.waited = 1
self.waited = "True"
def jsScriptSrc (self, url, language): assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) debug(HURT_ME_PLENTY, "JS: self.url", self.url) debug(HURT_ME_PLENTY, "JS: url", url) url = urlparse.urljoin(self.url, url) debug(HURT_ME_PLENTY, "JS: jsScriptSrc", url, ver) self.state = 'wait' client = HttpProxyClient(self.jsScriptData, (url, ver)) ClientServerMatchmaker(client, "GET %s HTTP/1.1" % url, #request {}, #headers '', #content {'nofilter': None}, 'identity', # compress ) self.waited = 1
self.buf += self.jsfilter.buf
self.buf[-2:-2] = self.jsfilter.buf
def jsScript (self, script, ver): """execute given script with javascript version ver return True if the script generates any output, else False""" #debug(HURT_ME_PLENTY, "JS: jsScript", ver, `script`) self.output_counter = 0 self.jsEnv.attachListener(self) self.jsfilter = HtmlFilter(self.rules, self.url, comments=self.comments, javascript=self.javascript) self.jsEnv.executeScript(script, ver) self.jsEnv.detachListener(self) if self.output_counter: self.jsfilter.flush() self.outbuf.write(self.jsfilter.flushbuf()) self.buf += self.jsfilter.buf self.rulestack += self.jsfilter.rulestack self.jsfilter = None return self.output_counter
return self.output_counter
return self.popup_counter + self.output_counter
def jsScript (self, script, ver): """execute given script with javascript version ver return True if the script generates any output, else False""" #debug(HURT_ME_PLENTY, "JS: jsScript", ver, `script`) self.output_counter = 0 self.jsEnv.attachListener(self) self.jsfilter = HtmlFilter(self.rules, self.url, comments=self.comments, javascript=self.javascript) self.jsEnv.executeScript(script, ver) self.jsEnv.detachListener(self) if self.output_counter: self.jsfilter.flush() self.outbuf.write(self.jsfilter.flushbuf()) self.buf += self.jsfilter.buf self.rulestack += self.jsfilter.rulestack self.jsfilter = None return self.output_counter
self.fd.write(" %s=\"%s\"" % (key, quote_attrval(val)))
if val is None: self.fd.write(" %s"%key) else: self.fd.write(" %s=\"%s\"" % (key, quote_attrval(val)))
def startElement (self, tag, attrs): self.fd.write("<%s"%tag) for key, val in attrs.iteritems(): self.fd.write(" %s=\"%s\"" % (key, quote_attrval(val))) self.fd.write(">")
self.min_bufsize = 256
self.min_bufsize = 6000
def __init__ (self, mimelist): super(ImageSize, self).__init__(mimelist) # minimal amount of image data for PIL to read header info self.min_bufsize = 256
if attrs['buffer'].closed:
if not attrs.has_key('buffer') or attrs['buffer'].closed:
def filter (self, data, **attrs): if attrs['buffer'].closed: # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() return data return ''
if self.check_sizes(buf):
if self.check_sizes(buf, attrs['sizes']):
def filter (self, data, **attrs): if attrs['buffer'].closed: # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() return data return ''
pass
if not attrs.has_key('buffer') or attrs['buffer'].closed: return data buf = attrs['buffer'] buf.write(data) if self.check_sizes(buf, attrs['sizes']): data = buf.getvalue() buf.close() return data return ''
def finish (self, data, **attrs): # XXX pass
def check_sizes (self, buf):
def check_sizes (self, buf, sizes):
def check_sizes (self, buf): try: img = Image.open(buf, 'r') for size, formats in sizes: if size==img.size: # size matches, look for format restriction if not formats: return False elif img.format.lower() in formats: return False except IOError: # XXX print error pass return True
pass
exception(FILTER, "Could not get image size")
def check_sizes (self, buf): try: img = Image.open(buf, 'r') for size, formats in sizes: if size==img.size: # size matches, look for format restriction if not formats: return False elif img.format.lower() in formats: return False except IOError: # XXX print error pass return True
try: self.addr = sock.getpeername() except socket.error: pass
self.addr = sock.getpeername()
def __init__ (self, sock=None, socktype=None): """ Initialize connection.
def _debug (self, msg, *args): """debug with recursion level and state""" debug(FILTER, "%s[%d]: %s"%(self.__class__.__name__, self.level, msg), *args)
def _str__ (self): return "%s[%d]" % (self.__class__.__name__, self.level)
def _debug (self, msg, *args): """debug with recursion level and state""" debug(FILTER, "%s[%d]: %s"%(self.__class__.__name__, self.level, msg), *args)
self._debug("JS document.write %r", data)
debug(FILTER, "%s jsProcessData %r", self, data)
def jsProcessData (self, data): """process data produced by document.write() JavaScript""" self._debug("JS document.write %r", data) self.js_output += 1 # parse recursively self.js_htmlparser.feed(data)
self._debug("JS: popup")
debug(FILTER, "%s jsProcessPopup", self)
def jsProcessPopup (self): """process javascript popup""" self._debug("JS: popup") self.js_popup += 1
def jsForm (self, name, action, target): """when hitting a named form, add it to the JS environment""" if not name: return self._debug("jsForm %r action %r %r", name, action, target) self.js_env.addForm(name, action, target)
def jsForm (self, name, action, target): """when hitting a named form, add it to the JS environment""" if not name: return self._debug("jsForm %r action %r %r", name, action, target) self.js_env.addForm(name, action, target)
self._debug("JS: jsPopup")
debug(FILTER, "%s jsPopup", self)
def jsPopup (self, attrs, name): """check if attrs[name] javascript opens a popup window""" self._debug("JS: jsPopup") val = resolve_html_entities(attrs[name]) if not val: return self.js_env.listeners.append(self) try: self.js_env.executeScriptAsFunction(val, 0.0) except jslib.error: pass self.js_env.listeners.remove(self) res, self.js_popup = self.js_popup, 0 return res
def new_instance (self, opts):
def new_instance (self, **opts):
def new_instance (self, opts): JSFilter(self.url, **opts)
self._debug("JS: jsScript %s %r", ver, script) assert self.htmlparser.state[0]=='parse', "non-parse state" assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf
debug(FILTER, "%s jsScript %s %r", self, ver, script) assert self.htmlparser.state[0]=='parse', "parser %s not in parse state" % self.htmlparser assert len(self.htmlparser.tagbuf) >= 2, "parser %s must have script start and content tags in tag buffer" % self.htmlparser
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug("JS: jsScript %s %r", ver, script) assert self.htmlparser.state[0]=='parse', "non-parse state" assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf self.js_output = 0 self.js_env.listeners.append(self) # start recursive html filter (used by jsProcessData) opts = dict(comments=self.comments, javascript=self.javascript, level=self.level+1) handler = self.new_instance(opts) self.js_htmlparser = HtmlParser(handler) handler.htmlparser = self.js_htmlparser # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.listeners.remove(self) # wait for recursive filter to finish self.jsEndScript(item)
opts = dict(comments=self.comments, javascript=self.javascript, level=self.level+1) handler = self.new_instance(opts)
handler = self.new_instance(comments=self.comments, javascript=self.javascript, level=self.level+1)
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug("JS: jsScript %s %r", ver, script) assert self.htmlparser.state[0]=='parse', "non-parse state" assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf self.js_output = 0 self.js_env.listeners.append(self) # start recursive html filter (used by jsProcessData) opts = dict(comments=self.comments, javascript=self.javascript, level=self.level+1) handler = self.new_instance(opts) self.js_htmlparser = HtmlParser(handler) handler.htmlparser = self.js_htmlparser # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.listeners.remove(self) # wait for recursive filter to finish self.jsEndScript(item)
debug(FILTER, "%s jsScriptSrc %r", self, url)
def jsScriptSrc (self, url, language): """Start a background download for <script src=""> tags""" assert self.htmlparser.state[0]=='parse', "non-parse state %s" % self.htmlparser.state ver = get_js_ver(language) if self.base_url: url = urlparse.urljoin(self.base_url, url) else: url = urlparse.urljoin(self.url, url) # unquote and norm url = url_norm(url) host = stripsite(url)[0] if not host: error(FILTER, "invalid script src url %r at %s (base %r)", url, self.url, self.base_url) return self.htmlparser.state = ('wait', url) self.htmlparser.waited = 1 self.js_src = True self.js_client = HttpProxyClient(self.jsScriptData, (url, ver)) headers = get_wc_client_headers(host) # note: some javascript servers do not specify content encoding # so only accept non-encoded content here headers['Accept-Encoding'] = 'identity\r' ClientServerMatchmaker(self.js_client, "GET %s HTTP/1.0" % url_quote(url), # request headers, '', # content mime="application/x-javascript", )
warn(FILTER, "HtmlParser[%d]: empty JS src %s", self.level, url)
warn(FILTER, "empty JavaScript src %s", url)
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0]=='wait', "non-wait state" if data is None: if not self.js_script: warn(FILTER, "HtmlParser[%d]: empty JS src %s", self.level, url) self.js_script = "// error fetching script from %r"%url self.htmlparser.tagbuf.append([STARTTAG, "script", {'type': 'text/javascript'}]) script = "\n<!--\n%s\n//-->\n"%escape_js(self.js_script) self.htmlparser.tagbuf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.htmlparser.tagbuf.append([ENDTAG, "script"]) self.js_script = '' self.htmlparser.state = ('parse',) self._debug("switching back to parse") else: self._debug("JS read %d <= %s", len(data), url) self.js_script += data
self._debug("switching back to parse")
debug(FILTER, "%s switching back to parse with", self) self.htmlparser.debugbuf()
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0]=='wait', "non-wait state" if data is None: if not self.js_script: warn(FILTER, "HtmlParser[%d]: empty JS src %s", self.level, url) self.js_script = "// error fetching script from %r"%url self.htmlparser.tagbuf.append([STARTTAG, "script", {'type': 'text/javascript'}]) script = "\n<!--\n%s\n//-->\n"%escape_js(self.js_script) self.htmlparser.tagbuf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.htmlparser.tagbuf.append([ENDTAG, "script"]) self.js_script = '' self.htmlparser.state = ('parse',) self._debug("switching back to parse") else: self._debug("JS read %d <= %s", len(data), url) self.js_script += data
self._debug("JS read %d <= %s", len(data), url)
debug(FILTER, "JS read %d <= %s", len(data), url)
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0]=='wait', "non-wait state" if data is None: if not self.js_script: warn(FILTER, "HtmlParser[%d]: empty JS src %s", self.level, url) self.js_script = "// error fetching script from %r"%url self.htmlparser.tagbuf.append([STARTTAG, "script", {'type': 'text/javascript'}]) script = "\n<!--\n%s\n//-->\n"%escape_js(self.js_script) self.htmlparser.tagbuf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.htmlparser.tagbuf.append([ENDTAG, "script"]) self.js_script = '' self.htmlparser.state = ('parse',) self._debug("switching back to parse") else: self._debug("JS read %d <= %s", len(data), url) self.js_script += data
def jsEndScript (self, item): self._debug("JS: endScript") assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf if self.js_output: try: self.js_htmlparser.feed('') data = self.js_htmlparser.flush() except FilterWait: self._debug("JS: subprocessor is waiting") self.htmlparser.state = ('wait', 'recursive script') self.htmlparser.waited = 1 make_timer(0.2, lambda : self.jsEndScript(item)) return self.js_htmlparser._debugbuf() assert not self.js_htmlparser.inbuf.getvalue() assert not self.js_htmlparser.waitbuf assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf self.htmlparser.tagbuf[-2:-2] = [[DATA, data]]+self.js_htmlparser.tagbuf self.js_htmlparser = None if (self.js_popup + self.js_output) > 0: del self.htmlparser.tagbuf[-1] del self.htmlparser.tagbuf[-1] elif not self._filterEndElement(item[1]): self.htmlparser.tagbuf.append(item) self._debug("JS: switching back to parse with") self.htmlparser.state = ('parse',) def _jsEndElement (self, item): """parse generated html for scripts""" self._debug("jsEndElement buf %r", self.htmlparser.tagbuf) if len(self.htmlparser.tagbuf)<2: warn(FILTER, "JS syntax error, self.tagbuf %r", self.htmlparser.tagbuf) return if self.js_src: debug(FILTER, "JS src, self.tagbuf %r", self.htmlparser.tagbuf) del self.htmlparser.tagbuf[-1] if len(self.htmlparser.tagbuf)<2: warn(FILTER, "JS end, self.tagbuf %s", self.htmlparser.tagbuf) return if len(self.htmlparser.tagbuf) > 2 and \ self.htmlparser.tagbuf[-3][0]==STARTTAG and \ self.htmlparser.tagbuf[-3][1]=='script': del self.htmlparser.tagbuf[-1] if len(self.htmlparser.tagbuf)<2 or \ self.htmlparser.tagbuf[-1][0]!=DATA or \ self.htmlparser.tagbuf[-2][0]!=STARTTAG or \ self.htmlparser.tagbuf[-2][1]!='script': return js_ok, js_lang = get_js_data(self.htmlparser.tagbuf[-2][2]) if not js_ok: self.htmlparser.tagbuf.append(item) return ver = get_js_ver(js_lang) script = self.htmlparser.tagbuf[-1][1].strip() mo = _start_js_comment(script) if mo: script = script[mo.end():] mo = _end_js_comment(script) if mo: script = script[:mo.start()] script = script.strip() if not script: del self.htmlparser.tagbuf[-1] del self.htmlparser.tagbuf[-1] return self.htmlparser.tagbuf[-1][1] = "\n<!--\n%s\n//-->\n"%escape_js(script) self.jsScript(script, ver, item) def _jsStartElement (self, tag, attrs): """Check popups for onmouseout and onmouseover. Inline extern javascript sources""" self.js_src = False self.js_output = 0 self.js_popup = 0 for name in ('onmouseover', 'onmouseout'): if attrs.has_key(name) and self.jsPopup(attrs, name): self._debug("JS: del %r from %r", name, tag) del attrs[name] if tag=='form': name = attrs.get('name', attrs.get('id')) self.jsForm(name, attrs.get('action', ''), attrs.get('target', '')) elif tag=='script': js_ok, js_lang = get_js_data(attrs) url = attrs.get('src', '') if js_ok and url: self.jsScriptSrc(url, js_lang) return self.htmlparser.tagbuf.append([STARTTAG, tag, attrs])
def jsEndScript (self, item): self._debug("JS: endScript") assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf if self.js_output: try: self.js_htmlparser.feed('') data = self.js_htmlparser.flush() except FilterWait: self._debug("JS: subprocessor is waiting") self.htmlparser.state = ('wait', 'recursive script') self.htmlparser.waited = 1 make_timer(0.2, lambda : self.jsEndScript(item)) return self.js_htmlparser._debugbuf() assert not self.js_htmlparser.inbuf.getvalue() assert not self.js_htmlparser.waitbuf assert len(self.htmlparser.tagbuf) >= 2, "too small buffer %s" % self.htmlparser.tagbuf self.htmlparser.tagbuf[-2:-2] = [[DATA, data]]+self.js_htmlparser.tagbuf self.js_htmlparser = None if (self.js_popup + self.js_output) > 0: # delete old script del self.htmlparser.tagbuf[-1] del self.htmlparser.tagbuf[-1] elif not self._filterEndElement(item[1]): self.htmlparser.tagbuf.append(item) self._debug("JS: switching back to parse with") self.htmlparser.state = ('parse',)
_in, _out = os.popen4(cc_cmd) _in.write(prog) _in.close() while _out.read(): pass return _out.close() is None
pipe = popen2.Popen4(cc_cmd) pipe.tochild.write(prog) pipe.tochild.close() status = pipe.wait() if os.WIFEXITED(status): return os.WEXITSTATUS(status)==0 return False
def cc_supports_option (cc, option): prog = "int main(){}\n" cc_cmd = "%s -E %s -" % (cc[0], option) _in, _out = os.popen4(cc_cmd) _in.write(prog) _in.close() while _out.read(): pass return _out.close() is None
'wc.webgui.PageTemplates', 'wc.webgui.TAL', 'wc.webgui.ZTUtils',
'wc.webgui.pagetemplate', 'wc.webgui.tal', 'wc.webgui.tales',
def run (self): if self.all: # remove share directory directory = os.path.join("build", "share") if os.path.exists(directory): remove_tree(directory, dry_run=self.dry_run) else: log.warn("'%s' does not exist -- can't clean it", directory) clean.run(self)
chals = [get_digest_challenge(), get_basic_challenge(), ]
chals = [ get_basic_challenge(), get_digest_challenge(), ]
def get_challenges (**args): """ Return list of challenges for WebCleaner proxy authentication Note that HTTP/1.1 allows multiple authentication challenges either as multiple headers with the same key, or as one single header whose value list is separated by commas. """ if wc.configuration.config['auth_ntlm'] and wc.HasCrypto: chals = [get_ntlm_challenge(**args)] else: chals = [get_digest_challenge(), get_basic_challenge(), ] wc.log.debug(wc.LOG_AUTH, "challenges %s", chals) return chals
file_logs=self.file_logs)
filelogs=self.filelogs)
def SvcDoRun (self): """start this service""" import servicemanager # Log a "started" message to the event log. servicemanager.LogMsg( servicemanager.EVENTLOG_INFORMATION_TYPE, servicemanager.PYS_SERVICE_STARTED, (self._svc_name_, '')) wc.wstartfunc(handle=self.hWaitStop, confdir=self.configdir, file_logs=self.file_logs) # Now log a "service stopped" message servicemanager.LogMsg( servicemanager.EVENTLOG_INFORMATION_TYPE, servicemanager.PYS_SERVICE_STOPPED, (self._svc_name_,''))
self._debug(NIGHTMARE, "XXX", self.buf)
self._debug(NIGHTMARE, "jsEndElement buf", self.buf)
def jsEndElement (self, item): """parse generated html for scripts""" self._debug(NIGHTMARE, "XXX", self.buf) if len(self.buf)<2: # syntax error, ignore return if self.js_src: del self.buf[-1] if self.buf[-3][0]==STARTTAG and self.buf[-3][1]=='script': del self.buf[-1] if len(self.buf)<2 or self.buf[-1][0]!=DATA or \ self.buf[-2][0]!=STARTTAG or self.buf[-2][1]!='script': # syntax error, ignore return # get script data script = self.buf[-1][1].strip() # remove html comments if script.startswith("<!--"): i = script.find('\n') if i==-1: script = script[4:] else: script = script[(i+1):] if script.endswith("-->"): script = script[:-3] if not script: # again, ignore an empty script del self.buf[-1] del self.buf[-1] else: self.jsScript(script, 0.0, item)
self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`)
def _debugbuf (self): """print debugging information about buffer status""" self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`)
self._debug(NIGHTMARE, "feed", `data`)
def feed (self, data): if self.state=='parse': if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': return data = self.inbuf.getvalue() self.inbuf.close() self.inbuf = StringIO() if data: self._debug(NIGHTMARE, "feed", `data`) HtmlParser.feed(self, data) else: self._debug(NIGHTMARE, "feed") else: self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
self._debug(NIGHTMARE, "feed")
def feed (self, data): if self.state=='parse': if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': return data = self.inbuf.getvalue() self.inbuf.close() self.inbuf = StringIO() if data: self._debug(NIGHTMARE, "feed", `data`) HtmlParser.feed(self, data) else: self._debug(NIGHTMARE, "feed") else: self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
self._debug(NIGHTMARE, "wait")
def feed (self, data): if self.state=='parse': if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': return data = self.inbuf.getvalue() self.inbuf.close() self.inbuf = StringIO() if data: self._debug(NIGHTMARE, "feed", `data`) HtmlParser.feed(self, data) else: self._debug(NIGHTMARE, "feed") else: self._debug(NIGHTMARE, "wait") self.inbuf.write(data)
self._debug(HURT_ME_PLENTY, "flush")
def flush (self): self._debug(HURT_ME_PLENTY, "flush") if self.state=='wait': raise FilterException("HtmlFilter[%d]: still waiting for data"%self.level) HtmlParser.flush(self)
self._debug(NIGHTMARE, "buf_append_data")
def buf_append_data (self, data): """we have to make sure that we have no two following DATA things in the buffer. Why? To be 100% sure that an ENCLOSED match really matches enclosed data. """ self._debug(NIGHTMARE, "buf_append_data") if data[0]==DATA and self.buf and self.buf[-1][0]==DATA: self.buf[-1][1] += data[1] else: self.buf.append(data)
self._debug(NIGHTMARE, "flushbuf")
def flushbuf (self): """flush internal data buffer""" self._debug(NIGHTMARE, "flushbuf") data = self.outbuf.getvalue() self.outbuf.close() self.outbuf = StringIO() return data
self._debug(NIGHTMARE, "startElement", `tag`)
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" # give'em a chance to replace more than one attribute if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`))
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" # give'em a chance to replace more than one attribute if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
self._debug(NIGHTMARE, "put on buffer")
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0 # look for filter rules which apply for rule in self.rules: if rule.match_tag(tag) and rule.match_attrs(attrs): self._debug(NIGHTMARE, "matched rule %s on tag %s" % (`rule.title`, `tag`)) if rule.start_sufficient: item = rule.filter_tag(tag, attrs) filtered = "True" # give'em a chance to replace more than one attribute if item[0]==STARTTAG and item[1]==tag: foo,tag,attrs = item continue else: break else: self._debug(NIGHTMARE, "put on buffer") rulelist.append(rule) if rulelist: # remember buffer position for end tag matching pos = len(self.buf) self.rulestack.append((pos, rulelist)) if filtered: self.buf_append_data(item) elif self.js_filter: # if its not yet filtered, try filter javascript self.jsStartElement(tag, attrs) else: self.buf.append(item) # if rule stack is empty, write out the buffered data if not self.rulestack and not self.js_filter: self.buf2data()
self._debug(NIGHTMARE, "JS: jsStartElement")
def jsStartElement (self, tag, attrs): """Check popups for onmouseout and onmouseover. Inline extern javascript sources""" self._debug(NIGHTMARE, "JS: jsStartElement") changed = 0 self.js_src = 0 self.js_output = 0 self.js_popup = 0 for name in ('onmouseover', 'onmouseout'): if attrs.has_key(name) and self.jsPopup(attrs, name): del attrs[name] changed = 1 if tag=='form': name = attrs.get('name', attrs.get('id')) self.jsForm(name, attrs.get('action', ''), attrs.get('target', '')) elif tag=='script': lang = attrs.get('language', '').lower() url = attrs.get('src', '') scrtype = attrs.get('type', '').lower() is_js = scrtype=='text/javascript' or \ lang.startswith('javascript') or \ not (lang or scrtype) if is_js and url: return self.jsScriptSrc(url, lang) self.buf.append([STARTTAG, tag, attrs])
self._debug(NIGHTMARE, "switching back to parse with")
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlFilter[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) if self.js_script.find("<!--")==-1: script = "<!--\n%s\n//-->"%self.js_script else: script = self.js_script self.buf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.buf.append([ENDTAG, "script"]) self.state = 'parse' self.js_script = '' self._debug(NIGHTMARE, "switching back to parse with") self._debugbuf() else: self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url) self.js_script += data
self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url)
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlFilter[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "script", {'type': 'text/javascript'}]) if self.js_script.find("<!--")==-1: script = "<!--\n%s\n//-->"%self.js_script else: script = self.js_script self.buf.append([DATA, script]) # Note: <script src=""> could be missing an end tag, # but now we need one. Look later for a duplicate </script>. self.buf.append([ENDTAG, "script"]) self.state = 'parse' self.js_script = '' self._debug(NIGHTMARE, "switching back to parse with") self._debugbuf() else: self._debug(HURT_ME_PLENTY, "JS read", len(data), "<=", url) self.js_script += data
self._debug(HURT_ME_PLENTY, "JS jsScriptSrc", url, ver)
def jsScriptSrc (self, url, language): """Start a background download for <script src=""> tags""" assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) url = urlparse.urljoin(self.url, url) self._debug(HURT_ME_PLENTY, "JS jsScriptSrc", url, ver) self.state = 'wait' self.js_src = 'True' client = HttpProxyClient(self.jsScriptData, (url, ver)) ClientServerMatchmaker(client, "GET %s HTTP/1.1" % url, #request {}, #headers '', #content {'nofilter': None}, 'identity', # compress ) self.waited = "True"
self._debug(NIGHTMARE, "JS: jsScript", ver, `script`)
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = HtmlFilter(self.rules, self.url, comments=self.comments, javascript=self.js_filter, level=self.level+1) # execute self.js_env.executeScript(script, ver) self.js_env.detachListener(self) # wait for recursive filter to finish self.jsEndScript(item)
self._debug(NIGHTMARE, "JS: endScript")
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterException: self.state = 'wait' self.waited = "True" make_timer(0.1, lambda : HtmlFilter.jsEndScript(self, item)) return self.js_html._debugbuf() assert not self.js_html.inbuf.getvalue() assert not self.js_html.waitbuf assert len(self.buf) >= 2 self.buf[-2:-2] = [[DATA, self.js_html.outbuf.getvalue()]]+self.js_html.buf self.js_html = None if (self.js_popup + self.js_output) > 0: # delete old script del self.buf[-1] del self.buf[-1] else: self.buf.append(item) self._debug(NIGHTMARE, "JS: switching back to parse with") self._debugbuf() self.state = 'parse'
self._debug(NIGHTMARE, "JS: switching back to parse with")
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterException: self.state = 'wait' self.waited = "True" make_timer(0.1, lambda : HtmlFilter.jsEndScript(self, item)) return self.js_html._debugbuf() assert not self.js_html.inbuf.getvalue() assert not self.js_html.waitbuf assert len(self.buf) >= 2 self.buf[-2:-2] = [[DATA, self.js_html.outbuf.getvalue()]]+self.js_html.buf self.js_html = None if (self.js_popup + self.js_output) > 0: # delete old script del self.buf[-1] del self.buf[-1] else: self.buf.append(item) self._debug(NIGHTMARE, "JS: switching back to parse with") self._debugbuf() self.state = 'parse'
self._debug(NIGHTMARE, "jsEndElement buf", self.buf)
def jsEndElement (self, item): """parse generated html for scripts""" self._debug(NIGHTMARE, "jsEndElement buf", self.buf) if len(self.buf)<2: # syntax error, ignore return if self.js_src: del self.buf[-1] if self.buf[-3][0]==STARTTAG and self.buf[-3][1]=='script': del self.buf[-1] if len(self.buf)<2 or self.buf[-1][0]!=DATA or \ self.buf[-2][0]!=STARTTAG or self.buf[-2][1]!='script': # syntax error, ignore return # get script data script = self.buf[-1][1].strip() # remove html comments if script.startswith("<!--"): i = script.find('\n') if i==-1: script = script[4:] else: script = script[(i+1):] if script.endswith("-->"): script = script[:-3] if not script: # again, ignore an empty script del self.buf[-1] del self.buf[-1] else: self.jsScript(script, 0.0, item)
"""<?xml version="1.0" encoding="UTF8"?>
"""<?xml version="1.0" encoding="UTF-8"?>
def testRdfDescription (self): self.filt("""<?xml version="1.0" encoding="ISO-8859-1"?>
"""<?xml version="1.0" encoding="UTF8"?>
"""<?xml version="1.0" encoding="UTF-8"?>
def testRdfDescription2 (self): self.filt("""<?xml version="1.0" encoding="ISO-8859-1"?>
"""<?xml version="1.0" encoding="UTF8"?>
"""<?xml version="1.0" encoding="UTF-8"?>
def testRdfDescription3 (self): self.filt("""<?xml version="1.0" encoding="ISO-8859-1"?>
"""compress the string s. Note that compression state is saved outside of this function in the compression object.
"""Compress the string s. Note that compression state is saved outside of this function in the compression object.
def filter (self, data, **attrs): """compress the string s. Note that compression state is saved outside of this function in the compression object. """ if not attrs.has_key('compressobj'): return data compobj = attrs['compressobj'] if compobj: header = compobj['header'] if header: compobj['header'] = '' wc.log.debug(wc.LOG_FILTER, 'writing gzip header') compobj['size'] += len(data) compobj['crc'] = zlib.crc32(data, compobj['crc']) data = "%s%s" % (header, compobj['compressor'].compress(data)) return data
if docontinue and serverpool.http_versions.get(addr, 1.1) < 1.1: self.client.error(417, i18n._("Expectation failed"), i18n._("Server does not understand HTTP/1.1")) return if expect:
if docontinue: if serverpool.http_versions.get(addr, 1.1) < 1.1: self.client.error(417, i18n._("Expectation failed"), i18n._("Server does not understand HTTP/1.1")) return elif expect:
def server_connected (self, server): assert self.state == 'connect' if not self.client.connected: # The client has aborted, so let's return this server # connection to the pool server.reuse() return self.server = server addr = (self.ipaddr, self.port) # check expectations expect = self.headers.get('Expect', '').lower().strip() docontinue = expect.startswith('100-continue') or \ expect.startswith('0100-continue') if docontinue and serverpool.http_versions.get(addr, 1.1) < 1.1: self.client.error(417, i18n._("Expectation failed"), i18n._("Server does not understand HTTP/1.1")) return if expect: self.client.error(417, i18n._("Expectation failed"), i18n._("Unsupported expectation `%s'")%expect) return # ok, assign server object self.state = 'response' # At this point, we tell the server that we are the client. # Once we get a response, we transfer to the real client. self.server.client_send_request(self.method, self.hostname, self.document, self.headers, self.content, self, self.nofilter, self.url)
ext.extra_compile_args.extend(extra)
for opt in extra: if opt not in ext.extra_compile_args: ext.extra_compile_args.append(opt)
def build_extensions (self): # For gcc 3.x we can add -std=gnu99 to get rid of warnings. extra = [] if self.compiler.compiler_type == 'unix': option = "-std=gnu99" if cc_supports_option(self.compiler.compiler, option): extra.append(option) # First, sanity-check the 'extensions' list self.check_extensions_list(self.extensions) for ext in self.extensions: ext.extra_compile_args.extend(extra) self.build_extension(ext)
if part==COMPLETE: return [DATA, ""]
def filter_tag (self, tag, attrs): #debug(NIGHTMARE, "rule %s filter_tag" % self.title) part = self.replace[0] #debug(NIGHTMARE, "original tag", `tag`, "attrs", attrs) #debug(NIGHTMARE, "replace", num_part(part), "with", `self.replace[1]`) if part==TAGNAME: return (STARTTAG, self.replace[1], attrs) if part==TAG: return (DATA, self.replace[1]) if part==ENCLOSED: return (STARTTAG, tag, attrs) if part==COMPLETE: return [DATA, ""] newattrs = {} # look for matching tag attributes for attr,val in attrs.items(): ro = self.attrs.get(attr) if ro: mo = ro.search(val) if mo: if part==ATTR: # replace complete attr if self.replace[1]: newattrs[self.replace[1][0]] = self.replace[1][1] else: # part has to be ATTRVAL # Python has named submatches, and we can use them # the name 'replace' replaces the value, # all other names are given as format strings dict = mo.groupdict() if dict.has_key('replace'): newattrs[attr] = dict['replace'] else: newattrs[attr] = self.replace[1] % dict continue # nothing matched, just append the attribute as is newattrs[attr] = val #debug(NIGHTMARE, "filtered tag", tag, "attrs", newattrs) return (STARTTAG, tag, newattrs)