rem
stringlengths 0
322k
| add
stringlengths 0
2.05M
| context
stringlengths 8
228k
|
|---|---|---|
goutput = gerror.strip('\n') + output
|
goutput = gerror.strip('\n') + goutput print goutput
|
def checking_loop(wiki): url = wiki.host while True: #Get all new history pages with pending status info('Lookig for pages') picked_pages = wiki.getMeta('CategoryHistory, overallvalue=pending') info('Found %d pages' % len(picked_pages)) if not picked_pages: info('No pages. Sleeping') time.sleep(10) continue #go thgrough all new pages for page in picked_pages: info('%s: picked %s' % (url, page)) path = tempfile.mkdtemp() os.chdir(path) info("Created tempdir %s" % path) #change the status to picked wiki.setMeta(page, {'overallvalue' : ['picked']}, True) metas = picked_pages[page] user = metas['user'].single().strip('[]') # get the attachment filename from the file meta info('Writing files') for filename in metas['file']: attachment_file = removeLink(filename) #get the source code info("Fetching sourcode from %s" % attachment_file) try: code = wiki.getAttachment(page, attachment_file) except opencollab.wiki.WikiFault, e: if 'There was an error in the wiki side (Nonexisting attachment' in e.args[0]: code = '' else: raise # get rid of the _rev<number> in filenames open(re.sub('(_rev\d+)', '', removeLink(filename)), 'w').write(code) #if there is wrong amount of question page linksd, leave #the returned assignment as picked so that other #assignments can be checked. if len(metas['question']) != 1: error('Invalid meta data in %s! There we %d values!\n' % (page, len(metas['question']))) continue #get the question pagenmae question = metas['question'].single(None) question = question.strip('[]') #find associataed answerpages answer_pages = wiki.getMeta(question +'/options').values()[0]['answer'] info("Found %d answer pages" % len(answer_pages)) regex = re.compile('{{{\s*(.*)\s*}}}', re.DOTALL) wrong = list() right = list() outputs = list() #TODO file handling for apage in [x.strip('[]') for x in answer_pages]: info('getting answers from %s' % apage) answer_meta = wiki.getMeta(apage).values()[0] testname = answer_meta['testname'].single() outputpage = None inputpage = None if 'output' in answer_meta: outputpage = answer_meta['output'].single().strip('[]') outfilesatt = wiki.listAttachments if 'input' in answer_meta: inputpage = answer_meta['input'].single().strip('[]') args = answer_meta['parameters'].single() input = '' if inputpage: content = wiki.getPage(inputpage) input = regex.search(content).group(1) input_meta = wiki.getMeta(inputpage) filelist = input_meta[inputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(inputpage, filename) info('Writing input file %s' % filename) open(os.path.join(path, filename), 'w').write(content) output = '' if outputpage: content = wiki.getPage(outputpage) output = regex.search(content).group(1) goutput, gerror, timeout = run(args, input, path) goutput = goutput.strip('\n') output = output.strip('\n') goutput = gerror.strip('\n') + output if timeout: goutput = "***** TIMEOUT *****\nYOUR PROGRAM TIMED OUT!\n\n" + goutput if goutput != output: info("Test %s failed" % testname) wrong.append(testname) else: info("Test %s succeeded" % testname) right.append(testname) #put user output to wiki outputs.append('[[%s]]' % (user + '/' + outputpage,)) try: wiki.putPage(user + '/' + outputpage, outputtemplate % (re.sub(ur'[\x00-\x08\x0b-\x19]', '?', goutput), testname)) except opencollab.wiki.WikiFault, error_message: # It's ok if the comment does not change if 'There was an error in the wiki side (You did not change the page content, not saved!)' in error_message: pass else: raise info('Removing ' + path) shutil.rmtree(path) metas = dict() #clear old info info('Clearing old metas') wiki.setMeta(page, {'wrong': [], 'right': []}, True) if len(wrong) == 0: metas['overallvalue'] = ['success'] else: metas['overallvalue'] = ['failure'] if outputs: metas['output'] = outputs if wrong: metas['wrong'] = wrong if right: metas['right'] = right info('Setting new metas') #add metas wiki.setMeta(page, metas, True) info('Done') time.sleep(5)
|
wiki.putPage(user + '/' + outputpage, outputtemplate % (re.sub(ur'[\x00-\x08\x0b-\x19]', '?', goutput), testname))
|
wiki.putPage(user + '/' + outputpage, outputtemplate % (esc(goutput), testname))
|
def checking_loop(wiki): url = wiki.host while True: #Get all new history pages with pending status info('Lookig for pages') picked_pages = wiki.getMeta('CategoryHistory, overallvalue=pending') info('Found %d pages' % len(picked_pages)) if not picked_pages: info('No pages. Sleeping') time.sleep(10) continue #go thgrough all new pages for page in picked_pages: info('%s: picked %s' % (url, page)) path = tempfile.mkdtemp() os.chdir(path) info("Created tempdir %s" % path) #change the status to picked wiki.setMeta(page, {'overallvalue' : ['picked']}, True) metas = picked_pages[page] user = metas['user'].single().strip('[]') # get the attachment filename from the file meta info('Writing files') for filename in metas['file']: attachment_file = removeLink(filename) #get the source code info("Fetching sourcode from %s" % attachment_file) try: code = wiki.getAttachment(page, attachment_file) except opencollab.wiki.WikiFault, e: if 'There was an error in the wiki side (Nonexisting attachment' in e.args[0]: code = '' else: raise # get rid of the _rev<number> in filenames open(re.sub('(_rev\d+)', '', removeLink(filename)), 'w').write(code) #if there is wrong amount of question page linksd, leave #the returned assignment as picked so that other #assignments can be checked. if len(metas['question']) != 1: error('Invalid meta data in %s! There we %d values!\n' % (page, len(metas['question']))) continue #get the question pagenmae question = metas['question'].single(None) question = question.strip('[]') #find associataed answerpages answer_pages = wiki.getMeta(question +'/options').values()[0]['answer'] info("Found %d answer pages" % len(answer_pages)) regex = re.compile('{{{\s*(.*)\s*}}}', re.DOTALL) wrong = list() right = list() outputs = list() #TODO file handling for apage in [x.strip('[]') for x in answer_pages]: info('getting answers from %s' % apage) answer_meta = wiki.getMeta(apage).values()[0] testname = answer_meta['testname'].single() outputpage = None inputpage = None if 'output' in answer_meta: outputpage = answer_meta['output'].single().strip('[]') outfilesatt = wiki.listAttachments if 'input' in answer_meta: inputpage = answer_meta['input'].single().strip('[]') args = answer_meta['parameters'].single() input = '' if inputpage: content = wiki.getPage(inputpage) input = regex.search(content).group(1) input_meta = wiki.getMeta(inputpage) filelist = input_meta[inputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(inputpage, filename) info('Writing input file %s' % filename) open(os.path.join(path, filename), 'w').write(content) output = '' if outputpage: content = wiki.getPage(outputpage) output = regex.search(content).group(1) goutput, gerror, timeout = run(args, input, path) goutput = goutput.strip('\n') output = output.strip('\n') goutput = gerror.strip('\n') + output if timeout: goutput = "***** TIMEOUT *****\nYOUR PROGRAM TIMED OUT!\n\n" + goutput if goutput != output: info("Test %s failed" % testname) wrong.append(testname) else: info("Test %s succeeded" % testname) right.append(testname) #put user output to wiki outputs.append('[[%s]]' % (user + '/' + outputpage,)) try: wiki.putPage(user + '/' + outputpage, outputtemplate % (re.sub(ur'[\x00-\x08\x0b-\x19]', '?', goutput), testname)) except opencollab.wiki.WikiFault, error_message: # It's ok if the comment does not change if 'There was an error in the wiki side (You did not change the page content, not saved!)' in error_message: pass else: raise info('Removing ' + path) shutil.rmtree(path) metas = dict() #clear old info info('Clearing old metas') wiki.setMeta(page, {'wrong': [], 'right': []}, True) if len(wrong) == 0: metas['overallvalue'] = ['success'] else: metas['overallvalue'] = ['failure'] if outputs: metas['output'] = outputs if wrong: metas['wrong'] = wrong if right: metas['right'] = right info('Setting new metas') #add metas wiki.setMeta(page, metas, True) info('Done') time.sleep(5)
|
pagename = self.formatter.page.page_name if AttachFile.exists(self.request, pagename, url): href = AttachFile.getAttachUrl(pagename, url, self.request)
|
current_pagename = self.formatter.page.page_name pagename, filename = AttachFile.absoluteName(url, current_pagename) if AttachFile.exists(self.request, pagename, filename): href = AttachFile.getAttachUrl(pagename, filename, self.request)
|
def _transclude_repl(self, word, groups): """Handles transcluding content, usually embedding images.""" target = groups.get('transclude_target', '') target = wikiutil.url_unquote(target) desc = groups.get('transclude_desc', '') or '' params = groups.get('transclude_params', u'') or u'' acceptable_attrs_img = ['class', 'title', 'longdesc', 'width', 'height', 'align', ] # no style because of JS acceptable_attrs_object = ['class', 'title', 'width', 'height', # no style because of JS 'type', 'standby', ] # we maybe need a hack for <PARAM> here m = self.link_target_re.match(target) if m: if m.group('extern_addr'): # currently only supports ext. image inclusion target = m.group('extern_addr') desc = self._transclude_description(desc, target) tag_attrs, query_args = self._get_params(params, tag_attrs={'class': 'external_image', 'alt': desc, 'title': desc, }, acceptable_attrs=acceptable_attrs_img) return self.formatter.image(src=target, **tag_attrs) # FF2 has a bug with target mimetype detection, it looks at the url path # and expects to find some "filename extension" there (like .png) and this # (not the response http headers) will set the default content-type of # the object. This will often work for staticly served files, but # fails for MoinMoin attachments (they don't have the filename.ext in the # path, but in the query string). FF3 seems to have this bug fixed, opera 9.2 # also works. #return (self.formatter.transclusion(1, data=target) + # desc + # self.formatter.transclusion(0))
|
req, escaped=1)
|
req)
|
def render(req, f, content, parent='', depth=0): listfmt = {'class': 'attachtree_list'} direntryfmt = {'class': 'attachtree_direntry'} fileentryfmt = {'class': 'attachtree_fileentry'} result = f.bullet_list(1, **listfmt) for entry, values in sorted(content.items()): fullname = os.path.join(parent, unicode(entry)) if not values.items(): result += f.listitem(1, **fileentryfmt) link = getAttachUrl(f.page.page_name, quote(fullname.encode("iso-8859-15"), safe=""), req, escaped=1) result += f.url(1, link, 'attachtree_link') result += f.text(entry) result += f.url(0) result += f.listitem(0) else: result += f.listitem(1, **direntryfmt) result += f.text(entry) result += f.listitem(0) result += render(req, f, values, fullname, depth + 1) result += f.bullet_list(0) return result
|
kw['ticket'] = wikiutil.createTicket(request)
|
kw['ticket'] = wikiutil.createTicket(request, pagename=pagename, action=action_name)
|
def attachUrl(request, pagename, filename=None, **kw): # filename is not used yet, but should be used later to make a sub-item url if not (kw.get('do') in ['get', 'view', None] and kw.get('rename') is None): # create a ticket for the not so harmless operations kw['ticket'] = wikiutil.createTicket(request) if kw: qs = '?%s' % wikiutil.makeQueryString(kw, want_unicode=False) else: qs = '' return "%s/%s%s" % (request.getScriptname(), wikiutil.quoteWikinameURL(pagename), qs)
|
DCENABLED = 1
|
DCENABLED = 0 dc_deprecated = "dircache function calls (dcdisable,dclistdir,dcreset) are deprecated, please fix caller"
|
def realPathCase(path): return None
|
request.theme.add_msg(_("[Content of new page loaded from %s]") % (template_page, ), 'info')
|
request.theme.add_msg(_("[Content of new page loaded from %s]") % (template_page_escaped, ), 'info')
|
def sendEditor(self, **kw): """ Send the editor form page.
|
request.theme.add_msg(_("[Template %s not found]") % (template_page, ), 'warning')
|
request.theme.add_msg(_("[Template %s not found]") % (template_page_escaped, ), 'warning')
|
def sendEditor(self, **kw): """ Send the editor form page.
|
request.theme.add_msg(_("[You may not read %s]") % (template_page, ), 'error')
|
request.theme.add_msg(_("[You may not read %s]") % (template_page_escaped, ), 'error')
|
def sendEditor(self, **kw): """ Send the editor form page.
|
user_agent = self.request.user_agent
|
user_agent = self.environ.get('HTTP_USER_AGENT')
|
def isSpiderAgent(self): """ Simple check if useragent is a spider bot. """ cfg = self.cfg user_agent = self.request.user_agent if user_agent and cfg.cache.ua_spiders: return cfg.cache.ua_spiders.search(user_agent.browser) is not None return False
|
return cfg.cache.ua_spiders.search(user_agent.browser) is not None
|
return cfg.cache.ua_spiders.search(user_agent) is not None
|
def isSpiderAgent(self): """ Simple check if useragent is a spider bot. """ cfg = self.cfg user_agent = self.request.user_agent if user_agent and cfg.cache.ua_spiders: return cfg.cache.ua_spiders.search(user_agent.browser) is not None return False
|
self.timeout = timeout
|
def communicate(self, input=None, timeout=None): """Interact with process: Send data to stdin. Read data from stdout and stderr, until end-of-file is reached. Wait for process to terminate. The optional input argument should be a string to be sent to the child process, or None, if no data should be sent to the child.
|
|
stdout_thread.join(self.timeout)
|
stdout_thread.join(timeout)
|
def _communicate(self, input): stdout = None # Return stderr = None # Return
|
stderr_thread.join(self.timeout)
|
stderr_thread.join(timeout)
|
def _communicate(self, input): stdout = None # Return stderr = None # Return
|
rlist, wlist, xlist = select.select(read_set, write_set, [], self.timeout)
|
rlist, wlist, xlist = select.select(read_set, write_set, [], timeout)
|
def _communicate(self, input): timed_out = False read_set = [] write_set = [] stdout = None # Return stderr = None # Return
|
TODO: make it a generator
|
def get_all_session_ids(self, request): """ Return a list of all session ids known to the SessionService.
|
|
return os.listdir(request.cfg.session_dir)
|
store = self._store_get(request) return store.get_all_sids()
|
def get_all_session_ids(self, request): # TODO: this should be done by werkzeug's FilesystemSessionStore # the sids are the same as the filenames, see filename_template above return os.listdir(request.cfg.session_dir)
|
self._request.session['trail'] = trail[-(self._cfg.trail_size-1):]
|
self._request.session['trail'] = trail[-self._cfg.trail_size:]
|
def addTrail(self, page): """ Add page to trail.
|
vals = [x.strip() for x in form[form_escape(oldkey)]
|
vals = [x.strip() for x in form[oldkey]
|
def wr(fmt, *args): args = tuple(map(form_escape, args)) request.write(fmt % args)
|
py.test.skip("Platform don't provide crypt module!")
|
py.test.skip("Platform does not provide crypt module!")
|
def test_auth_with_des_stored_password(self): """ Create user with {DES} password and check that user can login. """ # Create test user name = u'Test User' password = '{DES}sajEeYaHYyeSU' # 12345 self.createUser(name, password, True) try: import crypt # Try to "login" theuser = user.User(self.request, name=name, password='12345') assert theuser.valid except ImportError: py.test.skip("Platform don't provide crypt module!")
|
msg = page.saveText(pagecontent, page.get_real_rev())
|
try: msg = page.saveText(pagecontent, page.get_real_rev()) except PageEditor.Unchanged: pass
|
def save(self, answer_data, options): old_answers = self.answers() answertype = options.get("answertype", [u""])[0]
|
request.setContentLanguage(request.lang) request.theme.send_title(_('Diff for "%s"') % (pagename, ), pagename=pagename, allow_doubleclick=1)
|
def execute(pagename, request): """ Handle "action=diff" checking for either a "rev=formerrevision" parameter or rev1 and rev2 parameters """ if not request.user.may.read(pagename): Page(request, pagename).send_page() return try: date = request.values['date'] try: date = long(date) # must be long for py 2.2.x except StandardError: date = 0 except KeyError: date = 0 try: rev1 = int(request.values.get('rev1', -1)) except StandardError: rev1 = 0 try: rev2 = int(request.values.get('rev2', 0)) except StandardError: rev2 = 0 if rev1 == -1 and rev2 == 0: rev1 = request.rev if rev1 is None: rev1 = -1 # spacing flag? ignorews = int(request.values.get('ignorews', 0)) _ = request.getText # get a list of old revisions, and back out if none are available currentpage = Page(request, pagename) currentrev = currentpage.current_rev() if currentrev < 2: request.theme.add_msg(_("No older revisions available!"), "error") currentpage.send_page() return if date: # this is how we get called from RecentChanges rev1 = 0 log = editlog.EditLog(request, rootpagename=pagename) for line in log.reverse(): if date >= line.ed_time_usecs and int(line.rev) != 99999999: rev1 = int(line.rev) break else: rev1 = 1 rev2 = 0 # Start output # This action generates content in the user language request.setContentLanguage(request.lang) request.theme.send_title(_('Diff for "%s"') % (pagename, ), pagename=pagename, allow_doubleclick=1) if rev1 > 0 and rev2 > 0 and rev1 > rev2 or rev1 == 0 and rev2 > 0: rev1, rev2 = rev2, rev1 if rev1 == -1: oldrev = currentrev - 1 oldpage = Page(request, pagename, rev=oldrev) elif rev1 == 0: oldrev = currentrev oldpage = currentpage else: oldrev = rev1 oldpage = Page(request, pagename, rev=oldrev) if rev2 == 0: newrev = currentrev newpage = currentpage else: newrev = rev2 newpage = Page(request, pagename, rev=newrev) edit_count = abs(newrev - oldrev) f = request.formatter request.write(f.div(1, id="content")) oldrev = oldpage.get_real_rev() newrev = newpage.get_real_rev() title = _('Differences between revisions %d and %d') % (oldrev, newrev) if edit_count > 1: title += ' ' + _('(spanning %d versions)') % (edit_count, ) title = f.text(title) page_url = wikiutil.escape(currentpage.url(request), True) def enabled(val): return not val and u' disabled="disabled"' or u'' revert_html = "" if request.user.may.revert(pagename): revert_html = """ <td style="border:0"> <form action="%s" method="get"> <div style="text-align:center"> <input name="action" value="revert" type="hidden"> <input name="rev" value="%d" type="hidden"> <input value="%s" type="submit"%s> </div> </form> </td> """ % (page_url, rev2, _("Revert to this revision"), enabled(newrev < currentrev)) other_diff_button_html = """ <td style="border:0; width:1%%"> <form action="%s" method="get"> <div style="text-align:left"> <input name="action" value="diff" type="hidden"> <input name="rev1" value="%d" type="hidden"> <input name="rev2" value="%d" type="hidden"> <input value="%s" type="submit"%s> </div> </form> </td>
|
|
oldlog = oldpage.editlog_entry() newlog = newpage.editlog_entry()
|
def enabled(val): return not val and u' disabled="disabled"' or u''
|
|
('history_paging', True, "Enable paging for history. Warning: because event log functions are rather unefficient, enabling this option may lead to possibility of producing heavy requests!"),
|
('history_paging', True, "Enable paging functionality for info action's history display."),
|
def __init__(self, exprstr): self.text = exprstr self.value = eval(exprstr)
|
for key in request.form: query[key] = request.form[key]
|
for key in request.values: query[key] = request.values[key]
|
def _handle_oidreturn(self): request = self.request _ = request.getText
|
session = store.get(sid)
|
session = store.get(str(sid))
|
def get_session(self, request, sid=None): if sid is None: cookie_name = get_cookie_name(request, name=request.cfg.cookie_name, usage=self.cookie_usage) sid = request.cookies.get(cookie_name, None) store = self._store_get(request) if sid is None: session = store.new() else: session = store.get(sid) return session
|
text = FormatterBase.macro(self, macro_obj, name, args)
|
_arg_editlink = r'(,\s*(?P<editlink>editlink))?' macro_args = re.sub(_arg_editlink, '', args) text = FormatterBase.macro(self, macro_obj, name, macro_args)
|
def macro(self, macro_obj, name, args, markup=None): """As far as the DocBook formatter is conserned there are three kinds of macros: Bad, Handled and Unknown.
|
now = time.time() expiry = session.get('expires', 0) if expiry < now: self.destroy_session(request, session) session = store.new()
|
expiry = session.get('expires') if expiry is not None: now = time.time() if expiry < now: logging.debug("session has expired (expiry: %r now: %r)" % (expiry, now)) self.destroy_session(request, session) session = store.new()
|
def get_session(self, request, sid=None): if sid is None: cookie_name = get_cookie_name(request, name=request.cfg.cookie_name, usage=self.cookie_usage) sid = request.cookies.get(cookie_name, None) logging.debug("get_session for sid %r" % sid) store = self._store_get(request) if sid is None: session = store.new() else: session = store.get(sid) now = time.time() expiry = session.get('expires', 0) if expiry < now: # the browser should've killed that cookie already. # clock not in sync? trying to cheat? self.destroy_session(request, session) session = store.new() logging.debug("get_session returns session %r" % session) return session
|
request.write("Pages to revert:<br>%s" % "<br>".join(revertpages))
|
request.write("Pages to revert:<br>%s" % "<br>".join([wikiutil.escape(p) for p in revertpages]))
|
def revert_pages(request, editor, timestamp): _ = request.getText editor = wikiutil.url_unquote(editor, want_unicode=False) timestamp = int(timestamp * 1000000) log = editlog.EditLog(request) pages = {} revertpages = [] for line in log.reverse(): if line.ed_time_usecs < timestamp: break if not request.user.may.read(line.pagename): continue if not line.pagename in pages: pages[line.pagename] = 1 if repr(line.getInterwikiEditorData(request)) == editor: revertpages.append(line.pagename) request.write("Pages to revert:<br>%s" % "<br>".join(revertpages)) for pagename in revertpages: request.write("Begin reverting %s ...<br>" % pagename) msg = revert_page(request, pagename, editor) if msg: request.write("<p>%s: %s</p>" % ( Page.Page(request, pagename).link_to(request), msg)) request.write("Finished reverting %s.<br>" % pagename)
|
request.write("Begin reverting %s ...<br>" % pagename)
|
request.write("Begin reverting %s ...<br>" % wikiutil.escape(pagename))
|
def revert_pages(request, editor, timestamp): _ = request.getText editor = wikiutil.url_unquote(editor, want_unicode=False) timestamp = int(timestamp * 1000000) log = editlog.EditLog(request) pages = {} revertpages = [] for line in log.reverse(): if line.ed_time_usecs < timestamp: break if not request.user.may.read(line.pagename): continue if not line.pagename in pages: pages[line.pagename] = 1 if repr(line.getInterwikiEditorData(request)) == editor: revertpages.append(line.pagename) request.write("Pages to revert:<br>%s" % "<br>".join(revertpages)) for pagename in revertpages: request.write("Begin reverting %s ...<br>" % pagename) msg = revert_page(request, pagename, editor) if msg: request.write("<p>%s: %s</p>" % ( Page.Page(request, pagename).link_to(request), msg)) request.write("Finished reverting %s.<br>" % pagename)
|
request.write("Finished reverting %s.<br>" % pagename)
|
request.write("Finished reverting %s.<br>" % wikiutil.escape(pagename))
|
def revert_pages(request, editor, timestamp): _ = request.getText editor = wikiutil.url_unquote(editor, want_unicode=False) timestamp = int(timestamp * 1000000) log = editlog.EditLog(request) pages = {} revertpages = [] for line in log.reverse(): if line.ed_time_usecs < timestamp: break if not request.user.may.read(line.pagename): continue if not line.pagename in pages: pages[line.pagename] = 1 if repr(line.getInterwikiEditorData(request)) == editor: revertpages.append(line.pagename) request.write("Pages to revert:<br>%s" % "<br>".join(revertpages)) for pagename in revertpages: request.write("Begin reverting %s ...<br>" % pagename) msg = revert_page(request, pagename, editor) if msg: request.write("<p>%s: %s</p>" % ( Page.Page(request, pagename).link_to(request), msg)) request.write("Finished reverting %s.<br>" % pagename)
|
def _parser_content_repl(self, line):
|
def _parser_content(self, line):
|
def _parser_content_repl(self, line): if self.in_pre == 'search_parser' and line.strip(): if line.strip().startswith("#!"): parser_name = line.strip()[2:].split()[0] if parser_name == 'wiki': self.in_pre = 'found_parser' self.parser_name = 'wiki_form'
|
return apply(wikiParser._parser_content_repl, (self, word, groups))
|
return apply(wikiParser._parser_content, (self, line))
|
def _parser_content_repl(self, line): if self.in_pre == 'search_parser' and line.strip(): if line.strip().startswith("#!"): parser_name = line.strip()[2:].split()[0] if parser_name == 'wiki': self.in_pre = 'found_parser' self.parser_name = 'wiki_form'
|
except opencollab.wiki.WikiFault, error_message:
|
except opencollab.wiki.WikiFault, error_message.args[0]:
|
def checking_loop(wiki): wikiname = wiki.host + ''.join(wiki.path.rsplit('?')[:-1]) while True: #Get all new history pages with pending status info('Looking for pages in %s' % wikiname) picked_pages = wiki.getMeta('CategoryHistory, overallvalue=pending') info('Found %d pages' % len(picked_pages)) if not picked_pages: info('Sleeping') time.sleep(10) continue #go thgrough all new pages for page in picked_pages: info('%s: picked %s' % (wikiname, page)) tempdir = tempfile.mkdtemp() info("Created tempdir %s" % tempdir) os.chdir(tempdir) #change the status to picked wiki.setMeta(page, {'overallvalue' : ['picked']}, True) metas = picked_pages[page] user = metas['user'].single().strip('[]') # get the attachment filename from the file meta info('Writing files') codes = dict() for filename in metas['file']: attachment_file = removeLink(filename) #get the source code info("Fetching sourcode from %s" % attachment_file) try: code = wiki.getAttachment(page, attachment_file) except opencollab.wiki.WikiFault, e: if 'There was an error in the wiki side (Nonexisting attachment' in e.args[0]: code = '' else: raise # get rid of the _rev<number> in filenames codes[re.sub('(_rev\d+)', '', removeLink(filename))] = code revision = re.search('_rev(\d+)', removeLink(filename)).group(1) #if there is wrong amount of question page linksd, leave #the returned assignment as picked so that other #assignments can be checked. if len(metas['question']) != 1: error('Invalid meta data in %s! There we %d values!\n' % (page, len(metas['question']))) continue #get the question pagenmae question = metas['question'].single(None) if not question: error('NO QUESTION PAGE IN HISTORY %s!' % page) continue question = question.strip('[]') #find associataed answerpages answer_pages = wiki.getMeta(question +'/options').values()[0]['answer'] info("Found %d answer pages" % len(answer_pages)) regex = re.compile('{{{\s*(.*)\s?}}}', re.DOTALL) wrong = list() right = list() outputs = list() for answer_page in [x.strip('[]') for x in answer_pages]: info('getting answers from %s' % answer_page) answer_meta = wiki.getMeta(answer_page).values()[0] testname = answer_meta['testname'].single() outputpage = None inputpage = None if 'output' in answer_meta: outputpage = answer_meta['output'].single().strip('[]') outfilesatt = wiki.listAttachments if 'input' in answer_meta: inputpage = answer_meta['input'].single().strip('[]') try: args = answer_meta['parameters'].single() except ValueError: error('No params!') continue input = '' input_files = dict() if inputpage: content = wiki.getPage(inputpage) input = regex.search(content).group(1) input_meta = wiki.getMeta(inputpage) filelist = input_meta[inputpage]['file'] for attachment in filelist: filename = removeLink(attachment) try: content = wiki.getAttachment(inputpage, filename) except opencollab.wiki.WikiFault, error_message: if "There was an error in the wiki side (Nonexisting attachment:" in error_message: content = '' else: raise input_files[filename] = content output = '' if outputpage: content = wiki.getPage(outputpage) output = regex.search(content).group(1) output_meta = wiki.getMeta(outputpage) # get output files output_files = dict() filelist = output_meta[outputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(outputpage, filename) output_files[filename] = content info('Running test') stu_output, stu_error, timeout, stu_files = run_test(codes, args, input, input_files, tempdir) #FIXME. Must check that editors in raippa do not add #newlines in output. If not. These lines can be removed stu_output = stu_error + stu_output if timeout: stu_output = stu_output + "\n***** TIMEOUT *****\nYOUR PROGRAM TIMED OUT!\n\n" if len(stu_output) > 1024*100: stu_output = "***** Your program produced more than 100kB of output data *****\n(Meaning that your program failed)\nPlease check your code before returning it\n" info('Excess output!') passed = True if stu_output.rstrip('\n') != output.rstrip('\n'): passed = False # compare output files for filename, content in output_files.items(): if filename not in stu_files: info("%s is missing" % filename) passed = False break if content != stu_files[filename]: info("Output file does not match") passed = False break if passed: info("Test %s succeeded" % testname) right.append(testname) else: info("Test %s failed" % testname) wrong.append(testname) #put user output to wiki. stu_outputpage = user + '/' + outputpage outputs.append('[[%s]]' % stu_outputpage) try: wiki.putPage(stu_outputpage, outputtemplate % (esc(stu_output), testname)) #clean old attachments before adding new ones for old_attachment in wiki.listAttachments(stu_outputpage): wiki.deleteAttachment(stu_outputpage, old_attachment) for ofilename, ocontent in stu_files.items(): wiki.putAttachment(stu_outputpage, ofilename, esc(ocontent), True) except opencollab.wiki.WikiFault, error_message: # It's ok if the comment does not change if 'There was an error in the wiki side (You did not change the page content, not saved!)' in error_message: pass elif 'There was an error in the wiki side (Attachment not saved, file exists)' in error_message: pass else: raise # put output file metas to output page wiki.setMeta(stu_outputpage, {'file' : ['[[attachment:%s]]' % esc(x) for x in stu_files.keys()]}) info('Removing ' + tempdir) shutil.rmtree(tempdir) metas = dict() #clear old info info('Clearing old metas') wiki.setMeta(page, {'wrong': [], 'right': []}, True) if len(wrong) == 0: metas['overallvalue'] = ['success'] else: metas['overallvalue'] = ['failure'] if outputs: metas['output'] = outputs if wrong: metas['wrong'] = wrong if right: metas['right'] = right info('Setting new metas') #add metas wiki.setMeta(page, metas, True) info('Done')
|
def __init__(self, formatter):
|
def __init__(self, formatter, **kw):
|
def __init__(self, formatter): pygments.formatter.Formatter.__init__(self) self.result = [] self.formatter = formatter
|
u = self.request.handle_jid_auth(jid)
|
request = self.request request.session = request.cfg.session_service.get_session(request) logging.debug("getJabberAuthToken: got session %r" % request.session) u = user.get_by_jabber_id(request, jid) logging.debug("getJabberAuthToken: got user %r" % u)
|
def xmlrpc_getJabberAuthToken(self, jid, secret): """ Returns a token which can be used for authentication.
|
return self._generate_auth_token(u)
|
u.auth_method = 'moin' request.user = u request.cfg.session_service.finalize(request, request.session) logging.debug("getJabberAuthToken: returning sid %r" % request.session.sid) return request.session.sid
|
def xmlrpc_getJabberAuthToken(self, jid, secret): """ Returns a token which can be used for authentication.
|
text = _("""Dear Superuser, a new user has just been created on %(sitename)s". Details follow:
|
text = _("""Dear Superuser, a new user has just been created on "%(sitename)s". Details follow:
|
def user_created_message(request, _, sitename, username, email): """Formats a message used to notify about accounts being created @return: a dict containing message body and subject """ subject = _("New user account created on %(sitename)s") % {'sitename': sitename or "Wiki"} text = _("""Dear Superuser, a new user has just been created on %(sitename)s". Details follow: User name: %(username)s Email address: %(useremail)s""") % { 'username': username, 'useremail': email, 'sitename': sitename or "Wiki", } return {'subject': subject, 'text': text}
|
if avs.subset(avset): vector += "AV:" + av[0] + "/"
|
if len(avs) > 0 and avs <= avset: avv = av.pop() vector += "AV:" + avv[0] + "/"
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) av = base_metas["Access Vector"][:1] avs = set(av) if avs.subset(avset): vector += "AV:" + av[0] + "/" else: return None ac = base_metas["Access Complexity"][:1] acs = set(ac) if acs.subset(acset): vector += "AC:" + ac[0] + "/" else: return None au = base_metas["Authentication"][:1] aus = set(au) if aus.subset(auset): vector += "Au:" + au[0] + "/C:C/I:C/A:C" else: return None return vector
|
if acs.subset(acset): vector += "AC:" + ac[0] + "/"
|
if len(acs) > 0 and acs <= acset: acv = ac.pop() vector += "AC:" + acv[0] + "/"
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) av = base_metas["Access Vector"][:1] avs = set(av) if avs.subset(avset): vector += "AV:" + av[0] + "/" else: return None ac = base_metas["Access Complexity"][:1] acs = set(ac) if acs.subset(acset): vector += "AC:" + ac[0] + "/" else: return None au = base_metas["Authentication"][:1] aus = set(au) if aus.subset(auset): vector += "Au:" + au[0] + "/C:C/I:C/A:C" else: return None return vector
|
if aus.subset(auset): vector += "Au:" + au[0] + "/C:C/I:C/A:C"
|
if len(aus) > 0 and aus <= auset: auv = au.pop() vector += "Au:" + auv[0] + "/C:C/I:C/A:C"
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) av = base_metas["Access Vector"][:1] avs = set(av) if avs.subset(avset): vector += "AV:" + av[0] + "/" else: return None ac = base_metas["Access Complexity"][:1] acs = set(ac) if acs.subset(acset): vector += "AC:" + ac[0] + "/" else: return None au = base_metas["Authentication"][:1] aus = set(au) if aus.subset(auset): vector += "Au:" + au[0] + "/C:C/I:C/A:C" else: return None return vector
|
new_name = form.get('name', [request.user.name])[0]
|
new_name = wikiutil.clean_input(form.get('name', [request.user.name])[0]).strip()
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
new_email = wikiutil.clean_input(form.get('email', [request.user.email])[0]) new_email = new_email.strip()
|
new_email = wikiutil.clean_input(form.get('email', [request.user.email])[0]).strip()
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
request.user.aliasname = wikiutil.clean_input(form.get('aliasname', [''])[0])
|
request.user.aliasname = wikiutil.clean_input(form.get('aliasname', [''])[0]).strip()
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
request.user.editor_default = form.get('editor_default', [self.cfg.editor_default])[0] request.user.editor_ui = form.get('editor_ui', [self.cfg.editor_ui])[0]
|
request.user.editor_default = wikiutil.clean_input(form.get('editor_default', [self.cfg.editor_default])[0]) request.user.editor_ui = wikiutil.clean_input(form.get('editor_ui', [self.cfg.editor_ui])[0])
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
theme_name = form.get('theme_name', [self.cfg.theme_default])[0]
|
theme_name = wikiutil.clean_input(form.get('theme_name', [self.cfg.theme_default])[0])
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
request.user.language = form.get('language', [''])[0]
|
request.user.language = wikiutil.clean_input(form.get('language', [''])[0])
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
pass
|
value = 0
|
def _save_user_prefs(self): _ = self._ form = self.request.form request = self.request
|
page = request.graphdata.getpage(arg) newpages = page.get_in().get(CATEGORY_KEY, list())
|
newpages = request.graphdata.get_in(arg).get(CATEGORY_KEY, list())
|
def can_be_read(name): return request.user.may.read(name)
|
leftpane = '<br>\n'.join([indent(x) for x in leftresult.splitlines()]) rightpane = '<br>\n'.join([indent(x) for x in rightresult.splitlines()])
|
leftpane = '<br>'.join([indent(x) for x in leftresult.splitlines()]) rightpane = '<br>'.join([indent(x) for x in rightresult.splitlines()])
|
def diff(request, old, new, old_top='', new_top='', old_bottom='', new_bottom='', old_top_class='', new_top_class='', old_bottom_class='', new_bottom_class=''): """ Find changes between old and new and return HTML markup visualising them. @param old: old text [unicode] @param new: new text [unicode] @param old_top: Custom html for adding ontop of old revision column (optional) @param old_bottom: Custom html for adding at bottom of old revision column (optional) @param new_top: Custom html for adding ontop of new revision column (optional) @param new_bottom: Custom html for adding at bottom of new revision column (optional) @param old_top_class: Custom class for <td> with old_top content (optional) @param new_top_class: Custom class for <td> with new_top content (optional) @param old_bottom_class: Custom class for <td> with old_bottom content (optional) @param new_bottom_class: Custom class for <td> with new_bottom content (optional) """ _ = request.getText t_line = _("Line") + " %d" seq1 = old.splitlines() seq2 = new.splitlines() seqobj = difflib.SequenceMatcher(None, seq1, seq2) linematch = seqobj.get_matching_blocks() result = """
|
timestamp_until = self.request.user.getFormattedDateTime(self.timestamp+secs_valid)
|
def acquire(self): """ Begin an edit lock depending on the mode chosen in the config.
|
|
wiki=True) % {'owner': owner, 'timestamp': timestamp, 'mins_valid': mins_valid}
|
wiki=True) % {'owner': owner, 'timestamp': timestamp_until, 'mins_valid': mins_valid}
|
def acquire(self): """ Begin an edit lock depending on the mode chosen in the config.
|
avv = base_metas["Access Vector"][:1] if avv not in avset:
|
av = base_metas["Access Vector"][:1] avs = set(av) if avs.subset(avset): vector += "AV:" + av[0] + "/" else:
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = base_metas["Access Vector"][:1] if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = base_metas["Access Complexity"][:1] if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = base_metas["Authentication"][:1] if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
else: vector += "AV:" + avv[0] + "/" acv = base_metas["Access Complexity"][:1] if acv not in acset:
|
ac = base_metas["Access Complexity"][:1] acs = set(ac) if acs.subset(acset): vector += "AC:" + ac[0] + "/" else:
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = base_metas["Access Vector"][:1] if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = base_metas["Access Complexity"][:1] if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = base_metas["Authentication"][:1] if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
else: vector += "AC:" + acv[0] + "/" auv = base_metas["Authentication"][:1] if auv not in auset:
|
au = base_metas["Authentication"][:1] aus = set(au) if aus.subset(auset): vector += "Au:" + au[0] + "/C:C/I:C/A:C" else:
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = base_metas["Access Vector"][:1] if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = base_metas["Access Complexity"][:1] if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = base_metas["Authentication"][:1] if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C"
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = base_metas["Access Vector"][:1] if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = base_metas["Access Complexity"][:1] if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = base_metas["Authentication"][:1] if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
|
wrong.append(testname)
|
def checking_loop(wiki): url = wiki.host while True: #Get all new history pages with pending status info('Lookig for pages') picked_pages = wiki.getMeta('CategoryHistory, overallvalue=pending') info('Found %d pages' % len(picked_pages)) if not picked_pages: info('No pages. Sleeping') time.sleep(10) continue #go thgrough all new pages for page in picked_pages: info('%s: picked %s' % (url, page)) path = tempfile.mkdtemp() os.chdir(path) info("Created tempdir %s" % path) #change the status to picked wiki.setMeta(page, {'overallvalue' : ['picked']}, True) metas = picked_pages[page] user = metas['user'].single().strip('[]') # get the attachment filename from the file meta info('Writing files') for filename in metas['file']: attachment_file = removeLink(filename) #get the source code info("Fetching sourcode from %s" % attachment_file) try: code = wiki.getAttachment(page, attachment_file) except opencollab.wiki.WikiFault, e: if 'There was an error in the wiki side (Nonexisting attachment' in e.args[0]: code = '' else: raise # get rid of the _rev<number> in filenames open(re.sub('(_rev\d+)', '', removeLink(filename)), 'w').write(code) revision = re.search('_rev(\d+)', removeLink(filename)).group(1) #if there is wrong amount of question page linksd, leave #the returned assignment as picked so that other #assignments can be checked. if len(metas['question']) != 1: error('Invalid meta data in %s! There we %d values!\n' % (page, len(metas['question']))) continue #get the question pagenmae question = metas['question'].single(None) question = question.strip('[]') #find associataed answerpages answer_pages = wiki.getMeta(question +'/options').values()[0]['answer'] info("Found %d answer pages" % len(answer_pages)) regex = re.compile('{{{\s*(.*)\s*}}}', re.DOTALL) wrong = list() right = list() outputs = list() for apage in [x.strip('[]') for x in answer_pages]: info('getting answers from %s' % apage) answer_meta = wiki.getMeta(apage).values()[0] testname = answer_meta['testname'].single() outputpage = None inputpage = None if 'output' in answer_meta: outputpage = answer_meta['output'].single().strip('[]') outfilesatt = wiki.listAttachments if 'input' in answer_meta: inputpage = answer_meta['input'].single().strip('[]') args = answer_meta['parameters'].single() input = '' if inputpage: content = wiki.getPage(inputpage) input = regex.search(content).group(1) input_meta = wiki.getMeta(inputpage) filelist = input_meta[inputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(inputpage, filename) info('Writing input file %s' % filename) open(os.path.join(path, filename), 'w').write(content) output = '' if outputpage: content = wiki.getPage(outputpage) output = regex.search(content).group(1) output_meta = wiki.getMeta(outputpage) # get output files output_files = dict() filelist = output_meta[outputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(outputpage, filename) output_files[filename] = content info('Running test') goutput, gerror, timeout, gfiles = run(args, input, path) goutput = goutput.strip('\n') output = output.strip('\n') goutput = gerror.strip('\n') + goutput if timeout: goutput = goutput + "\n***** TIMEOUT *****\nYOUR PROGRAM TIMED OUT!\n\n" + goutput if len(goutput) > 1024*100: goutput = "***** Your program produced more than 100kB of output data *****\n(Meaning that your program failed)\nPlease check your code before returning it\n" info('Excess output!') passed = True if goutput != output: info("Test %s failed" % testname) wrong.append(testname) failed = False # compare output files for filename, content in output_files.items(): if filename not in gfiles: info("A file is missing") passed = False break if content != gfiles[filename]: info("Output file does not match") passed = False break if passed: info("Test %s succeeded" % testname) right.append(testname) else: info("Test %s failed" % testname) wrong.append(testname) #put user output to wiki. outputs.append('[[%s]]' % (user + '/' + outputpage,)) try: wiki.putPage(user + '/' + outputpage, outputtemplate % (esc(goutput), testname)) for ofilename, ocontent in gfiles.items(): wiki.putAttachment(user + '/' + outputpage, ofilename, ocontent) except opencollab.wiki.WikiFault, error_message: # It's ok if the comment does not change if 'There was an error in the wiki side (You did not change the page content, not saved!)' in error_message: pass elif 'There was an error in the wiki side (Attachment not saved, file exists)' in error_message: pass else: raise # put output file metas to output page wiki.setMeta(user + '/' + outputpage, {'file' : ['[[attachment:%s]]' % x for x in gfiles.keys()]}) info('Removing ' + path) shutil.rmtree(path) metas = dict() #clear old info info('Clearing old metas') wiki.setMeta(page, {'wrong': [], 'right': []}, True) if len(wrong) == 0: metas['overallvalue'] = ['success'] else: metas['overallvalue'] = ['failure'] if outputs: metas['output'] = outputs if wrong: metas['wrong'] = wrong if right: metas['right'] = right info('Setting new metas') #add metas wiki.setMeta(page, metas, True) info('Done') time.sleep(5)
|
|
else
|
else:
|
def add_user_to_group(request, myuser, group, create_link=True, comment=""): if not wikiutil.isGroupPage(request, group): raise GroupException("Page '%s' is not a group page." % group) if not (request.user.may.read(group) and request.user.may.write(group)): raise GroupException("No permissions to write to page '%s'." % group) member_rex = re.compile(ur"^ \* +(?:\[\[)?(.+?)(?:\]\])? *$", re.UNICODE) page = PageEditor(request, group) text = page.get_raw_body() _, head, tail = parse_categories(request, text) insertion_point = len(head) for lineno, line in enumerate(head): match = member_rex.match(line) if not match: continue if match.group(1).lower() == myuser.name.lower(): return insertion_point = lineno + 1 if create_link: template = " * [[%s]]" else: template = " * %s" head.insert(insertion_point, template % myuser.name) text = "\n".join(head + tail) if comment: page.saveText(text, 0, comment=comment) else page.saveText(text, 0)
|
_ = request.getText
|
def _handle_name_continuation(self, request): if not 'openid.id' in request.session: return CancelLogin(_('No OpenID found in session.'))
|
|
def_date = edit.split('_')[0]
|
defdate_rexp = re.compile('(\d{4}-\d{2}-\d{2})_') match = defdate_rexp.search(edit) if match: def_date = match.groups()[0]
|
def show_entryform(request): time_now = datetime.datetime.now() + datetime.timedelta(minutes=30) time_now -= datetime.timedelta(minutes=int(time_now.strftime("%M"))%30) def_date = time_now.strftime("%Y-%m-%d") def_time = time_now.strftime("%H:%M") duration = u'00:00' edit_page = u'' title = u'' time_opts = unicode() categories = categories = request.form.get('categories', [rc['timetrack']])[0] if request.form.has_key('date'): def_date = request.form.get('date')[0].encode() elif request.form.has_key('edit'): edit = request.form.get('edit')[0].encode() edit_page = u'<input type="hidden" name="edit" value="%s">' % edit def_date = edit.split('_')[0] #categories = ','.join(categories) pagelist, metakeys, styles = metatable_parseargs(request, categories, get_all_keys=True) meta = get_metas(request, edit, metakeys, display=False, checkAccess=True) if meta[u'Date']: if meta.has_key(u'Duration'): try: duration = meta[u'Duration'][0] except: None if meta.has_key(u'Time'): try: def_time = meta[u'Time'][0] except: None body = PageEditor(request, edit).get_raw_body() if '----' in body: title = body.split('----')[0] temp = list() for line in title.split("\n"): if not line.startswith("#acl"): temp.append(line) title = " ".join(temp) for h in range(24): for m in ['00','30']: t = u'"%02d:%s"' % (h,m) if t.find(def_time) != -1: t += ' selected' time_opts += u'<option value=%s>%02d:%s</option>\n' % (t,h,m) tasklist = unicode() for task in pages_in_category(request, rc['task']): tasktitle = Task(request, task).title() if not tasktitle: tasktitle = task tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle) pass html = u'''
|
None
|
pass
|
def show_entryform(request): time_now = datetime.datetime.now() + datetime.timedelta(minutes=30) time_now -= datetime.timedelta(minutes=int(time_now.strftime("%M"))%30) def_date = time_now.strftime("%Y-%m-%d") def_time = time_now.strftime("%H:%M") duration = u'00:00' edit_page = u'' title = u'' time_opts = unicode() categories = categories = request.form.get('categories', [rc['timetrack']])[0] if request.form.has_key('date'): def_date = request.form.get('date')[0].encode() elif request.form.has_key('edit'): edit = request.form.get('edit')[0].encode() edit_page = u'<input type="hidden" name="edit" value="%s">' % edit def_date = edit.split('_')[0] #categories = ','.join(categories) pagelist, metakeys, styles = metatable_parseargs(request, categories, get_all_keys=True) meta = get_metas(request, edit, metakeys, display=False, checkAccess=True) if meta[u'Date']: if meta.has_key(u'Duration'): try: duration = meta[u'Duration'][0] except: None if meta.has_key(u'Time'): try: def_time = meta[u'Time'][0] except: None body = PageEditor(request, edit).get_raw_body() if '----' in body: title = body.split('----')[0] temp = list() for line in title.split("\n"): if not line.startswith("#acl"): temp.append(line) title = " ".join(temp) for h in range(24): for m in ['00','30']: t = u'"%02d:%s"' % (h,m) if t.find(def_time) != -1: t += ' selected' time_opts += u'<option value=%s>%02d:%s</option>\n' % (t,h,m) tasklist = unicode() for task in pages_in_category(request, rc['task']): tasktitle = Task(request, task).title() if not tasktitle: tasktitle = task tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle) pass html = u'''
|
None
|
pass if meta.has_key(u'Task'): try: def_task = removelink(meta[u'Task'][0]) except: pass
|
def show_entryform(request): time_now = datetime.datetime.now() + datetime.timedelta(minutes=30) time_now -= datetime.timedelta(minutes=int(time_now.strftime("%M"))%30) def_date = time_now.strftime("%Y-%m-%d") def_time = time_now.strftime("%H:%M") duration = u'00:00' edit_page = u'' title = u'' time_opts = unicode() categories = categories = request.form.get('categories', [rc['timetrack']])[0] if request.form.has_key('date'): def_date = request.form.get('date')[0].encode() elif request.form.has_key('edit'): edit = request.form.get('edit')[0].encode() edit_page = u'<input type="hidden" name="edit" value="%s">' % edit def_date = edit.split('_')[0] #categories = ','.join(categories) pagelist, metakeys, styles = metatable_parseargs(request, categories, get_all_keys=True) meta = get_metas(request, edit, metakeys, display=False, checkAccess=True) if meta[u'Date']: if meta.has_key(u'Duration'): try: duration = meta[u'Duration'][0] except: None if meta.has_key(u'Time'): try: def_time = meta[u'Time'][0] except: None body = PageEditor(request, edit).get_raw_body() if '----' in body: title = body.split('----')[0] temp = list() for line in title.split("\n"): if not line.startswith("#acl"): temp.append(line) title = " ".join(temp) for h in range(24): for m in ['00','30']: t = u'"%02d:%s"' % (h,m) if t.find(def_time) != -1: t += ' selected' time_opts += u'<option value=%s>%02d:%s</option>\n' % (t,h,m) tasklist = unicode() for task in pages_in_category(request, rc['task']): tasktitle = Task(request, task).title() if not tasktitle: tasktitle = task tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle) pass html = u'''
|
tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle)
|
tasklist += " <option value='%s' %s>%s</option>\n" % (task, selected, tasktitle)
|
def show_entryform(request): time_now = datetime.datetime.now() + datetime.timedelta(minutes=30) time_now -= datetime.timedelta(minutes=int(time_now.strftime("%M"))%30) def_date = time_now.strftime("%Y-%m-%d") def_time = time_now.strftime("%H:%M") duration = u'00:00' edit_page = u'' title = u'' time_opts = unicode() categories = categories = request.form.get('categories', [rc['timetrack']])[0] if request.form.has_key('date'): def_date = request.form.get('date')[0].encode() elif request.form.has_key('edit'): edit = request.form.get('edit')[0].encode() edit_page = u'<input type="hidden" name="edit" value="%s">' % edit def_date = edit.split('_')[0] #categories = ','.join(categories) pagelist, metakeys, styles = metatable_parseargs(request, categories, get_all_keys=True) meta = get_metas(request, edit, metakeys, display=False, checkAccess=True) if meta[u'Date']: if meta.has_key(u'Duration'): try: duration = meta[u'Duration'][0] except: None if meta.has_key(u'Time'): try: def_time = meta[u'Time'][0] except: None body = PageEditor(request, edit).get_raw_body() if '----' in body: title = body.split('----')[0] temp = list() for line in title.split("\n"): if not line.startswith("#acl"): temp.append(line) title = " ".join(temp) for h in range(24): for m in ['00','30']: t = u'"%02d:%s"' % (h,m) if t.find(def_time) != -1: t += ' selected' time_opts += u'<option value=%s>%02d:%s</option>\n' % (t,h,m) tasklist = unicode() for task in pages_in_category(request, rc['task']): tasktitle = Task(request, task).title() if not tasktitle: tasktitle = task tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle) pass html = u'''
|
$('end-tr').setStyle('display', '');
|
def show_entryform(request): time_now = datetime.datetime.now() + datetime.timedelta(minutes=30) time_now -= datetime.timedelta(minutes=int(time_now.strftime("%M"))%30) def_date = time_now.strftime("%Y-%m-%d") def_time = time_now.strftime("%H:%M") duration = u'00:00' edit_page = u'' title = u'' time_opts = unicode() categories = categories = request.form.get('categories', [rc['timetrack']])[0] if request.form.has_key('date'): def_date = request.form.get('date')[0].encode() elif request.form.has_key('edit'): edit = request.form.get('edit')[0].encode() edit_page = u'<input type="hidden" name="edit" value="%s">' % edit def_date = edit.split('_')[0] #categories = ','.join(categories) pagelist, metakeys, styles = metatable_parseargs(request, categories, get_all_keys=True) meta = get_metas(request, edit, metakeys, display=False, checkAccess=True) if meta[u'Date']: if meta.has_key(u'Duration'): try: duration = meta[u'Duration'][0] except: None if meta.has_key(u'Time'): try: def_time = meta[u'Time'][0] except: None body = PageEditor(request, edit).get_raw_body() if '----' in body: title = body.split('----')[0] temp = list() for line in title.split("\n"): if not line.startswith("#acl"): temp.append(line) title = " ".join(temp) for h in range(24): for m in ['00','30']: t = u'"%02d:%s"' % (h,m) if t.find(def_time) != -1: t += ' selected' time_opts += u'<option value=%s>%02d:%s</option>\n' % (t,h,m) tasklist = unicode() for task in pages_in_category(request, rc['task']): tasktitle = Task(request, task).title() if not tasktitle: tasktitle = task tasklist += " <option value='%s'>%s</option>\n" % (task, tasktitle) pass html = u'''
|
|
request.write(u'''<script type="text/javascript" src="%s/raippajs/mootools-1.2-core-yc.js"></script> <script type="text/javascript"> function toggle(el){
|
request.write(u'''<script type="text/javascript"> function toggleParticipants(el){
|
def writeCell(stuff, width=''): if width: request.write('<td width="%s">' % width) else: request.write('<td>') request.write(stuff) request.write('</td>')
|
'''% request.cfg.url_prefix_static )
|
''')
|
def writeCell(stuff, width=''): if width: request.write('<td width="%s">' % width) else: request.write('<td>') request.write(stuff) request.write('</td>')
|
onclick="toggle(this);">show</a></span><br
|
onclick="toggleParticipants(this);">show</a></span><br
|
def writeCell(stuff, width=''): if width: request.write('<td width="%s">' % width) else: request.write('<td>') request.write(stuff) request.write('</td>')
|
for lineno in range(1, self.start_line + 1): result.append(fmt.line_anchordef(lineno))
|
def format(self, tokensource, outfile): line_ready = False fmt = self.formatter result = self.result self.lineno = self.start_line
|
|
if value[-1] == "%":
|
if value and value[-1] == "%":
|
def _cell_style(self, node): # TODO: attrs = get_attrs(node) if node.hasAttribute("rowspan"): rowspan = ("|%s" % node.getAttribute("rowspan")) else: rowspan = ""
|
user_head.append('<meta name="edit_on_doubleclick" content="1">\n')
|
user_head.append('<meta name="edit_on_doubleclick" content="%s">\n' % (request.script_root or '/'))
|
def send_title(self, text, **keywords): """ Output the page header (and title).
|
else:
|
elif self._lists:
|
def bullet_list(self, on, **kw): if on: if self._lists: # No empty lines between sublists self._paragraph_ended = False self.paragraph_begin() self._lists.append(-1) self._listitem_on.append(False) else: self.paragraph_end() self._lists.pop() listitem_on = self._listitem_on.pop() if listitem_on: self._indent -= 3 return ''
|
else: if self.colorby: outgraph = self.color_nodes(outgraph) if self.shapeby: outgraph, warn = self.shape_nodes(outgraph) if warn: warnings.append(warn) outgraph = self.color_edges(outgraph) outgraph = self.edge_tooltips(outgraph) outgraph = self.circle_start_nodes(outgraph) outgraph = self.fix_node_urls(outgraph) key_parts = [outgraph, self.graphengine, self.shapefiles, self.orderby]
|
if self.colorby: outgraph = self.color_nodes(outgraph) if self.shapeby: outgraph, warn = self.shape_nodes(outgraph) if warn: warnings.append(warn) outgraph = self.color_edges(outgraph) outgraph = self.edge_tooltips(outgraph) outgraph = self.circle_start_nodes(outgraph) outgraph = self.fix_node_urls(outgraph) key_parts = [outgraph, self.graphengine, self.shapefiles, self.orderby]
|
def execute(self): cl.start('execute') _ = self.request.getText
|
pagename = self.formatter.page.page_name if AttachFile.exists(self.request, pagename, url): href = AttachFile.getAttachUrl(pagename, url, self.request, escaped=0)
|
current_pagename = self.formatter.page.page_name pagename, filename = AttachFile.absoluteName(url, current_pagename) if AttachFile.exists(self.request, pagename, filename): href = AttachFile.getAttachUrl(pagename, filename, self.request, escaped=0)
|
def _transclude_repl(self, word, groups): """Handles transcluding content, usually embedding images.""" target = groups.get('transclude_target', '') target = wikiutil.url_unquote(target, want_unicode=True) desc = groups.get('transclude_desc', '') or '' params = groups.get('transclude_params', u'') or u'' acceptable_attrs_img = ['class', 'title', 'longdesc', 'width', 'height', 'align', ] # no style because of JS acceptable_attrs_object = ['class', 'title', 'width', 'height', # no style because of JS 'type', 'standby', ] # we maybe need a hack for <PARAM> here m = self.link_target_re.match(target) if m: if m.group('extern_addr'): # currently only supports ext. image inclusion target = m.group('extern_addr') desc = self._transclude_description(desc, target) tag_attrs, query_args = self._get_params(params, tag_attrs={'class': 'external_image', 'alt': desc, 'title': desc, }, acceptable_attrs=acceptable_attrs_img) return self.formatter.image(src=target, **tag_attrs) # FF2 has a bug with target mimetype detection, it looks at the url path # and expects to find some "filename extension" there (like .png) and this # (not the response http headers) will set the default content-type of # the object. This will often work for staticly served files, but # fails for MoinMoin attachments (they don't have the filename.ext in the # path, but in the query string). FF3 seems to have this bug fixed, opera 9.2 # also works. #return (self.formatter.transclusion(1, data=target) + # desc + # self.formatter.transclusion(0))
|
def buildVector(av,ac,au):
|
def buildVector(base_metas):
|
def buildVector(av,ac,au): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = av["Access Vector"].pop() if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = ac["Access Complexity"].pop() if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = au["Authentication"].pop() if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
avv = av["Access Vector"].pop()
|
avv = av["Access Vector"][:1]
|
def buildVector(av,ac,au): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = av["Access Vector"].pop() if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = ac["Access Complexity"].pop() if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = au["Authentication"].pop() if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
acv = ac["Access Complexity"].pop()
|
acv = ac["Access Complexity"][:1]
|
def buildVector(av,ac,au): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = av["Access Vector"].pop() if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = ac["Access Complexity"].pop() if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = au["Authentication"].pop() if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
auv = au["Authentication"].pop()
|
auv = au["Authentication"][:1]
|
def buildVector(av,ac,au): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) avv = av["Access Vector"].pop() if avv not in avset: return None else: vector += "AV:" + avv[0] + "/" acv = ac["Access Complexity"].pop() if acv not in acset: return None else: vector += "AC:" + acv[0] + "/" auv = au["Authentication"].pop() if auv not in auset: return None else: vector += "Au:" + auv[0] + "/C:C/I:C/A:C" return vector
|
@param index: column to sort. By a given list it does a multiple sort
|
@param sort_columns: column to sort. By a given list it does a multiple sort
|
def sort_table(rows, sort_columns=None, reverse=False): """ sorts table rows @param rows: table rows to sort @param index: column to sort. By a given list it does a multiple sort @param reverse: reverse sort """ if not (sort_columns and isinstance(sort_columns, (list, tuple))): # don't sort if no list is given return rows for idx in reversed(sort_columns): rows = sorted(rows, key=lambda x: _compare(idx, x), reverse=reverse) return rows
|
@param index: list of column index number for sorting
|
@param sort_columns: list of column index number for sorting
|
def setData(self, dataset, sort_columns=None, reverse=False): """ Sets the data for the browser (see MoinMoin.util.dataset).
|
subject = _("New user account created on %(sitename)s") % {'sitename': sitename or "Wiki"}
|
subject = _("[%(sitename)s] New user account created") % {'sitename': sitename or "Wiki"}
|
def user_created_message(request, _, sitename, username, email): """Formats a message used to notify about accounts being created @return: a dict containing message body and subject """ subject = _("New user account created on %(sitename)s") % {'sitename': sitename or "Wiki"} text = _("""Dear Superuser, a new user has just been created on "%(sitename)s". Details follow: User name: %(username)s Email address: %(useremail)s""") % { 'username': username, 'useremail': email, 'sitename': sitename or "Wiki", } return {'subject': subject, 'text': text}
|
data['subject'] = _("New attachment added to page %(pagename)s on %(sitename)s") % {
|
data['subject'] = _("[%(sitename)s] New attachment added to page %(pagename)s") % {
|
def attachment_added(request, _, page_name, attach_name, attach_size): """Formats a message used to notify about new attachments @param _: a gettext function @return: a dict with notification data """ data = {} data['subject'] = _("New attachment added to page %(pagename)s on %(sitename)s") % { 'pagename': page_name, 'sitename': request.cfg.sitename or request.url_root, } data['text'] = _("Dear Wiki user,\n\n" 'You have subscribed to a wiki page "%(page_name)s" for change notification. ' "An attachment has been added to that page by %(editor)s. " "Following detailed information is available:\n\n" "Attachment name: %(attach_name)s\n" "Attachment size: %(attach_size)s\n") % { 'editor': user.getUserIdentification(request), 'page_name': page_name, 'attach_name': attach_name, 'attach_size': attach_size, } data['editor'] = user.getUserIdentification(request) data['page_name'] = page_name data['attach_size'] = attach_size data['attach_name'] = attach_name return data
|
data['subject'] = _("Removed attachment from page %(pagename)s on %(sitename)s") % {
|
data['subject'] = _("[%(sitename)s] Removed attachment from page %(pagename)s") % {
|
def attachment_removed(request, _, page_name, attach_name, attach_size): """Formats a message used to notify about removed attachments @param _: a gettext function @return: a dict with notification data """ data = {} data['subject'] = _("Removed attachment from page %(pagename)s on %(sitename)s") % { 'pagename': page_name, 'sitename': request.cfg.sitename or request.url_root, } data['text'] = _("Dear Wiki user,\n\n" 'You have subscribed to a wiki page "%(page_name)s" for change notification. ' "An attachment has been removed from that page by %(editor)s. " "Following detailed information is available:\n\n" "Attachment name: %(attach_name)s\n" "Attachment size: %(attach_size)s\n") % { 'editor': user.getUserIdentification(request), 'page_name': page_name, 'attach_name': attach_name, 'attach_size': attach_size, } data['editor'] = user.getUserIdentification(request) data['page_name'] = page_name data['attach_size'] = attach_size data['attach_name'] = attach_name return data
|
logging.traceback("Filter %s threw error '%s' for file %s" % (modulename, str(err), filename))
|
logging.exception("Filter %s threw error '%s' for file %s" % (modulename, str(err), filename))
|
def contentfilter(self, filename): """ Get a filter for content of filename and return unicode content.
|
print goutput
|
def checking_loop(wiki): url = wiki.host while True: #Get all new history pages with pending status info('Lookig for pages') picked_pages = wiki.getMeta('CategoryHistory, overallvalue=pending') info('Found %d pages' % len(picked_pages)) if not picked_pages: info('No pages. Sleeping') time.sleep(10) continue #go thgrough all new pages for page in picked_pages: info('%s: picked %s' % (url, page)) path = tempfile.mkdtemp() os.chdir(path) info("Created tempdir %s" % path) #change the status to picked wiki.setMeta(page, {'overallvalue' : ['picked']}, True) metas = picked_pages[page] user = metas['user'].single().strip('[]') # get the attachment filename from the file meta info('Writing files') for filename in metas['file']: attachment_file = removeLink(filename) #get the source code info("Fetching sourcode from %s" % attachment_file) try: code = wiki.getAttachment(page, attachment_file) except opencollab.wiki.WikiFault, e: if 'There was an error in the wiki side (Nonexisting attachment' in e.args[0]: code = '' else: raise # get rid of the _rev<number> in filenames open(re.sub('(_rev\d+)', '', removeLink(filename)), 'w').write(code) #if there is wrong amount of question page linksd, leave #the returned assignment as picked so that other #assignments can be checked. if len(metas['question']) != 1: error('Invalid meta data in %s! There we %d values!\n' % (page, len(metas['question']))) continue #get the question pagenmae question = metas['question'].single(None) question = question.strip('[]') #find associataed answerpages answer_pages = wiki.getMeta(question +'/options').values()[0]['answer'] info("Found %d answer pages" % len(answer_pages)) regex = re.compile('{{{\s*(.*)\s*}}}', re.DOTALL) wrong = list() right = list() outputs = list() #TODO file handling for apage in [x.strip('[]') for x in answer_pages]: info('getting answers from %s' % apage) answer_meta = wiki.getMeta(apage).values()[0] testname = answer_meta['testname'].single() outputpage = None inputpage = None if 'output' in answer_meta: outputpage = answer_meta['output'].single().strip('[]') outfilesatt = wiki.listAttachments if 'input' in answer_meta: inputpage = answer_meta['input'].single().strip('[]') args = answer_meta['parameters'].single() input = '' if inputpage: content = wiki.getPage(inputpage) input = regex.search(content).group(1) input_meta = wiki.getMeta(inputpage) filelist = input_meta[inputpage]['file'] for attachment in filelist: filename = removeLink(attachment) content = wiki.getAttachment(inputpage, filename) info('Writing input file %s' % filename) open(os.path.join(path, filename), 'w').write(content) output = '' if outputpage: content = wiki.getPage(outputpage) output = regex.search(content).group(1) goutput, gerror, timeout = run(args, input, path) goutput = goutput.strip('\n') output = output.strip('\n') goutput = gerror.strip('\n') + goutput print goutput if timeout: goutput = "***** TIMEOUT *****\nYOUR PROGRAM TIMED OUT!\n\n" + goutput if goutput != output: info("Test %s failed" % testname) wrong.append(testname) else: info("Test %s succeeded" % testname) right.append(testname) #put user output to wiki outputs.append('[[%s]]' % (user + '/' + outputpage,)) try: wiki.putPage(user + '/' + outputpage, outputtemplate % (esc(goutput), testname)) except opencollab.wiki.WikiFault, error_message: # It's ok if the comment does not change if 'There was an error in the wiki side (You did not change the page content, not saved!)' in error_message: pass else: raise info('Removing ' + path) shutil.rmtree(path) metas = dict() #clear old info info('Clearing old metas') wiki.setMeta(page, {'wrong': [], 'right': []}, True) if len(wrong) == 0: metas['overallvalue'] = ['success'] else: metas['overallvalue'] = ['failure'] if outputs: metas['output'] = outputs if wrong: metas['wrong'] = wrong if right: metas['right'] = right info('Setting new metas') #add metas wiki.setMeta(page, metas, True) info('Done') time.sleep(5)
|
|
if cookie_lifetime: cfg = request.cfg cookie_name = get_cookie_name(request, name=cfg.cookie_name, usage=self.cookie_usage) cookie_path = cfg.cookie_path or '/' cookie_expires = int(time.time() / 60) * 60 + cookie_lifetime cookie_secure = (cfg.cookie_secure or cfg.cookie_secure is None and request.is_secure) logging.debug("user: %r, setting session cookie: %r" % (userobj, session.sid))
|
cookie_expires = int(time.time() / 60) * 60 + cookie_lifetime kill_session = not userobj.valid and 'user.id' in session if kill_session: logging.debug("logout detected, will kill session") if cookie_lifetime and not kill_session: logging.debug("setting session cookie: %r" % (session.sid, ))
|
def finalize(self, request, session): if request.user.auth_method == 'setuid': userobj = request._setuid_real_user setuid = request.user.id else: userobj = request.user setuid = None logging.debug("finalize userobj = %r, setuid = %r" % (userobj, setuid))
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError:
|
elif not session.new: logging.debug("deleting session cookie!") request.delete_cookie(cookie_name, path=cookie_path, domain=cfg.cookie_domain) def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val:
|
def finalize(self, request, session): if request.user.auth_method == 'setuid': userobj = request._setuid_real_user setuid = request.user.id else: userobj = request.user setuid = None logging.debug("finalize userobj = %r, setuid = %r" % (userobj, setuid))
|
else: if val != current_val: session[key] = val
|
if not session.new:
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
if userobj and userobj.valid:
|
if cookie_lifetime and not kill_session: if userobj.valid:
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
logging.debug("after auth: storing valid user into session: %r" % userobj.name)
|
logging.debug("storing valid user into session: %r" % userobj.name)
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
logging.debug("after auth: user is invalid") if 'user.id' in session: logging.debug("after auth: destroying session: %r" % session) self.destroy_session(request, session) logging.debug("after auth: deleting session cookie!") request.delete_cookie(cookie_name, path=cookie_path, domain=cfg.cookie_domain)
|
for key in ['user.id', 'user.auth_method', 'user.auth_attribs', 'setuid', ]: if key in session: del session[key] logging.debug("no valid user, cleaned user info from session")
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
and session
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
|
elif not session.new: logging.debug("destroying session: %r" % session) self.destroy_session(request, session)
|
def update_session(key, val): """ put key/val into session, avoid writing if it is unchanged """ try: current_val = session[key] except KeyError: session[key] = val else: if val != current_val: session[key] = val
|
|
text = text.encode("utf-8")
|
def replace_variables(text, variables): text = text.encode("utf-8") for name, variable in variables.iteritems(): variable = unicode(variable).encode("utf-8") name = unicode(name).encode("utf-8") text = re.sub("@%s@" % name, variable, text) text = text.decode("utf-8") return text
|
|
variable = unicode(variable).encode("utf-8") name = unicode(name).encode("utf-8")
|
def replace_variables(text, variables): text = text.encode("utf-8") for name, variable in variables.iteritems(): variable = unicode(variable).encode("utf-8") name = unicode(name).encode("utf-8") text = re.sub("@%s@" % name, variable, text) text = text.decode("utf-8") return text
|
|
text = text.decode("utf-8")
|
def replace_variables(text, variables): text = text.encode("utf-8") for name, variable in variables.iteritems(): variable = unicode(variable).encode("utf-8") name = unicode(name).encode("utf-8") text = re.sub("@%s@" % name, variable, text) text = text.decode("utf-8") return text
|
|
def encode_address_field(message, key, charset): all_values = message.get_all(key, list()) all_values = [encodeAddress(value, charset) for value in all_values]
|
def encode_address_field(message, key, encoding, charset): values = message.get_all(key, list())
|
def encode_address_field(message, key, charset): all_values = message.get_all(key, list()) all_values = [encodeAddress(value, charset) for value in all_values] del message[key] for value in all_values: message[key] = value
|
for value in all_values: message[key] = value def sendmail(request, template, variables, recipient_filter=lambda x: True):
|
for value in values: if not isinstance(value, unicode): value = value.decode(encoding) message[key] = encodeAddress(value, charset) def prepare_message(template, variables, encoding="utf-8"): r"""Return a prepared email.Message object. >>> template = u"Subject: @SUBJECT@\nFrom: @FROM@\nTo: @TO@\nBCC: @FROM@\n\nHello, @GREETED@!" >>> variables = dict(SUBJECT="Test", FROM="from@example.com", TO="to@example.com", GREETED="World") >>> message = prepare_message(template, variables) >>> message["SUBJECT"] == variables["SUBJECT"] True >>> message["TO"] == variables["TO"] True >>> message["FROM"] == message["BCC"] == variables["FROM"] True >>> message.get_payload() 'Hello, World!' """
|
def encode_address_field(message, key, charset): all_values = message.get_all(key, list()) all_values = [encodeAddress(value, charset) for value in all_values] del message[key] for value in all_values: message[key] = value
|
ins = set(c)
|
ins = set(i)
|
def buildVector(base_metas): vector = "" avset = set(['Local', 'Network', 'Adjacent Network']) acset = set(['High', 'Medium', 'Low']) auset = set(['Multiple', 'Single', 'None']) ciaset = set(['None', 'Partial', 'Complete']) av = base_metas["Access Vector"][:1] avs = set(av) if len(avs) > 0 and avs <= avset: avv = av.pop() vector += "AV:" + avv[0] + "/" else: return None ac = base_metas["Access Complexity"][:1] acs = set(ac) if len(acs) > 0 and acs <= acset: acv = ac.pop() vector += "AC:" + acv[0] + "/" else: return None au = base_metas["Authentication"][:1] aus = set(au) if len(aus) > 0 and aus <= auset: auv = au.pop() vector += "Au:" + auv[0] + "/" else: return None c = base_metas["Confidentiality"][:1] cs = set(c) if len(cs) > 0 and cs <= ciaset: cv = c.pop() vector += "C:" + cv[0] + "/" else: vector += "C:C/" i = base_metas["Integrity"][:1] ins = set(c) if len(ins) > 0 and ins <= ciaset: inv = i.pop() vector += "I:" + inv[0] + "/" else: vector += "I:C/" a = base_metas["Availability"][:1] avas = set(a) if len(avas) > 0 and avas <= ciaset: avav = a.pop() vector += "A:" + avav[0] else: vector += "A:C" return vector
|
def get_choice(request, arg, name=None, choices=[None]):
|
def get_choice(request, arg, name=None, choices=[None], default_none=False):
|
def get_choice(request, arg, name=None, choices=[None]): """ For use with values returned from parse_quoted_separated or given as macro parameters, return a unicode string that must be in the choices given. None is a valid input and yields first of the valid choices. @param request: A request instance @param arg: The argument, may be None or a unicode string @param name: Name of the argument, for error messages @param choices: the possible choices @rtype: unicode or None @returns: the unicode string (or default value) """ assert isinstance(choices, (tuple, list)) if arg is None: return choices[0] elif not isinstance(arg, unicode): raise TypeError('Argument must be None or unicode') elif not arg in choices: _ = request.getText if name: raise ValueError( _('Argument "%s" must be one of "%s", not "%s"') % ( name, '", "'.join(choices), arg)) else: raise ValueError( _('Argument must be one of "%s", not "%s"') % ( '", "'.join(choices), arg)) return arg
|
return choices[0]
|
if default_none: return None else: return choices[0]
|
def get_choice(request, arg, name=None, choices=[None]): """ For use with values returned from parse_quoted_separated or given as macro parameters, return a unicode string that must be in the choices given. None is a valid input and yields first of the valid choices. @param request: A request instance @param arg: The argument, may be None or a unicode string @param name: Name of the argument, for error messages @param choices: the possible choices @rtype: unicode or None @returns: the unicode string (or default value) """ assert isinstance(choices, (tuple, list)) if arg is None: return choices[0] elif not isinstance(arg, unicode): raise TypeError('Argument must be None or unicode') elif not arg in choices: _ = request.getText if name: raise ValueError( _('Argument "%s" must be one of "%s", not "%s"') % ( name, '", "'.join(choices), arg)) else: raise ValueError( _('Argument must be one of "%s", not "%s"') % ( '", "'.join(choices), arg)) return arg
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.