X-Git-Url: https://codewiz.org/gitweb?p=geekigeeki.git;a=blobdiff_plain;f=geekigeeki.py;h=1891fdd8c5842cea4357d5339009ad40f90134d6;hp=d3ce2203bdb6ef379d4fdcec2ebf6e83494e5a0c;hb=ba86d6242b2d2ff713f3db4725adcabbfb9f1cce;hpb=8a6eee8a2bd33f522824afba25feac7c3ac74db3 diff --git a/geekigeeki.py b/geekigeeki.py index d3ce220..1891fdd 100755 --- a/geekigeeki.py +++ b/geekigeeki.py @@ -1,44 +1,44 @@ #!/usr/bin/python # -*- coding: utf-8 -*- # -# Copyright 1999, 2000 Martin Pool -# Copyright 2002 Gerardo Poggiali -# Copyright 2007, 2008, 2009 Bernie Innocenti +# Copyright (C) 1999, 2000 Martin Pool +# Copyright (C) 2002 Gerardo Poggiali +# Copyright (C) 2007, 2008, 2009, 2010 Bernie Innocenti # # This program is free software: you can redistribute it and/or modify -# it under the terms of the GNU General Public License as published by -# the Free Software Foundation, either version 3 of the License, or -# (at your option) any later version. -# -# This program is distributed in the hope that it will be useful, but -# WITHOUT ANY WARRANTY; without even the implied warranty of -# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -# General Public License for more details. -# -# You should have received a copy of the GNU General Public License +# it under the terms of the GNU Affero General Public License as +# published by the Free Software Foundation, either version 3 of the +# License, or (at your option) any later version. +# You should have received a copy of the GNU Affero General Public License # along with this program. If not, see . __version__ = '4.0-' + '$Id$'[4:11] -from time import clock +from time import clock, localtime, gmtime, strftime start_time = clock() title_done = False -import cgi, sys, os, re, errno, stat +import cgi, sys, os, re, errno, stat, glob image_ext = 'png|gif|jpg|jpeg|bmp|ico' -video_ext = "ogg|ogv|oga" # Not supported by Firefox 3.5: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt -image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")", re.IGNORECASE) -video_re = re.compile(r".*\.(" + video_ext + ")", re.IGNORECASE) +video_ext = "ogg|ogv|oga|webm" # Not supported by Firefox 3.6: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt +image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")$", re.IGNORECASE) +video_re = re.compile(r".*\.(" + video_ext + ")$", re.IGNORECASE) # FIXME: we accept stuff like foo/../bar and we shouldn't -file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)") -url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S") +file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/ ]*)$") +url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S$") ext_re = re.compile(r"\.([^\./]+)$") -# CGI stuff --------------------------------------------------------- +def config_get(key, default=None): + return globals().get(key, default) + def script_name(): return os.environ.get('SCRIPT_NAME', '') +#TODO: move post-edit hook into wiki, then kill this +def script_path(): + return os.path.split(os.environ.get('SCRIPT_FILENAME', ''))[0] + def query_string(): path_info = os.environ.get('PATH_INFO', '') if len(path_info) and path_info[0] == '/': @@ -46,8 +46,9 @@ def query_string(): else: return os.environ.get('QUERY_STRING', '') or 'FrontPage' -def privileged_path(): - return privileged_url or script_name() +def is_privileged(): + purl = config_get('privileged_url') + return (purl is not None) and os.environ.get('SCRIPT_URI', '').startswith(purl) def remote_user(): user = os.environ.get('REMOTE_USER', '') @@ -71,7 +72,7 @@ def is_external_url(pathname): def relative_url(pathname, privileged=False): if not is_external_url(pathname): if privileged: - url = privileged_path() + url = config_get('privileged_url') or script_name() else: url = script_name() pathname = url + '/' + pathname @@ -83,18 +84,18 @@ def permalink(s): def humanlink(s): return re.sub(r'(?:.*[/:]|)([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' ')) -# Split arg lists like "blah| blah blah| width=100 | align = center", +# Split arg lists like "blah|blah blah| width=100 | align = center", # return a list containing anonymous arguments and a map containing the named arguments def parse_args(s): args = [] - kwargs = {} + kvargs = {} for arg in s.strip('<[{}]>').split('|'): m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg) if m is not None: - kwargs[m.group(1)] = m.group(2) + kvargs[m.group(1)] = m.group(2) else: args.append(arg.strip()) - return (args, kwargs) + return (args, kvargs) def url_args(kvargs): argv = [] @@ -104,103 +105,41 @@ def url_args(kvargs): return '?' + '&'.join(argv) return '' -# Formatting stuff -------------------------------------------------- -def emit_header(mime_type="text/html"): - print("Content-type: " + mime_type + "; charset=utf-8\n") +def emit_header(mtime=None, mime_type="text/html"): + if mtime: + # Prevent caching when the wiki engine gets updated + mtime = max(mtime, os.stat(__file__).st_mtime) + print("Last-Modified: " + strftime("%a, %d %b %Y %H:%M:%S GMT", gmtime(mtime))) + if mime_type: + print("Content-type: " + mime_type + "; charset=utf-8") + print('') def send_guru(msg_text, msg_type): if not msg_text: return print('
')
     if msg_type == 'error':
         print('    Software Failure.  Press left mouse button to continue.\n')
-    print(msg_text)
+    print(cgi.escape(msg_text))
     if msg_type == 'error':
         print '\n           Guru Meditation #DEADBEEF.ABADC0DE'
-    print('
' \ + print('' \ % relative_url('sys/GuruMeditation.js')) -def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=False): - global title_done - if title_done: return - - # Head - emit_header() - print('') - print('') - - print("%s: %s" % (site_name, text)) - print(' ') - if not name: - print(' ') - - for meta in meta_urls: - http_equiv, content = meta - print(' ' % (http_equiv, relative_url(content))) - - for link in link_urls: - rel, href = link - print(' ' % (rel, relative_url(href))) - - if name and writable and privileged_url is not None: - print(' ' \ - % (privileged_path() + '?a=edit&q=' + name)) - - if history_url is not None: - print(' ' \ - % relative_url(history_url + '?a=rss')) - - print('') - - # Body - if name and writable and privileged_url is not None: - print('') - else: - print('') - - title_done = True - send_guru(msg_text, msg_type) - - # Navbar - print('') - -def send_httperror(status="403 Not Found", query=""): +def send_httperror(status="403 Not Found", query="", trace=False): print("Status: %s" % status) - send_title(None, msg_text=("%s: on query '%s'" % (status, query))) - send_footer() + msg_text = "%s: on query '%s'" % (status, query) + if trace: + import traceback + msg_text += '\n\n' + traceback.format_exc() + page = Page() + page.send_title(msg_text=msg_text) + page.send_footer() def link_tag(dest, text=None, privileged=False, **kvargs): if text is None: text = humanlink(dest) elif image_re.match(text): - text = '' + text + '' + text = '' + text + '' link_class = kvargs.get('class', kvargs.get('cssclass', None)) if not link_class: @@ -209,7 +148,7 @@ def link_tag(dest, text=None, privileged=False, **kvargs): elif file_re.match(dest) and Page(dest).exists(): link_class = 'wikilink' else: - text = nonexist_pfx + text + text = config_get('nonexist_pfx', '') + text link_class = 'nonexistent' # Prevent crawlers from following links potentially added by spammers or to generated pages @@ -223,23 +162,31 @@ def link_inline(name, descr=None, kvargs={}): if not descr: descr = humanlink(name) url = relative_url(name) if video_re.match(name): - return '' % url + return '' % url elif image_re.match(name): return '%s' % (url, url + url_args(kvargs), descr) elif file_re.match(name) and not ext_re.search(name): # FIXME: this guesses a wiki page - return Page(name).send_naked() + Page(name).send_naked(kvargs) # FIXME: we should return the page as a string rather than print it + return '' else: return '' \ % (url, url, name) -# Search --------------------------------------------------- +def link_inline_glob(pattern, descr=None, kvargs={}): + if not url_re.match(pattern) and bool(set(pattern) & set('?*[')): + s = '' + for name in sorted(glob.glob(pattern), reverse=bool(int(kvargs.get('reverse', '0'))) ): + s += link_inline(name, descr, kvargs) + return s + else: + return link_inline(pattern, descr, kvargs) -def print_search_stats(hits, searched): - print("

%d hits out of %d pages searched.

" % (hits, searched)) +def search_stats(hits, searched): + return "%d hits out of %d pages searched.\n" % (hits, searched) def handle_fullsearch(query, form): needle = form['q'].value - send_title(None, 'Full text search for "' + needle + '"') + Page().send_title(text='Full text search for "' + needle + '"') needle_re = re.compile(needle, re.IGNORECASE) hits = [] @@ -250,59 +197,50 @@ def handle_fullsearch(query, form): if count: hits.append((count, page_name)) - # The default comparison for tuples compares elements in order, - # so this sorts by number of hits + # The default comparison for tuples compares elements in order, so this sorts by number of hits hits.sort() hits.reverse() - print("
    ") + out = '' for (count, page_name) in hits: - print('
  • ' + link_tag(page_name)) - print(' . . . . ' + `count`) - print(['match', 'matches'][count != 1]) - print('

  • ') - print("
") + out += ' * [[' + page_name + ']] . . . ' + str(count) + ' ' + ['match', 'matches'][count != 1] + '\n' - print_search_stats(len(hits), len(all_pages)) + out += search_stats(len(hits), len(all_pages)) + WikiFormatter(out).print_html() def handle_titlesearch(query, form): needle = form['q'].value - send_title(None, 'Title search for "' + needle + '"') + Page().send_title(text='Title search for "' + needle + '"') needle_re = re.compile(needle, re.IGNORECASE) all_pages = page_list() hits = list(filter(needle_re.search, all_pages)) - print("
    ") + out = '' for filename in hits: - print('
  • ' + link_tag(filename) + "

  • ") - print("
") + out += ' * [[' + filename + ']]\n' - print_search_stats(len(hits), len(all_pages)) + out += search_stats(len(hits), len(all_pages)) + WikiFormatter(out).print_html() def handle_raw(pagename, form): - if not file_re.match(pagename): - send_httperror("403 Forbidden", pagename) - return - Page(pagename).send_raw() -def handle_edit(pagename, form): - if not file_re.match(pagename): - send_httperror("403 Forbidden", pagename) - return +def handle_atom(pagename, form): + Page(pagename).send_atom() - pg = Page(form['q'].value) +def handle_edit(pagename, form): + pg = Page(form['q'].value) if 'save' in form: if form['file'].value: pg.save(form['file'].file.read(), form['changelog'].value) else: pg.save(form['savetext'].value.replace('\r\n', '\n'), form['changelog'].value) - pg.format() + pg.send() elif 'cancel' in form: pg.msg_text = 'Editing canceled' pg.msg_type = 'notice' - pg.format() + pg.send() else: # preview or edit text = None if 'preview' in form: @@ -310,56 +248,40 @@ def handle_edit(pagename, form): pg.send_editor(text) def handle_get(pagename, form): - if file_re.match(pagename): - # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension! - from mimetypes import MimeTypes - mimetype, encoding = MimeTypes().guess_type(pagename) - if mimetype: - Page(pagename).send_raw(mimetype=mimetype, args=form) - else: - Page(pagename).format() - else: - send_httperror("403 Forbidden", pagename) + if not ext_re.search(pagename): # FIXME: no extension guesses a wiki page + Page(pagename).send() + else: + # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension! + from mimetypes import MimeTypes + mimetype, encoding = MimeTypes().guess_type(pagename) + Page(pagename).send_raw(mimetype=mimetype, args=form) -# Used by macros/WordIndex and macros/TitleIndex +# Used by sys/macros/WordIndex and sys/macros/TitleIndex def make_index_key(): links = ['%s' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz'] return '

' + ' | '.join(links) + '

' -def page_list(dirname=None, re=None): - if re is None: +def page_list(dirname=None, search_re=None): + if search_re is None: # FIXME: WikiWord is too restrictive now! - re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") - return sorted(filter(re.match, os.listdir(dirname or data_dir))) - -def send_footer(mod_string=None): - if globals().get('debug_cgi', False): - cgi.print_arguments() - cgi.print_form(form) - cgi.print_environ() - print(''' -') + search_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") + return sorted(filter(search_re.match, os.listdir(dirname or '.'))) -class WikiFormatter: - """Object that turns Wiki markup into HTML. +def _macro_ELAPSED_TIME(*args, **kvargs): + return "%03f" % (clock() - start_time) + +def _macro_VERSION(*args, **kvargs): + return __version__ - All formatting commands can be parsed one line at a time, though - some state is carried over between lines. - """ - def __init__(self, raw): +class WikiFormatter: + """Object that turns Wiki markup into HTML.""" + def __init__(self, raw, kvargs=None): self.raw = raw + self.kvargs = kvargs or {} self.h_level = 0 self.in_pre = self.in_html = self.in_table = self.in_li = False self.in_header = True - self.list_indents = [] + self.list_indents = [] # a list of pairs (indent_level, list_type) to track nested lists self.tr_cnt = 0 self.styles = { #wiki html enabled? @@ -367,11 +289,11 @@ class WikiFormatter: "**": ["b", False], "##": ["tt", False], "__": ["u", False], + "--": ["del", False], "^^": ["sup", False], ",,": ["sub", False], "''": ["em", False], # LEGACY "'''": ["b", False], # LEGACY - "``": ["tt", False], # LEGACY } def _b_repl(self, word): @@ -379,6 +301,9 @@ class WikiFormatter: style[1] = not style[1] return ['' + def _glyph_repl(self, word): + return '—' + def _tit_repl(self, word): if self.h_level: result = '

\n' % self.h_level @@ -393,18 +318,25 @@ class WikiFormatter: return '
' def _rule_repl(self, word): - return self._undent() + '\n


\n' % (len(word) - 2) + return '\n
\n' % (len(word) - 2) def _macro_repl(self, word): try: - args, kwargs = parse_args(word) + args, macro_kvargs = parse_args(word) + # Is this a parameter given to the current page? + if args[0] in self.kvargs: + return self.kvargs[args[0]] + # Is this an internal macro? macro = globals().get('_macro_' + args[0]) if not macro: - exec(open("macros/" + name + ".py").read(), globals()) - macro = globals().get('_macro_' + name) - return macro(*args, **kwargs) - except Exception: - msg = cgi.escape(word) + # Can we load (and cache) an external macro? + exec(open("sys/macros/" + args[0] + ".py").read(), globals()) + macro = globals().get('_macro_' + args[0]) + # Invoke macro passing both macro args augmented by page args + macro_kvargs.update(self.kvargs) + return macro(*args, **macro_kvargs) + except Exception, e: + msg = cgi.escape(word) + ": " + cgi.escape(str(e)) if not self.in_html: msg = '' + msg + '' return msg @@ -421,10 +353,10 @@ class WikiFormatter: # This double div nonsense works around a limitation of the HTML block model return '
' \ + '
' \ - + link_inline(name, descr, kvargs) \ + + link_inline_glob(name, descr, kvargs) \ + '
' + descr + '
' else: - return link_inline(name, None, kvargs) + return link_inline_glob(name, None, kvargs) def _html_repl(self, word): if not self.in_html and word.startswith(' new_level: - del(self.list_indents[-1]) if self.in_li: s += '' self.in_li = False # FIXME - s += '\n' + s += '\n' + del(self.list_indents[-1]) + + list_type = ('ul', 'ol')[list_type == '#'] while self._indent_level() < new_level: - self.list_indents.append(new_level) - s += '
    \n' + self.list_indents.append((new_level, list_type)) + s += '<' + list_type + '>\n' s += '

    ' return s - def _undent(self): - res = '

    ' - res += '
' * len(self.list_indents) - res += '

' - self.list_indents = [] - return res - def replace(self, match): for rule, hit in list(match.groupdict().items()): if hit: return getattr(self, '_' + rule + '_repl')(hit) else: - raise "Can't handle match " + repr(match) + raise Exception("Can't handle match " + repr(match)) def print_html(self): print('

') scan_re = re.compile(r"""(?: - # Styles and formatting - (?P \*\*|'''|//|''|\#\#|``|__|\^\^|,,) + # Styles and formatting ("--" must cling to a word to disambiguate it from the dash) + (?P \*\* | // | \#\# | __ | --\b | \b-- | \^\^ | ,, | ''' | '' ) | (?P \={2,6}) | (?P
\\\\) | (?P ^-{3,}) | (?P \b( FIXME | TODO | DONE )\b ) + | (?P --) # Links - | (?P \<\<([^\s\|\>]+)(?:\s*\|\s*([^\>]+)|)\>\>) - | (?P \[\[([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\]\]) + | (?P \<\<[^\>]+\>\>) + | (?P \[\[[^\]]+\]\]) # Inline HTML - | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b ) + | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b ) | (?P ( /\s*> | ) ) | (?P [<>&] ) @@ -556,10 +484,10 @@ class WikiFormatter: | (?P (http|https|ftp|mailto)\:[^\s'\"]+\S) | (?P [-\w._+]+\@[\w.-]+) - # Lists, divs, spans + # Lists, divs, spans and inline objects | (?P

  • ^\s+[\*\#]\s+) | (?P
       \{\{\{|\s*\}\}\})
    -            | (?P   \{\{([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\}\})
    +            | (?P   \{\{[^\}]+\}\})
     
                 # Tables
                 | (?P    ^\s*\|\|(=|)\s*)
    @@ -573,7 +501,7 @@ class WikiFormatter:
                 | (?P[<>&])"
                 )""", re.VERBOSE)
             blank_re = re.compile(r"^\s*$")
    -        indent_re = re.compile(r"^\s*")
    +        indent_re = re.compile(r"^(\s*)(\*|\#|)")
             tr_re = re.compile(r"^\s*\|\|")
             eol_re = re.compile(r"\r?\n")
             # For each line, we scan through looking for magic strings, outputting verbatim any intervening text
    @@ -596,40 +524,51 @@ class WikiFormatter:
                         print('

    ') else: indent = indent_re.match(self.line) - #3.0: print(self._indent_to(len(indent.group(0))), end=' ') - print(self._indent_to(len(indent.group(0)))) + print(self._indent_to(len(indent.group(1)), indent.group(2))) + # Stand back! Here we apply the monster regex that does all the parsing print(re.sub(scan_re, self.replace, self.line)) if self.in_pre: print('

    ') if self.in_table: print('

    ') - print(self._undent()) + print(self._indent_to(0)) print('

  • ') +class HttpException(Exception): + def __init__(self, error, query): + self.error = error + self.query = query + class Page: - def __init__(self, page_name): - self.page_name = page_name + def __init__(self, page_name="Limbo"): + self.page_name = page_name.rstrip('/'); self.msg_text = '' self.msg_type = 'error' + if not file_re.match(self.page_name): + raise HttpException("403 Forbidden", self.page_name) def split_title(self): # look for the end of words and the start of a new word and insert a space there return re.sub('([a-z])([A-Z])', r'\1 \2', self.page_name) def _filename(self): - return os.path.join(data_dir, self.page_name) + return self.page_name def _tmp_filename(self): - return os.path.join(data_dir, ('#' + self.page_name.replace('/','_') + '.' + str(os.getpid()) + '#')) + return self.page_name + '.tmp' + str(os.getpid()) + '#' - def exists(self): + def _mtime(self): try: - os.stat(self._filename()) - return True + return os.stat(self._filename()).st_mtime except OSError, err: if err.errno == errno.ENOENT: - return False + return None raise err + def exists(self): + if self._mtime(): + return True + return False + def get_raw_body(self, default=None): try: return open(self._filename(), 'rb').read() @@ -645,19 +584,21 @@ class Page: def format_dir(self): out = '== ' pathname = '' - for dirname in self.page_name.split('/'): - pathname = (pathname + '/' + dirname) if pathname else dirname + for dirname in self.page_name.strip('/').split('/'): + pathname = (pathname and pathname + '/' ) + dirname out += '[[' + pathname + '|' + dirname + ']]/' out += ' ==\n' + images_out = '\n' for filename in page_list(self._filename(), file_re): if image_re.match(filename): - if image_maxwidth: - maxwidth_arg = ' | maxwidth=' + str(image_maxwidth) - out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth_arg + ' | class=thumbleft}}\n' + maxwidth = config_get('image_maxwidth', '400') + if maxwidth: + maxwidth = ' | maxwidth=' + str(maxwidth) + images_out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth + ' | class=thumbleft}}\n' else: out += ' * [[' + self.page_name + '/' + filename + ']]\n' - return out + return out + images_out def pragmas(self): if not '_pragmas' in self.__dict__: @@ -673,7 +614,7 @@ class Page: #print "bernie: pragmas[" + m.group(1) + "] = " + m.group(2) + "
    \n" except IOError, err: if err.errno != errno.ENOENT and err.errno != errno.EISDIR: - raise er + raise err return self._pragmas def pragma(self, name, default): @@ -700,89 +641,150 @@ class Page: def can_read(self): return self.can("read", True) - def send_naked(self): + def send_title(self, name=None, text="Limbo", msg_text=None, msg_type='error'): + global title_done + if title_done: return + + # HEAD + emit_header(self._mtime()) + print('\n') + print("%s: %s" % (config_get('site_name', "Unconfigured Wiki"), text)) + print(' ') + if not name: + print(' ') + + for http_equiv, content in config_get('meta_urls', {}): + print(' ' % (http_equiv, relative_url(content))) + + for link in config_get('link_urls', {}): + rel, href = link + print(' ' % (rel, relative_url(href))) + + editable = name and self.can_write() and is_privileged() + if editable: + print(' ' \ + % relative_url('?a=edit&q=' + name, privileged=True)) + + print(' ' \ + % relative_url(name + '?a=atom')) + + print('') + + # BODY + if editable: + print('') + else: + print('') + + title_done = True + send_guru(msg_text, msg_type) + + if self.pragma("navbar", "on") != "on": + return + + # NAVBAR + print('') + + def send_footer(self): + if config_get('debug_cgi', False): + cgi.print_arguments() + cgi.print_form(form) + cgi.print_environ() + footer = self.pragma("footer", "sys/footer") + if footer != "off": + link_inline(footer, kvargs = { + 'LAST_MODIFIED': strftime(config_get('datetime_fmt', '%Y-%m-%dT%I:%M:%S%p'), localtime(self._mtime())) + }) + print("") + + def send_naked(self, kvargs=None): if self.can_read(): - WikiFormatter(self.get_raw_body()).print_html() + WikiFormatter(self.get_raw_body(), kvargs).print_html() else: send_guru("Read access denied by ACLs", "notice") - def format(self): + def send(self): #css foo.css value = self.pragma("css", None) if value: global link_urls link_urls += [ [ "stylesheet", value ] ] - send_title(self.page_name, self.split_title(), - msg_text=self.msg_text, msg_type=self.msg_type, writable=self.can_write()) + self.send_title(name=self.page_name, text=self.split_title(), msg_text=self.msg_text, msg_type=self.msg_type) self.send_naked() - send_footer(self._last_modified()) - - def _last_modified(self): - try: - from time import localtime, strftime - modtime = localtime(os.stat(self._filename())[stat.ST_MTIME]) - except OSError, err: - if err.errno != errno.ENOENT: - raise err - return None - return strftime(datetime_fmt, modtime) + self.send_footer() + + def send_atom(self): + emit_header(self._mtime(), 'application/atom+xml') + self.in_html = True + link_inline("sys/atom_header", kvargs = { + 'LAST_MODIFIED': strftime(config_get('datetime_fmt', '%a, %d %b %Y %I:%M:%S %p'), localtime(self._mtime())) + }) + self.in_html = False + self.send_naked() + self.in_html = True + link_inline("sys/atom_footer") + self.in_html = False def send_editor(self, preview=None): - send_title(None, 'Edit ' + self.split_title(), msg_text=self.msg_text, msg_type=self.msg_type) + self.send_title(text='Edit ' + self.split_title(), msg_text=self.msg_text, msg_type=self.msg_type) if not self.can_write(): send_guru("Write access denied by ACLs", "error") return - filename = '' - if 'file' in form: - filename = form['file'].value - - print(('

    Editing ' + self.page_name - + ' for ' + cgi.escape(remote_user()) - + ' from ' + cgi.escape(get_hostname(remote_host())) - + '

    ')) - print('
    ' % relative_url(self.page_name)) - print('') - print('
    ' % (self.page_name)) - print('' \ - % cgi.escape(preview or self.get_raw_body(default=''))) - print(' ' % filename) - print(""" -
    - - - - -
    -
    - - """) - print("

    " + link_tag('EditingTips') + "

    ") + if preview is None: + preview = self.get_raw_body(default='') + + link_inline("sys/EditPage", kvargs = { + 'EDIT_BODY': cgi.escape(preview), + #'EDIT_PREVIEW': WikiFormatter(preview).print_html(), + }) + if preview: print("
    ") WikiFormatter(preview).print_html() print("
    ") - send_footer() + self.send_footer() def send_raw(self, mimetype='text/plain', args=[]): if not self.can_read(): - send_title(None, msg_text='Read access denied by ACLs', msg_type='notice') + self.send_title(msg_text='Read access denied by ACLs', msg_type='notice') return + emit_header(self._mtime(), mimetype) if 'maxwidth' in args: import subprocess - emit_header(mimetype) sys.stdout.flush() - subprocess.check_call(['gm', 'convert', self._filename(), + subprocess.check_call(['convert', self._filename(), + '-auto-orient', '-orient', 'TopLeft', '-scale', args['maxwidth'].value + ' >', '-']) else: body = self.get_raw_body() - emit_header(mimetype) print(body) def _write_file(self, data): @@ -795,20 +797,27 @@ class Page: os.remove(name) except OSError, err: if err.errno != errno.ENOENT: raise err - os.makedirs(os.path.split(name)[0]) + path = os.path.split(name)[0] + if path and not os.path.exists(path): + os.makedirs(path) os.rename(tmp_filename, name) def save(self, newdata, changelog): if not self.can_write(): - self.msg_text = 'Write access denied by ACLs' - self.msg_type = 'error' + self.msg_text = 'Write access denied by Access Control List' + return + if not is_privileged(): + self.msg_text = 'Unauthenticated access denied' return self._write_file(newdata) rc = 0 - if post_edit_hook: + if config_get('post_edit_hook'): import subprocess - cmd = [ post_edit_hook, data_dir + '/' + self.page_name, remote_user(), remote_host(), changelog] + cmd = [ + config_get('post_edit_hook'), + self.page_name, remote_user(), + remote_host(), changelog ] child = subprocess.Popen(cmd, stdout=subprocess.PIPE, close_fds=True) output = child.stdout.read() rc = child.wait() @@ -822,6 +831,7 @@ class Page: try: exec(open("geekigeeki.conf.py").read()) + os.chdir(config_get('data_dir', 'data')) form = cgi.FieldStorage() action = form.getvalue('a', 'get') handler = globals().get('handle_' + action) @@ -830,13 +840,9 @@ try: else: send_httperror("403 Forbidden", query_string()) +except HttpException, e: + send_httperror(e.error, query=e.query) except Exception: - import traceback - msg_text = traceback.format_exc() - if title_done: - send_guru(msg_text, "error") - else: - send_title(None, msg_text=msg_text) - send_footer() + send_httperror("500 Internal Server Error", query=query_string(), trace=True) sys.stdout.flush()