X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;f=geekigeeki.py;h=a0dc72ca0629280be7792e5c57d3dee60c5b2525;hb=6002848303f10b453acd930707ea436ec1183725;hp=4450aead0d3617848117f6c7ad6098a758b8877d;hpb=042882ca5918a6f84a176e4ef9dc19428ac4566f;p=geekigeeki.git
diff --git a/geekigeeki.py b/geekigeeki.py
index 4450aea..a0dc72c 100755
--- a/geekigeeki.py
+++ b/geekigeeki.py
@@ -3,7 +3,7 @@
#
# Copyright 1999, 2000 Martin Pool
# Copyright 2002 Gerardo Poggiali
-# Copyright 2007, 2008 Bernie Innocenti
+# Copyright 2007, 2008, 2009 Bernie Innocenti
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
@@ -18,24 +18,22 @@
# You should have received a copy of the GNU General Public License
# along with this program. If not, see .
-__version__ = '$Id$'[4:12]
+__version__ = '4.0-' + '$Id$'[4:11]
from time import clock
start_time = clock()
+title_done = False
import cgi, sys, os, re, errno, stat
-# Regular expression defining a WikiWord
-# (but this definition is also assumed in other places)
-word_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
+image_ext = 'png|gif|jpg|jpeg|bmp|ico'
+video_ext = "ogg|ogv|oga" # Not supported by Firefox 3.5: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt
+image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")", re.IGNORECASE)
+video_re = re.compile(r".*\.(" + video_ext + ")", re.IGNORECASE)
# FIXME: we accept stuff like foo/../bar and we shouldn't
-file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$")
-img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-video_re = re.compile(r"^.*\.(ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$")
-link_re = re.compile("(?:\[\[|{{)([^\s\|]+)(?:\s*\|\s*([^\]]+)|)(?:\]\]|}})")
-
-title_done = False
+file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)")
+url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S")
+ext_re = re.compile(r"\.([^\./]+)$")
# CGI stuff ---------------------------------------------------------
def script_name():
@@ -77,35 +75,49 @@ def relative_url(pathname, privileged=False):
else:
url = script_name()
pathname = url + '/' + pathname
- return pathname
+ return cgi.escape(pathname, quote=True)
def permalink(s):
return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip())
+def humanlink(s):
+ return re.sub(r'(?:.*[/:]|)([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' '))
+
+# Split arg lists like "blah| blah blah| width=100 | align = center",
+# return a list containing anonymous arguments and a map containing the named arguments
+def parse_args(s):
+ args = []
+ kwargs = {}
+ for arg in s.strip('<[{}]>').split('|'):
+ m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg)
+ if m is not None:
+ kwargs[m.group(1)] = m.group(2)
+ else:
+ args.append(arg.strip())
+ return (args, kwargs)
+
+def url_args(kvargs):
+ argv = []
+ for k, v in kvargs.items():
+ argv.append(k + '=' + v)
+ if argv:
+ return '?' + '&'.join(argv)
+ return ''
+
# Formatting stuff --------------------------------------------------
def emit_header(mime_type="text/html"):
- print "Content-type: " + mime_type + "; charset=utf-8\n"
-
-def sendfile(dest_file, src_file):
- """Efficiently copy file data between file descriptors"""
- while 1:
- data = src_file.read(65536)
- if not data: break
- dest_file.write(data)
+ print("Content-type: " + mime_type + "; charset=utf-8\n")
def send_guru(msg_text, msg_type):
if not msg_text: return
- print ''
+ print('')
if msg_type == 'error':
- print ' Software Failure. Press left mouse button to continue.\n'
- print msg_text
+ print(' Software Failure. Press left mouse button to continue.\n')
+ print(msg_text)
if msg_type == 'error':
- print '\n Guru Meditation #DEADBEEF.ABADC0DE'
- print ' '
- try:
- sendfile(sys.stdout, open('gurumeditation.js', 'rb'))
- except IOError, err:
- pass
+ print '\n Guru Meditation #DEADBEEF.ABADC0DE'
+ print(' ' \
+ % relative_url('sys/GuruMeditation.js'))
def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=False):
global title_done
@@ -113,103 +125,121 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal
# Head
emit_header()
- print ''
- print ''
+ print('')
+ print('')
- print "%s: %s " % (site_name, text)
- print ' '
+ print("%s: %s " % (site_name, text))
+ print(' ')
if not name:
- print ' '
+ print(' ')
for meta in meta_urls:
http_equiv, content = meta
- print ' ' % (http_equiv, relative_url(content))
+ print(' ' % (http_equiv, relative_url(content)))
for link in link_urls:
rel, href = link
- print ' ' % (rel, relative_url(href))
+ print(' ' % (rel, relative_url(href)))
if name and writable and privileged_url is not None:
- print ' ' \
- % (privileged_path() + '?edit=' + name)
+ print(' ' \
+ % (privileged_path() + '?a=edit&q=' + name))
if history_url is not None:
- print ' ' \
- % relative_url(history_url + '?a=rss')
+ print(' ' \
+ % relative_url(history_url + '?a=rss'))
- print ''
+ print('')
# Body
if name and writable and privileged_url is not None:
- print ''
+ print('')
else:
- print ''
+ print('')
title_done = True
send_guru(msg_text, msg_type)
# Navbar
- print ''
+ print('
')
+ print link_tag('FrontPage', site_icon or 'Home', cssclass='navlink')
if name:
- print '
' + link_tag('?fullsearch=' + name, text, 'navlink') + ' '
+ print('
' + link_tag('?fullsearch=' + name, text, cssclass='navlink') + ' ')
else:
- print '
' + text + ' '
- print ' | ' + link_tag('FrontPage', 'Home', 'navlink')
- print ' | ' + link_tag('FindPage', 'Find Page', 'navlink')
+ print('
' + text + ' ')
+ print(' | ' + link_tag('FindPage', 'Find Page', cssclass='navlink'))
if 'history_url' in globals():
- print ' |
Recent Changes '
+ print(' |
Recent Changes ')
if name:
- print ' |
Page History '
+ print(' |
Page History ')
if name:
- print ' | ' + link_tag('?raw=' + name, 'Raw Text', 'navlink')
+ print(' | ' + link_tag(name + '?a=raw', 'Raw Text', cssclass='navlink'))
if privileged_url is not None:
if writable:
- print ' | ' + link_tag('?edit=' + name, 'Edit', 'navlink', privileged=True)
+ print(' | ' + link_tag('?a=edit&q=' + name, 'Edit', cssclass='navlink', privileged=True))
else:
- print ' | ' + link_tag(name, 'Login', 'navlink', privileged=True)
+ print(' | ' + link_tag(name, 'Login', cssclass='navlink', privileged=True))
else:
- print ' |
Immutable Page '
+ print(' |
Immutable Page ')
user = remote_user()
if user != 'AnonymousCoward':
- print ' |
logged in as ' + cgi.escape(user) + ' '
+ print(' |
' + link_tag('User/' + user, user) + ' ')
- print '
'
+ print('
')
def send_httperror(status="403 Not Found", query=""):
- print "Status: %s" % status
+ print("Status: %s" % status)
send_title(None, msg_text=("%s: on query '%s'" % (status, query)))
send_footer()
-def link_tag(params, text=None, link_class=None, privileged=False):
+def link_tag(dest, text=None, privileged=False, **kvargs):
if text is None:
- text = params # default
- elif img_re.match(text):
- text = ' '
+ text = humanlink(dest)
+ elif image_re.match(text):
+ text = ' '
+ link_class = kvargs.get('class', kvargs.get('cssclass', None))
if not link_class:
- if is_external_url(params):
+ if is_external_url(dest):
link_class = 'external'
- elif file_re.match(params) and Page(params).exists():
+ elif file_re.match(dest) and Page(dest).exists():
link_class = 'wikilink'
else:
- params = nonexist_pfx + params
+ text = nonexist_pfx + text
link_class = 'nonexistent'
- classattr = 'class="%s" ' % link_class
# Prevent crawlers from following links potentially added by spammers or to generated pages
+ nofollow = ''
if link_class == 'external' or link_class == 'navlink':
- classattr += 'rel="nofollow"'
-
- return '%s ' % (classattr, relative_url(params, privileged=privileged), text)
+ nofollow = 'rel="nofollow" '
+
+ return '%s ' % (link_class, nofollow, relative_url(dest, privileged=privileged), text)
+
+def link_inline(name, descr=None, kvargs={}):
+ if not descr: descr = humanlink(name)
+ url = relative_url(name)
+ if video_re.match(name):
+ return 'Your browser does not support the HTML5 video tag ' % url
+ elif image_re.match(name):
+ return ' ' % (url, url + url_args(kvargs), descr)
+ elif file_re.match(name) and not ext_re.search(name): # FIXME: this guesses a wiki page
+ return Page(name).send_naked()
+ else:
+ return '' \
+ % (url, url, name)
# Search ---------------------------------------------------
-def handle_fullsearch(needle):
- send_title(None, 'Full text search for "%s"' % (needle))
+def print_search_stats(hits, searched):
+ print("%d hits out of %d pages searched.
" % (hits, searched))
+
+def handle_fullsearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Full text search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
hits = []
@@ -225,47 +255,44 @@ def handle_fullsearch(needle):
hits.sort()
hits.reverse()
- print "")
print_search_stats(len(hits), len(all_pages))
-def handle_titlesearch(needle):
- # TODO: check needle is legal -- but probably we can just accept any RE
- send_title(None, "Title search for \"" + needle + '"')
+def handle_titlesearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Title search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
all_pages = page_list()
- hits = filter(needle_re.search, all_pages)
+ hits = list(filter(needle_re.search, all_pages))
- print "")
print_search_stats(len(hits), len(all_pages))
-def print_search_stats(hits, searched):
- print "%d hits out of %d pages searched.
" % (hits, searched)
-
-def handle_raw(pagename):
+def handle_raw(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
Page(pagename).send_raw()
-def handle_edit(pagename):
+def handle_edit(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
- pg = Page(pagename)
+ pg = Page(form['q'].value)
if 'save' in form:
if form['file'].value:
pg.save(form['file'].file.read(), form['changelog'].value)
@@ -282,11 +309,27 @@ def handle_edit(pagename):
text = form['savetext'].value
pg.send_editor(text)
+def handle_get(pagename, form):
+ if file_re.match(pagename):
+ # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension!
+ from mimetypes import MimeTypes
+ mimetype, encoding = MimeTypes().guess_type(pagename)
+ if mimetype:
+ Page(pagename).send_raw(mimetype=mimetype, args=form)
+ else:
+ Page(pagename).format()
+ else:
+ send_httperror("403 Forbidden", pagename)
+
+# Used by macros/WordIndex and macros/TitleIndex
def make_index_key():
- links = map(lambda ch: '%s ' % (ch, ch), 'abcdefghijklmnopqrstuvwxyz')
- return '
'+ ' | '.join(links) + '
'
+ links = ['%s ' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz']
+ return '' + ' | '.join(links) + '
'
-def page_list(dirname = None, re = word_re):
+def page_list(dirname=None, re=None):
+ if re is None:
+ # FIXME: WikiWord is too restrictive now!
+ re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
return sorted(filter(re.match, os.listdir(dirname or data_dir)))
def send_footer(mod_string=None):
@@ -294,16 +337,17 @@ def send_footer(mod_string=None):
cgi.print_arguments()
cgi.print_form(form)
cgi.print_environ()
- print '''
+ #FIXME link_inline("sys/footer")
+ print('''
'
+ print('last modified %s
' % mod_string)
+ print('')
class WikiFormatter:
"""Object that turns Wiki markup into HTML.
@@ -321,14 +365,15 @@ class WikiFormatter:
self.styles = {
#wiki html enabled?
"//": ["em", False],
- "''": ["em", False],
"**": ["b", False],
- "'''": ["b", False],
"##": ["tt", False],
- "``": ["tt", False],
"__": ["u", False],
+ "--": ["del", False],
"^^": ["sup", False],
- ",,": ["sub", False]
+ ",,": ["sub", False],
+ "''": ["em", False], # LEGACY
+ "'''": ["b", False], # LEGACY
+ "``": ["tt", False], # LEGACY
}
def _b_repl(self, word):
@@ -336,6 +381,9 @@ class WikiFormatter:
style[1] = not style[1]
return ['', '<'][style[1]] + style[0] + '>'
+ def _glyph_repl(self, word):
+ return '—'
+
def _tit_repl(self, word):
if self.h_level:
result = '\n' % self.h_level
@@ -353,70 +401,44 @@ class WikiFormatter:
return self._undent() + '\n
\n' % (len(word) - 2)
def _macro_repl(self, word):
- m = re.compile("\<\<([^\s\|\>]+)(?:\s*\|\s*([^\>]+)|)\>\>").match(word)
- name = m.group(1)
- argv = [name]
- if m.group(2):
- argv.extend(m.group(2).split('|'))
- argv = map(str.strip, argv)
-
- macro = globals().get('_macro_' + name)
- if not macro:
- try:
- execfile("macros/" + name + ".py", globals())
- except IOError, err:
- if err.errno == errno.ENOENT: pass
- macro = globals().get('_macro_' + name)
- if macro:
- return macro(argv)
- else:
- msg = '<<' + '|'.join(argv) + '>>'
+ try:
+ args, kwargs = parse_args(word)
+ macro = globals().get('_macro_' + args[0])
+ if not macro:
+ exec(open("macros/" + name + ".py").read(), globals())
+ macro = globals().get('_macro_' + name)
+ return macro(*args, **kwargs)
+ except Exception:
+ msg = cgi.escape(word)
if not self.in_html:
msg = '' + msg + ' '
return msg
def _hurl_repl(self, word):
- m = link_re.match(word)
- return link_tag(m.group(1), m.group(2))
-
- def _url_repl(self, word):
- return link_tag(word)
-
- def _word_repl(self, word):
- return link_tag(word)
+ args, kvargs = parse_args(word)
+ return link_tag(*args, **kvargs)
def _inl_repl(self, word):
- m = link_re.match(word)
- name = relative_url(m.group(1))
- descr = m.group(2)
-
- if descr:
- argv = descr.split('|')
- descr = argv.pop(0)
- args = ''
- if argv:
- args = '?' + '&'.join(argv)
-
- # The "extthumb" nonsense works around a limitation of the HTML block model
- return '%s
' \
- % (name, name + args, descr, descr)
- elif video_re.match(name):
- return 'Your browser does not support the HTML5 video tag ' % name
+ args, kvargs = parse_args(word)
+ name = args.pop(0)
+ if len(args):
+ descr = args.pop(0)
+ # This double div nonsense works around a limitation of the HTML block model
+ return '' \
+ + '
' \
+ + link_inline(name, descr, kvargs) \
+ + '
' + descr + '
'
else:
- return ' ' % (name, name)
-
- def _img_repl(self, word):
- return self._inl_repl('{{' + word + '}}')
-
- def _email_repl(self, word):
- return '%s ' % (word, word)
+ return link_inline(name, None, kvargs)
def _html_repl(self, word):
+ if not self.in_html and word.startswith(''
return word; # Pass through
def _ent_repl(self, s):
@@ -426,6 +448,21 @@ class WikiFormatter:
'<': '<',
'>': '>'}[s]
+ def _img_repl(self, word): # LEGACY
+ return self._inl_repl('{{' + word + '}}')
+
+ def _word_repl(self, word): # LEGACY
+ if self.in_html: return word # pass through
+ return link_tag(word)
+
+ def _url_repl(self, word): # LEGACY
+ if self.in_html: return word # pass through
+ return link_tag(word)
+
+ def _email_repl(self, word): # LEGACY
+ if self.in_html: return word # pass through
+ return '
%s ' % (word, word)
+
def _li_repl(self, match):
if self.in_li:
return '
'
@@ -492,18 +529,19 @@ class WikiFormatter:
return res
def replace(self, match):
- for rule, hit in match.groupdict().items():
+ for rule, hit in list(match.groupdict().items()):
if hit:
return getattr(self, '_' + rule + '_repl')(hit)
else:
raise "Can't handle match " + repr(match)
def print_html(self):
- print ''
+ print('
')
scan_re = re.compile(r"""(?:
- # Styles and formatting
- (?P \*\*|'''|//|''|\#\#|``|__|\^\^|,,)
+ # Styles and formatting ("--" must cling to a word to disambiguate it from the dash)
+ (?P \*\* | // | \#\# | __ | --\b | \b-- | \^\^ | ,, | ''' | '' | `` )
+ | (?P --)
| (?P \={2,6})
| (?P \\\\)
| (?P ^-{3,})
@@ -514,17 +552,17 @@ class WikiFormatter:
| (?P \[\[([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\]\])
# Inline HTML
- | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
+ | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
| (?P ( /\s*> | (br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])> ) )
| (?P [<>&] )
# Auto links (LEGACY)
- | (?P \b[a-zA-Z0-9_/-]+\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt))
+ | (?P \b[a-zA-Z0-9_/-]+\.(""" + image_ext + "|" + video_ext + r"""))
| (?P \b(?:[A-Z][a-z]+){2,}\b)
| (?P (http|https|ftp|mailto)\:[^\s'\"]+\S)
| (?P [-\w._+]+\@[\w.-]+)
- # Lists, divs, spans
+ # Lists, divs, spans and inline objects
| (?P ^\s+[\*\#]\s+)
| (?P \{\{\{|\s*\}\}\})
| (?P \{\{([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\}\})
@@ -544,9 +582,9 @@ class WikiFormatter:
indent_re = re.compile(r"^\s*")
tr_re = re.compile(r"^\s*\|\|")
eol_re = re.compile(r"\r?\n")
-
# For each line, we scan through looking for magic strings, outputting verbatim any intervening text
- for self.line in eol_re.split(self.raw.expandtabs()):
+ #3.0: for self.line in eol_re.split(str(self.raw.expandtabs(), 'utf-8')):
+ for self.line in eol_re.split(str(self.raw.expandtabs())):
# Skip pragmas
if self.in_header:
if self.line.startswith('#'):
@@ -554,23 +592,24 @@ class WikiFormatter:
self.in_header = False
if self.in_pre:
- print re.sub(pre_re, self.replace, self.line)
+ print(re.sub(pre_re, self.replace, self.line))
else:
if self.in_table and not tr_re.match(self.line):
self.in_table = False
- print ''
+ print('
')
if blank_re.match(self.line):
- print '
'
+ print('
')
else:
indent = indent_re.match(self.line)
- print self._indent_to(len(indent.group(0))) ,
- print re.sub(scan_re, self.replace, self.line)
+ #3.0: print(self._indent_to(len(indent.group(0))), end=' ')
+ print(self._indent_to(len(indent.group(0))))
+ print(re.sub(scan_re, self.replace, self.line))
- if self.in_pre: print '
'
- if self.in_table: print ''
- print self._undent()
- print '
'
+ if self.in_pre: print('')
+ if self.in_table: print('
')
+ print(self._undent())
+ print('
')
class Page:
def __init__(self, page_name):
@@ -586,7 +625,7 @@ class Page:
return os.path.join(data_dir, self.page_name)
def _tmp_filename(self):
- return os.path.join(data_dir, ('#' + self.page_name.replace('/','_') + '.' + `os.getpid()` + '#'))
+ return os.path.join(data_dir, ('#' + self.page_name.replace('/','_') + '.' + str(os.getpid()) + '#'))
def exists(self):
try:
@@ -597,12 +636,14 @@ class Page:
return False
raise err
- def get_raw_body(self):
+ def get_raw_body(self, default=None):
try:
return open(self._filename(), 'rb').read()
except IOError, err:
if err.errno == errno.ENOENT:
- return '' # just doesn't exist, use default
+ if default is None:
+ default = '//[[?a=edit&q=%s|Describe %s]]//' % (self.page_name, self.page_name)
+ return default
if err.errno == errno.EISDIR:
return self.format_dir()
raise err
@@ -610,16 +651,16 @@ class Page:
def format_dir(self):
out = '== '
pathname = ''
- for dirname in self.page_name.split('/'):
+ for dirname in self.page_name.strip('/').split('/'):
pathname = (pathname + '/' + dirname) if pathname else dirname
out += '[[' + pathname + '|' + dirname + ']]/'
out += ' ==\n'
for filename in page_list(self._filename(), file_re):
- if img_re.match(filename):
+ if image_re.match(filename):
if image_maxwidth:
- maxwidth_arg = '|maxwidth=' + str(image_maxwidth)
- out += '{{' + self.page_name + '/' + filename + '|' + filename + maxwidth_arg + '}}\n'
+ maxwidth_arg = ' | maxwidth=' + str(image_maxwidth)
+ out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth_arg + ' | class=thumbleft}}\n'
else:
out += ' * [[' + self.page_name + '/' + filename + ']]\n'
return out
@@ -628,17 +669,17 @@ class Page:
if not '_pragmas' in self.__dict__:
self._pragmas = {}
try:
- f = open(self._filename(), 'rt')
+ file = open(self._filename(), 'rt')
attr_re = re.compile(r"^#(\S*)(.*)$")
- for line in f:
+ for line in file:
m = attr_re.match(line)
if not m:
break
self._pragmas[m.group(1)] = m.group(2).strip()
- #print "bernie: _pragmas[" + m.group(1) + "] = " + m.group(2) + " \n"
+ #print "bernie: pragmas[" + m.group(1) + "] = " + m.group(2) + " \n"
except IOError, err:
if err.errno != errno.ENOENT and err.errno != errno.EISDIR:
- raise err
+ raise er
return self._pragmas
def pragma(self, name, default):
@@ -703,16 +744,17 @@ class Page:
if 'file' in form:
filename = form['file'].value
- print ('Editing ' + self.page_name
+ print(('
Editing ' + self.page_name
+ ' for ' + cgi.escape(remote_user())
+ ' from ' + cgi.escape(get_hostname(remote_host()))
- + '
')
- print '