X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;f=geekigeeki.py;h=4b95368aaa6759eb60c0009bfa6fd913308883e2;hb=6225c1ca1fbc0f268d4b10cc391cd9e951171a4d;hp=fbf3d93e96b3e1ac322b19055aed56c7d7b3e159;hpb=4b7c0eb2e64ba9ac910dd57c545c2a455077632b;p=geekigeeki.git
diff --git a/geekigeeki.py b/geekigeeki.py
index fbf3d93..4b95368 100755
--- a/geekigeeki.py
+++ b/geekigeeki.py
@@ -18,27 +18,27 @@
# You should have received a copy of the GNU General Public License
# along with this program. If not, see .
-__version__ = '$Id$'[4:12]
+__version__ = '4.0-' + '$Id$'[4:11]
from time import clock
start_time = clock()
+title_done = False
import cgi, sys, os, re, errno, stat
-# Regular expression defining a WikiWord
-# (but this definition is also assumed in other places)
-word_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
+image_ext = 'png|gif|jpg|jpeg|bmp|ico'
+video_ext = "ogg|ogv|oga" # Not supported by Firefox 3.5: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt
+image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")", re.IGNORECASE)
+video_re = re.compile(r".*\.(" + video_ext + ")", re.IGNORECASE)
# FIXME: we accept stuff like foo/../bar and we shouldn't
-file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$")
-img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-video_re = re.compile(r"^.*\.(ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$")
-link_re = re.compile(r"(?:\[\[|{{)([^\s\|]+)(?:\s*\|\s*([^\]]+)|)(?:\]\]|}})")
-ext_re = re.compile(r"\.([^\./]+)$")
-
-title_done = False
+file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)")
+url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S")
+ext_re = re.compile(r"\.([^\./]+)$")
# CGI stuff ---------------------------------------------------------
+def config_get(key, default=None):
+ return globals().get(key, default)
+
def script_name():
return os.environ.get('SCRIPT_NAME', '')
@@ -50,7 +50,7 @@ def query_string():
return os.environ.get('QUERY_STRING', '') or 'FrontPage'
def privileged_path():
- return privileged_url or script_name()
+ return config_get('privileged_url') or script_name()
def remote_user():
user = os.environ.get('REMOTE_USER', '')
@@ -83,6 +83,30 @@ def relative_url(pathname, privileged=False):
def permalink(s):
return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip())
+def humanlink(s):
+ return re.sub(r'(?:.*[/:]|)([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' '))
+
+# Split arg lists like "blah|blah blah| width=100 | align = center",
+# return a list containing anonymous arguments and a map containing the named arguments
+def parse_args(s):
+ args = []
+ kvargs = {}
+ for arg in s.strip('<[{}]>').split('|'):
+ m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg)
+ if m is not None:
+ kvargs[m.group(1)] = m.group(2)
+ else:
+ args.append(arg.strip())
+ return (args, kvargs)
+
+def url_args(kvargs):
+ argv = []
+ for k, v in kvargs.items():
+ argv.append(k + '=' + v)
+ if argv:
+ return '?' + '&'.join(argv)
+ return ''
+
# Formatting stuff --------------------------------------------------
def emit_header(mime_type="text/html"):
print("Content-type: " + mime_type + "; charset=utf-8\n")
@@ -108,32 +132,33 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal
print(' "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">')
print('')
- print("
')
- print link_tag('FrontPage', site_icon or 'Home', 'navlink')
+ print link_tag('FrontPage', config_get('site_icon', 'Home'), cssclass='navlink')
if name:
- print('
' + link_tag('?fullsearch=' + name, text, 'navlink') + ' ')
+ print('
' + link_tag('?fullsearch=' + name, text, cssclass='navlink') + ' ')
else:
print('
' + text + ' ')
- print(' | ' + link_tag('FindPage', 'Find Page', 'navlink'))
- if 'history_url' in globals():
- print(' |
Recent Changes')
+ print(' | ' + link_tag('FindPage', 'Find Page', cssclass='navlink'))
+ if history:
+ print(' |
Recent Changes')
if name:
- print(' |
Page History')
+ print(' |
Page History')
if name:
- print(' | ' + link_tag('?raw=' + name, 'Raw Text', 'navlink'))
- if privileged_url is not None:
+ print(' | ' + link_tag(name + '?a=raw', 'Raw Text', cssclass='navlink'))
+ if config_get('privileged_url') is not None:
if writable:
- print(' | ' + link_tag('?edit=' + name, 'Edit', 'navlink', privileged=True))
+ print(' | ' + link_tag('?a=edit&q=' + name, 'Edit', cssclass='navlink', privileged=True))
else:
- print(' | ' + link_tag(name, 'Login', 'navlink', privileged=True))
+ print(' | ' + link_tag(name, 'Login', cssclass='navlink', privileged=True))
else:
print(' |
Immutable Page')
@@ -175,45 +200,50 @@ def send_httperror(status="403 Not Found", query=""):
send_title(None, msg_text=("%s: on query '%s'" % (status, query)))
send_footer()
-def link_tag(params, text=None, link_class=None, privileged=False):
+def link_tag(dest, text=None, privileged=False, **kvargs):
if text is None:
- text = params # default
- elif img_re.match(text):
+ text = humanlink(dest)
+ elif image_re.match(text):
text = '
'
+ link_class = kvargs.get('class', kvargs.get('cssclass', None))
if not link_class:
- if is_external_url(params):
+ if is_external_url(dest):
link_class = 'external'
- elif file_re.match(params) and Page(params).exists():
+ elif file_re.match(dest) and Page(dest).exists():
link_class = 'wikilink'
else:
- params = nonexist_pfx + params
+ text = config_get('nonexist_pfx', '') + text
link_class = 'nonexistent'
- classattr = 'class="%s" ' % link_class
# Prevent crawlers from following links potentially added by spammers or to generated pages
+ nofollow = ''
if link_class == 'external' or link_class == 'navlink':
- classattr += 'rel="nofollow"'
+ nofollow = 'rel="nofollow" '
- return '
%s' % (classattr, relative_url(params, privileged=privileged), text)
+ return '
%s' % (link_class, nofollow, relative_url(dest, privileged=privileged), text)
-def link_inline(name, descr=None, args=''):
- if not descr: descr = name
+def link_inline(name, descr=None, kvargs={}):
+ if not descr: descr = humanlink(name)
url = relative_url(name)
if video_re.match(name):
- return '
' % url
- elif img_re.match(name):
- return '
' % (url, url + args, descr)
+ return '
' % url
+ elif image_re.match(name):
+ return '
' % (url, url + url_args(kvargs), descr)
elif file_re.match(name) and not ext_re.search(name): # FIXME: this guesses a wiki page
- return Page(name).send_naked()
+ return Page(name).send_naked(kvargs)
else:
return '
' \
% (url, url, name)
# Search ---------------------------------------------------
-def handle_fullsearch(needle):
- send_title(None, 'Full text search for "%s"' % (needle))
+def print_search_stats(hits, searched):
+ print("
%d hits out of %d pages searched.
" % (hits, searched))
+
+def handle_fullsearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Full text search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
hits = []
@@ -239,9 +269,9 @@ def handle_fullsearch(needle):
print_search_stats(len(hits), len(all_pages))
-def handle_titlesearch(needle):
- # TODO: check needle is legal -- but probably we can just accept any RE
- send_title(None, "Title search for \"" + needle + '"')
+def handle_titlesearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Title search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
all_pages = page_list()
@@ -254,22 +284,19 @@ def handle_titlesearch(needle):
print_search_stats(len(hits), len(all_pages))
-def print_search_stats(hits, searched):
- print("
%d hits out of %d pages searched.
" % (hits, searched))
-
-def handle_raw(pagename):
+def handle_raw(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
Page(pagename).send_raw()
-def handle_edit(pagename):
+def handle_edit(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
- pg = Page(pagename)
+ pg = Page(form['q'].value)
if 'save' in form:
if form['file'].value:
pg.save(form['file'].file.read(), form['changelog'].value)
@@ -286,29 +313,42 @@ def handle_edit(pagename):
text = form['savetext'].value
pg.send_editor(text)
+def handle_get(pagename, form):
+ if file_re.match(pagename):
+ # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension!
+ from mimetypes import MimeTypes
+ mimetype, encoding = MimeTypes().guess_type(pagename)
+ if mimetype:
+ Page(pagename).send_raw(mimetype=mimetype, args=form)
+ else:
+ Page(pagename).format()
+ else:
+ send_httperror("403 Forbidden", pagename)
+
# Used by macros/WordIndex and macros/TitleIndex
def make_index_key():
links = ['
%s' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz']
return '
' + ' | '.join(links) + '
'
-def page_list(dirname = None, re = word_re):
- return sorted(filter(re.match, os.listdir(dirname or data_dir)))
+def page_list(dirname=None, search_re=None):
+ if search_re is None:
+ # FIXME: WikiWord is too restrictive now!
+ search_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
+ return sorted(filter(search_re.match, os.listdir(dirname or config_get('data_dir', 'data'))))
def send_footer(mod_string=None):
- if globals().get('debug_cgi', False):
+ if config_get('debug_cgi', False):
cgi.print_arguments()
cgi.print_form(form)
cgi.print_environ()
- print('''
-')
+ link_inline("sys/footer", kvargs= { 'LAST_MODIFIED': mod_string })
+ print("")
+
+def _macro_ELAPSED_TIME(*args, **kvargs):
+ return "%03f" % (clock() - start_time)
+
+def _macro_VERSION(*args, **kvargs):
+ return __version__
class WikiFormatter:
"""Object that turns Wiki markup into HTML.
@@ -316,8 +356,9 @@ class WikiFormatter:
All formatting commands can be parsed one line at a time, though
some state is carried over between lines.
"""
- def __init__(self, raw):
+ def __init__(self, raw, kvargs=None):
self.raw = raw
+ self.kvargs = kvargs or {}
self.h_level = 0
self.in_pre = self.in_html = self.in_table = self.in_li = False
self.in_header = True
@@ -329,6 +370,7 @@ class WikiFormatter:
"**": ["b", False],
"##": ["tt", False],
"__": ["u", False],
+ "--": ["del", False],
"^^": ["sup", False],
",,": ["sub", False],
"''": ["em", False], # LEGACY
@@ -341,6 +383,9 @@ class WikiFormatter:
style[1] = not style[1]
return ['', '<'][style[1]] + style[0] + '>'
+ def _glyph_repl(self, word):
+ return '—'
+
def _tit_repl(self, word):
if self.h_level:
result = '
\n' % self.h_level
@@ -358,48 +403,37 @@ class WikiFormatter:
return self._undent() + '\n
\n' % (len(word) - 2)
def _macro_repl(self, word):
- m = re.compile("\<\<([^\s\|\>]+)(?:\s*\|\s*([^\>]+)|)\>\>").match(word)
- name = m.group(1)
- argv = [name]
- if m.group(2):
- argv.extend(m.group(2).split('|'))
- argv = list(map(str.strip, argv))
-
- macro = globals().get('_macro_' + name)
- if not macro:
- try:
- exec(open("macros/" + name + ".py").read(), globals())
- except IOError as err:
- if err.errno == errno.ENOENT: pass
- macro = globals().get('_macro_' + name)
- if macro:
- return macro(argv)
- else:
- msg = '<<' + '|'.join(argv) + '>>'
+ try:
+ args, kvargs = parse_args(word)
+ if args[0] in self.kvargs:
+ return self.kvargs[args[0]]
+ macro = globals().get('_macro_' + args[0])
+ if not macro:
+ exec(open("macros/" + args[0] + ".py").read(), globals())
+ macro = globals().get('_macro_' + args[0])
+ return macro(*args, **kvargs)
+ except Exception, e:
+ msg = cgi.escape(word) + ": " + cgi.escape(e.message)
if not self.in_html:
msg = '
' + msg + ''
return msg
def _hurl_repl(self, word):
- m = link_re.match(word)
- return link_tag(m.group(1), m.group(2))
+ args, kvargs = parse_args(word)
+ return link_tag(*args, **kvargs)
def _inl_repl(self, word):
- (name, descr) = link_re.match(word).groups()
-
- if descr:
- argv = descr.split('|')
- descr = argv.pop(0)
- args = ''
- if argv:
- args = '?' + '&'.join(argv)
-
- # The "extthumb" nonsense works around a limitation of the HTML block model
- return '
' \
- + link_inline(name, descr, args) \
+ args, kvargs = parse_args(word)
+ name = args.pop(0)
+ if len(args):
+ descr = args.pop(0)
+ # This double div nonsense works around a limitation of the HTML block model
+ return '
' \
+ + '
' \
+ + link_inline(name, descr, kvargs) \
+ '
' + descr + '
'
else:
- return link_inline(name, name)
+ return link_inline(name, None, kvargs)
def _html_repl(self, word):
if not self.in_html and word.startswith('
')
scan_re = re.compile(r"""(?:
- # Styles and formatting
- (?P \*\*|'''|//|''|\#\#|``|__|\^\^|,,)
+ # Styles and formatting ("--" must cling to a word to disambiguate it from the dash)
+ (?P \*\* | // | \#\# | __ | --\b | \b-- | \^\^ | ,, | ''' | '' | `` )
| (?P \={2,6})
| (?P
\\\\)
| (?P ^-{3,})
| (?P \b( FIXME | TODO | DONE )\b )
+ | (?P --)
# Links
| (?P \<\<([^\s\|\>]+)(?:\s*\|\s*([^\>]+)|)\>\>)
| (?P \[\[([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\]\])
# Inline HTML
- | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
+ | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
| (?P ( /\s*> | (br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])> ) )
| (?P [<>&] )
# Auto links (LEGACY)
- | (?P \b[a-zA-Z0-9_/-]+\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt))
+ | (?P \b[a-zA-Z0-9_/-]+\.(""" + image_ext + "|" + video_ext + r"""))
| (?P \b(?:[A-Z][a-z]+){2,}\b)
| (?P (http|https|ftp|mailto)\:[^\s'\"]+\S)
| (?P [-\w._+]+\@[\w.-]+)
- # Lists, divs, spans
+ # Lists, divs, spans and inline objects
| (?P ^\s+[\*\#]\s+)
| (?P \{\{\{|\s*\}\}\})
| (?P \{\{([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\}\})
@@ -591,16 +626,16 @@ class Page:
return re.sub('([a-z])([A-Z])', r'\1 \2', self.page_name)
def _filename(self):
- return os.path.join(data_dir, self.page_name)
+ return os.path.join(config_get('data_dir', 'data'), self.page_name)
def _tmp_filename(self):
- return os.path.join(data_dir, ('#' + self.page_name.replace('/','_') + '.' + str(os.getpid()) + '#'))
+ return os.path.join(config_get('data_dir', 'data'), ('#' + self.page_name.replace('/','_') + '.' + str(os.getpid()) + '#'))
def exists(self):
try:
os.stat(self._filename())
return True
- except OSError as err:
+ except OSError, err:
if err.errno == errno.ENOENT:
return False
raise err
@@ -608,10 +643,10 @@ class Page:
def get_raw_body(self, default=None):
try:
return open(self._filename(), 'rb').read()
- except IOError as err:
+ except IOError, err:
if err.errno == errno.ENOENT:
if default is None:
- default = '//[[?edit=%s|Describe %s]]//' % (self.page_name, self.page_name)
+ default = '//[[?a=edit&q=%s|Describe %s]]//' % (self.page_name, self.page_name)
return default
if err.errno == errno.EISDIR:
return self.format_dir()
@@ -620,16 +655,17 @@ class Page:
def format_dir(self):
out = '== '
pathname = ''
- for dirname in self.page_name.split('/'):
+ for dirname in self.page_name.strip('/').split('/'):
pathname = (pathname + '/' + dirname) if pathname else dirname
out += '[[' + pathname + '|' + dirname + ']]/'
out += ' ==\n'
for filename in page_list(self._filename(), file_re):
- if img_re.match(filename):
- if image_maxwidth:
- maxwidth_arg = '|maxwidth=' + str(image_maxwidth)
- out += '{{' + self.page_name + '/' + filename + '|' + filename + maxwidth_arg + '}}\n'
+ if image_re.match(filename):
+ maxwidth = config_get(image_maxwidth, '')
+ if maxwidth:
+ maxwidth = ' | maxwidth=' + str(maxwidth)
+ out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth + ' | class=thumbleft}}\n'
else:
out += ' * [[' + self.page_name + '/' + filename + ']]\n'
return out
@@ -646,7 +682,7 @@ class Page:
break
self._pragmas[m.group(1)] = m.group(2).strip()
#print "bernie: pragmas[" + m.group(1) + "] = " + m.group(2) + "
\n"
- except IOError as err:
+ except IOError, err:
if err.errno != errno.ENOENT and err.errno != errno.EISDIR:
raise er
return self._pragmas
@@ -675,9 +711,9 @@ class Page:
def can_read(self):
return self.can("read", True)
- def send_naked(self):
+ def send_naked(self, kvargs=None):
if self.can_read():
- WikiFormatter(self.get_raw_body()).print_html()
+ WikiFormatter(self.get_raw_body(), kvargs).print_html()
else:
send_guru("Read access denied by ACLs", "notice")
@@ -697,11 +733,11 @@ class Page:
try:
from time import localtime, strftime
modtime = localtime(os.stat(self._filename())[stat.ST_MTIME])
- except OSError as err:
+ except OSError, err:
if err.errno != errno.ENOENT:
raise err
return None
- return strftime(datetime_fmt, modtime)
+ return strftime(config_get(datetime_fmt, '%a %d %b %Y %I:%M %p'), modtime)
def send_editor(self, preview=None):
send_title(None, 'Edit ' + self.split_title(), msg_text=self.msg_text, msg_type=self.msg_type)
@@ -718,7 +754,7 @@ class Page:
+ ' from ' + cgi.escape(get_hostname(remote_host()))
+ '
'))
print('