X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;ds=sidebyside;f=geekigeeki.py;h=eb51489cddb74e6bf2185f6a0e05ac62ebfb6685;hb=ce5a4dbd3de540928b7019a716859a551961105c;hp=ca16283d5ea2f2c78c573b3863a371add9c15aef;hpb=7ce906e26b7c2c91e232508ccd0af6fe9ec4aeec;p=geekigeeki.git
diff --git a/geekigeeki.py b/geekigeeki.py
index ca16283..eb51489 100755
--- a/geekigeeki.py
+++ b/geekigeeki.py
@@ -1,9 +1,9 @@
-#!/usr/bin/python3.0
+#!/usr/bin/python
# -*- coding: utf-8 -*-
#
# Copyright 1999, 2000 Martin Pool
# Copyright 2002 Gerardo Poggiali
-# Copyright 2007, 2008 Bernie Innocenti
+# Copyright 2007, 2008, 2009 Bernie Innocenti
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
@@ -18,30 +18,34 @@
# You should have received a copy of the GNU General Public License
# along with this program. If not, see .
-__version__ = '$Id$'[4:12]
+__version__ = '4.0-' + '$Id$'[4:11]
from time import clock
start_time = clock()
+title_done = False
import cgi, sys, os, re, errno, stat
-# Regular expression defining a WikiWord
-# (but this definition is also assumed in other places)
-word_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
+image_ext = 'png|gif|jpg|jpeg|bmp|ico'
+video_ext = "ogg|ogv|oga" # Not supported by Firefox 3.5: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt
+image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")", re.IGNORECASE)
+video_re = re.compile(r".*\.(" + video_ext + ")", re.IGNORECASE)
# FIXME: we accept stuff like foo/../bar and we shouldn't
-file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$")
-img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico)$", re.IGNORECASE)
-url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$")
-link_re = re.compile("(?:\[\[|{{)([^\s\|]+)(?:\s*\|\s*([^\]]+)|)(?:\]\]|}})")
-
-title_done = False
-
+file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)")
+url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S")
+ext_re = re.compile(r"\.([^\./]+)$")
# CGI stuff ---------------------------------------------------------
-
def script_name():
return os.environ.get('SCRIPT_NAME', '')
+def query_string():
+ path_info = os.environ.get('PATH_INFO', '')
+ if len(path_info) and path_info[0] == '/':
+ return path_info[1:] or 'FrontPage'
+ else:
+ return os.environ.get('QUERY_STRING', '') or 'FrontPage'
+
def privileged_path():
return privileged_url or script_name()
@@ -61,18 +65,45 @@ def get_hostname(addr):
except Exception:
return addr
+def is_external_url(pathname):
+ return (url_re.match(pathname) or pathname.startswith('/'))
+
def relative_url(pathname, privileged=False):
- if not (url_re.match(pathname) or pathname.startswith('/')):
+ if not is_external_url(pathname):
if privileged:
url = privileged_path()
else:
url = script_name()
pathname = url + '/' + pathname
- return pathname
+ return cgi.escape(pathname, quote=True)
def permalink(s):
return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip())
+def humanlink(s):
+ return re.sub(r'(?:.*[/:]|)([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' '))
+
+# Split arg lists like "blah|blah blah| width=100 | align = center",
+# return a list containing anonymous arguments and a map containing the named arguments
+def parse_args(s):
+ args = []
+ kvargs = {}
+ for arg in s.strip('<[{}]>').split('|'):
+ m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg)
+ if m is not None:
+ kvargs[m.group(1)] = m.group(2)
+ else:
+ args.append(arg.strip())
+ return (args, kvargs)
+
+def url_args(kvargs):
+ argv = []
+ for k, v in kvargs.items():
+ argv.append(k + '=' + v)
+ if argv:
+ return '?' + '&'.join(argv)
+ return ''
+
# Formatting stuff --------------------------------------------------
def emit_header(mime_type="text/html"):
print("Content-type: " + mime_type + "; charset=utf-8\n")
@@ -84,30 +115,9 @@ def send_guru(msg_text, msg_type):
print(' Software Failure. Press left mouse button to continue.\n')
print(msg_text)
if msg_type == 'error':
- print('\n Guru Meditation #DEADBEEF.ABADC0DE')
- print('')
- # FIXME: This little JS snippet is harder to pass than ACID 3.0
- print("""
- """)
+ print '\n Guru Meditation #DEADBEEF.ABADC0DE'
+ print('' \
+ % relative_url('sys/GuruMeditation.js'))
def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=False):
global title_done
@@ -134,7 +144,7 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal
if name and writable and privileged_url is not None:
print(' ' \
- % (privileged_path() + '?edit=' + name))
+ % (privileged_path() + '?a=edit&q=' + name))
if history_url is not None:
print(' ' \
@@ -144,7 +154,7 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal
# Body
if name and writable and privileged_url is not None:
- print('')
+ print('')
else:
print('')
@@ -153,31 +163,31 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal
# Navbar
print('')
+ print link_tag('FrontPage', site_icon or 'Home', cssclass='navlink')
if name:
- print('
' + link_tag('?fullsearch=' + name, text, 'navlink') + ' ')
+ print('
' + link_tag('?fullsearch=' + name, text, cssclass='navlink') + ' ')
else:
print('
' + text + ' ')
- print(' | ' + link_tag('FrontPage', 'Home', 'navlink'))
- print(' | ' + link_tag('FindPage', 'Find Page', 'navlink'))
+ print(' | ' + link_tag('FindPage', 'Find Page', cssclass='navlink'))
if 'history_url' in globals():
print(' |
Recent Changes')
if name:
print(' |
Page History')
if name:
- print(' | ' + link_tag('?raw=' + name, 'Raw Text', 'navlink'))
+ print(' | ' + link_tag(name + '?a=raw', 'Raw Text', cssclass='navlink'))
if privileged_url is not None:
if writable:
- print(' | ' + link_tag('?edit=' + name, 'Edit', 'navlink', privileged=True))
+ print(' | ' + link_tag('?a=edit&q=' + name, 'Edit', cssclass='navlink', privileged=True))
else:
- print(' | ' + link_tag(name, 'Login', 'navlink', privileged=True))
+ print(' | ' + link_tag(name, 'Login', cssclass='navlink', privileged=True))
else:
print(' |
Immutable Page')
user = remote_user()
if user != 'AnonymousCoward':
- print(' |
logged in as ' + cgi.escape(user) + '')
+ print(' |
' + link_tag('User/' + user, user) + '')
print('
')
@@ -186,23 +196,50 @@ def send_httperror(status="403 Not Found", query=""):
send_title(None, msg_text=("%s: on query '%s'" % (status, query)))
send_footer()
-def link_tag(params, text=None, ss_class=None, privileged=False):
+def link_tag(dest, text=None, privileged=False, **kvargs):
if text is None:
- text = params # default
- classattr = ''
- if ss_class:
- classattr += 'class="%s" ' % ss_class
- # Prevent crawlers from following links potentially added by spammers or to generated pages
- if ss_class == 'external' or ss_class == 'navlink':
- classattr += 'rel="nofollow" '
- elif url_re.match(params):
- classattr += 'rel="nofollow" '
- return '%s' % (classattr, relative_url(params, privileged=privileged), text)
+ text = humanlink(dest)
+ elif image_re.match(text):
+ text = ''
+
+ link_class = kvargs.get('class', kvargs.get('cssclass', None))
+ if not link_class:
+ if is_external_url(dest):
+ link_class = 'external'
+ elif file_re.match(dest) and Page(dest).exists():
+ link_class = 'wikilink'
+ else:
+ text = nonexist_pfx + text
+ link_class = 'nonexistent'
+
+ # Prevent crawlers from following links potentially added by spammers or to generated pages
+ nofollow = ''
+ if link_class == 'external' or link_class == 'navlink':
+ nofollow = 'rel="nofollow" '
+
+ return '%s' % (link_class, nofollow, relative_url(dest, privileged=privileged), text)
+
+def link_inline(name, descr=None, kvargs={}):
+ if not descr: descr = humanlink(name)
+ url = relative_url(name)
+ if video_re.match(name):
+ return '' % url
+ elif image_re.match(name):
+ return '' % (url, url + url_args(kvargs), descr)
+ elif file_re.match(name) and not ext_re.search(name): # FIXME: this guesses a wiki page
+ return Page(name).send_naked(kvargs)
+ else:
+ return '' \
+ % (url, url, name)
# Search ---------------------------------------------------
-def handle_fullsearch(needle):
- send_title(None, 'Full text search for "%s"' % (needle))
+def print_search_stats(hits, searched):
+ print("%d hits out of %d pages searched.
" % (hits, searched))
+
+def handle_fullsearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Full text search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
hits = []
@@ -220,17 +257,17 @@ def handle_fullsearch(needle):
print("")
for (count, page_name) in hits:
- print('' + Page(page_name).link_to())
- print(' . . . . ' + repr(count))
+ print('
' + link_tag(page_name))
+ print(' . . . . ' + `count`)
print(['match', 'matches'][count != 1])
print('
')
print("
")
print_search_stats(len(hits), len(all_pages))
-def handle_titlesearch(needle):
- # TODO: check needle is legal -- but probably we can just accept any RE
- send_title(None, "Title search for \"" + needle + '"')
+def handle_titlesearch(query, form):
+ needle = form['q'].value
+ send_title(None, 'Title search for "' + needle + '"')
needle_re = re.compile(needle, re.IGNORECASE)
all_pages = page_list()
@@ -238,27 +275,24 @@ def handle_titlesearch(needle):
print("")
print_search_stats(len(hits), len(all_pages))
-def print_search_stats(hits, searched):
- print("%d hits out of %d pages searched.
" % (hits, searched))
-
-def handle_raw(pagename):
+def handle_raw(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
Page(pagename).send_raw()
-def handle_edit(pagename):
+def handle_edit(pagename, form):
if not file_re.match(pagename):
send_httperror("403 Forbidden", pagename)
return
- pg = Page(pagename)
+ pg = Page(form['q'].value)
if 'save' in form:
if form['file'].value:
pg.save(form['file'].file.read(), form['changelog'].value)
@@ -275,11 +309,27 @@ def handle_edit(pagename):
text = form['savetext'].value
pg.send_editor(text)
+def handle_get(pagename, form):
+ if file_re.match(pagename):
+ # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension!
+ from mimetypes import MimeTypes
+ mimetype, encoding = MimeTypes().guess_type(pagename)
+ if mimetype:
+ Page(pagename).send_raw(mimetype=mimetype, args=form)
+ else:
+ Page(pagename).format()
+ else:
+ send_httperror("403 Forbidden", pagename)
+
+# Used by macros/WordIndex and macros/TitleIndex
def make_index_key():
links = ['%s' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz']
- return ''+ ' | '.join(links) + '
'
+ return '' + ' | '.join(links) + '
'
-def page_list(dirname = None, re = word_re):
+def page_list(dirname=None, re=None):
+ if re is None:
+ # FIXME: WikiWord is too restrictive now!
+ re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
return sorted(filter(re.match, os.listdir(dirname or data_dir)))
def send_footer(mod_string=None):
@@ -287,16 +337,14 @@ def send_footer(mod_string=None):
cgi.print_arguments()
cgi.print_form(form)
cgi.print_environ()
- print('''
-')
+ link_inline("sys/footer", kvargs= { 'LAST_MODIFIED': mod_string })
+ print("