X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;ds=sidebyside;f=geekigeeki.py;h=9d2bf9e12e8dea0617622960871baa95a990d753;hb=bfa2f63dc09c936f10b7cc3c1ee0c1c136af3591;hp=678754e3a9633d29fc779eeed48eeddb8fa9fa87;hpb=8d7f457924b6c4bde8486d1fd413516082ceedf9;p=geekigeeki.git
diff --git a/geekigeeki.py b/geekigeeki.py
index 678754e..9d2bf9e 100755
--- a/geekigeeki.py
+++ b/geekigeeki.py
@@ -8,35 +8,38 @@
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
-# (at your option) any later version.
-#
-# This program is distributed in the hope that it will be useful, but
-# WITHOUT ANY WARRANTY; without even the implied warranty of
-# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-# General Public License for more details.
-#
-# You should have received a copy of the GNU General Public License
-# along with this program. If not, see .
+# (at your option) any later version. You should have received a copy
+# of the GNU General Public License along with this program.
+# If not, see .
-__version__ = '$Id$'[4:12]
+__version__ = '4.0-' + '$Id$'[4:11]
-from time import clock
+from time import clock, localtime, gmtime, strftime
start_time = clock()
title_done = False
-import cgi, sys, os, re, errno, stat
+import cgi, sys, os, re, errno, stat, glob
+image_ext = 'png|gif|jpg|jpeg|bmp|ico'
+video_ext = "ogg|ogv|oga" # Not supported by Firefox 3.5: mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt
+image_re = re.compile(r".*\.(" + image_ext + "|" + video_ext + ")", re.IGNORECASE)
+video_re = re.compile(r".*\.(" + video_ext + ")", re.IGNORECASE)
# FIXME: we accept stuff like foo/../bar and we shouldn't
-file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$")
-img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-video_re = re.compile(r"^.*\.(ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE)
-url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$")
-ext_re = re.compile(r"\.([^\./]+)$")
+file_re = re.compile(r"([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)")
+url_re = re.compile(r"[a-z]{3,8}://[^\s'\"]+\S")
+ext_re = re.compile(r"\.([^\./]+)$")
# CGI stuff ---------------------------------------------------------
+def config_get(key, default=None):
+ return globals().get(key, default)
+
def script_name():
return os.environ.get('SCRIPT_NAME', '')
+#TODO: move post-edit hook into wiki, then kill this
+def script_path():
+ return os.path.split(os.environ.get('SCRIPT_FILENAME', ''))[0]
+
def query_string():
path_info = os.environ.get('PATH_INFO', '')
if len(path_info) and path_info[0] == '/':
@@ -44,8 +47,9 @@ def query_string():
else:
return os.environ.get('QUERY_STRING', '') or 'FrontPage'
-def privileged_path():
- return privileged_url or script_name()
+def is_privileged():
+ purl = config_get('privileged_url')
+ return (purl is not None) and os.environ.get('SCRIPT_URI', '').startswith(purl)
def remote_user():
user = os.environ.get('REMOTE_USER', '')
@@ -69,7 +73,7 @@ def is_external_url(pathname):
def relative_url(pathname, privileged=False):
if not is_external_url(pathname):
if privileged:
- url = privileged_path()
+ url = config_get('privileged_url') or script_name()
else:
url = script_name()
pathname = url + '/' + pathname
@@ -79,20 +83,20 @@ def permalink(s):
return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip())
def humanlink(s):
- return re.sub(r'([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' '))
+ return re.sub(r'(?:.*[/:]|)([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' '))
-# Split arg lists like "blah| blah blah| width=100 | align = center",
+# Split arg lists like "blah|blah blah| width=100 | align = center",
# return a list containing anonymous arguments and a map containing the named arguments
def parse_args(s):
args = []
- kwargs = {}
+ kvargs = {}
for arg in s.strip('<[{}]>').split('|'):
m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg)
if m is not None:
- kwargs[m.group(1)] = m.group(2)
+ kvargs[m.group(1)] = m.group(2)
else:
args.append(arg.strip())
- return (args, kwargs)
+ return (args, kvargs)
def url_args(kvargs):
argv = []
@@ -103,7 +107,9 @@ def url_args(kvargs):
return ''
# Formatting stuff --------------------------------------------------
-def emit_header(mime_type="text/html"):
+def emit_header(mtime=None, mime_type="text/html"):
+ if mtime:
+ print("Last-Modified: " + strftime("%a, %d %b %Y %H:%M:%S GMT", gmtime(mtime)))
print("Content-type: " + mime_type + "; charset=utf-8\n")
def send_guru(msg_text, msg_type):
@@ -111,48 +117,49 @@ def send_guru(msg_text, msg_type):
print('
')
- print link_tag('FrontPage', site_icon or 'Home', cssclass='navlink')
+ print link_tag('FrontPage', config_get('site_icon', 'Home'), cssclass='navlink')
if name:
print('
' + link_tag('?fullsearch=' + name, text, cssclass='navlink') + ' ')
else:
print('
' + text + ' ')
print(' | ' + link_tag('FindPage', 'Find Page', cssclass='navlink'))
- if 'history_url' in globals():
- print(' |
Recent Changes')
+ if history:
+ print(' |
Recent Changes')
if name:
- print(' |
Page History')
+ print(' |
Page History')
if name:
print(' | ' + link_tag(name + '?a=raw', 'Raw Text', cssclass='navlink'))
- if privileged_url is not None:
+ if config_get('privileged_url') is not None:
if writable:
print(' | ' + link_tag('?a=edit&q=' + name, 'Edit', cssclass='navlink', privileged=True))
else:
@@ -197,7 +204,7 @@ def send_httperror(status="403 Not Found", query=""):
def link_tag(dest, text=None, privileged=False, **kvargs):
if text is None:
text = humanlink(dest)
- elif img_re.match(text):
+ elif image_re.match(text):
text = '
'
link_class = kvargs.get('class', kvargs.get('cssclass', None))
@@ -207,7 +214,7 @@ def link_tag(dest, text=None, privileged=False, **kvargs):
elif file_re.match(dest) and Page(dest).exists():
link_class = 'wikilink'
else:
- text = nonexist_pfx + text
+ text = config_get('nonexist_pfx', '') + text
link_class = 'nonexistent'
# Prevent crawlers from following links potentially added by spammers or to generated pages
@@ -221,15 +228,25 @@ def link_inline(name, descr=None, kvargs={}):
if not descr: descr = humanlink(name)
url = relative_url(name)
if video_re.match(name):
- return '
' % url
- elif img_re.match(name):
+ return '
' % url
+ elif image_re.match(name):
return '
' % (url, url + url_args(kvargs), descr)
elif file_re.match(name) and not ext_re.search(name): # FIXME: this guesses a wiki page
- return Page(name).send_naked()
+ Page(name).send_naked(kvargs) # FIXME: we should return the page as a string rather than print it
+ return ''
else:
return '
' \
% (url, url, name)
+def link_inline_glob(pattern, descr=None, kvargs={}):
+ if not url_re.match(pattern) and bool(set(pattern) & set('?*[')):
+ s = ''
+ for name in glob.glob(pattern):
+ s += link_inline(name, descr, kvargs)
+ return s
+ else:
+ return link_inline(pattern, descr, kvargs)
+
# Search ---------------------------------------------------
def print_search_stats(hits, searched):
@@ -319,32 +336,32 @@ def handle_get(pagename, form):
else:
send_httperror("403 Forbidden", pagename)
-# Used by macros/WordIndex and macros/TitleIndex
+# Used by sys/macros/WordIndex and sys/macros/TitleIndex
def make_index_key():
links = ['
%s' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz']
return '
' + ' | '.join(links) + '
'
-def page_list(dirname = None, re = None):
- if re is None:
+def page_list(dirname=None, search_re=None):
+ if search_re is None:
# FIXME: WikiWord is too restrictive now!
- re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
- return sorted(filter(re.match, os.listdir(dirname or data_dir)))
+ search_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$")
+ return sorted(filter(search_re.match, os.listdir(dirname or '.')))
-def send_footer(mod_string=None):
- if globals().get('debug_cgi', False):
+def send_footer(mtime=None):
+ if config_get('debug_cgi', False):
cgi.print_arguments()
cgi.print_form(form)
cgi.print_environ()
- print('''
-')
+ link_inline("sys/footer", kvargs = {
+ 'LAST_MODIFIED': strftime(config_get('datetime_fmt', '%a %d %b %Y %I:%M %p'), localtime(mtime))
+ })
+ print("")
+
+def _macro_ELAPSED_TIME(*args, **kvargs):
+ return "%03f" % (clock() - start_time)
+
+def _macro_VERSION(*args, **kvargs):
+ return __version__
class WikiFormatter:
"""Object that turns Wiki markup into HTML.
@@ -352,12 +369,13 @@ class WikiFormatter:
All formatting commands can be parsed one line at a time, though
some state is carried over between lines.
"""
- def __init__(self, raw):
+ def __init__(self, raw, kvargs=None):
self.raw = raw
+ self.kvargs = kvargs or {}
self.h_level = 0
self.in_pre = self.in_html = self.in_table = self.in_li = False
self.in_header = True
- self.list_indents = []
+ self.list_indents = [] # a list of pairs (indent_level, list_type) to track nested lists
self.tr_cnt = 0
self.styles = {
#wiki html enabled?
@@ -365,6 +383,7 @@ class WikiFormatter:
"**": ["b", False],
"##": ["tt", False],
"__": ["u", False],
+ "--": ["del", False],
"^^": ["sup", False],
",,": ["sub", False],
"''": ["em", False], # LEGACY
@@ -377,6 +396,9 @@ class WikiFormatter:
style[1] = not style[1]
return ['', '<'][style[1]] + style[0] + '>'
+ def _glyph_repl(self, word):
+ return '—'
+
def _tit_repl(self, word):
if self.h_level:
result = '
\n' % self.h_level
@@ -391,18 +413,20 @@ class WikiFormatter:
return '
'
def _rule_repl(self, word):
- return self._undent() + '\n
\n' % (len(word) - 2)
+ return '\n
\n' % (len(word) - 2)
def _macro_repl(self, word):
try:
- args, kwargs = parse_args(word)
+ args, kvargs = parse_args(word)
+ if args[0] in self.kvargs:
+ return self.kvargs[args[0]]
macro = globals().get('_macro_' + args[0])
if not macro:
- exec(open("macros/" + name + ".py").read(), globals())
- macro = globals().get('_macro_' + name)
- return macro(*args, **kwargs)
- except Exception:
- msg = cgi.escape(word)
+ exec(open("sys/macros/" + args[0] + ".py").read(), globals())
+ macro = globals().get('_macro_' + args[0])
+ return macro(*args, **kvargs)
+ except Exception, e:
+ msg = cgi.escape(word) + ": " + cgi.escape(str(e))
if not self.in_html:
msg = '
' + msg + ''
return msg
@@ -419,10 +443,10 @@ class WikiFormatter:
# This double div nonsense works around a limitation of the HTML block model
return '
' \
+ '
' \
- + link_inline(name, descr, kvargs) \
+ + link_inline_glob(name, descr, kvargs) \
+ '
' + descr + '
'
else:
- return link_inline(name, None, kvargs)
+ return link_inline_glob(name, None, kvargs)
def _html_repl(self, word):
if not self.in_html and word.startswith('
new_level:
- del(self.list_indents[-1])
if self.in_li:
s += ''
self.in_li = False # FIXME
- s += '\n'
+ s += '' + self.list_indents[-1][1] + '>\n'
+ del(self.list_indents[-1])
+
+ list_type = ('ul', 'ol')[list_type == '#']
while self._indent_level() < new_level:
- self.list_indents.append(new_level)
- s += '
' * len(self.list_indents)
- res += '
'
- self.list_indents = []
- return res
-
def replace(self, match):
for rule, hit in list(match.groupdict().items()):
if hit:
return getattr(self, '_' + rule + '_repl')(hit)
else:
- raise "Can't handle match " + repr(match)
+ raise Exception("Can't handle match " + repr(match))
def print_html(self):
print('
')
scan_re = re.compile(r"""(?:
- # Styles and formatting
- (?P \*\*|'''|//|''|\#\#|``|__|\^\^|,,)
+ # Styles and formatting ("--" must cling to a word to disambiguate it from the dash)
+ (?P \*\* | // | \#\# | __ | --\b | \b-- | \^\^ | ,, | ''' | '' | `` )
| (?P \={2,6})
| (?P
\\\\)
| (?P ^-{3,})
| (?P \b( FIXME | TODO | DONE )\b )
+ | (?P --)
# Links
| (?P \<\<([^\s\|\>]+)(?:\s*\|\s*([^\>]+)|)\>\>)
| (?P \[\[([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\]\])
# Inline HTML
- | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
+ | (?P <(br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])\b )
| (?P ( /\s*> | (br|hr|div|span|form|iframe|input|textarea|a|img|h[1-5])> ) )
| (?P [<>&] )
# Auto links (LEGACY)
- | (?P \b[a-zA-Z0-9_/-]+\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt))
+ | (?P \b[a-zA-Z0-9_/-]+\.(""" + image_ext + "|" + video_ext + r"""))
| (?P \b(?:[A-Z][a-z]+){2,}\b)
| (?P (http|https|ftp|mailto)\:[^\s'\"]+\S)
| (?P [-\w._+]+\@[\w.-]+)
- # Lists, divs, spans
+ # Lists, divs, spans and inline objects
| (?P ^\s+[\*\#]\s+)
| (?P \{\{\{|\s*\}\}\})
| (?P \{\{([^\s\|]+)(?:\s*\|\s*([^\]]+)|)\}\})
@@ -571,7 +591,7 @@ class WikiFormatter:
| (?P[<>&])"
)""", re.VERBOSE)
blank_re = re.compile(r"^\s*$")
- indent_re = re.compile(r"^\s*")
+ indent_re = re.compile(r"^(\s*)(\*|\#|)")
tr_re = re.compile(r"^\s*\|\|")
eol_re = re.compile(r"\r?\n")
# For each line, we scan through looking for magic strings, outputting verbatim any intervening text
@@ -594,13 +614,13 @@ class WikiFormatter:
print('
')
else:
indent = indent_re.match(self.line)
- #3.0: print(self._indent_to(len(indent.group(0))), end=' ')
- print(self._indent_to(len(indent.group(0))))
+ print(self._indent_to(len(indent.group(1)), indent.group(2)))
+ # Stand back! Here we apply the monster regex that does all the parsing
print(re.sub(scan_re, self.replace, self.line))
if self.in_pre: print('')
if self.in_table: print('
')
- print(self._undent())
+ print(self._indent_to(0))
print('
')
class Page:
@@ -614,27 +634,31 @@ class Page:
return re.sub('([a-z])([A-Z])', r'\1 \2', self.page_name)
def _filename(self):
- return os.path.join(data_dir, self.page_name)
+ return self.page_name
def _tmp_filename(self):
- return os.path.join(data_dir, ('#' + self.page_name.replace('/','_') + '.' + str(os.getpid()) + '#'))
+ return self.page_name + '.tmp' + str(os.getpid()) + '#'
- def exists(self):
+ def _mtime(self):
try:
- os.stat(self._filename())
- return True
- except OSError as err:
+ return os.stat(self._filename()).st_mtime
+ except OSError, err:
if err.errno == errno.ENOENT:
- return False
+ return None
raise err
+ def exists(self):
+ if self._mtime():
+ return True
+ return False
+
def get_raw_body(self, default=None):
try:
return open(self._filename(), 'rb').read()
- except IOError as err:
+ except IOError, err:
if err.errno == errno.ENOENT:
if default is None:
- default = '//[[%s|Describe %s|action=edit]]//' % (self.page_name, self.page_name)
+ default = '//[[?a=edit&q=%s|Describe %s]]//' % (self.page_name, self.page_name)
return default
if err.errno == errno.EISDIR:
return self.format_dir()
@@ -643,16 +667,17 @@ class Page:
def format_dir(self):
out = '== '
pathname = ''
- for dirname in self.page_name.split('/'):
- pathname = (pathname + '/' + dirname) if pathname else dirname
+ for dirname in self.page_name.strip('/').split('/'):
+ pathname = (pathname and pathname + '/' ) + dirname
out += '[[' + pathname + '|' + dirname + ']]/'
out += ' ==\n'
for filename in page_list(self._filename(), file_re):
- if img_re.match(filename):
- if image_maxwidth:
- maxwidth_arg = ' | maxwidth=' + str(image_maxwidth)
- out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth_arg + ' | class=thumbleft}}\n'
+ if image_re.match(filename):
+ maxwidth = config_get('image_maxwidth', '400')
+ if maxwidth:
+ maxwidth = ' | maxwidth=' + str(maxwidth)
+ out += '{{' + self.page_name + '/' + filename + ' | ' + humanlink(filename) + maxwidth + ' | class=thumbleft}}\n'
else:
out += ' * [[' + self.page_name + '/' + filename + ']]\n'
return out
@@ -669,9 +694,9 @@ class Page:
break
self._pragmas[m.group(1)] = m.group(2).strip()
#print "bernie: pragmas[" + m.group(1) + "] = " + m.group(2) + "
\n"
- except IOError as err:
+ except IOError, err:
if err.errno != errno.ENOENT and err.errno != errno.EISDIR:
- raise er
+ raise err
return self._pragmas
def pragma(self, name, default):
@@ -698,9 +723,9 @@ class Page:
def can_read(self):
return self.can("read", True)
- def send_naked(self):
+ def send_naked(self, kvargs=None):
if self.can_read():
- WikiFormatter(self.get_raw_body()).print_html()
+ WikiFormatter(self.get_raw_body(), kvargs).print_html()
else:
send_guru("Read access denied by ACLs", "notice")
@@ -712,19 +737,9 @@ class Page:
link_urls += [ [ "stylesheet", value ] ]
send_title(self.page_name, self.split_title(),
- msg_text=self.msg_text, msg_type=self.msg_type, writable=self.can_write())
+ msg_text=self.msg_text, msg_type=self.msg_type, writable=self.can_write(), mtime=self._mtime())
self.send_naked()
- send_footer(self._last_modified())
-
- def _last_modified(self):
- try:
- from time import localtime, strftime
- modtime = localtime(os.stat(self._filename())[stat.ST_MTIME])
- except OSError as err:
- if err.errno != errno.ENOENT:
- raise err
- return None
- return strftime(datetime_fmt, modtime)
+ send_footer(mtime=self._mtime())
def send_editor(self, preview=None):
send_title(None, 'Edit ' + self.split_title(), msg_text=self.msg_text, msg_type=self.msg_type)
@@ -732,35 +747,14 @@ class Page:
send_guru("Write access denied by ACLs", "error")
return
- filename = ''
- if 'file' in form:
- filename = form['file'].value
-
- print(('
Editing ' + self.page_name
- + ' for ' + cgi.escape(remote_user())
- + ' from ' + cgi.escape(get_hostname(remote_host()))
- + '
'))
- print('
-
- """)
- print("
" + link_tag('EditingTips') + "
")
+ if preview is None:
+ preview = self.get_raw_body(default='')
+
+ link_inline("sys/EditPage", kvargs = {
+ 'EDIT_BODY': cgi.escape(preview),
+ #'EDIT_PREVIEW': WikiFormatter(preview).print_html(),
+ })
+
if preview:
print("
")
WikiFormatter(preview).print_html()
@@ -769,18 +763,17 @@ class Page:
def send_raw(self, mimetype='text/plain', args=[]):
if not self.can_read():
- send_title(None, msg_text='Read access denied by ACLs', msg_type='notice')
+ send_title(None, msg_text='Read access denied by ACLs', msg_type='notice', mtime=self._mtime())
return
+ emit_header(self._mtime(), mimetype)
if 'maxwidth' in args:
import subprocess
- emit_header(mimetype)
sys.stdout.flush()
subprocess.check_call(['gm', 'convert', self._filename(),
'-scale', args['maxwidth'].value + ' >', '-'])
else:
body = self.get_raw_body()
- emit_header(mimetype)
print(body)
def _write_file(self, data):
@@ -791,21 +784,29 @@ class Page:
# Bad Bill! POSIX rename ought to replace. :-(
try:
os.remove(name)
- except OSError as err:
+ except OSError, err:
if err.errno != errno.ENOENT: raise err
+ path = os.path.split(name)[0]
+ if path and not os.path.exists(path):
+ os.makedirs(path)
os.rename(tmp_filename, name)
def save(self, newdata, changelog):
if not self.can_write():
- self.msg_text = 'Write access denied by ACLs'
- self.msg_type = 'error'
+ self.msg_text = 'Write access denied by Access Control List'
+ return
+ if not is_privileged():
+ self.msg_text = 'Unauthenticated access denied'
return
self._write_file(newdata)
rc = 0
- if post_edit_hook:
+ if config_get('post_edit_hook'):
import subprocess
- cmd = [ post_edit_hook, data_dir + '/' + self.page_name, remote_user(), remote_host(), changelog]
+ cmd = [
+ config_get('post_edit_hook'),
+ self.page_name, remote_user(),
+ remote_host(), changelog ]
child = subprocess.Popen(cmd, stdout=subprocess.PIPE, close_fds=True)
output = child.stdout.read()
rc = child.wait()
@@ -819,6 +820,7 @@ class Page:
try:
exec(open("geekigeeki.conf.py").read())
+ os.chdir(config_get('data_dir', 'data'))
form = cgi.FieldStorage()
action = form.getvalue('a', 'get')
handler = globals().get('handle_' + action)