X-Git-Url: https://codewiz.org/gitweb?p=geekigeeki.git;a=blobdiff_plain;f=geekigeeki.py;h=4d50a3af5c0fd4aa29c24ec1918e59f801a2d047;hp=faf59027217f4c59f0300d99c22f4a22f487e4b5;hb=41e94725555d64df1b6e56ffd29a733c28e0bfdf;hpb=d810022b52bf2aa2c7a4b4c530dd7a7f03dc9299 diff --git a/geekigeeki.py b/geekigeeki.py index faf5902..4d50a3a 100755 --- a/geekigeeki.py +++ b/geekigeeki.py @@ -22,22 +22,17 @@ __version__ = '$Id$'[4:12] from time import clock start_time = clock() +title_done = False import cgi, sys, os, re, errno, stat -# Regular expression defining a WikiWord -# (but this definition is also assumed in other places) -word_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") # FIXME: we accept stuff like foo/../bar and we shouldn't file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$") img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE) video_re = re.compile(r"^.*\.(ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE) url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$") -link_re = re.compile(r"(?:\[\[|{{)([^\s\|]+)(?:\s*\|\s*([^\]]+)|)(?:\]\]|}})") ext_re = re.compile(r"\.([^\./]+)$") -title_done = False - # CGI stuff --------------------------------------------------------- def script_name(): return os.environ.get('SCRIPT_NAME', '') @@ -83,21 +78,44 @@ def relative_url(pathname, privileged=False): def permalink(s): return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip()) +def humanlink(s): + return re.search('([^:/\.]+)(?:\.[^/:]+|)$', s).group(1).replace('_', ' ') + +# Split arg lists like "blah| blah blah| width=100 | align = center", +# return a list containing anonymous arguments and a map containing the named arguments +def parse_args(s): + args = [] + kwargs = {} + for arg in s.strip('<[{}]>').split('|'): + try: + key, val = arg.split('=', 1) + kwargs[key.strip()] = val.strip() + except ValueError: + args.append(arg.strip()) + return (args, kwargs) + +def url_args(kvargs): + argv = [] + for k, v in kvargs.items(): + argv.append(k + '=' + v) + if argv: + return '?' + '&'.join(argv) + return '' + # Formatting stuff -------------------------------------------------- def emit_header(mime_type="text/html"): - print "Content-type: " + mime_type + "; charset=utf-8\n" + print("Content-type: " + mime_type + "; charset=utf-8\n") def send_guru(msg_text, msg_type): if not msg_text: return - print '
' + print('' \ + % relative_url('sys/GuruMeditation.js')) def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=False): global title_done @@ -105,80 +123,80 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal # Head emit_header() - print '' - print '' + print('') + print('') - print "') if msg_type == 'error': - print ' Software Failure. Press left mouse button to continue.\n' - print msg_text + print(' Software Failure. Press left mouse button to continue.\n') + print(msg_text) if msg_type == 'error': print '\n Guru Meditation #DEADBEEF.ABADC0DE' - print '' - print '' \ - % relative_url('sys/GuruMeditation.js') + print('
%d hits out of %d pages searched.
" % (hits, searched)) + +def handle_fullsearch(query, form): + needle = form['q'].value + send_title(None, 'Full text search for "' + needle + '"') needle_re = re.compile(needle, re.IGNORECASE) hits = [] @@ -230,47 +252,44 @@ def handle_fullsearch(needle): hits.sort() hits.reverse() - print "' + link_tag(page_name) - print ' . . . . ' + `count` - print ['match', 'matches'][count != 1] - print '
' + link_tag(page_name)) + print(' . . . . ' + `count`) + print(['match', 'matches'][count != 1]) + print('
' + link_tag(filename) + "
' + link_tag(filename) + "
%d hits out of %d pages searched.
" % (hits, searched) - -def handle_raw(pagename): +def handle_raw(pagename, form): if not file_re.match(pagename): send_httperror("403 Forbidden", pagename) return Page(pagename).send_raw() -def handle_edit(pagename): +def handle_edit(pagename, form): if not file_re.match(pagename): send_httperror("403 Forbidden", pagename) return - pg = Page(pagename) + pg = Page(form['q'].value) if 'save' in form: if form['file'].value: pg.save(form['file'].file.read(), form['changelog'].value) @@ -287,12 +306,27 @@ def handle_edit(pagename): text = form['savetext'].value pg.send_editor(text) +def handle_get(pagename, form): + if file_re.match(pagename): + # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension! + from mimetypes import MimeTypes + mimetype, encoding = MimeTypes().guess_type(pagename) + if mimetype: + Page(pagename).send_raw(mimetype=mimetype, args=form) + else: + Page(pagename).format() + else: + send_httperror("403 Forbidden", pagename) + # Used by macros/WordIndex and macros/TitleIndex def make_index_key(): - links = map(lambda ch: '%s' % (ch, ch), 'abcdefghijklmnopqrstuvwxyz') - return ''+ ' | '.join(links) + '
' + links = ['%s' % (ch, ch) for ch in 'abcdefghijklmnopqrstuvwxyz'] + return '' + ' | '.join(links) + '
' -def page_list(dirname = None, re = word_re): +def page_list(dirname = None, re = None): + if re is None: + # FIXME: WikiWord is too restrictive now! + re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") return sorted(filter(re.match, os.listdir(dirname or data_dir))) def send_footer(mod_string=None): @@ -300,16 +334,16 @@ def send_footer(mod_string=None): cgi.print_arguments() cgi.print_form(form) cgi.print_environ() - print ''' + print(''' ' + print('last modified %s
' % mod_string) + print('') class WikiFormatter: """Object that turns Wiki markup into HTML. @@ -359,48 +393,35 @@ class WikiFormatter: return self._undent() + '\n' + print('
') scan_re = re.compile(r"""(?: # Styles and formatting @@ -552,9 +573,9 @@ class WikiFormatter: indent_re = re.compile(r"^\s*") tr_re = re.compile(r"^\s*\|\|") eol_re = re.compile(r"\r?\n") - # For each line, we scan through looking for magic strings, outputting verbatim any intervening text - for self.line in eol_re.split(self.raw.expandtabs()): + #3.0: for self.line in eol_re.split(str(self.raw.expandtabs(), 'utf-8')): + for self.line in eol_re.split(str(self.raw.expandtabs())): # Skip pragmas if self.in_header: if self.line.startswith('#'): @@ -562,23 +583,24 @@ class WikiFormatter: self.in_header = False if self.in_pre: - print re.sub(pre_re, self.replace, self.line) + print(re.sub(pre_re, self.replace, self.line)) else: if self.in_table and not tr_re.match(self.line): self.in_table = False - print '
' + print('
') if blank_re.match(self.line): - print '
' + print('
') else: indent = indent_re.match(self.line) - print self._indent_to(len(indent.group(0))) , - print re.sub(scan_re, self.replace, self.line) + #3.0: print(self._indent_to(len(indent.group(0))), end=' ') + print(self._indent_to(len(indent.group(0)))) + print(re.sub(scan_re, self.replace, self.line)) - if self.in_pre: print '' - if self.in_table: print '
' - print self._undent() - print '
') + print(self._undent()) + print('
Editing ' + self.page_name + print(('
Editing ' + self.page_name + ' for ' + cgi.escape(remote_user()) + ' from ' + cgi.escape(get_hostname(remote_host())) - + '
') - print '