X-Git-Url: https://codewiz.org/gitweb?a=blobdiff_plain;f=geekigeeki.py;h=678754e3a9633d29fc779eeed48eeddb8fa9fa87;hb=8d7f457924b6c4bde8486d1fd413516082ceedf9;hp=eba446715e1e6f08bcdb19d01fbcfe7f3b3c2025;hpb=06653e2489969feaf8ec69fc449187eaf30ea525;p=geekigeeki.git diff --git a/geekigeeki.py b/geekigeeki.py index eba4467..678754e 100755 --- a/geekigeeki.py +++ b/geekigeeki.py @@ -22,20 +22,16 @@ __version__ = '$Id$'[4:12] from time import clock start_time = clock() +title_done = False import cgi, sys, os, re, errno, stat -# Regular expression defining a WikiWord -# (but this definition is also assumed in other places) -word_re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") # FIXME: we accept stuff like foo/../bar and we shouldn't file_re = re.compile(r"^\b([A-Za-z0-9_\-][A-Za-z0-9_\.\-/]*)\b$") img_re = re.compile(r"^.*\.(png|gif|jpg|jpeg|bmp|ico|ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE) video_re = re.compile(r"^.*\.(ogm|ogg|mkv|mpg|mpeg|mp4|avi|asf|flv|wmv|qt)$", re.IGNORECASE) url_re = re.compile(r"^[a-z]{3,8}://[^\s'\"]+\S$") -link_re = re.compile("(?:\[\[|{{)([^\s\|]+)(?:\s*\|\s*([^\]]+)|)(?:\]\]|}})") - -title_done = False +ext_re = re.compile(r"\.([^\./]+)$") # CGI stuff --------------------------------------------------------- def script_name(): @@ -77,35 +73,49 @@ def relative_url(pathname, privileged=False): else: url = script_name() pathname = url + '/' + pathname - return pathname + return cgi.escape(pathname, quote=True) def permalink(s): return re.sub(' ', '-', re.sub('[^a-z0-9_ ]', '', s.lower()).strip()) +def humanlink(s): + return re.sub(r'([^:/\.]+)(?:\.[^/:]+|)$', r'\1', s.replace('_', ' ')) + +# Split arg lists like "blah| blah blah| width=100 | align = center", +# return a list containing anonymous arguments and a map containing the named arguments +def parse_args(s): + args = [] + kwargs = {} + for arg in s.strip('<[{}]>').split('|'): + m = re.match('\s*(\w+)\s*=\s*(.+)\s*', arg) + if m is not None: + kwargs[m.group(1)] = m.group(2) + else: + args.append(arg.strip()) + return (args, kwargs) + +def url_args(kvargs): + argv = [] + for k, v in kvargs.items(): + argv.append(k + '=' + v) + if argv: + return '?' + '&'.join(argv) + return '' + # Formatting stuff -------------------------------------------------- def emit_header(mime_type="text/html"): - print "Content-type: " + mime_type + "; charset=utf-8\n" - -def sendfile(dest_file, src_file): - """Efficiently copy file data between file descriptors""" - while 1: - data = src_file.read(65536) - if not data: break - dest_file.write(data) + print("Content-type: " + mime_type + "; charset=utf-8\n") def send_guru(msg_text, msg_type): if not msg_text: return - print '
' + print('' \ + % relative_url('sys/GuruMeditation.js')) def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=False): global title_done @@ -113,110 +123,121 @@ def send_title(name, text="Limbo", msg_text=None, msg_type='error', writable=Fal # Head emit_header() - print '' - print '' + print('') + print('') - print "') if msg_type == 'error': - print ' Software Failure. Press left mouse button to continue.\n' - print msg_text + print(' Software Failure. Press left mouse button to continue.\n') + print(msg_text) if msg_type == 'error': - print '\n Guru Meditation #DEADBEEF.ABADC0DE' - print '' - try: - sendfile(sys.stdout, open('gurumeditation.js', 'rb')) - except IOError, err: - pass + print '\n Guru Meditation #DEADBEEF.ABADC0DE' + print('
%d hits out of %d pages searched.
" % (hits, searched)) + +def handle_fullsearch(query, form): + needle = form['q'].value + send_title(None, 'Full text search for "' + needle + '"') needle_re = re.compile(needle, re.IGNORECASE) hits = [] @@ -232,47 +253,44 @@ def handle_fullsearch(needle): hits.sort() hits.reverse() - print "' + link_tag(page_name) - print ' . . . . ' + `count` - print ['match', 'matches'][count != 1] - print '
' + link_tag(page_name)) + print(' . . . . ' + `count`) + print(['match', 'matches'][count != 1]) + print('
' + link_tag(filename) + "
' + link_tag(filename) + "
%d hits out of %d pages searched.
" % (hits, searched) - -def handle_raw(pagename): +def handle_raw(pagename, form): if not file_re.match(pagename): send_httperror("403 Forbidden", pagename) return Page(pagename).send_raw() -def handle_edit(pagename): +def handle_edit(pagename, form): if not file_re.match(pagename): send_httperror("403 Forbidden", pagename) return - pg = Page(pagename) + pg = Page(form['q'].value) if 'save' in form: if form['file'].value: pg.save(form['file'].file.read(), form['changelog'].value) @@ -289,11 +307,27 @@ def handle_edit(pagename): text = form['savetext'].value pg.send_editor(text) +def handle_get(pagename, form): + if file_re.match(pagename): + # FIMXE: this is all bullshit, MimeTypes bases its guess on the extension! + from mimetypes import MimeTypes + mimetype, encoding = MimeTypes().guess_type(pagename) + if mimetype: + Page(pagename).send_raw(mimetype=mimetype, args=form) + else: + Page(pagename).format() + else: + send_httperror("403 Forbidden", pagename) + +# Used by macros/WordIndex and macros/TitleIndex def make_index_key(): - links = map(lambda ch: '%s' % (ch, ch), 'abcdefghijklmnopqrstuvwxyz') - return '' + ' | '.join(links) + '
' -def page_list(dirname = None, re = word_re): +def page_list(dirname = None, re = None): + if re is None: + # FIXME: WikiWord is too restrictive now! + re = re.compile(r"^\b((([A-Z][a-z0-9]+){2,}/)*([A-Z][a-z0-9]+){2,})\b$") return sorted(filter(re.match, os.listdir(dirname or data_dir))) def send_footer(mod_string=None): @@ -301,16 +335,16 @@ def send_footer(mod_string=None): cgi.print_arguments() cgi.print_form(form) cgi.print_environ() - print ''' + print(''' ' + print('last modified %s
' % mod_string) + print('') class WikiFormatter: """Object that turns Wiki markup into HTML. @@ -360,57 +394,44 @@ class WikiFormatter: return self._undent() + '\n' + print('
') scan_re = re.compile(r"""(?: # Styles and formatting @@ -553,9 +574,9 @@ class WikiFormatter: indent_re = re.compile(r"^\s*") tr_re = re.compile(r"^\s*\|\|") eol_re = re.compile(r"\r?\n") - # For each line, we scan through looking for magic strings, outputting verbatim any intervening text - for self.line in eol_re.split(self.raw.expandtabs()): + #3.0: for self.line in eol_re.split(str(self.raw.expandtabs(), 'utf-8')): + for self.line in eol_re.split(str(self.raw.expandtabs())): # Skip pragmas if self.in_header: if self.line.startswith('#'): @@ -563,23 +584,24 @@ class WikiFormatter: self.in_header = False if self.in_pre: - print re.sub(pre_re, self.replace, self.line) + print(re.sub(pre_re, self.replace, self.line)) else: if self.in_table and not tr_re.match(self.line): self.in_table = False - print '
' + print('
') if blank_re.match(self.line): - print '
' + print('
') else: indent = indent_re.match(self.line) - print self._indent_to(len(indent.group(0))) , - print re.sub(scan_re, self.replace, self.line) + #3.0: print(self._indent_to(len(indent.group(0))), end=' ') + print(self._indent_to(len(indent.group(0)))) + print(re.sub(scan_re, self.replace, self.line)) - if self.in_pre: print '' - if self.in_table: print '
' - print self._undent() - print '
') + print(self._undent()) + print('
Editing ' + self.page_name + print(('
Editing ' + self.page_name + ' for ' + cgi.escape(remote_user()) + ' from ' + cgi.escape(get_hostname(remote_host())) - + '
') - print '