165 lines
4.9 KiB
Python
Executable File
165 lines
4.9 KiB
Python
Executable File
#!/usr/bin/env python
|
|
"""
|
|
wikipedia.py - Phenny Wikipedia Module
|
|
Copyright 2008, Sean B. Palmer, inamidst.com
|
|
Licensed under the Eiffel Forum License 2.
|
|
|
|
http://inamidst.com/phenny/
|
|
"""
|
|
|
|
import re, urllib
|
|
import web
|
|
|
|
wikiuri = 'http://en.wikipedia.org/wiki/%s'
|
|
wikisearch = 'http://en.wikipedia.org/wiki/Special:Search?' \
|
|
+ 'search=%s&fulltext=Search'
|
|
|
|
r_tr = re.compile(r'(?ims)<tr[^>]*>.*?</tr>')
|
|
r_paragraph = re.compile(r'(?ims)<p[^>]*>.*?</p>|<li(?!n)[^>]*>.*?</li>')
|
|
r_tag = re.compile(r'<(?!!)[^>]+>')
|
|
r_whitespace = re.compile(r'[\t\r\n ]+')
|
|
r_redirect = re.compile(
|
|
r'(?ims)class=.redirectText.>\s*<a\s*href=./wiki/([^"/]+)'
|
|
)
|
|
|
|
abbrs = ['etc', 'ca', 'cf', 'Co', 'Ltd', 'Inc', 'Mt', 'Mr', 'Mrs',
|
|
'Dr', 'Ms', 'Rev', 'Fr', 'St', 'Sgt', 'pron', 'approx', 'lit',
|
|
'syn'] \
|
|
+ list('ABCDEFGHIJKLMNOPQRSTUVWXYZ') \
|
|
+ list('abcdefghijklmnopqrstuvwxyz')
|
|
t_sentence = r'^.{5,}?(?<!\b%s)(?:\.(?= [A-Z0-9]|\Z)|\Z)'
|
|
r_sentence = re.compile(t_sentence % r')(?<!\b'.join(abbrs))
|
|
|
|
def unescape(s):
|
|
s = s.replace('>', '>')
|
|
s = s.replace('<', '<')
|
|
s = s.replace('&', '&')
|
|
s = s.replace(' ', ' ')
|
|
return s
|
|
|
|
def text(html):
|
|
html = r_tag.sub('', html)
|
|
html = r_whitespace.sub(' ', html)
|
|
return unescape(html).strip()
|
|
|
|
def search(term):
|
|
try: import search
|
|
except ImportError, e:
|
|
print e
|
|
return term
|
|
|
|
if isinstance(term, unicode):
|
|
term = term.encode('utf-8')
|
|
else: term = term.decode('utf-8')
|
|
|
|
term = term.replace('_', ' ')
|
|
try: uri = search.result('site:en.wikipedia.org %s' % term)
|
|
except IndexError: return term
|
|
if uri:
|
|
return uri[len('http://en.wikipedia.org/wiki/'):]
|
|
else: return term
|
|
|
|
def wikipedia(term, last=False):
|
|
global wikiuri
|
|
if not '%' in term:
|
|
if isinstance(term, unicode):
|
|
t = term.encode('utf-8')
|
|
else: t = term
|
|
q = urllib.quote(t)
|
|
u = wikiuri % q
|
|
bytes = web.get(u)
|
|
else: bytes = web.get(wikiuri % term)
|
|
bytes = r_tr.sub('', bytes)
|
|
|
|
if not last:
|
|
r = r_redirect.search(bytes[:4096])
|
|
if r:
|
|
term = urllib.unquote(r.group(1))
|
|
return wikipedia(term, last=True)
|
|
|
|
paragraphs = r_paragraph.findall(bytes)
|
|
|
|
if not paragraphs:
|
|
if not last:
|
|
term = search(term)
|
|
return wikipedia(term, last=True)
|
|
return None
|
|
|
|
# Pre-process
|
|
paragraphs = [para for para in paragraphs
|
|
if (para and 'technical limitations' not in para
|
|
and 'window.showTocToggle' not in para
|
|
and 'Deletion_policy' not in para
|
|
and 'Template:AfD_footer' not in para
|
|
and not (para.startswith('<p><i>') and
|
|
para.endswith('</i></p>'))
|
|
and not 'disambiguation)"' in para)
|
|
and not '(images and media)' in para
|
|
and not 'This article contains a' in para
|
|
and not 'id="coordinates"' in para
|
|
and not 'class="thumb' in para]
|
|
|
|
for i, para in enumerate(paragraphs):
|
|
para = para.replace('<sup>', '|')
|
|
para = para.replace('</sup>', '|')
|
|
paragraphs[i] = text(para).strip()
|
|
|
|
# Post-process
|
|
paragraphs = [para for para in paragraphs if
|
|
(para and not (para.endswith(':') and len(para) < 150))]
|
|
|
|
para = text(paragraphs[0])
|
|
m = r_sentence.match(para)
|
|
|
|
if not m:
|
|
if not last:
|
|
term = search(term)
|
|
return wikipedia(term, last=True)
|
|
return None
|
|
sentence = m.group(0)
|
|
|
|
maxlength = 275
|
|
if len(sentence) > maxlength:
|
|
sentence = sentence[:maxlength]
|
|
words = sentence[:-5].split(' ')
|
|
words.pop()
|
|
sentence = ' '.join(words) + ' [...]'
|
|
|
|
if ((sentence == 'Wikipedia does not have an article with this exact name.')
|
|
or (sentence == 'Wikipedia does not have a page with this exact name.')):
|
|
if not last:
|
|
term = search(term)
|
|
return wikipedia(term, last=True)
|
|
return None
|
|
|
|
sentence = '"' + sentence.replace('"', "'") + '"'
|
|
sentence = sentence.decode('utf-8').encode('utf-8')
|
|
wikiuri = wikiuri.decode('utf-8').encode('utf-8')
|
|
term = term.decode('utf-8').encode('utf-8')
|
|
return sentence + ' - ' + (wikiuri % term)
|
|
|
|
def wik(phenny, input):
|
|
origterm = input.groups()[1]
|
|
if not origterm:
|
|
return phenny.say('Perhaps you meant ".wik Zen"?')
|
|
origterm = origterm.encode('utf-8')
|
|
|
|
term = urllib.unquote(origterm)
|
|
term = term[0].upper() + term[1:]
|
|
term = term.replace(' ', '_')
|
|
|
|
try: result = wikipedia(term)
|
|
except IOError:
|
|
error = "Can't connect to en.wikipedia.org (%s)" % (wikiuri % term)
|
|
return phenny.say(error)
|
|
|
|
if result is not None:
|
|
phenny.say(result)
|
|
else: phenny.say('Can\'t find anything in Wikipedia for "%s".' % origterm)
|
|
|
|
wik.commands = ['wik']
|
|
wik.priority = 'high'
|
|
|
|
if __name__ == '__main__':
|
|
print __doc__.strip()
|