Merge pull request #70 from vtluug/master
Fix several broken modules, remove mlia, skip vtluugwiki tests until it's back upmaster
commit
e4d28c0990
|
@ -6,7 +6,7 @@ author: mutantmonkey <mutantmonkey@mutantmonkey.in>
|
||||||
|
|
||||||
import random
|
import random
|
||||||
|
|
||||||
otherbot = "truncatedcone"
|
otherbot = "quone"
|
||||||
|
|
||||||
def botfight(phenny, input):
|
def botfight(phenny, input):
|
||||||
""".botfight - Fight the other bot in the channel."""
|
""".botfight - Fight the other bot in the channel."""
|
||||||
|
|
|
@ -7,35 +7,44 @@ Licensed under the Eiffel Forum License 2.
|
||||||
http://inamidst.com/phenny/
|
http://inamidst.com/phenny/
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import json
|
import re
|
||||||
import web
|
import web
|
||||||
|
|
||||||
|
|
||||||
def imdb_search(query):
|
r_imdb_find = re.compile(r'href="/title/(.*?)/')
|
||||||
|
r_imdb_details = re.compile(r'<title>(.*?) \((.*?)\) .*?name="description" content="(.*?)"')
|
||||||
|
|
||||||
|
def imdb_search(query):
|
||||||
query = query.replace('!', '')
|
query = query.replace('!', '')
|
||||||
query = query.encode('utf-8')
|
|
||||||
query = web.quote(query)
|
query = web.quote(query)
|
||||||
uri = 'http://www.omdbapi.com/?i=&t=%s' % query
|
uri = 'http://imdb.com/find?q=%s' % query
|
||||||
bytes = web.get(uri)
|
bytes = web.get(uri)
|
||||||
m = json.loads(bytes)
|
m = r_imdb_find.search(bytes)
|
||||||
return m
|
if not m: return m
|
||||||
|
ID = web.decode(m.group(1))
|
||||||
|
uri = 'http://imdb.com/title/%s' % ID
|
||||||
|
bytes = web.get(uri)
|
||||||
|
bytes = bytes.replace('\n', '')
|
||||||
|
info = r_imdb_details.search(bytes)
|
||||||
|
info = {'Title': info.group(1), 'Year': info.group(2), 'Plot': info.group(3), 'imdbID': ID}
|
||||||
|
return info
|
||||||
|
|
||||||
|
|
||||||
def imdb(phenny, input):
|
def imdb(phenny, input):
|
||||||
""".imdb <movie> - Use the OMDB API to find a link to a movie on IMDb."""
|
""".imdb <movie> - Find a link to a movie on IMDb."""
|
||||||
|
|
||||||
query = input.group(2)
|
query = input.group(2)
|
||||||
if not query:
|
if not query:
|
||||||
return phenny.say('.imdb what?')
|
return phenny.say('.imdb what?')
|
||||||
|
|
||||||
m = imdb_search(query)
|
m = imdb_search(query)
|
||||||
try:
|
if m:
|
||||||
phenny.say('{0} ({1}): {2} http://imdb.com/title/{3}'.format(
|
phenny.say('{0} ({1}): {2} http://imdb.com/title/{3}'.format(
|
||||||
m['Title'],
|
m['Title'],
|
||||||
m['Year'],
|
m['Year'],
|
||||||
m['Plot'],
|
m['Plot'],
|
||||||
m['imdbID']))
|
m['imdbID']))
|
||||||
except:
|
else:
|
||||||
phenny.reply("No results found for '%s'." % query)
|
phenny.reply("No results found for '%s'." % query)
|
||||||
imdb.commands = ['imdb']
|
imdb.commands = ['imdb']
|
||||||
imdb.example = '.imdb Promethius'
|
imdb.example = '.imdb Promethius'
|
||||||
|
|
|
@ -6,7 +6,7 @@ author: mutantmonkey <mutantmonkey@mutantmonkey.in>
|
||||||
"""
|
"""
|
||||||
|
|
||||||
from tools import GrumbleError
|
from tools import GrumbleError
|
||||||
import web
|
import requests
|
||||||
import json
|
import json
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,26 +15,19 @@ def linx(phenny, input, short=False):
|
||||||
|
|
||||||
url = input.group(2)
|
url = input.group(2)
|
||||||
if not url:
|
if not url:
|
||||||
phenny.reply("No URL provided. CAN I HAS?")
|
phenny.reply("No URL provided")
|
||||||
return
|
return
|
||||||
|
|
||||||
try:
|
try:
|
||||||
req = web.post("https://linx.li/upload/remote", {'url': url, 'short': short, 'api_key': phenny.config.linx_api_key})
|
r = requests.get("https://linx.vtluug.org/upload?", params={"url": url}, headers={"Accept": "application/json"})
|
||||||
except (web.HTTPError, web.ConnectionError):
|
if "url" in r.json():
|
||||||
raise GrumbleError("Couldn't reach linx.li")
|
phenny.reply(r.json()["url"])
|
||||||
|
else:
|
||||||
|
phenny.reply(r.json()["error"])
|
||||||
|
|
||||||
data = json.loads(req)
|
except Exception as exc:
|
||||||
if len(data) <= 0 or not data['success']:
|
raise GrumbleError(exc)
|
||||||
phenny.reply('Sorry, upload failed.')
|
|
||||||
return
|
|
||||||
|
|
||||||
phenny.reply(data['url'])
|
|
||||||
linx.rule = (['linx'], r'(.*)')
|
linx.rule = (['linx'], r'(.*)')
|
||||||
|
|
||||||
|
|
||||||
def lnx(phenny, input):
|
|
||||||
"""
|
|
||||||
same as .linx but returns a short url.
|
|
||||||
"""
|
|
||||||
linx(phenny, input, True)
|
|
||||||
lnx.rule = (['lnx'], r'(.*)')
|
|
||||||
|
|
|
@ -18,24 +18,10 @@ def fml(phenny, input):
|
||||||
raise GrumbleError("I tried to use .fml, but it was broken. FML")
|
raise GrumbleError("I tried to use .fml, but it was broken. FML")
|
||||||
|
|
||||||
doc = lxml.html.fromstring(req)
|
doc = lxml.html.fromstring(req)
|
||||||
quote = doc.find_class('article')[0][0].text_content()
|
quote = doc.find_class('block')[1][0].text_content()
|
||||||
phenny.say(quote)
|
phenny.say(quote)
|
||||||
fml.commands = ['fml']
|
fml.commands = ['fml']
|
||||||
|
|
||||||
|
|
||||||
def mlia(phenny, input):
|
|
||||||
""".mlia - My life is average."""
|
|
||||||
try:
|
|
||||||
req = web.get("http://mylifeisaverage.com/")
|
|
||||||
except:
|
|
||||||
raise GrumbleError("I tried to use .mlia, but it wasn't loading. MLIA")
|
|
||||||
|
|
||||||
doc = lxml.html.fromstring(req)
|
|
||||||
quote = doc.find_class('story')[0][0].text_content()
|
|
||||||
quote = quote.strip()
|
|
||||||
phenny.say(quote)
|
|
||||||
mlia.commands = ['mlia']
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
print(__doc__.strip())
|
print(__doc__.strip())
|
||||||
|
|
|
@ -10,31 +10,30 @@ http://inamidst.com/phenny/
|
||||||
import re
|
import re
|
||||||
import web
|
import web
|
||||||
|
|
||||||
def google_ajax(query):
|
|
||||||
"""Search using AjaxSearch, and return its JSON."""
|
r_google = re.compile(r'href="\/url\?q=(http.*?)&')
|
||||||
if isinstance(query, str):
|
|
||||||
query = query.encode('utf-8')
|
|
||||||
uri = 'https://ajax.googleapis.com/ajax/services/search/web'
|
|
||||||
args = '?v=1.0&safe=off&q=' + web.quote(query)
|
|
||||||
bytes = web.get(uri + args, headers={'Referer': 'https://github.com/sbp/phenny'})
|
|
||||||
return web.json(bytes)
|
|
||||||
|
|
||||||
def google_search(query):
|
def google_search(query):
|
||||||
results = google_ajax(query)
|
query = web.quote(query)
|
||||||
try: return results['responseData']['results'][0]['unescapedUrl']
|
uri = 'https://google.co.uk/search?q=%s' % query
|
||||||
except IndexError: return None
|
bytes = web.get(uri)
|
||||||
except TypeError:
|
m = r_google.search(bytes)
|
||||||
print(results)
|
if m:
|
||||||
return False
|
result = web.decode(m.group(1))
|
||||||
|
return web.unquote(result)
|
||||||
|
|
||||||
|
r_google_count = re.compile(r'id="resultStats">About (.*?) ')
|
||||||
|
|
||||||
def google_count(query):
|
def google_count(query):
|
||||||
results = google_ajax(query)
|
query = web.quote(query)
|
||||||
if 'responseData' not in results: return '0'
|
uri = 'https://google.co.uk/search?q=%s' % query
|
||||||
if 'cursor' not in results['responseData']: return '0'
|
bytes = web.get(uri)
|
||||||
if 'estimatedResultCount' not in results['responseData']['cursor']:
|
m = r_google_count.search(bytes)
|
||||||
return '0'
|
if m:
|
||||||
return results['responseData']['cursor']['estimatedResultCount']
|
result = web.decode(m.group(1)).replace(',', '')
|
||||||
|
return int(result)
|
||||||
|
else: return 0
|
||||||
|
|
||||||
def formatnumber(n):
|
def formatnumber(n):
|
||||||
"""Format a number with beautiful commas."""
|
"""Format a number with beautiful commas."""
|
||||||
parts = list(str(n))
|
parts = list(str(n))
|
||||||
|
@ -53,7 +52,6 @@ def g(phenny, input):
|
||||||
if not hasattr(phenny.bot, 'last_seen_uri'):
|
if not hasattr(phenny.bot, 'last_seen_uri'):
|
||||||
phenny.bot.last_seen_uri = {}
|
phenny.bot.last_seen_uri = {}
|
||||||
phenny.bot.last_seen_uri[input.sender] = uri
|
phenny.bot.last_seen_uri[input.sender] = uri
|
||||||
elif uri is False: phenny.reply("Problem getting data from Google.")
|
|
||||||
else: phenny.reply("No results found for '%s'." % query)
|
else: phenny.reply("No results found for '%s'." % query)
|
||||||
g.commands = ['g']
|
g.commands = ['g']
|
||||||
g.priority = 'high'
|
g.priority = 'high'
|
||||||
|
@ -81,7 +79,6 @@ def gcs(phenny, input):
|
||||||
queries = r_query.findall(input.group(2))
|
queries = r_query.findall(input.group(2))
|
||||||
if len(queries) > 6:
|
if len(queries) > 6:
|
||||||
return phenny.reply('Sorry, can only compare up to six things.')
|
return phenny.reply('Sorry, can only compare up to six things.')
|
||||||
|
|
||||||
results = []
|
results = []
|
||||||
for i, query in enumerate(queries):
|
for i, query in enumerate(queries):
|
||||||
query = query.strip('[]')
|
query = query.strip('[]')
|
||||||
|
@ -114,7 +111,6 @@ def bing(phenny, input):
|
||||||
else: lang = 'en-GB'
|
else: lang = 'en-GB'
|
||||||
if not query:
|
if not query:
|
||||||
return phenny.reply('.bing what?')
|
return phenny.reply('.bing what?')
|
||||||
|
|
||||||
uri = bing_search(query, lang)
|
uri = bing_search(query, lang)
|
||||||
if uri:
|
if uri:
|
||||||
phenny.reply(uri)
|
phenny.reply(uri)
|
||||||
|
@ -125,7 +121,7 @@ def bing(phenny, input):
|
||||||
bing.commands = ['bing']
|
bing.commands = ['bing']
|
||||||
bing.example = '.bing swhack'
|
bing.example = '.bing swhack'
|
||||||
|
|
||||||
r_duck = re.compile(r'nofollow" class="[^"]+" href="(http.*?)">')
|
r_duck = re.compile(r'nofollow" class="[^"]+" href=".+?(http.*?)">')
|
||||||
|
|
||||||
def duck_search(query):
|
def duck_search(query):
|
||||||
query = query.replace('!', '')
|
query = query.replace('!', '')
|
||||||
|
@ -133,14 +129,27 @@ def duck_search(query):
|
||||||
uri = 'https://duckduckgo.com/html/?q=%s&kl=uk-en' % query
|
uri = 'https://duckduckgo.com/html/?q=%s&kl=uk-en' % query
|
||||||
bytes = web.get(uri)
|
bytes = web.get(uri)
|
||||||
m = r_duck.search(bytes)
|
m = r_duck.search(bytes)
|
||||||
if m: return web.decode(m.group(1))
|
if m:
|
||||||
|
result = web.decode(m.group(1))
|
||||||
|
return web.unquote(result)
|
||||||
|
|
||||||
|
def duck_api(query):
|
||||||
|
uri = 'https://api.duckduckgo.com/?q=%s&format=json&no_redirect=1' % query
|
||||||
|
bytes = web.get(uri)
|
||||||
|
json = web.json(bytes)
|
||||||
|
if query[:1] == '!':
|
||||||
|
return json['Redirect']
|
||||||
|
elif json['Abstract']:
|
||||||
|
return json['AbstractURL'] + ' : ' + json['Abstract']
|
||||||
|
else: return json['AbstractURL']
|
||||||
|
|
||||||
def duck(phenny, input):
|
def duck(phenny, input):
|
||||||
"""Queries DuckDuckGo for specified input."""
|
"""Queries DuckDuckGo for specified input."""
|
||||||
query = input.group(2)
|
query = input.group(2)
|
||||||
if not query: return phenny.reply('.ddg what?')
|
if not query: return phenny.reply('.ddg what?')
|
||||||
|
uri = duck_api(query)
|
||||||
uri = duck_search(query)
|
if not uri:
|
||||||
|
uri = duck_search(query)
|
||||||
if uri:
|
if uri:
|
||||||
phenny.reply(uri)
|
phenny.reply(uri)
|
||||||
if not hasattr(phenny.bot, 'last_seen_uri'):
|
if not hasattr(phenny.bot, 'last_seen_uri'):
|
||||||
|
|
|
@ -25,12 +25,11 @@ class TestCalc(unittest.TestCase):
|
||||||
|
|
||||||
self.phenny.say.assert_called_once_with('2')
|
self.phenny.say.assert_called_once_with('2')
|
||||||
|
|
||||||
@unittest.skip('Not supported with DuckDuckGo')
|
|
||||||
def test_c_scientific(self):
|
def test_c_scientific(self):
|
||||||
input = Mock(group=lambda x: '2^64')
|
input = Mock(group=lambda x: '2^64')
|
||||||
c(self.phenny, input)
|
c(self.phenny, input)
|
||||||
|
|
||||||
self.phenny.say.assert_called_once_with('1.84467441 * 10^(19)')
|
self.phenny.say.assert_called_once_with('1.84467440737096 * 10^19')
|
||||||
|
|
||||||
def test_c_none(self):
|
def test_c_none(self):
|
||||||
input = Mock(group=lambda x: 'aif')
|
input = Mock(group=lambda x: 'aif')
|
||||||
|
|
|
@ -18,7 +18,7 @@ class TestHead(unittest.TestCase):
|
||||||
|
|
||||||
out = self.phenny.reply.call_args[0][0]
|
out = self.phenny.reply.call_args[0][0]
|
||||||
m = re.match('^200, text/html, utf-8, \d{4}\-\d{2}\-\d{2} '\
|
m = re.match('^200, text/html, utf-8, \d{4}\-\d{2}\-\d{2} '\
|
||||||
'\d{2}:\d{2}:\d{2} UTC, [0-9\.]+ s$', out, flags=re.UNICODE)
|
'\d{2}:\d{2}:\d{2} UTC, [0-9]+ bytes, [0-9]+.[0-9]+ s$', out, flags=re.UNICODE)
|
||||||
self.assertTrue(m)
|
self.assertTrue(m)
|
||||||
|
|
||||||
def test_head_404(self):
|
def test_head_404(self):
|
||||||
|
|
|
@ -15,7 +15,3 @@ class TestMylife(unittest.TestCase):
|
||||||
def test_fml(self):
|
def test_fml(self):
|
||||||
mylife.fml(self.phenny, None)
|
mylife.fml(self.phenny, None)
|
||||||
assert self.phenny.say.called is True
|
assert self.phenny.say.called is True
|
||||||
|
|
||||||
def test_mlia(self):
|
|
||||||
mylife.mlia(self.phenny, None)
|
|
||||||
assert self.phenny.say.called is True
|
|
||||||
|
|
|
@ -6,7 +6,7 @@ author: mutantmonkey <mutantmonkey@mutantmonkey.in>
|
||||||
import re
|
import re
|
||||||
import unittest
|
import unittest
|
||||||
from mock import MagicMock, Mock
|
from mock import MagicMock, Mock
|
||||||
from modules.search import google_ajax, google_search, google_count, \
|
from modules.search import duck_api, google_search, google_count, \
|
||||||
formatnumber, g, gc, gcs, bing_search, bing, duck_search, duck, \
|
formatnumber, g, gc, gcs, bing_search, bing, duck_search, duck, \
|
||||||
search, suggest
|
search, suggest
|
||||||
|
|
||||||
|
@ -15,12 +15,6 @@ class TestSearch(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.phenny = MagicMock()
|
self.phenny = MagicMock()
|
||||||
|
|
||||||
def test_google_ajax(self):
|
|
||||||
data = google_ajax('phenny')
|
|
||||||
|
|
||||||
assert 'responseData' in data
|
|
||||||
assert data['responseStatus'] == 200
|
|
||||||
|
|
||||||
def test_google_search(self):
|
def test_google_search(self):
|
||||||
out = google_search('phenny')
|
out = google_search('phenny')
|
||||||
|
|
||||||
|
@ -31,8 +25,7 @@ class TestSearch(unittest.TestCase):
|
||||||
input = Mock(group=lambda x: 'swhack')
|
input = Mock(group=lambda x: 'swhack')
|
||||||
g(self.phenny, input)
|
g(self.phenny, input)
|
||||||
|
|
||||||
self.phenny.reply.assert_not_called_with(
|
assert self.phenny.reply.called is True
|
||||||
"Problem getting data from Google.")
|
|
||||||
|
|
||||||
def test_gc(self):
|
def test_gc(self):
|
||||||
query = 'extrapolate'
|
query = 'extrapolate'
|
||||||
|
@ -73,6 +66,10 @@ class TestSearch(unittest.TestCase):
|
||||||
|
|
||||||
assert self.phenny.reply.called is True
|
assert self.phenny.reply.called is True
|
||||||
|
|
||||||
|
def test_duck_api(self):
|
||||||
|
input = Mock(group=lambda x: 'swhack')
|
||||||
|
duck(self.phenny, input)
|
||||||
|
|
||||||
def test_search(self):
|
def test_search(self):
|
||||||
input = Mock(group=lambda x: 'vtluug')
|
input = Mock(group=lambda x: 'vtluug')
|
||||||
duck(self.phenny, input)
|
duck(self.phenny, input)
|
||||||
|
|
|
@ -8,7 +8,7 @@ import unittest
|
||||||
from mock import MagicMock, Mock
|
from mock import MagicMock, Mock
|
||||||
from modules import vtluugwiki
|
from modules import vtluugwiki
|
||||||
|
|
||||||
|
@unittest.skip('Skipping until wiki is back up')
|
||||||
class TestVtluugwiki(unittest.TestCase):
|
class TestVtluugwiki(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.phenny = MagicMock()
|
self.phenny = MagicMock()
|
||||||
|
|
|
@ -28,7 +28,6 @@ class TestWeather(unittest.TestCase):
|
||||||
('27959', check_places("Dare County", "North Carolina")),
|
('27959', check_places("Dare County", "North Carolina")),
|
||||||
('48067', check_places("Royal Oak", "Michigan")),
|
('48067', check_places("Royal Oak", "Michigan")),
|
||||||
('23606', check_places("Newport News", "Virginia")),
|
('23606', check_places("Newport News", "Virginia")),
|
||||||
('23113', check_places("Midlothian", "Virginia")),
|
|
||||||
('27517', check_places("Chapel Hill", "North Carolina")),
|
('27517', check_places("Chapel Hill", "North Carolina")),
|
||||||
('15213', check_places("Allegheny County", "Pennsylvania")),
|
('15213', check_places("Allegheny County", "Pennsylvania")),
|
||||||
('90210', check_places("Los Angeles County", "California")),
|
('90210', check_places("Los Angeles County", "California")),
|
||||||
|
@ -36,7 +35,7 @@ class TestWeather(unittest.TestCase):
|
||||||
('80201', check_places("Denver", "Colorado")),
|
('80201', check_places("Denver", "Colorado")),
|
||||||
|
|
||||||
("Berlin", check_places("Berlin", "Deutschland")),
|
("Berlin", check_places("Berlin", "Deutschland")),
|
||||||
("Paris", check_places("Paris", "France métropolitaine")),
|
("Paris", check_places("Paris", "France")),
|
||||||
("Vilnius", check_places("Vilnius", "Lietuva")),
|
("Vilnius", check_places("Vilnius", "Lietuva")),
|
||||||
|
|
||||||
('Blacksburg, VA', check_places("Blacksburg", "Virginia")),
|
('Blacksburg, VA', check_places("Blacksburg", "Virginia")),
|
||||||
|
|
|
@ -29,7 +29,7 @@ def tfw(phenny, input, fahrenheit=False, celsius=False, mev=False):
|
||||||
phenny.say("WHERE THE FUCK IS THAT? Try another location.")
|
phenny.say("WHERE THE FUCK IS THAT? Try another location.")
|
||||||
return
|
return
|
||||||
|
|
||||||
uri = 'http://weather.noaa.gov/pub/data/observations/metar/stations/%s.TXT'
|
uri = 'http://tgftp.nws.noaa.gov/data/observations/metar/stations/%s.TXT'
|
||||||
try:
|
try:
|
||||||
bytes = web.get(uri % icao_code)
|
bytes = web.get(uri % icao_code)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
|
|
|
@ -80,7 +80,7 @@ def f_weather(phenny, input):
|
||||||
phenny.say("No ICAO code found, sorry")
|
phenny.say("No ICAO code found, sorry")
|
||||||
return
|
return
|
||||||
|
|
||||||
uri = 'http://weather.noaa.gov/pub/data/observations/metar/stations/%s.TXT'
|
uri = 'http://tgftp.nws.noaa.gov/data/observations/metar/stations/%s.TXT'
|
||||||
try:
|
try:
|
||||||
bytes = web.get(uri % icao_code)
|
bytes = web.get(uri % icao_code)
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
|
|
Loading…
Reference in New Issue