Last active
March 26, 2018 08:57
-
-
Save m0n5t3r/b35c81aaa0ca249d00a8e58ec1c51b1e to your computer and use it in GitHub Desktop.
supybot web plugin using bs4 instead of HtmlParser
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
### | |
# Copyright (c) 2005, Jeremiah Fincher | |
# Copyright (c) 2009, James McCoy | |
# All rights reserved. | |
# | |
# Redistribution and use in source and binary forms, with or without | |
# modification, are permitted provided that the following conditions are met: | |
# | |
# * Redistributions of source code must retain the above copyright notice, | |
# this list of conditions, and the following disclaimer. | |
# * Redistributions in binary form must reproduce the above copyright notice, | |
# this list of conditions, and the following disclaimer in the | |
# documentation and/or other materials provided with the distribution. | |
# * Neither the name of the author of this software nor the name of | |
# contributors to this software may be used to endorse or promote products | |
# derived from this software without specific prior written consent. | |
# | |
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | |
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE | |
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR | |
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF | |
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS | |
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN | |
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) | |
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | |
# POSSIBILITY OF SUCH DAMAGE. | |
### | |
import re | |
from bs4 import BeautifulSoup | |
import supybot.conf as conf | |
import supybot.utils as utils | |
from supybot.commands import * | |
import supybot.plugins as plugins | |
import supybot.ircutils as ircutils | |
import supybot.callbacks as callbacks | |
class Web(callbacks.PluginRegexp): | |
"""Add the help for "@help Web" here.""" | |
threaded = True | |
regexps = ['titleSnarfer'] | |
def callCommand(self, command, irc, msg, *args, **kwargs): | |
try: | |
super(Web, self).callCommand(command, irc, msg, *args, **kwargs) | |
except utils.web.Error, e: | |
irc.reply(str(e)) | |
def titleSnarfer(self, irc, msg, match): | |
channel = msg.args[0] | |
if not irc.isChannel(channel): | |
return | |
if callbacks.addressed(irc.nick, msg): | |
return | |
if self.registryValue('titleSnarfer', channel): | |
url = match.group(0) | |
r = self.registryValue('nonSnarfingRegexp', channel) | |
if r and r.search(url): | |
self.log.debug('Not titleSnarfing %q.', url) | |
return | |
try: | |
size = conf.supybot.protocols.http.peekSize() | |
text = utils.web.getUrl(url, size=size) | |
except utils.web.Error, e: | |
self.log.info('Couldn\'t snarf title of %u: %s.', url, e) | |
return | |
parser = BeautifulSoup(text) | |
try: | |
title = parser.find(name='title').text | |
except: | |
self.log.debug('Encountered a problem parsing %u.', url) | |
if title: | |
domain = utils.web.getDomain(url) | |
s = format('Title: %s (at %s)', title, domain) | |
irc.reply(s, prefixNick=False) | |
titleSnarfer = urlSnarfer(titleSnarfer) | |
titleSnarfer.__doc__ = utils.web._httpUrlRe | |
def headers(self, irc, msg, args, url): | |
"""<url> | |
Returns the HTTP headers of <url>. Only HTTP urls are valid, of | |
course. | |
""" | |
fd = utils.web.getUrlFd(url) | |
try: | |
s = ', '.join([format('%s: %s', k, v) | |
for (k, v) in fd.headers.items()]) | |
irc.reply(s) | |
finally: | |
fd.close() | |
headers = wrap(headers, ['httpUrl']) | |
_doctypeRe = re.compile(r'(<!DOCTYPE[^>]+>)', re.M) | |
def doctype(self, irc, msg, args, url): | |
"""<url> | |
Returns the DOCTYPE string of <url>. Only HTTP urls are valid, of | |
course. | |
""" | |
size = conf.supybot.protocols.http.peekSize() | |
s = utils.web.getUrl(url, size=size) | |
m = self._doctypeRe.search(s) | |
if m: | |
s = utils.str.normalizeWhitespace(m.group(0)) | |
irc.reply(s) | |
else: | |
irc.reply('That URL has no specified doctype.') | |
doctype = wrap(doctype, ['httpUrl']) | |
def size(self, irc, msg, args, url): | |
"""<url> | |
Returns the Content-Length header of <url>. Only HTTP urls are valid, | |
of course. | |
""" | |
fd = utils.web.getUrlFd(url) | |
try: | |
try: | |
size = fd.headers['Content-Length'] | |
irc.reply(format('%u is %i bytes long.', url, size)) | |
except KeyError: | |
size = conf.supybot.protocols.http.peekSize() | |
s = fd.read(size) | |
if len(s) != size: | |
irc.reply(format('%u is %i bytes long.', url, len(s))) | |
else: | |
irc.reply(format('The server didn\'t tell me how long %u ' | |
'is but it\'s longer than %i bytes.', | |
url, size)) | |
finally: | |
fd.close() | |
size = wrap(size, ['httpUrl']) | |
def title(self, irc, msg, args, url): | |
"""<url> | |
Returns the HTML <title>...</title> of a URL. | |
""" | |
size = conf.supybot.protocols.http.peekSize() | |
text = utils.web.getUrl(url, size=size) | |
parser = bs4.BeautifulSoup(text) | |
try: | |
title = parser.find(name='title').text | |
except: | |
self.log.debug('Encountered a problem parsing %u.', url) | |
if title: | |
irc.reply(utils.web.htmlToText(title.strip())) | |
elif len(text) < size: | |
irc.reply('That URL appears to have no HTML title.') | |
else: | |
irc.reply(format('That URL appears to have no HTML title ' | |
'within the first %i bytes.', size)) | |
title = wrap(title, ['httpUrl']) | |
_netcraftre = re.compile(r'td align="left">\s+<a[^>]+>(.*?)<a href', | |
re.S | re.I) | |
def netcraft(self, irc, msg, args, hostname): | |
"""<hostname|ip> | |
Returns Netcraft.com's determination of what operating system and | |
webserver is running on the host given. | |
""" | |
url = 'http://uptime.netcraft.com/up/graph/?host=' + hostname | |
html = utils.web.getUrl(url) | |
m = self._netcraftre.search(html) | |
if m: | |
html = m.group(1) | |
s = utils.web.htmlToText(html, tagReplace='').strip() | |
s = s.rstrip('-').strip() | |
irc.reply(s) # Snip off "the site" | |
elif 'We could not get any results' in html: | |
irc.reply('No results found for %s.' % hostname) | |
else: | |
irc.error('The format of page the was odd.') | |
netcraft = wrap(netcraft, ['text']) | |
def urlquote(self, irc, msg, args, text): | |
"""<text> | |
Returns the URL quoted form of the text. | |
""" | |
irc.reply(utils.web.urlquote(text)) | |
urlquote = wrap(urlquote, ['text']) | |
def urlunquote(self, irc, msg, args, text): | |
"""<text> | |
Returns the text un-URL quoted. | |
""" | |
s = utils.web.urlunquote(text) | |
irc.reply(s) | |
urlunquote = wrap(urlunquote, ['text']) | |
def fetch(self, irc, msg, args, url): | |
"""<url> | |
Returns the contents of <url>, or as much as is configured in | |
supybot.plugins.Web.fetch.maximum. If that configuration variable is | |
set to 0, this command will be effectively disabled. | |
""" | |
max = self.registryValue('fetch.maximum') | |
if not max: | |
irc.error('This command is disabled ' | |
'(supybot.plugins.Web.fetch.maximum is set to 0).', | |
Raise=True) | |
timeout = self.registryValue('fetch.timeout') | |
if timeout == 0: | |
timeout = None | |
s = utils.web.getUrl(url, timeout=timeout, size=max) | |
irc.reply(s) | |
fetch = wrap(fetch, ['url']) | |
Class = Web | |
# vim:set shiftwidth=4 softtabstop=4 expandtab textwidth=79: |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment