mirror of
https://github.com/Mikaela/Limnoria.git
synced 2024-11-14 22:49:23 +01:00
b533290c7a
Upstream bug: http://bugs.python.org/issue3932 Rather than override the unescape method with the patch posted, we just convert the page text to unicode before passing it to the HTMLParser. UTF8 and Latin1 will eat just about anything.
257 lines
9.2 KiB
Python
257 lines
9.2 KiB
Python
###
|
|
# Copyright (c) 2005, Jeremiah Fincher
|
|
# Copyright (c) 2009, James Vega
|
|
# All rights reserved.
|
|
#
|
|
# Redistribution and use in source and binary forms, with or without
|
|
# modification, are permitted provided that the following conditions are met:
|
|
#
|
|
# * Redistributions of source code must retain the above copyright notice,
|
|
# this list of conditions, and the following disclaimer.
|
|
# * Redistributions in binary form must reproduce the above copyright notice,
|
|
# this list of conditions, and the following disclaimer in the
|
|
# documentation and/or other materials provided with the distribution.
|
|
# * Neither the name of the author of this software nor the name of
|
|
# contributors to this software may be used to endorse or promote products
|
|
# derived from this software without specific prior written consent.
|
|
#
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
# POSSIBILITY OF SUCH DAMAGE.
|
|
###
|
|
|
|
import re
|
|
import HTMLParser
|
|
import htmlentitydefs
|
|
|
|
import supybot.conf as conf
|
|
import supybot.utils as utils
|
|
from supybot.commands import *
|
|
import supybot.plugins as plugins
|
|
import supybot.ircutils as ircutils
|
|
import supybot.callbacks as callbacks
|
|
|
|
class Title(HTMLParser.HTMLParser):
|
|
entitydefs = htmlentitydefs.entitydefs.copy()
|
|
entitydefs['nbsp'] = ' '
|
|
entitydefs['apos'] = '\''
|
|
def __init__(self):
|
|
self.inTitle = False
|
|
self.title = ''
|
|
HTMLParser.HTMLParser.__init__(self)
|
|
|
|
def handle_starttag(self, tag, attrs):
|
|
if tag == 'title':
|
|
self.inTitle = True
|
|
|
|
def handle_endtag(self, tag):
|
|
if tag == 'title':
|
|
self.inTitle = False
|
|
|
|
def handle_data(self, data):
|
|
if self.inTitle:
|
|
self.title += data
|
|
|
|
def handle_entityref(self, name):
|
|
if self.inTitle:
|
|
if name in self.entitydefs:
|
|
self.title += self.entitydefs[name]
|
|
|
|
class Web(callbacks.PluginRegexp):
|
|
"""Add the help for "@help Web" here."""
|
|
threaded = True
|
|
regexps = ['titleSnarfer']
|
|
def callCommand(self, command, irc, msg, *args, **kwargs):
|
|
try:
|
|
super(Web, self).callCommand(command, irc, msg, *args, **kwargs)
|
|
except utils.web.Error, e:
|
|
irc.reply(str(e))
|
|
|
|
def titleSnarfer(self, irc, msg, match):
|
|
channel = msg.args[0]
|
|
if not irc.isChannel(channel):
|
|
return
|
|
if callbacks.addressed(irc.nick, msg):
|
|
return
|
|
if self.registryValue('titleSnarfer', channel):
|
|
url = match.group(0)
|
|
r = self.registryValue('nonSnarfingRegexp', channel)
|
|
if r and r.search(url):
|
|
self.log.debug('Not titleSnarfing %q.', url)
|
|
return
|
|
try:
|
|
size = conf.supybot.protocols.http.peekSize()
|
|
text = utils.web.getUrl(url, size=size)
|
|
try:
|
|
text = text.decode('utf8')
|
|
except UnicodeDecodeError:
|
|
text = text.decode('latin1')
|
|
except utils.web.Error, e:
|
|
self.log.info('Couldn\'t snarf title of %u: %s.', url, e)
|
|
return
|
|
parser = Title()
|
|
try:
|
|
parser.feed(text)
|
|
except HTMLParser.HTMLParseError:
|
|
self.log.debug('Encountered a problem parsing %u. Title may '
|
|
'already be set, though', url)
|
|
if parser.title:
|
|
domain = utils.web.getDomain(url)
|
|
title = utils.web.htmlToText(parser.title.strip())
|
|
s = format('Title: %s (at %s)', title, domain)
|
|
irc.reply(s, prefixNick=False)
|
|
titleSnarfer = urlSnarfer(titleSnarfer)
|
|
titleSnarfer.__doc__ = utils.web._httpUrlRe
|
|
|
|
def headers(self, irc, msg, args, url):
|
|
"""<url>
|
|
|
|
Returns the HTTP headers of <url>. Only HTTP urls are valid, of
|
|
course.
|
|
"""
|
|
fd = utils.web.getUrlFd(url)
|
|
try:
|
|
s = ', '.join([format('%s: %s', k, v)
|
|
for (k, v) in fd.headers.items()])
|
|
irc.reply(s)
|
|
finally:
|
|
fd.close()
|
|
headers = wrap(headers, ['httpUrl'])
|
|
|
|
_doctypeRe = re.compile(r'(<!DOCTYPE[^>]+>)', re.M)
|
|
def doctype(self, irc, msg, args, url):
|
|
"""<url>
|
|
|
|
Returns the DOCTYPE string of <url>. Only HTTP urls are valid, of
|
|
course.
|
|
"""
|
|
size = conf.supybot.protocols.http.peekSize()
|
|
s = utils.web.getUrl(url, size=size)
|
|
m = self._doctypeRe.search(s)
|
|
if m:
|
|
s = utils.str.normalizeWhitespace(m.group(0))
|
|
irc.reply(s)
|
|
else:
|
|
irc.reply('That URL has no specified doctype.')
|
|
doctype = wrap(doctype, ['httpUrl'])
|
|
|
|
def size(self, irc, msg, args, url):
|
|
"""<url>
|
|
|
|
Returns the Content-Length header of <url>. Only HTTP urls are valid,
|
|
of course.
|
|
"""
|
|
fd = utils.web.getUrlFd(url)
|
|
try:
|
|
try:
|
|
size = fd.headers['Content-Length']
|
|
irc.reply(format('%u is %i bytes long.', url, size))
|
|
except KeyError:
|
|
size = conf.supybot.protocols.http.peekSize()
|
|
s = fd.read(size)
|
|
if len(s) != size:
|
|
irc.reply(format('%u is %i bytes long.', url, len(s)))
|
|
else:
|
|
irc.reply(format('The server didn\'t tell me how long %u '
|
|
'is but it\'s longer than %i bytes.',
|
|
url, size))
|
|
finally:
|
|
fd.close()
|
|
size = wrap(size, ['httpUrl'])
|
|
|
|
def title(self, irc, msg, args, url):
|
|
"""<url>
|
|
|
|
Returns the HTML <title>...</title> of a URL.
|
|
"""
|
|
size = conf.supybot.protocols.http.peekSize()
|
|
text = utils.web.getUrl(url, size=size)
|
|
try:
|
|
text = text.decode('utf8')
|
|
except UnicodeDecodeError:
|
|
text = text.decode('latin1')
|
|
parser = Title()
|
|
try:
|
|
parser.feed(text)
|
|
except HTMLParser.HTMLParseError:
|
|
self.log.debug('Encountered a problem parsing %u. Title may '
|
|
'already be set, though', url)
|
|
if parser.title:
|
|
irc.reply(utils.web.htmlToText(parser.title.strip()))
|
|
elif len(text) < size:
|
|
irc.reply('That URL appears to have no HTML title.')
|
|
else:
|
|
irc.reply(format('That URL appears to have no HTML title '
|
|
'within the first %i bytes.', size))
|
|
title = wrap(title, ['httpUrl'])
|
|
|
|
_netcraftre = re.compile(r'td align="left">\s+<a[^>]+>(.*?)<a href',
|
|
re.S | re.I)
|
|
def netcraft(self, irc, msg, args, hostname):
|
|
"""<hostname|ip>
|
|
|
|
Returns Netcraft.com's determination of what operating system and
|
|
webserver is running on the host given.
|
|
"""
|
|
url = 'http://uptime.netcraft.com/up/graph/?host=' + hostname
|
|
html = utils.web.getUrl(url)
|
|
m = self._netcraftre.search(html)
|
|
if m:
|
|
html = m.group(1)
|
|
s = utils.web.htmlToText(html, tagReplace='').strip()
|
|
s = s.rstrip('-').strip()
|
|
irc.reply(s) # Snip off "the site"
|
|
elif 'We could not get any results' in html:
|
|
irc.reply('No results found for %s.' % hostname)
|
|
else:
|
|
irc.error('The format of page the was odd.')
|
|
netcraft = wrap(netcraft, ['text'])
|
|
|
|
def urlquote(self, irc, msg, args, text):
|
|
"""<text>
|
|
|
|
Returns the URL quoted form of the text.
|
|
"""
|
|
irc.reply(utils.web.urlquote(text))
|
|
urlquote = wrap(urlquote, ['text'])
|
|
|
|
def urlunquote(self, irc, msg, args, text):
|
|
"""<text>
|
|
|
|
Returns the text un-URL quoted.
|
|
"""
|
|
s = utils.web.urlunquote(text)
|
|
irc.reply(s)
|
|
urlunquote = wrap(urlunquote, ['text'])
|
|
|
|
def fetch(self, irc, msg, args, url):
|
|
"""<url>
|
|
|
|
Returns the contents of <url>, or as much as is configured in
|
|
supybot.plugins.Web.fetch.maximum. If that configuration variable is
|
|
set to 0, this command will be effectively disabled.
|
|
"""
|
|
max = self.registryValue('fetch.maximum')
|
|
if not max:
|
|
irc.error('This command is disabled '
|
|
'(supybot.plugins.Web.fetch.maximum is set to 0).',
|
|
Raise=True)
|
|
timeout = self.registryValue('fetch.timeout')
|
|
if timeout == 0:
|
|
timeout = None
|
|
fd = utils.web.getUrlFd(url, timeout=timeout)
|
|
irc.reply(fd.read(max))
|
|
fetch = wrap(fetch, ['url'])
|
|
|
|
Class = Web
|
|
|
|
# vim:set shiftwidth=4 softtabstop=4 expandtab textwidth=79:
|