mirror of
				https://github.com/python/cpython.git
				synced 2025-10-31 02:15:10 +00:00 
			
		
		
		
	 545006259d
			
		
	
	
		545006259d
		
	
	
	
	
		
			
			Unfortunately his code breaks wcgui.py in a way that's not easy to fix. I expect that this is a temporary situation -- eventually Sam's changes will be merged back in. (The changes add a -t option to specify exceptions to the -x option, and explicit checking for #foo style fragment ids.)
		
			
				
	
	
		
			884 lines
		
	
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable file
		
	
	
	
	
			
		
		
	
	
			884 lines
		
	
	
	
		
			28 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable file
		
	
	
	
	
| #! /usr/bin/env python
 | |
| 
 | |
| # Modifications where indicated by Samuel Bayer (SLB), 2/24/99,
 | |
| # to support checking of links to internal NAME anchors. Key 
 | |
| # modifications are in the dopage() method of the Checker class,
 | |
| # where the fragments (internal NAME anchors) associated with
 | |
| # a page are checked, and in the getlinkinfos() method of the
 | |
| # Page class, where the fragment element of the result of
 | |
| # urlparse() is saved, rather than discarded. The global effect
 | |
| # of this change is to make the entries in the "done" and 
 | |
| # "todo" instance variables of Checker into a pair (URL, fragment)
 | |
| # rather than simply a URL. This change has implications in 
 | |
| # the following methods:
 | |
| 
 | |
| # addroot() of Checker
 | |
| # report_errors() of Checker
 | |
| # dopage() of Checker
 | |
| # newdonelink() of Checker
 | |
| # newtodolink() of Checker
 | |
| # format_url() of Checker (new)
 | |
| # getpage() of Checker
 | |
| # readhtml() of Checker
 | |
| # openhtml() of Checker
 | |
| # openpage() of Checker
 | |
| # seterror() of Checker
 | |
| # __init__() of Page()
 | |
| # getnames() of Page() (new)
 | |
| # getlinkinfos() of Page()
 | |
| # start_a() of MyHTMLParser()
 | |
| 
 | |
| # Also added new command line argument to support specification of
 | |
| # additional roots, for the circumstance where a root the root index.html
 | |
| # page is not in a directory which dominates the entire tree of 
 | |
| # relevant internal references.
 | |
| 
 | |
| """Web tree checker.
 | |
| 
 | |
| This utility is handy to check a subweb of the world-wide web for
 | |
| errors.  A subweb is specified by giving one or more ``root URLs''; a
 | |
| page belongs to the subweb if one of the root URLs is an initial
 | |
| prefix of it.
 | |
| 
 | |
| File URL extension:
 | |
| 
 | |
| In order to easy the checking of subwebs via the local file system,
 | |
| the interpretation of ``file:'' URLs is extended to mimic the behavior
 | |
| of your average HTTP daemon: if a directory pathname is given, the
 | |
| file index.html in that directory is returned if it exists, otherwise
 | |
| a directory listing is returned.  Now, you can point webchecker to the
 | |
| document tree in the local file system of your HTTP daemon, and have
 | |
| most of it checked.  In fact the default works this way if your local
 | |
| web tree is located at /usr/local/etc/httpd/htdpcs (the default for
 | |
| the NCSA HTTP daemon and probably others).
 | |
| 
 | |
| Report printed:
 | |
| 
 | |
| When done, it reports pages with bad links within the subweb.  When
 | |
| interrupted, it reports for the pages that it has checked already.
 | |
| 
 | |
| In verbose mode, additional messages are printed during the
 | |
| information gathering phase.  By default, it prints a summary of its
 | |
| work status every 50 URLs (adjustable with the -r option), and it
 | |
| reports errors as they are encountered.  Use the -q option to disable
 | |
| this output.
 | |
| 
 | |
| Checkpoint feature:
 | |
| 
 | |
| Whether interrupted or not, it dumps its state (a Python pickle) to a
 | |
| checkpoint file and the -R option allows it to restart from the
 | |
| checkpoint (assuming that the pages on the subweb that were already
 | |
| processed haven't changed).  Even when it has run till completion, -R
 | |
| can still be useful -- it will print the reports again, and -Rq prints
 | |
| the errors only.  In this case, the checkpoint file is not written
 | |
| again.  The checkpoint file can be set with the -d option.
 | |
| 
 | |
| The checkpoint file is written as a Python pickle.  Remember that
 | |
| Python's pickle module is currently quite slow.  Give it the time it
 | |
| needs to load and save the checkpoint file.  When interrupted while
 | |
| writing the checkpoint file, the old checkpoint file is not
 | |
| overwritten, but all work done in the current run is lost.
 | |
| 
 | |
| Miscellaneous:
 | |
| 
 | |
| - You may find the (Tk-based) GUI version easier to use.  See wcgui.py.
 | |
| 
 | |
| - Webchecker honors the "robots.txt" convention.  Thanks to Skip
 | |
| Montanaro for his robotparser.py module (included in this directory)!
 | |
| The agent name is hardwired to "webchecker".  URLs that are disallowed
 | |
| by the robots.txt file are reported as external URLs.
 | |
| 
 | |
| - Because the SGML parser is a bit slow, very large SGML files are
 | |
| skipped.  The size limit can be set with the -m option.
 | |
| 
 | |
| - When the server or protocol does not tell us a file's type, we guess
 | |
| it based on the URL's suffix.  The mimetypes.py module (also in this
 | |
| directory) has a built-in table mapping most currently known suffixes,
 | |
| and in addition attempts to read the mime.types configuration files in
 | |
| the default locations of Netscape and the NCSA HTTP daemon.
 | |
| 
 | |
| - We follow links indicated by <A>, <FRAME> and <IMG> tags.  We also
 | |
| honor the <BASE> tag.
 | |
| 
 | |
| - (SLB 2/24/99) We now check internal NAME anchor links, as well
 | |
| as toplevel links.
 | |
| 
 | |
| - Checking external links is now done by default; use -x to *disable*
 | |
| this feature.  External links are now checked during normal
 | |
| processing.  (XXX The status of a checked link could be categorized
 | |
| better.  Later...)
 | |
| 
 | |
| - (SLB 2/24/99) If external links are not checked, you can use the -t
 | |
| flag to provide specific overrides to -x.
 | |
| 
 | |
| Usage: webchecker.py [option] ... [rooturl] ...
 | |
| 
 | |
| Options:
 | |
| 
 | |
| -R        -- restart from checkpoint file
 | |
| -d file   -- checkpoint filename (default %(DUMPFILE)s)
 | |
| -m bytes  -- skip HTML pages larger than this size (default %(MAXPAGE)d)
 | |
| -n        -- reports only, no checking (use with -R)
 | |
| -q        -- quiet operation (also suppresses external links report)
 | |
| -r number -- number of links processed per round (default %(ROUNDSIZE)d)
 | |
| -t root   -- specify root dir which should be treated as internal (can repeat)
 | |
| -v        -- verbose operation; repeating -v will increase verbosity
 | |
| -x        -- don't check external links (these are often slow to check)
 | |
| 
 | |
| Arguments:
 | |
| 
 | |
| rooturl   -- URL to start checking
 | |
|              (default %(DEFROOT)s)
 | |
| 
 | |
| """
 | |
| 
 | |
| 
 | |
| __version__ = "$Revision$"
 | |
| 
 | |
| 
 | |
| import sys
 | |
| import os
 | |
| from types import *
 | |
| import string
 | |
| import StringIO
 | |
| import getopt
 | |
| import pickle
 | |
| 
 | |
| import urllib
 | |
| import urlparse
 | |
| import sgmllib
 | |
| 
 | |
| import mimetypes
 | |
| import robotparser
 | |
| 
 | |
| # Extract real version number if necessary
 | |
| if __version__[0] == '$':
 | |
|     _v = string.split(__version__)
 | |
|     if len(_v) == 3:
 | |
|         __version__ = _v[1]
 | |
| 
 | |
| 
 | |
| # Tunable parameters
 | |
| DEFROOT = "file:/usr/local/etc/httpd/htdocs/"   # Default root URL
 | |
| CHECKEXT = 1                            # Check external references (1 deep)
 | |
| VERBOSE = 1                             # Verbosity level (0-3)
 | |
| MAXPAGE = 150000                        # Ignore files bigger than this
 | |
| ROUNDSIZE = 50                          # Number of links processed per round
 | |
| DUMPFILE = "@webchecker.pickle"         # Pickled checkpoint
 | |
| AGENTNAME = "webchecker"                # Agent name for robots.txt parser
 | |
| 
 | |
| 
 | |
| # Global variables
 | |
| 
 | |
| 
 | |
| def main():
 | |
|     checkext = CHECKEXT
 | |
|     verbose = VERBOSE
 | |
|     maxpage = MAXPAGE
 | |
|     roundsize = ROUNDSIZE
 | |
|     dumpfile = DUMPFILE
 | |
|     restart = 0
 | |
|     norun = 0
 | |
| 
 | |
|     try:
 | |
| 
 | |
| 	# Begin SLB 2/24/99: Added -t option here.
 | |
|         opts, args = getopt.getopt(sys.argv[1:], 'Rd:m:nqr:t:vx')
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     except getopt.error, msg:
 | |
|         sys.stdout = sys.stderr
 | |
|         print msg
 | |
|         print __doc__%globals()
 | |
|         sys.exit(2)
 | |
| 
 | |
|     # Begin SLB 2/24/99: Added extra_roots variable to
 | |
|     # collect extra roots.
 | |
|     extra_roots = []
 | |
|     # End SLB 2/24/99
 | |
| 
 | |
|     for o, a in opts:
 | |
|         if o == '-R':
 | |
|             restart = 1
 | |
|         if o == '-d':
 | |
|             dumpfile = a
 | |
|         if o == '-m':
 | |
|             maxpage = string.atoi(a)
 | |
|         if o == '-n':
 | |
|             norun = 1
 | |
|         if o == '-q':
 | |
|             verbose = 0
 | |
|         if o == '-r':
 | |
|             roundsize = string.atoi(a)
 | |
| 
 | |
| 	# Begin SLB 2/24/99: Added processing for
 | |
| 	# -t option.
 | |
| 	if o == '-t':
 | |
| 	    extra_roots.append(a)
 | |
| 	# End SLB 2/24/99
 | |
| 	
 | |
|         if o == '-v':
 | |
|             verbose = verbose + 1
 | |
|         if o == '-x':
 | |
|             checkext = not checkext
 | |
| 
 | |
|     if verbose > 0:
 | |
|         print AGENTNAME, "version", __version__
 | |
| 
 | |
|     if restart:
 | |
|         c = load_pickle(dumpfile=dumpfile, verbose=verbose)
 | |
|     else:
 | |
|         c = Checker()
 | |
| 
 | |
|     c.setflags(checkext=checkext, verbose=verbose,
 | |
|                maxpage=maxpage, roundsize=roundsize)
 | |
| 
 | |
|     if not restart and not args:
 | |
|         args.append(DEFROOT)
 | |
| 
 | |
|     for arg in args:
 | |
|         c.addroot(arg)
 | |
| 
 | |
|     # Begin SLB 2/24/99. The -t flag is only needed if external
 | |
|     # links are not to be checked. So -t values are ignored unless
 | |
|     # -x was specified.
 | |
|     if not checkext:
 | |
| 	for root in extra_roots:
 | |
| 	    # Make sure it's terminated by a slash,
 | |
| 	    # so that addroot doesn't discard the last
 | |
| 	    # directory component.
 | |
| 	    if root[-1] != "/":
 | |
| 		root = root + "/"
 | |
| 	    c.addroot(root)
 | |
|     # End SLB 2/24/99
 | |
| 
 | |
|     try:
 | |
| 
 | |
|         if not norun:
 | |
|             try:
 | |
|                 c.run()
 | |
|             except KeyboardInterrupt:
 | |
|                 if verbose > 0:
 | |
|                     print "[run interrupted]"
 | |
| 
 | |
|         try:
 | |
|             c.report()
 | |
|         except KeyboardInterrupt:
 | |
|             if verbose > 0:
 | |
|                 print "[report interrupted]"
 | |
| 
 | |
|     finally:
 | |
|         if c.save_pickle(dumpfile):
 | |
|             if dumpfile == DUMPFILE:
 | |
|                 print "Use ``%s -R'' to restart." % sys.argv[0]
 | |
|             else:
 | |
|                 print "Use ``%s -R -d %s'' to restart." % (sys.argv[0],
 | |
|                                                            dumpfile)
 | |
| 
 | |
| 
 | |
| def load_pickle(dumpfile=DUMPFILE, verbose=VERBOSE):
 | |
|     if verbose > 0:
 | |
|         print "Loading checkpoint from %s ..." % dumpfile
 | |
|     f = open(dumpfile, "rb")
 | |
|     c = pickle.load(f)
 | |
|     f.close()
 | |
|     if verbose > 0:
 | |
|         print "Done."
 | |
|         print "Root:", string.join(c.roots, "\n      ")
 | |
|     return c
 | |
| 
 | |
| 
 | |
| class Checker:
 | |
| 
 | |
|     checkext = CHECKEXT
 | |
|     verbose = VERBOSE
 | |
|     maxpage = MAXPAGE
 | |
|     roundsize = ROUNDSIZE
 | |
| 
 | |
|     validflags = tuple(dir())
 | |
| 
 | |
|     def __init__(self):
 | |
|         self.reset()
 | |
| 
 | |
|     def setflags(self, **kw):
 | |
|         for key in kw.keys():
 | |
|             if key not in self.validflags:
 | |
|                 raise NameError, "invalid keyword argument: %s" % str(key)
 | |
|         for key, value in kw.items():
 | |
|             setattr(self, key, value)
 | |
| 
 | |
|     def reset(self):
 | |
|         self.roots = []
 | |
|         self.todo = {}
 | |
|         self.done = {}
 | |
|         self.bad = {}
 | |
| 
 | |
| 	# Begin SLB 2/24/99: Added a name table, so that the
 | |
| 	# name URLs can be checked. Also serves as an implicit
 | |
| 	# cache for which URLs are done.
 | |
| 	self.name_table = {}
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|         self.round = 0
 | |
|         # The following are not pickled:
 | |
|         self.robots = {}
 | |
|         self.errors = {}
 | |
|         self.urlopener = MyURLopener()
 | |
|         self.changed = 0
 | |
|         
 | |
|     def note(self, level, format, *args):
 | |
|         if self.verbose > level:
 | |
|             if args:
 | |
|                 format = format%args
 | |
|             self.message(format)
 | |
|     
 | |
|     def message(self, format, *args):
 | |
|         if args:
 | |
|             format = format%args
 | |
|         print format 
 | |
| 
 | |
|     def __getstate__(self):
 | |
|         return (self.roots, self.todo, self.done, self.bad, self.round)
 | |
| 
 | |
|     def __setstate__(self, state):
 | |
|         self.reset()
 | |
|         (self.roots, self.todo, self.done, self.bad, self.round) = state
 | |
|         for root in self.roots:
 | |
|             self.addrobot(root)
 | |
|         for url in self.bad.keys():
 | |
|             self.markerror(url)
 | |
| 
 | |
|     def addroot(self, root):
 | |
|         if root not in self.roots:
 | |
|             troot = root
 | |
|             scheme, netloc, path, params, query, fragment = \
 | |
|                     urlparse.urlparse(root)
 | |
|             i = string.rfind(path, "/") + 1
 | |
|             if 0 < i < len(path):
 | |
|                 path = path[:i]
 | |
|                 troot = urlparse.urlunparse((scheme, netloc, path,
 | |
|                                              params, query, fragment))
 | |
|             self.roots.append(troot)
 | |
|             self.addrobot(root)
 | |
| 	    # Begin SLB 2/24/99: Modified this call to respect
 | |
| 	    # the fact that the "done" and "todo" dictionaries
 | |
| 	    # are now (URL, fragment) pairs
 | |
| 	    self.newlink((root, ""), ("<root>", root))
 | |
| 	    # End SLB 2/24/99
 | |
| 
 | |
|     def addrobot(self, root):
 | |
|         root = urlparse.urljoin(root, "/")
 | |
|         if self.robots.has_key(root): return
 | |
|         url = urlparse.urljoin(root, "/robots.txt")
 | |
|         self.robots[root] = rp = robotparser.RobotFileParser()
 | |
|         self.note(2, "Parsing %s", url)
 | |
|         rp.debug = self.verbose > 3
 | |
|         rp.set_url(url)
 | |
|         try:
 | |
|             rp.read()
 | |
|         except IOError, msg:
 | |
|             self.note(1, "I/O error parsing %s: %s", url, msg)
 | |
| 
 | |
|     def run(self):
 | |
|         while self.todo:
 | |
|             self.round = self.round + 1
 | |
|             self.note(0, "\nRound %d (%s)\n", self.round, self.status())
 | |
|             urls = self.todo.keys()
 | |
|             urls.sort()
 | |
|             del urls[self.roundsize:]
 | |
|             for url in urls:
 | |
|                 self.dopage(url)
 | |
| 
 | |
|     def status(self):
 | |
|         return "%d total, %d to do, %d done, %d bad" % (
 | |
|             len(self.todo)+len(self.done),
 | |
|             len(self.todo), len(self.done),
 | |
|             len(self.bad))
 | |
| 
 | |
|     def report(self):
 | |
|         self.message("")
 | |
|         if not self.todo: s = "Final"
 | |
|         else: s = "Interim"
 | |
|         self.message("%s Report (%s)", s, self.status())
 | |
|         self.report_errors()
 | |
| 
 | |
|     def report_errors(self):
 | |
|         if not self.bad:
 | |
|             self.message("\nNo errors")
 | |
|             return
 | |
|         self.message("\nError Report:")
 | |
|         sources = self.errors.keys()
 | |
|         sources.sort()
 | |
|         for source in sources:
 | |
|             triples = self.errors[source]
 | |
|             self.message("")
 | |
|             if len(triples) > 1:
 | |
|                 self.message("%d Errors in %s", len(triples), source)
 | |
|             else:
 | |
|                 self.message("Error in %s", source)
 | |
| 	    # Begin SLB 2/24/99: Modified this loop to
 | |
| 	    # call self.format_url() instead of referring
 | |
| 	    # to the URL directly, since the URLs in these
 | |
| 	    # triples is now a (URL, fragment) pair. The value
 | |
| 	    # of the "source" variable comes from the list of
 | |
| 	    # origins, and is a URL, not a pair.
 | |
|             for url, rawlink, msg in triples:		
 | |
|                 if rawlink != self.format_url(url): s = " (%s)" % rawlink
 | |
|                 else: s = ""
 | |
|                 self.message("  HREF %s%s\n    msg %s",
 | |
| 			     self.format_url(url), s, msg)
 | |
| 	    # End SLB 2/24/99.
 | |
| 
 | |
|     def dopage(self, url_pair):
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: Substantial modification to reflect the change
 | |
| 	# to dictionary entries being (URL, fragment) pairs. All
 | |
| 	# printing of URLs uses format_url(); argument changed to
 | |
| 	# url_pair for clarity.
 | |
|         if self.verbose > 1:
 | |
|             if self.verbose > 2:
 | |
|                 self.show("Check ", self.format_url(url_pair),
 | |
| 			  "  from", self.todo[url_pair])
 | |
|             else:
 | |
|                 self.message("Check %s", self.format_url(url_pair))
 | |
|         page = self.getpage(url_pair)
 | |
|         if page:
 | |
| 	    url, local_fragment = url_pair
 | |
| 	    # Store the page which corresponds to this URL.
 | |
| 	    self.name_table[url] = page
 | |
| 	    # If there is a fragment in this url_pair, and it's not
 | |
| 	    # in the list of names for the page, call setbad(), since
 | |
| 	    # it's a missing anchor.
 | |
| 	    if local_fragment and local_fragment not in page.getnames():
 | |
| 		self.setbad(url_pair, ("Missing name anchor `%s'" % local_fragment))
 | |
| 	    for info in page.getlinkinfos():
 | |
| 		# getlinkinfos() now returns the fragment as well,
 | |
| 		# and we store that fragment here in the "todo" dictionary.
 | |
|                 link, rawlink, fragment = info
 | |
| 		# However, we don't want the fragment as the origin, since
 | |
| 		# the origin is logically a page.
 | |
|                 origin = url, rawlink
 | |
|                 self.newlink((link, fragment), origin)
 | |
| 	else:
 | |
| 	    # If no page has been created yet, we want to
 | |
| 	    # record that fact.
 | |
| 	    self.name_table[url_pair[0]] = None
 | |
|         self.markdone(url_pair)
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     def newlink(self, url, origin):
 | |
|         if self.done.has_key(url):
 | |
|             self.newdonelink(url, origin)
 | |
|         else:
 | |
|             self.newtodolink(url, origin)
 | |
| 
 | |
|     def newdonelink(self, url, origin):
 | |
|         self.done[url].append(origin)
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: changed reference to URL
 | |
| 	# to call self.format_url(), since the URL here
 | |
| 	# is now a (URL, fragment) pair.
 | |
|         self.note(3, "  Done link %s", self.format_url(url))
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     def newtodolink(self, url, origin):
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: changed reference to URL
 | |
| 	# to call self.format_url(), since the URL here
 | |
| 	# is now a (URL, fragment) pair.
 | |
|         if self.todo.has_key(url):
 | |
|             self.todo[url].append(origin)
 | |
|             self.note(3, "  Seen todo link %s", self.format_url(url))
 | |
|         else:
 | |
|             self.todo[url] = [origin]
 | |
|             self.note(3, "  New todo link %s", self.format_url(url))
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     # Begin SLB 2/24/99: Added method format_url().
 | |
|     def format_url(self, url):	
 | |
| 	link, fragment = url
 | |
| 	if fragment: return link + "#" + fragment
 | |
| 	else: return link
 | |
|     # End SLB 2/24/99
 | |
| 
 | |
|     def markdone(self, url):
 | |
|         self.done[url] = self.todo[url]
 | |
|         del self.todo[url]
 | |
|         self.changed = 1
 | |
| 
 | |
|     def inroots(self, url):
 | |
|         for root in self.roots:
 | |
|             if url[:len(root)] == root:
 | |
|                 return self.isallowed(root, url)
 | |
|         return 0
 | |
|     
 | |
|     def isallowed(self, root, url):
 | |
|         root = urlparse.urljoin(root, "/")
 | |
|         return self.robots[root].can_fetch(AGENTNAME, url)
 | |
| 
 | |
|     def getpage(self, url_pair):
 | |
| 
 | |
| 	# Begin SLB 2/24/99: changed incoming argument name to 
 | |
| 	# url_pair, since it's now a (URL, fragment) pair.
 | |
| 	# The page may have been cached in the name_table variable.
 | |
| 	url, fragment = url_pair
 | |
| 	if self.name_table.has_key(url):
 | |
| 	    return self.name_table[url]
 | |
| 	# End SLB 2/24/99
 | |
| 	
 | |
|         if url[:7] == 'mailto:' or url[:5] == 'news:':
 | |
|             self.note(1, " Not checking mailto/news URL")
 | |
|             return None
 | |
|         isint = self.inroots(url)
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: Changed calls to openpage and readhtml
 | |
| 	# in order to ensure that openpage gets the URL pair to
 | |
| 	# print out its error message and record the error pair
 | |
| 	# correctly.
 | |
|         if not isint:
 | |
|             if not self.checkext:
 | |
|                 self.note(1, " Not checking ext link")
 | |
|                 return None
 | |
|             f = self.openpage(url_pair)
 | |
|             if f:
 | |
|                 self.safeclose(f)
 | |
|             return None
 | |
|         text, nurl = self.readhtml(url_pair)
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|         if nurl != url:
 | |
|             self.note(1, " Redirected to %s", nurl)
 | |
|             url = nurl
 | |
|         if text:
 | |
|             return Page(text, url, maxpage=self.maxpage, checker=self)
 | |
| 
 | |
|     # Begin SLB 2/24/99: Modified these next three functions
 | |
|     # to take (URL, fragment) pairs as arguments, so that openpage()
 | |
|     # receives the appropriate tuple to record error messages.
 | |
|     def readhtml(self, url_pair):
 | |
| 	url, fragment = url_pair
 | |
|         text = None
 | |
|         f, url = self.openhtml(url_pair)
 | |
|         if f:
 | |
|             text = f.read()
 | |
|             f.close()
 | |
|         return text, url
 | |
| 
 | |
|     def openhtml(self, url_pair):
 | |
| 	url, fragment = url_pair
 | |
|         f = self.openpage(url_pair)
 | |
|         if f:
 | |
|             url = f.geturl()
 | |
|             info = f.info()
 | |
|             if not self.checkforhtml(info, url):
 | |
|                 self.safeclose(f)
 | |
|                 f = None
 | |
|         return f, url
 | |
| 
 | |
|     def openpage(self, url_pair):
 | |
| 	url, fragment = url_pair
 | |
|         try:
 | |
|             return self.urlopener.open(url)
 | |
|         except IOError, msg:
 | |
|             msg = self.sanitize(msg)
 | |
|             self.note(0, "Error %s", msg)
 | |
|             if self.verbose > 0:
 | |
|                 self.show(" HREF ", url, "  from", self.todo[url_pair])
 | |
|             self.setbad(url_pair, msg)
 | |
|             return None
 | |
|     # End SLB 2/24/99
 | |
| 
 | |
|     def checkforhtml(self, info, url):
 | |
|         if info.has_key('content-type'):
 | |
|             ctype = string.lower(info['content-type'])
 | |
|         else:
 | |
|             if url[-1:] == "/":
 | |
|                 return 1
 | |
|             ctype, encoding = mimetypes.guess_type(url)
 | |
|         if ctype == 'text/html':
 | |
|             return 1
 | |
|         else:
 | |
|             self.note(1, " Not HTML, mime type %s", ctype)
 | |
|             return 0
 | |
| 
 | |
|     def setgood(self, url):
 | |
|         if self.bad.has_key(url):
 | |
|             del self.bad[url]
 | |
|             self.changed = 1
 | |
|             self.note(0, "(Clear previously seen error)")
 | |
| 
 | |
|     def setbad(self, url, msg):
 | |
|         if self.bad.has_key(url) and self.bad[url] == msg:
 | |
|             self.note(0, "(Seen this error before)")
 | |
|             return
 | |
|         self.bad[url] = msg
 | |
|         self.changed = 1
 | |
|         self.markerror(url)
 | |
|         
 | |
|     def markerror(self, url):
 | |
|         try:
 | |
|             origins = self.todo[url]
 | |
|         except KeyError:
 | |
|             origins = self.done[url]
 | |
|         for source, rawlink in origins:
 | |
|             triple = url, rawlink, self.bad[url]
 | |
|             self.seterror(source, triple)
 | |
| 
 | |
|     def seterror(self, url, triple):
 | |
|         try:
 | |
| 	    
 | |
| 	    # Begin SLB 2/24/99: Because of the way the
 | |
| 	    # URLs are now processed, I need to check to make
 | |
| 	    # sure the URL hasn't been entered in the error list.
 | |
| 	    # The first element of the triple here is a (URL, fragment)
 | |
| 	    # pair, but the URL key is not, since it's from the
 | |
| 	    # list of origins.
 | |
| 	    if triple not in self.errors[url]:
 | |
| 		self.errors[url].append(triple)
 | |
| 	    # End SLB 2/24/99
 | |
| 	    
 | |
|         except KeyError:
 | |
|             self.errors[url] = [triple]
 | |
| 
 | |
|     # The following used to be toplevel functions; they have been
 | |
|     # changed into methods so they can be overridden in subclasses.
 | |
| 
 | |
|     def show(self, p1, link, p2, origins):
 | |
|         self.message("%s %s", p1, link)
 | |
|         i = 0
 | |
|         for source, rawlink in origins:
 | |
|             i = i+1
 | |
|             if i == 2:
 | |
|                 p2 = ' '*len(p2)
 | |
|             if rawlink != link: s = " (%s)" % rawlink
 | |
|             else: s = ""
 | |
|             self.message("%s %s%s", p2, source, s)
 | |
| 
 | |
|     def sanitize(self, msg):
 | |
|         if isinstance(IOError, ClassType) and isinstance(msg, IOError):
 | |
|             # Do the other branch recursively
 | |
|             msg.args = self.sanitize(msg.args)
 | |
|         elif isinstance(msg, TupleType):
 | |
|             if len(msg) >= 4 and msg[0] == 'http error' and \
 | |
|                isinstance(msg[3], InstanceType):
 | |
|                 # Remove the Message instance -- it may contain
 | |
|                 # a file object which prevents pickling.
 | |
|                 msg = msg[:3] + msg[4:]
 | |
|         return msg
 | |
| 
 | |
|     def safeclose(self, f):
 | |
|         try:
 | |
|             url = f.geturl()
 | |
|         except AttributeError:
 | |
|             pass
 | |
|         else:
 | |
|             if url[:4] == 'ftp:' or url[:7] == 'file://':
 | |
|                 # Apparently ftp connections don't like to be closed
 | |
|                 # prematurely...
 | |
|                 text = f.read()
 | |
|         f.close()
 | |
| 
 | |
|     def save_pickle(self, dumpfile=DUMPFILE):
 | |
|         if not self.changed:
 | |
|             self.note(0, "\nNo need to save checkpoint")
 | |
|         elif not dumpfile:
 | |
|             self.note(0, "No dumpfile, won't save checkpoint")
 | |
|         else:
 | |
|             self.note(0, "\nSaving checkpoint to %s ...", dumpfile)
 | |
|             newfile = dumpfile + ".new"
 | |
|             f = open(newfile, "wb")
 | |
|             pickle.dump(self, f)
 | |
|             f.close()
 | |
|             try:
 | |
|                 os.unlink(dumpfile)
 | |
|             except os.error:
 | |
|                 pass
 | |
|             os.rename(newfile, dumpfile)
 | |
|             self.note(0, "Done.")
 | |
|             return 1
 | |
| 
 | |
| 
 | |
| class Page:
 | |
| 
 | |
|     def __init__(self, text, url, verbose=VERBOSE, maxpage=MAXPAGE, checker=None):
 | |
|         self.text = text
 | |
|         self.url = url
 | |
|         self.verbose = verbose
 | |
|         self.maxpage = maxpage
 | |
|         self.checker = checker
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: Moved the parsing of the page into
 | |
| 	# the __init__() routine in order to initialize the list of
 | |
| 	# names the file contains. Stored the parser in an instance
 | |
| 	# variable. Passed the URL to MyHTMLParser().
 | |
| 	size = len(self.text)
 | |
|         if size > self.maxpage:
 | |
|             self.note(0, "Skip huge file %s (%.0f Kbytes)", self.url, (size*0.001))
 | |
|             self.parser = None
 | |
| 	    return
 | |
|         self.checker.note(2, "  Parsing %s (%d bytes)", self.url, size)
 | |
|         self.parser = MyHTMLParser(url, verbose=self.verbose,
 | |
| 				   checker=self.checker)
 | |
|         self.parser.feed(self.text)
 | |
|         self.parser.close()
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     def note(self, level, msg, *args):
 | |
|         if self.checker:
 | |
|             apply(self.checker.note, (level, msg) + args)
 | |
|         else:
 | |
|             if self.verbose >= level:
 | |
|                 if args:
 | |
|                     msg = msg%args
 | |
|                 print msg
 | |
| 
 | |
|     # Begin SLB 2/24/99: Added method to retrieve names.
 | |
|     def getnames(self):
 | |
| 	return self.parser.names
 | |
|     # End SLB 2/24/99
 | |
| 
 | |
|     def getlinkinfos(self):
 | |
| 
 | |
| 	# Begin SLB 2/24/99: Moved file reading to __init__() routine.
 | |
| 	# Stored parser in local variable to indicate success of
 | |
| 	# parsing.
 | |
| 	
 | |
| 	# If no parser was stored, fail.
 | |
| 	if not self.parser: return []
 | |
| 	
 | |
|         rawlinks = self.parser.getlinks()
 | |
|         base = urlparse.urljoin(self.url, self.parser.getbase() or "")
 | |
|         infos = []
 | |
|         for rawlink in rawlinks:
 | |
|             t = urlparse.urlparse(rawlink)
 | |
| 	    # DON'T DISCARD THE FRAGMENT! Instead, include
 | |
| 	    # it in the tuples which are returned. See Checker.dopage().
 | |
| 	    fragment = t[-1]
 | |
|             t = t[:-1] + ('',)
 | |
|             rawlink = urlparse.urlunparse(t)
 | |
|             link = urlparse.urljoin(base, rawlink)
 | |
|             infos.append((link, rawlink, fragment))	
 | |
|         # End SLB 2/24/99
 | |
| 	
 | |
|         return infos
 | |
| 
 | |
| 
 | |
| class MyStringIO(StringIO.StringIO):
 | |
| 
 | |
|     def __init__(self, url, info):
 | |
|         self.__url = url
 | |
|         self.__info = info
 | |
|         StringIO.StringIO.__init__(self)
 | |
| 
 | |
|     def info(self):
 | |
|         return self.__info
 | |
| 
 | |
|     def geturl(self):
 | |
|         return self.__url
 | |
| 
 | |
| 
 | |
| class MyURLopener(urllib.FancyURLopener):
 | |
| 
 | |
|     http_error_default = urllib.URLopener.http_error_default
 | |
| 
 | |
|     def __init__(*args):
 | |
|         self = args[0]
 | |
|         apply(urllib.FancyURLopener.__init__, args)
 | |
|         self.addheaders = [
 | |
|             ('User-agent', 'Python-webchecker/%s' % __version__),
 | |
|             ]
 | |
| 
 | |
|     def http_error_401(self, url, fp, errcode, errmsg, headers):
 | |
|         return None
 | |
| 
 | |
|     def open_file(self, url):
 | |
|         path = urllib.url2pathname(urllib.unquote(url))
 | |
|         if path[-1] != os.sep:
 | |
|             url = url + '/'
 | |
|         if os.path.isdir(path):
 | |
|             indexpath = os.path.join(path, "index.html")
 | |
|             if os.path.exists(indexpath):
 | |
|                 return self.open_file(url + "index.html")
 | |
|             try:
 | |
|                 names = os.listdir(path)
 | |
|             except os.error, msg:
 | |
|                 raise IOError, msg, sys.exc_traceback
 | |
|             names.sort()
 | |
|             s = MyStringIO("file:"+url, {'content-type': 'text/html'})
 | |
|             s.write('<BASE HREF="file:%s">\n' %
 | |
|                     urllib.quote(os.path.join(path, "")))
 | |
|             for name in names:
 | |
|                 q = urllib.quote(name)
 | |
|                 s.write('<A HREF="%s">%s</A>\n' % (q, q))
 | |
|             s.seek(0)
 | |
|             return s
 | |
|         return urllib.FancyURLopener.open_file(self, path)
 | |
| 
 | |
| 
 | |
| class MyHTMLParser(sgmllib.SGMLParser):
 | |
| 
 | |
|     def __init__(self, url, verbose=VERBOSE, checker=None):
 | |
|         self.myverbose = verbose # now unused
 | |
|         self.checker = checker
 | |
|         self.base = None
 | |
|         self.links = {}
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: Added names instance variable.
 | |
| 	# Modified arglist to take the URL as an argument.
 | |
| 	self.names = []
 | |
| 	self.url = url
 | |
| 	# End SLB 2/24/99
 | |
| 	
 | |
|         sgmllib.SGMLParser.__init__(self)
 | |
| 
 | |
|     def start_a(self, attributes):
 | |
|         self.link_attr(attributes, 'href')
 | |
| 	
 | |
| 	# Begin SLB 2/24/99: We must rescue the NAME
 | |
| 	# attributes from the anchor, in order to
 | |
| 	# cache the internal anchors which are made
 | |
| 	# available in the page.
 | |
| 	for name, value in attributes:
 | |
| 	    if name == "name":
 | |
| 		if value in self.names:
 | |
| 		    self.checker.message("WARNING: duplicate name %s in %s",
 | |
| 					 value, self.url)
 | |
| 		else: self.names.append(value)
 | |
| 		break
 | |
| 	# End SLB 2/24/99
 | |
| 
 | |
|     def end_a(self): pass
 | |
| 
 | |
|     def do_area(self, attributes):
 | |
|         self.link_attr(attributes, 'href')
 | |
| 
 | |
|     def do_img(self, attributes):
 | |
|         self.link_attr(attributes, 'src', 'lowsrc')
 | |
| 
 | |
|     def do_frame(self, attributes):
 | |
|         self.link_attr(attributes, 'src')
 | |
| 
 | |
|     def link_attr(self, attributes, *args):
 | |
|         for name, value in attributes:
 | |
|             if name in args:
 | |
|                 if value: value = string.strip(value)
 | |
|                 if value: self.links[value] = None
 | |
| 
 | |
|     def do_base(self, attributes):
 | |
|         for name, value in attributes:
 | |
|             if name == 'href':
 | |
|                 if value: value = string.strip(value)
 | |
|                 if value:
 | |
|                     if self.checker:
 | |
|                         self.checker.note(1, "  Base %s", value)
 | |
|                     self.base = value
 | |
| 
 | |
|     def getlinks(self):
 | |
|         return self.links.keys()
 | |
| 
 | |
|     def getbase(self):
 | |
|         return self.base
 | |
| 
 | |
| 
 | |
| if __name__ == '__main__':
 | |
|     main()
 |