SF patch #911431: robot.txt must be robots.txt
(Contributed by George Yoshida.)
diff --git a/Lib/robotparser.py b/Lib/robotparser.py
index e2af545..6b23188 100644
--- a/Lib/robotparser.py
+++ b/Lib/robotparser.py
@@ -83,7 +83,7 @@
             self.entries.append(entry)
 
     def parse(self, lines):
-        """parse the input lines from a robot.txt file.
+        """parse the input lines from a robots.txt file.
            We allow that a user-agent: line is not preceded by
            one or more blank lines."""
         state = 0
@@ -148,7 +148,7 @@
 
     def can_fetch(self, useragent, url):
         """using the parsed robots.txt decide if useragent can fetch url"""
-        _debug("Checking robot.txt allowance for:\n  user agent: %s\n  url: %s" %
+        _debug("Checking robots.txt allowance for:\n  user agent: %s\n  url: %s" %
                (useragent, url))
         if self.disallow_all:
             return False
diff --git a/Misc/cheatsheet b/Misc/cheatsheet
index 0c16ddb..487949a 100644
--- a/Misc/cheatsheet
+++ b/Misc/cheatsheet
@@ -1962,7 +1962,7 @@
 rexec            Restricted execution facilities ("safe" exec, eval, etc).
 rfc822           RFC-822 message manipulation class.
 rlcompleter      Word completion for GNU readline 2.0.
-robotparser      Parse robot.txt files, useful for web spiders.
+robotparser      Parse robots.txt files, useful for web spiders.
 sched            A generally useful event scheduler class.
 sets             Module for a set datatype.
 sgmllib          A parser for SGML.