reformat some documentation of classes so methods and attributes are under the class directive
diff --git a/Doc/library/robotparser.rst b/Doc/library/robotparser.rst
index 6cc7df8..b3a9a60 100644
--- a/Doc/library/robotparser.rst
+++ b/Doc/library/robotparser.rst
@@ -25,35 +25,35 @@
single :file:`robots.txt` file.
- .. method:: RobotFileParser.set_url(url)
+ .. method:: set_url(url)
Sets the URL referring to a :file:`robots.txt` file.
- .. method:: RobotFileParser.read()
+ .. method:: read()
Reads the :file:`robots.txt` URL and feeds it to the parser.
- .. method:: RobotFileParser.parse(lines)
+ .. method:: parse(lines)
Parses the lines argument.
- .. method:: RobotFileParser.can_fetch(useragent, url)
+ .. method:: can_fetch(useragent, url)
Returns ``True`` if the *useragent* is allowed to fetch the *url* according to
the rules contained in the parsed :file:`robots.txt` file.
- .. method:: RobotFileParser.mtime()
+ .. method:: mtime()
Returns the time the ``robots.txt`` file was last fetched. This is useful for
long-running web spiders that need to check for new ``robots.txt`` files
periodically.
- .. method:: RobotFileParser.modified()
+ .. method:: modified()
Sets the time the ``robots.txt`` file was last fetched to the current time.