| """HTTP cookie handling for web clients. |
| |
| This module has (now fairly distant) origins in Gisle Aas' Perl module |
| HTTP::Cookies, from the libwww-perl library. |
| |
| Docstrings, comments and debug strings in this code refer to the |
| attributes of the HTTP cookie system as cookie-attributes, to distinguish |
| them clearly from Python attributes. |
| |
| Class diagram (note that the classes which do not derive from |
| FileCookieJar are not distributed with the Python standard library, but |
| are available from http://wwwsearch.sf.net/): |
| |
| CookieJar____ |
| / \ \ |
| FileCookieJar \ \ |
| / | \ \ \ |
| MozillaCookieJar | LWPCookieJar \ \ |
| | | \ |
| | ---MSIEBase | \ |
| | / | | \ |
| | / MSIEDBCookieJar BSDDBCookieJar |
| |/ |
| MSIECookieJar |
| |
| """ |
| |
| import sys, re, urlparse, copy, time, urllib, logging |
| from types import StringTypes |
| try: |
| import threading as _threading |
| except ImportError: |
| import dummy_threading as _threading |
| import httplib # only for the default HTTP port |
| from calendar import timegm |
| |
| debug = logging.getLogger("cookielib").debug |
| |
| DEFAULT_HTTP_PORT = str(httplib.HTTP_PORT) |
| MISSING_FILENAME_TEXT = ("a filename was not supplied (nor was the CookieJar " |
| "instance initialised with one)") |
| |
| def reraise_unmasked_exceptions(unmasked=()): |
| # There are a few catch-all except: statements in this module, for |
| # catching input that's bad in unexpected ways. |
| # This function re-raises some exceptions we don't want to trap. |
| unmasked = unmasked + (KeyboardInterrupt, SystemExit, MemoryError) |
| etype = sys.exc_info()[0] |
| if issubclass(etype, unmasked): |
| raise |
| # swallowed an exception |
| import warnings, traceback, StringIO |
| f = StringIO.StringIO() |
| traceback.print_exc(None, f) |
| msg = f.getvalue() |
| warnings.warn("cookielib bug!\n%s" % msg, stacklevel=2) |
| |
| |
| # Date/time conversion |
| # ----------------------------------------------------------------------------- |
| |
| EPOCH_YEAR = 1970 |
| def _timegm(tt): |
| year, month, mday, hour, min, sec = tt[:6] |
| if ((year >= EPOCH_YEAR) and (1 <= month <= 12) and (1 <= mday <= 31) and |
| (0 <= hour <= 24) and (0 <= min <= 59) and (0 <= sec <= 61)): |
| return timegm(tt) |
| else: |
| return None |
| |
| DAYS = ["Mon", "Tue", "Wed", "Thu", "Fri", "Sat", "Sun"] |
| MONTHS = ["Jan", "Feb", "Mar", "Apr", "May", "Jun", |
| "Jul", "Aug", "Sep", "Oct", "Nov", "Dec"] |
| MONTHS_LOWER = [] |
| for month in MONTHS: MONTHS_LOWER.append(month.lower()) |
| |
| def time2isoz(t=None): |
| """Return a string representing time in seconds since epoch, t. |
| |
| If the function is called without an argument, it will use the current |
| time. |
| |
| The format of the returned string is like "YYYY-MM-DD hh:mm:ssZ", |
| representing Universal Time (UTC, aka GMT). An example of this format is: |
| |
| 1994-11-24 08:49:37Z |
| |
| """ |
| if t is None: t = time.time() |
| year, mon, mday, hour, min, sec = time.gmtime(t)[:6] |
| return "%04d-%02d-%02d %02d:%02d:%02dZ" % ( |
| year, mon, mday, hour, min, sec) |
| |
| def time2netscape(t=None): |
| """Return a string representing time in seconds since epoch, t. |
| |
| If the function is called without an argument, it will use the current |
| time. |
| |
| The format of the returned string is like this: |
| |
| Wed, DD-Mon-YYYY HH:MM:SS GMT |
| |
| """ |
| if t is None: t = time.time() |
| year, mon, mday, hour, min, sec, wday = time.gmtime(t)[:7] |
| return "%s %02d-%s-%04d %02d:%02d:%02d GMT" % ( |
| DAYS[wday], mday, MONTHS[mon-1], year, hour, min, sec) |
| |
| |
| UTC_ZONES = {"GMT": None, "UTC": None, "UT": None, "Z": None} |
| |
| TIMEZONE_RE = re.compile(r"^([-+])?(\d\d?):?(\d\d)?$") |
| def offset_from_tz_string(tz): |
| offset = None |
| if tz in UTC_ZONES: |
| offset = 0 |
| else: |
| m = TIMEZONE_RE.search(tz) |
| if m: |
| offset = 3600 * int(m.group(2)) |
| if m.group(3): |
| offset = offset + 60 * int(m.group(3)) |
| if m.group(1) == '-': |
| offset = -offset |
| return offset |
| |
| def _str2time(day, mon, yr, hr, min, sec, tz): |
| # translate month name to number |
| # month numbers start with 1 (January) |
| try: |
| mon = MONTHS_LOWER.index(mon.lower())+1 |
| except ValueError: |
| # maybe it's already a number |
| try: |
| imon = int(mon) |
| except ValueError: |
| return None |
| if 1 <= imon <= 12: |
| mon = imon |
| else: |
| return None |
| |
| # make sure clock elements are defined |
| if hr is None: hr = 0 |
| if min is None: min = 0 |
| if sec is None: sec = 0 |
| |
| yr = int(yr) |
| day = int(day) |
| hr = int(hr) |
| min = int(min) |
| sec = int(sec) |
| |
| if yr < 1000: |
| # find "obvious" year |
| cur_yr = time.localtime(time.time())[0] |
| m = cur_yr % 100 |
| tmp = yr |
| yr = yr + cur_yr - m |
| m = m - tmp |
| if abs(m) > 50: |
| if m > 0: yr = yr + 100 |
| else: yr = yr - 100 |
| |
| # convert UTC time tuple to seconds since epoch (not timezone-adjusted) |
| t = _timegm((yr, mon, day, hr, min, sec, tz)) |
| |
| if t is not None: |
| # adjust time using timezone string, to get absolute time since epoch |
| if tz is None: |
| tz = "UTC" |
| tz = tz.upper() |
| offset = offset_from_tz_string(tz) |
| if offset is None: |
| return None |
| t = t - offset |
| |
| return t |
| |
| STRICT_DATE_RE = re.compile( |
| r"^[SMTWF][a-z][a-z], (\d\d) ([JFMASOND][a-z][a-z]) " |
| "(\d\d\d\d) (\d\d):(\d\d):(\d\d) GMT$") |
| WEEKDAY_RE = re.compile( |
| r"^(?:Sun|Mon|Tue|Wed|Thu|Fri|Sat)[a-z]*,?\s*", re.I) |
| LOOSE_HTTP_DATE_RE = re.compile( |
| r"""^ |
| (\d\d?) # day |
| (?:\s+|[-\/]) |
| (\w+) # month |
| (?:\s+|[-\/]) |
| (\d+) # year |
| (?: |
| (?:\s+|:) # separator before clock |
| (\d\d?):(\d\d) # hour:min |
| (?::(\d\d))? # optional seconds |
| )? # optional clock |
| \s* |
| ([-+]?\d{2,4}|(?![APap][Mm]\b)[A-Za-z]+)? # timezone |
| \s* |
| (?:\(\w+\))? # ASCII representation of timezone in parens. |
| \s*$""", re.X) |
| def http2time(text): |
| """Returns time in seconds since epoch of time represented by a string. |
| |
| Return value is an integer. |
| |
| None is returned if the format of str is unrecognized, the time is outside |
| the representable range, or the timezone string is not recognized. If the |
| string contains no timezone, UTC is assumed. |
| |
| The timezone in the string may be numerical (like "-0800" or "+0100") or a |
| string timezone (like "UTC", "GMT", "BST" or "EST"). Currently, only the |
| timezone strings equivalent to UTC (zero offset) are known to the function. |
| |
| The function loosely parses the following formats: |
| |
| Wed, 09 Feb 1994 22:23:32 GMT -- HTTP format |
| Tuesday, 08-Feb-94 14:15:29 GMT -- old rfc850 HTTP format |
| Tuesday, 08-Feb-1994 14:15:29 GMT -- broken rfc850 HTTP format |
| 09 Feb 1994 22:23:32 GMT -- HTTP format (no weekday) |
| 08-Feb-94 14:15:29 GMT -- rfc850 format (no weekday) |
| 08-Feb-1994 14:15:29 GMT -- broken rfc850 format (no weekday) |
| |
| The parser ignores leading and trailing whitespace. The time may be |
| absent. |
| |
| If the year is given with only 2 digits, the function will select the |
| century that makes the year closest to the current date. |
| |
| """ |
| # fast exit for strictly conforming string |
| m = STRICT_DATE_RE.search(text) |
| if m: |
| g = m.groups() |
| mon = MONTHS_LOWER.index(g[1].lower()) + 1 |
| tt = (int(g[2]), mon, int(g[0]), |
| int(g[3]), int(g[4]), float(g[5])) |
| return _timegm(tt) |
| |
| # No, we need some messy parsing... |
| |
| # clean up |
| text = text.lstrip() |
| text = WEEKDAY_RE.sub("", text, 1) # Useless weekday |
| |
| # tz is time zone specifier string |
| day, mon, yr, hr, min, sec, tz = [None]*7 |
| |
| # loose regexp parse |
| m = LOOSE_HTTP_DATE_RE.search(text) |
| if m is not None: |
| day, mon, yr, hr, min, sec, tz = m.groups() |
| else: |
| return None # bad format |
| |
| return _str2time(day, mon, yr, hr, min, sec, tz) |
| |
| ISO_DATE_RE = re.compile( |
| """^ |
| (\d{4}) # year |
| [-\/]? |
| (\d\d?) # numerical month |
| [-\/]? |
| (\d\d?) # day |
| (?: |
| (?:\s+|[-:Tt]) # separator before clock |
| (\d\d?):?(\d\d) # hour:min |
| (?::?(\d\d(?:\.\d*)?))? # optional seconds (and fractional) |
| )? # optional clock |
| \s* |
| ([-+]?\d\d?:?(:?\d\d)? |
| |Z|z)? # timezone (Z is "zero meridian", i.e. GMT) |
| \s*$""", re.X) |
| def iso2time(text): |
| """ |
| As for http2time, but parses the ISO 8601 formats: |
| |
| 1994-02-03 14:15:29 -0100 -- ISO 8601 format |
| 1994-02-03 14:15:29 -- zone is optional |
| 1994-02-03 -- only date |
| 1994-02-03T14:15:29 -- Use T as separator |
| 19940203T141529Z -- ISO 8601 compact format |
| 19940203 -- only date |
| |
| """ |
| # clean up |
| text = text.lstrip() |
| |
| # tz is time zone specifier string |
| day, mon, yr, hr, min, sec, tz = [None]*7 |
| |
| # loose regexp parse |
| m = ISO_DATE_RE.search(text) |
| if m is not None: |
| # XXX there's an extra bit of the timezone I'm ignoring here: is |
| # this the right thing to do? |
| yr, mon, day, hr, min, sec, tz, _ = m.groups() |
| else: |
| return None # bad format |
| |
| return _str2time(day, mon, yr, hr, min, sec, tz) |
| |
| |
| # Header parsing |
| # ----------------------------------------------------------------------------- |
| |
| def unmatched(match): |
| """Return unmatched part of re.Match object.""" |
| start, end = match.span(0) |
| return match.string[:start]+match.string[end:] |
| |
| HEADER_TOKEN_RE = re.compile(r"^\s*([^=\s;,]+)") |
| HEADER_QUOTED_VALUE_RE = re.compile(r"^\s*=\s*\"([^\"\\]*(?:\\.[^\"\\]*)*)\"") |
| HEADER_VALUE_RE = re.compile(r"^\s*=\s*([^\s;,]*)") |
| HEADER_ESCAPE_RE = re.compile(r"\\(.)") |
| def split_header_words(header_values): |
| r"""Parse header values into a list of lists containing key,value pairs. |
| |
| The function knows how to deal with ",", ";" and "=" as well as quoted |
| values after "=". A list of space separated tokens are parsed as if they |
| were separated by ";". |
| |
| If the header_values passed as argument contains multiple values, then they |
| are treated as if they were a single value separated by comma ",". |
| |
| This means that this function is useful for parsing header fields that |
| follow this syntax (BNF as from the HTTP/1.1 specification, but we relax |
| the requirement for tokens). |
| |
| headers = #header |
| header = (token | parameter) *( [";"] (token | parameter)) |
| |
| token = 1*<any CHAR except CTLs or separators> |
| separators = "(" | ")" | "<" | ">" | "@" |
| | "," | ";" | ":" | "\" | <"> |
| | "/" | "[" | "]" | "?" | "=" |
| | "{" | "}" | SP | HT |
| |
| quoted-string = ( <"> *(qdtext | quoted-pair ) <"> ) |
| qdtext = <any TEXT except <">> |
| quoted-pair = "\" CHAR |
| |
| parameter = attribute "=" value |
| attribute = token |
| value = token | quoted-string |
| |
| Each header is represented by a list of key/value pairs. The value for a |
| simple token (not part of a parameter) is None. Syntactically incorrect |
| headers will not necessarily be parsed as you would want. |
| |
| This is easier to describe with some examples: |
| |
| >>> split_header_words(['foo="bar"; port="80,81"; discard, bar=baz']) |
| [[('foo', 'bar'), ('port', '80,81'), ('discard', None)], [('bar', 'baz')]] |
| >>> split_header_words(['text/html; charset="iso-8859-1"']) |
| [[('text/html', None), ('charset', 'iso-8859-1')]] |
| >>> split_header_words([r'Basic realm="\"foo\bar\""']) |
| [[('Basic', None), ('realm', '"foobar"')]] |
| |
| """ |
| assert type(header_values) not in StringTypes |
| result = [] |
| for text in header_values: |
| orig_text = text |
| pairs = [] |
| while text: |
| m = HEADER_TOKEN_RE.search(text) |
| if m: |
| text = unmatched(m) |
| name = m.group(1) |
| m = HEADER_QUOTED_VALUE_RE.search(text) |
| if m: # quoted value |
| text = unmatched(m) |
| value = m.group(1) |
| value = HEADER_ESCAPE_RE.sub(r"\1", value) |
| else: |
| m = HEADER_VALUE_RE.search(text) |
| if m: # unquoted value |
| text = unmatched(m) |
| value = m.group(1) |
| value = value.rstrip() |
| else: |
| # no value, a lone token |
| value = None |
| pairs.append((name, value)) |
| elif text.lstrip().startswith(","): |
| # concatenated headers, as per RFC 2616 section 4.2 |
| text = text.lstrip()[1:] |
| if pairs: result.append(pairs) |
| pairs = [] |
| else: |
| # skip junk |
| non_junk, nr_junk_chars = re.subn("^[=\s;]*", "", text) |
| assert nr_junk_chars > 0, ( |
| "split_header_words bug: '%s', '%s', %s" % |
| (orig_text, text, pairs)) |
| text = non_junk |
| if pairs: result.append(pairs) |
| return result |
| |
| HEADER_JOIN_ESCAPE_RE = re.compile(r"([\"\\])") |
| def join_header_words(lists): |
| """Do the inverse (almost) of the conversion done by split_header_words. |
| |
| Takes a list of lists of (key, value) pairs and produces a single header |
| value. Attribute values are quoted if needed. |
| |
| >>> join_header_words([[("text/plain", None), ("charset", "iso-8859/1")]]) |
| 'text/plain; charset="iso-8859/1"' |
| >>> join_header_words([[("text/plain", None)], [("charset", "iso-8859/1")]]) |
| 'text/plain, charset="iso-8859/1"' |
| |
| """ |
| headers = [] |
| for pairs in lists: |
| attr = [] |
| for k, v in pairs: |
| if v is not None: |
| if not re.search(r"^\w+$", v): |
| v = HEADER_JOIN_ESCAPE_RE.sub(r"\\\1", v) # escape " and \ |
| v = '"%s"' % v |
| k = "%s=%s" % (k, v) |
| attr.append(k) |
| if attr: headers.append("; ".join(attr)) |
| return ", ".join(headers) |
| |
| def parse_ns_headers(ns_headers): |
| """Ad-hoc parser for Netscape protocol cookie-attributes. |
| |
| The old Netscape cookie format for Set-Cookie can for instance contain |
| an unquoted "," in the expires field, so we have to use this ad-hoc |
| parser instead of split_header_words. |
| |
| XXX This may not make the best possible effort to parse all the crap |
| that Netscape Cookie headers contain. Ronald Tschalar's HTTPClient |
| parser is probably better, so could do worse than following that if |
| this ever gives any trouble. |
| |
| Currently, this is also used for parsing RFC 2109 cookies. |
| |
| """ |
| known_attrs = ("expires", "domain", "path", "secure", |
| # RFC 2109 attrs (may turn up in Netscape cookies, too) |
| "port", "max-age") |
| |
| result = [] |
| for ns_header in ns_headers: |
| pairs = [] |
| version_set = False |
| for param in re.split(r";\s*", ns_header): |
| param = param.rstrip() |
| if param == "": continue |
| if "=" not in param: |
| if param.lower() in known_attrs: |
| k, v = param, None |
| else: |
| # cookie with missing value |
| k, v = param, None |
| else: |
| k, v = re.split(r"\s*=\s*", param, 1) |
| k = k.lstrip() |
| if k is not None: |
| lc = k.lower() |
| if lc in known_attrs: |
| k = lc |
| if k == "version": |
| # This is an RFC 2109 cookie. Will be treated as RFC 2965 |
| # cookie in rest of code. |
| # Probably it should be parsed with split_header_words, but |
| # that's too much hassle. |
| version_set = True |
| if k == "expires": |
| # convert expires date to seconds since epoch |
| if v.startswith('"'): v = v[1:] |
| if v.endswith('"'): v = v[:-1] |
| v = http2time(v) # None if invalid |
| pairs.append((k, v)) |
| |
| if pairs: |
| if not version_set: |
| pairs.append(("version", "0")) |
| result.append(pairs) |
| |
| return result |
| |
| |
| IPV4_RE = re.compile(r"\.\d+$") |
| def is_HDN(text): |
| """Return True if text is a host domain name.""" |
| # XXX |
| # This may well be wrong. Which RFC is HDN defined in, if any (for |
| # the purposes of RFC 2965)? |
| # For the current implementation, what about IPv6? Remember to look |
| # at other uses of IPV4_RE also, if change this. |
| if IPV4_RE.search(text): |
| return False |
| if text == "": |
| return False |
| if text[0] == "." or text[-1] == ".": |
| return False |
| return True |
| |
| def domain_match(A, B): |
| """Return True if domain A domain-matches domain B, according to RFC 2965. |
| |
| A and B may be host domain names or IP addresses. |
| |
| RFC 2965, section 1: |
| |
| Host names can be specified either as an IP address or a HDN string. |
| Sometimes we compare one host name with another. (Such comparisons SHALL |
| be case-insensitive.) Host A's name domain-matches host B's if |
| |
| * their host name strings string-compare equal; or |
| |
| * A is a HDN string and has the form NB, where N is a non-empty |
| name string, B has the form .B', and B' is a HDN string. (So, |
| x.y.com domain-matches .Y.com but not Y.com.) |
| |
| Note that domain-match is not a commutative operation: a.b.c.com |
| domain-matches .c.com, but not the reverse. |
| |
| """ |
| # Note that, if A or B are IP addresses, the only relevant part of the |
| # definition of the domain-match algorithm is the direct string-compare. |
| A = A.lower() |
| B = B.lower() |
| if A == B: |
| return True |
| if not is_HDN(A): |
| return False |
| i = A.rfind(B) |
| if i == -1 or i == 0: |
| # A does not have form NB, or N is the empty string |
| return False |
| if not B.startswith("."): |
| return False |
| if not is_HDN(B[1:]): |
| return False |
| return True |
| |
| def liberal_is_HDN(text): |
| """Return True if text is a sort-of-like a host domain name. |
| |
| For accepting/blocking domains. |
| |
| """ |
| if IPV4_RE.search(text): |
| return False |
| return True |
| |
| def user_domain_match(A, B): |
| """For blocking/accepting domains. |
| |
| A and B may be host domain names or IP addresses. |
| |
| """ |
| A = A.lower() |
| B = B.lower() |
| if not (liberal_is_HDN(A) and liberal_is_HDN(B)): |
| if A == B: |
| # equal IP addresses |
| return True |
| return False |
| initial_dot = B.startswith(".") |
| if initial_dot and A.endswith(B): |
| return True |
| if not initial_dot and A == B: |
| return True |
| return False |
| |
| cut_port_re = re.compile(r":\d+$") |
| def request_host(request): |
| """Return request-host, as defined by RFC 2965. |
| |
| Variation from RFC: returned value is lowercased, for convenient |
| comparison. |
| |
| """ |
| url = request.get_full_url() |
| host = urlparse.urlparse(url)[1] |
| if host == "": |
| host = request.get_header("Host", "") |
| |
| # remove port, if present |
| host = cut_port_re.sub("", host, 1) |
| return host.lower() |
| |
| def eff_request_host(request): |
| """Return a tuple (request-host, effective request-host name). |
| |
| As defined by RFC 2965, except both are lowercased. |
| |
| """ |
| erhn = req_host = request_host(request) |
| if req_host.find(".") == -1 and not IPV4_RE.search(req_host): |
| erhn = req_host + ".local" |
| return req_host, erhn |
| |
| def request_path(request): |
| """request-URI, as defined by RFC 2965.""" |
| url = request.get_full_url() |
| #scheme, netloc, path, parameters, query, frag = urlparse.urlparse(url) |
| #req_path = escape_path("".join(urlparse.urlparse(url)[2:])) |
| path, parameters, query, frag = urlparse.urlparse(url)[2:] |
| if parameters: |
| path = "%s;%s" % (path, parameters) |
| path = escape_path(path) |
| req_path = urlparse.urlunparse(("", "", path, "", query, frag)) |
| if not req_path.startswith("/"): |
| # fix bad RFC 2396 absoluteURI |
| req_path = "/"+req_path |
| return req_path |
| |
| def request_port(request): |
| host = request.get_host() |
| i = host.find(':') |
| if i >= 0: |
| port = host[i+1:] |
| try: |
| int(port) |
| except ValueError: |
| debug("nonnumeric port: '%s'", port) |
| return None |
| else: |
| port = DEFAULT_HTTP_PORT |
| return port |
| |
| # Characters in addition to A-Z, a-z, 0-9, '_', '.', and '-' that don't |
| # need to be escaped to form a valid HTTP URL (RFCs 2396 and 1738). |
| HTTP_PATH_SAFE = "%/;:@&=+$,!~*'()" |
| ESCAPED_CHAR_RE = re.compile(r"%([0-9a-fA-F][0-9a-fA-F])") |
| def uppercase_escaped_char(match): |
| return "%%%s" % match.group(1).upper() |
| def escape_path(path): |
| """Escape any invalid characters in HTTP URL, and uppercase all escapes.""" |
| # There's no knowing what character encoding was used to create URLs |
| # containing %-escapes, but since we have to pick one to escape invalid |
| # path characters, we pick UTF-8, as recommended in the HTML 4.0 |
| # specification: |
| # http://www.w3.org/TR/REC-html40/appendix/notes.html#h-B.2.1 |
| # And here, kind of: draft-fielding-uri-rfc2396bis-03 |
| # (And in draft IRI specification: draft-duerst-iri-05) |
| # (And here, for new URI schemes: RFC 2718) |
| if isinstance(path, unicode): |
| path = path.encode("utf-8") |
| path = urllib.quote(path, HTTP_PATH_SAFE) |
| path = ESCAPED_CHAR_RE.sub(uppercase_escaped_char, path) |
| return path |
| |
| def reach(h): |
| """Return reach of host h, as defined by RFC 2965, section 1. |
| |
| The reach R of a host name H is defined as follows: |
| |
| * If |
| |
| - H is the host domain name of a host; and, |
| |
| - H has the form A.B; and |
| |
| - A has no embedded (that is, interior) dots; and |
| |
| - B has at least one embedded dot, or B is the string "local". |
| then the reach of H is .B. |
| |
| * Otherwise, the reach of H is H. |
| |
| >>> reach("www.acme.com") |
| '.acme.com' |
| >>> reach("acme.com") |
| 'acme.com' |
| >>> reach("acme.local") |
| '.local' |
| |
| """ |
| i = h.find(".") |
| if i >= 0: |
| #a = h[:i] # this line is only here to show what a is |
| b = h[i+1:] |
| i = b.find(".") |
| if is_HDN(h) and (i >= 0 or b == "local"): |
| return "."+b |
| return h |
| |
| def is_third_party(request): |
| """ |
| |
| RFC 2965, section 3.3.6: |
| |
| An unverifiable transaction is to a third-party host if its request- |
| host U does not domain-match the reach R of the request-host O in the |
| origin transaction. |
| |
| """ |
| req_host = request_host(request) |
| if not domain_match(req_host, reach(request.get_origin_req_host())): |
| return True |
| else: |
| return False |
| |
| |
| class Cookie: |
| """HTTP Cookie. |
| |
| This class represents both Netscape and RFC 2965 cookies. |
| |
| This is deliberately a very simple class. It just holds attributes. It's |
| possible to construct Cookie instances that don't comply with the cookie |
| standards. CookieJar.make_cookies is the factory function for Cookie |
| objects -- it deals with cookie parsing, supplying defaults, and |
| normalising to the representation used in this class. CookiePolicy is |
| responsible for checking them to see whether they should be accepted from |
| and returned to the server. |
| |
| Note that the port may be present in the headers, but unspecified ("Port" |
| rather than"Port=80", for example); if this is the case, port is None. |
| |
| """ |
| |
| def __init__(self, version, name, value, |
| port, port_specified, |
| domain, domain_specified, domain_initial_dot, |
| path, path_specified, |
| secure, |
| expires, |
| discard, |
| comment, |
| comment_url, |
| rest): |
| |
| if version is not None: version = int(version) |
| if expires is not None: expires = int(expires) |
| if port is None and port_specified is True: |
| raise ValueError("if port is None, port_specified must be false") |
| |
| self.version = version |
| self.name = name |
| self.value = value |
| self.port = port |
| self.port_specified = port_specified |
| # normalise case, as per RFC 2965 section 3.3.3 |
| self.domain = domain.lower() |
| self.domain_specified = domain_specified |
| # Sigh. We need to know whether the domain given in the |
| # cookie-attribute had an initial dot, in order to follow RFC 2965 |
| # (as clarified in draft errata). Needed for the returned $Domain |
| # value. |
| self.domain_initial_dot = domain_initial_dot |
| self.path = path |
| self.path_specified = path_specified |
| self.secure = secure |
| self.expires = expires |
| self.discard = discard |
| self.comment = comment |
| self.comment_url = comment_url |
| |
| self._rest = copy.copy(rest) |
| |
| def has_nonstandard_attr(self, name): |
| return name in self._rest |
| def get_nonstandard_attr(self, name, default=None): |
| return self._rest.get(name, default) |
| def set_nonstandard_attr(self, name, value): |
| self._rest[name] = value |
| |
| def is_expired(self, now=None): |
| if now is None: now = time.time() |
| if (self.expires is not None) and (self.expires <= now): |
| return True |
| return False |
| |
| def __str__(self): |
| if self.port is None: p = "" |
| else: p = ":"+self.port |
| limit = self.domain + p + self.path |
| if self.value is not None: |
| namevalue = "%s=%s" % (self.name, self.value) |
| else: |
| namevalue = self.name |
| return "<Cookie %s for %s>" % (namevalue, limit) |
| |
| def __repr__(self): |
| args = [] |
| for name in ["version", "name", "value", |
| "port", "port_specified", |
| "domain", "domain_specified", "domain_initial_dot", |
| "path", "path_specified", |
| "secure", "expires", "discard", "comment", "comment_url", |
| ]: |
| attr = getattr(self, name) |
| args.append("%s=%s" % (name, repr(attr))) |
| args.append("rest=%s" % repr(self._rest)) |
| return "Cookie(%s)" % ", ".join(args) |
| |
| |
| class CookiePolicy: |
| """Defines which cookies get accepted from and returned to server. |
| |
| May also modify cookies, though this is probably a bad idea. |
| |
| The subclass DefaultCookiePolicy defines the standard rules for Netscape |
| and RFC 2965 cookies -- override that if you want a customised policy. |
| |
| """ |
| def set_ok(self, cookie, request): |
| """Return true if (and only if) cookie should be accepted from server. |
| |
| Currently, pre-expired cookies never get this far -- the CookieJar |
| class deletes such cookies itself. |
| |
| """ |
| raise NotImplementedError() |
| |
| def return_ok(self, cookie, request): |
| """Return true if (and only if) cookie should be returned to server.""" |
| raise NotImplementedError() |
| |
| def domain_return_ok(self, domain, request): |
| """Return false if cookies should not be returned, given cookie domain. |
| """ |
| return True |
| |
| def path_return_ok(self, path, request): |
| """Return false if cookies should not be returned, given cookie path. |
| """ |
| return True |
| |
| |
| class DefaultCookiePolicy(CookiePolicy): |
| """Implements the standard rules for accepting and returning cookies.""" |
| |
| DomainStrictNoDots = 1 |
| DomainStrictNonDomain = 2 |
| DomainRFC2965Match = 4 |
| |
| DomainLiberal = 0 |
| DomainStrict = DomainStrictNoDots|DomainStrictNonDomain |
| |
| def __init__(self, |
| blocked_domains=None, allowed_domains=None, |
| netscape=True, rfc2965=False, |
| hide_cookie2=False, |
| strict_domain=False, |
| strict_rfc2965_unverifiable=True, |
| strict_ns_unverifiable=False, |
| strict_ns_domain=DomainLiberal, |
| strict_ns_set_initial_dollar=False, |
| strict_ns_set_path=False, |
| ): |
| """Constructor arguments should be passed as keyword arguments only.""" |
| self.netscape = netscape |
| self.rfc2965 = rfc2965 |
| self.hide_cookie2 = hide_cookie2 |
| self.strict_domain = strict_domain |
| self.strict_rfc2965_unverifiable = strict_rfc2965_unverifiable |
| self.strict_ns_unverifiable = strict_ns_unverifiable |
| self.strict_ns_domain = strict_ns_domain |
| self.strict_ns_set_initial_dollar = strict_ns_set_initial_dollar |
| self.strict_ns_set_path = strict_ns_set_path |
| |
| if blocked_domains is not None: |
| self._blocked_domains = tuple(blocked_domains) |
| else: |
| self._blocked_domains = () |
| |
| if allowed_domains is not None: |
| allowed_domains = tuple(allowed_domains) |
| self._allowed_domains = allowed_domains |
| |
| def blocked_domains(self): |
| """Return the sequence of blocked domains (as a tuple).""" |
| return self._blocked_domains |
| def set_blocked_domains(self, blocked_domains): |
| """Set the sequence of blocked domains.""" |
| self._blocked_domains = tuple(blocked_domains) |
| |
| def is_blocked(self, domain): |
| for blocked_domain in self._blocked_domains: |
| if user_domain_match(domain, blocked_domain): |
| return True |
| return False |
| |
| def allowed_domains(self): |
| """Return None, or the sequence of allowed domains (as a tuple).""" |
| return self._allowed_domains |
| def set_allowed_domains(self, allowed_domains): |
| """Set the sequence of allowed domains, or None.""" |
| if allowed_domains is not None: |
| allowed_domains = tuple(allowed_domains) |
| self._allowed_domains = allowed_domains |
| |
| def is_not_allowed(self, domain): |
| if self._allowed_domains is None: |
| return False |
| for allowed_domain in self._allowed_domains: |
| if user_domain_match(domain, allowed_domain): |
| return False |
| return True |
| |
| def set_ok(self, cookie, request): |
| """ |
| If you override .set_ok(), be sure to call this method. If it returns |
| false, so should your subclass (assuming your subclass wants to be more |
| strict about which cookies to accept). |
| |
| """ |
| debug(" - checking cookie %s=%s", cookie.name, cookie.value) |
| |
| assert cookie.name is not None |
| |
| for n in "version", "verifiability", "name", "path", "domain", "port": |
| fn_name = "set_ok_"+n |
| fn = getattr(self, fn_name) |
| if not fn(cookie, request): |
| return False |
| |
| return True |
| |
| def set_ok_version(self, cookie, request): |
| if cookie.version is None: |
| # Version is always set to 0 by parse_ns_headers if it's a Netscape |
| # cookie, so this must be an invalid RFC 2965 cookie. |
| debug(" Set-Cookie2 without version attribute (%s=%s)", |
| cookie.name, cookie.value) |
| return False |
| if cookie.version > 0 and not self.rfc2965: |
| debug(" RFC 2965 cookies are switched off") |
| return False |
| elif cookie.version == 0 and not self.netscape: |
| debug(" Netscape cookies are switched off") |
| return False |
| return True |
| |
| def set_ok_verifiability(self, cookie, request): |
| if request.is_unverifiable() and is_third_party(request): |
| if cookie.version > 0 and self.strict_rfc2965_unverifiable: |
| debug(" third-party RFC 2965 cookie during " |
| "unverifiable transaction") |
| return False |
| elif cookie.version == 0 and self.strict_ns_unverifiable: |
| debug(" third-party Netscape cookie during " |
| "unverifiable transaction") |
| return False |
| return True |
| |
| def set_ok_name(self, cookie, request): |
| # Try and stop servers setting V0 cookies designed to hack other |
| # servers that know both V0 and V1 protocols. |
| if (cookie.version == 0 and self.strict_ns_set_initial_dollar and |
| cookie.name.startswith("$")): |
| debug(" illegal name (starts with '$'): '%s'", cookie.name) |
| return False |
| return True |
| |
| def set_ok_path(self, cookie, request): |
| if cookie.path_specified: |
| req_path = request_path(request) |
| if ((cookie.version > 0 or |
| (cookie.version == 0 and self.strict_ns_set_path)) and |
| not req_path.startswith(cookie.path)): |
| debug(" path attribute %s is not a prefix of request " |
| "path %s", cookie.path, req_path) |
| return False |
| return True |
| |
| def set_ok_domain(self, cookie, request): |
| if self.is_blocked(cookie.domain): |
| debug(" domain %s is in user block-list", cookie.domain) |
| return False |
| if self.is_not_allowed(cookie.domain): |
| debug(" domain %s is not in user allow-list", cookie.domain) |
| return False |
| if cookie.domain_specified: |
| req_host, erhn = eff_request_host(request) |
| domain = cookie.domain |
| if self.strict_domain and (domain.count(".") >= 2): |
| i = domain.rfind(".") |
| j = domain.rfind(".", 0, i) |
| if j == 0: # domain like .foo.bar |
| tld = domain[i+1:] |
| sld = domain[j+1:i] |
| if (sld.lower() in [ |
| "co", "ac", |
| "com", "edu", "org", "net", "gov", "mil", "int"] and |
| len(tld) == 2): |
| # domain like .co.uk |
| debug(" country-code second level domain %s", domain) |
| return False |
| if domain.startswith("."): |
| undotted_domain = domain[1:] |
| else: |
| undotted_domain = domain |
| embedded_dots = (undotted_domain.find(".") >= 0) |
| if not embedded_dots and domain != ".local": |
| debug(" non-local domain %s contains no embedded dot", |
| domain) |
| return False |
| if cookie.version == 0: |
| if (not erhn.endswith(domain) and |
| (not erhn.startswith(".") and |
| not ("."+erhn).endswith(domain))): |
| debug(" effective request-host %s (even with added " |
| "initial dot) does not end end with %s", |
| erhn, domain) |
| return False |
| if (cookie.version > 0 or |
| (self.strict_ns_domain & self.DomainRFC2965Match)): |
| if not domain_match(erhn, domain): |
| debug(" effective request-host %s does not domain-match " |
| "%s", erhn, domain) |
| return False |
| if (cookie.version > 0 or |
| (self.strict_ns_domain & self.DomainStrictNoDots)): |
| host_prefix = req_host[:-len(domain)] |
| if (host_prefix.find(".") >= 0 and |
| not IPV4_RE.search(req_host)): |
| debug(" host prefix %s for domain %s contains a dot", |
| host_prefix, domain) |
| return False |
| return True |
| |
| def set_ok_port(self, cookie, request): |
| if cookie.port_specified: |
| req_port = request_port(request) |
| if req_port is None: |
| req_port = "80" |
| else: |
| req_port = str(req_port) |
| for p in cookie.port.split(","): |
| try: |
| int(p) |
| except ValueError: |
| debug(" bad port %s (not numeric)", p) |
| return False |
| if p == req_port: |
| break |
| else: |
| debug(" request port (%s) not found in %s", |
| req_port, cookie.port) |
| return False |
| return True |
| |
| def return_ok(self, cookie, request): |
| """ |
| If you override .return_ok(), be sure to call this method. If it |
| returns false, so should your subclass (assuming your subclass wants to |
| be more strict about which cookies to return). |
| |
| """ |
| # Path has already been checked by .path_return_ok(), and domain |
| # blocking done by .domain_return_ok(). |
| debug(" - checking cookie %s=%s", cookie.name, cookie.value) |
| |
| for n in "version", "verifiability", "secure", "expires", "port", "domain": |
| fn_name = "return_ok_"+n |
| fn = getattr(self, fn_name) |
| if not fn(cookie, request): |
| return False |
| return True |
| |
| def return_ok_version(self, cookie, request): |
| if cookie.version > 0 and not self.rfc2965: |
| debug(" RFC 2965 cookies are switched off") |
| return False |
| elif cookie.version == 0 and not self.netscape: |
| debug(" Netscape cookies are switched off") |
| return False |
| return True |
| |
| def return_ok_verifiability(self, cookie, request): |
| if request.is_unverifiable() and is_third_party(request): |
| if cookie.version > 0 and self.strict_rfc2965_unverifiable: |
| debug(" third-party RFC 2965 cookie during unverifiable " |
| "transaction") |
| return False |
| elif cookie.version == 0 and self.strict_ns_unverifiable: |
| debug(" third-party Netscape cookie during unverifiable " |
| "transaction") |
| return False |
| return True |
| |
| def return_ok_secure(self, cookie, request): |
| if cookie.secure and request.get_type() != "https": |
| debug(" secure cookie with non-secure request") |
| return False |
| return True |
| |
| def return_ok_expires(self, cookie, request): |
| if cookie.is_expired(self._now): |
| debug(" cookie expired") |
| return False |
| return True |
| |
| def return_ok_port(self, cookie, request): |
| if cookie.port: |
| req_port = request_port(request) |
| if req_port is None: |
| req_port = "80" |
| for p in cookie.port.split(","): |
| if p == req_port: |
| break |
| else: |
| debug(" request port %s does not match cookie port %s", |
| req_port, cookie.port) |
| return False |
| return True |
| |
| def return_ok_domain(self, cookie, request): |
| req_host, erhn = eff_request_host(request) |
| domain = cookie.domain |
| |
| # strict check of non-domain cookies: Mozilla does this, MSIE5 doesn't |
| if (cookie.version == 0 and |
| (self.strict_ns_domain & self.DomainStrictNonDomain) and |
| not cookie.domain_specified and domain != erhn): |
| debug(" cookie with unspecified domain does not string-compare " |
| "equal to request domain") |
| return False |
| |
| if cookie.version > 0 and not domain_match(erhn, domain): |
| debug(" effective request-host name %s does not domain-match " |
| "RFC 2965 cookie domain %s", erhn, domain) |
| return False |
| if cookie.version == 0 and not ("."+erhn).endswith(domain): |
| debug(" request-host %s does not match Netscape cookie domain " |
| "%s", req_host, domain) |
| return False |
| return True |
| |
| def domain_return_ok(self, domain, request): |
| # Liberal check of. This is here as an optimization to avoid |
| # having to load lots of MSIE cookie files unless necessary. |
| req_host, erhn = eff_request_host(request) |
| if not req_host.startswith("."): |
| dotted_req_host = "."+req_host |
| if not erhn.startswith("."): |
| dotted_erhn = "."+erhn |
| if not (dotted_req_host.endswith(domain) or |
| dotted_erhn.endswith(domain)): |
| #debug(" request domain %s does not match cookie domain %s", |
| # req_host, domain) |
| return False |
| |
| if self.is_blocked(domain): |
| debug(" domain %s is in user block-list", domain) |
| return False |
| if self.is_not_allowed(domain): |
| debug(" domain %s is not in user allow-list", domain) |
| return False |
| |
| return True |
| |
| def path_return_ok(self, path, request): |
| debug("- checking cookie path=%s", path) |
| req_path = request_path(request) |
| if not req_path.startswith(path): |
| debug(" %s does not path-match %s", req_path, path) |
| return False |
| return True |
| |
| |
| def vals_sorted_by_key(adict): |
| keys = adict.keys() |
| keys.sort() |
| return map(adict.get, keys) |
| |
| def deepvalues(mapping): |
| """Iterates over nested mapping, depth-first, in sorted order by key.""" |
| values = vals_sorted_by_key(mapping) |
| for obj in values: |
| mapping = False |
| try: |
| obj.items |
| except AttributeError: |
| pass |
| else: |
| mapping = True |
| for subobj in deepvalues(obj): |
| yield subobj |
| if not mapping: |
| yield obj |
| |
| |
| # Used as second parameter to dict.get() method, to distinguish absent |
| # dict key from one with a None value. |
| class Absent: pass |
| |
| class CookieJar: |
| """Collection of HTTP cookies. |
| |
| You may not need to know about this class: try |
| urllib2.build_opener(HTTPCookieProcessor).open(url). |
| |
| """ |
| |
| non_word_re = re.compile(r"\W") |
| quote_re = re.compile(r"([\"\\])") |
| strict_domain_re = re.compile(r"\.?[^.]*") |
| domain_re = re.compile(r"[^.]*") |
| dots_re = re.compile(r"^\.+") |
| |
| magic_re = r"^\#LWP-Cookies-(\d+\.\d+)" |
| |
| def __init__(self, policy=None): |
| if policy is None: |
| policy = DefaultCookiePolicy() |
| self._policy = policy |
| |
| self._cookies_lock = _threading.RLock() |
| self._cookies = {} |
| |
| def set_policy(self, policy): |
| self._policy = policy |
| |
| def _cookies_for_domain(self, domain, request): |
| cookies = [] |
| if not self._policy.domain_return_ok(domain, request): |
| return [] |
| debug("Checking %s for cookies to return", domain) |
| cookies_by_path = self._cookies[domain] |
| for path in cookies_by_path.keys(): |
| if not self._policy.path_return_ok(path, request): |
| continue |
| cookies_by_name = cookies_by_path[path] |
| for cookie in cookies_by_name.values(): |
| if not self._policy.return_ok(cookie, request): |
| debug(" not returning cookie") |
| continue |
| debug(" it's a match") |
| cookies.append(cookie) |
| return cookies |
| |
| def _cookies_for_request(self, request): |
| """Return a list of cookies to be returned to server.""" |
| cookies = [] |
| for domain in self._cookies.keys(): |
| cookies.extend(self._cookies_for_domain(domain, request)) |
| return cookies |
| |
| def _cookie_attrs(self, cookies): |
| """Return a list of cookie-attributes to be returned to server. |
| |
| like ['foo="bar"; $Path="/"', ...] |
| |
| The $Version attribute is also added when appropriate (currently only |
| once per request). |
| |
| """ |
| # add cookies in order of most specific (ie. longest) path first |
| def decreasing_size(a, b): return cmp(len(b.path), len(a.path)) |
| cookies.sort(decreasing_size) |
| |
| version_set = False |
| |
| attrs = [] |
| for cookie in cookies: |
| # set version of Cookie header |
| # XXX |
| # What should it be if multiple matching Set-Cookie headers have |
| # different versions themselves? |
| # Answer: there is no answer; was supposed to be settled by |
| # RFC 2965 errata, but that may never appear... |
| version = cookie.version |
| if not version_set: |
| version_set = True |
| if version > 0: |
| attrs.append("$Version=%s" % version) |
| |
| # quote cookie value if necessary |
| # (not for Netscape protocol, which already has any quotes |
| # intact, due to the poorly-specified Netscape Cookie: syntax) |
| if ((cookie.value is not None) and |
| self.non_word_re.search(cookie.value) and version > 0): |
| value = self.quote_re.sub(r"\\\1", cookie.value) |
| else: |
| value = cookie.value |
| |
| # add cookie-attributes to be returned in Cookie header |
| if cookie.value is None: |
| attrs.append(cookie.name) |
| else: |
| attrs.append("%s=%s" % (cookie.name, value)) |
| if version > 0: |
| if cookie.path_specified: |
| attrs.append('$Path="%s"' % cookie.path) |
| if cookie.domain.startswith("."): |
| domain = cookie.domain |
| if (not cookie.domain_initial_dot and |
| domain.startswith(".")): |
| domain = domain[1:] |
| attrs.append('$Domain="%s"' % domain) |
| if cookie.port is not None: |
| p = "$Port" |
| if cookie.port_specified: |
| p = p + ('="%s"' % cookie.port) |
| attrs.append(p) |
| |
| return attrs |
| |
| def add_cookie_header(self, request): |
| """Add correct Cookie: header to request (urllib2.Request object). |
| |
| The Cookie2 header is also added unless policy.hide_cookie2 is true. |
| |
| """ |
| debug("add_cookie_header") |
| self._cookies_lock.acquire() |
| |
| self._policy._now = self._now = int(time.time()) |
| |
| req_host, erhn = eff_request_host(request) |
| strict_non_domain = ( |
| self._policy.strict_ns_domain & self._policy.DomainStrictNonDomain) |
| |
| cookies = self._cookies_for_request(request) |
| |
| attrs = self._cookie_attrs(cookies) |
| if attrs: |
| if not request.has_header("Cookie"): |
| request.add_unredirected_header( |
| "Cookie", "; ".join(attrs)) |
| |
| # if necessary, advertise that we know RFC 2965 |
| if (self._policy.rfc2965 and not self._policy.hide_cookie2 and |
| not request.has_header("Cookie2")): |
| for cookie in cookies: |
| if cookie.version != 1: |
| request.add_unredirected_header("Cookie2", '$Version="1"') |
| break |
| |
| self._cookies_lock.release() |
| |
| self.clear_expired_cookies() |
| |
| def _normalized_cookie_tuples(self, attrs_set): |
| """Return list of tuples containing normalised cookie information. |
| |
| attrs_set is the list of lists of key,value pairs extracted from |
| the Set-Cookie or Set-Cookie2 headers. |
| |
| Tuples are name, value, standard, rest, where name and value are the |
| cookie name and value, standard is a dictionary containing the standard |
| cookie-attributes (discard, secure, version, expires or max-age, |
| domain, path and port) and rest is a dictionary containing the rest of |
| the cookie-attributes. |
| |
| """ |
| cookie_tuples = [] |
| |
| boolean_attrs = "discard", "secure" |
| value_attrs = ("version", |
| "expires", "max-age", |
| "domain", "path", "port", |
| "comment", "commenturl") |
| |
| for cookie_attrs in attrs_set: |
| name, value = cookie_attrs[0] |
| |
| # Build dictionary of standard cookie-attributes (standard) and |
| # dictionary of other cookie-attributes (rest). |
| |
| # Note: expiry time is normalised to seconds since epoch. V0 |
| # cookies should have the Expires cookie-attribute, and V1 cookies |
| # should have Max-Age, but since V1 includes RFC 2109 cookies (and |
| # since V0 cookies may be a mish-mash of Netscape and RFC 2109), we |
| # accept either (but prefer Max-Age). |
| max_age_set = False |
| |
| bad_cookie = False |
| |
| standard = {} |
| rest = {} |
| for k, v in cookie_attrs[1:]: |
| lc = k.lower() |
| # don't lose case distinction for unknown fields |
| if lc in value_attrs or lc in boolean_attrs: |
| k = lc |
| if k in boolean_attrs and v is None: |
| # boolean cookie-attribute is present, but has no value |
| # (like "discard", rather than "port=80") |
| v = True |
| if k in standard: |
| # only first value is significant |
| continue |
| if k == "domain": |
| if v is None: |
| debug(" missing value for domain attribute") |
| bad_cookie = True |
| break |
| # RFC 2965 section 3.3.3 |
| v = v.lower() |
| if k == "expires": |
| if max_age_set: |
| # Prefer max-age to expires (like Mozilla) |
| continue |
| if v is None: |
| debug(" missing or invalid value for expires " |
| "attribute: treating as session cookie") |
| continue |
| if k == "max-age": |
| max_age_set = True |
| try: |
| v = int(v) |
| except ValueError: |
| debug(" missing or invalid (non-numeric) value for " |
| "max-age attribute") |
| bad_cookie = True |
| break |
| # convert RFC 2965 Max-Age to seconds since epoch |
| # XXX Strictly you're supposed to follow RFC 2616 |
| # age-calculation rules. Remember that zero Max-Age is a |
| # is a request to discard (old and new) cookie, though. |
| k = "expires" |
| v = self._now + v |
| if (k in value_attrs) or (k in boolean_attrs): |
| if (v is None and |
| k not in ["port", "comment", "commenturl"]): |
| debug(" missing value for %s attribute" % k) |
| bad_cookie = True |
| break |
| standard[k] = v |
| else: |
| rest[k] = v |
| |
| if bad_cookie: |
| continue |
| |
| cookie_tuples.append((name, value, standard, rest)) |
| |
| return cookie_tuples |
| |
| def _cookie_from_cookie_tuple(self, tup, request): |
| # standard is dict of standard cookie-attributes, rest is dict of the |
| # rest of them |
| name, value, standard, rest = tup |
| |
| domain = standard.get("domain", Absent) |
| path = standard.get("path", Absent) |
| port = standard.get("port", Absent) |
| expires = standard.get("expires", Absent) |
| |
| # set the easy defaults |
| version = standard.get("version", None) |
| if version is not None: version = int(version) |
| secure = standard.get("secure", False) |
| # (discard is also set if expires is Absent) |
| discard = standard.get("discard", False) |
| comment = standard.get("comment", None) |
| comment_url = standard.get("commenturl", None) |
| |
| # set default path |
| if path is not Absent and path != "": |
| path_specified = True |
| path = escape_path(path) |
| else: |
| path_specified = False |
| path = request_path(request) |
| i = path.rfind("/") |
| if i != -1: |
| if version == 0: |
| # Netscape spec parts company from reality here |
| path = path[:i] |
| else: |
| path = path[:i+1] |
| if len(path) == 0: path = "/" |
| |
| # set default domain |
| domain_specified = domain is not Absent |
| # but first we have to remember whether it starts with a dot |
| domain_initial_dot = False |
| if domain_specified: |
| domain_initial_dot = bool(domain.startswith(".")) |
| if domain is Absent: |
| req_host, erhn = eff_request_host(request) |
| domain = erhn |
| elif not domain.startswith("."): |
| domain = "."+domain |
| |
| # set default port |
| port_specified = False |
| if port is not Absent: |
| if port is None: |
| # Port attr present, but has no value: default to request port. |
| # Cookie should then only be sent back on that port. |
| port = request_port(request) |
| else: |
| port_specified = True |
| port = re.sub(r"\s+", "", port) |
| else: |
| # No port attr present. Cookie can be sent back on any port. |
| port = None |
| |
| # set default expires and discard |
| if expires is Absent: |
| expires = None |
| discard = True |
| elif expires <= self._now: |
| # Expiry date in past is request to delete cookie. This can't be |
| # in DefaultCookiePolicy, because can't delete cookies there. |
| try: |
| self.clear(domain, path, name) |
| except KeyError: |
| pass |
| debug("Expiring cookie, domain='%s', path='%s', name='%s'", |
| domain, path, name) |
| return None |
| |
| return Cookie(version, |
| name, value, |
| port, port_specified, |
| domain, domain_specified, domain_initial_dot, |
| path, path_specified, |
| secure, |
| expires, |
| discard, |
| comment, |
| comment_url, |
| rest) |
| |
| def _cookies_from_attrs_set(self, attrs_set, request): |
| cookie_tuples = self._normalized_cookie_tuples(attrs_set) |
| |
| cookies = [] |
| for tup in cookie_tuples: |
| cookie = self._cookie_from_cookie_tuple(tup, request) |
| if cookie: cookies.append(cookie) |
| return cookies |
| |
| def make_cookies(self, response, request): |
| """Return sequence of Cookie objects extracted from response object.""" |
| # get cookie-attributes for RFC 2965 and Netscape protocols |
| headers = response.info() |
| rfc2965_hdrs = headers.getheaders("Set-Cookie2") |
| ns_hdrs = headers.getheaders("Set-Cookie") |
| |
| rfc2965 = self._policy.rfc2965 |
| netscape = self._policy.netscape |
| |
| if ((not rfc2965_hdrs and not ns_hdrs) or |
| (not ns_hdrs and not rfc2965) or |
| (not rfc2965_hdrs and not netscape) or |
| (not netscape and not rfc2965)): |
| return [] # no relevant cookie headers: quick exit |
| |
| try: |
| cookies = self._cookies_from_attrs_set( |
| split_header_words(rfc2965_hdrs), request) |
| except: |
| reraise_unmasked_exceptions() |
| cookies = [] |
| |
| if ns_hdrs and netscape: |
| try: |
| ns_cookies = self._cookies_from_attrs_set( |
| parse_ns_headers(ns_hdrs), request) |
| except: |
| reraise_unmasked_exceptions() |
| ns_cookies = [] |
| |
| # Look for Netscape cookies (from Set-Cookie headers) that match |
| # corresponding RFC 2965 cookies (from Set-Cookie2 headers). |
| # For each match, keep the RFC 2965 cookie and ignore the Netscape |
| # cookie (RFC 2965 section 9.1). Actually, RFC 2109 cookies are |
| # bundled in with the Netscape cookies for this purpose, which is |
| # reasonable behaviour. |
| if rfc2965: |
| lookup = {} |
| for cookie in cookies: |
| lookup[(cookie.domain, cookie.path, cookie.name)] = None |
| |
| def no_matching_rfc2965(ns_cookie, lookup=lookup): |
| key = ns_cookie.domain, ns_cookie.path, ns_cookie.name |
| return key not in lookup |
| ns_cookies = filter(no_matching_rfc2965, ns_cookies) |
| |
| if ns_cookies: |
| cookies.extend(ns_cookies) |
| |
| return cookies |
| |
| def set_cookie_if_ok(self, cookie, request): |
| """Set a cookie if policy says it's OK to do so.""" |
| self._cookies_lock.acquire() |
| self._policy._now = self._now = int(time.time()) |
| |
| if self._policy.set_ok(cookie, request): |
| self.set_cookie(cookie) |
| |
| self._cookies_lock.release() |
| |
| def set_cookie(self, cookie): |
| """Set a cookie, without checking whether or not it should be set.""" |
| c = self._cookies |
| self._cookies_lock.acquire() |
| try: |
| if cookie.domain not in c: c[cookie.domain] = {} |
| c2 = c[cookie.domain] |
| if cookie.path not in c2: c2[cookie.path] = {} |
| c3 = c2[cookie.path] |
| c3[cookie.name] = cookie |
| finally: |
| self._cookies_lock.release() |
| |
| def extract_cookies(self, response, request): |
| """Extract cookies from response, where allowable given the request.""" |
| debug("extract_cookies: %s", response.info()) |
| self._cookies_lock.acquire() |
| self._policy._now = self._now = int(time.time()) |
| |
| for cookie in self.make_cookies(response, request): |
| if self._policy.set_ok(cookie, request): |
| debug(" setting cookie: %s", cookie) |
| self.set_cookie(cookie) |
| self._cookies_lock.release() |
| |
| def clear(self, domain=None, path=None, name=None): |
| """Clear some cookies. |
| |
| Invoking this method without arguments will clear all cookies. If |
| given a single argument, only cookies belonging to that domain will be |
| removed. If given two arguments, cookies belonging to the specified |
| path within that domain are removed. If given three arguments, then |
| the cookie with the specified name, path and domain is removed. |
| |
| Raises KeyError if no matching cookie exists. |
| |
| """ |
| if name is not None: |
| if (domain is None) or (path is None): |
| raise ValueError( |
| "domain and path must be given to remove a cookie by name") |
| del self._cookies[domain][path][name] |
| elif path is not None: |
| if domain is None: |
| raise ValueError( |
| "domain must be given to remove cookies by path") |
| del self._cookies[domain][path] |
| elif domain is not None: |
| del self._cookies[domain] |
| else: |
| self._cookies = {} |
| |
| def clear_session_cookies(self): |
| """Discard all session cookies. |
| |
| Note that the .save() method won't save session cookies anyway, unless |
| you ask otherwise by passing a true ignore_discard argument. |
| |
| """ |
| self._cookies_lock.acquire() |
| for cookie in self: |
| if cookie.discard: |
| self.clear(cookie.domain, cookie.path, cookie.name) |
| self._cookies_lock.release() |
| |
| def clear_expired_cookies(self): |
| """Discard all expired cookies. |
| |
| You probably don't need to call this method: expired cookies are never |
| sent back to the server (provided you're using DefaultCookiePolicy), |
| this method is called by CookieJar itself every so often, and the |
| .save() method won't save expired cookies anyway (unless you ask |
| otherwise by passing a true ignore_expires argument). |
| |
| """ |
| self._cookies_lock.acquire() |
| now = time.time() |
| for cookie in self: |
| if cookie.is_expired(now): |
| self.clear(cookie.domain, cookie.path, cookie.name) |
| self._cookies_lock.release() |
| |
| def __iter__(self): |
| return deepvalues(self._cookies) |
| |
| def __len__(self): |
| """Return number of contained cookies.""" |
| i = 0 |
| for cookie in self: i = i + 1 |
| return i |
| |
| def __repr__(self): |
| r = [] |
| for cookie in self: r.append(repr(cookie)) |
| return "<%s[%s]>" % (self.__class__, ", ".join(r)) |
| |
| def __str__(self): |
| r = [] |
| for cookie in self: r.append(str(cookie)) |
| return "<%s[%s]>" % (self.__class__, ", ".join(r)) |
| |
| |
| class LoadError(Exception): pass |
| |
| class FileCookieJar(CookieJar): |
| """CookieJar that can be loaded from and saved to a file.""" |
| |
| def __init__(self, filename=None, delayload=False, policy=None): |
| """ |
| Cookies are NOT loaded from the named file until either the .load() or |
| .revert() method is called. |
| |
| """ |
| CookieJar.__init__(self, policy) |
| if filename is not None: |
| try: |
| filename+"" |
| except: |
| raise ValueError("filename must be string-like") |
| self.filename = filename |
| self.delayload = bool(delayload) |
| |
| def save(self, filename=None, ignore_discard=False, ignore_expires=False): |
| """Save cookies to a file.""" |
| raise NotImplementedError() |
| |
| def load(self, filename=None, ignore_discard=False, ignore_expires=False): |
| """Load cookies from a file.""" |
| if filename is None: |
| if self.filename is not None: filename = self.filename |
| else: raise ValueError(MISSING_FILENAME_TEXT) |
| |
| f = open(filename) |
| try: |
| self._really_load(f, filename, ignore_discard, ignore_expires) |
| finally: |
| f.close() |
| |
| def revert(self, filename=None, |
| ignore_discard=False, ignore_expires=False): |
| """Clear all cookies and reload cookies from a saved file. |
| |
| Raises LoadError (or IOError) if reversion is not successful; the |
| object's state will not be altered if this happens. |
| |
| """ |
| if filename is None: |
| if self.filename is not None: filename = self.filename |
| else: raise ValueError(MISSING_FILENAME_TEXT) |
| |
| self._cookies_lock.acquire() |
| |
| old_state = copy.deepcopy(self._cookies) |
| self._cookies = {} |
| try: |
| self.load(filename, ignore_discard, ignore_expires) |
| except (LoadError, IOError): |
| self._cookies = old_state |
| raise |
| |
| self._cookies_lock.release() |
| |
| from _LWPCookieJar import LWPCookieJar, lwp_cookie_str |
| from _MozillaCookieJar import MozillaCookieJar |