blob: 0a083b859d1a0c9d2617d6ca615288af4346969f [file] [log] [blame]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001"""An extensible library for opening URLs using a variety of protocols
2
3The simplest way to use this module is to call the urlopen function,
4which accepts a string containing a URL or a Request object (described
5below). It opens the URL and returns the results as file-like
6object; the returned object has some extra methods described below.
7
8The OpenerDirector manages a collection of Handler objects that do
9all the actual work. Each Handler implements a particular protocol or
10option. The OpenerDirector is a composite object that invokes the
11Handlers needed to open the requested URL. For example, the
12HTTPHandler performs HTTP GET and POST requests and deals with
13non-error returns. The HTTPRedirectHandler automatically deals with
14HTTP 301, 302, 303 and 307 redirect errors, and the HTTPDigestAuthHandler
15deals with digest authentication.
16
17urlopen(url, data=None) -- Basic usage is the same as original
18urllib. pass the url and optionally data to post to an HTTP URL, and
19get a file-like object back. One difference is that you can also pass
20a Request instance instead of URL. Raises a URLError (subclass of
21IOError); for HTTP errors, raises an HTTPError, which can also be
22treated as a valid response.
23
24build_opener -- Function that creates a new OpenerDirector instance.
25Will install the default handlers. Accepts one or more Handlers as
26arguments, either instances or Handler classes that it will
27instantiate. If one of the argument is a subclass of the default
28handler, the argument will be installed instead of the default.
29
30install_opener -- Installs a new opener as the default opener.
31
32objects of interest:
Senthil Kumaran1107c5d2009-11-15 06:20:55 +000033
Senthil Kumaran47fff872009-12-20 07:10:31 +000034OpenerDirector -- Sets up the User Agent as the Python-urllib client and manages
35the Handler classes, while dealing with requests and responses.
Jeremy Hylton1afc1692008-06-18 20:49:58 +000036
37Request -- An object that encapsulates the state of a request. The
38state can be as simple as the URL. It can also include extra HTTP
39headers, e.g. a User-Agent.
40
41BaseHandler --
42
43internals:
44BaseHandler and parent
45_call_chain conventions
46
47Example usage:
48
Georg Brandl029986a2008-06-23 11:44:14 +000049import urllib.request
Jeremy Hylton1afc1692008-06-18 20:49:58 +000050
51# set up authentication info
Georg Brandl029986a2008-06-23 11:44:14 +000052authinfo = urllib.request.HTTPBasicAuthHandler()
Jeremy Hylton1afc1692008-06-18 20:49:58 +000053authinfo.add_password(realm='PDQ Application',
54 uri='https://mahler:8092/site-updates.py',
55 user='klem',
56 passwd='geheim$parole')
57
Georg Brandl029986a2008-06-23 11:44:14 +000058proxy_support = urllib.request.ProxyHandler({"http" : "http://ahad-haam:3128"})
Jeremy Hylton1afc1692008-06-18 20:49:58 +000059
60# build a new opener that adds authentication and caching FTP handlers
Georg Brandl029986a2008-06-23 11:44:14 +000061opener = urllib.request.build_opener(proxy_support, authinfo,
62 urllib.request.CacheFTPHandler)
Jeremy Hylton1afc1692008-06-18 20:49:58 +000063
64# install it
Georg Brandl029986a2008-06-23 11:44:14 +000065urllib.request.install_opener(opener)
Jeremy Hylton1afc1692008-06-18 20:49:58 +000066
Georg Brandl029986a2008-06-23 11:44:14 +000067f = urllib.request.urlopen('http://www.python.org/')
Jeremy Hylton1afc1692008-06-18 20:49:58 +000068"""
69
70# XXX issues:
71# If an authentication error handler that tries to perform
72# authentication for some reason but fails, how should the error be
73# signalled? The client needs to know the HTTP error code. But if
74# the handler knows that the problem was, e.g., that it didn't know
75# that hash algo that requested in the challenge, it would be good to
76# pass that information along to the client, too.
77# ftp errors aren't handled cleanly
78# check digest against correct (i.e. non-apache) implementation
79
80# Possible extensions:
81# complex proxies XXX not sure what exactly was meant by this
82# abstract factory for opener
83
84import base64
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +000085import bisect
Jeremy Hylton1afc1692008-06-18 20:49:58 +000086import email
87import hashlib
88import http.client
89import io
90import os
91import posixpath
92import random
93import re
94import socket
95import sys
96import time
Jeremy Hylton1afc1692008-06-18 20:49:58 +000097
Georg Brandl13e89462008-07-01 19:56:00 +000098from urllib.error import URLError, HTTPError, ContentTooShortError
99from urllib.parse import (
100 urlparse, urlsplit, urljoin, unwrap, quote, unquote,
101 splittype, splithost, splitport, splituser, splitpasswd,
Senthil Kumarand95cc752010-08-08 11:27:53 +0000102 splitattr, splitquery, splitvalue, splittag, to_bytes, urlunparse)
Georg Brandl13e89462008-07-01 19:56:00 +0000103from urllib.response import addinfourl, addclosehook
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000104
105# check for SSL
106try:
107 import ssl
108except:
109 _have_ssl = False
110else:
111 _have_ssl = True
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000112
113# used in User-Agent header sent
114__version__ = sys.version[:3]
115
116_opener = None
117def urlopen(url, data=None, timeout=socket._GLOBAL_DEFAULT_TIMEOUT):
118 global _opener
119 if _opener is None:
120 _opener = build_opener()
121 return _opener.open(url, data, timeout)
122
123def install_opener(opener):
124 global _opener
125 _opener = opener
126
127# TODO(jhylton): Make this work with the same global opener.
128_urlopener = None
129def urlretrieve(url, filename=None, reporthook=None, data=None):
130 global _urlopener
131 if not _urlopener:
132 _urlopener = FancyURLopener()
133 return _urlopener.retrieve(url, filename, reporthook, data)
134
135def urlcleanup():
136 if _urlopener:
137 _urlopener.cleanup()
138 global _opener
139 if _opener:
140 _opener = None
141
142# copied from cookielib.py
Antoine Pitroufd036452008-08-19 17:56:33 +0000143_cut_port_re = re.compile(r":\d+$", re.ASCII)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000144def request_host(request):
145 """Return request-host, as defined by RFC 2965.
146
147 Variation from RFC: returned value is lowercased, for convenient
148 comparison.
149
150 """
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000151 url = request.full_url
Georg Brandl13e89462008-07-01 19:56:00 +0000152 host = urlparse(url)[1]
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000153 if host == "":
154 host = request.get_header("Host", "")
155
156 # remove port, if present
157 host = _cut_port_re.sub("", host, 1)
158 return host.lower()
159
160class Request:
161
162 def __init__(self, url, data=None, headers={},
163 origin_req_host=None, unverifiable=False):
164 # unwrap('<URL:type://host/path>') --> 'type://host/path'
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000165 self.full_url = unwrap(url)
Senthil Kumarand95cc752010-08-08 11:27:53 +0000166 self.full_url, fragment = splittag(self.full_url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000167 self.data = data
168 self.headers = {}
Senthil Kumaran97f0c6b2009-07-25 04:24:38 +0000169 self._tunnel_host = None
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000170 for key, value in headers.items():
171 self.add_header(key, value)
172 self.unredirected_hdrs = {}
173 if origin_req_host is None:
174 origin_req_host = request_host(self)
175 self.origin_req_host = origin_req_host
176 self.unverifiable = unverifiable
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000177 self._parse()
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000178
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000179 def _parse(self):
180 self.type, rest = splittype(self.full_url)
181 if self.type is None:
182 raise ValueError("unknown url type: %s" % self.full_url)
183 self.host, self.selector = splithost(rest)
184 if self.host:
185 self.host = unquote(self.host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000186
187 def get_method(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000188 if self.data is not None:
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000189 return "POST"
190 else:
191 return "GET"
192
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000193 # Begin deprecated methods
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000194
195 def add_data(self, data):
196 self.data = data
197
198 def has_data(self):
199 return self.data is not None
200
201 def get_data(self):
202 return self.data
203
204 def get_full_url(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000205 return self.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000206
207 def get_type(self):
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000208 return self.type
209
210 def get_host(self):
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000211 return self.host
212
213 def get_selector(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000214 return self.selector
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000215
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000216 def is_unverifiable(self):
217 return self.unverifiable
Facundo Batista72dc1ea2008-08-16 14:44:32 +0000218
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000219 def get_origin_req_host(self):
220 return self.origin_req_host
221
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000222 # End deprecated methods
223
224 def set_proxy(self, host, type):
Senthil Kumaran97f0c6b2009-07-25 04:24:38 +0000225 if self.type == 'https' and not self._tunnel_host:
226 self._tunnel_host = self.host
227 else:
228 self.type= type
229 self.selector = self.full_url
230 self.host = host
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000231
232 def has_proxy(self):
233 return self.selector == self.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000234
235 def add_header(self, key, val):
236 # useful for something like authentication
237 self.headers[key.capitalize()] = val
238
239 def add_unredirected_header(self, key, val):
240 # will not be added to a redirected request
241 self.unredirected_hdrs[key.capitalize()] = val
242
243 def has_header(self, header_name):
244 return (header_name in self.headers or
245 header_name in self.unredirected_hdrs)
246
247 def get_header(self, header_name, default=None):
248 return self.headers.get(
249 header_name,
250 self.unredirected_hdrs.get(header_name, default))
251
252 def header_items(self):
253 hdrs = self.unredirected_hdrs.copy()
254 hdrs.update(self.headers)
255 return list(hdrs.items())
256
257class OpenerDirector:
258 def __init__(self):
259 client_version = "Python-urllib/%s" % __version__
260 self.addheaders = [('User-agent', client_version)]
261 # manage the individual handlers
262 self.handlers = []
263 self.handle_open = {}
264 self.handle_error = {}
265 self.process_response = {}
266 self.process_request = {}
267
268 def add_handler(self, handler):
269 if not hasattr(handler, "add_parent"):
270 raise TypeError("expected BaseHandler instance, got %r" %
271 type(handler))
272
273 added = False
274 for meth in dir(handler):
275 if meth in ["redirect_request", "do_open", "proxy_open"]:
276 # oops, coincidental match
277 continue
278
279 i = meth.find("_")
280 protocol = meth[:i]
281 condition = meth[i+1:]
282
283 if condition.startswith("error"):
284 j = condition.find("_") + i + 1
285 kind = meth[j+1:]
286 try:
287 kind = int(kind)
288 except ValueError:
289 pass
290 lookup = self.handle_error.get(protocol, {})
291 self.handle_error[protocol] = lookup
292 elif condition == "open":
293 kind = protocol
294 lookup = self.handle_open
295 elif condition == "response":
296 kind = protocol
297 lookup = self.process_response
298 elif condition == "request":
299 kind = protocol
300 lookup = self.process_request
301 else:
302 continue
303
304 handlers = lookup.setdefault(kind, [])
305 if handlers:
306 bisect.insort(handlers, handler)
307 else:
308 handlers.append(handler)
309 added = True
310
311 if added:
312 # the handlers must work in an specific order, the order
313 # is specified in a Handler attribute
314 bisect.insort(self.handlers, handler)
315 handler.add_parent(self)
316
317 def close(self):
318 # Only exists for backwards compatibility.
319 pass
320
321 def _call_chain(self, chain, kind, meth_name, *args):
322 # Handlers raise an exception if no one else should try to handle
323 # the request, or return None if they can't but another handler
324 # could. Otherwise, they return the response.
325 handlers = chain.get(kind, ())
326 for handler in handlers:
327 func = getattr(handler, meth_name)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000328 result = func(*args)
329 if result is not None:
330 return result
331
332 def open(self, fullurl, data=None, timeout=socket._GLOBAL_DEFAULT_TIMEOUT):
333 # accept a URL or a Request object
334 if isinstance(fullurl, str):
335 req = Request(fullurl, data)
336 else:
337 req = fullurl
338 if data is not None:
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000339 req.data = data
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000340
341 req.timeout = timeout
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000342 protocol = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000343
344 # pre-process request
345 meth_name = protocol+"_request"
346 for processor in self.process_request.get(protocol, []):
347 meth = getattr(processor, meth_name)
348 req = meth(req)
349
350 response = self._open(req, data)
351
352 # post-process response
353 meth_name = protocol+"_response"
354 for processor in self.process_response.get(protocol, []):
355 meth = getattr(processor, meth_name)
356 response = meth(req, response)
357
358 return response
359
360 def _open(self, req, data=None):
361 result = self._call_chain(self.handle_open, 'default',
362 'default_open', req)
363 if result:
364 return result
365
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000366 protocol = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000367 result = self._call_chain(self.handle_open, protocol, protocol +
368 '_open', req)
369 if result:
370 return result
371
372 return self._call_chain(self.handle_open, 'unknown',
373 'unknown_open', req)
374
375 def error(self, proto, *args):
376 if proto in ('http', 'https'):
377 # XXX http[s] protocols are special-cased
378 dict = self.handle_error['http'] # https is not different than http
379 proto = args[2] # YUCK!
380 meth_name = 'http_error_%s' % proto
381 http_err = 1
382 orig_args = args
383 else:
384 dict = self.handle_error
385 meth_name = proto + '_error'
386 http_err = 0
387 args = (dict, proto, meth_name) + args
388 result = self._call_chain(*args)
389 if result:
390 return result
391
392 if http_err:
393 args = (dict, 'default', 'http_error_default') + orig_args
394 return self._call_chain(*args)
395
396# XXX probably also want an abstract factory that knows when it makes
397# sense to skip a superclass in favor of a subclass and when it might
398# make sense to include both
399
400def build_opener(*handlers):
401 """Create an opener object from a list of handlers.
402
403 The opener will use several default handlers, including support
Senthil Kumaran1107c5d2009-11-15 06:20:55 +0000404 for HTTP, FTP and when applicable HTTPS.
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000405
406 If any of the handlers passed as arguments are subclasses of the
407 default handlers, the default handlers will not be used.
408 """
409 def isclass(obj):
410 return isinstance(obj, type) or hasattr(obj, "__bases__")
411
412 opener = OpenerDirector()
413 default_classes = [ProxyHandler, UnknownHandler, HTTPHandler,
414 HTTPDefaultErrorHandler, HTTPRedirectHandler,
415 FTPHandler, FileHandler, HTTPErrorProcessor]
416 if hasattr(http.client, "HTTPSConnection"):
417 default_classes.append(HTTPSHandler)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000418 skip = set()
419 for klass in default_classes:
420 for check in handlers:
421 if isclass(check):
422 if issubclass(check, klass):
423 skip.add(klass)
424 elif isinstance(check, klass):
425 skip.add(klass)
426 for klass in skip:
427 default_classes.remove(klass)
428
429 for klass in default_classes:
430 opener.add_handler(klass())
431
432 for h in handlers:
433 if isclass(h):
434 h = h()
435 opener.add_handler(h)
436 return opener
437
438class BaseHandler:
439 handler_order = 500
440
441 def add_parent(self, parent):
442 self.parent = parent
443
444 def close(self):
445 # Only exists for backwards compatibility
446 pass
447
448 def __lt__(self, other):
449 if not hasattr(other, "handler_order"):
450 # Try to preserve the old behavior of having custom classes
451 # inserted after default ones (works only for custom user
452 # classes which are not aware of handler_order).
453 return True
454 return self.handler_order < other.handler_order
455
456
457class HTTPErrorProcessor(BaseHandler):
458 """Process HTTP error responses."""
459 handler_order = 1000 # after all other processing
460
461 def http_response(self, request, response):
462 code, msg, hdrs = response.code, response.msg, response.info()
463
464 # According to RFC 2616, "2xx" code indicates that the client's
465 # request was successfully received, understood, and accepted.
466 if not (200 <= code < 300):
467 response = self.parent.error(
468 'http', request, response, code, msg, hdrs)
469
470 return response
471
472 https_response = http_response
473
474class HTTPDefaultErrorHandler(BaseHandler):
475 def http_error_default(self, req, fp, code, msg, hdrs):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000476 raise HTTPError(req.full_url, code, msg, hdrs, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000477
478class HTTPRedirectHandler(BaseHandler):
479 # maximum number of redirections to any single URL
480 # this is needed because of the state that cookies introduce
481 max_repeats = 4
482 # maximum total number of redirections (regardless of URL) before
483 # assuming we're in a loop
484 max_redirections = 10
485
486 def redirect_request(self, req, fp, code, msg, headers, newurl):
487 """Return a Request or None in response to a redirect.
488
489 This is called by the http_error_30x methods when a
490 redirection response is received. If a redirection should
491 take place, return a new Request to allow http_error_30x to
492 perform the redirect. Otherwise, raise HTTPError if no-one
493 else should try to handle this url. Return None if you can't
494 but another Handler might.
495 """
496 m = req.get_method()
497 if (not (code in (301, 302, 303, 307) and m in ("GET", "HEAD")
498 or code in (301, 302, 303) and m == "POST")):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000499 raise HTTPError(req.full_url, code, msg, headers, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000500
501 # Strictly (according to RFC 2616), 301 or 302 in response to
502 # a POST MUST NOT cause a redirection without confirmation
Georg Brandl029986a2008-06-23 11:44:14 +0000503 # from the user (of urllib.request, in this case). In practice,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000504 # essentially all clients do redirect in this case, so we do
505 # the same.
506 # be conciliant with URIs containing a space
507 newurl = newurl.replace(' ', '%20')
508 CONTENT_HEADERS = ("content-length", "content-type")
509 newheaders = dict((k, v) for k, v in req.headers.items()
510 if k.lower() not in CONTENT_HEADERS)
511 return Request(newurl,
512 headers=newheaders,
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000513 origin_req_host=req.origin_req_host,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000514 unverifiable=True)
515
516 # Implementation note: To avoid the server sending us into an
517 # infinite loop, the request object needs to track what URLs we
518 # have already seen. Do this by adding a handler-specific
519 # attribute to the Request object.
520 def http_error_302(self, req, fp, code, msg, headers):
521 # Some servers (incorrectly) return multiple Location headers
522 # (so probably same goes for URI). Use first header.
523 if "location" in headers:
524 newurl = headers["location"]
525 elif "uri" in headers:
526 newurl = headers["uri"]
527 else:
528 return
Facundo Batistaf24802c2008-08-17 03:36:03 +0000529
530 # fix a possible malformed URL
531 urlparts = urlparse(newurl)
532 if not urlparts.path:
533 urlparts = list(urlparts)
534 urlparts[2] = "/"
535 newurl = urlunparse(urlparts)
536
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000537 newurl = urljoin(req.full_url, newurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000538
539 # XXX Probably want to forget about the state of the current
540 # request, although that might interact poorly with other
541 # handlers that also use handler-specific request attributes
542 new = self.redirect_request(req, fp, code, msg, headers, newurl)
543 if new is None:
544 return
545
546 # loop detection
547 # .redirect_dict has a key url if url was previously visited.
548 if hasattr(req, 'redirect_dict'):
549 visited = new.redirect_dict = req.redirect_dict
550 if (visited.get(newurl, 0) >= self.max_repeats or
551 len(visited) >= self.max_redirections):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000552 raise HTTPError(req.full_url, code,
Georg Brandl13e89462008-07-01 19:56:00 +0000553 self.inf_msg + msg, headers, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000554 else:
555 visited = new.redirect_dict = req.redirect_dict = {}
556 visited[newurl] = visited.get(newurl, 0) + 1
557
558 # Don't close the fp until we are sure that we won't use it
559 # with HTTPError.
560 fp.read()
561 fp.close()
562
Senthil Kumaranfb8cc2f2009-07-19 02:44:19 +0000563 return self.parent.open(new, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000564
565 http_error_301 = http_error_303 = http_error_307 = http_error_302
566
567 inf_msg = "The HTTP server returned a redirect error that would " \
568 "lead to an infinite loop.\n" \
569 "The last 30x error message was:\n"
570
571
572def _parse_proxy(proxy):
573 """Return (scheme, user, password, host/port) given a URL or an authority.
574
575 If a URL is supplied, it must have an authority (host:port) component.
576 According to RFC 3986, having an authority component means the URL must
577 have two slashes after the scheme:
578
579 >>> _parse_proxy('file:/ftp.example.com/')
580 Traceback (most recent call last):
581 ValueError: proxy URL with no authority: 'file:/ftp.example.com/'
582
583 The first three items of the returned tuple may be None.
584
585 Examples of authority parsing:
586
587 >>> _parse_proxy('proxy.example.com')
588 (None, None, None, 'proxy.example.com')
589 >>> _parse_proxy('proxy.example.com:3128')
590 (None, None, None, 'proxy.example.com:3128')
591
592 The authority component may optionally include userinfo (assumed to be
593 username:password):
594
595 >>> _parse_proxy('joe:password@proxy.example.com')
596 (None, 'joe', 'password', 'proxy.example.com')
597 >>> _parse_proxy('joe:password@proxy.example.com:3128')
598 (None, 'joe', 'password', 'proxy.example.com:3128')
599
600 Same examples, but with URLs instead:
601
602 >>> _parse_proxy('http://proxy.example.com/')
603 ('http', None, None, 'proxy.example.com')
604 >>> _parse_proxy('http://proxy.example.com:3128/')
605 ('http', None, None, 'proxy.example.com:3128')
606 >>> _parse_proxy('http://joe:password@proxy.example.com/')
607 ('http', 'joe', 'password', 'proxy.example.com')
608 >>> _parse_proxy('http://joe:password@proxy.example.com:3128')
609 ('http', 'joe', 'password', 'proxy.example.com:3128')
610
611 Everything after the authority is ignored:
612
613 >>> _parse_proxy('ftp://joe:password@proxy.example.com/rubbish:3128')
614 ('ftp', 'joe', 'password', 'proxy.example.com')
615
616 Test for no trailing '/' case:
617
618 >>> _parse_proxy('http://joe:password@proxy.example.com')
619 ('http', 'joe', 'password', 'proxy.example.com')
620
621 """
Georg Brandl13e89462008-07-01 19:56:00 +0000622 scheme, r_scheme = splittype(proxy)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000623 if not r_scheme.startswith("/"):
624 # authority
625 scheme = None
626 authority = proxy
627 else:
628 # URL
629 if not r_scheme.startswith("//"):
630 raise ValueError("proxy URL with no authority: %r" % proxy)
631 # We have an authority, so for RFC 3986-compliant URLs (by ss 3.
632 # and 3.3.), path is empty or starts with '/'
633 end = r_scheme.find("/", 2)
634 if end == -1:
635 end = None
636 authority = r_scheme[2:end]
Georg Brandl13e89462008-07-01 19:56:00 +0000637 userinfo, hostport = splituser(authority)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000638 if userinfo is not None:
Georg Brandl13e89462008-07-01 19:56:00 +0000639 user, password = splitpasswd(userinfo)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000640 else:
641 user = password = None
642 return scheme, user, password, hostport
643
644class ProxyHandler(BaseHandler):
645 # Proxies must be in front
646 handler_order = 100
647
648 def __init__(self, proxies=None):
649 if proxies is None:
650 proxies = getproxies()
651 assert hasattr(proxies, 'keys'), "proxies must be a mapping"
652 self.proxies = proxies
653 for type, url in proxies.items():
654 setattr(self, '%s_open' % type,
655 lambda r, proxy=url, type=type, meth=self.proxy_open: \
656 meth(r, proxy, type))
657
658 def proxy_open(self, req, proxy, type):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000659 orig_type = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000660 proxy_type, user, password, hostport = _parse_proxy(proxy)
661 if proxy_type is None:
662 proxy_type = orig_type
Senthil Kumaran7bb04972009-10-11 04:58:55 +0000663
664 if req.host and proxy_bypass(req.host):
665 return None
666
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000667 if user and password:
Georg Brandl13e89462008-07-01 19:56:00 +0000668 user_pass = '%s:%s' % (unquote(user),
669 unquote(password))
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000670 creds = base64.b64encode(user_pass.encode()).decode("ascii")
671 req.add_header('Proxy-authorization', 'Basic ' + creds)
Georg Brandl13e89462008-07-01 19:56:00 +0000672 hostport = unquote(hostport)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000673 req.set_proxy(hostport, proxy_type)
Senthil Kumaran97f0c6b2009-07-25 04:24:38 +0000674 if orig_type == proxy_type or orig_type == 'https':
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000675 # let other handlers take care of it
676 return None
677 else:
678 # need to start over, because the other handlers don't
679 # grok the proxy's URL type
680 # e.g. if we have a constructor arg proxies like so:
681 # {'http': 'ftp://proxy.example.com'}, we may end up turning
682 # a request for http://acme.example.com/a into one for
683 # ftp://proxy.example.com/a
Senthil Kumaranfb8cc2f2009-07-19 02:44:19 +0000684 return self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000685
686class HTTPPasswordMgr:
687
688 def __init__(self):
689 self.passwd = {}
690
691 def add_password(self, realm, uri, user, passwd):
692 # uri could be a single URI or a sequence
693 if isinstance(uri, str):
694 uri = [uri]
695 if not realm in self.passwd:
696 self.passwd[realm] = {}
697 for default_port in True, False:
698 reduced_uri = tuple(
699 [self.reduce_uri(u, default_port) for u in uri])
700 self.passwd[realm][reduced_uri] = (user, passwd)
701
702 def find_user_password(self, realm, authuri):
703 domains = self.passwd.get(realm, {})
704 for default_port in True, False:
705 reduced_authuri = self.reduce_uri(authuri, default_port)
706 for uris, authinfo in domains.items():
707 for uri in uris:
708 if self.is_suburi(uri, reduced_authuri):
709 return authinfo
710 return None, None
711
712 def reduce_uri(self, uri, default_port=True):
713 """Accept authority or URI and extract only the authority and path."""
714 # note HTTP URLs do not have a userinfo component
Georg Brandl13e89462008-07-01 19:56:00 +0000715 parts = urlsplit(uri)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000716 if parts[1]:
717 # URI
718 scheme = parts[0]
719 authority = parts[1]
720 path = parts[2] or '/'
721 else:
722 # host or host:port
723 scheme = None
724 authority = uri
725 path = '/'
Georg Brandl13e89462008-07-01 19:56:00 +0000726 host, port = splitport(authority)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000727 if default_port and port is None and scheme is not None:
728 dport = {"http": 80,
729 "https": 443,
730 }.get(scheme)
731 if dport is not None:
732 authority = "%s:%d" % (host, dport)
733 return authority, path
734
735 def is_suburi(self, base, test):
736 """Check if test is below base in a URI tree
737
738 Both args must be URIs in reduced form.
739 """
740 if base == test:
741 return True
742 if base[0] != test[0]:
743 return False
744 common = posixpath.commonprefix((base[1], test[1]))
745 if len(common) == len(base[1]):
746 return True
747 return False
748
749
750class HTTPPasswordMgrWithDefaultRealm(HTTPPasswordMgr):
751
752 def find_user_password(self, realm, authuri):
753 user, password = HTTPPasswordMgr.find_user_password(self, realm,
754 authuri)
755 if user is not None:
756 return user, password
757 return HTTPPasswordMgr.find_user_password(self, None, authuri)
758
759
760class AbstractBasicAuthHandler:
761
762 # XXX this allows for multiple auth-schemes, but will stupidly pick
763 # the last one with a realm specified.
764
765 # allow for double- and single-quoted realm values
766 # (single quotes are a violation of the RFC, but appear in the wild)
767 rx = re.compile('(?:.*,)*[ \t]*([^ \t]+)[ \t]+'
768 'realm=(["\'])(.*?)\\2', re.I)
769
770 # XXX could pre-emptively send auth info already accepted (RFC 2617,
771 # end of section 2, and section 1.2 immediately after "credentials"
772 # production).
773
774 def __init__(self, password_mgr=None):
775 if password_mgr is None:
776 password_mgr = HTTPPasswordMgr()
777 self.passwd = password_mgr
778 self.add_password = self.passwd.add_password
Senthil Kumaranf4998ac2010-06-01 12:53:48 +0000779 self.retried = 0
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000780
781 def http_error_auth_reqed(self, authreq, host, req, headers):
782 # host may be an authority (without userinfo) or a URL with an
783 # authority
784 # XXX could be multiple headers
785 authreq = headers.get(authreq, None)
Senthil Kumaranf4998ac2010-06-01 12:53:48 +0000786
787 if self.retried > 5:
788 # retry sending the username:password 5 times before failing.
789 raise HTTPError(req.get_full_url(), 401, "basic auth failed",
790 headers, None)
791 else:
792 self.retried += 1
793
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000794 if authreq:
795 mo = AbstractBasicAuthHandler.rx.search(authreq)
796 if mo:
797 scheme, quote, realm = mo.groups()
798 if scheme.lower() == 'basic':
799 return self.retry_http_basic_auth(host, req, realm)
800
801 def retry_http_basic_auth(self, host, req, realm):
802 user, pw = self.passwd.find_user_password(realm, host)
803 if pw is not None:
804 raw = "%s:%s" % (user, pw)
805 auth = "Basic " + base64.b64encode(raw.encode()).decode("ascii")
806 if req.headers.get(self.auth_header, None) == auth:
807 return None
Senthil Kumaranca2fc9e2010-02-24 16:53:16 +0000808 req.add_unredirected_header(self.auth_header, auth)
Senthil Kumaranfb8cc2f2009-07-19 02:44:19 +0000809 return self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000810 else:
811 return None
812
813
814class HTTPBasicAuthHandler(AbstractBasicAuthHandler, BaseHandler):
815
816 auth_header = 'Authorization'
817
818 def http_error_401(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000819 url = req.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000820 return self.http_error_auth_reqed('www-authenticate',
821 url, req, headers)
822
823
824class ProxyBasicAuthHandler(AbstractBasicAuthHandler, BaseHandler):
825
826 auth_header = 'Proxy-authorization'
827
828 def http_error_407(self, req, fp, code, msg, headers):
829 # http_error_auth_reqed requires that there is no userinfo component in
Georg Brandl029986a2008-06-23 11:44:14 +0000830 # authority. Assume there isn't one, since urllib.request does not (and
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000831 # should not, RFC 3986 s. 3.2.1) support requests for URLs containing
832 # userinfo.
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000833 authority = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000834 return self.http_error_auth_reqed('proxy-authenticate',
835 authority, req, headers)
836
837
838def randombytes(n):
839 """Return n random bytes."""
840 return os.urandom(n)
841
842class AbstractDigestAuthHandler:
843 # Digest authentication is specified in RFC 2617.
844
845 # XXX The client does not inspect the Authentication-Info header
846 # in a successful response.
847
848 # XXX It should be possible to test this implementation against
849 # a mock server that just generates a static set of challenges.
850
851 # XXX qop="auth-int" supports is shaky
852
853 def __init__(self, passwd=None):
854 if passwd is None:
855 passwd = HTTPPasswordMgr()
856 self.passwd = passwd
857 self.add_password = self.passwd.add_password
858 self.retried = 0
859 self.nonce_count = 0
Senthil Kumaran4c7eaee2009-11-15 08:43:45 +0000860 self.last_nonce = None
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000861
862 def reset_retry_count(self):
863 self.retried = 0
864
865 def http_error_auth_reqed(self, auth_header, host, req, headers):
866 authreq = headers.get(auth_header, None)
867 if self.retried > 5:
868 # Don't fail endlessly - if we failed once, we'll probably
869 # fail a second time. Hm. Unless the Password Manager is
870 # prompting for the information. Crap. This isn't great
871 # but it's better than the current 'repeat until recursion
872 # depth exceeded' approach <wink>
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000873 raise HTTPError(req.full_url, 401, "digest auth failed",
Georg Brandl13e89462008-07-01 19:56:00 +0000874 headers, None)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000875 else:
876 self.retried += 1
877 if authreq:
878 scheme = authreq.split()[0]
879 if scheme.lower() == 'digest':
880 return self.retry_http_digest_auth(req, authreq)
881
882 def retry_http_digest_auth(self, req, auth):
883 token, challenge = auth.split(' ', 1)
884 chal = parse_keqv_list(filter(None, parse_http_list(challenge)))
885 auth = self.get_authorization(req, chal)
886 if auth:
887 auth_val = 'Digest %s' % auth
888 if req.headers.get(self.auth_header, None) == auth_val:
889 return None
890 req.add_unredirected_header(self.auth_header, auth_val)
Senthil Kumaranfb8cc2f2009-07-19 02:44:19 +0000891 resp = self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000892 return resp
893
894 def get_cnonce(self, nonce):
895 # The cnonce-value is an opaque
896 # quoted string value provided by the client and used by both client
897 # and server to avoid chosen plaintext attacks, to provide mutual
898 # authentication, and to provide some message integrity protection.
899 # This isn't a fabulous effort, but it's probably Good Enough.
900 s = "%s:%s:%s:" % (self.nonce_count, nonce, time.ctime())
901 b = s.encode("ascii") + randombytes(8)
902 dig = hashlib.sha1(b).hexdigest()
903 return dig[:16]
904
905 def get_authorization(self, req, chal):
906 try:
907 realm = chal['realm']
908 nonce = chal['nonce']
909 qop = chal.get('qop')
910 algorithm = chal.get('algorithm', 'MD5')
911 # mod_digest doesn't send an opaque, even though it isn't
912 # supposed to be optional
913 opaque = chal.get('opaque', None)
914 except KeyError:
915 return None
916
917 H, KD = self.get_algorithm_impls(algorithm)
918 if H is None:
919 return None
920
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000921 user, pw = self.passwd.find_user_password(realm, req.full_url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000922 if user is None:
923 return None
924
925 # XXX not implemented yet
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000926 if req.data is not None:
927 entdig = self.get_entity_digest(req.data, chal)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000928 else:
929 entdig = None
930
931 A1 = "%s:%s:%s" % (user, realm, pw)
932 A2 = "%s:%s" % (req.get_method(),
933 # XXX selector: what about proxies and full urls
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000934 req.selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000935 if qop == 'auth':
Senthil Kumaran4c7eaee2009-11-15 08:43:45 +0000936 if nonce == self.last_nonce:
937 self.nonce_count += 1
938 else:
939 self.nonce_count = 1
940 self.last_nonce = nonce
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000941 ncvalue = '%08x' % self.nonce_count
942 cnonce = self.get_cnonce(nonce)
943 noncebit = "%s:%s:%s:%s:%s" % (nonce, ncvalue, cnonce, qop, H(A2))
944 respdig = KD(H(A1), noncebit)
945 elif qop is None:
946 respdig = KD(H(A1), "%s:%s" % (nonce, H(A2)))
947 else:
948 # XXX handle auth-int.
Georg Brandl13e89462008-07-01 19:56:00 +0000949 raise URLError("qop '%s' is not supported." % qop)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000950
951 # XXX should the partial digests be encoded too?
952
953 base = 'username="%s", realm="%s", nonce="%s", uri="%s", ' \
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000954 'response="%s"' % (user, realm, nonce, req.selector,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000955 respdig)
956 if opaque:
957 base += ', opaque="%s"' % opaque
958 if entdig:
959 base += ', digest="%s"' % entdig
960 base += ', algorithm="%s"' % algorithm
961 if qop:
962 base += ', qop=auth, nc=%s, cnonce="%s"' % (ncvalue, cnonce)
963 return base
964
965 def get_algorithm_impls(self, algorithm):
966 # lambdas assume digest modules are imported at the top level
967 if algorithm == 'MD5':
968 H = lambda x: hashlib.md5(x.encode("ascii")).hexdigest()
969 elif algorithm == 'SHA':
970 H = lambda x: hashlib.sha1(x.encode("ascii")).hexdigest()
971 # XXX MD5-sess
972 KD = lambda s, d: H("%s:%s" % (s, d))
973 return H, KD
974
975 def get_entity_digest(self, data, chal):
976 # XXX not implemented yet
977 return None
978
979
980class HTTPDigestAuthHandler(BaseHandler, AbstractDigestAuthHandler):
981 """An authentication protocol defined by RFC 2069
982
983 Digest authentication improves on basic authentication because it
984 does not transmit passwords in the clear.
985 """
986
987 auth_header = 'Authorization'
988 handler_order = 490 # before Basic auth
989
990 def http_error_401(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000991 host = urlparse(req.full_url)[1]
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000992 retry = self.http_error_auth_reqed('www-authenticate',
993 host, req, headers)
994 self.reset_retry_count()
995 return retry
996
997
998class ProxyDigestAuthHandler(BaseHandler, AbstractDigestAuthHandler):
999
1000 auth_header = 'Proxy-Authorization'
1001 handler_order = 490 # before Basic auth
1002
1003 def http_error_407(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001004 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001005 retry = self.http_error_auth_reqed('proxy-authenticate',
1006 host, req, headers)
1007 self.reset_retry_count()
1008 return retry
1009
1010class AbstractHTTPHandler(BaseHandler):
1011
1012 def __init__(self, debuglevel=0):
1013 self._debuglevel = debuglevel
1014
1015 def set_http_debuglevel(self, level):
1016 self._debuglevel = level
1017
1018 def do_request_(self, request):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001019 host = request.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001020 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001021 raise URLError('no host given')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001022
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001023 if request.data is not None: # POST
1024 data = request.data
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001025 if not request.has_header('Content-type'):
1026 request.add_unredirected_header(
1027 'Content-type',
1028 'application/x-www-form-urlencoded')
1029 if not request.has_header('Content-length'):
1030 request.add_unredirected_header(
1031 'Content-length', '%d' % len(data))
1032
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001033 sel_host = host
1034 if request.has_proxy():
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001035 scheme, sel = splittype(request.selector)
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001036 sel_host, sel_path = splithost(sel)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001037 if not request.has_header('Host'):
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001038 request.add_unredirected_header('Host', sel_host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001039 for name, value in self.parent.addheaders:
1040 name = name.capitalize()
1041 if not request.has_header(name):
1042 request.add_unredirected_header(name, value)
1043
1044 return request
1045
1046 def do_open(self, http_class, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001047 """Return an HTTPResponse object for the request, using http_class.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001048
1049 http_class must implement the HTTPConnection API from http.client.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001050 """
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001051 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001052 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001053 raise URLError('no host given')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001054
1055 h = http_class(host, timeout=req.timeout) # will parse host:port
1056 headers = dict(req.headers)
1057 headers.update(req.unredirected_hdrs)
1058
1059 # TODO(jhylton): Should this be redesigned to handle
1060 # persistent connections?
1061
1062 # We want to make an HTTP/1.1 request, but the addinfourl
1063 # class isn't prepared to deal with a persistent connection.
1064 # It will try to read all remaining data from the socket,
1065 # which will block while the server waits for the next request.
1066 # So make sure the connection gets closed after the (only)
1067 # request.
1068 headers["Connection"] = "close"
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001069 headers = dict((name.title(), val) for name, val in headers.items())
Senthil Kumaran97f0c6b2009-07-25 04:24:38 +00001070
1071 if req._tunnel_host:
Senthil Kumaran47fff872009-12-20 07:10:31 +00001072 tunnel_headers = {}
1073 proxy_auth_hdr = "Proxy-Authorization"
1074 if proxy_auth_hdr in headers:
1075 tunnel_headers[proxy_auth_hdr] = headers[proxy_auth_hdr]
1076 # Proxy-Authorization should not be sent to origin
1077 # server.
1078 del headers[proxy_auth_hdr]
1079 h.set_tunnel(req._tunnel_host, headers=tunnel_headers)
Senthil Kumaran97f0c6b2009-07-25 04:24:38 +00001080
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001081 try:
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001082 h.request(req.get_method(), req.selector, req.data, headers)
1083 r = h.getresponse() # an HTTPResponse instance
1084 except socket.error as err:
Georg Brandl13e89462008-07-01 19:56:00 +00001085 raise URLError(err)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001086
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001087 r.url = req.full_url
1088 # This line replaces the .msg attribute of the HTTPResponse
1089 # with .headers, because urllib clients expect the response to
1090 # have the reason in .msg. It would be good to mark this
1091 # attribute is deprecated and get then to use info() or
1092 # .headers.
1093 r.msg = r.reason
1094 return r
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001095
1096
1097class HTTPHandler(AbstractHTTPHandler):
1098
1099 def http_open(self, req):
1100 return self.do_open(http.client.HTTPConnection, req)
1101
1102 http_request = AbstractHTTPHandler.do_request_
1103
1104if hasattr(http.client, 'HTTPSConnection'):
1105 class HTTPSHandler(AbstractHTTPHandler):
1106
1107 def https_open(self, req):
1108 return self.do_open(http.client.HTTPSConnection, req)
1109
1110 https_request = AbstractHTTPHandler.do_request_
1111
1112class HTTPCookieProcessor(BaseHandler):
1113 def __init__(self, cookiejar=None):
1114 import http.cookiejar
1115 if cookiejar is None:
1116 cookiejar = http.cookiejar.CookieJar()
1117 self.cookiejar = cookiejar
1118
1119 def http_request(self, request):
1120 self.cookiejar.add_cookie_header(request)
1121 return request
1122
1123 def http_response(self, request, response):
1124 self.cookiejar.extract_cookies(response, request)
1125 return response
1126
1127 https_request = http_request
1128 https_response = http_response
1129
1130class UnknownHandler(BaseHandler):
1131 def unknown_open(self, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001132 type = req.type
Georg Brandl13e89462008-07-01 19:56:00 +00001133 raise URLError('unknown url type: %s' % type)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001134
1135def parse_keqv_list(l):
1136 """Parse list of key=value strings where keys are not duplicated."""
1137 parsed = {}
1138 for elt in l:
1139 k, v = elt.split('=', 1)
1140 if v[0] == '"' and v[-1] == '"':
1141 v = v[1:-1]
1142 parsed[k] = v
1143 return parsed
1144
1145def parse_http_list(s):
1146 """Parse lists as described by RFC 2068 Section 2.
1147
1148 In particular, parse comma-separated lists where the elements of
1149 the list may include quoted-strings. A quoted-string could
1150 contain a comma. A non-quoted string could have quotes in the
1151 middle. Neither commas nor quotes count if they are escaped.
1152 Only double-quotes count, not single-quotes.
1153 """
1154 res = []
1155 part = ''
1156
1157 escape = quote = False
1158 for cur in s:
1159 if escape:
1160 part += cur
1161 escape = False
1162 continue
1163 if quote:
1164 if cur == '\\':
1165 escape = True
1166 continue
1167 elif cur == '"':
1168 quote = False
1169 part += cur
1170 continue
1171
1172 if cur == ',':
1173 res.append(part)
1174 part = ''
1175 continue
1176
1177 if cur == '"':
1178 quote = True
1179
1180 part += cur
1181
1182 # append last part
1183 if part:
1184 res.append(part)
1185
1186 return [part.strip() for part in res]
1187
1188class FileHandler(BaseHandler):
1189 # Use local file or FTP depending on form of URL
1190 def file_open(self, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001191 url = req.selector
Senthil Kumaran2ef16322010-07-11 03:12:43 +00001192 if url[:2] == '//' and url[2:3] != '/' and (req.host and
1193 req.host != 'localhost'):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001194 req.type = 'ftp'
1195 return self.parent.open(req)
1196 else:
1197 return self.open_local_file(req)
1198
1199 # names for the localhost
1200 names = None
1201 def get_names(self):
1202 if FileHandler.names is None:
1203 try:
Senthil Kumaran99b2c8f2009-12-27 10:13:39 +00001204 FileHandler.names = tuple(
1205 socket.gethostbyname_ex('localhost')[2] +
1206 socket.gethostbyname_ex(socket.gethostname())[2])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001207 except socket.gaierror:
1208 FileHandler.names = (socket.gethostbyname('localhost'),)
1209 return FileHandler.names
1210
1211 # not entirely sure what the rules are here
1212 def open_local_file(self, req):
1213 import email.utils
1214 import mimetypes
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001215 host = req.host
Senthil Kumaran06f5a532010-05-08 05:12:05 +00001216 filename = req.selector
1217 localfile = url2pathname(filename)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001218 try:
1219 stats = os.stat(localfile)
1220 size = stats.st_size
1221 modified = email.utils.formatdate(stats.st_mtime, usegmt=True)
Senthil Kumaran06f5a532010-05-08 05:12:05 +00001222 mtype = mimetypes.guess_type(filename)[0]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001223 headers = email.message_from_string(
1224 'Content-type: %s\nContent-length: %d\nLast-modified: %s\n' %
1225 (mtype or 'text/plain', size, modified))
1226 if host:
Georg Brandl13e89462008-07-01 19:56:00 +00001227 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001228 if not host or \
1229 (not port and _safe_gethostbyname(host) in self.get_names()):
Senthil Kumaran06f5a532010-05-08 05:12:05 +00001230 if host:
1231 origurl = 'file://' + host + filename
1232 else:
1233 origurl = 'file://' + filename
1234 return addinfourl(open(localfile, 'rb'), headers, origurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001235 except OSError as msg:
Georg Brandl029986a2008-06-23 11:44:14 +00001236 # users shouldn't expect OSErrors coming from urlopen()
Georg Brandl13e89462008-07-01 19:56:00 +00001237 raise URLError(msg)
1238 raise URLError('file not on local host')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001239
1240def _safe_gethostbyname(host):
1241 try:
1242 return socket.gethostbyname(host)
1243 except socket.gaierror:
1244 return None
1245
1246class FTPHandler(BaseHandler):
1247 def ftp_open(self, req):
1248 import ftplib
1249 import mimetypes
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001250 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001251 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001252 raise URLError('ftp error: no host given')
1253 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001254 if port is None:
1255 port = ftplib.FTP_PORT
1256 else:
1257 port = int(port)
1258
1259 # username/password handling
Georg Brandl13e89462008-07-01 19:56:00 +00001260 user, host = splituser(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001261 if user:
Georg Brandl13e89462008-07-01 19:56:00 +00001262 user, passwd = splitpasswd(user)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001263 else:
1264 passwd = None
Georg Brandl13e89462008-07-01 19:56:00 +00001265 host = unquote(host)
1266 user = unquote(user or '')
1267 passwd = unquote(passwd or '')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001268
1269 try:
1270 host = socket.gethostbyname(host)
1271 except socket.error as msg:
Georg Brandl13e89462008-07-01 19:56:00 +00001272 raise URLError(msg)
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001273 path, attrs = splitattr(req.selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001274 dirs = path.split('/')
Georg Brandl13e89462008-07-01 19:56:00 +00001275 dirs = list(map(unquote, dirs))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001276 dirs, file = dirs[:-1], dirs[-1]
1277 if dirs and not dirs[0]:
1278 dirs = dirs[1:]
1279 try:
1280 fw = self.connect_ftp(user, passwd, host, port, dirs, req.timeout)
1281 type = file and 'I' or 'D'
1282 for attr in attrs:
Georg Brandl13e89462008-07-01 19:56:00 +00001283 attr, value = splitvalue(attr)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001284 if attr.lower() == 'type' and \
1285 value in ('a', 'A', 'i', 'I', 'd', 'D'):
1286 type = value.upper()
1287 fp, retrlen = fw.retrfile(file, type)
1288 headers = ""
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001289 mtype = mimetypes.guess_type(req.full_url)[0]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001290 if mtype:
1291 headers += "Content-type: %s\n" % mtype
1292 if retrlen is not None and retrlen >= 0:
1293 headers += "Content-length: %d\n" % retrlen
1294 headers = email.message_from_string(headers)
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001295 return addinfourl(fp, headers, req.full_url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001296 except ftplib.all_errors as msg:
Georg Brandl13e89462008-07-01 19:56:00 +00001297 exc = URLError('ftp error: %s' % msg)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001298 raise exc.with_traceback(sys.exc_info()[2])
1299
1300 def connect_ftp(self, user, passwd, host, port, dirs, timeout):
1301 fw = ftpwrapper(user, passwd, host, port, dirs, timeout)
1302 return fw
1303
1304class CacheFTPHandler(FTPHandler):
1305 # XXX would be nice to have pluggable cache strategies
1306 # XXX this stuff is definitely not thread safe
1307 def __init__(self):
1308 self.cache = {}
1309 self.timeout = {}
1310 self.soonest = 0
1311 self.delay = 60
1312 self.max_conns = 16
1313
1314 def setTimeout(self, t):
1315 self.delay = t
1316
1317 def setMaxConns(self, m):
1318 self.max_conns = m
1319
1320 def connect_ftp(self, user, passwd, host, port, dirs, timeout):
1321 key = user, host, port, '/'.join(dirs), timeout
1322 if key in self.cache:
1323 self.timeout[key] = time.time() + self.delay
1324 else:
1325 self.cache[key] = ftpwrapper(user, passwd, host, port,
1326 dirs, timeout)
1327 self.timeout[key] = time.time() + self.delay
1328 self.check_cache()
1329 return self.cache[key]
1330
1331 def check_cache(self):
1332 # first check for old ones
1333 t = time.time()
1334 if self.soonest <= t:
1335 for k, v in list(self.timeout.items()):
1336 if v < t:
1337 self.cache[k].close()
1338 del self.cache[k]
1339 del self.timeout[k]
1340 self.soonest = min(list(self.timeout.values()))
1341
1342 # then check the size
1343 if len(self.cache) == self.max_conns:
1344 for k, v in list(self.timeout.items()):
1345 if v == self.soonest:
1346 del self.cache[k]
1347 del self.timeout[k]
1348 break
1349 self.soonest = min(list(self.timeout.values()))
1350
1351# Code move from the old urllib module
1352
1353MAXFTPCACHE = 10 # Trim the ftp cache beyond this size
1354
1355# Helper for non-unix systems
Ronald Oussoren94f25282010-05-05 19:11:21 +00001356if os.name == 'nt':
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001357 from nturl2path import url2pathname, pathname2url
1358else:
1359 def url2pathname(pathname):
1360 """OS-specific conversion from a relative URL of the 'file' scheme
1361 to a file system path; not recommended for general use."""
Georg Brandl13e89462008-07-01 19:56:00 +00001362 return unquote(pathname)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001363
1364 def pathname2url(pathname):
1365 """OS-specific conversion from a file system path to a relative URL
1366 of the 'file' scheme; not recommended for general use."""
Georg Brandl13e89462008-07-01 19:56:00 +00001367 return quote(pathname)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001368
1369# This really consists of two pieces:
1370# (1) a class which handles opening of all sorts of URLs
1371# (plus assorted utilities etc.)
1372# (2) a set of functions for parsing URLs
1373# XXX Should these be separated out into different modules?
1374
1375
1376ftpcache = {}
1377class URLopener:
1378 """Class to open URLs.
1379 This is a class rather than just a subroutine because we may need
1380 more than one set of global protocol-specific options.
1381 Note -- this is a base class for those who don't want the
1382 automatic handling of errors type 302 (relocated) and 401
1383 (authorization needed)."""
1384
1385 __tempfiles = None
1386
1387 version = "Python-urllib/%s" % __version__
1388
1389 # Constructor
1390 def __init__(self, proxies=None, **x509):
1391 if proxies is None:
1392 proxies = getproxies()
1393 assert hasattr(proxies, 'keys'), "proxies must be a mapping"
1394 self.proxies = proxies
1395 self.key_file = x509.get('key_file')
1396 self.cert_file = x509.get('cert_file')
1397 self.addheaders = [('User-Agent', self.version)]
1398 self.__tempfiles = []
1399 self.__unlink = os.unlink # See cleanup()
1400 self.tempcache = None
1401 # Undocumented feature: if you assign {} to tempcache,
1402 # it is used to cache files retrieved with
1403 # self.retrieve(). This is not enabled by default
1404 # since it does not work for changing documents (and I
1405 # haven't got the logic to check expiration headers
1406 # yet).
1407 self.ftpcache = ftpcache
1408 # Undocumented feature: you can use a different
1409 # ftp cache by assigning to the .ftpcache member;
1410 # in case you want logically independent URL openers
1411 # XXX This is not threadsafe. Bah.
1412
1413 def __del__(self):
1414 self.close()
1415
1416 def close(self):
1417 self.cleanup()
1418
1419 def cleanup(self):
1420 # This code sometimes runs when the rest of this module
1421 # has already been deleted, so it can't use any globals
1422 # or import anything.
1423 if self.__tempfiles:
1424 for file in self.__tempfiles:
1425 try:
1426 self.__unlink(file)
1427 except OSError:
1428 pass
1429 del self.__tempfiles[:]
1430 if self.tempcache:
1431 self.tempcache.clear()
1432
1433 def addheader(self, *args):
1434 """Add a header to be used by the HTTP interface only
1435 e.g. u.addheader('Accept', 'sound/basic')"""
1436 self.addheaders.append(args)
1437
1438 # External interface
1439 def open(self, fullurl, data=None):
1440 """Use URLopener().open(file) instead of open(file, 'r')."""
Georg Brandl13e89462008-07-01 19:56:00 +00001441 fullurl = unwrap(to_bytes(fullurl))
Senthil Kumaran734f0592010-02-20 22:19:04 +00001442 fullurl = quote(fullurl, safe="%/:=&?~#+!$,;'@()*[]|")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001443 if self.tempcache and fullurl in self.tempcache:
1444 filename, headers = self.tempcache[fullurl]
1445 fp = open(filename, 'rb')
Georg Brandl13e89462008-07-01 19:56:00 +00001446 return addinfourl(fp, headers, fullurl)
1447 urltype, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001448 if not urltype:
1449 urltype = 'file'
1450 if urltype in self.proxies:
1451 proxy = self.proxies[urltype]
Georg Brandl13e89462008-07-01 19:56:00 +00001452 urltype, proxyhost = splittype(proxy)
1453 host, selector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001454 url = (host, fullurl) # Signal special case to open_*()
1455 else:
1456 proxy = None
1457 name = 'open_' + urltype
1458 self.type = urltype
1459 name = name.replace('-', '_')
1460 if not hasattr(self, name):
1461 if proxy:
1462 return self.open_unknown_proxy(proxy, fullurl, data)
1463 else:
1464 return self.open_unknown(fullurl, data)
1465 try:
1466 if data is None:
1467 return getattr(self, name)(url)
1468 else:
1469 return getattr(self, name)(url, data)
1470 except socket.error as msg:
1471 raise IOError('socket error', msg).with_traceback(sys.exc_info()[2])
1472
1473 def open_unknown(self, fullurl, data=None):
1474 """Overridable interface to open unknown URL type."""
Georg Brandl13e89462008-07-01 19:56:00 +00001475 type, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001476 raise IOError('url error', 'unknown url type', type)
1477
1478 def open_unknown_proxy(self, proxy, fullurl, data=None):
1479 """Overridable interface to open unknown URL type."""
Georg Brandl13e89462008-07-01 19:56:00 +00001480 type, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001481 raise IOError('url error', 'invalid proxy for %s' % type, proxy)
1482
1483 # External interface
1484 def retrieve(self, url, filename=None, reporthook=None, data=None):
1485 """retrieve(url) returns (filename, headers) for a local object
1486 or (tempfilename, headers) for a remote object."""
Georg Brandl13e89462008-07-01 19:56:00 +00001487 url = unwrap(to_bytes(url))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001488 if self.tempcache and url in self.tempcache:
1489 return self.tempcache[url]
Georg Brandl13e89462008-07-01 19:56:00 +00001490 type, url1 = splittype(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001491 if filename is None and (not type or type == 'file'):
1492 try:
1493 fp = self.open_local_file(url1)
1494 hdrs = fp.info()
Philip Jenveycb134d72009-12-03 02:45:01 +00001495 fp.close()
Georg Brandl13e89462008-07-01 19:56:00 +00001496 return url2pathname(splithost(url1)[1]), hdrs
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001497 except IOError as msg:
1498 pass
1499 fp = self.open(url, data)
Benjamin Peterson5f28b7b2009-03-26 21:49:58 +00001500 try:
1501 headers = fp.info()
1502 if filename:
1503 tfp = open(filename, 'wb')
1504 else:
1505 import tempfile
1506 garbage, path = splittype(url)
1507 garbage, path = splithost(path or "")
1508 path, garbage = splitquery(path or "")
1509 path, garbage = splitattr(path or "")
1510 suffix = os.path.splitext(path)[1]
1511 (fd, filename) = tempfile.mkstemp(suffix)
1512 self.__tempfiles.append(filename)
1513 tfp = os.fdopen(fd, 'wb')
1514 try:
1515 result = filename, headers
1516 if self.tempcache is not None:
1517 self.tempcache[url] = result
1518 bs = 1024*8
1519 size = -1
1520 read = 0
1521 blocknum = 0
1522 if reporthook:
1523 if "content-length" in headers:
1524 size = int(headers["Content-Length"])
1525 reporthook(blocknum, bs, size)
1526 while 1:
1527 block = fp.read(bs)
1528 if not block:
1529 break
1530 read += len(block)
1531 tfp.write(block)
1532 blocknum += 1
1533 if reporthook:
1534 reporthook(blocknum, bs, size)
1535 finally:
1536 tfp.close()
1537 finally:
1538 fp.close()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001539
1540 # raise exception if actual size does not match content-length header
1541 if size >= 0 and read < size:
Georg Brandl13e89462008-07-01 19:56:00 +00001542 raise ContentTooShortError(
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001543 "retrieval incomplete: got only %i out of %i bytes"
1544 % (read, size), result)
1545
1546 return result
1547
1548 # Each method named open_<type> knows how to open that type of URL
1549
1550 def _open_generic_http(self, connection_factory, url, data):
1551 """Make an HTTP connection using connection_class.
1552
1553 This is an internal method that should be called from
1554 open_http() or open_https().
1555
1556 Arguments:
1557 - connection_factory should take a host name and return an
1558 HTTPConnection instance.
1559 - url is the url to retrieval or a host, relative-path pair.
1560 - data is payload for a POST request or None.
1561 """
1562
1563 user_passwd = None
1564 proxy_passwd= None
1565 if isinstance(url, str):
Georg Brandl13e89462008-07-01 19:56:00 +00001566 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001567 if host:
Georg Brandl13e89462008-07-01 19:56:00 +00001568 user_passwd, host = splituser(host)
1569 host = unquote(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001570 realhost = host
1571 else:
1572 host, selector = url
1573 # check whether the proxy contains authorization information
Georg Brandl13e89462008-07-01 19:56:00 +00001574 proxy_passwd, host = splituser(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001575 # now we proceed with the url we want to obtain
Georg Brandl13e89462008-07-01 19:56:00 +00001576 urltype, rest = splittype(selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001577 url = rest
1578 user_passwd = None
1579 if urltype.lower() != 'http':
1580 realhost = None
1581 else:
Georg Brandl13e89462008-07-01 19:56:00 +00001582 realhost, rest = splithost(rest)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001583 if realhost:
Georg Brandl13e89462008-07-01 19:56:00 +00001584 user_passwd, realhost = splituser(realhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001585 if user_passwd:
1586 selector = "%s://%s%s" % (urltype, realhost, rest)
1587 if proxy_bypass(realhost):
1588 host = realhost
1589
1590 #print "proxy via http:", host, selector
1591 if not host: raise IOError('http error', 'no host given')
1592
1593 if proxy_passwd:
1594 import base64
Senthil Kumaran5626eec2010-08-04 17:46:23 +00001595 proxy_auth = base64.b64encode(proxy_passwd.encode()).decode('ascii')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001596 else:
1597 proxy_auth = None
1598
1599 if user_passwd:
1600 import base64
Senthil Kumaran5626eec2010-08-04 17:46:23 +00001601 auth = base64.b64encode(user_passwd.encode()).decode('ascii')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001602 else:
1603 auth = None
1604 http_conn = connection_factory(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001605 headers = {}
1606 if proxy_auth:
1607 headers["Proxy-Authorization"] = "Basic %s" % proxy_auth
1608 if auth:
1609 headers["Authorization"] = "Basic %s" % auth
1610 if realhost:
1611 headers["Host"] = realhost
1612 for header, value in self.addheaders:
1613 headers[header] = value
1614
1615 if data is not None:
1616 headers["Content-Type"] = "application/x-www-form-urlencoded"
1617 http_conn.request("POST", selector, data, headers)
1618 else:
1619 http_conn.request("GET", selector, headers=headers)
1620
1621 try:
1622 response = http_conn.getresponse()
1623 except http.client.BadStatusLine:
1624 # something went wrong with the HTTP status line
Georg Brandl13e89462008-07-01 19:56:00 +00001625 raise URLError("http protocol error: bad status line")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001626
1627 # According to RFC 2616, "2xx" code indicates that the client's
1628 # request was successfully received, understood, and accepted.
1629 if 200 <= response.status < 300:
Antoine Pitroub353c122009-02-11 00:39:14 +00001630 return addinfourl(response, response.msg, "http:" + url,
Georg Brandl13e89462008-07-01 19:56:00 +00001631 response.status)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001632 else:
1633 return self.http_error(
1634 url, response.fp,
1635 response.status, response.reason, response.msg, data)
1636
1637 def open_http(self, url, data=None):
1638 """Use HTTP protocol."""
1639 return self._open_generic_http(http.client.HTTPConnection, url, data)
1640
1641 def http_error(self, url, fp, errcode, errmsg, headers, data=None):
1642 """Handle http errors.
1643
1644 Derived class can override this, or provide specific handlers
1645 named http_error_DDD where DDD is the 3-digit error code."""
1646 # First check if there's a specific handler for this error
1647 name = 'http_error_%d' % errcode
1648 if hasattr(self, name):
1649 method = getattr(self, name)
1650 if data is None:
1651 result = method(url, fp, errcode, errmsg, headers)
1652 else:
1653 result = method(url, fp, errcode, errmsg, headers, data)
1654 if result: return result
1655 return self.http_error_default(url, fp, errcode, errmsg, headers)
1656
1657 def http_error_default(self, url, fp, errcode, errmsg, headers):
1658 """Default error handler: close the connection and raise IOError."""
1659 void = fp.read()
1660 fp.close()
Georg Brandl13e89462008-07-01 19:56:00 +00001661 raise HTTPError(url, errcode, errmsg, headers, None)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001662
1663 if _have_ssl:
1664 def _https_connection(self, host):
1665 return http.client.HTTPSConnection(host,
1666 key_file=self.key_file,
1667 cert_file=self.cert_file)
1668
1669 def open_https(self, url, data=None):
1670 """Use HTTPS protocol."""
1671 return self._open_generic_http(self._https_connection, url, data)
1672
1673 def open_file(self, url):
1674 """Use local file or FTP depending on form of URL."""
1675 if not isinstance(url, str):
1676 raise URLError('file error', 'proxy support for file protocol currently not implemented')
1677 if url[:2] == '//' and url[2:3] != '/' and url[2:12].lower() != 'localhost/':
1678 return self.open_ftp(url)
1679 else:
1680 return self.open_local_file(url)
1681
1682 def open_local_file(self, url):
1683 """Use local file."""
1684 import mimetypes, email.utils
1685 from io import StringIO
Georg Brandl13e89462008-07-01 19:56:00 +00001686 host, file = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001687 localname = url2pathname(file)
1688 try:
1689 stats = os.stat(localname)
1690 except OSError as e:
1691 raise URLError(e.errno, e.strerror, e.filename)
1692 size = stats.st_size
1693 modified = email.utils.formatdate(stats.st_mtime, usegmt=True)
1694 mtype = mimetypes.guess_type(url)[0]
1695 headers = email.message_from_string(
1696 'Content-Type: %s\nContent-Length: %d\nLast-modified: %s\n' %
1697 (mtype or 'text/plain', size, modified))
1698 if not host:
1699 urlfile = file
1700 if file[:1] == '/':
1701 urlfile = 'file://' + file
Georg Brandl13e89462008-07-01 19:56:00 +00001702 return addinfourl(open(localname, 'rb'), headers, urlfile)
1703 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001704 if (not port
Senthil Kumaran99b2c8f2009-12-27 10:13:39 +00001705 and socket.gethostbyname(host) in (localhost() + thishost())):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001706 urlfile = file
1707 if file[:1] == '/':
1708 urlfile = 'file://' + file
Georg Brandl13e89462008-07-01 19:56:00 +00001709 return addinfourl(open(localname, 'rb'), headers, urlfile)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001710 raise URLError('local file error', 'not on local host')
1711
1712 def open_ftp(self, url):
1713 """Use FTP protocol."""
1714 if not isinstance(url, str):
1715 raise URLError('ftp error', 'proxy support for ftp protocol currently not implemented')
1716 import mimetypes
1717 from io import StringIO
Georg Brandl13e89462008-07-01 19:56:00 +00001718 host, path = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001719 if not host: raise URLError('ftp error', 'no host given')
Georg Brandl13e89462008-07-01 19:56:00 +00001720 host, port = splitport(host)
1721 user, host = splituser(host)
1722 if user: user, passwd = splitpasswd(user)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001723 else: passwd = None
Georg Brandl13e89462008-07-01 19:56:00 +00001724 host = unquote(host)
1725 user = unquote(user or '')
1726 passwd = unquote(passwd or '')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001727 host = socket.gethostbyname(host)
1728 if not port:
1729 import ftplib
1730 port = ftplib.FTP_PORT
1731 else:
1732 port = int(port)
Georg Brandl13e89462008-07-01 19:56:00 +00001733 path, attrs = splitattr(path)
1734 path = unquote(path)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001735 dirs = path.split('/')
1736 dirs, file = dirs[:-1], dirs[-1]
1737 if dirs and not dirs[0]: dirs = dirs[1:]
1738 if dirs and not dirs[0]: dirs[0] = '/'
1739 key = user, host, port, '/'.join(dirs)
1740 # XXX thread unsafe!
1741 if len(self.ftpcache) > MAXFTPCACHE:
1742 # Prune the cache, rather arbitrarily
1743 for k in self.ftpcache.keys():
1744 if k != key:
1745 v = self.ftpcache[k]
1746 del self.ftpcache[k]
1747 v.close()
1748 try:
1749 if not key in self.ftpcache:
1750 self.ftpcache[key] = \
1751 ftpwrapper(user, passwd, host, port, dirs)
1752 if not file: type = 'D'
1753 else: type = 'I'
1754 for attr in attrs:
Georg Brandl13e89462008-07-01 19:56:00 +00001755 attr, value = splitvalue(attr)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001756 if attr.lower() == 'type' and \
1757 value in ('a', 'A', 'i', 'I', 'd', 'D'):
1758 type = value.upper()
1759 (fp, retrlen) = self.ftpcache[key].retrfile(file, type)
1760 mtype = mimetypes.guess_type("ftp:" + url)[0]
1761 headers = ""
1762 if mtype:
1763 headers += "Content-Type: %s\n" % mtype
1764 if retrlen is not None and retrlen >= 0:
1765 headers += "Content-Length: %d\n" % retrlen
1766 headers = email.message_from_string(headers)
Georg Brandl13e89462008-07-01 19:56:00 +00001767 return addinfourl(fp, headers, "ftp:" + url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001768 except ftperrors() as msg:
1769 raise URLError('ftp error', msg).with_traceback(sys.exc_info()[2])
1770
1771 def open_data(self, url, data=None):
1772 """Use "data" URL."""
1773 if not isinstance(url, str):
1774 raise URLError('data error', 'proxy support for data protocol currently not implemented')
1775 # ignore POSTed data
1776 #
1777 # syntax of data URLs:
1778 # dataurl := "data:" [ mediatype ] [ ";base64" ] "," data
1779 # mediatype := [ type "/" subtype ] *( ";" parameter )
1780 # data := *urlchar
1781 # parameter := attribute "=" value
1782 try:
1783 [type, data] = url.split(',', 1)
1784 except ValueError:
1785 raise IOError('data error', 'bad data URL')
1786 if not type:
1787 type = 'text/plain;charset=US-ASCII'
1788 semi = type.rfind(';')
1789 if semi >= 0 and '=' not in type[semi:]:
1790 encoding = type[semi+1:]
1791 type = type[:semi]
1792 else:
1793 encoding = ''
1794 msg = []
Senthil Kumaranf6c456d2010-05-01 08:29:18 +00001795 msg.append('Date: %s'%time.strftime('%a, %d %b %Y %H:%M:%S GMT',
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001796 time.gmtime(time.time())))
1797 msg.append('Content-type: %s' % type)
1798 if encoding == 'base64':
1799 import base64
Georg Brandl706824f2009-06-04 09:42:55 +00001800 # XXX is this encoding/decoding ok?
1801 data = base64.decodebytes(data.encode('ascii')).decode('latin1')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001802 else:
Georg Brandl13e89462008-07-01 19:56:00 +00001803 data = unquote(data)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001804 msg.append('Content-Length: %d' % len(data))
1805 msg.append('')
1806 msg.append(data)
1807 msg = '\n'.join(msg)
Georg Brandl13e89462008-07-01 19:56:00 +00001808 headers = email.message_from_string(msg)
1809 f = io.StringIO(msg)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001810 #f.fileno = None # needed for addinfourl
Georg Brandl13e89462008-07-01 19:56:00 +00001811 return addinfourl(f, headers, url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001812
1813
1814class FancyURLopener(URLopener):
1815 """Derived class with handlers for errors we can handle (perhaps)."""
1816
1817 def __init__(self, *args, **kwargs):
1818 URLopener.__init__(self, *args, **kwargs)
1819 self.auth_cache = {}
1820 self.tries = 0
1821 self.maxtries = 10
1822
1823 def http_error_default(self, url, fp, errcode, errmsg, headers):
1824 """Default error handling -- don't raise an exception."""
Georg Brandl13e89462008-07-01 19:56:00 +00001825 return addinfourl(fp, headers, "http:" + url, errcode)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001826
1827 def http_error_302(self, url, fp, errcode, errmsg, headers, data=None):
1828 """Error 302 -- relocated (temporarily)."""
1829 self.tries += 1
1830 if self.maxtries and self.tries >= self.maxtries:
1831 if hasattr(self, "http_error_500"):
1832 meth = self.http_error_500
1833 else:
1834 meth = self.http_error_default
1835 self.tries = 0
1836 return meth(url, fp, 500,
1837 "Internal Server Error: Redirect Recursion", headers)
1838 result = self.redirect_internal(url, fp, errcode, errmsg, headers,
1839 data)
1840 self.tries = 0
1841 return result
1842
1843 def redirect_internal(self, url, fp, errcode, errmsg, headers, data):
1844 if 'location' in headers:
1845 newurl = headers['location']
1846 elif 'uri' in headers:
1847 newurl = headers['uri']
1848 else:
1849 return
1850 void = fp.read()
1851 fp.close()
1852 # In case the server sent a relative URL, join with original:
Georg Brandl13e89462008-07-01 19:56:00 +00001853 newurl = urljoin(self.type + ":" + url, newurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001854 return self.open(newurl)
1855
1856 def http_error_301(self, url, fp, errcode, errmsg, headers, data=None):
1857 """Error 301 -- also relocated (permanently)."""
1858 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1859
1860 def http_error_303(self, url, fp, errcode, errmsg, headers, data=None):
1861 """Error 303 -- also relocated (essentially identical to 302)."""
1862 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1863
1864 def http_error_307(self, url, fp, errcode, errmsg, headers, data=None):
1865 """Error 307 -- relocated, but turn POST into error."""
1866 if data is None:
1867 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1868 else:
1869 return self.http_error_default(url, fp, errcode, errmsg, headers)
1870
Senthil Kumaran80f1b052010-06-18 15:08:18 +00001871 def http_error_401(self, url, fp, errcode, errmsg, headers, data=None,
1872 retry=False):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001873 """Error 401 -- authentication required.
1874 This function supports Basic authentication only."""
1875 if not 'www-authenticate' in headers:
1876 URLopener.http_error_default(self, url, fp,
1877 errcode, errmsg, headers)
1878 stuff = headers['www-authenticate']
1879 import re
1880 match = re.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff)
1881 if not match:
1882 URLopener.http_error_default(self, url, fp,
1883 errcode, errmsg, headers)
1884 scheme, realm = match.groups()
1885 if scheme.lower() != 'basic':
1886 URLopener.http_error_default(self, url, fp,
1887 errcode, errmsg, headers)
Senthil Kumaran80f1b052010-06-18 15:08:18 +00001888 if not retry:
1889 URLopener.http_error_default(self, url, fp, errcode, errmsg,
1890 headers)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001891 name = 'retry_' + self.type + '_basic_auth'
1892 if data is None:
1893 return getattr(self,name)(url, realm)
1894 else:
1895 return getattr(self,name)(url, realm, data)
1896
Senthil Kumaran80f1b052010-06-18 15:08:18 +00001897 def http_error_407(self, url, fp, errcode, errmsg, headers, data=None,
1898 retry=False):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001899 """Error 407 -- proxy authentication required.
1900 This function supports Basic authentication only."""
1901 if not 'proxy-authenticate' in headers:
1902 URLopener.http_error_default(self, url, fp,
1903 errcode, errmsg, headers)
1904 stuff = headers['proxy-authenticate']
1905 import re
1906 match = re.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff)
1907 if not match:
1908 URLopener.http_error_default(self, url, fp,
1909 errcode, errmsg, headers)
1910 scheme, realm = match.groups()
1911 if scheme.lower() != 'basic':
1912 URLopener.http_error_default(self, url, fp,
1913 errcode, errmsg, headers)
Senthil Kumaran80f1b052010-06-18 15:08:18 +00001914 if not retry:
1915 URLopener.http_error_default(self, url, fp, errcode, errmsg,
1916 headers)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001917 name = 'retry_proxy_' + self.type + '_basic_auth'
1918 if data is None:
1919 return getattr(self,name)(url, realm)
1920 else:
1921 return getattr(self,name)(url, realm, data)
1922
1923 def retry_proxy_http_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001924 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001925 newurl = 'http://' + host + selector
1926 proxy = self.proxies['http']
Georg Brandl13e89462008-07-01 19:56:00 +00001927 urltype, proxyhost = splittype(proxy)
1928 proxyhost, proxyselector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001929 i = proxyhost.find('@') + 1
1930 proxyhost = proxyhost[i:]
1931 user, passwd = self.get_user_passwd(proxyhost, realm, i)
1932 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001933 proxyhost = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001934 quote(passwd, safe=''), proxyhost)
1935 self.proxies['http'] = 'http://' + proxyhost + proxyselector
1936 if data is None:
1937 return self.open(newurl)
1938 else:
1939 return self.open(newurl, data)
1940
1941 def retry_proxy_https_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001942 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001943 newurl = 'https://' + host + selector
1944 proxy = self.proxies['https']
Georg Brandl13e89462008-07-01 19:56:00 +00001945 urltype, proxyhost = splittype(proxy)
1946 proxyhost, proxyselector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001947 i = proxyhost.find('@') + 1
1948 proxyhost = proxyhost[i:]
1949 user, passwd = self.get_user_passwd(proxyhost, realm, i)
1950 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001951 proxyhost = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001952 quote(passwd, safe=''), proxyhost)
1953 self.proxies['https'] = 'https://' + proxyhost + proxyselector
1954 if data is None:
1955 return self.open(newurl)
1956 else:
1957 return self.open(newurl, data)
1958
1959 def retry_http_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001960 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001961 i = host.find('@') + 1
1962 host = host[i:]
1963 user, passwd = self.get_user_passwd(host, realm, i)
1964 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001965 host = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001966 quote(passwd, safe=''), host)
1967 newurl = 'http://' + host + selector
1968 if data is None:
1969 return self.open(newurl)
1970 else:
1971 return self.open(newurl, data)
1972
1973 def retry_https_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001974 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001975 i = host.find('@') + 1
1976 host = host[i:]
1977 user, passwd = self.get_user_passwd(host, realm, i)
1978 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001979 host = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001980 quote(passwd, safe=''), host)
1981 newurl = 'https://' + host + selector
1982 if data is None:
1983 return self.open(newurl)
1984 else:
1985 return self.open(newurl, data)
1986
Florent Xicluna757445b2010-05-17 17:24:07 +00001987 def get_user_passwd(self, host, realm, clear_cache=0):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001988 key = realm + '@' + host.lower()
1989 if key in self.auth_cache:
1990 if clear_cache:
1991 del self.auth_cache[key]
1992 else:
1993 return self.auth_cache[key]
1994 user, passwd = self.prompt_user_passwd(host, realm)
1995 if user or passwd: self.auth_cache[key] = (user, passwd)
1996 return user, passwd
1997
1998 def prompt_user_passwd(self, host, realm):
1999 """Override this in a GUI environment!"""
2000 import getpass
2001 try:
2002 user = input("Enter username for %s at %s: " % (realm, host))
2003 passwd = getpass.getpass("Enter password for %s in %s at %s: " %
2004 (user, realm, host))
2005 return user, passwd
2006 except KeyboardInterrupt:
2007 print()
2008 return None, None
2009
2010
2011# Utility functions
2012
2013_localhost = None
2014def localhost():
2015 """Return the IP address of the magic hostname 'localhost'."""
2016 global _localhost
2017 if _localhost is None:
2018 _localhost = socket.gethostbyname('localhost')
2019 return _localhost
2020
2021_thishost = None
2022def thishost():
Senthil Kumaran99b2c8f2009-12-27 10:13:39 +00002023 """Return the IP addresses of the current host."""
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002024 global _thishost
2025 if _thishost is None:
Senthil Kumaran99b2c8f2009-12-27 10:13:39 +00002026 _thishost = tuple(socket.gethostbyname_ex(socket.gethostname()[2]))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002027 return _thishost
2028
2029_ftperrors = None
2030def ftperrors():
2031 """Return the set of errors raised by the FTP class."""
2032 global _ftperrors
2033 if _ftperrors is None:
2034 import ftplib
2035 _ftperrors = ftplib.all_errors
2036 return _ftperrors
2037
2038_noheaders = None
2039def noheaders():
Georg Brandl13e89462008-07-01 19:56:00 +00002040 """Return an empty email Message object."""
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002041 global _noheaders
2042 if _noheaders is None:
Georg Brandl13e89462008-07-01 19:56:00 +00002043 _noheaders = email.message_from_string("")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002044 return _noheaders
2045
2046
2047# Utility classes
2048
2049class ftpwrapper:
2050 """Class used by open_ftp() for cache of open FTP connections."""
2051
2052 def __init__(self, user, passwd, host, port, dirs, timeout=None):
2053 self.user = user
2054 self.passwd = passwd
2055 self.host = host
2056 self.port = port
2057 self.dirs = dirs
2058 self.timeout = timeout
2059 self.init()
2060
2061 def init(self):
2062 import ftplib
2063 self.busy = 0
2064 self.ftp = ftplib.FTP()
2065 self.ftp.connect(self.host, self.port, self.timeout)
2066 self.ftp.login(self.user, self.passwd)
2067 for dir in self.dirs:
2068 self.ftp.cwd(dir)
2069
2070 def retrfile(self, file, type):
2071 import ftplib
2072 self.endtransfer()
2073 if type in ('d', 'D'): cmd = 'TYPE A'; isdir = 1
2074 else: cmd = 'TYPE ' + type; isdir = 0
2075 try:
2076 self.ftp.voidcmd(cmd)
2077 except ftplib.all_errors:
2078 self.init()
2079 self.ftp.voidcmd(cmd)
2080 conn = None
2081 if file and not isdir:
2082 # Try to retrieve as a file
2083 try:
2084 cmd = 'RETR ' + file
2085 conn = self.ftp.ntransfercmd(cmd)
2086 except ftplib.error_perm as reason:
2087 if str(reason)[:3] != '550':
Georg Brandl13e89462008-07-01 19:56:00 +00002088 raise URLError('ftp error', reason).with_traceback(
2089 sys.exc_info()[2])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002090 if not conn:
2091 # Set transfer mode to ASCII!
2092 self.ftp.voidcmd('TYPE A')
2093 # Try a directory listing. Verify that directory exists.
2094 if file:
2095 pwd = self.ftp.pwd()
2096 try:
2097 try:
2098 self.ftp.cwd(file)
2099 except ftplib.error_perm as reason:
Georg Brandl13e89462008-07-01 19:56:00 +00002100 raise URLError('ftp error', reason) from reason
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002101 finally:
2102 self.ftp.cwd(pwd)
2103 cmd = 'LIST ' + file
2104 else:
2105 cmd = 'LIST'
2106 conn = self.ftp.ntransfercmd(cmd)
2107 self.busy = 1
2108 # Pass back both a suitably decorated object and a retrieval length
Georg Brandl13e89462008-07-01 19:56:00 +00002109 return (addclosehook(conn[0].makefile('rb'), self.endtransfer), conn[1])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002110 def endtransfer(self):
2111 if not self.busy:
2112 return
2113 self.busy = 0
2114 try:
2115 self.ftp.voidresp()
2116 except ftperrors():
2117 pass
2118
2119 def close(self):
2120 self.endtransfer()
2121 try:
2122 self.ftp.close()
2123 except ftperrors():
2124 pass
2125
2126# Proxy handling
2127def getproxies_environment():
2128 """Return a dictionary of scheme -> proxy server URL mappings.
2129
2130 Scan the environment for variables named <scheme>_proxy;
2131 this seems to be the standard convention. If you need a
2132 different way, you can pass a proxies dictionary to the
2133 [Fancy]URLopener constructor.
2134
2135 """
2136 proxies = {}
2137 for name, value in os.environ.items():
2138 name = name.lower()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002139 if value and name[-6:] == '_proxy':
2140 proxies[name[:-6]] = value
2141 return proxies
2142
2143def proxy_bypass_environment(host):
2144 """Test if proxies should not be used for a particular host.
2145
2146 Checks the environment for a variable named no_proxy, which should
2147 be a list of DNS suffixes separated by commas, or '*' for all hosts.
2148 """
2149 no_proxy = os.environ.get('no_proxy', '') or os.environ.get('NO_PROXY', '')
2150 # '*' is special case for always bypass
2151 if no_proxy == '*':
2152 return 1
2153 # strip port off host
Georg Brandl13e89462008-07-01 19:56:00 +00002154 hostonly, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002155 # check if the host ends with any of the DNS suffixes
2156 for name in no_proxy.split(','):
2157 if name and (hostonly.endswith(name) or host.endswith(name)):
2158 return 1
2159 # otherwise, don't bypass
2160 return 0
2161
2162
2163if sys.platform == 'darwin':
Ronald Oussoren84151202010-04-18 20:46:11 +00002164 from _scproxy import _get_proxy_settings, _get_proxies
2165
2166 def proxy_bypass_macosx_sysconf(host):
2167 """
2168 Return True iff this host shouldn't be accessed using a proxy
2169
2170 This function uses the MacOSX framework SystemConfiguration
2171 to fetch the proxy information.
2172 """
2173 import re
2174 import socket
2175 from fnmatch import fnmatch
2176
2177 hostonly, port = splitport(host)
2178
2179 def ip2num(ipAddr):
2180 parts = ipAddr.split('.')
Mark Dickinsonc3f45c22010-05-09 12:16:29 +00002181 parts = list(map(int, parts))
Ronald Oussoren84151202010-04-18 20:46:11 +00002182 if len(parts) != 4:
2183 parts = (parts + [0, 0, 0, 0])[:4]
2184 return (parts[0] << 24) | (parts[1] << 16) | (parts[2] << 8) | parts[3]
2185
2186 proxy_settings = _get_proxy_settings()
2187
2188 # Check for simple host names:
2189 if '.' not in host:
2190 if proxy_settings['exclude_simple']:
2191 return True
2192
2193 hostIP = None
2194
2195 for value in proxy_settings.get('exceptions', ()):
2196 # Items in the list are strings like these: *.local, 169.254/16
2197 if not value: continue
2198
2199 m = re.match(r"(\d+(?:\.\d+)*)(/\d+)?", value)
2200 if m is not None:
2201 if hostIP is None:
2202 try:
2203 hostIP = socket.gethostbyname(hostonly)
2204 hostIP = ip2num(hostIP)
2205 except socket.error:
2206 continue
2207
2208 base = ip2num(m.group(1))
Ronald Oussorenab90f8e2010-06-27 14:26:30 +00002209 mask = m.group(2)
2210 if mask is None:
2211 mask = 8 * (m.group(1).count('.') + 1)
2212
2213 else:
2214 mask = int(mask[1:])
2215 mask = 32 - mask
Ronald Oussoren84151202010-04-18 20:46:11 +00002216
2217 if (hostIP >> mask) == (base >> mask):
2218 return True
2219
2220 elif fnmatch(host, value):
2221 return True
2222
2223 return False
2224
2225
2226 def getproxies_macosx_sysconf():
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002227 """Return a dictionary of scheme -> proxy server URL mappings.
2228
Ronald Oussoren84151202010-04-18 20:46:11 +00002229 This function uses the MacOSX framework SystemConfiguration
2230 to fetch the proxy information.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002231 """
Ronald Oussoren84151202010-04-18 20:46:11 +00002232 return _get_proxies()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002233
Ronald Oussoren84151202010-04-18 20:46:11 +00002234
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002235
2236 def proxy_bypass(host):
2237 if getproxies_environment():
2238 return proxy_bypass_environment(host)
2239 else:
Ronald Oussoren84151202010-04-18 20:46:11 +00002240 return proxy_bypass_macosx_sysconf(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002241
2242 def getproxies():
Ronald Oussoren84151202010-04-18 20:46:11 +00002243 return getproxies_environment() or getproxies_macosx_sysconf()
2244
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002245
2246elif os.name == 'nt':
2247 def getproxies_registry():
2248 """Return a dictionary of scheme -> proxy server URL mappings.
2249
2250 Win32 uses the registry to store proxies.
2251
2252 """
2253 proxies = {}
2254 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002255 import winreg
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002256 except ImportError:
2257 # Std module, so should be around - but you never know!
2258 return proxies
2259 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002260 internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002261 r'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002262 proxyEnable = winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002263 'ProxyEnable')[0]
2264 if proxyEnable:
2265 # Returned as Unicode but problems if not converted to ASCII
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002266 proxyServer = str(winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002267 'ProxyServer')[0])
2268 if '=' in proxyServer:
2269 # Per-protocol settings
2270 for p in proxyServer.split(';'):
2271 protocol, address = p.split('=', 1)
2272 # See if address has a type:// prefix
2273 import re
2274 if not re.match('^([^/:]+)://', address):
2275 address = '%s://%s' % (protocol, address)
2276 proxies[protocol] = address
2277 else:
2278 # Use one setting for all protocols
2279 if proxyServer[:5] == 'http:':
2280 proxies['http'] = proxyServer
2281 else:
2282 proxies['http'] = 'http://%s' % proxyServer
Senthil Kumaran04f31b82010-07-14 20:10:52 +00002283 proxies['https'] = 'https://%s' % proxyServer
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002284 proxies['ftp'] = 'ftp://%s' % proxyServer
2285 internetSettings.Close()
2286 except (WindowsError, ValueError, TypeError):
2287 # Either registry key not found etc, or the value in an
2288 # unexpected format.
2289 # proxies already set up to be empty so nothing to do
2290 pass
2291 return proxies
2292
2293 def getproxies():
2294 """Return a dictionary of scheme -> proxy server URL mappings.
2295
2296 Returns settings gathered from the environment, if specified,
2297 or the registry.
2298
2299 """
2300 return getproxies_environment() or getproxies_registry()
2301
2302 def proxy_bypass_registry(host):
2303 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002304 import winreg
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002305 import re
2306 except ImportError:
2307 # Std modules, so should be around - but you never know!
2308 return 0
2309 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002310 internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002311 r'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002312 proxyEnable = winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002313 'ProxyEnable')[0]
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002314 proxyOverride = str(winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002315 'ProxyOverride')[0])
2316 # ^^^^ Returned as Unicode but problems if not converted to ASCII
2317 except WindowsError:
2318 return 0
2319 if not proxyEnable or not proxyOverride:
2320 return 0
2321 # try to make a host list from name and IP address.
Georg Brandl13e89462008-07-01 19:56:00 +00002322 rawHost, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002323 host = [rawHost]
2324 try:
2325 addr = socket.gethostbyname(rawHost)
2326 if addr != rawHost:
2327 host.append(addr)
2328 except socket.error:
2329 pass
2330 try:
2331 fqdn = socket.getfqdn(rawHost)
2332 if fqdn != rawHost:
2333 host.append(fqdn)
2334 except socket.error:
2335 pass
2336 # make a check value list from the registry entry: replace the
2337 # '<local>' string by the localhost entry and the corresponding
2338 # canonical entry.
2339 proxyOverride = proxyOverride.split(';')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002340 # now check if we match one of the registry values.
2341 for test in proxyOverride:
Senthil Kumaran49476062009-05-01 06:00:23 +00002342 if test == '<local>':
2343 if '.' not in rawHost:
2344 return 1
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002345 test = test.replace(".", r"\.") # mask dots
2346 test = test.replace("*", r".*") # change glob sequence
2347 test = test.replace("?", r".") # change glob char
2348 for val in host:
2349 # print "%s <--> %s" %( test, val )
2350 if re.match(test, val, re.I):
2351 return 1
2352 return 0
2353
2354 def proxy_bypass(host):
2355 """Return a dictionary of scheme -> proxy server URL mappings.
2356
2357 Returns settings gathered from the environment, if specified,
2358 or the registry.
2359
2360 """
2361 if getproxies_environment():
2362 return proxy_bypass_environment(host)
2363 else:
2364 return proxy_bypass_registry(host)
2365
2366else:
2367 # By default use environment variables
2368 getproxies = getproxies_environment
2369 proxy_bypass = proxy_bypass_environment