blob: 5440efb905420cbc35832164384c097f45658a99 [file] [log] [blame]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001"""An extensible library for opening URLs using a variety of protocols
2
3The simplest way to use this module is to call the urlopen function,
4which accepts a string containing a URL or a Request object (described
5below). It opens the URL and returns the results as file-like
6object; the returned object has some extra methods described below.
7
8The OpenerDirector manages a collection of Handler objects that do
9all the actual work. Each Handler implements a particular protocol or
10option. The OpenerDirector is a composite object that invokes the
11Handlers needed to open the requested URL. For example, the
12HTTPHandler performs HTTP GET and POST requests and deals with
13non-error returns. The HTTPRedirectHandler automatically deals with
14HTTP 301, 302, 303 and 307 redirect errors, and the HTTPDigestAuthHandler
15deals with digest authentication.
16
17urlopen(url, data=None) -- Basic usage is the same as original
18urllib. pass the url and optionally data to post to an HTTP URL, and
19get a file-like object back. One difference is that you can also pass
20a Request instance instead of URL. Raises a URLError (subclass of
21IOError); for HTTP errors, raises an HTTPError, which can also be
22treated as a valid response.
23
24build_opener -- Function that creates a new OpenerDirector instance.
25Will install the default handlers. Accepts one or more Handlers as
26arguments, either instances or Handler classes that it will
27instantiate. If one of the argument is a subclass of the default
28handler, the argument will be installed instead of the default.
29
30install_opener -- Installs a new opener as the default opener.
31
32objects of interest:
Senthil Kumaran04454cd2009-11-15 07:27:02 +000033
Senthil Kumaran4b9fbeb2009-12-20 07:18:22 +000034OpenerDirector -- Sets up the User Agent as the Python-urllib client and manages
35the Handler classes, while dealing with requests and responses.
Jeremy Hylton1afc1692008-06-18 20:49:58 +000036
37Request -- An object that encapsulates the state of a request. The
38state can be as simple as the URL. It can also include extra HTTP
39headers, e.g. a User-Agent.
40
41BaseHandler --
42
43internals:
44BaseHandler and parent
45_call_chain conventions
46
47Example usage:
48
Georg Brandl029986a2008-06-23 11:44:14 +000049import urllib.request
Jeremy Hylton1afc1692008-06-18 20:49:58 +000050
51# set up authentication info
Georg Brandl029986a2008-06-23 11:44:14 +000052authinfo = urllib.request.HTTPBasicAuthHandler()
Jeremy Hylton1afc1692008-06-18 20:49:58 +000053authinfo.add_password(realm='PDQ Application',
54 uri='https://mahler:8092/site-updates.py',
55 user='klem',
56 passwd='geheim$parole')
57
Georg Brandl029986a2008-06-23 11:44:14 +000058proxy_support = urllib.request.ProxyHandler({"http" : "http://ahad-haam:3128"})
Jeremy Hylton1afc1692008-06-18 20:49:58 +000059
60# build a new opener that adds authentication and caching FTP handlers
Georg Brandl029986a2008-06-23 11:44:14 +000061opener = urllib.request.build_opener(proxy_support, authinfo,
62 urllib.request.CacheFTPHandler)
Jeremy Hylton1afc1692008-06-18 20:49:58 +000063
64# install it
Georg Brandl029986a2008-06-23 11:44:14 +000065urllib.request.install_opener(opener)
Jeremy Hylton1afc1692008-06-18 20:49:58 +000066
Georg Brandl029986a2008-06-23 11:44:14 +000067f = urllib.request.urlopen('http://www.python.org/')
Jeremy Hylton1afc1692008-06-18 20:49:58 +000068"""
69
70# XXX issues:
71# If an authentication error handler that tries to perform
72# authentication for some reason but fails, how should the error be
73# signalled? The client needs to know the HTTP error code. But if
74# the handler knows that the problem was, e.g., that it didn't know
75# that hash algo that requested in the challenge, it would be good to
76# pass that information along to the client, too.
77# ftp errors aren't handled cleanly
78# check digest against correct (i.e. non-apache) implementation
79
80# Possible extensions:
81# complex proxies XXX not sure what exactly was meant by this
82# abstract factory for opener
83
84import base64
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +000085import bisect
Jeremy Hylton1afc1692008-06-18 20:49:58 +000086import email
87import hashlib
88import http.client
89import io
90import os
91import posixpath
92import random
93import re
94import socket
95import sys
96import time
Jeremy Hylton1afc1692008-06-18 20:49:58 +000097
Georg Brandl13e89462008-07-01 19:56:00 +000098from urllib.error import URLError, HTTPError, ContentTooShortError
99from urllib.parse import (
100 urlparse, urlsplit, urljoin, unwrap, quote, unquote,
101 splittype, splithost, splitport, splituser, splitpasswd,
Facundo Batistaf24802c2008-08-17 03:36:03 +0000102 splitattr, splitquery, splitvalue, to_bytes, urlunparse)
Georg Brandl13e89462008-07-01 19:56:00 +0000103from urllib.response import addinfourl, addclosehook
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000104
105# check for SSL
106try:
107 import ssl
108except:
109 _have_ssl = False
110else:
111 _have_ssl = True
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000112
113# used in User-Agent header sent
114__version__ = sys.version[:3]
115
116_opener = None
117def urlopen(url, data=None, timeout=socket._GLOBAL_DEFAULT_TIMEOUT):
118 global _opener
119 if _opener is None:
120 _opener = build_opener()
121 return _opener.open(url, data, timeout)
122
123def install_opener(opener):
124 global _opener
125 _opener = opener
126
127# TODO(jhylton): Make this work with the same global opener.
128_urlopener = None
129def urlretrieve(url, filename=None, reporthook=None, data=None):
130 global _urlopener
131 if not _urlopener:
132 _urlopener = FancyURLopener()
133 return _urlopener.retrieve(url, filename, reporthook, data)
134
135def urlcleanup():
136 if _urlopener:
137 _urlopener.cleanup()
138 global _opener
139 if _opener:
140 _opener = None
141
142# copied from cookielib.py
Antoine Pitroufd036452008-08-19 17:56:33 +0000143_cut_port_re = re.compile(r":\d+$", re.ASCII)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000144def request_host(request):
145 """Return request-host, as defined by RFC 2965.
146
147 Variation from RFC: returned value is lowercased, for convenient
148 comparison.
149
150 """
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000151 url = request.full_url
Georg Brandl13e89462008-07-01 19:56:00 +0000152 host = urlparse(url)[1]
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000153 if host == "":
154 host = request.get_header("Host", "")
155
156 # remove port, if present
157 host = _cut_port_re.sub("", host, 1)
158 return host.lower()
159
160class Request:
161
162 def __init__(self, url, data=None, headers={},
163 origin_req_host=None, unverifiable=False):
164 # unwrap('<URL:type://host/path>') --> 'type://host/path'
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000165 self.full_url = unwrap(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000166 self.data = data
167 self.headers = {}
Senthil Kumaran0ac1f832009-07-26 12:39:47 +0000168 self._tunnel_host = None
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000169 for key, value in headers.items():
170 self.add_header(key, value)
171 self.unredirected_hdrs = {}
172 if origin_req_host is None:
173 origin_req_host = request_host(self)
174 self.origin_req_host = origin_req_host
175 self.unverifiable = unverifiable
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000176 self._parse()
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000177
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000178 def _parse(self):
179 self.type, rest = splittype(self.full_url)
180 if self.type is None:
181 raise ValueError("unknown url type: %s" % self.full_url)
182 self.host, self.selector = splithost(rest)
183 if self.host:
184 self.host = unquote(self.host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000185
186 def get_method(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000187 if self.data is not None:
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000188 return "POST"
189 else:
190 return "GET"
191
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000192 # Begin deprecated methods
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000193
194 def add_data(self, data):
195 self.data = data
196
197 def has_data(self):
198 return self.data is not None
199
200 def get_data(self):
201 return self.data
202
203 def get_full_url(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000204 return self.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000205
206 def get_type(self):
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000207 return self.type
208
209 def get_host(self):
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000210 return self.host
211
212 def get_selector(self):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000213 return self.selector
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000214
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000215 def is_unverifiable(self):
216 return self.unverifiable
Facundo Batista72dc1ea2008-08-16 14:44:32 +0000217
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000218 def get_origin_req_host(self):
219 return self.origin_req_host
220
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000221 # End deprecated methods
222
223 def set_proxy(self, host, type):
Senthil Kumaran0ac1f832009-07-26 12:39:47 +0000224 if self.type == 'https' and not self._tunnel_host:
225 self._tunnel_host = self.host
226 else:
227 self.type= type
228 self.selector = self.full_url
229 self.host = host
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000230
231 def has_proxy(self):
232 return self.selector == self.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000233
234 def add_header(self, key, val):
235 # useful for something like authentication
236 self.headers[key.capitalize()] = val
237
238 def add_unredirected_header(self, key, val):
239 # will not be added to a redirected request
240 self.unredirected_hdrs[key.capitalize()] = val
241
242 def has_header(self, header_name):
243 return (header_name in self.headers or
244 header_name in self.unredirected_hdrs)
245
246 def get_header(self, header_name, default=None):
247 return self.headers.get(
248 header_name,
249 self.unredirected_hdrs.get(header_name, default))
250
251 def header_items(self):
252 hdrs = self.unredirected_hdrs.copy()
253 hdrs.update(self.headers)
254 return list(hdrs.items())
255
256class OpenerDirector:
257 def __init__(self):
258 client_version = "Python-urllib/%s" % __version__
259 self.addheaders = [('User-agent', client_version)]
260 # manage the individual handlers
261 self.handlers = []
262 self.handle_open = {}
263 self.handle_error = {}
264 self.process_response = {}
265 self.process_request = {}
266
267 def add_handler(self, handler):
268 if not hasattr(handler, "add_parent"):
269 raise TypeError("expected BaseHandler instance, got %r" %
270 type(handler))
271
272 added = False
273 for meth in dir(handler):
274 if meth in ["redirect_request", "do_open", "proxy_open"]:
275 # oops, coincidental match
276 continue
277
278 i = meth.find("_")
279 protocol = meth[:i]
280 condition = meth[i+1:]
281
282 if condition.startswith("error"):
283 j = condition.find("_") + i + 1
284 kind = meth[j+1:]
285 try:
286 kind = int(kind)
287 except ValueError:
288 pass
289 lookup = self.handle_error.get(protocol, {})
290 self.handle_error[protocol] = lookup
291 elif condition == "open":
292 kind = protocol
293 lookup = self.handle_open
294 elif condition == "response":
295 kind = protocol
296 lookup = self.process_response
297 elif condition == "request":
298 kind = protocol
299 lookup = self.process_request
300 else:
301 continue
302
303 handlers = lookup.setdefault(kind, [])
304 if handlers:
305 bisect.insort(handlers, handler)
306 else:
307 handlers.append(handler)
308 added = True
309
310 if added:
311 # the handlers must work in an specific order, the order
312 # is specified in a Handler attribute
313 bisect.insort(self.handlers, handler)
314 handler.add_parent(self)
315
316 def close(self):
317 # Only exists for backwards compatibility.
318 pass
319
320 def _call_chain(self, chain, kind, meth_name, *args):
321 # Handlers raise an exception if no one else should try to handle
322 # the request, or return None if they can't but another handler
323 # could. Otherwise, they return the response.
324 handlers = chain.get(kind, ())
325 for handler in handlers:
326 func = getattr(handler, meth_name)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000327 result = func(*args)
328 if result is not None:
329 return result
330
331 def open(self, fullurl, data=None, timeout=socket._GLOBAL_DEFAULT_TIMEOUT):
332 # accept a URL or a Request object
333 if isinstance(fullurl, str):
334 req = Request(fullurl, data)
335 else:
336 req = fullurl
337 if data is not None:
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000338 req.data = data
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000339
340 req.timeout = timeout
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000341 protocol = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000342
343 # pre-process request
344 meth_name = protocol+"_request"
345 for processor in self.process_request.get(protocol, []):
346 meth = getattr(processor, meth_name)
347 req = meth(req)
348
349 response = self._open(req, data)
350
351 # post-process response
352 meth_name = protocol+"_response"
353 for processor in self.process_response.get(protocol, []):
354 meth = getattr(processor, meth_name)
355 response = meth(req, response)
356
357 return response
358
359 def _open(self, req, data=None):
360 result = self._call_chain(self.handle_open, 'default',
361 'default_open', req)
362 if result:
363 return result
364
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000365 protocol = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000366 result = self._call_chain(self.handle_open, protocol, protocol +
367 '_open', req)
368 if result:
369 return result
370
371 return self._call_chain(self.handle_open, 'unknown',
372 'unknown_open', req)
373
374 def error(self, proto, *args):
375 if proto in ('http', 'https'):
376 # XXX http[s] protocols are special-cased
377 dict = self.handle_error['http'] # https is not different than http
378 proto = args[2] # YUCK!
379 meth_name = 'http_error_%s' % proto
380 http_err = 1
381 orig_args = args
382 else:
383 dict = self.handle_error
384 meth_name = proto + '_error'
385 http_err = 0
386 args = (dict, proto, meth_name) + args
387 result = self._call_chain(*args)
388 if result:
389 return result
390
391 if http_err:
392 args = (dict, 'default', 'http_error_default') + orig_args
393 return self._call_chain(*args)
394
395# XXX probably also want an abstract factory that knows when it makes
396# sense to skip a superclass in favor of a subclass and when it might
397# make sense to include both
398
399def build_opener(*handlers):
400 """Create an opener object from a list of handlers.
401
402 The opener will use several default handlers, including support
Senthil Kumaran04454cd2009-11-15 07:27:02 +0000403 for HTTP, FTP and when applicable HTTPS.
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000404
405 If any of the handlers passed as arguments are subclasses of the
406 default handlers, the default handlers will not be used.
407 """
408 def isclass(obj):
409 return isinstance(obj, type) or hasattr(obj, "__bases__")
410
411 opener = OpenerDirector()
412 default_classes = [ProxyHandler, UnknownHandler, HTTPHandler,
413 HTTPDefaultErrorHandler, HTTPRedirectHandler,
414 FTPHandler, FileHandler, HTTPErrorProcessor]
415 if hasattr(http.client, "HTTPSConnection"):
416 default_classes.append(HTTPSHandler)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000417 skip = set()
418 for klass in default_classes:
419 for check in handlers:
420 if isclass(check):
421 if issubclass(check, klass):
422 skip.add(klass)
423 elif isinstance(check, klass):
424 skip.add(klass)
425 for klass in skip:
426 default_classes.remove(klass)
427
428 for klass in default_classes:
429 opener.add_handler(klass())
430
431 for h in handlers:
432 if isclass(h):
433 h = h()
434 opener.add_handler(h)
435 return opener
436
437class BaseHandler:
438 handler_order = 500
439
440 def add_parent(self, parent):
441 self.parent = parent
442
443 def close(self):
444 # Only exists for backwards compatibility
445 pass
446
447 def __lt__(self, other):
448 if not hasattr(other, "handler_order"):
449 # Try to preserve the old behavior of having custom classes
450 # inserted after default ones (works only for custom user
451 # classes which are not aware of handler_order).
452 return True
453 return self.handler_order < other.handler_order
454
455
456class HTTPErrorProcessor(BaseHandler):
457 """Process HTTP error responses."""
458 handler_order = 1000 # after all other processing
459
460 def http_response(self, request, response):
461 code, msg, hdrs = response.code, response.msg, response.info()
462
463 # According to RFC 2616, "2xx" code indicates that the client's
464 # request was successfully received, understood, and accepted.
465 if not (200 <= code < 300):
466 response = self.parent.error(
467 'http', request, response, code, msg, hdrs)
468
469 return response
470
471 https_response = http_response
472
473class HTTPDefaultErrorHandler(BaseHandler):
474 def http_error_default(self, req, fp, code, msg, hdrs):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000475 raise HTTPError(req.full_url, code, msg, hdrs, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000476
477class HTTPRedirectHandler(BaseHandler):
478 # maximum number of redirections to any single URL
479 # this is needed because of the state that cookies introduce
480 max_repeats = 4
481 # maximum total number of redirections (regardless of URL) before
482 # assuming we're in a loop
483 max_redirections = 10
484
485 def redirect_request(self, req, fp, code, msg, headers, newurl):
486 """Return a Request or None in response to a redirect.
487
488 This is called by the http_error_30x methods when a
489 redirection response is received. If a redirection should
490 take place, return a new Request to allow http_error_30x to
491 perform the redirect. Otherwise, raise HTTPError if no-one
492 else should try to handle this url. Return None if you can't
493 but another Handler might.
494 """
495 m = req.get_method()
496 if (not (code in (301, 302, 303, 307) and m in ("GET", "HEAD")
497 or code in (301, 302, 303) and m == "POST")):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000498 raise HTTPError(req.full_url, code, msg, headers, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000499
500 # Strictly (according to RFC 2616), 301 or 302 in response to
501 # a POST MUST NOT cause a redirection without confirmation
Georg Brandl029986a2008-06-23 11:44:14 +0000502 # from the user (of urllib.request, in this case). In practice,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000503 # essentially all clients do redirect in this case, so we do
504 # the same.
505 # be conciliant with URIs containing a space
506 newurl = newurl.replace(' ', '%20')
507 CONTENT_HEADERS = ("content-length", "content-type")
508 newheaders = dict((k, v) for k, v in req.headers.items()
509 if k.lower() not in CONTENT_HEADERS)
510 return Request(newurl,
511 headers=newheaders,
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000512 origin_req_host=req.origin_req_host,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000513 unverifiable=True)
514
515 # Implementation note: To avoid the server sending us into an
516 # infinite loop, the request object needs to track what URLs we
517 # have already seen. Do this by adding a handler-specific
518 # attribute to the Request object.
519 def http_error_302(self, req, fp, code, msg, headers):
520 # Some servers (incorrectly) return multiple Location headers
521 # (so probably same goes for URI). Use first header.
522 if "location" in headers:
523 newurl = headers["location"]
524 elif "uri" in headers:
525 newurl = headers["uri"]
526 else:
527 return
Facundo Batistaf24802c2008-08-17 03:36:03 +0000528
529 # fix a possible malformed URL
530 urlparts = urlparse(newurl)
531 if not urlparts.path:
532 urlparts = list(urlparts)
533 urlparts[2] = "/"
534 newurl = urlunparse(urlparts)
535
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000536 newurl = urljoin(req.full_url, newurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000537
538 # XXX Probably want to forget about the state of the current
539 # request, although that might interact poorly with other
540 # handlers that also use handler-specific request attributes
541 new = self.redirect_request(req, fp, code, msg, headers, newurl)
542 if new is None:
543 return
544
545 # loop detection
546 # .redirect_dict has a key url if url was previously visited.
547 if hasattr(req, 'redirect_dict'):
548 visited = new.redirect_dict = req.redirect_dict
549 if (visited.get(newurl, 0) >= self.max_repeats or
550 len(visited) >= self.max_redirections):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000551 raise HTTPError(req.full_url, code,
Georg Brandl13e89462008-07-01 19:56:00 +0000552 self.inf_msg + msg, headers, fp)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000553 else:
554 visited = new.redirect_dict = req.redirect_dict = {}
555 visited[newurl] = visited.get(newurl, 0) + 1
556
557 # Don't close the fp until we are sure that we won't use it
558 # with HTTPError.
559 fp.read()
560 fp.close()
561
Senthil Kumarane9da06f2009-07-19 04:20:12 +0000562 return self.parent.open(new, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000563
564 http_error_301 = http_error_303 = http_error_307 = http_error_302
565
566 inf_msg = "The HTTP server returned a redirect error that would " \
567 "lead to an infinite loop.\n" \
568 "The last 30x error message was:\n"
569
570
571def _parse_proxy(proxy):
572 """Return (scheme, user, password, host/port) given a URL or an authority.
573
574 If a URL is supplied, it must have an authority (host:port) component.
575 According to RFC 3986, having an authority component means the URL must
576 have two slashes after the scheme:
577
578 >>> _parse_proxy('file:/ftp.example.com/')
579 Traceback (most recent call last):
580 ValueError: proxy URL with no authority: 'file:/ftp.example.com/'
581
582 The first three items of the returned tuple may be None.
583
584 Examples of authority parsing:
585
586 >>> _parse_proxy('proxy.example.com')
587 (None, None, None, 'proxy.example.com')
588 >>> _parse_proxy('proxy.example.com:3128')
589 (None, None, None, 'proxy.example.com:3128')
590
591 The authority component may optionally include userinfo (assumed to be
592 username:password):
593
594 >>> _parse_proxy('joe:password@proxy.example.com')
595 (None, 'joe', 'password', 'proxy.example.com')
596 >>> _parse_proxy('joe:password@proxy.example.com:3128')
597 (None, 'joe', 'password', 'proxy.example.com:3128')
598
599 Same examples, but with URLs instead:
600
601 >>> _parse_proxy('http://proxy.example.com/')
602 ('http', None, None, 'proxy.example.com')
603 >>> _parse_proxy('http://proxy.example.com:3128/')
604 ('http', None, None, 'proxy.example.com:3128')
605 >>> _parse_proxy('http://joe:password@proxy.example.com/')
606 ('http', 'joe', 'password', 'proxy.example.com')
607 >>> _parse_proxy('http://joe:password@proxy.example.com:3128')
608 ('http', 'joe', 'password', 'proxy.example.com:3128')
609
610 Everything after the authority is ignored:
611
612 >>> _parse_proxy('ftp://joe:password@proxy.example.com/rubbish:3128')
613 ('ftp', 'joe', 'password', 'proxy.example.com')
614
615 Test for no trailing '/' case:
616
617 >>> _parse_proxy('http://joe:password@proxy.example.com')
618 ('http', 'joe', 'password', 'proxy.example.com')
619
620 """
Georg Brandl13e89462008-07-01 19:56:00 +0000621 scheme, r_scheme = splittype(proxy)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000622 if not r_scheme.startswith("/"):
623 # authority
624 scheme = None
625 authority = proxy
626 else:
627 # URL
628 if not r_scheme.startswith("//"):
629 raise ValueError("proxy URL with no authority: %r" % proxy)
630 # We have an authority, so for RFC 3986-compliant URLs (by ss 3.
631 # and 3.3.), path is empty or starts with '/'
632 end = r_scheme.find("/", 2)
633 if end == -1:
634 end = None
635 authority = r_scheme[2:end]
Georg Brandl13e89462008-07-01 19:56:00 +0000636 userinfo, hostport = splituser(authority)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000637 if userinfo is not None:
Georg Brandl13e89462008-07-01 19:56:00 +0000638 user, password = splitpasswd(userinfo)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000639 else:
640 user = password = None
641 return scheme, user, password, hostport
642
643class ProxyHandler(BaseHandler):
644 # Proxies must be in front
645 handler_order = 100
646
647 def __init__(self, proxies=None):
648 if proxies is None:
649 proxies = getproxies()
650 assert hasattr(proxies, 'keys'), "proxies must be a mapping"
651 self.proxies = proxies
652 for type, url in proxies.items():
653 setattr(self, '%s_open' % type,
654 lambda r, proxy=url, type=type, meth=self.proxy_open: \
655 meth(r, proxy, type))
656
657 def proxy_open(self, req, proxy, type):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000658 orig_type = req.type
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000659 proxy_type, user, password, hostport = _parse_proxy(proxy)
660 if proxy_type is None:
661 proxy_type = orig_type
Senthil Kumaran11301632009-10-11 06:07:46 +0000662
663 if req.host and proxy_bypass(req.host):
664 return None
665
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000666 if user and password:
Georg Brandl13e89462008-07-01 19:56:00 +0000667 user_pass = '%s:%s' % (unquote(user),
668 unquote(password))
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000669 creds = base64.b64encode(user_pass.encode()).decode("ascii")
670 req.add_header('Proxy-authorization', 'Basic ' + creds)
Georg Brandl13e89462008-07-01 19:56:00 +0000671 hostport = unquote(hostport)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000672 req.set_proxy(hostport, proxy_type)
Senthil Kumaran0ac1f832009-07-26 12:39:47 +0000673 if orig_type == proxy_type or orig_type == 'https':
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000674 # let other handlers take care of it
675 return None
676 else:
677 # need to start over, because the other handlers don't
678 # grok the proxy's URL type
679 # e.g. if we have a constructor arg proxies like so:
680 # {'http': 'ftp://proxy.example.com'}, we may end up turning
681 # a request for http://acme.example.com/a into one for
682 # ftp://proxy.example.com/a
Senthil Kumarane9da06f2009-07-19 04:20:12 +0000683 return self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000684
685class HTTPPasswordMgr:
686
687 def __init__(self):
688 self.passwd = {}
689
690 def add_password(self, realm, uri, user, passwd):
691 # uri could be a single URI or a sequence
692 if isinstance(uri, str):
693 uri = [uri]
694 if not realm in self.passwd:
695 self.passwd[realm] = {}
696 for default_port in True, False:
697 reduced_uri = tuple(
698 [self.reduce_uri(u, default_port) for u in uri])
699 self.passwd[realm][reduced_uri] = (user, passwd)
700
701 def find_user_password(self, realm, authuri):
702 domains = self.passwd.get(realm, {})
703 for default_port in True, False:
704 reduced_authuri = self.reduce_uri(authuri, default_port)
705 for uris, authinfo in domains.items():
706 for uri in uris:
707 if self.is_suburi(uri, reduced_authuri):
708 return authinfo
709 return None, None
710
711 def reduce_uri(self, uri, default_port=True):
712 """Accept authority or URI and extract only the authority and path."""
713 # note HTTP URLs do not have a userinfo component
Georg Brandl13e89462008-07-01 19:56:00 +0000714 parts = urlsplit(uri)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000715 if parts[1]:
716 # URI
717 scheme = parts[0]
718 authority = parts[1]
719 path = parts[2] or '/'
720 else:
721 # host or host:port
722 scheme = None
723 authority = uri
724 path = '/'
Georg Brandl13e89462008-07-01 19:56:00 +0000725 host, port = splitport(authority)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000726 if default_port and port is None and scheme is not None:
727 dport = {"http": 80,
728 "https": 443,
729 }.get(scheme)
730 if dport is not None:
731 authority = "%s:%d" % (host, dport)
732 return authority, path
733
734 def is_suburi(self, base, test):
735 """Check if test is below base in a URI tree
736
737 Both args must be URIs in reduced form.
738 """
739 if base == test:
740 return True
741 if base[0] != test[0]:
742 return False
743 common = posixpath.commonprefix((base[1], test[1]))
744 if len(common) == len(base[1]):
745 return True
746 return False
747
748
749class HTTPPasswordMgrWithDefaultRealm(HTTPPasswordMgr):
750
751 def find_user_password(self, realm, authuri):
752 user, password = HTTPPasswordMgr.find_user_password(self, realm,
753 authuri)
754 if user is not None:
755 return user, password
756 return HTTPPasswordMgr.find_user_password(self, None, authuri)
757
758
759class AbstractBasicAuthHandler:
760
761 # XXX this allows for multiple auth-schemes, but will stupidly pick
762 # the last one with a realm specified.
763
764 # allow for double- and single-quoted realm values
765 # (single quotes are a violation of the RFC, but appear in the wild)
766 rx = re.compile('(?:.*,)*[ \t]*([^ \t]+)[ \t]+'
767 'realm=(["\'])(.*?)\\2', re.I)
768
769 # XXX could pre-emptively send auth info already accepted (RFC 2617,
770 # end of section 2, and section 1.2 immediately after "credentials"
771 # production).
772
773 def __init__(self, password_mgr=None):
774 if password_mgr is None:
775 password_mgr = HTTPPasswordMgr()
776 self.passwd = password_mgr
777 self.add_password = self.passwd.add_password
778
779 def http_error_auth_reqed(self, authreq, host, req, headers):
780 # host may be an authority (without userinfo) or a URL with an
781 # authority
782 # XXX could be multiple headers
783 authreq = headers.get(authreq, None)
784 if authreq:
785 mo = AbstractBasicAuthHandler.rx.search(authreq)
786 if mo:
787 scheme, quote, realm = mo.groups()
788 if scheme.lower() == 'basic':
789 return self.retry_http_basic_auth(host, req, realm)
790
791 def retry_http_basic_auth(self, host, req, realm):
792 user, pw = self.passwd.find_user_password(realm, host)
793 if pw is not None:
794 raw = "%s:%s" % (user, pw)
795 auth = "Basic " + base64.b64encode(raw.encode()).decode("ascii")
796 if req.headers.get(self.auth_header, None) == auth:
797 return None
Senthil Kumaranefcd8832010-02-24 16:56:20 +0000798 req.add_unredirected_header(self.auth_header, auth)
Senthil Kumarane9da06f2009-07-19 04:20:12 +0000799 return self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000800 else:
801 return None
802
803
804class HTTPBasicAuthHandler(AbstractBasicAuthHandler, BaseHandler):
805
806 auth_header = 'Authorization'
807
808 def http_error_401(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000809 url = req.full_url
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000810 return self.http_error_auth_reqed('www-authenticate',
811 url, req, headers)
812
813
814class ProxyBasicAuthHandler(AbstractBasicAuthHandler, BaseHandler):
815
816 auth_header = 'Proxy-authorization'
817
818 def http_error_407(self, req, fp, code, msg, headers):
819 # http_error_auth_reqed requires that there is no userinfo component in
Georg Brandl029986a2008-06-23 11:44:14 +0000820 # authority. Assume there isn't one, since urllib.request does not (and
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000821 # should not, RFC 3986 s. 3.2.1) support requests for URLs containing
822 # userinfo.
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000823 authority = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000824 return self.http_error_auth_reqed('proxy-authenticate',
825 authority, req, headers)
826
827
828def randombytes(n):
829 """Return n random bytes."""
830 return os.urandom(n)
831
832class AbstractDigestAuthHandler:
833 # Digest authentication is specified in RFC 2617.
834
835 # XXX The client does not inspect the Authentication-Info header
836 # in a successful response.
837
838 # XXX It should be possible to test this implementation against
839 # a mock server that just generates a static set of challenges.
840
841 # XXX qop="auth-int" supports is shaky
842
843 def __init__(self, passwd=None):
844 if passwd is None:
845 passwd = HTTPPasswordMgr()
846 self.passwd = passwd
847 self.add_password = self.passwd.add_password
848 self.retried = 0
849 self.nonce_count = 0
Senthil Kumaranb58474f2009-11-15 08:45:27 +0000850 self.last_nonce = None
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000851
852 def reset_retry_count(self):
853 self.retried = 0
854
855 def http_error_auth_reqed(self, auth_header, host, req, headers):
856 authreq = headers.get(auth_header, None)
857 if self.retried > 5:
858 # Don't fail endlessly - if we failed once, we'll probably
859 # fail a second time. Hm. Unless the Password Manager is
860 # prompting for the information. Crap. This isn't great
861 # but it's better than the current 'repeat until recursion
862 # depth exceeded' approach <wink>
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000863 raise HTTPError(req.full_url, 401, "digest auth failed",
Georg Brandl13e89462008-07-01 19:56:00 +0000864 headers, None)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000865 else:
866 self.retried += 1
867 if authreq:
868 scheme = authreq.split()[0]
869 if scheme.lower() == 'digest':
870 return self.retry_http_digest_auth(req, authreq)
871
872 def retry_http_digest_auth(self, req, auth):
873 token, challenge = auth.split(' ', 1)
874 chal = parse_keqv_list(filter(None, parse_http_list(challenge)))
875 auth = self.get_authorization(req, chal)
876 if auth:
877 auth_val = 'Digest %s' % auth
878 if req.headers.get(self.auth_header, None) == auth_val:
879 return None
880 req.add_unredirected_header(self.auth_header, auth_val)
Senthil Kumarane9da06f2009-07-19 04:20:12 +0000881 resp = self.parent.open(req, timeout=req.timeout)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000882 return resp
883
884 def get_cnonce(self, nonce):
885 # The cnonce-value is an opaque
886 # quoted string value provided by the client and used by both client
887 # and server to avoid chosen plaintext attacks, to provide mutual
888 # authentication, and to provide some message integrity protection.
889 # This isn't a fabulous effort, but it's probably Good Enough.
890 s = "%s:%s:%s:" % (self.nonce_count, nonce, time.ctime())
891 b = s.encode("ascii") + randombytes(8)
892 dig = hashlib.sha1(b).hexdigest()
893 return dig[:16]
894
895 def get_authorization(self, req, chal):
896 try:
897 realm = chal['realm']
898 nonce = chal['nonce']
899 qop = chal.get('qop')
900 algorithm = chal.get('algorithm', 'MD5')
901 # mod_digest doesn't send an opaque, even though it isn't
902 # supposed to be optional
903 opaque = chal.get('opaque', None)
904 except KeyError:
905 return None
906
907 H, KD = self.get_algorithm_impls(algorithm)
908 if H is None:
909 return None
910
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000911 user, pw = self.passwd.find_user_password(realm, req.full_url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000912 if user is None:
913 return None
914
915 # XXX not implemented yet
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000916 if req.data is not None:
917 entdig = self.get_entity_digest(req.data, chal)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000918 else:
919 entdig = None
920
921 A1 = "%s:%s:%s" % (user, realm, pw)
922 A2 = "%s:%s" % (req.get_method(),
923 # XXX selector: what about proxies and full urls
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000924 req.selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000925 if qop == 'auth':
Senthil Kumaranb58474f2009-11-15 08:45:27 +0000926 if nonce == self.last_nonce:
927 self.nonce_count += 1
928 else:
929 self.nonce_count = 1
930 self.last_nonce = nonce
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000931 ncvalue = '%08x' % self.nonce_count
932 cnonce = self.get_cnonce(nonce)
933 noncebit = "%s:%s:%s:%s:%s" % (nonce, ncvalue, cnonce, qop, H(A2))
934 respdig = KD(H(A1), noncebit)
935 elif qop is None:
936 respdig = KD(H(A1), "%s:%s" % (nonce, H(A2)))
937 else:
938 # XXX handle auth-int.
Georg Brandl13e89462008-07-01 19:56:00 +0000939 raise URLError("qop '%s' is not supported." % qop)
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000940
941 # XXX should the partial digests be encoded too?
942
943 base = 'username="%s", realm="%s", nonce="%s", uri="%s", ' \
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000944 'response="%s"' % (user, realm, nonce, req.selector,
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000945 respdig)
946 if opaque:
947 base += ', opaque="%s"' % opaque
948 if entdig:
949 base += ', digest="%s"' % entdig
950 base += ', algorithm="%s"' % algorithm
951 if qop:
952 base += ', qop=auth, nc=%s, cnonce="%s"' % (ncvalue, cnonce)
953 return base
954
955 def get_algorithm_impls(self, algorithm):
956 # lambdas assume digest modules are imported at the top level
957 if algorithm == 'MD5':
958 H = lambda x: hashlib.md5(x.encode("ascii")).hexdigest()
959 elif algorithm == 'SHA':
960 H = lambda x: hashlib.sha1(x.encode("ascii")).hexdigest()
961 # XXX MD5-sess
962 KD = lambda s, d: H("%s:%s" % (s, d))
963 return H, KD
964
965 def get_entity_digest(self, data, chal):
966 # XXX not implemented yet
967 return None
968
969
970class HTTPDigestAuthHandler(BaseHandler, AbstractDigestAuthHandler):
971 """An authentication protocol defined by RFC 2069
972
973 Digest authentication improves on basic authentication because it
974 does not transmit passwords in the clear.
975 """
976
977 auth_header = 'Authorization'
978 handler_order = 490 # before Basic auth
979
980 def http_error_401(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000981 host = urlparse(req.full_url)[1]
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000982 retry = self.http_error_auth_reqed('www-authenticate',
983 host, req, headers)
984 self.reset_retry_count()
985 return retry
986
987
988class ProxyDigestAuthHandler(BaseHandler, AbstractDigestAuthHandler):
989
990 auth_header = 'Proxy-Authorization'
991 handler_order = 490 # before Basic auth
992
993 def http_error_407(self, req, fp, code, msg, headers):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +0000994 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +0000995 retry = self.http_error_auth_reqed('proxy-authenticate',
996 host, req, headers)
997 self.reset_retry_count()
998 return retry
999
1000class AbstractHTTPHandler(BaseHandler):
1001
1002 def __init__(self, debuglevel=0):
1003 self._debuglevel = debuglevel
1004
1005 def set_http_debuglevel(self, level):
1006 self._debuglevel = level
1007
1008 def do_request_(self, request):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001009 host = request.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001010 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001011 raise URLError('no host given')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001012
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001013 if request.data is not None: # POST
1014 data = request.data
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001015 if not request.has_header('Content-type'):
1016 request.add_unredirected_header(
1017 'Content-type',
1018 'application/x-www-form-urlencoded')
1019 if not request.has_header('Content-length'):
1020 request.add_unredirected_header(
1021 'Content-length', '%d' % len(data))
1022
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001023 sel_host = host
1024 if request.has_proxy():
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001025 scheme, sel = splittype(request.selector)
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001026 sel_host, sel_path = splithost(sel)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001027 if not request.has_header('Host'):
Facundo Batista72dc1ea2008-08-16 14:44:32 +00001028 request.add_unredirected_header('Host', sel_host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001029 for name, value in self.parent.addheaders:
1030 name = name.capitalize()
1031 if not request.has_header(name):
1032 request.add_unredirected_header(name, value)
1033
1034 return request
1035
1036 def do_open(self, http_class, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001037 """Return an HTTPResponse object for the request, using http_class.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001038
1039 http_class must implement the HTTPConnection API from http.client.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001040 """
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001041 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001042 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001043 raise URLError('no host given')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001044
1045 h = http_class(host, timeout=req.timeout) # will parse host:port
1046 headers = dict(req.headers)
1047 headers.update(req.unredirected_hdrs)
1048
1049 # TODO(jhylton): Should this be redesigned to handle
1050 # persistent connections?
1051
1052 # We want to make an HTTP/1.1 request, but the addinfourl
1053 # class isn't prepared to deal with a persistent connection.
1054 # It will try to read all remaining data from the socket,
1055 # which will block while the server waits for the next request.
1056 # So make sure the connection gets closed after the (only)
1057 # request.
1058 headers["Connection"] = "close"
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001059 headers = dict((name.title(), val) for name, val in headers.items())
Senthil Kumaran0ac1f832009-07-26 12:39:47 +00001060
1061 if req._tunnel_host:
Senthil Kumaran4b9fbeb2009-12-20 07:18:22 +00001062 tunnel_headers = {}
1063 proxy_auth_hdr = "Proxy-Authorization"
1064 if proxy_auth_hdr in headers:
1065 tunnel_headers[proxy_auth_hdr] = headers[proxy_auth_hdr]
1066 # Proxy-Authorization should not be sent to origin
1067 # server.
1068 del headers[proxy_auth_hdr]
1069 h._set_tunnel(req._tunnel_host, headers=tunnel_headers)
Senthil Kumaran0ac1f832009-07-26 12:39:47 +00001070
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001071 try:
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001072 h.request(req.get_method(), req.selector, req.data, headers)
1073 r = h.getresponse() # an HTTPResponse instance
1074 except socket.error as err:
Georg Brandl13e89462008-07-01 19:56:00 +00001075 raise URLError(err)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001076
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001077 r.url = req.full_url
1078 # This line replaces the .msg attribute of the HTTPResponse
1079 # with .headers, because urllib clients expect the response to
1080 # have the reason in .msg. It would be good to mark this
1081 # attribute is deprecated and get then to use info() or
1082 # .headers.
1083 r.msg = r.reason
1084 return r
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001085
1086
1087class HTTPHandler(AbstractHTTPHandler):
1088
1089 def http_open(self, req):
1090 return self.do_open(http.client.HTTPConnection, req)
1091
1092 http_request = AbstractHTTPHandler.do_request_
1093
1094if hasattr(http.client, 'HTTPSConnection'):
1095 class HTTPSHandler(AbstractHTTPHandler):
1096
1097 def https_open(self, req):
1098 return self.do_open(http.client.HTTPSConnection, req)
1099
1100 https_request = AbstractHTTPHandler.do_request_
1101
1102class HTTPCookieProcessor(BaseHandler):
1103 def __init__(self, cookiejar=None):
1104 import http.cookiejar
1105 if cookiejar is None:
1106 cookiejar = http.cookiejar.CookieJar()
1107 self.cookiejar = cookiejar
1108
1109 def http_request(self, request):
1110 self.cookiejar.add_cookie_header(request)
1111 return request
1112
1113 def http_response(self, request, response):
1114 self.cookiejar.extract_cookies(response, request)
1115 return response
1116
1117 https_request = http_request
1118 https_response = http_response
1119
1120class UnknownHandler(BaseHandler):
1121 def unknown_open(self, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001122 type = req.type
Georg Brandl13e89462008-07-01 19:56:00 +00001123 raise URLError('unknown url type: %s' % type)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001124
1125def parse_keqv_list(l):
1126 """Parse list of key=value strings where keys are not duplicated."""
1127 parsed = {}
1128 for elt in l:
1129 k, v = elt.split('=', 1)
1130 if v[0] == '"' and v[-1] == '"':
1131 v = v[1:-1]
1132 parsed[k] = v
1133 return parsed
1134
1135def parse_http_list(s):
1136 """Parse lists as described by RFC 2068 Section 2.
1137
1138 In particular, parse comma-separated lists where the elements of
1139 the list may include quoted-strings. A quoted-string could
1140 contain a comma. A non-quoted string could have quotes in the
1141 middle. Neither commas nor quotes count if they are escaped.
1142 Only double-quotes count, not single-quotes.
1143 """
1144 res = []
1145 part = ''
1146
1147 escape = quote = False
1148 for cur in s:
1149 if escape:
1150 part += cur
1151 escape = False
1152 continue
1153 if quote:
1154 if cur == '\\':
1155 escape = True
1156 continue
1157 elif cur == '"':
1158 quote = False
1159 part += cur
1160 continue
1161
1162 if cur == ',':
1163 res.append(part)
1164 part = ''
1165 continue
1166
1167 if cur == '"':
1168 quote = True
1169
1170 part += cur
1171
1172 # append last part
1173 if part:
1174 res.append(part)
1175
1176 return [part.strip() for part in res]
1177
1178class FileHandler(BaseHandler):
1179 # Use local file or FTP depending on form of URL
1180 def file_open(self, req):
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001181 url = req.selector
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001182 if url[:2] == '//' and url[2:3] != '/':
1183 req.type = 'ftp'
1184 return self.parent.open(req)
1185 else:
1186 return self.open_local_file(req)
1187
1188 # names for the localhost
1189 names = None
1190 def get_names(self):
1191 if FileHandler.names is None:
1192 try:
Senthil Kumaran88a495d2009-12-27 10:15:45 +00001193 FileHandler.names = tuple(
1194 socket.gethostbyname_ex('localhost')[2] +
1195 socket.gethostbyname_ex(socket.gethostname())[2])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001196 except socket.gaierror:
1197 FileHandler.names = (socket.gethostbyname('localhost'),)
1198 return FileHandler.names
1199
1200 # not entirely sure what the rules are here
1201 def open_local_file(self, req):
1202 import email.utils
1203 import mimetypes
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001204 host = req.host
Senthil Kumaran1e72bd32010-05-08 05:14:29 +00001205 filename = req.selector
1206 localfile = url2pathname(filename)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001207 try:
1208 stats = os.stat(localfile)
1209 size = stats.st_size
1210 modified = email.utils.formatdate(stats.st_mtime, usegmt=True)
Senthil Kumaran1e72bd32010-05-08 05:14:29 +00001211 mtype = mimetypes.guess_type(filename)[0]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001212 headers = email.message_from_string(
1213 'Content-type: %s\nContent-length: %d\nLast-modified: %s\n' %
1214 (mtype or 'text/plain', size, modified))
1215 if host:
Georg Brandl13e89462008-07-01 19:56:00 +00001216 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001217 if not host or \
1218 (not port and _safe_gethostbyname(host) in self.get_names()):
Senthil Kumaran1e72bd32010-05-08 05:14:29 +00001219 if host:
1220 origurl = 'file://' + host + filename
1221 else:
1222 origurl = 'file://' + filename
1223 return addinfourl(open(localfile, 'rb'), headers, origurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001224 except OSError as msg:
Georg Brandl029986a2008-06-23 11:44:14 +00001225 # users shouldn't expect OSErrors coming from urlopen()
Georg Brandl13e89462008-07-01 19:56:00 +00001226 raise URLError(msg)
1227 raise URLError('file not on local host')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001228
1229def _safe_gethostbyname(host):
1230 try:
1231 return socket.gethostbyname(host)
1232 except socket.gaierror:
1233 return None
1234
1235class FTPHandler(BaseHandler):
1236 def ftp_open(self, req):
1237 import ftplib
1238 import mimetypes
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001239 host = req.host
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001240 if not host:
Georg Brandl13e89462008-07-01 19:56:00 +00001241 raise URLError('ftp error: no host given')
1242 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001243 if port is None:
1244 port = ftplib.FTP_PORT
1245 else:
1246 port = int(port)
1247
1248 # username/password handling
Georg Brandl13e89462008-07-01 19:56:00 +00001249 user, host = splituser(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001250 if user:
Georg Brandl13e89462008-07-01 19:56:00 +00001251 user, passwd = splitpasswd(user)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001252 else:
1253 passwd = None
Georg Brandl13e89462008-07-01 19:56:00 +00001254 host = unquote(host)
1255 user = unquote(user or '')
1256 passwd = unquote(passwd or '')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001257
1258 try:
1259 host = socket.gethostbyname(host)
1260 except socket.error as msg:
Georg Brandl13e89462008-07-01 19:56:00 +00001261 raise URLError(msg)
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001262 path, attrs = splitattr(req.selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001263 dirs = path.split('/')
Georg Brandl13e89462008-07-01 19:56:00 +00001264 dirs = list(map(unquote, dirs))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001265 dirs, file = dirs[:-1], dirs[-1]
1266 if dirs and not dirs[0]:
1267 dirs = dirs[1:]
1268 try:
1269 fw = self.connect_ftp(user, passwd, host, port, dirs, req.timeout)
1270 type = file and 'I' or 'D'
1271 for attr in attrs:
Georg Brandl13e89462008-07-01 19:56:00 +00001272 attr, value = splitvalue(attr)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001273 if attr.lower() == 'type' and \
1274 value in ('a', 'A', 'i', 'I', 'd', 'D'):
1275 type = value.upper()
1276 fp, retrlen = fw.retrfile(file, type)
1277 headers = ""
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001278 mtype = mimetypes.guess_type(req.full_url)[0]
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001279 if mtype:
1280 headers += "Content-type: %s\n" % mtype
1281 if retrlen is not None and retrlen >= 0:
1282 headers += "Content-length: %d\n" % retrlen
1283 headers = email.message_from_string(headers)
Jeremy Hylton6c5e28c2009-03-31 14:35:53 +00001284 return addinfourl(fp, headers, req.full_url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001285 except ftplib.all_errors as msg:
Georg Brandl13e89462008-07-01 19:56:00 +00001286 exc = URLError('ftp error: %s' % msg)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001287 raise exc.with_traceback(sys.exc_info()[2])
1288
1289 def connect_ftp(self, user, passwd, host, port, dirs, timeout):
1290 fw = ftpwrapper(user, passwd, host, port, dirs, timeout)
1291 return fw
1292
1293class CacheFTPHandler(FTPHandler):
1294 # XXX would be nice to have pluggable cache strategies
1295 # XXX this stuff is definitely not thread safe
1296 def __init__(self):
1297 self.cache = {}
1298 self.timeout = {}
1299 self.soonest = 0
1300 self.delay = 60
1301 self.max_conns = 16
1302
1303 def setTimeout(self, t):
1304 self.delay = t
1305
1306 def setMaxConns(self, m):
1307 self.max_conns = m
1308
1309 def connect_ftp(self, user, passwd, host, port, dirs, timeout):
1310 key = user, host, port, '/'.join(dirs), timeout
1311 if key in self.cache:
1312 self.timeout[key] = time.time() + self.delay
1313 else:
1314 self.cache[key] = ftpwrapper(user, passwd, host, port,
1315 dirs, timeout)
1316 self.timeout[key] = time.time() + self.delay
1317 self.check_cache()
1318 return self.cache[key]
1319
1320 def check_cache(self):
1321 # first check for old ones
1322 t = time.time()
1323 if self.soonest <= t:
1324 for k, v in list(self.timeout.items()):
1325 if v < t:
1326 self.cache[k].close()
1327 del self.cache[k]
1328 del self.timeout[k]
1329 self.soonest = min(list(self.timeout.values()))
1330
1331 # then check the size
1332 if len(self.cache) == self.max_conns:
1333 for k, v in list(self.timeout.items()):
1334 if v == self.soonest:
1335 del self.cache[k]
1336 del self.timeout[k]
1337 break
1338 self.soonest = min(list(self.timeout.values()))
1339
1340# Code move from the old urllib module
1341
1342MAXFTPCACHE = 10 # Trim the ftp cache beyond this size
1343
1344# Helper for non-unix systems
1345if os.name == 'mac':
1346 from macurl2path import url2pathname, pathname2url
1347elif os.name == 'nt':
1348 from nturl2path import url2pathname, pathname2url
1349else:
1350 def url2pathname(pathname):
1351 """OS-specific conversion from a relative URL of the 'file' scheme
1352 to a file system path; not recommended for general use."""
Georg Brandl13e89462008-07-01 19:56:00 +00001353 return unquote(pathname)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001354
1355 def pathname2url(pathname):
1356 """OS-specific conversion from a file system path to a relative URL
1357 of the 'file' scheme; not recommended for general use."""
Georg Brandl13e89462008-07-01 19:56:00 +00001358 return quote(pathname)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001359
1360# This really consists of two pieces:
1361# (1) a class which handles opening of all sorts of URLs
1362# (plus assorted utilities etc.)
1363# (2) a set of functions for parsing URLs
1364# XXX Should these be separated out into different modules?
1365
1366
1367ftpcache = {}
1368class URLopener:
1369 """Class to open URLs.
1370 This is a class rather than just a subroutine because we may need
1371 more than one set of global protocol-specific options.
1372 Note -- this is a base class for those who don't want the
1373 automatic handling of errors type 302 (relocated) and 401
1374 (authorization needed)."""
1375
1376 __tempfiles = None
1377
1378 version = "Python-urllib/%s" % __version__
1379
1380 # Constructor
1381 def __init__(self, proxies=None, **x509):
1382 if proxies is None:
1383 proxies = getproxies()
1384 assert hasattr(proxies, 'keys'), "proxies must be a mapping"
1385 self.proxies = proxies
1386 self.key_file = x509.get('key_file')
1387 self.cert_file = x509.get('cert_file')
1388 self.addheaders = [('User-Agent', self.version)]
1389 self.__tempfiles = []
1390 self.__unlink = os.unlink # See cleanup()
1391 self.tempcache = None
1392 # Undocumented feature: if you assign {} to tempcache,
1393 # it is used to cache files retrieved with
1394 # self.retrieve(). This is not enabled by default
1395 # since it does not work for changing documents (and I
1396 # haven't got the logic to check expiration headers
1397 # yet).
1398 self.ftpcache = ftpcache
1399 # Undocumented feature: you can use a different
1400 # ftp cache by assigning to the .ftpcache member;
1401 # in case you want logically independent URL openers
1402 # XXX This is not threadsafe. Bah.
1403
1404 def __del__(self):
1405 self.close()
1406
1407 def close(self):
1408 self.cleanup()
1409
1410 def cleanup(self):
1411 # This code sometimes runs when the rest of this module
1412 # has already been deleted, so it can't use any globals
1413 # or import anything.
1414 if self.__tempfiles:
1415 for file in self.__tempfiles:
1416 try:
1417 self.__unlink(file)
1418 except OSError:
1419 pass
1420 del self.__tempfiles[:]
1421 if self.tempcache:
1422 self.tempcache.clear()
1423
1424 def addheader(self, *args):
1425 """Add a header to be used by the HTTP interface only
1426 e.g. u.addheader('Accept', 'sound/basic')"""
1427 self.addheaders.append(args)
1428
1429 # External interface
1430 def open(self, fullurl, data=None):
1431 """Use URLopener().open(file) instead of open(file, 'r')."""
Georg Brandl13e89462008-07-01 19:56:00 +00001432 fullurl = unwrap(to_bytes(fullurl))
Senthil Kumaran0e7e9ae2010-02-20 22:30:21 +00001433 fullurl = quote(fullurl, safe="%/:=&?~#+!$,;'@()*[]|")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001434 if self.tempcache and fullurl in self.tempcache:
1435 filename, headers = self.tempcache[fullurl]
1436 fp = open(filename, 'rb')
Georg Brandl13e89462008-07-01 19:56:00 +00001437 return addinfourl(fp, headers, fullurl)
1438 urltype, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001439 if not urltype:
1440 urltype = 'file'
1441 if urltype in self.proxies:
1442 proxy = self.proxies[urltype]
Georg Brandl13e89462008-07-01 19:56:00 +00001443 urltype, proxyhost = splittype(proxy)
1444 host, selector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001445 url = (host, fullurl) # Signal special case to open_*()
1446 else:
1447 proxy = None
1448 name = 'open_' + urltype
1449 self.type = urltype
1450 name = name.replace('-', '_')
1451 if not hasattr(self, name):
1452 if proxy:
1453 return self.open_unknown_proxy(proxy, fullurl, data)
1454 else:
1455 return self.open_unknown(fullurl, data)
1456 try:
1457 if data is None:
1458 return getattr(self, name)(url)
1459 else:
1460 return getattr(self, name)(url, data)
1461 except socket.error as msg:
1462 raise IOError('socket error', msg).with_traceback(sys.exc_info()[2])
1463
1464 def open_unknown(self, fullurl, data=None):
1465 """Overridable interface to open unknown URL type."""
Georg Brandl13e89462008-07-01 19:56:00 +00001466 type, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001467 raise IOError('url error', 'unknown url type', type)
1468
1469 def open_unknown_proxy(self, proxy, fullurl, data=None):
1470 """Overridable interface to open unknown URL type."""
Georg Brandl13e89462008-07-01 19:56:00 +00001471 type, url = splittype(fullurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001472 raise IOError('url error', 'invalid proxy for %s' % type, proxy)
1473
1474 # External interface
1475 def retrieve(self, url, filename=None, reporthook=None, data=None):
1476 """retrieve(url) returns (filename, headers) for a local object
1477 or (tempfilename, headers) for a remote object."""
Georg Brandl13e89462008-07-01 19:56:00 +00001478 url = unwrap(to_bytes(url))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001479 if self.tempcache and url in self.tempcache:
1480 return self.tempcache[url]
Georg Brandl13e89462008-07-01 19:56:00 +00001481 type, url1 = splittype(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001482 if filename is None and (not type or type == 'file'):
1483 try:
1484 fp = self.open_local_file(url1)
1485 hdrs = fp.info()
1486 del fp
Georg Brandl13e89462008-07-01 19:56:00 +00001487 return url2pathname(splithost(url1)[1]), hdrs
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001488 except IOError as msg:
1489 pass
1490 fp = self.open(url, data)
Benjamin Peterson5f28b7b2009-03-26 21:49:58 +00001491 try:
1492 headers = fp.info()
1493 if filename:
1494 tfp = open(filename, 'wb')
1495 else:
1496 import tempfile
1497 garbage, path = splittype(url)
1498 garbage, path = splithost(path or "")
1499 path, garbage = splitquery(path or "")
1500 path, garbage = splitattr(path or "")
1501 suffix = os.path.splitext(path)[1]
1502 (fd, filename) = tempfile.mkstemp(suffix)
1503 self.__tempfiles.append(filename)
1504 tfp = os.fdopen(fd, 'wb')
1505 try:
1506 result = filename, headers
1507 if self.tempcache is not None:
1508 self.tempcache[url] = result
1509 bs = 1024*8
1510 size = -1
1511 read = 0
1512 blocknum = 0
1513 if reporthook:
1514 if "content-length" in headers:
1515 size = int(headers["Content-Length"])
1516 reporthook(blocknum, bs, size)
1517 while 1:
1518 block = fp.read(bs)
1519 if not block:
1520 break
1521 read += len(block)
1522 tfp.write(block)
1523 blocknum += 1
1524 if reporthook:
1525 reporthook(blocknum, bs, size)
1526 finally:
1527 tfp.close()
1528 finally:
1529 fp.close()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001530 del fp
1531 del tfp
1532
1533 # raise exception if actual size does not match content-length header
1534 if size >= 0 and read < size:
Georg Brandl13e89462008-07-01 19:56:00 +00001535 raise ContentTooShortError(
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001536 "retrieval incomplete: got only %i out of %i bytes"
1537 % (read, size), result)
1538
1539 return result
1540
1541 # Each method named open_<type> knows how to open that type of URL
1542
1543 def _open_generic_http(self, connection_factory, url, data):
1544 """Make an HTTP connection using connection_class.
1545
1546 This is an internal method that should be called from
1547 open_http() or open_https().
1548
1549 Arguments:
1550 - connection_factory should take a host name and return an
1551 HTTPConnection instance.
1552 - url is the url to retrieval or a host, relative-path pair.
1553 - data is payload for a POST request or None.
1554 """
1555
1556 user_passwd = None
1557 proxy_passwd= None
1558 if isinstance(url, str):
Georg Brandl13e89462008-07-01 19:56:00 +00001559 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001560 if host:
Georg Brandl13e89462008-07-01 19:56:00 +00001561 user_passwd, host = splituser(host)
1562 host = unquote(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001563 realhost = host
1564 else:
1565 host, selector = url
1566 # check whether the proxy contains authorization information
Georg Brandl13e89462008-07-01 19:56:00 +00001567 proxy_passwd, host = splituser(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001568 # now we proceed with the url we want to obtain
Georg Brandl13e89462008-07-01 19:56:00 +00001569 urltype, rest = splittype(selector)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001570 url = rest
1571 user_passwd = None
1572 if urltype.lower() != 'http':
1573 realhost = None
1574 else:
Georg Brandl13e89462008-07-01 19:56:00 +00001575 realhost, rest = splithost(rest)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001576 if realhost:
Georg Brandl13e89462008-07-01 19:56:00 +00001577 user_passwd, realhost = splituser(realhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001578 if user_passwd:
1579 selector = "%s://%s%s" % (urltype, realhost, rest)
1580 if proxy_bypass(realhost):
1581 host = realhost
1582
1583 #print "proxy via http:", host, selector
1584 if not host: raise IOError('http error', 'no host given')
1585
1586 if proxy_passwd:
1587 import base64
1588 proxy_auth = base64.b64encode(proxy_passwd).strip()
1589 else:
1590 proxy_auth = None
1591
1592 if user_passwd:
1593 import base64
1594 auth = base64.b64encode(user_passwd).strip()
1595 else:
1596 auth = None
1597 http_conn = connection_factory(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001598 headers = {}
1599 if proxy_auth:
1600 headers["Proxy-Authorization"] = "Basic %s" % proxy_auth
1601 if auth:
1602 headers["Authorization"] = "Basic %s" % auth
1603 if realhost:
1604 headers["Host"] = realhost
1605 for header, value in self.addheaders:
1606 headers[header] = value
1607
1608 if data is not None:
1609 headers["Content-Type"] = "application/x-www-form-urlencoded"
1610 http_conn.request("POST", selector, data, headers)
1611 else:
1612 http_conn.request("GET", selector, headers=headers)
1613
1614 try:
1615 response = http_conn.getresponse()
1616 except http.client.BadStatusLine:
1617 # something went wrong with the HTTP status line
Georg Brandl13e89462008-07-01 19:56:00 +00001618 raise URLError("http protocol error: bad status line")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001619
1620 # According to RFC 2616, "2xx" code indicates that the client's
1621 # request was successfully received, understood, and accepted.
1622 if 200 <= response.status < 300:
Antoine Pitroub353c122009-02-11 00:39:14 +00001623 return addinfourl(response, response.msg, "http:" + url,
Georg Brandl13e89462008-07-01 19:56:00 +00001624 response.status)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001625 else:
1626 return self.http_error(
1627 url, response.fp,
1628 response.status, response.reason, response.msg, data)
1629
1630 def open_http(self, url, data=None):
1631 """Use HTTP protocol."""
1632 return self._open_generic_http(http.client.HTTPConnection, url, data)
1633
1634 def http_error(self, url, fp, errcode, errmsg, headers, data=None):
1635 """Handle http errors.
1636
1637 Derived class can override this, or provide specific handlers
1638 named http_error_DDD where DDD is the 3-digit error code."""
1639 # First check if there's a specific handler for this error
1640 name = 'http_error_%d' % errcode
1641 if hasattr(self, name):
1642 method = getattr(self, name)
1643 if data is None:
1644 result = method(url, fp, errcode, errmsg, headers)
1645 else:
1646 result = method(url, fp, errcode, errmsg, headers, data)
1647 if result: return result
1648 return self.http_error_default(url, fp, errcode, errmsg, headers)
1649
1650 def http_error_default(self, url, fp, errcode, errmsg, headers):
1651 """Default error handler: close the connection and raise IOError."""
1652 void = fp.read()
1653 fp.close()
Georg Brandl13e89462008-07-01 19:56:00 +00001654 raise HTTPError(url, errcode, errmsg, headers, None)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001655
1656 if _have_ssl:
1657 def _https_connection(self, host):
1658 return http.client.HTTPSConnection(host,
1659 key_file=self.key_file,
1660 cert_file=self.cert_file)
1661
1662 def open_https(self, url, data=None):
1663 """Use HTTPS protocol."""
1664 return self._open_generic_http(self._https_connection, url, data)
1665
1666 def open_file(self, url):
1667 """Use local file or FTP depending on form of URL."""
1668 if not isinstance(url, str):
1669 raise URLError('file error', 'proxy support for file protocol currently not implemented')
1670 if url[:2] == '//' and url[2:3] != '/' and url[2:12].lower() != 'localhost/':
1671 return self.open_ftp(url)
1672 else:
1673 return self.open_local_file(url)
1674
1675 def open_local_file(self, url):
1676 """Use local file."""
1677 import mimetypes, email.utils
1678 from io import StringIO
Georg Brandl13e89462008-07-01 19:56:00 +00001679 host, file = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001680 localname = url2pathname(file)
1681 try:
1682 stats = os.stat(localname)
1683 except OSError as e:
1684 raise URLError(e.errno, e.strerror, e.filename)
1685 size = stats.st_size
1686 modified = email.utils.formatdate(stats.st_mtime, usegmt=True)
1687 mtype = mimetypes.guess_type(url)[0]
1688 headers = email.message_from_string(
1689 'Content-Type: %s\nContent-Length: %d\nLast-modified: %s\n' %
1690 (mtype or 'text/plain', size, modified))
1691 if not host:
1692 urlfile = file
1693 if file[:1] == '/':
1694 urlfile = 'file://' + file
Georg Brandl13e89462008-07-01 19:56:00 +00001695 return addinfourl(open(localname, 'rb'), headers, urlfile)
1696 host, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001697 if (not port
Senthil Kumaran88a495d2009-12-27 10:15:45 +00001698 and socket.gethostbyname(host) in (localhost() + thishost())):
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001699 urlfile = file
1700 if file[:1] == '/':
1701 urlfile = 'file://' + file
Georg Brandl13e89462008-07-01 19:56:00 +00001702 return addinfourl(open(localname, 'rb'), headers, urlfile)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001703 raise URLError('local file error', 'not on local host')
1704
1705 def open_ftp(self, url):
1706 """Use FTP protocol."""
1707 if not isinstance(url, str):
1708 raise URLError('ftp error', 'proxy support for ftp protocol currently not implemented')
1709 import mimetypes
1710 from io import StringIO
Georg Brandl13e89462008-07-01 19:56:00 +00001711 host, path = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001712 if not host: raise URLError('ftp error', 'no host given')
Georg Brandl13e89462008-07-01 19:56:00 +00001713 host, port = splitport(host)
1714 user, host = splituser(host)
1715 if user: user, passwd = splitpasswd(user)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001716 else: passwd = None
Georg Brandl13e89462008-07-01 19:56:00 +00001717 host = unquote(host)
1718 user = unquote(user or '')
1719 passwd = unquote(passwd or '')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001720 host = socket.gethostbyname(host)
1721 if not port:
1722 import ftplib
1723 port = ftplib.FTP_PORT
1724 else:
1725 port = int(port)
Georg Brandl13e89462008-07-01 19:56:00 +00001726 path, attrs = splitattr(path)
1727 path = unquote(path)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001728 dirs = path.split('/')
1729 dirs, file = dirs[:-1], dirs[-1]
1730 if dirs and not dirs[0]: dirs = dirs[1:]
1731 if dirs and not dirs[0]: dirs[0] = '/'
1732 key = user, host, port, '/'.join(dirs)
1733 # XXX thread unsafe!
1734 if len(self.ftpcache) > MAXFTPCACHE:
1735 # Prune the cache, rather arbitrarily
1736 for k in self.ftpcache.keys():
1737 if k != key:
1738 v = self.ftpcache[k]
1739 del self.ftpcache[k]
1740 v.close()
1741 try:
1742 if not key in self.ftpcache:
1743 self.ftpcache[key] = \
1744 ftpwrapper(user, passwd, host, port, dirs)
1745 if not file: type = 'D'
1746 else: type = 'I'
1747 for attr in attrs:
Georg Brandl13e89462008-07-01 19:56:00 +00001748 attr, value = splitvalue(attr)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001749 if attr.lower() == 'type' and \
1750 value in ('a', 'A', 'i', 'I', 'd', 'D'):
1751 type = value.upper()
1752 (fp, retrlen) = self.ftpcache[key].retrfile(file, type)
1753 mtype = mimetypes.guess_type("ftp:" + url)[0]
1754 headers = ""
1755 if mtype:
1756 headers += "Content-Type: %s\n" % mtype
1757 if retrlen is not None and retrlen >= 0:
1758 headers += "Content-Length: %d\n" % retrlen
1759 headers = email.message_from_string(headers)
Georg Brandl13e89462008-07-01 19:56:00 +00001760 return addinfourl(fp, headers, "ftp:" + url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001761 except ftperrors() as msg:
1762 raise URLError('ftp error', msg).with_traceback(sys.exc_info()[2])
1763
1764 def open_data(self, url, data=None):
1765 """Use "data" URL."""
1766 if not isinstance(url, str):
1767 raise URLError('data error', 'proxy support for data protocol currently not implemented')
1768 # ignore POSTed data
1769 #
1770 # syntax of data URLs:
1771 # dataurl := "data:" [ mediatype ] [ ";base64" ] "," data
1772 # mediatype := [ type "/" subtype ] *( ";" parameter )
1773 # data := *urlchar
1774 # parameter := attribute "=" value
1775 try:
1776 [type, data] = url.split(',', 1)
1777 except ValueError:
1778 raise IOError('data error', 'bad data URL')
1779 if not type:
1780 type = 'text/plain;charset=US-ASCII'
1781 semi = type.rfind(';')
1782 if semi >= 0 and '=' not in type[semi:]:
1783 encoding = type[semi+1:]
1784 type = type[:semi]
1785 else:
1786 encoding = ''
1787 msg = []
Senthil Kumaran5a3bc652010-05-01 08:32:23 +00001788 msg.append('Date: %s'%time.strftime('%a, %d %b %Y %H:%M:%S GMT',
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001789 time.gmtime(time.time())))
1790 msg.append('Content-type: %s' % type)
1791 if encoding == 'base64':
1792 import base64
Georg Brandl706824f2009-06-04 09:42:55 +00001793 # XXX is this encoding/decoding ok?
1794 data = base64.decodebytes(data.encode('ascii')).decode('latin1')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001795 else:
Georg Brandl13e89462008-07-01 19:56:00 +00001796 data = unquote(data)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001797 msg.append('Content-Length: %d' % len(data))
1798 msg.append('')
1799 msg.append(data)
1800 msg = '\n'.join(msg)
Georg Brandl13e89462008-07-01 19:56:00 +00001801 headers = email.message_from_string(msg)
1802 f = io.StringIO(msg)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001803 #f.fileno = None # needed for addinfourl
Georg Brandl13e89462008-07-01 19:56:00 +00001804 return addinfourl(f, headers, url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001805
1806
1807class FancyURLopener(URLopener):
1808 """Derived class with handlers for errors we can handle (perhaps)."""
1809
1810 def __init__(self, *args, **kwargs):
1811 URLopener.__init__(self, *args, **kwargs)
1812 self.auth_cache = {}
1813 self.tries = 0
1814 self.maxtries = 10
1815
1816 def http_error_default(self, url, fp, errcode, errmsg, headers):
1817 """Default error handling -- don't raise an exception."""
Georg Brandl13e89462008-07-01 19:56:00 +00001818 return addinfourl(fp, headers, "http:" + url, errcode)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001819
1820 def http_error_302(self, url, fp, errcode, errmsg, headers, data=None):
1821 """Error 302 -- relocated (temporarily)."""
1822 self.tries += 1
1823 if self.maxtries and self.tries >= self.maxtries:
1824 if hasattr(self, "http_error_500"):
1825 meth = self.http_error_500
1826 else:
1827 meth = self.http_error_default
1828 self.tries = 0
1829 return meth(url, fp, 500,
1830 "Internal Server Error: Redirect Recursion", headers)
1831 result = self.redirect_internal(url, fp, errcode, errmsg, headers,
1832 data)
1833 self.tries = 0
1834 return result
1835
1836 def redirect_internal(self, url, fp, errcode, errmsg, headers, data):
1837 if 'location' in headers:
1838 newurl = headers['location']
1839 elif 'uri' in headers:
1840 newurl = headers['uri']
1841 else:
1842 return
1843 void = fp.read()
1844 fp.close()
1845 # In case the server sent a relative URL, join with original:
Georg Brandl13e89462008-07-01 19:56:00 +00001846 newurl = urljoin(self.type + ":" + url, newurl)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001847 return self.open(newurl)
1848
1849 def http_error_301(self, url, fp, errcode, errmsg, headers, data=None):
1850 """Error 301 -- also relocated (permanently)."""
1851 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1852
1853 def http_error_303(self, url, fp, errcode, errmsg, headers, data=None):
1854 """Error 303 -- also relocated (essentially identical to 302)."""
1855 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1856
1857 def http_error_307(self, url, fp, errcode, errmsg, headers, data=None):
1858 """Error 307 -- relocated, but turn POST into error."""
1859 if data is None:
1860 return self.http_error_302(url, fp, errcode, errmsg, headers, data)
1861 else:
1862 return self.http_error_default(url, fp, errcode, errmsg, headers)
1863
1864 def http_error_401(self, url, fp, errcode, errmsg, headers, data=None):
1865 """Error 401 -- authentication required.
1866 This function supports Basic authentication only."""
1867 if not 'www-authenticate' in headers:
1868 URLopener.http_error_default(self, url, fp,
1869 errcode, errmsg, headers)
1870 stuff = headers['www-authenticate']
1871 import re
1872 match = re.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff)
1873 if not match:
1874 URLopener.http_error_default(self, url, fp,
1875 errcode, errmsg, headers)
1876 scheme, realm = match.groups()
1877 if scheme.lower() != 'basic':
1878 URLopener.http_error_default(self, url, fp,
1879 errcode, errmsg, headers)
1880 name = 'retry_' + self.type + '_basic_auth'
1881 if data is None:
1882 return getattr(self,name)(url, realm)
1883 else:
1884 return getattr(self,name)(url, realm, data)
1885
1886 def http_error_407(self, url, fp, errcode, errmsg, headers, data=None):
1887 """Error 407 -- proxy authentication required.
1888 This function supports Basic authentication only."""
1889 if not 'proxy-authenticate' in headers:
1890 URLopener.http_error_default(self, url, fp,
1891 errcode, errmsg, headers)
1892 stuff = headers['proxy-authenticate']
1893 import re
1894 match = re.match('[ \t]*([^ \t]+)[ \t]+realm="([^"]*)"', stuff)
1895 if not match:
1896 URLopener.http_error_default(self, url, fp,
1897 errcode, errmsg, headers)
1898 scheme, realm = match.groups()
1899 if scheme.lower() != 'basic':
1900 URLopener.http_error_default(self, url, fp,
1901 errcode, errmsg, headers)
1902 name = 'retry_proxy_' + self.type + '_basic_auth'
1903 if data is None:
1904 return getattr(self,name)(url, realm)
1905 else:
1906 return getattr(self,name)(url, realm, data)
1907
1908 def retry_proxy_http_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001909 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001910 newurl = 'http://' + host + selector
1911 proxy = self.proxies['http']
Georg Brandl13e89462008-07-01 19:56:00 +00001912 urltype, proxyhost = splittype(proxy)
1913 proxyhost, proxyselector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001914 i = proxyhost.find('@') + 1
1915 proxyhost = proxyhost[i:]
1916 user, passwd = self.get_user_passwd(proxyhost, realm, i)
1917 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001918 proxyhost = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001919 quote(passwd, safe=''), proxyhost)
1920 self.proxies['http'] = 'http://' + proxyhost + proxyselector
1921 if data is None:
1922 return self.open(newurl)
1923 else:
1924 return self.open(newurl, data)
1925
1926 def retry_proxy_https_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001927 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001928 newurl = 'https://' + host + selector
1929 proxy = self.proxies['https']
Georg Brandl13e89462008-07-01 19:56:00 +00001930 urltype, proxyhost = splittype(proxy)
1931 proxyhost, proxyselector = splithost(proxyhost)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001932 i = proxyhost.find('@') + 1
1933 proxyhost = proxyhost[i:]
1934 user, passwd = self.get_user_passwd(proxyhost, realm, i)
1935 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001936 proxyhost = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001937 quote(passwd, safe=''), proxyhost)
1938 self.proxies['https'] = 'https://' + proxyhost + proxyselector
1939 if data is None:
1940 return self.open(newurl)
1941 else:
1942 return self.open(newurl, data)
1943
1944 def retry_http_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001945 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001946 i = host.find('@') + 1
1947 host = host[i:]
1948 user, passwd = self.get_user_passwd(host, realm, i)
1949 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001950 host = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001951 quote(passwd, safe=''), host)
1952 newurl = 'http://' + host + selector
1953 if data is None:
1954 return self.open(newurl)
1955 else:
1956 return self.open(newurl, data)
1957
1958 def retry_https_basic_auth(self, url, realm, data=None):
Georg Brandl13e89462008-07-01 19:56:00 +00001959 host, selector = splithost(url)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001960 i = host.find('@') + 1
1961 host = host[i:]
1962 user, passwd = self.get_user_passwd(host, realm, i)
1963 if not (user or passwd): return None
Georg Brandl13e89462008-07-01 19:56:00 +00001964 host = "%s:%s@%s" % (quote(user, safe=''),
Jeremy Hylton1afc1692008-06-18 20:49:58 +00001965 quote(passwd, safe=''), host)
1966 newurl = 'https://' + host + selector
1967 if data is None:
1968 return self.open(newurl)
1969 else:
1970 return self.open(newurl, data)
1971
1972 def get_user_passwd(self, host, realm, clear_cache = 0):
1973 key = realm + '@' + host.lower()
1974 if key in self.auth_cache:
1975 if clear_cache:
1976 del self.auth_cache[key]
1977 else:
1978 return self.auth_cache[key]
1979 user, passwd = self.prompt_user_passwd(host, realm)
1980 if user or passwd: self.auth_cache[key] = (user, passwd)
1981 return user, passwd
1982
1983 def prompt_user_passwd(self, host, realm):
1984 """Override this in a GUI environment!"""
1985 import getpass
1986 try:
1987 user = input("Enter username for %s at %s: " % (realm, host))
1988 passwd = getpass.getpass("Enter password for %s in %s at %s: " %
1989 (user, realm, host))
1990 return user, passwd
1991 except KeyboardInterrupt:
1992 print()
1993 return None, None
1994
1995
1996# Utility functions
1997
1998_localhost = None
1999def localhost():
2000 """Return the IP address of the magic hostname 'localhost'."""
2001 global _localhost
2002 if _localhost is None:
2003 _localhost = socket.gethostbyname('localhost')
2004 return _localhost
2005
2006_thishost = None
2007def thishost():
Senthil Kumaran88a495d2009-12-27 10:15:45 +00002008 """Return the IP addresses of the current host."""
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002009 global _thishost
2010 if _thishost is None:
Senthil Kumaran88a495d2009-12-27 10:15:45 +00002011 _thishost = tuple(socket.gethostbyname_ex(socket.gethostname()[2]))
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002012 return _thishost
2013
2014_ftperrors = None
2015def ftperrors():
2016 """Return the set of errors raised by the FTP class."""
2017 global _ftperrors
2018 if _ftperrors is None:
2019 import ftplib
2020 _ftperrors = ftplib.all_errors
2021 return _ftperrors
2022
2023_noheaders = None
2024def noheaders():
Georg Brandl13e89462008-07-01 19:56:00 +00002025 """Return an empty email Message object."""
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002026 global _noheaders
2027 if _noheaders is None:
Georg Brandl13e89462008-07-01 19:56:00 +00002028 _noheaders = email.message_from_string("")
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002029 return _noheaders
2030
2031
2032# Utility classes
2033
2034class ftpwrapper:
2035 """Class used by open_ftp() for cache of open FTP connections."""
2036
2037 def __init__(self, user, passwd, host, port, dirs, timeout=None):
2038 self.user = user
2039 self.passwd = passwd
2040 self.host = host
2041 self.port = port
2042 self.dirs = dirs
2043 self.timeout = timeout
2044 self.init()
2045
2046 def init(self):
2047 import ftplib
2048 self.busy = 0
2049 self.ftp = ftplib.FTP()
2050 self.ftp.connect(self.host, self.port, self.timeout)
2051 self.ftp.login(self.user, self.passwd)
2052 for dir in self.dirs:
2053 self.ftp.cwd(dir)
2054
2055 def retrfile(self, file, type):
2056 import ftplib
2057 self.endtransfer()
2058 if type in ('d', 'D'): cmd = 'TYPE A'; isdir = 1
2059 else: cmd = 'TYPE ' + type; isdir = 0
2060 try:
2061 self.ftp.voidcmd(cmd)
2062 except ftplib.all_errors:
2063 self.init()
2064 self.ftp.voidcmd(cmd)
2065 conn = None
2066 if file and not isdir:
2067 # Try to retrieve as a file
2068 try:
2069 cmd = 'RETR ' + file
2070 conn = self.ftp.ntransfercmd(cmd)
2071 except ftplib.error_perm as reason:
2072 if str(reason)[:3] != '550':
Georg Brandl13e89462008-07-01 19:56:00 +00002073 raise URLError('ftp error', reason).with_traceback(
2074 sys.exc_info()[2])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002075 if not conn:
2076 # Set transfer mode to ASCII!
2077 self.ftp.voidcmd('TYPE A')
2078 # Try a directory listing. Verify that directory exists.
2079 if file:
2080 pwd = self.ftp.pwd()
2081 try:
2082 try:
2083 self.ftp.cwd(file)
2084 except ftplib.error_perm as reason:
Georg Brandl13e89462008-07-01 19:56:00 +00002085 raise URLError('ftp error', reason) from reason
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002086 finally:
2087 self.ftp.cwd(pwd)
2088 cmd = 'LIST ' + file
2089 else:
2090 cmd = 'LIST'
2091 conn = self.ftp.ntransfercmd(cmd)
2092 self.busy = 1
2093 # Pass back both a suitably decorated object and a retrieval length
Georg Brandl13e89462008-07-01 19:56:00 +00002094 return (addclosehook(conn[0].makefile('rb'), self.endtransfer), conn[1])
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002095 def endtransfer(self):
2096 if not self.busy:
2097 return
2098 self.busy = 0
2099 try:
2100 self.ftp.voidresp()
2101 except ftperrors():
2102 pass
2103
2104 def close(self):
2105 self.endtransfer()
2106 try:
2107 self.ftp.close()
2108 except ftperrors():
2109 pass
2110
2111# Proxy handling
2112def getproxies_environment():
2113 """Return a dictionary of scheme -> proxy server URL mappings.
2114
2115 Scan the environment for variables named <scheme>_proxy;
2116 this seems to be the standard convention. If you need a
2117 different way, you can pass a proxies dictionary to the
2118 [Fancy]URLopener constructor.
2119
2120 """
2121 proxies = {}
2122 for name, value in os.environ.items():
2123 name = name.lower()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002124 if value and name[-6:] == '_proxy':
2125 proxies[name[:-6]] = value
2126 return proxies
2127
2128def proxy_bypass_environment(host):
2129 """Test if proxies should not be used for a particular host.
2130
2131 Checks the environment for a variable named no_proxy, which should
2132 be a list of DNS suffixes separated by commas, or '*' for all hosts.
2133 """
2134 no_proxy = os.environ.get('no_proxy', '') or os.environ.get('NO_PROXY', '')
2135 # '*' is special case for always bypass
2136 if no_proxy == '*':
2137 return 1
2138 # strip port off host
Georg Brandl13e89462008-07-01 19:56:00 +00002139 hostonly, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002140 # check if the host ends with any of the DNS suffixes
2141 for name in no_proxy.split(','):
2142 if name and (hostonly.endswith(name) or host.endswith(name)):
2143 return 1
2144 # otherwise, don't bypass
2145 return 0
2146
2147
2148if sys.platform == 'darwin':
Ronald Oussoren218cc582010-04-18 20:49:34 +00002149 from _scproxy import _get_proxy_settings, _get_proxies
2150
2151 def proxy_bypass_macosx_sysconf(host):
2152 """
2153 Return True iff this host shouldn't be accessed using a proxy
2154
2155 This function uses the MacOSX framework SystemConfiguration
2156 to fetch the proxy information.
2157 """
2158 import re
2159 import socket
2160 from fnmatch import fnmatch
2161
2162 hostonly, port = splitport(host)
2163
2164 def ip2num(ipAddr):
2165 parts = ipAddr.split('.')
Mark Dickinsonb7d94362010-05-09 12:17:58 +00002166 parts = list(map(int, parts))
Ronald Oussoren218cc582010-04-18 20:49:34 +00002167 if len(parts) != 4:
2168 parts = (parts + [0, 0, 0, 0])[:4]
2169 return (parts[0] << 24) | (parts[1] << 16) | (parts[2] << 8) | parts[3]
2170
2171 proxy_settings = _get_proxy_settings()
2172
2173 # Check for simple host names:
2174 if '.' not in host:
2175 if proxy_settings['exclude_simple']:
2176 return True
2177
2178 hostIP = None
2179
2180 for value in proxy_settings.get('exceptions', ()):
2181 # Items in the list are strings like these: *.local, 169.254/16
2182 if not value: continue
2183
2184 m = re.match(r"(\d+(?:\.\d+)*)(/\d+)?", value)
2185 if m is not None:
2186 if hostIP is None:
2187 try:
2188 hostIP = socket.gethostbyname(hostonly)
2189 hostIP = ip2num(hostIP)
2190 except socket.error:
2191 continue
2192
2193 base = ip2num(m.group(1))
2194 mask = int(m.group(2)[1:])
2195 mask = 32 - mask
2196
2197 if (hostIP >> mask) == (base >> mask):
2198 return True
2199
2200 elif fnmatch(host, value):
2201 return True
2202
2203 return False
2204
2205
2206 def getproxies_macosx_sysconf():
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002207 """Return a dictionary of scheme -> proxy server URL mappings.
2208
Ronald Oussoren218cc582010-04-18 20:49:34 +00002209 This function uses the MacOSX framework SystemConfiguration
2210 to fetch the proxy information.
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002211 """
Ronald Oussoren218cc582010-04-18 20:49:34 +00002212 return _get_proxies()
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002213
Ronald Oussoren218cc582010-04-18 20:49:34 +00002214
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002215
2216 def proxy_bypass(host):
2217 if getproxies_environment():
2218 return proxy_bypass_environment(host)
2219 else:
Ronald Oussoren218cc582010-04-18 20:49:34 +00002220 return proxy_bypass_macosx_sysconf(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002221
2222 def getproxies():
Ronald Oussoren218cc582010-04-18 20:49:34 +00002223 return getproxies_environment() or getproxies_macosx_sysconf()
2224
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002225
2226elif os.name == 'nt':
2227 def getproxies_registry():
2228 """Return a dictionary of scheme -> proxy server URL mappings.
2229
2230 Win32 uses the registry to store proxies.
2231
2232 """
2233 proxies = {}
2234 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002235 import winreg
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002236 except ImportError:
2237 # Std module, so should be around - but you never know!
2238 return proxies
2239 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002240 internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002241 r'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002242 proxyEnable = winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002243 'ProxyEnable')[0]
2244 if proxyEnable:
2245 # Returned as Unicode but problems if not converted to ASCII
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002246 proxyServer = str(winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002247 'ProxyServer')[0])
2248 if '=' in proxyServer:
2249 # Per-protocol settings
2250 for p in proxyServer.split(';'):
2251 protocol, address = p.split('=', 1)
2252 # See if address has a type:// prefix
2253 import re
2254 if not re.match('^([^/:]+)://', address):
2255 address = '%s://%s' % (protocol, address)
2256 proxies[protocol] = address
2257 else:
2258 # Use one setting for all protocols
2259 if proxyServer[:5] == 'http:':
2260 proxies['http'] = proxyServer
2261 else:
2262 proxies['http'] = 'http://%s' % proxyServer
2263 proxies['ftp'] = 'ftp://%s' % proxyServer
2264 internetSettings.Close()
2265 except (WindowsError, ValueError, TypeError):
2266 # Either registry key not found etc, or the value in an
2267 # unexpected format.
2268 # proxies already set up to be empty so nothing to do
2269 pass
2270 return proxies
2271
2272 def getproxies():
2273 """Return a dictionary of scheme -> proxy server URL mappings.
2274
2275 Returns settings gathered from the environment, if specified,
2276 or the registry.
2277
2278 """
2279 return getproxies_environment() or getproxies_registry()
2280
2281 def proxy_bypass_registry(host):
2282 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002283 import winreg
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002284 import re
2285 except ImportError:
2286 # Std modules, so should be around - but you never know!
2287 return 0
2288 try:
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002289 internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002290 r'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002291 proxyEnable = winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002292 'ProxyEnable')[0]
Georg Brandl4ed72ac2009-04-01 04:28:33 +00002293 proxyOverride = str(winreg.QueryValueEx(internetSettings,
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002294 'ProxyOverride')[0])
2295 # ^^^^ Returned as Unicode but problems if not converted to ASCII
2296 except WindowsError:
2297 return 0
2298 if not proxyEnable or not proxyOverride:
2299 return 0
2300 # try to make a host list from name and IP address.
Georg Brandl13e89462008-07-01 19:56:00 +00002301 rawHost, port = splitport(host)
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002302 host = [rawHost]
2303 try:
2304 addr = socket.gethostbyname(rawHost)
2305 if addr != rawHost:
2306 host.append(addr)
2307 except socket.error:
2308 pass
2309 try:
2310 fqdn = socket.getfqdn(rawHost)
2311 if fqdn != rawHost:
2312 host.append(fqdn)
2313 except socket.error:
2314 pass
2315 # make a check value list from the registry entry: replace the
2316 # '<local>' string by the localhost entry and the corresponding
2317 # canonical entry.
2318 proxyOverride = proxyOverride.split(';')
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002319 # now check if we match one of the registry values.
2320 for test in proxyOverride:
Senthil Kumaran49476062009-05-01 06:00:23 +00002321 if test == '<local>':
2322 if '.' not in rawHost:
2323 return 1
Jeremy Hylton1afc1692008-06-18 20:49:58 +00002324 test = test.replace(".", r"\.") # mask dots
2325 test = test.replace("*", r".*") # change glob sequence
2326 test = test.replace("?", r".") # change glob char
2327 for val in host:
2328 # print "%s <--> %s" %( test, val )
2329 if re.match(test, val, re.I):
2330 return 1
2331 return 0
2332
2333 def proxy_bypass(host):
2334 """Return a dictionary of scheme -> proxy server URL mappings.
2335
2336 Returns settings gathered from the environment, if specified,
2337 or the registry.
2338
2339 """
2340 if getproxies_environment():
2341 return proxy_bypass_environment(host)
2342 else:
2343 return proxy_bypass_registry(host)
2344
2345else:
2346 # By default use environment variables
2347 getproxies = getproxies_environment
2348 proxy_bypass = proxy_bypass_environment