summaryrefslogtreecommitdiff
path: root/lib/python2.7/test/test_urllib2_localnet.py
diff options
context:
space:
mode:
Diffstat (limited to 'lib/python2.7/test/test_urllib2_localnet.py')
-rw-r--r--lib/python2.7/test/test_urllib2_localnet.py552
1 files changed, 0 insertions, 552 deletions
diff --git a/lib/python2.7/test/test_urllib2_localnet.py b/lib/python2.7/test/test_urllib2_localnet.py
deleted file mode 100644
index 0802f69..0000000
--- a/lib/python2.7/test/test_urllib2_localnet.py
+++ /dev/null
@@ -1,552 +0,0 @@
-#!/usr/bin/env python
-
-import urlparse
-import urllib2
-import BaseHTTPServer
-import unittest
-import hashlib
-
-from test import test_support
-
-mimetools = test_support.import_module('mimetools', deprecated=True)
-threading = test_support.import_module('threading')
-
-# Loopback http server infrastructure
-
-class LoopbackHttpServer(BaseHTTPServer.HTTPServer):
- """HTTP server w/ a few modifications that make it useful for
- loopback testing purposes.
- """
-
- def __init__(self, server_address, RequestHandlerClass):
- BaseHTTPServer.HTTPServer.__init__(self,
- server_address,
- RequestHandlerClass)
-
- # Set the timeout of our listening socket really low so
- # that we can stop the server easily.
- self.socket.settimeout(1.0)
-
- def get_request(self):
- """BaseHTTPServer method, overridden."""
-
- request, client_address = self.socket.accept()
-
- # It's a loopback connection, so setting the timeout
- # really low shouldn't affect anything, but should make
- # deadlocks less likely to occur.
- request.settimeout(10.0)
-
- return (request, client_address)
-
-class LoopbackHttpServerThread(threading.Thread):
- """Stoppable thread that runs a loopback http server."""
-
- def __init__(self, request_handler):
- threading.Thread.__init__(self)
- self._stop = False
- self.ready = threading.Event()
- request_handler.protocol_version = "HTTP/1.0"
- self.httpd = LoopbackHttpServer(('127.0.0.1', 0),
- request_handler)
- #print "Serving HTTP on %s port %s" % (self.httpd.server_name,
- # self.httpd.server_port)
- self.port = self.httpd.server_port
-
- def stop(self):
- """Stops the webserver if it's currently running."""
-
- # Set the stop flag.
- self._stop = True
-
- self.join()
-
- def run(self):
- self.ready.set()
- while not self._stop:
- self.httpd.handle_request()
-
-# Authentication infrastructure
-
-class DigestAuthHandler:
- """Handler for performing digest authentication."""
-
- def __init__(self):
- self._request_num = 0
- self._nonces = []
- self._users = {}
- self._realm_name = "Test Realm"
- self._qop = "auth"
-
- def set_qop(self, qop):
- self._qop = qop
-
- def set_users(self, users):
- assert isinstance(users, dict)
- self._users = users
-
- def set_realm(self, realm):
- self._realm_name = realm
-
- def _generate_nonce(self):
- self._request_num += 1
- nonce = hashlib.md5(str(self._request_num)).hexdigest()
- self._nonces.append(nonce)
- return nonce
-
- def _create_auth_dict(self, auth_str):
- first_space_index = auth_str.find(" ")
- auth_str = auth_str[first_space_index+1:]
-
- parts = auth_str.split(",")
-
- auth_dict = {}
- for part in parts:
- name, value = part.split("=")
- name = name.strip()
- if value[0] == '"' and value[-1] == '"':
- value = value[1:-1]
- else:
- value = value.strip()
- auth_dict[name] = value
- return auth_dict
-
- def _validate_auth(self, auth_dict, password, method, uri):
- final_dict = {}
- final_dict.update(auth_dict)
- final_dict["password"] = password
- final_dict["method"] = method
- final_dict["uri"] = uri
- HA1_str = "%(username)s:%(realm)s:%(password)s" % final_dict
- HA1 = hashlib.md5(HA1_str).hexdigest()
- HA2_str = "%(method)s:%(uri)s" % final_dict
- HA2 = hashlib.md5(HA2_str).hexdigest()
- final_dict["HA1"] = HA1
- final_dict["HA2"] = HA2
- response_str = "%(HA1)s:%(nonce)s:%(nc)s:" \
- "%(cnonce)s:%(qop)s:%(HA2)s" % final_dict
- response = hashlib.md5(response_str).hexdigest()
-
- return response == auth_dict["response"]
-
- def _return_auth_challenge(self, request_handler):
- request_handler.send_response(407, "Proxy Authentication Required")
- request_handler.send_header("Content-Type", "text/html")
- request_handler.send_header(
- 'Proxy-Authenticate', 'Digest realm="%s", '
- 'qop="%s",'
- 'nonce="%s", ' % \
- (self._realm_name, self._qop, self._generate_nonce()))
- # XXX: Not sure if we're supposed to add this next header or
- # not.
- #request_handler.send_header('Connection', 'close')
- request_handler.end_headers()
- request_handler.wfile.write("Proxy Authentication Required.")
- return False
-
- def handle_request(self, request_handler):
- """Performs digest authentication on the given HTTP request
- handler. Returns True if authentication was successful, False
- otherwise.
-
- If no users have been set, then digest auth is effectively
- disabled and this method will always return True.
- """
-
- if len(self._users) == 0:
- return True
-
- if 'Proxy-Authorization' not in request_handler.headers:
- return self._return_auth_challenge(request_handler)
- else:
- auth_dict = self._create_auth_dict(
- request_handler.headers['Proxy-Authorization']
- )
- if auth_dict["username"] in self._users:
- password = self._users[ auth_dict["username"] ]
- else:
- return self._return_auth_challenge(request_handler)
- if not auth_dict.get("nonce") in self._nonces:
- return self._return_auth_challenge(request_handler)
- else:
- self._nonces.remove(auth_dict["nonce"])
-
- auth_validated = False
-
- # MSIE uses short_path in its validation, but Python's
- # urllib2 uses the full path, so we're going to see if
- # either of them works here.
-
- for path in [request_handler.path, request_handler.short_path]:
- if self._validate_auth(auth_dict,
- password,
- request_handler.command,
- path):
- auth_validated = True
-
- if not auth_validated:
- return self._return_auth_challenge(request_handler)
- return True
-
-# Proxy test infrastructure
-
-class FakeProxyHandler(BaseHTTPServer.BaseHTTPRequestHandler):
- """This is a 'fake proxy' that makes it look like the entire
- internet has gone down due to a sudden zombie invasion. It main
- utility is in providing us with authentication support for
- testing.
- """
-
- def __init__(self, digest_auth_handler, *args, **kwargs):
- # This has to be set before calling our parent's __init__(), which will
- # try to call do_GET().
- self.digest_auth_handler = digest_auth_handler
- BaseHTTPServer.BaseHTTPRequestHandler.__init__(self, *args, **kwargs)
-
- def log_message(self, format, *args):
- # Uncomment the next line for debugging.
- #sys.stderr.write(format % args)
- pass
-
- def do_GET(self):
- (scm, netloc, path, params, query, fragment) = urlparse.urlparse(
- self.path, 'http')
- self.short_path = path
- if self.digest_auth_handler.handle_request(self):
- self.send_response(200, "OK")
- self.send_header("Content-Type", "text/html")
- self.end_headers()
- self.wfile.write("You've reached %s!<BR>" % self.path)
- self.wfile.write("Our apologies, but our server is down due to "
- "a sudden zombie invasion.")
-
-# Test cases
-
-class BaseTestCase(unittest.TestCase):
- def setUp(self):
- self._threads = test_support.threading_setup()
-
- def tearDown(self):
- test_support.threading_cleanup(*self._threads)
-
-
-class ProxyAuthTests(BaseTestCase):
- URL = "http://localhost"
-
- USER = "tester"
- PASSWD = "test123"
- REALM = "TestRealm"
-
- def setUp(self):
- super(ProxyAuthTests, self).setUp()
- self.digest_auth_handler = DigestAuthHandler()
- self.digest_auth_handler.set_users({self.USER: self.PASSWD})
- self.digest_auth_handler.set_realm(self.REALM)
- def create_fake_proxy_handler(*args, **kwargs):
- return FakeProxyHandler(self.digest_auth_handler, *args, **kwargs)
-
- self.server = LoopbackHttpServerThread(create_fake_proxy_handler)
- self.server.start()
- self.server.ready.wait()
- proxy_url = "http://127.0.0.1:%d" % self.server.port
- handler = urllib2.ProxyHandler({"http" : proxy_url})
- self.proxy_digest_handler = urllib2.ProxyDigestAuthHandler()
- self.opener = urllib2.build_opener(handler, self.proxy_digest_handler)
-
- def tearDown(self):
- self.server.stop()
- super(ProxyAuthTests, self).tearDown()
-
- def test_proxy_with_bad_password_raises_httperror(self):
- self.proxy_digest_handler.add_password(self.REALM, self.URL,
- self.USER, self.PASSWD+"bad")
- self.digest_auth_handler.set_qop("auth")
- self.assertRaises(urllib2.HTTPError,
- self.opener.open,
- self.URL)
-
- def test_proxy_with_no_password_raises_httperror(self):
- self.digest_auth_handler.set_qop("auth")
- self.assertRaises(urllib2.HTTPError,
- self.opener.open,
- self.URL)
-
- def test_proxy_qop_auth_works(self):
- self.proxy_digest_handler.add_password(self.REALM, self.URL,
- self.USER, self.PASSWD)
- self.digest_auth_handler.set_qop("auth")
- result = self.opener.open(self.URL)
- while result.read():
- pass
- result.close()
-
- def test_proxy_qop_auth_int_works_or_throws_urlerror(self):
- self.proxy_digest_handler.add_password(self.REALM, self.URL,
- self.USER, self.PASSWD)
- self.digest_auth_handler.set_qop("auth-int")
- try:
- result = self.opener.open(self.URL)
- except urllib2.URLError:
- # It's okay if we don't support auth-int, but we certainly
- # shouldn't receive any kind of exception here other than
- # a URLError.
- result = None
- if result:
- while result.read():
- pass
- result.close()
-
-
-def GetRequestHandler(responses):
-
- class FakeHTTPRequestHandler(BaseHTTPServer.BaseHTTPRequestHandler):
-
- server_version = "TestHTTP/"
- requests = []
- headers_received = []
- port = 80
-
- def do_GET(self):
- body = self.send_head()
- if body:
- self.wfile.write(body)
-
- def do_POST(self):
- content_length = self.headers['Content-Length']
- post_data = self.rfile.read(int(content_length))
- self.do_GET()
- self.requests.append(post_data)
-
- def send_head(self):
- FakeHTTPRequestHandler.headers_received = self.headers
- self.requests.append(self.path)
- response_code, headers, body = responses.pop(0)
-
- self.send_response(response_code)
-
- for (header, value) in headers:
- self.send_header(header, value % self.port)
- if body:
- self.send_header('Content-type', 'text/plain')
- self.end_headers()
- return body
- self.end_headers()
-
- def log_message(self, *args):
- pass
-
-
- return FakeHTTPRequestHandler
-
-
-class TestUrlopen(BaseTestCase):
- """Tests urllib2.urlopen using the network.
-
- These tests are not exhaustive. Assuming that testing using files does a
- good job overall of some of the basic interface features. There are no
- tests exercising the optional 'data' and 'proxies' arguments. No tests
- for transparent redirection have been written.
- """
-
- def setUp(self):
- proxy_handler = urllib2.ProxyHandler({})
- opener = urllib2.build_opener(proxy_handler)
- urllib2.install_opener(opener)
- super(TestUrlopen, self).setUp()
-
- def start_server(self, responses):
- handler = GetRequestHandler(responses)
-
- self.server = LoopbackHttpServerThread(handler)
- self.server.start()
- self.server.ready.wait()
- port = self.server.port
- handler.port = port
- return handler
-
-
- def test_redirection(self):
- expected_response = 'We got here...'
- responses = [
- (302, [('Location', 'http://localhost:%s/somewhere_else')], ''),
- (200, [], expected_response)
- ]
-
- handler = self.start_server(responses)
-
- try:
- f = urllib2.urlopen('http://localhost:%s/' % handler.port)
- data = f.read()
- f.close()
-
- self.assertEqual(data, expected_response)
- self.assertEqual(handler.requests, ['/', '/somewhere_else'])
- finally:
- self.server.stop()
-
-
- def test_404(self):
- expected_response = 'Bad bad bad...'
- handler = self.start_server([(404, [], expected_response)])
-
- try:
- try:
- urllib2.urlopen('http://localhost:%s/weeble' % handler.port)
- except urllib2.URLError, f:
- pass
- else:
- self.fail('404 should raise URLError')
-
- data = f.read()
- f.close()
-
- self.assertEqual(data, expected_response)
- self.assertEqual(handler.requests, ['/weeble'])
- finally:
- self.server.stop()
-
-
- def test_200(self):
- expected_response = 'pycon 2008...'
- handler = self.start_server([(200, [], expected_response)])
-
- try:
- f = urllib2.urlopen('http://localhost:%s/bizarre' % handler.port)
- data = f.read()
- f.close()
-
- self.assertEqual(data, expected_response)
- self.assertEqual(handler.requests, ['/bizarre'])
- finally:
- self.server.stop()
-
- def test_200_with_parameters(self):
- expected_response = 'pycon 2008...'
- handler = self.start_server([(200, [], expected_response)])
-
- try:
- f = urllib2.urlopen('http://localhost:%s/bizarre' % handler.port, 'get=with_feeling')
- data = f.read()
- f.close()
-
- self.assertEqual(data, expected_response)
- self.assertEqual(handler.requests, ['/bizarre', 'get=with_feeling'])
- finally:
- self.server.stop()
-
-
- def test_sending_headers(self):
- handler = self.start_server([(200, [], "we don't care")])
-
- try:
- req = urllib2.Request("http://localhost:%s/" % handler.port,
- headers={'Range': 'bytes=20-39'})
- urllib2.urlopen(req)
- self.assertEqual(handler.headers_received['Range'], 'bytes=20-39')
- finally:
- self.server.stop()
-
- def test_basic(self):
- handler = self.start_server([(200, [], "we don't care")])
-
- try:
- open_url = urllib2.urlopen("http://localhost:%s" % handler.port)
- for attr in ("read", "close", "info", "geturl"):
- self.assertTrue(hasattr(open_url, attr), "object returned from "
- "urlopen lacks the %s attribute" % attr)
- try:
- self.assertTrue(open_url.read(), "calling 'read' failed")
- finally:
- open_url.close()
- finally:
- self.server.stop()
-
- def test_info(self):
- handler = self.start_server([(200, [], "we don't care")])
-
- try:
- open_url = urllib2.urlopen("http://localhost:%s" % handler.port)
- info_obj = open_url.info()
- self.assertIsInstance(info_obj, mimetools.Message,
- "object returned by 'info' is not an "
- "instance of mimetools.Message")
- self.assertEqual(info_obj.getsubtype(), "plain")
- finally:
- self.server.stop()
-
- def test_geturl(self):
- # Make sure same URL as opened is returned by geturl.
- handler = self.start_server([(200, [], "we don't care")])
-
- try:
- open_url = urllib2.urlopen("http://localhost:%s" % handler.port)
- url = open_url.geturl()
- self.assertEqual(url, "http://localhost:%s" % handler.port)
- finally:
- self.server.stop()
-
-
- def test_bad_address(self):
- # Make sure proper exception is raised when connecting to a bogus
- # address.
-
- # as indicated by the comment below, this might fail with some ISP,
- # so we run the test only when -unetwork/-uall is specified to
- # mitigate the problem a bit (see #17564)
- test_support.requires('network')
- self.assertRaises(IOError,
- # Given that both VeriSign and various ISPs have in
- # the past or are presently hijacking various invalid
- # domain name requests in an attempt to boost traffic
- # to their own sites, finding a domain name to use
- # for this test is difficult. RFC2606 leads one to
- # believe that '.invalid' should work, but experience
- # seemed to indicate otherwise. Single character
- # TLDs are likely to remain invalid, so this seems to
- # be the best choice. The trailing '.' prevents a
- # related problem: The normal DNS resolver appends
- # the domain names from the search path if there is
- # no '.' the end and, and if one of those domains
- # implements a '*' rule a result is returned.
- # However, none of this will prevent the test from
- # failing if the ISP hijacks all invalid domain
- # requests. The real solution would be to be able to
- # parameterize the framework with a mock resolver.
- urllib2.urlopen, "http://sadflkjsasf.i.nvali.d./")
-
- def test_iteration(self):
- expected_response = "pycon 2008..."
- handler = self.start_server([(200, [], expected_response)])
- try:
- data = urllib2.urlopen("http://localhost:%s" % handler.port)
- for line in data:
- self.assertEqual(line, expected_response)
- finally:
- self.server.stop()
-
- def ztest_line_iteration(self):
- lines = ["We\n", "got\n", "here\n", "verylong " * 8192 + "\n"]
- expected_response = "".join(lines)
- handler = self.start_server([(200, [], expected_response)])
- try:
- data = urllib2.urlopen("http://localhost:%s" % handler.port)
- for index, line in enumerate(data):
- self.assertEqual(line, lines[index],
- "Fetched line number %s doesn't match expected:\n"
- " Expected length was %s, got %s" %
- (index, len(lines[index]), len(line)))
- finally:
- self.server.stop()
- self.assertEqual(index + 1, len(lines))
-
-def test_main():
- # We will NOT depend on the network resource flag
- # (Lib/test/regrtest.py -u network) since all tests here are only
- # localhost. However, if this is a bad rationale, then uncomment
- # the next line.
- #test_support.requires("network")
-
- test_support.run_unittest(ProxyAuthTests, TestUrlopen)
-
-if __name__ == "__main__":
- test_main()