diff options
Diffstat (limited to 'Lib/test/test_hashlib.py')
-rw-r--r-- | Lib/test/test_hashlib.py | 201 |
1 files changed, 88 insertions, 113 deletions
diff --git a/Lib/test/test_hashlib.py b/Lib/test/test_hashlib.py index e6365c5c5f3..29d3a1cc44f 100644 --- a/Lib/test/test_hashlib.py +++ b/Lib/test/test_hashlib.py @@ -16,20 +16,19 @@ except ImportError: threading = None import unittest import warnings -from test import test_support -from test.test_support import _4G, precisionbigmemtest +from test import support +from test.support import _4G, bigmemtest # Were we compiled --with-pydebug or with #define Py_DEBUG? COMPILED_WITH_PYDEBUG = hasattr(sys, 'gettotalrefcount') def hexstr(s): - import string - h = string.hexdigits + assert isinstance(s, bytes), repr(s) + h = "0123456789abcdef" r = '' - for c in s: - i = ord(c) - r = r + h[(i >> 4) & 0xF] + h[i & 0xF] + for i in s: + r += h[(i >> 4) & 0xF] + h[i & 0xF] return r @@ -44,7 +43,7 @@ class HashLibTestCase(unittest.TestCase): """Import a module and return a reference to it or None on failure.""" try: exec('import '+module_name) - except ImportError, error: + except ImportError as error: if self._warn_on_extension_import: warnings.warn('Did a C extension fail to compile? %s' % error) return locals().get(module_name) @@ -80,10 +79,10 @@ class HashLibTestCase(unittest.TestCase): _md5 = self._conditional_import_module('_md5') if _md5: - self.constructors_to_test['md5'].add(_md5.new) - _sha = self._conditional_import_module('_sha') - if _sha: - self.constructors_to_test['sha1'].add(_sha.new) + self.constructors_to_test['md5'].add(_md5.md5) + _sha1 = self._conditional_import_module('_sha1') + if _sha1: + self.constructors_to_test['sha1'].add(_sha1.sha1) _sha256 = self._conditional_import_module('_sha256') if _sha256: self.constructors_to_test['sha224'].add(_sha256.sha224) @@ -97,15 +96,19 @@ class HashLibTestCase(unittest.TestCase): def test_hash_array(self): a = array.array("b", range(10)) - constructors = self.constructors_to_test.itervalues() + constructors = self.constructors_to_test.values() for cons in itertools.chain.from_iterable(constructors): c = cons(a) c.hexdigest() - def test_algorithms_attribute(self): - self.assertEqual(hashlib.algorithms, - tuple([_algo for _algo in self.supported_hash_names if - _algo.islower()])) + def test_algorithms_guaranteed(self): + self.assertEqual(hashlib.algorithms_guaranteed, + set(_algo for _algo in self.supported_hash_names + if _algo.islower())) + + def test_algorithms_available(self): + self.assertTrue(set(hashlib.algorithms_guaranteed). + issubset(hashlib.algorithms_available)) def test_unknown_hash(self): self.assertRaises(ValueError, hashlib.new, 'spam spam spam spam spam') @@ -133,13 +136,14 @@ class HashLibTestCase(unittest.TestCase): def test_hexdigest(self): for name in self.supported_hash_names: h = hashlib.new(name) - self.assertTrue(hexstr(h.digest()) == h.hexdigest()) + assert isinstance(h.digest(), bytes), name + self.assertEqual(hexstr(h.digest()), h.hexdigest()) + def test_large_update(self): - aas = 'a' * 128 - bees = 'b' * 127 - cees = 'c' * 126 - abcs = aas + bees + cees + aas = b'a' * 128 + bees = b'b' * 127 + cees = b'c' * 126 for name in self.supported_hash_names: m1 = hashlib.new(name) @@ -148,11 +152,8 @@ class HashLibTestCase(unittest.TestCase): m1.update(cees) m2 = hashlib.new(name) - m2.update(abcs) - self.assertEqual(m1.digest(), m2.digest(), name+' update problem.') - - m3 = hashlib.new(name, abcs) - self.assertEqual(m1.digest(), m3.digest(), name+' new problem.') + m2.update(aas + bees + cees) + self.assertEqual(m1.digest(), m2.digest()) def check(self, name, data, digest): constructors = self.constructors_to_test[name] @@ -167,68 +168,44 @@ class HashLibTestCase(unittest.TestCase): % (name, hash_object_constructor, computed, len(data), digest)) - def check_update(self, name, data, digest): - constructors = self.constructors_to_test[name] - # 2 is for hashlib.name(...) and hashlib.new(name, ...) - self.assertGreaterEqual(len(constructors), 2) + def check_no_unicode(self, algorithm_name): + # Unicode objects are not allowed as input. + constructors = self.constructors_to_test[algorithm_name] for hash_object_constructor in constructors: - h = hash_object_constructor() - h.update(data) - computed = h.hexdigest() - self.assertEqual( - computed, digest, - "Hash algorithm %s using %s when updated returned hexdigest" - " %r for %d byte input data that should have hashed to %r." - % (name, hash_object_constructor, - computed, len(data), digest)) + self.assertRaises(TypeError, hash_object_constructor, 'spam') - def check_unicode(self, algorithm_name): - # Unicode objects are not allowed as input. - expected = hashlib.new(algorithm_name, str(u'spam')).hexdigest() - self.check(algorithm_name, u'spam', expected) - - def test_unicode(self): - # In python 2.x unicode is auto-encoded to the system default encoding - # when passed to hashlib functions. - self.check_unicode('md5') - self.check_unicode('sha1') - self.check_unicode('sha224') - self.check_unicode('sha256') - self.check_unicode('sha384') - self.check_unicode('sha512') + def test_no_unicode(self): + self.check_no_unicode('md5') + self.check_no_unicode('sha1') + self.check_no_unicode('sha224') + self.check_no_unicode('sha256') + self.check_no_unicode('sha384') + self.check_no_unicode('sha512') def test_case_md5_0(self): - self.check('md5', '', 'd41d8cd98f00b204e9800998ecf8427e') + self.check('md5', b'', 'd41d8cd98f00b204e9800998ecf8427e') def test_case_md5_1(self): - self.check('md5', 'abc', '900150983cd24fb0d6963f7d28e17f72') + self.check('md5', b'abc', '900150983cd24fb0d6963f7d28e17f72') def test_case_md5_2(self): - self.check('md5', 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789', + self.check('md5', + b'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789', 'd174ab98d277d9f5a5611c2c9f419d9f') - @precisionbigmemtest(size=_4G + 5, memuse=1) + @bigmemtest(size=_4G + 5, memuse=1) def test_case_md5_huge(self, size): if size == _4G + 5: try: - self.check('md5', 'A'*size, 'c9af2dff37468ce5dfee8f2cfc0a9c6d') - except OverflowError: - pass # 32-bit arch - - @precisionbigmemtest(size=_4G + 5, memuse=1) - def test_case_md5_huge_update(self, size): - if size == _4G + 5: - try: - self.check_update('md5', 'A'*size, - 'c9af2dff37468ce5dfee8f2cfc0a9c6d') + self.check('md5', b'A'*size, 'c9af2dff37468ce5dfee8f2cfc0a9c6d') except OverflowError: pass # 32-bit arch - @precisionbigmemtest(size=_4G - 1, memuse=1) + @bigmemtest(size=_4G - 1, memuse=1) def test_case_md5_uintmax(self, size): if size == _4G - 1: try: - self.check('md5', 'A'*size, '28138d306ff1b8281f1a9067e1a1a2b3') + self.check('md5', b'A'*size, '28138d306ff1b8281f1a9067e1a1a2b3') except OverflowError: pass # 32-bit arch @@ -237,126 +214,124 @@ class HashLibTestCase(unittest.TestCase): # http://www.itl.nist.gov/div897/pubs/fip180-1.htm def test_case_sha1_0(self): - self.check('sha1', "", + self.check('sha1', b"", "da39a3ee5e6b4b0d3255bfef95601890afd80709") def test_case_sha1_1(self): - self.check('sha1', "abc", + self.check('sha1', b"abc", "a9993e364706816aba3e25717850c26c9cd0d89d") def test_case_sha1_2(self): - self.check('sha1', "abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", + self.check('sha1', + b"abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", "84983e441c3bd26ebaae4aa1f95129e5e54670f1") def test_case_sha1_3(self): - self.check('sha1', "a" * 1000000, + self.check('sha1', b"a" * 1000000, "34aa973cd4c4daa4f61eeb2bdbad27316534016f") - @precisionbigmemtest(size=_4G + 5, memuse=1) - def test_case_sha1_huge(self, size): - if size == _4G + 5: - try: - self.check('sha1', 'A'*size, - '87d745c50e6b2879ffa0fb2c930e9fbfe0dc9a5b') - except OverflowError: - pass # 32-bit arch - - @precisionbigmemtest(size=_4G + 5, memuse=1) - def test_case_sha1_huge_update(self, size): - if size == _4G + 5: - try: - self.check_update('sha1', 'A'*size, - '87d745c50e6b2879ffa0fb2c930e9fbfe0dc9a5b') - except OverflowError: - pass # 32-bit arch # use the examples from Federal Information Processing Standards # Publication 180-2, Secure Hash Standard, 2002 August 1 # http://csrc.nist.gov/publications/fips/fips180-2/fips180-2.pdf def test_case_sha224_0(self): - self.check('sha224', "", + self.check('sha224', b"", "d14a028c2a3a2bc9476102bb288234c415a2b01f828ea62ac5b3e42f") def test_case_sha224_1(self): - self.check('sha224', "abc", + self.check('sha224', b"abc", "23097d223405d8228642a477bda255b32aadbce4bda0b3f7e36c9da7") def test_case_sha224_2(self): self.check('sha224', - "abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", + b"abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", "75388b16512776cc5dba5da1fd890150b0c6455cb4f58b1952522525") def test_case_sha224_3(self): - self.check('sha224', "a" * 1000000, + self.check('sha224', b"a" * 1000000, "20794655980c91d8bbb4c1ea97618a4bf03f42581948b2ee4ee7ad67") def test_case_sha256_0(self): - self.check('sha256', "", + self.check('sha256', b"", "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855") def test_case_sha256_1(self): - self.check('sha256', "abc", + self.check('sha256', b"abc", "ba7816bf8f01cfea414140de5dae2223b00361a396177a9cb410ff61f20015ad") def test_case_sha256_2(self): self.check('sha256', - "abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", + b"abcdbcdecdefdefgefghfghighijhijkijkljklmklmnlmnomnopnopq", "248d6a61d20638b8e5c026930c3e6039a33ce45964ff2167f6ecedd419db06c1") def test_case_sha256_3(self): - self.check('sha256', "a" * 1000000, + self.check('sha256', b"a" * 1000000, "cdc76e5c9914fb9281a1c7e284d73e67f1809a48a497200e046d39ccc7112cd0") def test_case_sha384_0(self): - self.check('sha384', "", + self.check('sha384', b"", "38b060a751ac96384cd9327eb1b1e36a21fdb71114be07434c0cc7bf63f6e1da"+ "274edebfe76f65fbd51ad2f14898b95b") def test_case_sha384_1(self): - self.check('sha384', "abc", + self.check('sha384', b"abc", "cb00753f45a35e8bb5a03d699ac65007272c32ab0eded1631a8b605a43ff5bed"+ "8086072ba1e7cc2358baeca134c825a7") def test_case_sha384_2(self): self.check('sha384', - "abcdefghbcdefghicdefghijdefghijkefghijklfghijklmghijklmn"+ - "hijklmnoijklmnopjklmnopqklmnopqrlmnopqrsmnopqrstnopqrstu", + b"abcdefghbcdefghicdefghijdefghijkefghijklfghijklmghijklmn"+ + b"hijklmnoijklmnopjklmnopqklmnopqrlmnopqrsmnopqrstnopqrstu", "09330c33f71147e83d192fc782cd1b4753111b173b3b05d22fa08086e3b0f712"+ "fcc7c71a557e2db966c3e9fa91746039") def test_case_sha384_3(self): - self.check('sha384', "a" * 1000000, + self.check('sha384', b"a" * 1000000, "9d0e1809716474cb086e834e310a4a1ced149e9c00f248527972cec5704c2a5b"+ "07b8b3dc38ecc4ebae97ddd87f3d8985") def test_case_sha512_0(self): - self.check('sha512', "", + self.check('sha512', b"", "cf83e1357eefb8bdf1542850d66d8007d620e4050b5715dc83f4a921d36ce9ce"+ "47d0d13c5d85f2b0ff8318d2877eec2f63b931bd47417a81a538327af927da3e") def test_case_sha512_1(self): - self.check('sha512', "abc", + self.check('sha512', b"abc", "ddaf35a193617abacc417349ae20413112e6fa4e89a97ea20a9eeee64b55d39a"+ "2192992a274fc1a836ba3c23a3feebbd454d4423643ce80e2a9ac94fa54ca49f") def test_case_sha512_2(self): self.check('sha512', - "abcdefghbcdefghicdefghijdefghijkefghijklfghijklmghijklmn"+ - "hijklmnoijklmnopjklmnopqklmnopqrlmnopqrsmnopqrstnopqrstu", + b"abcdefghbcdefghicdefghijdefghijkefghijklfghijklmghijklmn"+ + b"hijklmnoijklmnopjklmnopqklmnopqrlmnopqrsmnopqrstnopqrstu", "8e959b75dae313da8cf4f72814fc143f8f7779c6eb9f7fa17299aeadb6889018"+ "501d289e4900f7e4331b99dec4b5433ac7d329eeb6dd26545e96e55b874be909") def test_case_sha512_3(self): - self.check('sha512', "a" * 1000000, + self.check('sha512', b"a" * 1000000, "e718483d0ce769644e2e42c7bc15b4638e1f98b13b2044285632a803afa973eb"+ "de0ff244877ea60a4cb0432ce577c31beb009c5c2c49aa2e4eadb217ad8cc09b") + def test_gil(self): + # Check things work fine with an input larger than the size required + # for multithreaded operation (which is hardwired to 2048). + gil_minsize = 2048 + + m = hashlib.md5() + m.update(b'1') + m.update(b'#' * gil_minsize) + m.update(b'1') + self.assertEqual(m.hexdigest(), 'cb1e1a2cbc80be75e19935d621fb9b21') + + m = hashlib.md5(b'x' * gil_minsize) + self.assertEqual(m.hexdigest(), 'cfb767f225d58469c5de3632a8803958') + @unittest.skipUnless(threading, 'Threading required for this test.') - @test_support.reap_threads + @support.reap_threads def test_threaded_hashing(self): # Updating the same hash object from several threads at once # using data chunk sizes containing the same byte sequences. @@ -366,7 +341,7 @@ class HashLibTestCase(unittest.TestCase): # hash will be the same as doing it single threaded upfront. hasher = hashlib.sha1() num_threads = 5 - smallest_data = 'swineflu' + smallest_data = b'swineflu' data = smallest_data*200000 expected_hash = hashlib.sha1(data*num_threads).hexdigest() @@ -378,7 +353,7 @@ class HashLibTestCase(unittest.TestCase): event.set() events = [] - for threadnum in xrange(num_threads): + for threadnum in range(num_threads): chunk_size = len(data) // (10**threadnum) assert chunk_size > 0 assert chunk_size % len(smallest_data) == 0 @@ -393,7 +368,7 @@ class HashLibTestCase(unittest.TestCase): self.assertEqual(expected_hash, hasher.hexdigest()) def test_main(): - test_support.run_unittest(HashLibTestCase) + support.run_unittest(HashLibTestCase) if __name__ == "__main__": test_main() |