mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-05-22 21:32:31 +00:00
Undiamond Python headers
This change gets the Python codebase into a state where it conforms to the conventions of this codebase. It's now possible to include headers from Python, without worrying about ordering. Python has traditionally solved that problem by "diamonding" everything in Python.h, but that's problematic since it means any change to any Python header invalidates all the build artifacts. Lastly it makes tooling not work. Since it is hard to explain to Emacs when I press C-c C-h to add an import line it shouldn't add the header that actually defines the symbol, and instead do follow the nonstandard Python convention. Progress has been made on letting Python load source code from the zip executable structure via the standard C library APIs. System calss now recognizes zip!FILENAME alternative URIs as equivalent to zip:FILENAME since Python uses colon as its delimiter. Some progress has been made on embedding the notice license terms into the Python object code. This is easier said than done since Python has an extremely complicated ownership story. - Some termios APIs have been added - Implement rewinddir() dirstream API - GetCpuCount() API added to Cosmopolitan Libc - More bugs in Cosmopolitan Libc have been fixed - zipobj.com now has flags for mangling the path - Fixed bug a priori with sendfile() on certain BSDs - Polyfill F_DUPFD and F_DUPFD_CLOEXEC across platforms - FIOCLEX / FIONCLEX now polyfilled for fast O_CLOEXEC changes - APE now supports a hybrid solution to no-self-modify for builds - Many BSD-only magnums added, e.g. O_SEARCH, O_SHLOCK, SF_NODISKIO
This commit is contained in:
parent
20bb8db9f8
commit
b420ed8248
762 changed files with 18410 additions and 53772 deletions
|
@ -235,13 +235,6 @@ def clear_caches():
|
|||
else:
|
||||
doctest.master = None
|
||||
|
||||
try:
|
||||
ctypes = sys.modules['ctypes']
|
||||
except KeyError:
|
||||
pass
|
||||
else:
|
||||
ctypes._reset_cache()
|
||||
|
||||
try:
|
||||
typing = sys.modules['typing']
|
||||
except KeyError:
|
||||
|
|
215
third_party/python/Lib/test/test_hashlib.py
vendored
215
third_party/python/Lib/test/test_hashlib.py
vendored
|
@ -30,13 +30,6 @@ COMPILED_WITH_PYDEBUG = hasattr(sys, 'gettotalrefcount')
|
|||
c_hashlib = import_fresh_module('hashlib', fresh=['_hashlib'])
|
||||
py_hashlib = import_fresh_module('hashlib', blocked=['_hashlib'])
|
||||
|
||||
try:
|
||||
import _blake2
|
||||
except ImportError:
|
||||
_blake2 = None
|
||||
|
||||
requires_blake2 = unittest.skipUnless(_blake2, 'requires _blake2')
|
||||
|
||||
try:
|
||||
import _sha3
|
||||
except ImportError:
|
||||
|
@ -76,7 +69,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
supported_hash_names = ( 'md5', 'MD5', 'sha1', 'SHA1',
|
||||
'sha224', 'SHA224', 'sha256', 'SHA256',
|
||||
'sha384', 'SHA384', 'sha512', 'SHA512',
|
||||
'blake2b', 'blake2s',
|
||||
'sha3_224', 'sha3_256', 'sha3_384', 'sha3_512',
|
||||
'shake_128', 'shake_256')
|
||||
|
||||
|
@ -99,10 +91,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
for algorithm in self.supported_hash_names:
|
||||
algorithms.add(algorithm.lower())
|
||||
|
||||
_blake2 = self._conditional_import_module('_blake2')
|
||||
if _blake2:
|
||||
algorithms.update({'blake2b', 'blake2s'})
|
||||
|
||||
self.constructors_to_test = {}
|
||||
for algorithm in algorithms:
|
||||
self.constructors_to_test[algorithm] = set()
|
||||
|
@ -146,10 +134,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
if _sha512:
|
||||
add_builtin_constructor('sha384')
|
||||
add_builtin_constructor('sha512')
|
||||
if _blake2:
|
||||
add_builtin_constructor('blake2s')
|
||||
add_builtin_constructor('blake2b')
|
||||
|
||||
_sha3 = self._conditional_import_module('_sha3')
|
||||
if _sha3:
|
||||
add_builtin_constructor('sha3_224')
|
||||
|
@ -325,11 +309,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
self.check_no_unicode('sha384')
|
||||
self.check_no_unicode('sha512')
|
||||
|
||||
@requires_blake2
|
||||
def test_no_unicode_blake2(self):
|
||||
self.check_no_unicode('blake2b')
|
||||
self.check_no_unicode('blake2s')
|
||||
|
||||
@requires_sha3
|
||||
def test_no_unicode_sha3(self):
|
||||
self.check_no_unicode('sha3_224')
|
||||
|
@ -393,11 +372,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
self.check_sha3('shake_128', 256, 1344, b'\x1f')
|
||||
self.check_sha3('shake_256', 512, 1088, b'\x1f')
|
||||
|
||||
@requires_blake2
|
||||
def test_blocksize_name_blake2(self):
|
||||
self.check_blocksize_name('blake2b', 128, 64)
|
||||
self.check_blocksize_name('blake2s', 64, 32)
|
||||
|
||||
def test_case_md5_0(self):
|
||||
self.check('md5', b'', 'd41d8cd98f00b204e9800998ecf8427e')
|
||||
|
||||
|
@ -526,195 +500,6 @@ class HashLibTestCase(unittest.TestCase):
|
|||
"e718483d0ce769644e2e42c7bc15b4638e1f98b13b2044285632a803afa973eb"+
|
||||
"de0ff244877ea60a4cb0432ce577c31beb009c5c2c49aa2e4eadb217ad8cc09b")
|
||||
|
||||
def check_blake2(self, constructor, salt_size, person_size, key_size,
|
||||
digest_size, max_offset):
|
||||
self.assertEqual(constructor.SALT_SIZE, salt_size)
|
||||
for i in range(salt_size + 1):
|
||||
constructor(salt=b'a' * i)
|
||||
salt = b'a' * (salt_size + 1)
|
||||
self.assertRaises(ValueError, constructor, salt=salt)
|
||||
|
||||
self.assertEqual(constructor.PERSON_SIZE, person_size)
|
||||
for i in range(person_size+1):
|
||||
constructor(person=b'a' * i)
|
||||
person = b'a' * (person_size + 1)
|
||||
self.assertRaises(ValueError, constructor, person=person)
|
||||
|
||||
self.assertEqual(constructor.MAX_DIGEST_SIZE, digest_size)
|
||||
for i in range(1, digest_size + 1):
|
||||
constructor(digest_size=i)
|
||||
self.assertRaises(ValueError, constructor, digest_size=-1)
|
||||
self.assertRaises(ValueError, constructor, digest_size=0)
|
||||
self.assertRaises(ValueError, constructor, digest_size=digest_size+1)
|
||||
|
||||
self.assertEqual(constructor.MAX_KEY_SIZE, key_size)
|
||||
for i in range(key_size+1):
|
||||
constructor(key=b'a' * i)
|
||||
key = b'a' * (key_size + 1)
|
||||
self.assertRaises(ValueError, constructor, key=key)
|
||||
self.assertEqual(constructor().hexdigest(),
|
||||
constructor(key=b'').hexdigest())
|
||||
|
||||
for i in range(0, 256):
|
||||
constructor(fanout=i)
|
||||
self.assertRaises(ValueError, constructor, fanout=-1)
|
||||
self.assertRaises(ValueError, constructor, fanout=256)
|
||||
|
||||
for i in range(1, 256):
|
||||
constructor(depth=i)
|
||||
self.assertRaises(ValueError, constructor, depth=-1)
|
||||
self.assertRaises(ValueError, constructor, depth=0)
|
||||
self.assertRaises(ValueError, constructor, depth=256)
|
||||
|
||||
for i in range(0, 256):
|
||||
constructor(node_depth=i)
|
||||
self.assertRaises(ValueError, constructor, node_depth=-1)
|
||||
self.assertRaises(ValueError, constructor, node_depth=256)
|
||||
|
||||
for i in range(0, digest_size + 1):
|
||||
constructor(inner_size=i)
|
||||
self.assertRaises(ValueError, constructor, inner_size=-1)
|
||||
self.assertRaises(ValueError, constructor, inner_size=digest_size+1)
|
||||
|
||||
constructor(leaf_size=0)
|
||||
constructor(leaf_size=(1<<32)-1)
|
||||
self.assertRaises(OverflowError, constructor, leaf_size=-1)
|
||||
self.assertRaises(OverflowError, constructor, leaf_size=1<<32)
|
||||
|
||||
constructor(node_offset=0)
|
||||
constructor(node_offset=max_offset)
|
||||
self.assertRaises(OverflowError, constructor, node_offset=-1)
|
||||
self.assertRaises(OverflowError, constructor, node_offset=max_offset+1)
|
||||
|
||||
self.assertRaises(TypeError, constructor, data=b'')
|
||||
self.assertRaises(TypeError, constructor, string=b'')
|
||||
self.assertRaises(TypeError, constructor, '')
|
||||
|
||||
constructor(
|
||||
b'',
|
||||
key=b'',
|
||||
salt=b'',
|
||||
person=b'',
|
||||
digest_size=17,
|
||||
fanout=1,
|
||||
depth=1,
|
||||
leaf_size=256,
|
||||
node_offset=512,
|
||||
node_depth=1,
|
||||
inner_size=7,
|
||||
last_node=True
|
||||
)
|
||||
|
||||
def blake2_rfc7693(self, constructor, md_len, in_len):
|
||||
def selftest_seq(length, seed):
|
||||
mask = (1<<32)-1
|
||||
a = (0xDEAD4BAD * seed) & mask
|
||||
b = 1
|
||||
out = bytearray(length)
|
||||
for i in range(length):
|
||||
t = (a + b) & mask
|
||||
a, b = b, t
|
||||
out[i] = (t >> 24) & 0xFF
|
||||
return out
|
||||
outer = constructor(digest_size=32)
|
||||
for outlen in md_len:
|
||||
for inlen in in_len:
|
||||
indata = selftest_seq(inlen, inlen)
|
||||
key = selftest_seq(outlen, outlen)
|
||||
unkeyed = constructor(indata, digest_size=outlen)
|
||||
outer.update(unkeyed.digest())
|
||||
keyed = constructor(indata, key=key, digest_size=outlen)
|
||||
outer.update(keyed.digest())
|
||||
return outer.hexdigest()
|
||||
|
||||
@requires_blake2
|
||||
def test_blake2b(self):
|
||||
self.check_blake2(hashlib.blake2b, 16, 16, 64, 64, (1<<64)-1)
|
||||
b2b_md_len = [20, 32, 48, 64]
|
||||
b2b_in_len = [0, 3, 128, 129, 255, 1024]
|
||||
self.assertEqual(
|
||||
self.blake2_rfc7693(hashlib.blake2b, b2b_md_len, b2b_in_len),
|
||||
"c23a7800d98123bd10f506c61e29da5603d763b8bbad2e737f5e765a7bccd475")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2b_0(self):
|
||||
self.check('blake2b', b"",
|
||||
"786a02f742015903c6c6fd852552d272912f4740e15847618a86e217f71f5419"+
|
||||
"d25e1031afee585313896444934eb04b903a685b1448b755d56f701afe9be2ce")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2b_1(self):
|
||||
self.check('blake2b', b"abc",
|
||||
"ba80a53f981c4d0d6a2797b69f12f6e94c212f14685ac4b74b12bb6fdbffa2d1"+
|
||||
"7d87c5392aab792dc252d5de4533cc9518d38aa8dbf1925ab92386edd4009923")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2b_all_parameters(self):
|
||||
# This checks that all the parameters work in general, and also that
|
||||
# parameter byte order doesn't get confused on big endian platforms.
|
||||
self.check('blake2b', b"foo",
|
||||
"920568b0c5873b2f0ab67bedb6cf1b2b",
|
||||
digest_size=16,
|
||||
key=b"bar",
|
||||
salt=b"baz",
|
||||
person=b"bing",
|
||||
fanout=2,
|
||||
depth=3,
|
||||
leaf_size=4,
|
||||
node_offset=5,
|
||||
node_depth=6,
|
||||
inner_size=7,
|
||||
last_node=True)
|
||||
|
||||
@requires_blake2
|
||||
def test_blake2b_vectors(self):
|
||||
for msg, key, md in read_vectors('blake2b'):
|
||||
key = bytes.fromhex(key)
|
||||
self.check('blake2b', msg, md, key=key)
|
||||
|
||||
@requires_blake2
|
||||
def test_blake2s(self):
|
||||
self.check_blake2(hashlib.blake2s, 8, 8, 32, 32, (1<<48)-1)
|
||||
b2s_md_len = [16, 20, 28, 32]
|
||||
b2s_in_len = [0, 3, 64, 65, 255, 1024]
|
||||
self.assertEqual(
|
||||
self.blake2_rfc7693(hashlib.blake2s, b2s_md_len, b2s_in_len),
|
||||
"6a411f08ce25adcdfb02aba641451cec53c598b24f4fc787fbdc88797f4c1dfe")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2s_0(self):
|
||||
self.check('blake2s', b"",
|
||||
"69217a3079908094e11121d042354a7c1f55b6482ca1a51e1b250dfd1ed0eef9")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2s_1(self):
|
||||
self.check('blake2s', b"abc",
|
||||
"508c5e8c327c14e2e1a72ba34eeb452f37458b209ed63a294d999b4c86675982")
|
||||
|
||||
@requires_blake2
|
||||
def test_case_blake2s_all_parameters(self):
|
||||
# This checks that all the parameters work in general, and also that
|
||||
# parameter byte order doesn't get confused on big endian platforms.
|
||||
self.check('blake2s', b"foo",
|
||||
"bf2a8f7fe3c555012a6f8046e646bc75",
|
||||
digest_size=16,
|
||||
key=b"bar",
|
||||
salt=b"baz",
|
||||
person=b"bing",
|
||||
fanout=2,
|
||||
depth=3,
|
||||
leaf_size=4,
|
||||
node_offset=5,
|
||||
node_depth=6,
|
||||
inner_size=7,
|
||||
last_node=True)
|
||||
|
||||
@requires_blake2
|
||||
def test_blake2s_vectors(self):
|
||||
for msg, key, md in read_vectors('blake2s'):
|
||||
key = bytes.fromhex(key)
|
||||
self.check('blake2s', msg, md, key=key)
|
||||
|
||||
@requires_sha3
|
||||
def test_case_sha3_224_0(self):
|
||||
self.check('sha3_224', b"",
|
||||
|
|
87
third_party/python/Lib/test/test_threading.py
vendored
87
third_party/python/Lib/test/test_threading.py
vendored
|
@ -177,93 +177,6 @@ class ThreadTests(BaseTestCase):
|
|||
self.assertRegex(repr(threading._active[tid]), '_DummyThread')
|
||||
del threading._active[tid]
|
||||
|
||||
# PyThreadState_SetAsyncExc() is a CPython-only gimmick, not (currently)
|
||||
# exposed at the Python level. This test relies on ctypes to get at it.
|
||||
def test_PyThreadState_SetAsyncExc(self):
|
||||
ctypes = import_module("ctypes")
|
||||
|
||||
set_async_exc = ctypes.pythonapi.PyThreadState_SetAsyncExc
|
||||
|
||||
class AsyncExc(Exception):
|
||||
pass
|
||||
|
||||
exception = ctypes.py_object(AsyncExc)
|
||||
|
||||
# First check it works when setting the exception from the same thread.
|
||||
tid = threading.get_ident()
|
||||
|
||||
try:
|
||||
result = set_async_exc(ctypes.c_long(tid), exception)
|
||||
# The exception is async, so we might have to keep the VM busy until
|
||||
# it notices.
|
||||
while True:
|
||||
pass
|
||||
except AsyncExc:
|
||||
pass
|
||||
else:
|
||||
# This code is unreachable but it reflects the intent. If we wanted
|
||||
# to be smarter the above loop wouldn't be infinite.
|
||||
self.fail("AsyncExc not raised")
|
||||
try:
|
||||
self.assertEqual(result, 1) # one thread state modified
|
||||
except UnboundLocalError:
|
||||
# The exception was raised too quickly for us to get the result.
|
||||
pass
|
||||
|
||||
# `worker_started` is set by the thread when it's inside a try/except
|
||||
# block waiting to catch the asynchronously set AsyncExc exception.
|
||||
# `worker_saw_exception` is set by the thread upon catching that
|
||||
# exception.
|
||||
worker_started = threading.Event()
|
||||
worker_saw_exception = threading.Event()
|
||||
|
||||
class Worker(threading.Thread):
|
||||
def run(self):
|
||||
self.id = threading.get_ident()
|
||||
self.finished = False
|
||||
|
||||
try:
|
||||
while True:
|
||||
worker_started.set()
|
||||
time.sleep(0.1)
|
||||
except AsyncExc:
|
||||
self.finished = True
|
||||
worker_saw_exception.set()
|
||||
|
||||
t = Worker()
|
||||
t.daemon = True # so if this fails, we don't hang Python at shutdown
|
||||
t.start()
|
||||
if verbose:
|
||||
print(" started worker thread")
|
||||
|
||||
# Try a thread id that doesn't make sense.
|
||||
if verbose:
|
||||
print(" trying nonsensical thread id")
|
||||
result = set_async_exc(ctypes.c_long(-1), exception)
|
||||
self.assertEqual(result, 0) # no thread states modified
|
||||
|
||||
# Now raise an exception in the worker thread.
|
||||
if verbose:
|
||||
print(" waiting for worker thread to get started")
|
||||
ret = worker_started.wait()
|
||||
self.assertTrue(ret)
|
||||
if verbose:
|
||||
print(" verifying worker hasn't exited")
|
||||
self.assertFalse(t.finished)
|
||||
if verbose:
|
||||
print(" attempting to raise asynch exception in worker")
|
||||
result = set_async_exc(ctypes.c_long(t.id), exception)
|
||||
self.assertEqual(result, 1) # one thread state modified
|
||||
if verbose:
|
||||
print(" waiting for worker to say it caught the exception")
|
||||
worker_saw_exception.wait(timeout=10)
|
||||
self.assertTrue(t.finished)
|
||||
if verbose:
|
||||
print(" all OK -- joining worker")
|
||||
if t.finished:
|
||||
t.join()
|
||||
# else the thread is still running, and we have no way to kill it
|
||||
|
||||
def test_limbo_cleanup(self):
|
||||
# Issue 7481: Failure to start thread should cleanup the limbo map.
|
||||
def fail_new_thread(*args):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue