diff options
Diffstat (limited to 'lib/python2.7/test/test_threading.py')
-rw-r--r-- | lib/python2.7/test/test_threading.py | 778 |
1 files changed, 0 insertions, 778 deletions
diff --git a/lib/python2.7/test/test_threading.py b/lib/python2.7/test/test_threading.py deleted file mode 100644 index ef04cd3..0000000 --- a/lib/python2.7/test/test_threading.py +++ /dev/null @@ -1,778 +0,0 @@ -# Very rudimentary test of threading module - -import test.test_support -from test.test_support import verbose -from test.script_helper import assert_python_ok - -import random -import re -import sys -thread = test.test_support.import_module('thread') -threading = test.test_support.import_module('threading') -import time -import unittest -import weakref -import os -import subprocess - -from test import lock_tests - -# A trivial mutable counter. -class Counter(object): - def __init__(self): - self.value = 0 - def inc(self): - self.value += 1 - def dec(self): - self.value -= 1 - def get(self): - return self.value - -class TestThread(threading.Thread): - def __init__(self, name, testcase, sema, mutex, nrunning): - threading.Thread.__init__(self, name=name) - self.testcase = testcase - self.sema = sema - self.mutex = mutex - self.nrunning = nrunning - - def run(self): - delay = random.random() / 10000.0 - if verbose: - print 'task %s will run for %.1f usec' % ( - self.name, delay * 1e6) - - with self.sema: - with self.mutex: - self.nrunning.inc() - if verbose: - print self.nrunning.get(), 'tasks are running' - self.testcase.assertTrue(self.nrunning.get() <= 3) - - time.sleep(delay) - if verbose: - print 'task', self.name, 'done' - - with self.mutex: - self.nrunning.dec() - self.testcase.assertTrue(self.nrunning.get() >= 0) - if verbose: - print '%s is finished. %d tasks are running' % ( - self.name, self.nrunning.get()) - -class BaseTestCase(unittest.TestCase): - def setUp(self): - self._threads = test.test_support.threading_setup() - - def tearDown(self): - test.test_support.threading_cleanup(*self._threads) - test.test_support.reap_children() - - -class ThreadTests(BaseTestCase): - - # Create a bunch of threads, let each do some work, wait until all are - # done. - def test_various_ops(self): - # This takes about n/3 seconds to run (about n/3 clumps of tasks, - # times about 1 second per clump). - NUMTASKS = 10 - - # no more than 3 of the 10 can run at once - sema = threading.BoundedSemaphore(value=3) - mutex = threading.RLock() - numrunning = Counter() - - threads = [] - - for i in range(NUMTASKS): - t = TestThread("<thread %d>"%i, self, sema, mutex, numrunning) - threads.append(t) - self.assertEqual(t.ident, None) - self.assertTrue(re.match('<TestThread\(.*, initial\)>', repr(t))) - t.start() - - if verbose: - print 'waiting for all tasks to complete' - for t in threads: - t.join(NUMTASKS) - self.assertTrue(not t.is_alive()) - self.assertNotEqual(t.ident, 0) - self.assertFalse(t.ident is None) - self.assertTrue(re.match('<TestThread\(.*, \w+ -?\d+\)>', repr(t))) - if verbose: - print 'all tasks done' - self.assertEqual(numrunning.get(), 0) - - def test_ident_of_no_threading_threads(self): - # The ident still must work for the main thread and dummy threads. - self.assertFalse(threading.currentThread().ident is None) - def f(): - ident.append(threading.currentThread().ident) - done.set() - done = threading.Event() - ident = [] - thread.start_new_thread(f, ()) - done.wait() - self.assertFalse(ident[0] is None) - # Kill the "immortal" _DummyThread - del threading._active[ident[0]] - - # run with a small(ish) thread stack size (256kB) - def test_various_ops_small_stack(self): - if verbose: - print 'with 256kB thread stack size...' - try: - threading.stack_size(262144) - except thread.error: - if verbose: - print 'platform does not support changing thread stack size' - return - self.test_various_ops() - threading.stack_size(0) - - # run with a large thread stack size (1MB) - def test_various_ops_large_stack(self): - if verbose: - print 'with 1MB thread stack size...' - try: - threading.stack_size(0x100000) - except thread.error: - if verbose: - print 'platform does not support changing thread stack size' - return - self.test_various_ops() - threading.stack_size(0) - - def test_foreign_thread(self): - # Check that a "foreign" thread can use the threading module. - def f(mutex): - # Calling current_thread() forces an entry for the foreign - # thread to get made in the threading._active map. - threading.current_thread() - mutex.release() - - mutex = threading.Lock() - mutex.acquire() - tid = thread.start_new_thread(f, (mutex,)) - # Wait for the thread to finish. - mutex.acquire() - self.assertIn(tid, threading._active) - self.assertIsInstance(threading._active[tid], threading._DummyThread) - del threading._active[tid] - - # PyThreadState_SetAsyncExc() is a CPython-only gimmick, not (currently) - # exposed at the Python level. This test relies on ctypes to get at it. - def test_PyThreadState_SetAsyncExc(self): - try: - import ctypes - except ImportError: - if verbose: - print "test_PyThreadState_SetAsyncExc can't import ctypes" - return # can't do anything - - set_async_exc = ctypes.pythonapi.PyThreadState_SetAsyncExc - - class AsyncExc(Exception): - pass - - exception = ctypes.py_object(AsyncExc) - - # First check it works when setting the exception from the same thread. - tid = thread.get_ident() - - try: - result = set_async_exc(ctypes.c_long(tid), exception) - # The exception is async, so we might have to keep the VM busy until - # it notices. - while True: - pass - except AsyncExc: - pass - else: - # This code is unreachable but it reflects the intent. If we wanted - # to be smarter the above loop wouldn't be infinite. - self.fail("AsyncExc not raised") - try: - self.assertEqual(result, 1) # one thread state modified - except UnboundLocalError: - # The exception was raised too quickly for us to get the result. - pass - - # `worker_started` is set by the thread when it's inside a try/except - # block waiting to catch the asynchronously set AsyncExc exception. - # `worker_saw_exception` is set by the thread upon catching that - # exception. - worker_started = threading.Event() - worker_saw_exception = threading.Event() - - class Worker(threading.Thread): - def run(self): - self.id = thread.get_ident() - self.finished = False - - try: - while True: - worker_started.set() - time.sleep(0.1) - except AsyncExc: - self.finished = True - worker_saw_exception.set() - - t = Worker() - t.daemon = True # so if this fails, we don't hang Python at shutdown - t.start() - if verbose: - print " started worker thread" - - # Try a thread id that doesn't make sense. - if verbose: - print " trying nonsensical thread id" - result = set_async_exc(ctypes.c_long(-1), exception) - self.assertEqual(result, 0) # no thread states modified - - # Now raise an exception in the worker thread. - if verbose: - print " waiting for worker thread to get started" - ret = worker_started.wait() - self.assertTrue(ret) - if verbose: - print " verifying worker hasn't exited" - self.assertTrue(not t.finished) - if verbose: - print " attempting to raise asynch exception in worker" - result = set_async_exc(ctypes.c_long(t.id), exception) - self.assertEqual(result, 1) # one thread state modified - if verbose: - print " waiting for worker to say it caught the exception" - worker_saw_exception.wait(timeout=10) - self.assertTrue(t.finished) - if verbose: - print " all OK -- joining worker" - if t.finished: - t.join() - # else the thread is still running, and we have no way to kill it - - def test_limbo_cleanup(self): - # Issue 7481: Failure to start thread should cleanup the limbo map. - def fail_new_thread(*args): - raise thread.error() - _start_new_thread = threading._start_new_thread - threading._start_new_thread = fail_new_thread - try: - t = threading.Thread(target=lambda: None) - self.assertRaises(thread.error, t.start) - self.assertFalse( - t in threading._limbo, - "Failed to cleanup _limbo map on failure of Thread.start().") - finally: - threading._start_new_thread = _start_new_thread - - def test_finalize_runnning_thread(self): - # Issue 1402: the PyGILState_Ensure / _Release functions may be called - # very late on python exit: on deallocation of a running thread for - # example. - try: - import ctypes - except ImportError: - if verbose: - print("test_finalize_with_runnning_thread can't import ctypes") - return # can't do anything - - rc = subprocess.call([sys.executable, "-c", """if 1: - import ctypes, sys, time, thread - - # This lock is used as a simple event variable. - ready = thread.allocate_lock() - ready.acquire() - - # Module globals are cleared before __del__ is run - # So we save the functions in class dict - class C: - ensure = ctypes.pythonapi.PyGILState_Ensure - release = ctypes.pythonapi.PyGILState_Release - def __del__(self): - state = self.ensure() - self.release(state) - - def waitingThread(): - x = C() - ready.release() - time.sleep(100) - - thread.start_new_thread(waitingThread, ()) - ready.acquire() # Be sure the other thread is waiting. - sys.exit(42) - """]) - self.assertEqual(rc, 42) - - def test_finalize_with_trace(self): - # Issue1733757 - # Avoid a deadlock when sys.settrace steps into threading._shutdown - p = subprocess.Popen([sys.executable, "-c", """if 1: - import sys, threading - - # A deadlock-killer, to prevent the - # testsuite to hang forever - def killer(): - import os, time - time.sleep(2) - print 'program blocked; aborting' - os._exit(2) - t = threading.Thread(target=killer) - t.daemon = True - t.start() - - # This is the trace function - def func(frame, event, arg): - threading.current_thread() - return func - - sys.settrace(func) - """], - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - self.addCleanup(p.stdout.close) - self.addCleanup(p.stderr.close) - stdout, stderr = p.communicate() - rc = p.returncode - self.assertFalse(rc == 2, "interpreted was blocked") - self.assertTrue(rc == 0, - "Unexpected error: " + repr(stderr)) - - def test_join_nondaemon_on_shutdown(self): - # Issue 1722344 - # Raising SystemExit skipped threading._shutdown - p = subprocess.Popen([sys.executable, "-c", """if 1: - import threading - from time import sleep - - def child(): - sleep(1) - # As a non-daemon thread we SHOULD wake up and nothing - # should be torn down yet - print "Woke up, sleep function is:", sleep - - threading.Thread(target=child).start() - raise SystemExit - """], - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - self.addCleanup(p.stdout.close) - self.addCleanup(p.stderr.close) - stdout, stderr = p.communicate() - self.assertEqual(stdout.strip(), - "Woke up, sleep function is: <built-in function sleep>") - stderr = re.sub(r"^\[\d+ refs\]", "", stderr, re.MULTILINE).strip() - self.assertEqual(stderr, "") - - def test_enumerate_after_join(self): - # Try hard to trigger #1703448: a thread is still returned in - # threading.enumerate() after it has been join()ed. - enum = threading.enumerate - old_interval = sys.getcheckinterval() - try: - for i in xrange(1, 100): - # Try a couple times at each thread-switching interval - # to get more interleavings. - sys.setcheckinterval(i // 5) - t = threading.Thread(target=lambda: None) - t.start() - t.join() - l = enum() - self.assertNotIn(t, l, - "#1703448 triggered after %d trials: %s" % (i, l)) - finally: - sys.setcheckinterval(old_interval) - - def test_no_refcycle_through_target(self): - class RunSelfFunction(object): - def __init__(self, should_raise): - # The links in this refcycle from Thread back to self - # should be cleaned up when the thread completes. - self.should_raise = should_raise - self.thread = threading.Thread(target=self._run, - args=(self,), - kwargs={'yet_another':self}) - self.thread.start() - - def _run(self, other_ref, yet_another): - if self.should_raise: - raise SystemExit - - cyclic_object = RunSelfFunction(should_raise=False) - weak_cyclic_object = weakref.ref(cyclic_object) - cyclic_object.thread.join() - del cyclic_object - self.assertEqual(None, weak_cyclic_object(), - msg=('%d references still around' % - sys.getrefcount(weak_cyclic_object()))) - - raising_cyclic_object = RunSelfFunction(should_raise=True) - weak_raising_cyclic_object = weakref.ref(raising_cyclic_object) - raising_cyclic_object.thread.join() - del raising_cyclic_object - self.assertEqual(None, weak_raising_cyclic_object(), - msg=('%d references still around' % - sys.getrefcount(weak_raising_cyclic_object()))) - - @unittest.skipUnless(hasattr(os, 'fork'), 'test needs fork()') - def test_dummy_thread_after_fork(self): - # Issue #14308: a dummy thread in the active list doesn't mess up - # the after-fork mechanism. - code = """if 1: - import thread, threading, os, time - - def background_thread(evt): - # Creates and registers the _DummyThread instance - threading.current_thread() - evt.set() - time.sleep(10) - - evt = threading.Event() - thread.start_new_thread(background_thread, (evt,)) - evt.wait() - assert threading.active_count() == 2, threading.active_count() - if os.fork() == 0: - assert threading.active_count() == 1, threading.active_count() - os._exit(0) - else: - os.wait() - """ - _, out, err = assert_python_ok("-c", code) - self.assertEqual(out, '') - self.assertEqual(err, '') - - -class ThreadJoinOnShutdown(BaseTestCase): - - # Between fork() and exec(), only async-safe functions are allowed (issues - # #12316 and #11870), and fork() from a worker thread is known to trigger - # problems with some operating systems (issue #3863): skip problematic tests - # on platforms known to behave badly. - platforms_to_skip = ('freebsd4', 'freebsd5', 'freebsd6', 'netbsd5', - 'os2emx') - - def _run_and_join(self, script): - script = """if 1: - import sys, os, time, threading - - # a thread, which waits for the main program to terminate - def joiningfunc(mainthread): - mainthread.join() - print 'end of thread' - \n""" + script - - p = subprocess.Popen([sys.executable, "-c", script], stdout=subprocess.PIPE) - rc = p.wait() - data = p.stdout.read().replace('\r', '') - p.stdout.close() - self.assertEqual(data, "end of main\nend of thread\n") - self.assertFalse(rc == 2, "interpreter was blocked") - self.assertTrue(rc == 0, "Unexpected error") - - def test_1_join_on_shutdown(self): - # The usual case: on exit, wait for a non-daemon thread - script = """if 1: - import os - t = threading.Thread(target=joiningfunc, - args=(threading.current_thread(),)) - t.start() - time.sleep(0.1) - print 'end of main' - """ - self._run_and_join(script) - - - @unittest.skipUnless(hasattr(os, 'fork'), "needs os.fork()") - @unittest.skipIf(sys.platform in platforms_to_skip, "due to known OS bug") - def test_2_join_in_forked_process(self): - # Like the test above, but from a forked interpreter - script = """if 1: - childpid = os.fork() - if childpid != 0: - os.waitpid(childpid, 0) - sys.exit(0) - - t = threading.Thread(target=joiningfunc, - args=(threading.current_thread(),)) - t.start() - print 'end of main' - """ - self._run_and_join(script) - - @unittest.skipUnless(hasattr(os, 'fork'), "needs os.fork()") - @unittest.skipIf(sys.platform in platforms_to_skip, "due to known OS bug") - def test_3_join_in_forked_from_thread(self): - # Like the test above, but fork() was called from a worker thread - # In the forked process, the main Thread object must be marked as stopped. - script = """if 1: - main_thread = threading.current_thread() - def worker(): - childpid = os.fork() - if childpid != 0: - os.waitpid(childpid, 0) - sys.exit(0) - - t = threading.Thread(target=joiningfunc, - args=(main_thread,)) - print 'end of main' - t.start() - t.join() # Should not block: main_thread is already stopped - - w = threading.Thread(target=worker) - w.start() - """ - self._run_and_join(script) - - def assertScriptHasOutput(self, script, expected_output): - p = subprocess.Popen([sys.executable, "-c", script], - stdout=subprocess.PIPE) - rc = p.wait() - data = p.stdout.read().decode().replace('\r', '') - self.assertEqual(rc, 0, "Unexpected error") - self.assertEqual(data, expected_output) - - @unittest.skipUnless(hasattr(os, 'fork'), "needs os.fork()") - @unittest.skipIf(sys.platform in platforms_to_skip, "due to known OS bug") - def test_4_joining_across_fork_in_worker_thread(self): - # There used to be a possible deadlock when forking from a child - # thread. See http://bugs.python.org/issue6643. - - # The script takes the following steps: - # - The main thread in the parent process starts a new thread and then - # tries to join it. - # - The join operation acquires the Lock inside the thread's _block - # Condition. (See threading.py:Thread.join().) - # - We stub out the acquire method on the condition to force it to wait - # until the child thread forks. (See LOCK ACQUIRED HERE) - # - The child thread forks. (See LOCK HELD and WORKER THREAD FORKS - # HERE) - # - The main thread of the parent process enters Condition.wait(), - # which releases the lock on the child thread. - # - The child process returns. Without the necessary fix, when the - # main thread of the child process (which used to be the child thread - # in the parent process) attempts to exit, it will try to acquire the - # lock in the Thread._block Condition object and hang, because the - # lock was held across the fork. - - script = """if 1: - import os, time, threading - - finish_join = False - start_fork = False - - def worker(): - # Wait until this thread's lock is acquired before forking to - # create the deadlock. - global finish_join - while not start_fork: - time.sleep(0.01) - # LOCK HELD: Main thread holds lock across this call. - childpid = os.fork() - finish_join = True - if childpid != 0: - # Parent process just waits for child. - os.waitpid(childpid, 0) - # Child process should just return. - - w = threading.Thread(target=worker) - - # Stub out the private condition variable's lock acquire method. - # This acquires the lock and then waits until the child has forked - # before returning, which will release the lock soon after. If - # someone else tries to fix this test case by acquiring this lock - # before forking instead of resetting it, the test case will - # deadlock when it shouldn't. - condition = w._block - orig_acquire = condition.acquire - call_count_lock = threading.Lock() - call_count = 0 - def my_acquire(): - global call_count - global start_fork - orig_acquire() # LOCK ACQUIRED HERE - start_fork = True - if call_count == 0: - while not finish_join: - time.sleep(0.01) # WORKER THREAD FORKS HERE - with call_count_lock: - call_count += 1 - condition.acquire = my_acquire - - w.start() - w.join() - print('end of main') - """ - self.assertScriptHasOutput(script, "end of main\n") - - @unittest.skipUnless(hasattr(os, 'fork'), "needs os.fork()") - @unittest.skipIf(sys.platform in platforms_to_skip, "due to known OS bug") - def test_5_clear_waiter_locks_to_avoid_crash(self): - # Check that a spawned thread that forks doesn't segfault on certain - # platforms, namely OS X. This used to happen if there was a waiter - # lock in the thread's condition variable's waiters list. Even though - # we know the lock will be held across the fork, it is not safe to - # release locks held across forks on all platforms, so releasing the - # waiter lock caused a segfault on OS X. Furthermore, since locks on - # OS X are (as of this writing) implemented with a mutex + condition - # variable instead of a semaphore, while we know that the Python-level - # lock will be acquired, we can't know if the internal mutex will be - # acquired at the time of the fork. - - script = """if True: - import os, time, threading - - start_fork = False - - def worker(): - # Wait until the main thread has attempted to join this thread - # before continuing. - while not start_fork: - time.sleep(0.01) - childpid = os.fork() - if childpid != 0: - # Parent process just waits for child. - (cpid, rc) = os.waitpid(childpid, 0) - assert cpid == childpid - assert rc == 0 - print('end of worker thread') - else: - # Child process should just return. - pass - - w = threading.Thread(target=worker) - - # Stub out the private condition variable's _release_save method. - # This releases the condition's lock and flips the global that - # causes the worker to fork. At this point, the problematic waiter - # lock has been acquired once by the waiter and has been put onto - # the waiters list. - condition = w._block - orig_release_save = condition._release_save - def my_release_save(): - global start_fork - orig_release_save() - # Waiter lock held here, condition lock released. - start_fork = True - condition._release_save = my_release_save - - w.start() - w.join() - print('end of main thread') - """ - output = "end of worker thread\nend of main thread\n" - self.assertScriptHasOutput(script, output) - - @unittest.skipUnless(hasattr(os, 'fork'), "needs os.fork()") - @unittest.skipIf(sys.platform in platforms_to_skip, "due to known OS bug") - def test_reinit_tls_after_fork(self): - # Issue #13817: fork() would deadlock in a multithreaded program with - # the ad-hoc TLS implementation. - - def do_fork_and_wait(): - # just fork a child process and wait it - pid = os.fork() - if pid > 0: - os.waitpid(pid, 0) - else: - os._exit(0) - - # start a bunch of threads that will fork() child processes - threads = [] - for i in range(16): - t = threading.Thread(target=do_fork_and_wait) - threads.append(t) - t.start() - - for t in threads: - t.join() - - -class ThreadingExceptionTests(BaseTestCase): - # A RuntimeError should be raised if Thread.start() is called - # multiple times. - def test_start_thread_again(self): - thread = threading.Thread() - thread.start() - self.assertRaises(RuntimeError, thread.start) - - def test_joining_current_thread(self): - current_thread = threading.current_thread() - self.assertRaises(RuntimeError, current_thread.join); - - def test_joining_inactive_thread(self): - thread = threading.Thread() - self.assertRaises(RuntimeError, thread.join) - - def test_daemonize_active_thread(self): - thread = threading.Thread() - thread.start() - self.assertRaises(RuntimeError, setattr, thread, "daemon", True) - - -class LockTests(lock_tests.LockTests): - locktype = staticmethod(threading.Lock) - -class RLockTests(lock_tests.RLockTests): - locktype = staticmethod(threading.RLock) - -class EventTests(lock_tests.EventTests): - eventtype = staticmethod(threading.Event) - -class ConditionAsRLockTests(lock_tests.RLockTests): - # An Condition uses an RLock by default and exports its API. - locktype = staticmethod(threading.Condition) - -class ConditionTests(lock_tests.ConditionTests): - condtype = staticmethod(threading.Condition) - -class SemaphoreTests(lock_tests.SemaphoreTests): - semtype = staticmethod(threading.Semaphore) - -class BoundedSemaphoreTests(lock_tests.BoundedSemaphoreTests): - semtype = staticmethod(threading.BoundedSemaphore) - - @unittest.skipUnless(sys.platform == 'darwin', 'test macosx problem') - def test_recursion_limit(self): - # Issue 9670 - # test that excessive recursion within a non-main thread causes - # an exception rather than crashing the interpreter on platforms - # like Mac OS X or FreeBSD which have small default stack sizes - # for threads - script = """if True: - import threading - - def recurse(): - return recurse() - - def outer(): - try: - recurse() - except RuntimeError: - pass - - w = threading.Thread(target=outer) - w.start() - w.join() - print('end of main thread') - """ - expected_output = "end of main thread\n" - p = subprocess.Popen([sys.executable, "-c", script], - stdout=subprocess.PIPE) - stdout, stderr = p.communicate() - data = stdout.decode().replace('\r', '') - self.assertEqual(p.returncode, 0, "Unexpected error") - self.assertEqual(data, expected_output) - -def test_main(): - test.test_support.run_unittest(LockTests, RLockTests, EventTests, - ConditionAsRLockTests, ConditionTests, - SemaphoreTests, BoundedSemaphoreTests, - ThreadTests, - ThreadJoinOnShutdown, - ThreadingExceptionTests, - ) - -if __name__ == "__main__": - test_main() |