mirror of
https://github.com/python/cpython.git
synced 2024-11-25 09:39:56 +01:00
bf2e7e55d7
In the case of multiprocessing.synchronize() being missing, the test_concurrent_futures test suite now skips only the tests that require multiprocessing.synchronize(). Validate that multiprocessing.synchronize exists as part of _check_system_limits(), allowing ProcessPoolExecutor to raise NotImplementedError during __init__, rather than crashing with ImportError during __init__ when creating a lock imported from multiprocessing.synchronize. Use _check_system_limits() to disable tests of ProcessPoolExecutor on systems without multiprocessing.synchronize. Running the test suite without multiprocessing.synchronize reveals that Lib/compileall.py crashes when it uses a ProcessPoolExecutor. Therefore, change Lib/compileall.py to call _check_system_limits() before creating the ProcessPoolExecutor. Note that both Lib/compileall.py and Lib/test/test_compileall.py were attempting to sanity-check ProcessPoolExecutor by expecting ImportError. In multiprocessing.resource_tracker, sem_unlink() is also absent on platforms where POSIX semaphores aren't available. Avoid using sem_unlink() if it, too, does not exist. Co-authored-by: Pablo Galindo <Pablogsal@gmail.com>
240 lines
8.8 KiB
Python
240 lines
8.8 KiB
Python
###############################################################################
|
|
# Server process to keep track of unlinked resources (like shared memory
|
|
# segments, semaphores etc.) and clean them.
|
|
#
|
|
# On Unix we run a server process which keeps track of unlinked
|
|
# resources. The server ignores SIGINT and SIGTERM and reads from a
|
|
# pipe. Every other process of the program has a copy of the writable
|
|
# end of the pipe, so we get EOF when all other processes have exited.
|
|
# Then the server process unlinks any remaining resource names.
|
|
#
|
|
# This is important because there may be system limits for such resources: for
|
|
# instance, the system only supports a limited number of named semaphores, and
|
|
# shared-memory segments live in the RAM. If a python process leaks such a
|
|
# resource, this resource will not be removed till the next reboot. Without
|
|
# this resource tracker process, "killall python" would probably leave unlinked
|
|
# resources.
|
|
|
|
import os
|
|
import signal
|
|
import sys
|
|
import threading
|
|
import warnings
|
|
|
|
from . import spawn
|
|
from . import util
|
|
|
|
__all__ = ['ensure_running', 'register', 'unregister']
|
|
|
|
_HAVE_SIGMASK = hasattr(signal, 'pthread_sigmask')
|
|
_IGNORED_SIGNALS = (signal.SIGINT, signal.SIGTERM)
|
|
|
|
_CLEANUP_FUNCS = {
|
|
'noop': lambda: None,
|
|
}
|
|
|
|
if os.name == 'posix':
|
|
import _multiprocessing
|
|
import _posixshmem
|
|
|
|
# Use sem_unlink() to clean up named semaphores.
|
|
#
|
|
# sem_unlink() may be missing if the Python build process detected the
|
|
# absence of POSIX named semaphores. In that case, no named semaphores were
|
|
# ever opened, so no cleanup would be necessary.
|
|
if hasattr(_multiprocessing, 'sem_unlink'):
|
|
_CLEANUP_FUNCS.update({
|
|
'semaphore': _multiprocessing.sem_unlink,
|
|
})
|
|
_CLEANUP_FUNCS.update({
|
|
'shared_memory': _posixshmem.shm_unlink,
|
|
})
|
|
|
|
|
|
class ResourceTracker(object):
|
|
|
|
def __init__(self):
|
|
self._lock = threading.Lock()
|
|
self._fd = None
|
|
self._pid = None
|
|
|
|
def _stop(self):
|
|
with self._lock:
|
|
if self._fd is None:
|
|
# not running
|
|
return
|
|
|
|
# closing the "alive" file descriptor stops main()
|
|
os.close(self._fd)
|
|
self._fd = None
|
|
|
|
os.waitpid(self._pid, 0)
|
|
self._pid = None
|
|
|
|
def getfd(self):
|
|
self.ensure_running()
|
|
return self._fd
|
|
|
|
def ensure_running(self):
|
|
'''Make sure that resource tracker process is running.
|
|
|
|
This can be run from any process. Usually a child process will use
|
|
the resource created by its parent.'''
|
|
with self._lock:
|
|
if self._fd is not None:
|
|
# resource tracker was launched before, is it still running?
|
|
if self._check_alive():
|
|
# => still alive
|
|
return
|
|
# => dead, launch it again
|
|
os.close(self._fd)
|
|
|
|
# Clean-up to avoid dangling processes.
|
|
try:
|
|
# _pid can be None if this process is a child from another
|
|
# python process, which has started the resource_tracker.
|
|
if self._pid is not None:
|
|
os.waitpid(self._pid, 0)
|
|
except ChildProcessError:
|
|
# The resource_tracker has already been terminated.
|
|
pass
|
|
self._fd = None
|
|
self._pid = None
|
|
|
|
warnings.warn('resource_tracker: process died unexpectedly, '
|
|
'relaunching. Some resources might leak.')
|
|
|
|
fds_to_pass = []
|
|
try:
|
|
fds_to_pass.append(sys.stderr.fileno())
|
|
except Exception:
|
|
pass
|
|
cmd = 'from multiprocessing.resource_tracker import main;main(%d)'
|
|
r, w = os.pipe()
|
|
try:
|
|
fds_to_pass.append(r)
|
|
# process will out live us, so no need to wait on pid
|
|
exe = spawn.get_executable()
|
|
args = [exe] + util._args_from_interpreter_flags()
|
|
args += ['-c', cmd % r]
|
|
# bpo-33613: Register a signal mask that will block the signals.
|
|
# This signal mask will be inherited by the child that is going
|
|
# to be spawned and will protect the child from a race condition
|
|
# that can make the child die before it registers signal handlers
|
|
# for SIGINT and SIGTERM. The mask is unregistered after spawning
|
|
# the child.
|
|
try:
|
|
if _HAVE_SIGMASK:
|
|
signal.pthread_sigmask(signal.SIG_BLOCK, _IGNORED_SIGNALS)
|
|
pid = util.spawnv_passfds(exe, args, fds_to_pass)
|
|
finally:
|
|
if _HAVE_SIGMASK:
|
|
signal.pthread_sigmask(signal.SIG_UNBLOCK, _IGNORED_SIGNALS)
|
|
except:
|
|
os.close(w)
|
|
raise
|
|
else:
|
|
self._fd = w
|
|
self._pid = pid
|
|
finally:
|
|
os.close(r)
|
|
|
|
def _check_alive(self):
|
|
'''Check that the pipe has not been closed by sending a probe.'''
|
|
try:
|
|
# We cannot use send here as it calls ensure_running, creating
|
|
# a cycle.
|
|
os.write(self._fd, b'PROBE:0:noop\n')
|
|
except OSError:
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
def register(self, name, rtype):
|
|
'''Register name of resource with resource tracker.'''
|
|
self._send('REGISTER', name, rtype)
|
|
|
|
def unregister(self, name, rtype):
|
|
'''Unregister name of resource with resource tracker.'''
|
|
self._send('UNREGISTER', name, rtype)
|
|
|
|
def _send(self, cmd, name, rtype):
|
|
self.ensure_running()
|
|
msg = '{0}:{1}:{2}\n'.format(cmd, name, rtype).encode('ascii')
|
|
if len(name) > 512:
|
|
# posix guarantees that writes to a pipe of less than PIPE_BUF
|
|
# bytes are atomic, and that PIPE_BUF >= 512
|
|
raise ValueError('name too long')
|
|
nbytes = os.write(self._fd, msg)
|
|
assert nbytes == len(msg), "nbytes {0:n} but len(msg) {1:n}".format(
|
|
nbytes, len(msg))
|
|
|
|
|
|
_resource_tracker = ResourceTracker()
|
|
ensure_running = _resource_tracker.ensure_running
|
|
register = _resource_tracker.register
|
|
unregister = _resource_tracker.unregister
|
|
getfd = _resource_tracker.getfd
|
|
|
|
def main(fd):
|
|
'''Run resource tracker.'''
|
|
# protect the process from ^C and "killall python" etc
|
|
signal.signal(signal.SIGINT, signal.SIG_IGN)
|
|
signal.signal(signal.SIGTERM, signal.SIG_IGN)
|
|
if _HAVE_SIGMASK:
|
|
signal.pthread_sigmask(signal.SIG_UNBLOCK, _IGNORED_SIGNALS)
|
|
|
|
for f in (sys.stdin, sys.stdout):
|
|
try:
|
|
f.close()
|
|
except Exception:
|
|
pass
|
|
|
|
cache = {rtype: set() for rtype in _CLEANUP_FUNCS.keys()}
|
|
try:
|
|
# keep track of registered/unregistered resources
|
|
with open(fd, 'rb') as f:
|
|
for line in f:
|
|
try:
|
|
cmd, name, rtype = line.strip().decode('ascii').split(':')
|
|
cleanup_func = _CLEANUP_FUNCS.get(rtype, None)
|
|
if cleanup_func is None:
|
|
raise ValueError(
|
|
f'Cannot register {name} for automatic cleanup: '
|
|
f'unknown resource type {rtype}')
|
|
|
|
if cmd == 'REGISTER':
|
|
cache[rtype].add(name)
|
|
elif cmd == 'UNREGISTER':
|
|
cache[rtype].remove(name)
|
|
elif cmd == 'PROBE':
|
|
pass
|
|
else:
|
|
raise RuntimeError('unrecognized command %r' % cmd)
|
|
except Exception:
|
|
try:
|
|
sys.excepthook(*sys.exc_info())
|
|
except:
|
|
pass
|
|
finally:
|
|
# all processes have terminated; cleanup any remaining resources
|
|
for rtype, rtype_cache in cache.items():
|
|
if rtype_cache:
|
|
try:
|
|
warnings.warn('resource_tracker: There appear to be %d '
|
|
'leaked %s objects to clean up at shutdown' %
|
|
(len(rtype_cache), rtype))
|
|
except Exception:
|
|
pass
|
|
for name in rtype_cache:
|
|
# For some reason the process which created and registered this
|
|
# resource has failed to unregister it. Presumably it has
|
|
# died. We therefore unlink it.
|
|
try:
|
|
try:
|
|
_CLEANUP_FUNCS[rtype](name)
|
|
except Exception as e:
|
|
warnings.warn('resource_tracker: %r: %s' % (name, e))
|
|
finally:
|
|
pass
|