peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/joblib
/_parallel_backends.py
""" | |
Backends for embarrassingly parallel code. | |
""" | |
import gc | |
import os | |
import warnings | |
import threading | |
import contextlib | |
from abc import ABCMeta, abstractmethod | |
from ._utils import ( | |
_TracebackCapturingWrapper, | |
_retrieve_traceback_capturing_wrapped_call | |
) | |
from ._multiprocessing_helpers import mp | |
if mp is not None: | |
from .pool import MemmappingPool | |
from multiprocessing.pool import ThreadPool | |
from .executor import get_memmapping_executor | |
# Import loky only if multiprocessing is present | |
from .externals.loky import process_executor, cpu_count | |
from .externals.loky.process_executor import ShutdownExecutorError | |
class ParallelBackendBase(metaclass=ABCMeta): | |
"""Helper abc which defines all methods a ParallelBackend must implement""" | |
supports_inner_max_num_threads = False | |
supports_retrieve_callback = False | |
default_n_jobs = 1 | |
def supports_return_generator(self): | |
return self.supports_retrieve_callback | |
def supports_timeout(self): | |
return self.supports_retrieve_callback | |
nesting_level = None | |
def __init__(self, nesting_level=None, inner_max_num_threads=None, | |
**kwargs): | |
super().__init__(**kwargs) | |
self.nesting_level = nesting_level | |
self.inner_max_num_threads = inner_max_num_threads | |
MAX_NUM_THREADS_VARS = [ | |
'OMP_NUM_THREADS', 'OPENBLAS_NUM_THREADS', 'MKL_NUM_THREADS', | |
'BLIS_NUM_THREADS', 'VECLIB_MAXIMUM_THREADS', 'NUMBA_NUM_THREADS', | |
'NUMEXPR_NUM_THREADS', | |
] | |
TBB_ENABLE_IPC_VAR = "ENABLE_IPC" | |
def effective_n_jobs(self, n_jobs): | |
"""Determine the number of jobs that can actually run in parallel | |
n_jobs is the number of workers requested by the callers. Passing | |
n_jobs=-1 means requesting all available workers for instance matching | |
the number of CPU cores on the worker host(s). | |
This method should return a guesstimate of the number of workers that | |
can actually perform work concurrently. The primary use case is to make | |
it possible for the caller to know in how many chunks to slice the | |
work. | |
In general working on larger data chunks is more efficient (less | |
scheduling overhead and better use of CPU cache prefetching heuristics) | |
as long as all the workers have enough work to do. | |
""" | |
def apply_async(self, func, callback=None): | |
"""Schedule a func to be run""" | |
def retrieve_result_callback(self, out): | |
"""Called within the callback function passed in apply_async. | |
The argument of this function is the argument given to a callback in | |
the considered backend. It is supposed to return the outcome of a task | |
if it succeeded or raise the exception if it failed. | |
""" | |
def configure(self, n_jobs=1, parallel=None, prefer=None, require=None, | |
**backend_args): | |
"""Reconfigure the backend and return the number of workers. | |
This makes it possible to reuse an existing backend instance for | |
successive independent calls to Parallel with different parameters. | |
""" | |
self.parallel = parallel | |
return self.effective_n_jobs(n_jobs) | |
def start_call(self): | |
"""Call-back method called at the beginning of a Parallel call""" | |
def stop_call(self): | |
"""Call-back method called at the end of a Parallel call""" | |
def terminate(self): | |
"""Shutdown the workers and free the shared memory.""" | |
def compute_batch_size(self): | |
"""Determine the optimal batch size""" | |
return 1 | |
def batch_completed(self, batch_size, duration): | |
"""Callback indicate how long it took to run a batch""" | |
def get_exceptions(self): | |
"""List of exception types to be captured.""" | |
return [] | |
def abort_everything(self, ensure_ready=True): | |
"""Abort any running tasks | |
This is called when an exception has been raised when executing a task | |
and all the remaining tasks will be ignored and can therefore be | |
aborted to spare computation resources. | |
If ensure_ready is True, the backend should be left in an operating | |
state as future tasks might be re-submitted via that same backend | |
instance. | |
If ensure_ready is False, the implementer of this method can decide | |
to leave the backend in a closed / terminated state as no new task | |
are expected to be submitted to this backend. | |
Setting ensure_ready to False is an optimization that can be leveraged | |
when aborting tasks via killing processes from a local process pool | |
managed by the backend it-self: if we expect no new tasks, there is no | |
point in re-creating new workers. | |
""" | |
# Does nothing by default: to be overridden in subclasses when | |
# canceling tasks is possible. | |
pass | |
def get_nested_backend(self): | |
"""Backend instance to be used by nested Parallel calls. | |
By default a thread-based backend is used for the first level of | |
nesting. Beyond, switch to sequential backend to avoid spawning too | |
many threads on the host. | |
""" | |
nesting_level = getattr(self, 'nesting_level', 0) + 1 | |
if nesting_level > 1: | |
return SequentialBackend(nesting_level=nesting_level), None | |
else: | |
return ThreadingBackend(nesting_level=nesting_level), None | |
def retrieval_context(self): | |
"""Context manager to manage an execution context. | |
Calls to Parallel.retrieve will be made inside this context. | |
By default, this does nothing. It may be useful for subclasses to | |
handle nested parallelism. In particular, it may be required to avoid | |
deadlocks if a backend manages a fixed number of workers, when those | |
workers may be asked to do nested Parallel calls. Without | |
'retrieval_context' this could lead to deadlock, as all the workers | |
managed by the backend may be "busy" waiting for the nested parallel | |
calls to finish, but the backend has no free workers to execute those | |
tasks. | |
""" | |
yield | |
def _prepare_worker_env(self, n_jobs): | |
"""Return environment variables limiting threadpools in external libs. | |
This function return a dict containing environment variables to pass | |
when creating a pool of process. These environment variables limit the | |
number of threads to `n_threads` for OpenMP, MKL, Accelerated and | |
OpenBLAS libraries in the child processes. | |
""" | |
explicit_n_threads = self.inner_max_num_threads | |
default_n_threads = max(cpu_count() // n_jobs, 1) | |
# Set the inner environment variables to self.inner_max_num_threads if | |
# it is given. Else, default to cpu_count // n_jobs unless the variable | |
# is already present in the parent process environment. | |
env = {} | |
for var in self.MAX_NUM_THREADS_VARS: | |
if explicit_n_threads is None: | |
var_value = os.environ.get(var, default_n_threads) | |
else: | |
var_value = explicit_n_threads | |
env[var] = str(var_value) | |
if self.TBB_ENABLE_IPC_VAR not in os.environ: | |
# To avoid over-subscription when using TBB, let the TBB schedulers | |
# use Inter Process Communication to coordinate: | |
env[self.TBB_ENABLE_IPC_VAR] = "1" | |
return env | |
def in_main_thread(): | |
return isinstance(threading.current_thread(), threading._MainThread) | |
class SequentialBackend(ParallelBackendBase): | |
"""A ParallelBackend which will execute all batches sequentially. | |
Does not use/create any threading objects, and hence has minimal | |
overhead. Used when n_jobs == 1. | |
""" | |
uses_threads = True | |
supports_timeout = False | |
supports_retrieve_callback = False | |
supports_sharedmem = True | |
def effective_n_jobs(self, n_jobs): | |
"""Determine the number of jobs which are going to run in parallel""" | |
if n_jobs == 0: | |
raise ValueError('n_jobs == 0 in Parallel has no meaning') | |
return 1 | |
def apply_async(self, func, callback=None): | |
"""Schedule a func to be run""" | |
raise RuntimeError("Should never be called for SequentialBackend.") | |
def retrieve_result_callback(self, out): | |
raise RuntimeError("Should never be called for SequentialBackend.") | |
def get_nested_backend(self): | |
# import is not top level to avoid cyclic import errors. | |
from .parallel import get_active_backend | |
# SequentialBackend should neither change the nesting level, the | |
# default backend or the number of jobs. Just return the current one. | |
return get_active_backend() | |
class PoolManagerMixin(object): | |
"""A helper class for managing pool of workers.""" | |
_pool = None | |
def effective_n_jobs(self, n_jobs): | |
"""Determine the number of jobs which are going to run in parallel""" | |
if n_jobs == 0: | |
raise ValueError('n_jobs == 0 in Parallel has no meaning') | |
elif mp is None or n_jobs is None: | |
# multiprocessing is not available or disabled, fallback | |
# to sequential mode | |
return 1 | |
elif n_jobs < 0: | |
n_jobs = max(cpu_count() + 1 + n_jobs, 1) | |
return n_jobs | |
def terminate(self): | |
"""Shutdown the process or thread pool""" | |
if self._pool is not None: | |
self._pool.close() | |
self._pool.terminate() # terminate does a join() | |
self._pool = None | |
def _get_pool(self): | |
"""Used by apply_async to make it possible to implement lazy init""" | |
return self._pool | |
def apply_async(self, func, callback=None): | |
"""Schedule a func to be run""" | |
# Here, we need a wrapper to avoid crashes on KeyboardInterruptErrors. | |
# We also call the callback on error, to make sure the pool does not | |
# wait on crashed jobs. | |
return self._get_pool().apply_async( | |
_TracebackCapturingWrapper(func), (), | |
callback=callback, error_callback=callback | |
) | |
def retrieve_result_callback(self, out): | |
"""Mimic concurrent.futures results, raising an error if needed.""" | |
return _retrieve_traceback_capturing_wrapped_call(out) | |
def abort_everything(self, ensure_ready=True): | |
"""Shutdown the pool and restart a new one with the same parameters""" | |
self.terminate() | |
if ensure_ready: | |
self.configure(n_jobs=self.parallel.n_jobs, parallel=self.parallel, | |
**self.parallel._backend_args) | |
class AutoBatchingMixin(object): | |
"""A helper class for automagically batching jobs.""" | |
# In seconds, should be big enough to hide multiprocessing dispatching | |
# overhead. | |
# This settings was found by running benchmarks/bench_auto_batching.py | |
# with various parameters on various platforms. | |
MIN_IDEAL_BATCH_DURATION = .2 | |
# Should not be too high to avoid stragglers: long jobs running alone | |
# on a single worker while other workers have no work to process any more. | |
MAX_IDEAL_BATCH_DURATION = 2 | |
# Batching counters default values | |
_DEFAULT_EFFECTIVE_BATCH_SIZE = 1 | |
_DEFAULT_SMOOTHED_BATCH_DURATION = 0.0 | |
def __init__(self, **kwargs): | |
super().__init__(**kwargs) | |
self._effective_batch_size = self._DEFAULT_EFFECTIVE_BATCH_SIZE | |
self._smoothed_batch_duration = self._DEFAULT_SMOOTHED_BATCH_DURATION | |
def compute_batch_size(self): | |
"""Determine the optimal batch size""" | |
old_batch_size = self._effective_batch_size | |
batch_duration = self._smoothed_batch_duration | |
if (batch_duration > 0 and | |
batch_duration < self.MIN_IDEAL_BATCH_DURATION): | |
# The current batch size is too small: the duration of the | |
# processing of a batch of task is not large enough to hide | |
# the scheduling overhead. | |
ideal_batch_size = int(old_batch_size * | |
self.MIN_IDEAL_BATCH_DURATION / | |
batch_duration) | |
# Multiply by two to limit oscilations between min and max. | |
ideal_batch_size *= 2 | |
# dont increase the batch size too fast to limit huge batch sizes | |
# potentially leading to starving worker | |
batch_size = min(2 * old_batch_size, ideal_batch_size) | |
batch_size = max(batch_size, 1) | |
self._effective_batch_size = batch_size | |
if self.parallel.verbose >= 10: | |
self.parallel._print( | |
f"Batch computation too fast ({batch_duration}s.) " | |
f"Setting batch_size={batch_size}." | |
) | |
elif (batch_duration > self.MAX_IDEAL_BATCH_DURATION and | |
old_batch_size >= 2): | |
# The current batch size is too big. If we schedule overly long | |
# running batches some CPUs might wait with nothing left to do | |
# while a couple of CPUs a left processing a few long running | |
# batches. Better reduce the batch size a bit to limit the | |
# likelihood of scheduling such stragglers. | |
# decrease the batch size quickly to limit potential starving | |
ideal_batch_size = int( | |
old_batch_size * self.MIN_IDEAL_BATCH_DURATION / batch_duration | |
) | |
# Multiply by two to limit oscilations between min and max. | |
batch_size = max(2 * ideal_batch_size, 1) | |
self._effective_batch_size = batch_size | |
if self.parallel.verbose >= 10: | |
self.parallel._print( | |
f"Batch computation too slow ({batch_duration}s.) " | |
f"Setting batch_size={batch_size}." | |
) | |
else: | |
# No batch size adjustment | |
batch_size = old_batch_size | |
if batch_size != old_batch_size: | |
# Reset estimation of the smoothed mean batch duration: this | |
# estimate is updated in the multiprocessing apply_async | |
# CallBack as long as the batch_size is constant. Therefore | |
# we need to reset the estimate whenever we re-tune the batch | |
# size. | |
self._smoothed_batch_duration = \ | |
self._DEFAULT_SMOOTHED_BATCH_DURATION | |
return batch_size | |
def batch_completed(self, batch_size, duration): | |
"""Callback indicate how long it took to run a batch""" | |
if batch_size == self._effective_batch_size: | |
# Update the smoothed streaming estimate of the duration of a batch | |
# from dispatch to completion | |
old_duration = self._smoothed_batch_duration | |
if old_duration == self._DEFAULT_SMOOTHED_BATCH_DURATION: | |
# First record of duration for this batch size after the last | |
# reset. | |
new_duration = duration | |
else: | |
# Update the exponentially weighted average of the duration of | |
# batch for the current effective size. | |
new_duration = 0.8 * old_duration + 0.2 * duration | |
self._smoothed_batch_duration = new_duration | |
def reset_batch_stats(self): | |
"""Reset batch statistics to default values. | |
This avoids interferences with future jobs. | |
""" | |
self._effective_batch_size = self._DEFAULT_EFFECTIVE_BATCH_SIZE | |
self._smoothed_batch_duration = self._DEFAULT_SMOOTHED_BATCH_DURATION | |
class ThreadingBackend(PoolManagerMixin, ParallelBackendBase): | |
"""A ParallelBackend which will use a thread pool to execute batches in. | |
This is a low-overhead backend but it suffers from the Python Global | |
Interpreter Lock if the called function relies a lot on Python objects. | |
Mostly useful when the execution bottleneck is a compiled extension that | |
explicitly releases the GIL (for instance a Cython loop wrapped in a "with | |
nogil" block or an expensive call to a library such as NumPy). | |
The actual thread pool is lazily initialized: the actual thread pool | |
construction is delayed to the first call to apply_async. | |
ThreadingBackend is used as the default backend for nested calls. | |
""" | |
supports_retrieve_callback = True | |
uses_threads = True | |
supports_sharedmem = True | |
def configure(self, n_jobs=1, parallel=None, **backend_args): | |
"""Build a process or thread pool and return the number of workers""" | |
n_jobs = self.effective_n_jobs(n_jobs) | |
if n_jobs == 1: | |
# Avoid unnecessary overhead and use sequential backend instead. | |
raise FallbackToBackend( | |
SequentialBackend(nesting_level=self.nesting_level)) | |
self.parallel = parallel | |
self._n_jobs = n_jobs | |
return n_jobs | |
def _get_pool(self): | |
"""Lazily initialize the thread pool | |
The actual pool of worker threads is only initialized at the first | |
call to apply_async. | |
""" | |
if self._pool is None: | |
self._pool = ThreadPool(self._n_jobs) | |
return self._pool | |
class MultiprocessingBackend(PoolManagerMixin, AutoBatchingMixin, | |
ParallelBackendBase): | |
"""A ParallelBackend which will use a multiprocessing.Pool. | |
Will introduce some communication and memory overhead when exchanging | |
input and output data with the with the worker Python processes. | |
However, does not suffer from the Python Global Interpreter Lock. | |
""" | |
supports_retrieve_callback = True | |
supports_return_generator = False | |
def effective_n_jobs(self, n_jobs): | |
"""Determine the number of jobs which are going to run in parallel. | |
This also checks if we are attempting to create a nested parallel | |
loop. | |
""" | |
if mp is None: | |
return 1 | |
if mp.current_process().daemon: | |
# Daemonic processes cannot have children | |
if n_jobs != 1: | |
if inside_dask_worker(): | |
msg = ( | |
"Inside a Dask worker with daemon=True, " | |
"setting n_jobs=1.\nPossible work-arounds:\n" | |
"- dask.config.set(" | |
"{'distributed.worker.daemon': False})" | |
"- set the environment variable " | |
"DASK_DISTRIBUTED__WORKER__DAEMON=False\n" | |
"before creating your Dask cluster." | |
) | |
else: | |
msg = ( | |
'Multiprocessing-backed parallel loops ' | |
'cannot be nested, setting n_jobs=1' | |
) | |
warnings.warn(msg, stacklevel=3) | |
return 1 | |
if process_executor._CURRENT_DEPTH > 0: | |
# Mixing loky and multiprocessing in nested loop is not supported | |
if n_jobs != 1: | |
warnings.warn( | |
'Multiprocessing-backed parallel loops cannot be nested,' | |
' below loky, setting n_jobs=1', | |
stacklevel=3) | |
return 1 | |
elif not (self.in_main_thread() or self.nesting_level == 0): | |
# Prevent posix fork inside in non-main posix threads | |
if n_jobs != 1: | |
warnings.warn( | |
'Multiprocessing-backed parallel loops cannot be nested' | |
' below threads, setting n_jobs=1', | |
stacklevel=3) | |
return 1 | |
return super(MultiprocessingBackend, self).effective_n_jobs(n_jobs) | |
def configure(self, n_jobs=1, parallel=None, prefer=None, require=None, | |
**memmappingpool_args): | |
"""Build a process or thread pool and return the number of workers""" | |
n_jobs = self.effective_n_jobs(n_jobs) | |
if n_jobs == 1: | |
raise FallbackToBackend( | |
SequentialBackend(nesting_level=self.nesting_level)) | |
# Make sure to free as much memory as possible before forking | |
gc.collect() | |
self._pool = MemmappingPool(n_jobs, **memmappingpool_args) | |
self.parallel = parallel | |
return n_jobs | |
def terminate(self): | |
"""Shutdown the process or thread pool""" | |
super(MultiprocessingBackend, self).terminate() | |
self.reset_batch_stats() | |
class LokyBackend(AutoBatchingMixin, ParallelBackendBase): | |
"""Managing pool of workers with loky instead of multiprocessing.""" | |
supports_retrieve_callback = True | |
supports_inner_max_num_threads = True | |
def configure(self, n_jobs=1, parallel=None, prefer=None, require=None, | |
idle_worker_timeout=300, **memmappingexecutor_args): | |
"""Build a process executor and return the number of workers""" | |
n_jobs = self.effective_n_jobs(n_jobs) | |
if n_jobs == 1: | |
raise FallbackToBackend( | |
SequentialBackend(nesting_level=self.nesting_level)) | |
self._workers = get_memmapping_executor( | |
n_jobs, timeout=idle_worker_timeout, | |
env=self._prepare_worker_env(n_jobs=n_jobs), | |
context_id=parallel._id, **memmappingexecutor_args) | |
self.parallel = parallel | |
return n_jobs | |
def effective_n_jobs(self, n_jobs): | |
"""Determine the number of jobs which are going to run in parallel""" | |
if n_jobs == 0: | |
raise ValueError('n_jobs == 0 in Parallel has no meaning') | |
elif mp is None or n_jobs is None: | |
# multiprocessing is not available or disabled, fallback | |
# to sequential mode | |
return 1 | |
elif mp.current_process().daemon: | |
# Daemonic processes cannot have children | |
if n_jobs != 1: | |
if inside_dask_worker(): | |
msg = ( | |
"Inside a Dask worker with daemon=True, " | |
"setting n_jobs=1.\nPossible work-arounds:\n" | |
"- dask.config.set(" | |
"{'distributed.worker.daemon': False})\n" | |
"- set the environment variable " | |
"DASK_DISTRIBUTED__WORKER__DAEMON=False\n" | |
"before creating your Dask cluster." | |
) | |
else: | |
msg = ( | |
'Loky-backed parallel loops cannot be called in a' | |
' multiprocessing, setting n_jobs=1' | |
) | |
warnings.warn(msg, stacklevel=3) | |
return 1 | |
elif not (self.in_main_thread() or self.nesting_level == 0): | |
# Prevent posix fork inside in non-main posix threads | |
if n_jobs != 1: | |
warnings.warn( | |
'Loky-backed parallel loops cannot be nested below ' | |
'threads, setting n_jobs=1', | |
stacklevel=3) | |
return 1 | |
elif n_jobs < 0: | |
n_jobs = max(cpu_count() + 1 + n_jobs, 1) | |
return n_jobs | |
def apply_async(self, func, callback=None): | |
"""Schedule a func to be run""" | |
future = self._workers.submit(func) | |
if callback is not None: | |
future.add_done_callback(callback) | |
return future | |
def retrieve_result_callback(self, out): | |
try: | |
return out.result() | |
except ShutdownExecutorError: | |
raise RuntimeError( | |
"The executor underlying Parallel has been shutdown. " | |
"This is likely due to the garbage collection of a previous " | |
"generator from a call to Parallel with return_as='generator'." | |
" Make sure the generator is not garbage collected when " | |
"submitting a new job or that it is first properly exhausted." | |
) | |
def terminate(self): | |
if self._workers is not None: | |
# Don't terminate the workers as we want to reuse them in later | |
# calls, but cleanup the temporary resources that the Parallel call | |
# created. This 'hack' requires a private, low-level operation. | |
self._workers._temp_folder_manager._clean_temporary_resources( | |
context_id=self.parallel._id, force=False | |
) | |
self._workers = None | |
self.reset_batch_stats() | |
def abort_everything(self, ensure_ready=True): | |
"""Shutdown the workers and restart a new one with the same parameters | |
""" | |
self._workers.terminate(kill_workers=True) | |
self._workers = None | |
if ensure_ready: | |
self.configure(n_jobs=self.parallel.n_jobs, parallel=self.parallel) | |
class FallbackToBackend(Exception): | |
"""Raised when configuration should fallback to another backend""" | |
def __init__(self, backend): | |
self.backend = backend | |
def inside_dask_worker(): | |
"""Check whether the current function is executed inside a Dask worker. | |
""" | |
# This function can not be in joblib._dask because there would be a | |
# circular import: | |
# _dask imports _parallel_backend that imports _dask ... | |
try: | |
from distributed import get_worker | |
except ImportError: | |
return False | |
try: | |
get_worker() | |
return True | |
except ValueError: | |
return False | |