peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/joblib
/executor.py
"""Utility function to construct a loky.ReusableExecutor with custom pickler. | |
This module provides efficient ways of working with data stored in | |
shared memory with numpy.memmap arrays without inducing any memory | |
copy between the parent and child processes. | |
""" | |
# Author: Thomas Moreau <[email protected]> | |
# Copyright: 2017, Thomas Moreau | |
# License: BSD 3 clause | |
from ._memmapping_reducer import get_memmapping_reducers | |
from ._memmapping_reducer import TemporaryResourcesManager | |
from .externals.loky.reusable_executor import _ReusablePoolExecutor | |
_executor_args = None | |
def get_memmapping_executor(n_jobs, **kwargs): | |
return MemmappingExecutor.get_memmapping_executor(n_jobs, **kwargs) | |
class MemmappingExecutor(_ReusablePoolExecutor): | |
def get_memmapping_executor(cls, n_jobs, timeout=300, initializer=None, | |
initargs=(), env=None, temp_folder=None, | |
context_id=None, **backend_args): | |
"""Factory for ReusableExecutor with automatic memmapping for large | |
numpy arrays. | |
""" | |
global _executor_args | |
# Check if we can reuse the executor here instead of deferring the test | |
# to loky as the reducers are objects that changes at each call. | |
executor_args = backend_args.copy() | |
executor_args.update(env if env else {}) | |
executor_args.update(dict( | |
timeout=timeout, initializer=initializer, initargs=initargs)) | |
reuse = _executor_args is None or _executor_args == executor_args | |
_executor_args = executor_args | |
manager = TemporaryResourcesManager(temp_folder) | |
# reducers access the temporary folder in which to store temporary | |
# pickles through a call to manager.resolve_temp_folder_name. resolving | |
# the folder name dynamically is useful to use different folders across | |
# calls of a same reusable executor | |
job_reducers, result_reducers = get_memmapping_reducers( | |
unlink_on_gc_collect=True, | |
temp_folder_resolver=manager.resolve_temp_folder_name, | |
**backend_args) | |
_executor, executor_is_reused = super().get_reusable_executor( | |
n_jobs, job_reducers=job_reducers, result_reducers=result_reducers, | |
reuse=reuse, timeout=timeout, initializer=initializer, | |
initargs=initargs, env=env | |
) | |
if not executor_is_reused: | |
# Only set a _temp_folder_manager for new executors. Reused | |
# executors already have a _temporary_folder_manager that must not | |
# be re-assigned like that because it is referenced in various | |
# places in the reducing machinery of the executor. | |
_executor._temp_folder_manager = manager | |
if context_id is not None: | |
# Only register the specified context once we know which manager | |
# the current executor is using, in order to not register an atexit | |
# finalizer twice for the same folder. | |
_executor._temp_folder_manager.register_new_context(context_id) | |
return _executor | |
def terminate(self, kill_workers=False): | |
self.shutdown(kill_workers=kill_workers) | |
# When workers are killed in a brutal manner, they cannot execute the | |
# finalizer of their shared memmaps. The refcount of those memmaps may | |
# be off by an unknown number, so instead of decref'ing them, we force | |
# delete the whole temporary folder, and unregister them. There is no | |
# risk of PermissionError at folder deletion because at this | |
# point, all child processes are dead, so all references to temporary | |
# memmaps are closed. Otherwise, just try to delete as much as possible | |
# with allow_non_empty=True but if we can't, it will be clean up later | |
# on by the resource_tracker. | |
with self._submit_resize_lock: | |
self._temp_folder_manager._clean_temporary_resources( | |
force=kill_workers, allow_non_empty=True | |
) | |
def _temp_folder(self): | |
# Legacy property in tests. could be removed if we refactored the | |
# memmapping tests. SHOULD ONLY BE USED IN TESTS! | |
# We cache this property because it is called late in the tests - at | |
# this point, all context have been unregistered, and | |
# resolve_temp_folder_name raises an error. | |
if getattr(self, '_cached_temp_folder', None) is not None: | |
return self._cached_temp_folder | |
else: | |
self._cached_temp_folder = self._temp_folder_manager.resolve_temp_folder_name() # noqa | |
return self._cached_temp_folder | |
class _TestingMemmappingExecutor(MemmappingExecutor): | |
"""Wrapper around ReusableExecutor to ease memmapping testing with Pool | |
and Executor. This is only for testing purposes. | |
""" | |
def apply_async(self, func, args): | |
"""Schedule a func to be run""" | |
future = self.submit(func, *args) | |
future.get = future.result | |
return future | |
def map(self, f, *args): | |
return list(super().map(f, *args)) | |