""" =============================== NumPy memmap in joblib.Parallel =============================== This example illustrates some features enabled by using a memory map (:class:`numpy.memmap`) within :class:`joblib.Parallel`. First, we show that dumping a huge data array ahead of passing it to :class:`joblib.Parallel` speeds up computation. Then, we show the possibility to provide write access to original data. """ ############################################################################## # Speed up processing of a large data array ############################################################################## # # We create a large data array for which the average is computed for several # slices. import numpy as np data = np.random.random((int(1e7),)) window_size = int(5e5) slices = [slice(start, start + window_size) for start in range(0, data.size - window_size, int(1e5))] ############################################################################### # The ``slow_mean`` function introduces a :func:`time.sleep` call to simulate a # more expensive computation cost for which parallel computing is beneficial. # Parallel may not be beneficial for very fast operation, due to extra overhead # (workers creations, communication, etc.). import time def slow_mean(data, sl): """Simulate a time consuming processing.""" time.sleep(0.01) return data[sl].mean() ############################################################################### # First, we will evaluate the sequential computing on our problem. tic = time.time() results = [slow_mean(data, sl) for sl in slices] toc = time.time() print('\nElapsed time computing the average of couple of slices {:.2f} s' .format(toc - tic)) ############################################################################### # :class:`joblib.Parallel` is used to compute in parallel the average of all # slices using 2 workers. from joblib import Parallel, delayed tic = time.time() results = Parallel(n_jobs=2)(delayed(slow_mean)(data, sl) for sl in slices) toc = time.time() print('\nElapsed time computing the average of couple of slices {:.2f} s' .format(toc - tic)) ############################################################################### # Parallel processing is already faster than the sequential processing. It is # also possible to remove a bit of overhead by dumping the ``data`` array to a # memmap and pass the memmap to :class:`joblib.Parallel`. import os from joblib import dump, load folder = './joblib_memmap' try: os.mkdir(folder) except FileExistsError: pass data_filename_memmap = os.path.join(folder, 'data_memmap') dump(data, data_filename_memmap) data = load(data_filename_memmap, mmap_mode='r') tic = time.time() results = Parallel(n_jobs=2)(delayed(slow_mean)(data, sl) for sl in slices) toc = time.time() print('\nElapsed time computing the average of couple of slices {:.2f} s\n' .format(toc - tic)) ############################################################################### # Therefore, dumping large ``data`` array ahead of calling # :class:`joblib.Parallel` can speed up the processing by removing some # overhead. ############################################################################### # Writable memmap for shared memory :class:`joblib.Parallel` ############################################################################### # # ``slow_mean_write_output`` will compute the mean for some given slices as in # the previous example. However, the resulting mean will be directly written on # the output array. def slow_mean_write_output(data, sl, output, idx): """Simulate a time consuming processing.""" time.sleep(0.005) res_ = data[sl].mean() print("[Worker %d] Mean for slice %d is %f" % (os.getpid(), idx, res_)) output[idx] = res_ ############################################################################### # Prepare the folder where the memmap will be dumped. output_filename_memmap = os.path.join(folder, 'output_memmap') ############################################################################### # Pre-allocate a writable shared memory map as a container for the results of # the parallel computation. output = np.memmap(output_filename_memmap, dtype=data.dtype, shape=len(slices), mode='w+') ############################################################################### # ``data`` is replaced by its memory mapped version. Note that the buffer has # already been dumped in the previous section. data = load(data_filename_memmap, mmap_mode='r') ############################################################################### # Fork the worker processes to perform computation concurrently Parallel(n_jobs=2)(delayed(slow_mean_write_output)(data, sl, output, idx) for idx, sl in enumerate(slices)) ############################################################################### # Compare the results from the output buffer with the expected results print("\nExpected means computed in the parent process:\n {}" .format(np.array(results))) print("\nActual means computed by the worker processes:\n {}" .format(output)) ############################################################################### # Clean-up the memmap ############################################################################### # # Remove the different memmap that we created. It might fail in Windows due # to file permissions. import shutil try: shutil.rmtree(folder) except: # noqa print('Could not clean-up automatically.')