""" ================================================== Checkpoint using joblib.Memory and joblib.Parallel ================================================== This example illustrates how to cache intermediate computing results using :class:`joblib.Memory` within :class:`joblib.Parallel`. """ ############################################################################### # Embed caching within parallel processing ############################################################################### # # It is possible to cache a computationally expensive function executed during # a parallel process. ``costly_compute`` emulates such time consuming function. import time def costly_compute(data, column): """Emulate a costly function by sleeping and returning a column.""" time.sleep(2) return data[column] def data_processing_mean(data, column): """Compute the mean of a column.""" return costly_compute(data, column).mean() ############################################################################### # Create some data. The random seed is fixed to generate deterministic data # across Python session. Note that this is not necessary for this specific # example since the memory cache is cleared at the end of the session. import numpy as np rng = np.random.RandomState(42) data = rng.randn(int(1e4), 4) ############################################################################### # It is first possible to make the processing without caching or parallel # processing. start = time.time() results = [data_processing_mean(data, col) for col in range(data.shape[1])] stop = time.time() print('\nSequential processing') print('Elapsed time for the entire processing: {:.2f} s' .format(stop - start)) ############################################################################### # ``costly_compute`` is expensive to compute and it is used as an intermediate # step in ``data_processing_mean``. Therefore, it is interesting to store the # intermediate results from ``costly_compute`` using :class:`joblib.Memory`. from joblib import Memory location = './cachedir' memory = Memory(location, verbose=0) costly_compute_cached = memory.cache(costly_compute) ############################################################################### # Now, we define ``data_processing_mean_using_cache`` which benefits from the # cache by calling ``costly_compute_cached`` def data_processing_mean_using_cache(data, column): """Compute the mean of a column.""" return costly_compute_cached(data, column).mean() ############################################################################### # Then, we execute the same processing in parallel and caching the intermediate # results. from joblib import Parallel, delayed start = time.time() results = Parallel(n_jobs=2)( delayed(data_processing_mean_using_cache)(data, col) for col in range(data.shape[1])) stop = time.time() print('\nFirst round - caching the data') print('Elapsed time for the entire processing: {:.2f} s' .format(stop - start)) ############################################################################### # By using 2 workers, the parallel processing gives a x2 speed-up compared to # the sequential case. By executing again the same process, the intermediate # results obtained by calling ``costly_compute_cached`` will be loaded from the # cache instead of executing the function. start = time.time() results = Parallel(n_jobs=2)( delayed(data_processing_mean_using_cache)(data, col) for col in range(data.shape[1])) stop = time.time() print('\nSecond round - reloading from the cache') print('Elapsed time for the entire processing: {:.2f} s' .format(stop - start)) ############################################################################### # Reuse intermediate checkpoints ############################################################################### # # Having cached the intermediate results of the ``costly_compute_cached`` # function, they are reusable by calling the function. We define a new # processing which will take the maximum of the array returned by # ``costly_compute_cached`` instead of previously the mean. def data_processing_max_using_cache(data, column): """Compute the max of a column.""" return costly_compute_cached(data, column).max() start = time.time() results = Parallel(n_jobs=2)( delayed(data_processing_max_using_cache)(data, col) for col in range(data.shape[1])) stop = time.time() print('\nReusing intermediate checkpoints') print('Elapsed time for the entire processing: {:.2f} s' .format(stop - start)) ############################################################################### # The processing time only corresponds to the execution of the ``max`` # function. The internal call to ``costly_compute_cached`` is reloading the # results from the cache. ############################################################################### # Clean-up the cache folder ############################################################################### memory.clear(warn=False)