"""
Using Dask for single-machine parallel computing
================================================
This example shows the simplest usage of the
`Dask `_
backend on your local machine.
This is useful for prototyping a solution, to later be run on a truly
`distributed Dask cluster
`_,
as the only change needed is the cluster class.
Another realistic usage scenario: combining dask code with joblib code,
for instance using dask for preprocessing data, and scikit-learn for
machine learning. In such a setting, it may be interesting to use
distributed as a backend scheduler for both dask and joblib, to
orchestrate the computation.
"""
###############################################################################
# Setup the distributed client
###############################################################################
from dask.distributed import Client, LocalCluster
# replace with whichever cluster class you're using
# https://docs.dask.org/en/stable/deploying.html#distributed-computing
cluster = LocalCluster()
# connect client to your cluster
client = Client(cluster)
# Monitor your computation with the Dask dashboard
print(client.dashboard_link)
###############################################################################
# Run parallel computation using dask.distributed
###############################################################################
import time
import joblib
def long_running_function(i):
time.sleep(0.1)
return i
###############################################################################
# The verbose messages below show that the backend is indeed the
# dask.distributed one
with joblib.parallel_config(backend="dask"):
joblib.Parallel(verbose=100)(
joblib.delayed(long_running_function)(i) for i in range(10)
)
###############################################################################