Last active
February 3, 2023 15:23
-
-
Save jyesselm/37d3a419aadf675c0cb8e839ddc4caba to your computer and use it in GitHub Desktop.
a simple example for getting dask to work on a cluster with SLURM job queue
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
""" | |
To install | |
conda create -n dask-tutorial python=3.7 anaconda | |
conda activate dask-tutorial | |
conda install dask | |
conda install -c conda-forge dask-jobqueue | |
""" | |
import dask | |
import time | |
from dask.distributed import Client, progress, get_worker | |
from dask_jobqueue import SLURMCluster | |
# create a SLURMCluster instance with extra lines to load the Anaconda module | |
cluster = SLURMCluster(processes=1, cores=1, memory='4GB', walltime='1:00:00', | |
job_script_prologue=['#!/bin/bash', 'module load anaconda']) | |
cluster.scale(10) # this may take a few seconds to launch | |
# define a Dask function | |
def slow_increment(x): | |
time.sleep(1) | |
return(x+1) | |
# start client | |
client = Client(cluster) | |
futures = client.map(slow_increment, range(500) ) | |
results = client.gather(futures) | |
print(results) | |
futures = client.map(slow_increment, range(500, 1000, 1)) | |
results = client.gather(futures) | |
print(results) | |
# wait for the task to complete and retrieve the result | |
# close the cluster and the client | |
cluster.close() | |
client.close() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment