Cluster keeps appending "interface" flag to job script
See original GitHub issueWhen I run the following code snippet:
from dask_jobqueue import SLURMCluster
cluster = SLURMCluster(memory="100GB", cores=40, interface="ib0")
for _ in range(20):
_ = cluster.job_script()
print(cluster.job_script())
I get the following output:
#!/usr/bin/env bash
#SBATCH -J dask-worker
#SBATCH -n 1
#SBATCH --cpus-per-task=40
#SBATCH --mem=94G
#SBATCH -t 00:30:00
/path/to/python -m distributed.cli.dask_worker tcp://xx.xx.xx.xx:pppp --nthreads 5 --nworkers 8 --memory-limit 11.64GiB --name dummy-name --nanny --death-timeout 60 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0 --interface ib0
It seems everytime the job script method is called, the interface flag is appended to it. I am running dask-jobqueue==0.8.0
off the conda-forge
channel. (Python 3.10 if relevant)
You can replicate this with PBSCluster
as well. The issue seems to be in how worker_extra_args
are handled in the __init__()
method of dask_jobqueue.core.Job
Issue Analytics
- State:
- Created a year ago
- Comments:7 (6 by maintainers)
Top Results From Across the Web
Cluster Execution — Snakemake 7.19.1 documentation
Consider using the --default-resources and --set-resources flags to define such resources on the command line. Additional custom job configuration¶. SLURM ...
Read more >How To Submit Parallel Jobs in O2 - Atlassian
The same job submission can be done using a script to pass slurm's flags and commands to be executed, see the main docs...
Read more >Job Scripts | Ohio Supercomputer Center
The sbatch flag --signal can be used to specify commands to be ran when these signals are received by the job. It is...
Read more >Cluster User Guide - Google Sites
Job partitions (job submission queues). Connecting to the cluster. Command-line access via SSH. Using the Open OnDemand interface. Using the FastX interface.
Read more >Configuring and managing high availability clusters Red Hat ...
The High Availability Add-On features two configuration tools for cluster deployment, monitoring, and management. pcs. The pcs command line interface ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Nice catch and many thanks @jolange.
Ok, this is a subtle combination of two things:
worker_extra_args
: https://github.com/dask/dask-jobqueue/blob/f79f9136542abd86566e0a36f7370c144052ee9d/dask_jobqueue/core.py#L226-L228X = X + NEW
toX += NEW
:This changes the config, whereas before
worker_extra_args
(orextra
then) was newly assigned. I will revert that!