I am trying to upload this job via a .sh script to a cluster with SLURM, using the COMSOL software: #!/bin/bash #SBATCH --job-name=my_work #SBATCH --nodes=1
Trying to run slurmd: sudo systemctl start slurmd I display the status of the daemon and an error is displayed on the screen: >>sudo systemctl status slu
I've recently started a new job and need to run some scripts on the HPC through Slurm. My scripts are written in Python, and therefore I want to execute these u
I have a folder structure which is like this: /home/01/01/script.R /home/01/02/script.R /home/01/03/script.R /home/02/01/script.R /home/02/02/script.R /home/02/
Is it possible to write a bash script that opens a remote node (i.e. through ssh and/or slurm) and starts an interactive session there after running some comman
I was previously using HTCondor as a cluster scheduler. Now even after reading Slurm documentation, I have no idea how to parallelize... What I want to achieve
Hi I get this message when I run my job in slurm what does it mean? tensorflow/core/platform/default/subprocess.cc:304] Start cannot spawn child process: No suc
I'm configuring a small cluster, controlled by SLURM. This cluster has one master node and two partitions. Users submit their jobs from worker nodes, I've rest
Let's say I have 6233 simulations to run. The commands are generated and stored in a file, one in each line. I would like to use Slurm to schedule and run these
Is there some way to configure an alternative default directory (other than the current directory) for sbatch to put the file slurm-%j.out (or slurm-%A_%a.out)
I'm running into a problem using a Xarray together with SLURMcluster from Dask. I'm using pandas_plink to load some data into a Xarray, then filtering it and ma
I was wondering if I could ask something about running slurm jobs in parallel.(Please note that I am new to slurm and linux and have only started using it 2 day
I am queuing multiple jobs in SLURM. Can I limit the number of parallel running jobs in slurm? Thanks in advance!
I am trying to run WRF (real.exe, wrf.exe) through the crontab using compute nodes but compute nodes are not able to run slurm job. I think there is some issue
When running a Python script via slurm srun --pty bash I get a cryptic error message Bus error: core dumped. I searched the slurm documentation and it doesn't m