-
Notifications
You must be signed in to change notification settings - Fork 2
Open
Description
I have been trying to launch an example job on a slurm cluster in CLI mode.
Slurm is configured to require --time to be specified in all jobs, so when the pyp tries to start workers the sbatch command fails because --time wasn't specified.
Error message:
sbatch: error: Please specify --time= for your job.
sbatch: error: Batch job submission failed: Requested time limit is invalid (missing or exceeds some limit).
I tried adding it to the pyp command as shown below but it wasn't successful
cd nextPyP
module load singularity
export PYP_CONFIG=${PWD}/config.toml
./pyp \
-data_mode spr \
-scope_voltage 300 \
-scope_pixel 0.66 \
-data_path="./T20S/nextpyp_spr_tutorial/14*.tif" \
-gain_reference="./T20S/nextpyp_spr_tutorial/Gain.mrc" \
-gain_flipv \
-detect_rad 80 \
-detect_method all \
-detect_dist 40 \
-slurm_tasks 7 \
-slurm_memory 70 \
-slurm_merge_tasks 7 \
-slurm_merge_memory 70 \
-slurm_merge_walltime 00:20:00 \
-slurm_verbose \
-slurm_profile
I have also tried using -slurm_walltime instead and got same error.
My config.toml file:
[pyp]
# path to the PYP container file, from a SLURM compute node's point of view
container = '/home/mgh/nextPyP/pyp.sif'
# fast storage for temporary files, ideally local to the compute node
scratch = '/home/mgh/tmp'
# Bind network locations as needed
binds = []
#containerExec = { module = 'singularity', exec = 'singularity' }
containerExec = '/sw/rl9g/singularity/3.9.7/rl9_go1.19.4/install/bin/singularity'
[slurm]
# user account for the SLURM login node
user = 'mgh'
# hostname for the SLURM login node
host = 'login.node'
# configure SLURM partitions here
queues = [ "batch" ]
# path to SLURM executables
path = '/opt/slurm/cluster/install/RedHat-9/bin/'
Is there another way to configure it in CLI?
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels