Skip to content

SLURM job time configuration in CLI mode #4

@mgharawy

Description

@mgharawy

I have been trying to launch an example job on a slurm cluster in CLI mode.
Slurm is configured to require --time to be specified in all jobs, so when the pyp tries to start workers the sbatch command fails because --time wasn't specified.

Error message:

sbatch: error: Please specify --time= for your job.
sbatch: error: Batch job submission failed: Requested time limit is invalid (missing or exceeds some limit).

I tried adding it to the pyp command as shown below but it wasn't successful

cd nextPyP
module load singularity
export PYP_CONFIG=${PWD}/config.toml
./pyp \
    -data_mode spr                                  \
    -scope_voltage 300                              \
    -scope_pixel 0.66                               \
    -data_path="./T20S/nextpyp_spr_tutorial/14*.tif"       \
    -gain_reference="./T20S/nextpyp_spr_tutorial/Gain.mrc" \
    -gain_flipv                                     \
    -detect_rad 80                                  \
    -detect_method all                              \
    -detect_dist 40                                 \
    -slurm_tasks 7                                  \
    -slurm_memory 70                                \
    -slurm_merge_tasks 7                            \
    -slurm_merge_memory 70			    \
    -slurm_merge_walltime 00:20:00                \
    -slurm_verbose                                \
    -slurm_profile

I have also tried using -slurm_walltime instead and got same error.

My config.toml file:

[pyp]

# path to the PYP container file, from a SLURM compute node's point of view
container = '/home/mgh/nextPyP/pyp.sif'

# fast storage for temporary files, ideally local to the compute node
scratch = '/home/mgh/tmp'

# Bind network locations as needed
binds = []
#containerExec = { module = 'singularity', exec = 'singularity' }
containerExec = '/sw/rl9g/singularity/3.9.7/rl9_go1.19.4/install/bin/singularity'

[slurm]
# user account for the SLURM login node
user = 'mgh'

# hostname for the SLURM login node
host = 'login.node'

# configure SLURM partitions here
queues = [ "batch" ]

# path to SLURM executables
path = '/opt/slurm/cluster/install/RedHat-9/bin/'

Is there another way to configure it in CLI?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions