HPC Cluster Job Scheduler: Difference between revisions

From HPC Docs
Jump to navigation Jump to search
Line 29: Line 29:
''Content to be created.''
''Content to be created.''


Using [[HPC_SLURM_Features|constraints]].
Available [[HPC_SLURM_Features|constraints]].


Example
Example

Revision as of 19:37, 31 May 2019

This content is under construction. Check back often for updates.

Submitting Your First HPC Job

Content to be created.

Anatomy of a SLURM Sbatch Submit Script

Content to be updated.

!/bin/bash

#SBATCH --workdir=./                     # Set the working directory
#SBATCH --mail-user=nobody@tcnj.edu      # Who to send emails to
#SBATCH --mail-type=ALL                  # Send emails on start, end and failure
#SBATCH --job-name=pi_dart               # Name to show in the job queue
#SBATCH --output=job.%j.out              # Name of stdout output file (%j expands to jobId)
#SBATCH --ntasks=4                       # Total number of mpi tasks requested
#SBATCH --nodes=1                        # Total number of nodes requested
#SBATCH --partition=test  		 # Partition (a.k.a. queue) to use

# Disable selecting Infiniband
export OMPI_MCA_btl=self,tcp

# Run MPI program
echo "Starting on "`date`
mpirun pi_dartboard
echo "Finished on "`date`

Advanced Submit Script Options

Content to be created.

Available constraints.

Example

#SBATCH --constraint=skylake

Example Submit Scripts

Content to be created.

ELSA Job Partitions/Queues

Parition/Queue Name Max Time Limit Resource Type
short 6 hours CPU
normal 24 hours CPU
long 7 days CPU
nolimit* none CPU
shortgpu 6 hours GPU
gpu 7 days GPU

* - Use of the nolimit partition is restricted to approved cluster users. Faculty may request access for themselves and students by emailing ssivy@tcnj.edu.