HPC Cluster Job Scheduler: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
No edit summary |
||
Line 43: | Line 43: | ||
</pre> | </pre> | ||
=== Job Arrays === | === [https://slurm.schedmd.com/job_array.html Job Arrays] === | ||
Example 1: | Example 1: | ||
<pre> | <pre> | ||
Line 50: | Line 50: | ||
</pre> | </pre> | ||
Example 2: | Example 2 (step size): | ||
<pre> | |||
#SBATCH --output=job.%A_%a.out | |||
#SBATCH --array=1-100:20 | |||
</pre> | |||
Example 3 (limit simultaneous task): | |||
<pre> | <pre> | ||
#SBATCH --output=job.%A_%a.out | #SBATCH --output=job.%A_%a.out |
Revision as of 18:10, 2 June 2019
This content is under construction. Check back often for updates.
Submitting Your First HPC Job
Content to be created.
Anatomy of a SLURM Sbatch Submit Script
Content to be updated.
!/bin/bash #SBATCH --workdir=./ # Set the working directory #SBATCH --mail-user=nobody@tcnj.edu # Who to send emails to #SBATCH --mail-type=ALL # Send emails on start, end and failure #SBATCH --job-name=pi_dart # Name to show in the job queue #SBATCH --output=job.%j.out # Name of stdout output file (%j expands to jobId) #SBATCH --ntasks=4 # Total number of mpi tasks requested #SBATCH --nodes=1 # Total number of nodes requested #SBATCH --partition=test # Partition (a.k.a. queue) to use # Disable selecting Infiniband export OMPI_MCA_btl=self,tcp # Run MPI program echo "Starting on "`date` mpirun pi_dartboard echo "Finished on "`date`
Advanced Submit Script Options
Content to be created.
Constraints
Available constraints.
Example:
#SBATCH --constraint=skylake
Node Exclusivity
Example:
#SBATCH --exclusive
Job Arrays
Example 1:
#SBATCH --output=job.%A_%a.out #SBATCH --array=1-100
Example 2 (step size):
#SBATCH --output=job.%A_%a.out #SBATCH --array=1-100:20
Example 3 (limit simultaneous task):
#SBATCH --output=job.%A_%a.out #SBATCH --array=1-100%5
Example Submit Scripts
Content to be created.
ELSA Job Partitions/Queues
Parition/Queue Name | Max Time Limit | Resource Type |
---|---|---|
short | 6 hours | CPU |
normal | 24 hours | CPU |
long | 7 days | CPU |
nolimit* | none | CPU |
shortgpu | 6 hours | GPU |
gpu | 7 days | GPU |
* - Use of the nolimit partition is restricted to approved cluster users. Faculty may request access for themselves and students by emailing ssivy@tcnj.edu.