Slurm Workload Manager
Overview
The Slurm Workload Manager is a powerful and flexible workload manager used to schedule jobs on high performance computing (HPV) clusters. The Slurm Workload Manager can be used to schedule jobs, control resource access, provide fairshare, implement preemption, and provide record keeping. All compute activity should be used from within a Slurm resource allocation (i.e., job). ARCC utilizes Slurm is on Teton and Loren.
Slurm is an open source, fault-tolerant, and highly scalable cluster management and job scheduling system for large and small Linux clusters. Slurm requires no kernel modifications for its operation and is relatively self-contained.
Please view the child pages for more details and examples on understanding, learning and using Slurm:
Job Submission
Getting started with SLURM, submitting jobs, common commands.
Submitting general CPU jobs.
Submitting jobs requesting specific allocations for memory and GPUs.
Submitting jobs using job arrays.
Submitting child jobs from a main “parent” job (jobs within a job).
HPC System Querying
HPC System Querying: Getting information about your jobs, available hardware, efficiency, etc.
Querying job efficiency
More In-Depth Training on SLURM
Want to learn more about using SLURM? Take a look at our in depth training module on SLURM here.