Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Request your desired resources without defining a partition

Assuming:
(tick) your project is associated with an investment
(tick) and you did NOT not define a partition in your request

  1. Slurm will identify a list of partitions to try to run your job against.

  2. Slurm will automatically prioritize your investment partition and hardware to try to run your job there. 

  3. Slurm checks for jobs running on your investment.

    1. If:
      (tick) there are no jobs running on your investment
      (tick) and the hardware requested in your job fits within the hardware confines of your partition
      Slurm immediately starts your job.

    2. If:
      (tick) there are jobs running on your investment
      (tick) and they were requested by HPC users who are not a member of any projects associated with your investment

      1. Slurm will pre-empt these jobs (ie stop them and add back to the queue)

      2. Slurm will immediately start your job.

    3. If:
      (tick) your investment is 'full' with jobs from users who are members of your project

      1. Slurm will try to allocate across the other partitions if resources are available.

        1. On MedicineBow: The list of other partitions tried in order are: mb, mb-a30, mb-l40s, mb-h100.

        2. On Beartooth: The list of other partitions tried in order are: moran, teton, teton-cascade, teton-gpu, teton-hugemem.

    4. If:
      (tick) there are no resources available to fit your job (i.e. cluster usage is very high)

      1. then your job will have a state of pending (i.e. waiting in the queue).

      2. Slurm will monitor the queue on regular intervals and run the job when any appropriate resources become available.

...