Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Expand
titleHow do I connect to Wildiris?

Wildiris is now part of the Medicinebow cluster. Users logging into Medicinebow with an arcc-only (non-UWYO) account should follow instructions here.

...

MedicineBow SLURM Job Scheduling (After Changes Jan 13, 2025)

Expand
titleWhy can't I submit a CPU job on a GPU node?

Users that are not associated with an investment cannot run CPU jobs on a GPU node unless their specified job nodes fall within their investment hardware.
In all other cases, users are restricted from running CPU jobs on a GPU node and those jobs should be requested with a --gres:gpu flag in their submission script or salloc command.

Expand
titleWill my job run if I don't specify a QoS?

If you do not specify a QoS as part of your job, a QoS will be assigned to that job based on partition or wall-time. Different partitions and wall-times are associated with different QoS, as detailed in our published Slurm Policy. Should no QoS, partition, or wall-time be specified, the job by default will be placed in the Normal queue with a 3 day wall-time by default.

Expand
titleWill my job run if I don’t specify wall-time?

Similar to jobs with unspecified QoS, wall-time is assigned to a job based on other job specifications, like QoS or partition. Specific QoS or partitions in a job submission will result in a default wall-time associated with those other flags. If no QoS, partition, or wall-time is specified, the job by default is placed in the Normal queue with a 3 day wall-time.

Expand
titleDo I need to specify a partition?

If you are requesting a GPU, you must also specify a partition with GPU nodes. Otherwise, you are not required to specify a partition. Users requesting GPUs should likely use a --gres=gpu:# or --gpus-per-node flag AND a --partition flag in their job submission.

Expand
titleWhy am I unable to request an OnDemand session for longer than 8 hours?

To encourage users to use only the time they need, all interactive jobs, including those requested through OnDemand have been limited to 8 hours in length. Please specify a time from the OnDemand webform under 8 hours.

Expand
titleMy job has been sitting in queue for a very long time without running. Why?

This is usually the result of specified walltime. If you have specified a 7 day walltime in your job using --time or -t flag over 3 days, you will be placed in the “long” queue which may result in a longer wait time. If your job doesn’t require 7 days, please try specifying a shorter walltime (ideally under 3 days). This should result in your job being placed in a queue with a shorter wait time.

Expand
titleI'm getting the following error:

sbatch/salloc: error: Interactive jobs cannot be longer than 8 hours

Post maintenance, interactive jobs are restricted to an 8 hour walltime. Please submit your salloc command with a walltime 8 hours or less.
Example:

salloc -A projectname -t 8:00:00

sbatch/salloc: error: You didn't specify a project account (-A,--account). Please open a ticket at arcc-help@uwyo.edu for help

If accompanied by “sbatch/salloc: error: Batch job submission failed: Invalid account or account/partition combination specified” it’s likely you need to specify an account in your batch script or salloc command, or the account name provided after the -A or --account flag is invalid. The account flag should specify the name of the project in which you’re running your job. Example:
salloc -A projectname -t 8:00:00

sbatch/salloc: error: Use of --mem=0 is not permitted. Consider using --exclusive instead

Users may no longer request all memory on a node using the --mem=0 flag and are encouraged to request only the memory they require to run their job. If you know you need the use of an entire node, replace your --mem=0 flag specification in your job with --exclusive to get use of an entire node an all it’s resources.

sbatch/salloc: error: QOSMinGRES

Users must specify a GPU device if requesting a GPU partition. Assuming you plan to use a GPU in your computations, please specify a GPU by including either the --gres=gpu:# or --gpus-per-node=# flag in your job submission.

sbatch/salloc: error: Job submit/allocate failed: Job violates accounting/QOS policy (job submit limit, user's size and/or time limits)

This may occur for a number of reasons. Please e-mail arcc-help@uwyo.edu with the location of the batch script you’re attempting to run, or salloc command you’re attempting to run, and the error message you receive.

salloc: error: Interactive jobs must be run under the 'interactive' QOS or 'debug' QOS, not 'fast'

Users must specify the interactive or debug queue, or a time under 8 hrs when requesting an interactive job.

sbatch/salloc: error: Job submit/allocate failed: Invalid qos specification

Users should specify a walltime that is available for their specified queue. i.e.,
Debug (<= 1 hr)
Interactive (<= 8 hrs)
Fast (< = 12 hrs)
Normal (<= 3 days)
Long (<= 7 days)

sbatch: error: Batch job submission failed: Requested node configuration is not available

This may occur for a number of reasons, but is likely due to the combination of nodes and hardware you’ve requested, and whether that hardware is available on the node/partition. If you need assistance please e-mail arcc-help@uwyo.edu with the location of the batch script you’re attempting to run, or salloc command you’re attempting to run, and the error message you receive.

 

...

HPC Migration Questions: Beartooth to MedicineBow

...

Expand
titleHow do I access my Alcova allocation when I am not on the UWyo network?

How do I access my Alcova allocation when I am not on the UWyo network?

Users will need to use the UWyo VPN (wyosecure) to get onto the UWyo network.

Expand
titleI would like to add user(s) to my Alcova Project

I would like to add user(s) to my Alcova Project

Only the project PI may make changes to Alcova Project permissions. If you are not the PI on the project, ARCC will need to obtain the PI’s approval to make changes to project permissions and add members. If you are the project PI, you may request a project change through our portal.

Expand
titleI can’t access my Alcova data at alcova.arcc.uwyo.edu/<projectname>

I can’t access my Alcova data at alcova.arcc.uwyo.edu/<projectname>

Alcova was migrated to the ARCC Data Portal effective June 2024. Please see this page for information on accessing the your new project space.

If this doesn’t help, please contact ARCC at arcc-help@uwyo.edu so we may troubleshoot your issue.

...