Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

The Teton Compute Environment (Teton) is a high performance computing (HPC) cluster that offers over 500 compute nodes and 1.2 PB of storage, with an expected uptime of 98%, allowing researchers to perform computation-intensive analysis analyses on datasets of various sizes.

...

This links to a summary of the Teton hardware.

Teton Storage

Teton’s storage is divided into three isolated filesystems to ensure that researchers have control of where their data are, and who can access it. 

...

User Accounts require a valid UWyo email address and a UWyo-Affiliated PI sponsor. UWyo faculty members can sponsor their own accounts, while students, post-doctoral researchers, or research associates must use their PI as their sponsor. Users with a valid UWyo email address can be added in the project request or added later, using the Request New Account Project form.

Non-UWyo external collaborators (Ex_Co) must be sponsored by a current UWyo faculty member. Ex_Co accounts can be requested here.

Once access is granted, connection to ARCC HPC resources may be established via SSH. Note that SSH connections require Two-Factor Authentication.

Condo Model

The model for sustaining the Condo program is premised on faculty and principal investigators using equipment purchase funds from their grants or other available funds to purchase compute nodes (individual servers) which are then added to the Teton compute cluster. Condo computing resources are used simultaneously by multiple users. Teton is a condo model resource and as such, investors do have priority on invested resources. This is implemented through preemption and jobs not associated with the investment could be requested on the system when investor submits jobs. However, if the investor chooses not to implement preemption on their resources, ARCC can disable preemption while offering next-in-line access if that mode is preferred.

  • There are default concurrent limits in place to prevent individual project accounts and users from saturating the cluster away from others. The default limits are listed below. To incentivize investments into the condo system, investors will have their limits increased.

  • The system leverages a fairshare mechanism to offer a mechanism for projects that execute jobs on a more rare occasion priority over those who continuously run jobs on the system. To incentivize investments into the condo system, investors will have their fairshare value increased as well.

  • Finally, individual jobs occur runtime limits based on a study that was performed in ~2014 such that our maximum walltime for a compute job is 7 days. ARCC is currently evaluating this to determine whether the orthogonal limits of CPU count and walltime are optimal operational modes. ARCC is considering concurrent usage limits based on a relational combination of CPU count, Memory, and walltime that would allow more flexibility for different areas of science. There will likely still be an upper limit on individual compute job walltime as ARCC will not allow infinite job walltime and due to possible hardware faults.

Citing Teton

For information on citing Teton, please reference the citing section in Documentation and Help.