Overview: Global Filesystems

The Teton global parallel filesystem configured with a 160 TB SSD tier for active data and 1.2 PB HDD capacity tier for less-used data. The system policy engine moves data automatically between pools (disks and tiers). The system will automatically migrate data to HDD when the SSD tier reaches 70% used capacity. Teton has several spaces that are available for users to access described in the table below.

home/home/username ($HOME)

project/project/project_name/[username]

gscratch - /gscratch/username ($SCRATCH)

Global Filesystems

Filesystem Name

Quota (GB)

Snapshots

Backups

Purge Policy

Additional Info

home

25

Yes

No

No

Always on SDD

project

1024

Yes

No

No

Aging Data will move to HDD

gscratch

5120

No

No

Yes

Aging Data will move to HDD

Snapshots vs Backups

Snapshots are a point in time reference of a specific filesystem, that can be referenced after changes are made. The data stays on the same storage system. This data would not be recoverable if there is an issue with the storage system.

Backups are the act of transporting data to another storage system for safe-keeping, recoverable if an issue occurs with the source file system.

Purge Policy

Storage Increases on Teton

Special Filesystems

Certain filesystems exist on different nodes of the cluster where specialized requirements exist. The table below summarizes these specialized filesystems.

Specialty Filesystems

Filesystem Name

Mount Location

Notes

petaLibrary

/petalibrary/homes

Only on login nodes

petaLibrary

/petalibrary/Commons

Only on login nodes

node local scratch

/lscratch

Only on compute nodes, Moran is 1 TB HDD, Teton is 240 GB SSD

memory filesystem

/dev/shm

RAM-based tmpfs available as part of RAM for very rapid I/O operations; small capacity

The node-local scratch or lscratch filesystem is purged at the end of each job.

The memory filesystems can really enhance the performance of small I/O operations. If you have a localized single node I/O jobs that have very intensive random access patterns, this filesystem may improve the performance of your compute job.

The petaLibrary filesystems are only available from the login nodes, not on the compute nodes. Storage space on the Teton global filesystems does not imply storage space on the ARCC petaLibrary or vice versa. For more information about the petaLibrary please see the following link petaLibrary.

The Bighorn filesystems will be provided for a limited amount of time in order for researchers to move data to either the petaLibrary, Teton storage, or to some other storage media. The actual date that these mounts will be removed is still TBD.