Skip to main content

File Systems on MonARCH

The MonARCH File System is arranged into 3 major parts:

  • your home directory
  • your project directory
  • project scratch space

Also, when a job is running via the Slurm scheduler, it is also given a unique /tmp and /var/tmp directory on a local (ephermeral) disk. See Extra Features Available for more information.

image

File SystemQuota> Description
/home5G per userNFS mount (CEPH)
/mnt/lustre50GB or 500GB. See textLUSTRE
/mnt/scratchNo quota but disks are >1TB each. (See below re use)Local to server
/tmp /var/tmpSame as /mnt/scratch. Or NVMe disk if presentLocal to server

In your home directory you should find a number of symbolic links (ls -al). These point to the project space and scratch you have been allocated. You can request additional space via mcc-help@monash.edu.au

For example if your project name is YourProject001 you will see the following two links:

ls -al ~/
YourProject001 -> /projects/YourProject001
YourProject001_scratch -> /scratch/YourProject001

The first link points to your project data that is backed up weekly (as with your home directory). The second one points to the faster system scratch space, which is not backed up and is used for temporary data.

File Systems​

/home​

This is where the user's home directory resides. It is a NFS-mounted CEPH file system. Currently (as of July, 2016), users' home directories are backed up nightly to a disk/tape archival system (See https://vicnode.org.au/products-4/vault-tape-mon/ for more information). Any changes in files are kept for 30 days.

/mnt/lustre​

This is our high-performance system-wide parallel LUSTRE file system, which has over 300TB of capacity. Each user has two links in their home directory that point to LUSTRE.

  • $HOME/scratch This is a link to a scratch directory which is unique for each user. This is for temporary use during the running of a program.
  • $HOME/<projectID> This is a pointer to a project directory, i.e. to store input and output files. Users within a group will have the same project directory. An example of a project directory is: p2015120001.

The LUSTRE file system is NOT backed up at the moment. A backup / archive solution is in preparation. Currently, each project has a default soft quota of 500 GB. This project requires a supervisor to lead. A PhD student may request a project space of 50 GB. We run a periodic check of disk usage to determine which projects have exceeded their quotas, and we will contact the affected users to resolve this. This avoids jobs being aborted due to exceeding disk quotas. Users may requests an increase in project quotas, and these requests will be decided on a case-by-case basis.

/mnt/scratch​

This is a scratch space that exists on each compute node. It uses the local hard disk of the machine that supports the Virtual Machine. Using Nectar Openstack terminology, these disks are ephemeral disks. The /mnt/scratch partition is different on each compute node, i.e. the /mnt/scratch on the MonARCH login node is different to the one on the compute nodes. For example, if a SLURM job running on a compute node writes to its /mnt/scratch, you will not see any of the files when you are on the MonARCH login node.

These disks are suitable for small, frequent IO operations. Any files written here should be copied to your /home or /mnt/lustre file systems before the job is finished. The disks are not backed up.

PLEASE CLEAN UP YOUR FILES AFTER USE. You should not assume the files will persist after your job has finished. If OS file system fills up, the machine will hang or crash.

What to put on each file system?​

That is up to you, but as a general guide:

Home directory (~2GB)​

This should contain all of your hidden files and configuration files. Things like personal settings for editors and other programs can go here. The home directory is backed up.

Project directory (Shared with everyone on your project)​

This area is not backed up but is faster and has more space. It should contain all of your input data, a copy of all your job scripts and final output data. It might be worth while to keep copies of any spreadsheets you might use to analyse results or any matlab or mathematica scripts/programs here as well. Basically anything that would be hard to regenerate.

Generally each user in the project should create a subdirectory in the project folder for themselves.

Scratch directory (Shared with everyone on your project)​

This area is not backed up. Generally all your intermediate data will go here. Anything that you can recreate by submitting a job script and waiting (even the job runs for quite a long time) can go here. Anything that you can't regenerate automatically, things that you have to think about and create rather than asking the computer to calculate should go in the project directory because that is backed up.

Disk Quotas​

MonARCH uses soft and hard quota limits.

A soft limit allows you to have more than your allocated space for a short period of "grace" time (in days). After the grace time has been exceeded, the filesystem will prevent further files being added until the excess is removed.

A hard limit prevents further files being added.

The quotas on the Project directories are much larger than the space users get in their own Home directories, so it is much better to use the Project space. Also the project space is available for all members of your project, so you can use it to share data with your colleagues.

Please use the user_info command to view your usage and quota on all your projects and /scratch.

If you need higher allocation for project storage spaces, please send your request to mcc-help@monash.edu.au

System Backups and File Recovery​

The data storage on MonARCH is based on Lustre which distributes data across a number of disks and provides a mechanism to manage disk failures. This means that the MonARCH file system is fault tolerant and provides a high level of data protection.

In addition to this, the home directories are backed up to tape every night. This means that if you create a file on Tuesday, on the following day there will be a copy of the file stored in the backup system. Backups are kept for 30 days, before the system permanently removes the file to make space for new data.

File SystemTypeQuotaPolicyHow long are backups kept?
Home DirectoryNFSYesDaily Backup30 days
Project DirectoryLustreNoNo
Scratch DirectoryLustreNoNo

File Recovery Procedure​

If you delete a file/directory by mistake, you will be able to recover the file by following the following the following procedure:

  • Email a request to mcc-help@monash.edu.au
  • Please include the the full path to the missing data, as well as information on when it was last seen and when it was deleted
  • We will be able to restore files modified within the 30 day window. Beyond that time, any changes in the file will be lost.

The project scratch space is not backed up. Files are purged from this space as new space is required.

Storage outside of MonARCH​

With your project, you have an allocation of storage on its high performance Lustre file system. This storage space is intended for data analyses and has a limited capacity. For large-scale, secure, and long-term research data storage, Monash University has the following offerings available through VicNode:

  • Vault, primarily used as archive, is a tape-based system specifically for long-term storage; this is best used to free up space on your MonARCH project, allowing for more data to be staged into your project for analyses. For further information, please visit: https://vicnode.org.au/products-4/vault-tape-mon/
  • Market, is backed-up storage intended for active data sets and is accessible through the Windows, Linux, or Mac desktop environments at your research laboratory for convenient sharing of data files. For further information, please visit: https://vicnode.org.au/products-4/market-mon

We invite all MonARCH project leaders to kindly make a storage request to mcc-help@monash.edu.au , which we will forward to VicNode.

All additional storage requests can be lodged with the Research Storage team via the Data Dashboard or contacting researchdata@monash.edu

Instructions to access Ceph Market​

The Market allocation is presented as a CIFS share with a given name, usually of the form: RDS-R-<Faculty>-<Name>. This share can be mounted within a Desktop session as follows:

  1. Open a Terminal window within your Desktop session and issue this command:

    gvfs-mount smb://storage.erc.monash.edu.au/shares/<sharename>
    • Replace the <sharename> with the one provided by your allocation;
    • Enter your Monash ID (previously known as Authcate) username, when prompted;
    • enter MONASH when prompted to enter the "Domain"; and
    • finally your Monash ID password on the "Password" prompt.
    info

    gvfs-mount is not available on MonARCH login nodes, use desktop (Strudel) to access the share.

  2. If successful, the mounted share will be visible through the file browser. If the user is not a member of the group, an "access denied" message will be displayed.

  3. It is best to cleanly unmount the share when it is no longer needed, by using this command:

gvfs-mount -e smb://storage.erc.monash.edu.au/shares/<sharename>

However, the share will be automatically unmounted once the desktop session terminates.

The collection owner will/should be able to add and/or remove collaborators who can mount the share; through the eSolutions Group Management page. On this page, a list of shares that you have admin privileges will appear, each of these shares will appear as: RDS-R-<Faculty>-<Name>-Shared.

Important Note

It is a known issue that the available storage for the share is incorrectly reported. Users are advised to simply ignore the warning, and allow a copy/move to proceed. We are unable to add non-Monash users to mount a share, since this authenticates against the Monash AD.

Instructions to access Vault​

Coming soon.

In the meantime, please contact: mcc-help@monash.edu.au.

Extra Features Available​

The following features are implemented using Slurm Plug-in Architecture for Node and job (K)control (SPANK).

/tmp directory​

During a running Slurm job, the /tmp and /var/tmp directories on each running slurm job are mapped to a unique portion of a local filesystem disk ( e.g. /mnt/scratch ) and is larger than normal /tmp allocations that you might see if you had logged into the computer via ssh.

warning

Be aware this /tmp will be deleted after the Slurm job ends. It is unique to your Slurm JOBID. If you need any data files stored by your program in /tmp , please move it to your file system before
before the job ends.

Non-Volatile Memory Express (NVMe) Disks​

The Slurm job will automatically use an NVMe drive as /tmp if there is one available, which means you may expect faster disk access on those nodes. Some of our mk machines contain NVMe drives.