Skip to main content

Home

MonARCH

Merge of MonARCH and M3

After over nine years of service, the MonARCH HPC cluster will be retired. Moving forward, the HPC workloads on MonARCH will be served by M3.

In the coming weeks, our eResearch HPC team will be reaching out to you on how your current MonARCH project(s) will be migrated to the M3 HPC system. In the meantime, please continue running your current workloads on MonARCH as per usual.

For any concerns and queries, please feel free to email us.

Training in 2025

For information on training, please see our new training page.

MonARCH is Retiring this July 2025

We cordially invite all current and prospective MonARCH users to apply for an appropriate allocation on M3. There are two options:

  • larger MonARCH projects requiring 3 or more TBs of storage are to apply for a Monash M3 allocation; or
  • smaller MonARCH projects requiring < 500 GB and involving one or two project members to apply for a startup (PEACH) allocation.

The PEACH (Personal Easy Allocation of Compute on HPC) is a new offering on M3 for small, personal-sized HPC projects. A PEACH project may be lead by a PhD student and is given a small allocation of space. Many existing MonARCH projects fall under this category.

This table shows the differences between a PEACH project and a Monash M3 project:

PEACH projectMonash M3 project
Project Leadercan be a studentacademic researcher, lab head, group leader
# of membersideal for one or tworesearch group / lab
CPU limits per user100 CPU coresMonash M3 standard
DesktopsYes (CPU only)Yes with GPUs
Accessible partitionsdedicated peach partitionall other M3 partitions
GPUsNo GPU accessup to four GPUs per user
Project Space (initial)50 GB of protected spaceMonash M3 standard: https://docs.erc.monash.edu/M3/Files/StorageQuotas
Scratch Space (initial)100 GB of /scratch2 spacesee above

Applying for your M3 allocation

Note that your current MonARCH project will not be carried over to M3.

You need to apply for a new project either under the PEACH allocation or a Monash M3 allocation.

The PEACH allocation is appropriate if you are a PhD student working on your own project and do not require large storage space and no GPUs. If you need more than the 50 GB protected and 100 GB /scratch2 space for PEACH, ask your research supervisor to join and co-lead your project; and we can increase your allocation as required. Please see: https://docs.erc.monash.edu/M3/Files/StorageQuotas

Larger MonARCH projects, like those with several project members and is using 10s of TBs of space, please ask your academic supervisor to apply for a Monash M3 allocation. This will provide a starting 500 GB of protected space and 3 TB of /scratch2 space. More quota may be requested as needed, please see: https://docs.erc.monash.edu/M3/Files/StorageQuotas

After submitting your M3 project request, we will notify you when the project is ready. Your M3 project will be given a new project ID different from that on your existing MonARCH project.

You will need to copy important data on monarch into your new M3 project space.

If your current MonARCH allocation is bigger than the one we will grant on M3, please advise us as to:

  • if you need the data on M3; or
  • if this data can be cleaned up and/or archived.

We recommend all projects to apply for a vault allocation to be used for archiving your valuable data. Please see: https://docs.erc.monash.edu/RDS/StorageProducts/VaultStorage on how to secure an allocation.

Contact: mcc-help@monash.edu if you need assistance with data migration.


[To be deprecated] Documentation on the current MonARCH System

Welcome to the MonARCH documentation. Please select from the options at the left to get started.

MonARCH (Monash Advanced Research Computing Hybrid) is the next-generation HPC/HTC Cluster, designed from the ground up to address the emergent and future needs of the Monash HPC community.

Through the use of advanced cloud technology, MonARCH is able to configure and grow dynamically. As with any HPC cluster, MonARCH presents a single point-of-access to computational researchers to run calculations on its constituent servers.

MonARCH aims to continually develop over time. Currently, it consists of the following servers

NameCPUNumber of Cores / ServerUsable Memory / ServerNotes
mi*Xeon-Gold 6150 @ 2.70GHz36158893MB
hi*Xeon-Gold 6150 @ 2.70GHz27131000MBSame hardware as mi* nodes, but with less cores/memory in the VM
ga*Xeon-Gold-6330 @ 3.10GHz56754178MBEach server has two A100 GPU devices
gd*Xeon-Gold-6448Y @ 4.1GHG64774551MBEach server has two A40 GPU devices
hm00Xeon-Gold-6150 @ 2.70GHz261419500MBSpecialist High Memory ~1.4TB machine. Please contact support to get access
md*Xeon(R) Gold 5220R @ 2.2GHz48735000MBThe most recent Monarch Nodes which are baremetal
mk*Xeon-Platinum-8260 @ 2.50GHz48342000MB
ms*Xeon-Gold-6338 @ 2.00GHz64505700MBThe most recent Monarch Nodes

For data storage, we have deployed a parallel file system service using Intel Enterprise Lustre; providing over 300 TB usable storage with room for future expansion.

The MonARCH service is operated by the Monash HPC team and continuing technical and operational support from the Monash Cloud team, and eSolutions Servers-and-Storage, and Networks teams.

Acknowledgement and Citation

If you have found MonARCH useful for your research, we will be very grateful if you kindly acknowledge us. Find out how to acknowledge MonARCH in your publications.