Status: (more...)

Spartan is High Performance Computing (HPC) system operated by Research Platform Services (ResPlat) at The University of Melbourne. It combines a high performance bare-metal compute with flexible cloud infrastructure and GPGPU to suit a wide range of use-cases.

If your computing jobs take too long on your desktop computer, or are simply not possible due to a lack of speed and memory, a HPC system like Spartan can help.

Use of this service is governed by the University's general regulations for IT resources and our HPC Support Service Policy.

Scheduled Outage

Please not that the University of Melbourne Research Cloud will have a scheduled maintenance from December 17, 9am, to December 18, 6pm

https://research.unimelb.edu.au/infrastructure/research-platform-services/news-events/scheduled-maintenance

This will also effect the spartan-rc nodes, including the "cloud" partition.

Walltimes have already been reduced to account for this.

Spartan Daily Weather Report (20181213)

  • CephFS usage: 846T Free 332T
  • Spartan is very busy on cloud partition, with close to 99% node allocation.
  • Spartan is very busy on physical partition, with close to 100% node allocation.
  • Spartan is busy on the GPGPU partition with 93% node allocation.
  • Many cloud nodes out (67), mainly due to qh2-uom migration.

Getting Help

Training

We run regular one-day courses on HPC, shell scripting and parallel programming. ResPlat also offer training in a wide range of other digital tools to accelerate your research.

Signup here: http://melbourne.resbaz.edu.au/participate

Helpdesk

If you can't find an answer here, need advice, or otherwise stuck, you can contact our support team at hpc-support@unimelb.edu.au

Please submit one topic per ticket. If you require a assistance with a separate matter, compose a new ticket. Do not reply to existing or closed tickets.

For password resets please see the FAQ or contact University Services on +61 3 8344 0999 or ext 40999 or email service-centre@unimelb.edu.au.

Specifications

Spartan has a number of partitions available for general usage. A full list of partitions can be viewed with the command sinfo -s.

Partition Nodes Cores/node Memory/node Processor Extra notes
cloud 210 8 62GB Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz
longcloud 2 8 62GB Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz Max walltime of 90 days
physical 19 12 254GB Intel(R) Xeon(R) CPU E5-2643 v3 @ 3.40GHz
6 32 508GB Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz
bigmem 2 36 1540GB Intel(R) Xeon(R) CPU E5-2697 v4 @ 2.30GHz
phi 4 256 190GB Intel(R) Xeon Phi(TM) CPU 7230 @ 1.30GHz Based on the Xeon Phi Knights Landing architecture
snowy 31 32 127GB Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz Currently reserved for Melbourne Bioinfomatics users
mig 12 32 127GB Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz Reserved for Melbourne Integrative Genomics users
gpgpu 73 24 127GB Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz 4 P100 Nvidia GPUs per node

Cloud

This partition is best suited for general-purpose single-node jobs. Multiple node jobs will work, but communication between nodes will be comparatively slow.

Physical

Each node is connected by high-speed 25Gb networking with 1.15 µsec latency, making this partition suited to multi-node jobs (e.g. those using OpenMPI).

GPGPU

See here for more details.

bigmem

This partition is suited to memory-intensive single-node workloads.

Other Partitions

There are also special partitions which are outside normal walltime constraints. In particular, shortcloud and shortgpgpu should be used for quick test cases; the partitions have a maximum time constraint of one hour.

By adding --time=00:59:00 to your cloud partition job, your job will run in a dedicated reservation, ensuring your job runs faster

Storage system

Spartan uses a storage system called CephFS. CephFS is a highly scalable, parallel and robust filesystem.

The total Spartan storage is broken up into 3 areas:

Location Capacity Disk type
/home 1760TB 7.2K SAS
/data/cephfs 1760TB 7.2K SAS
/scratch 650TB Sandisk Flash

NB: /home and /data/cephfs share the same pool of disks

Citing Spartan

If you use Spartan to obtain results for a publication, we'd appreciate if you'd cite our service, including the DOI below. This makes it easy for us demonstrate research impact, helping to secure ongoing funding for expansion and user support.

Lev Lafayette, Greg Sauter, Linh Vu, Bernard Meade, "Spartan Performance and Flexibility: An HPC-Cloud Chimera", OpenStack Summit, Barcelona, October 27, 2016. doi.org/10.4225/49/58ead90dceaaa

If you are using the LIEF GPGPU cluster for a publication, please include the following citation in the acknowledgements section of your paper:

This research was undertaken using the LIEF HPC-GPGPU Facility hosted at the University of Melbourne. This Facility was established with the assistance of LIEF Grant LE170100200.

Other Resources

Spartan is just one of many research IT resources offered by The University of Melbourne, or available from other institutions.

Nectar

Nectar is a national initiative to provide cloud-based Infrastructure as a Service (IaaS) resources to researchers. It's based on OpenStack, and allows researchers on-demand access to computation instances, storage, and a variety of application platforms and Virtual Laboratories.

Spartan runs some of it's computation resources in the Nectar cloud.

Melbourne Bioinformatics

Melbourne Bioinformatics run two large HPC systems for life sciences researchers. From 2019, Melbourne Bioinformatics systems will be incorporated into Spartan.

Multi-modal Australian ScienceS Imaging and Visualisation Environment (MASSIVE)

MASSIVE is a HPC system at Monash University and the Australian Synchrotron which is optimized for imaging and visualization. It can run batched jobs, as well as provide a desktop environment for interactive work.