Spartan is High Performance Computing (HPC) system operated by Research Platform Services (ResPlat) at The University of Melbourne. It combines a high performance bare-metal compute with flexible cloud infrastructure and GPGPU to suit a wide range of use-cases.
If your computing jobs take too long on your desktop computer, or are simply not possible due to a lack of speed and memory, a HPC system like Spartan can help.
Maintenance window announcement (8-9/July)
The next regular Spartan maintenance window will be 9am July 8 until 5pm July 9. During this time Spartan will be offline, and over the next month, maximum job walltimes will be adjusted to ensure no jobs will run during the maintenance window.
- During the maintenance window, work will include:
- migrating home directories from CephFS onto an Enterprise NFS server
- upgrading Linux kernel and security patching
- migrating the final parts of the Spartan networking to the new Research network
Spartan Daily Weather Report (20190614)
- CephFS usage: 939.02TB Free: 461.70TB (67%)
- Spartan is very busy on the cloud partition, with close to 100% node allocation.
- Spartan is very busy on the physical partition, with close to 100% node allocation.
- Spartan is not busy on the GPGPU partition, with 81% node allocation.
- GPU usage in the [ gpgpu ] partition: 148 / 232 cards in use (63.79%)
- Some nodes out (26), primarily due to not needed or moving to Thespian.
We run regular one-day courses on HPC, shell scripting, parallel programming, and GPU programming. ResPlat also offer training in a wide range of other digital tools to accelerate your research.
Signup here: http://melbourne.resbaz.edu.au/participate
If you can't find an answer here, need advice, or otherwise stuck, you can contact our support team at email@example.com
Please submit one topic per ticket. If you require a assistance with a separate matter, compose a new ticket. Do not reply to existing or closed tickets.
Spartan has a number of partitions available for general usage. A full list of partitions can be viewed with the command
|cloud||165||12||100GB||Intel(R) Xeon(R) Gold 6138 CPU @ 2.00GHz|
|longcloud||2||12||100GB||Intel(R) Xeon(R) Gold 6138 CPU @ 2.00GHz||Max walltime of 90 days|
|physical||19||12||254GB||Intel(R) Xeon(R) CPU E5-2643 v3 @ 3.40GHz||Group = physg1|
|6||32||508GB||Intel(R) Xeon(R) CPU E5-2683 v4 @ 2.10GHz||Group = physg3|
|12||72||1540GB||Intel(R) Xeon(R) Gold 6154 CPU @ 3.00GHz||Group = physg4|
|bigmem||2||36||1540GB||Intel(R) Xeon(R) CPU E5-2697 v4 @ 2.30GHz||Group = physg2|
|phi||4||256||190GB||Intel(R) Xeon Phi(TM) CPU 7230 @ 1.30GHz||Based on the Xeon Phi Knights Landing architecture|
|snowy||31||32||127GB||Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz|
|mig||12||32||127GB||Intel(R) Xeon(R) CPU E5-2698 v3 @ 2.30GHz||Reserved for Melbourne Integrative Genomics users|
|gpgpu||73||24||127GB||Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz||4 P100 Nvidia GPUs per node|
This partition is best suited for general-purpose single-node jobs. Multiple node jobs will work, but communication between nodes will be comparatively slow.
Each node is connected by high-speed 25Gb networking with 1.15 Âµsec latency, making this partition suited to multi-node jobs (e.g. those using OpenMPI).
You can constrain your jobs to use different groups of nodes (e.g. just the Intel(R) Xeon(R) Gold 6154 CPU @ 3.00GHz nodes) by adding
#SBATCH --constraint=physg4 to your submit script
See here for more details.
This partition is suited to memory-intensive single-node workloads.
There are also special partitions which are outside normal walltime constraints. In particular,
shortgpgpu should be used for quick test cases; the partitions have a maximum time constraint of one hour.
--time=00:59:00 to your cloud partition job, your job will run in a dedicated reservation, ensuring your job runs faster
Spartan uses a storage system called CephFS. CephFS is a highly scalable, parallel and robust filesystem.
The total Spartan storage is broken up into 3 areas:
NB: /home and /data/cephfs share the same pool of disks
If you use Spartan to obtain results for a publication, we'd appreciate if you'd cite our service, including the DOI below. This makes it easy for us demonstrate research impact, helping to secure ongoing funding for expansion and user support.
Lev Lafayette, Greg Sauter, Linh Vu, Bernard Meade, "Spartan Performance and Flexibility: An HPC-Cloud Chimera", OpenStack Summit, Barcelona, October 27, 2016. doi.org/10.4225/49/58ead90dceaaa
If you are using the LIEF GPGPU cluster for a publication, please include the following citation in the acknowledgements section of your paper:
This research was undertaken using the LIEF HPC-GPGPU Facility hosted at the University of Melbourne. This Facility was established with the assistance of LIEF Grant LE170100200.
Spartan is just one of many research IT resources offered by The University of Melbourne, or available from other institutions.
Nectar is a national initiative to provide cloud-based Infrastructure as a Service (IaaS) resources to researchers. It's based on OpenStack, and allows researchers on-demand access to computation instances, storage, and a variety of application platforms and Virtual Laboratories.
Spartan runs some of it's computation resources in the Nectar cloud.
Multi-modal Australian ScienceS Imaging and Visualisation Environment (MASSIVE)
MASSIVE is a HPC system at Monash University and the Australian Synchrotron which is optimized for imaging and visualization. It can run batched jobs, as well as provide a desktop environment for interactive work.