What's new?
Access
There is a new direct entry point to NT via nt.swin.edu.au. The existing login to the OzSTAR machine at ozstar.swin.edu.au remains accessible.
Hardware
Ngarrgu Tindebeek has 11,648 CPU cores and 88 GPUs in total.
160 standard compute nodes:
2x 32-core AMD EPYC 7543 CPUs
256 GB RAM
2 TB NVMe SSD
10 high-memory compute nodes:
1024 GB RAM
(other specs same as standard compute nodes)
22 GPU compute nodes:
512 or 1024 GB RAM
4x NVIDIA A100 80 GB GPUs
(other specs same as standard compute nodes)
Nodes are connected via NVIDIA Quantum-2 NDR InfiniBand switches with a data throughput of 200 Gbps.
Projects and filesystem
Ngarrgu Tindebeek and the existing OzSTAR cluster shares the same project management system and filesystem. The existing Lustre file system (/fred) will be upgraded to 19 PB, which is accessible on OzSTAR and Ngarrgu Tindebeek.
Limitations
Maximum wall time for a job is 7 days.
Maximum CPUs per project is 2500.
Maximum GPUs per project is 100.
Queues
As with with the OzSTAR cluster, you do not need to specify a Slurm queue. This is determined automatically by the node you are submitting from. Jobs submitted via tooarrana1/2 will be sent to the new nodes.
Modules
Ngarrgu Tindebeek continues to use software modules, but contains a new software hierarchy and a fresh installation of the latest applications. See Environment Modules for more information.
Environment
The default number of OpenMP threads is now set to 1 using the environment variable
OMP_NUM_THREADS=1(previously not defined by default). Users must explicitly change this value if they wish to use more than 1 thread.