High-Performance Computing¶
Introduction¶
High-Performance Computing (HPC) utilises several fast network-connected servers to perform computationally intensive calculations or process large data sets. Disciplines such as Physics, Medical-Physics, Chemistry, and Engineering benefit from HPC. However, due to the processing requirements of large data sets, such as genome sequencing in fields such as Genetics, Biochemistry, Microbiology, Virology, and Pathology, an increase in HPC utilisation has been observed in the past decade.
Implementing Artificial Intelligence and Machine Learning also demanded fields such as Social Sciences and Humanities to use HPC more frequently. HPC is also often used to automate or streamline scientific workflows or pipelines of data processing.
Using an HPC has several benefits, but the most renowned benefits are saving time, performing several tasks in parallel and processing datasets that were otherwise impossible to process on an ordinary computer or even in a physical laboratory.
Watch the video below to see how an HPC is used in various scientific fields.
These examples are a fraction of the potential use of an HPC and reflect a portion of research projects that utilised the UFS HPC in recent years.
Hardware Spesifications¶
Compute Nodes¶
To summarise the components listed in the table below:
-
There are eight compute nodes in the HPC.
-
There are two main different types of nodes. The first four nodes each have eight NVIDIATM RTX 4000 GPUs, and the subsequent four nodes each have four NVIDIATM RTX 5000 GPUs.
-
Each node has two AMDTM EPYC 64 Core CPUs with 256 CPU Core Threads per node.
- Each node has two terabytes (2 TB) of ECC registered system memory (RAM).
The following table shows the hardware specifications of the compute nodes in the HPC.
Host | CPU Model | GPUs | Total Threads | Total RAM |
---|---|---|---|---|
cn0601 | 2 x AMD EPYC 7773X 1 | 8 x NVIDIA RTX A4000 | 256 | 2 TB |
cn0602 | 2 x AMD EPYC 7773X 1 | 8 x NVIDIA RTX A4000 | 256 | 2 TB |
cn0603 | 2 x AMD EPYC 7773X 1 | 8 x NVIDIA RTX A4000 | 256 | 2 TB |
cn0604 | 2 x AMD EPYC 7773X 1 | 8 x NVIDIA RTX A4000 | 256 | 2 TB |
cn0605 | 2 x AMD EPYC 7773X 1 | 4 x NVIDIA RTX A5000 | 256 | 2 TB |
cn0606 | 2 x AMD EPYC 7773X 1 | 4 x NVIDIA RTX A5000 | 256 | 2 TB |
cn0607 | 2 x AMD EPYC 7773X 1 | 4 x NVIDIA RTX A5000 | 256 | 2 TB |
cn0608 | 2 x AMD EPYC 7773X 1 | 4 x NVIDIA RTX A5000 | 256 | 2 TB |
Total (8 Nodes) | 16 CPUS | 40 GPUs | 2 048 Threads | 16 TB RAM |
Host | System Model | CPU Model | Cores per CPU | Threads per CPU | GPUs | RAM | Total Threads | Total RAM |
---|---|---|---|---|---|---|---|---|
cn0601 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 8 x NVIDIA RTX A4000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0602 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 8 x NVIDIA RTX A4000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0603 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 8 x NVIDIA RTX A4000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0604 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 8 x NVIDIA RTX A4000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0605 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 4 x NVIDIA RTX A5000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0606 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 4 x NVIDIA RTX A5000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0607 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 4 x NVIDIA RTX A5000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
cn0608 | 4124GS-TNR | 2 x AMD EPYC 7773X 1 | 64 | 128 | 4 x NVIDIA RTX A5000 | 32 x DDR4 3200 MHz (64 GB) | 256 | 2 TB |
Total | 8 Nodes | 16 CPUs | 512 Cores | 2 x 1 024 Threads | 40 GPUs | 256 x 64 GB Modules | 2 048 Threads | 16 TB RAM |
Components not contained in the above mentioned table but of some importance:
- 2 x 800 GB SSD 6 GB/s hard drives per node in RAID 0 configuration.
- 2 x 3.84 TB NVMe hard drives used for local scratch space.
- 2 x Mellanox MCX512A-ACAT SFP28 25 GbE Network interfaces.
- 4 x Redundant 2 000 Watt Titanium Level power supplies per node.
GPU Nodes
Currently, all compute nodes contain GPUs; therefore, no specific GPU nodes are available.
Footnotes: