Linux-basierter Parallelrechner / HPC-Cluster Alex (Megware)

Model: GPU-Cluster 2022

Manufacturer: Megware (2022)

URL: https://hpc.fau.de/systems-services/systems-documentation-instructions/clusters/alex-cluster/

Location: Erlangen

Usage: For external users too

DFG Key: 7000 Datenverarbeitungsanlagen, zentrale Rechenanlagen

Description

FAU’s Alex cluster (system integrator: Megware) is a high-performance compute resource with Nvidia GPGPU accelerators and partially high speed interconnect. It is intended for single and multi GPGPU workloads, e.g. from molecular dynamics, or machine learning. Alex serves for both, FAU’s basic Tier3 resources as well as NHR’s project resources.

  • 2 front end nodes, each with two AMD EPYC 7713 “Milan” processors (64 cores per chip) running at 2.0 GHz with 256 MB Shared L3 cache per chip, 512 GB of RAM, and 100 GbE connection to RRZE’s network backbone but no GPGPUs.
  • 8 GPGPU nodes, each with two AMD EPYC 7662 “Rome” processors (64 cores per chip) running at 2.0 GHz with 256 MB Shared L3 cache per chip, 512 GB of DDR4-RAM, four Nvidia A100 (each 40 GB HBM2 @ 1,555 GB/s; DGX board with NVLink; 9.7 TFlop/s in FP64 or 19.5 TFlop/s in FP32), one HDR200 Infiniband HCAs, 25 GbE, and 6 TB on local NVMe SSDs. (During the year 2021 and early 2022, these nodes have previously been part of TinyGPU.)
  • 20 GPGPU nodes, each with two AMD EPYC 7713 “Milan” processors (64 cores per chip) running at 2.0 GHz with 256 MB Shared L3 cache per chip, 1,024 GB of DDR4-RAM, eight Nvidia A100 (each 40 GB HBM2 @ 1,555 GB/s; HGX board with NVLink; 9.7 TFlop/s in FP64 or 19.5 TFlop/s in FP32), two HDR200 Infiniband HCAs, 25 GbE, and 14 TB on local NVMe SSDs.
  • 12 GPGPU nodes, each with two AMD EPYC 7713 “Milan” processors (64 cores per chip) running at 2.0 GHz with 256 MB Shared L3 cache per chip, 2,048 GB of DDR4-RAM, eight Nvidia A100 (each 80 GB HBM2 @ 1,555 GB/s; HGX board with NVLink; 9.7 TFlop/s in FP64 or 19.5 TFlop/s in FP32), two HDR200 Infiniband HCAs, 25 GbE, and 14 TB on local NVMe SSDs.
  • 38 GPGPU nodes, each with two AMD EPYC 7713 “Milan” processors (64 cores per chip) running at 2.0 GHz with 256 MB Shared L3Cache per chip, 512 GB of DDR4-RAM, eight Nvidia A40 (each with 48 GB DDR6 @ 696 GB/s; 37.42 TFlop/s in FP32), 25 GbE, and 7 TB on local NVMe SSDs.

Involved Person(s)

Organisation(s)

Research Areas

Funding Sources