.. _UGentT2 hardware: ######################### HPC-UGent Tier-2 Clusters ######################### The Stevin computing infrastructure consists of several Tier2 clusters which are hosted in the S10 datacenter of Ghent University. This infrastructure is co-financed by FWO and Department of Economy, Science and Innovation (EWI). Login nodes =========== Log in to the HPC-UGent Tier-2 infrastructure via https://login.hpc.ugent.be or using SSH via ``login.hpc.ugent.be``. Compute clusters ================ For most recent information about the available resources and cluster status, please consult https://www.ugent.be/hpc/en/infrastructure . CPU clusters ------------ The HPC-UGent Tier-2 infrastructure currently included several standard CPU-only clusters, of different generations (listed from old to new). For basic information on using these clusters, see our `documentation `__. =============== ========== =========================================================== ===================== =========================== ======================= =================== cluster name # nodes Processor architecture Usable memory/node Local diskspace/node Interconnect Operating system =============== ========== =========================================================== ===================== =========================== ======================= =================== doduo (default) 128 2 x 48-core AMD EPYC 7552 (Rome @ 2.2 GHz) 250 GiB 180 GB SSD HDR-100 InfiniBand RHEL 9 gallade 16 2 x 64-core AMD EPYC 7773X (Milan-X @ 2.2 GHz) 940 GiB 1.5 TB NVME HDR-100 InfiniBand RHEL 9 shinx 48 2 x 96-core AMD EPYC 9654 (Genoa @ 2.4 GHz) 370 GiB 500GB NVME NDR-200 InfiniBand RHEL 9 =============== ========== =========================================================== ===================== =========================== ======================= =================== Interactive debug cluster ------------------------- A special-purpose interactive debug cluster is available, where you should always be able to get a job running quickly, without waiting in the queue. Intended usage is mainly for interactive work, either via an interactive job or using the `HPC-UGent web portal `__. This cluster is heavily over-provisioned, so jobs may run slower if the cluster is used more heavily. Strict limits are in place per user: * max. 5 jobs in queue * max. 3 jobs running * max. of 8 cores and 27GB of memory in total for running jobs For more information, see our `documentation `__. =============== ========== =========================================================== ===================== =========================== ======================= =================== cluster name # nodes Processor architecture Usable memory/node Local diskspace/node Interconnect Operating system =============== ========== =========================================================== ===================== =========================== ======================= =================== donphan 16 2 x 18-core Intel Xeon Gold 6240 (Cascade Lake @ 2.6 GHz) 738 GiB 1.6 TB NVME HDR-100 InfiniBand RHEL 9 + 1x shared NVIDIA Ampere A2 GPU (16GB GPU memory) =============== ========== =========================================================== ===================== =========================== ======================= =================== GPU clusters ------------ GPU clusters are available in the HPC-UGent Tier-2 infrastructure, with different generations of NVIDIA GPUs. These are well suited for specific workloads, with software that can leverage the GPU resources (like TensorFlow, PyTorch, GROMACS, AlphaFold, etc.). =============== ========== =========================================================== ===================== =========================== ======================= =================== cluster name # nodes Processor architecture Usable memory/node Local diskspace/node Interconnect Operating system =============== ========== =========================================================== ===================== =========================== ======================= =================== joltik 10 2 x 16-core Intel Xeon Gold 6242 (Cascade Lake @ 2.8 GHz) 256 GiB 800 GB SSD double EDR Infiniband RHEL 9 + 4x NVIDIA Volta V100 GPUs (32GB GPU memory) accelgor 9 2 x 24-core AMD EPYC 7413 (Milan @ 2.2 GHz) 500 GiB 180GB SSD HDR-100 InfiniBand RHEL 9 + 4x NVIDIA Ampere A100 GPUs (80GB GPU memory) litleo 8 1x 48 core AMD EPYC 9454P (Genoa @ 2.75 GHz) 315 GiB 1.4TB SSD NDR-200 Infiniband RHEL 9 + 2x NVIDIA H100 NVL (96GB GPU memory) =============== ========== =========================================================== ===================== =========================== ======================= =================== .. _UGent storage: Shared storage ============== .. include:: storage_quota_table.rst For more information, see our HPC-UGent tutorial: https://www.ugent.be/hpc/en/support/documentation.htm . User documentation ================== Please consult https://www.ugent.be/hpc/en/support/documentation.htm . In case of questions or problems, don't hesitate to contact the HPC-UGent support team via hpc@ugent.be, see also https://www.ugent.be/hpc/en/support .