DiRAC: Tursa (GPU)

Tursa is an Extreme Scaling GPU based DiRAC system.

DiRAC Tursa 2021

EPCC hosts the Extreme Scaling component of the DiRAC facility. The Extreme Scaling GPU based system, Tursa, is a 120-node Atos Sequana XH2000 system housed at EPCC's Advanced Computing Facility. EPCC also provides service support for the entire DiRAC consortium.

Technology

# Nodes
120: 114 GPU nodes, 6 CPU nodes

CPUs
GPU Nodes: 2x AMD EPYC Rome 7272, 2.9GHz, 12-core per node
CPU Nodes: 2x AMD EPYC ROME 7H12, 2.66GHz, 64-core per node

Total CPU cores
3,504

GPUs
4x Nvidia RedStone A100-40 per node, 640 Tensor cores, 6,912 CUDA cores

Total GPU cores
291,840 Tensor cores, 3,151,872 CUDA cores

System Memory details
GPU Nodes: 1,024 GB per node
CPU Nodes: 256 GB per node
Total of 117TB of RAM

Storage technologies and specs
4PB of DDN Lustre backed up to a Spectra tape library

Interconnect technologies and specs
Mellanox 200 Gbit/s HDR Infiniband Interconnect with a fat tree topology

Layout/Physical system scale
Tursa is composed of 5 Atos Sequana XH2000 racks, two management and storage racks and a tape library.

Cooling tech and specs
The Tursa compute nodes are stored in water cooled Atos Sequana XH2000 racks.

Scheduler details
Bull Slurm

System OS Details
Red Hat Enterprise Linux 8

Science and applications

DiRAC is recognised as the primary provider of HPC resources to the STFC Particle Physics, Astroparticle physics, Astrophysics, Cosmology, Solar System and Planetary Science and Nuclear physics (PPAN: STFC Frontier Science) theory community. It provides the modelling, simulation, data analysis and data storage capability that underpins the STFC Science Challenges and our researcher’s world-leading science outcomes.

The STFC Science Challenges are three fundamental questions in frontier physics:

  1. How did the Universe begin and how is it evolving?
  2. How do stars and planetary systems develop and how do they support the existence of life?
  3. What are the basic constituents of matter and how do they interact?

Access

Academic access

Access to DiRAC is coordinated by The STFC’s DiRAC Resource Allocation Committee, which puts out an annual Call for Proposals to request time as well as a Director’s Discretionary Call.

Details of access to DiRAC can be found on the DiRAC website.

Commercial access

DiRAC has a long track record of collaborating with Industry on bleeding-edge technology and we are recognised as a global pioneer of scientific software and computing hardware co-design. 

We specialise in the design, deployment, management and utilisation of HPC for simulation and large-scale data analytics. We work closely with our industrial partners on the challenges of data intensive science, machine learning and artificial intelligence that are increasingly important in the modern world.

Details of industrial access to DiRAC can be found on the DiRAC website.

Trial access

If you are a researcher wishing to try the DiRAC resources, get a feel for HPC, test codes, benchmark or see what the DiRAC resources can do for you before making a full application for resources, an application can be made for seedcorn time.

Details of seedcorn access to DiRAC can be found on the DiRAC website.

People

The system is managed by DiRAC and is maintained by EPCC and the hardware provider Atos.

Support

The DiRAC helpdesk is the first point of contact for all questions relating to the DiRAC Extreme Scaling services. Support is available Monday to Friday from 08:00 until 18:00 UK time, excluding UK public holidays.

Resources