HPC Installation at ZIH

ZIH offers several HPC components tailored towards different usage scenarios

  • Data-intensive and compute-intensive HPC applications
  • HPC Data Analytics and Machine Learning
  • Processing of extremely large data sets

Access is free of charge for academic researchers from TU Dresden, Saxony, and all over Germany as well as their project partners. HPC project proposals and resource allocations are subject to a scientific review.

Taurus © Robert Gommlich

Computing center and HPC components

LZR © Robert Gommlich

LZR computing center

The LZR computing center at TU Dresden hosts the main campus IT services as well as the HPC installations. It combines state-of-the-art infrastructure with a highly energy efficient operation and cooling concept including warm water cooling an energy reuse for heating nearby campus buildings.

It was inaugurated in 2015 and was awarded the 1st place of Deutscher Rechenzentrumspreis 2014 (German Award for Computing Centers) in the category for energy- and resource-efficient computing centers.


The High Performance Computing and Storage Complex (HRSK-II) by Bull/Atos provides the major part of the computing capacity available at ZIH, especially for highly parallel, data-intensive and compute-intensive HPC applications.

Typical applications: FEM simulations, CFD simulations with Ansys or OpenFOAM, molecular dynamics with GROMACS or NAMD, computations with Matlab or R

  • over 40 000 CPU cores, mostly Intel Haswell and Broadwell
  • 2.6 to 36 GB main memory per core
Taurus HPC cluster © Robert Gommlich

HPC Data Analytics and Machine Learning partition

The HPC-DA partition combines different hardware components and allows to combine them into tailored research environments. For machine learning and deep learning it offers 192 powerful Nvidia V100 GPUs. For data analytics it has a CPU partition with high memory bandwidth. To efficiently access large active data sets 2 petabytes of NVME memory with a nominal bandwidth of 2 Terabytes/s are available.

Typical applications: training of neural nets with Tensorflow, data analytics with Big Data frameworks such as Apache Spark

  • 32 IBM Power 9 nodes with 6 Nvidia V100 GPUs each for machine learning
  • 192 AMD Rome nodes, each with 128 cores, 512 GB RAM
  • 2 PB fast NVME storage
  • 10 PB warm archive storage
Data analytics and Machine Learning partition © Robert Gommlich
© Robert Gommlich

HPE SuperdomeFlex

The shared memory system HPE Superdome Flex is especially well suited for data intensive application scenarios, for example to process extremely large data sets completely in main memory or in very fast NVMe memory.

Typical applications: applications that require a very large shared memory, such as genome analysis

  • Shared memory system with 32 Intel CascadeLake CPUs, 896 cores in total
  • 48 TB main memory in a shared address space
  • 400 TB NVMe cards as very fast local storage

More Information

Please send us a message if you have any questions!

Contact: Andreas Knüpfer, Ulf Markwardt

Find further information about

or contact us with questions or project proposals.