SlideShare a Scribd company logo
PROMETHEUS
ENERGY EFFICIENT
SUPERCOMPUTING
Marek Magryś
ACC Cyfronet AGH-UST
established in 1973
part of AGH University of Science and Technology
in Krakow, Poland
member of PIONIER consortium
operator of Krakow MAN
centre of competence in HPC and Grid Computing
provides free computing resources for scientific
institutions
home for supercomputers
Prometheus 3
Liquid Cooling
Water: up to 1000x more efficient heat exchange
than air
Less energy needed to move the coolant
Hardware (CPUs, DIMMs) can handle ~80C
CPU/GPU vendors show TDPs up to 300W
Challenge: cool 100% of HW with liquid
network switches
PSUs
MTBF
The less movement the better
less pumps
less fans
less HDDs
Example
pump MTBF: 50 000 hrs
fan MTBF: 50 000 hrs
1800 node system MTBF: 7 hrs
Prometheus
HP Apollo 8000
16 m2, 20 racks (4 CDU, 16 compute)
2.4 PFLOPS
PUE <1.05, 800 kW peak power
2232 nodes:
2160 CPU nodes: 2x Intel Haswell E5-2680v3, 128 GB
RAM, IB FDR 56 Gb/s, Ethernet 1 Gb/s
72 GPU nodes: +2 NVIDIA Tesla K40d
53568 cores, up to 13824 per island
279 TB DDR4 RAM
CentOS 7 + SLURM
Prometheus storage
Diskless compute nodes
Separate project for storage
DDN SFA12kx hardware
Lustre-based
2 file systems:
Scratch: 120 GB/s, 5 PB usable space
Archive: 60 GB/s, 5 PB usable space
HSM-ready
NFS for $HOME and software
Why Apollo 8000?
Most energy efficient
The only solution with
100% warm water cooling
Highest density
Lowest TCO
Even more Apollo
Focuses also on ‘1’ in PUE!
Power distribution
Less fans
Detailed monitoring
‘energy to solution’
Dry node maintenance
Less cables
Prefabricated piping
Simplified management
Deployment timeline
Day 0 - Contract signed (20.10.2014)
Day 23 - Installation of the primary loop starts
Day 35 - First delivery (service island)
Day 56 - Apollo piping arrives
Day 98 - 1st and 2nd island delivered
Day 101 - 3rd island delivered
Day 111 - basic acceptance ends
Official launch event on 27.04.2015
Facility preparation
Primary loop installation took 5 weeks
Secondary (prefabricated) just 1 week
Upgrade of the raised floor done „just in case”
Additional pipes for leakage/condensation drain
Water dam with emergency drain
Lot of space needed for the hardware deliveries (over
100 pallets)
Facility monitoring
Secondary loop
Prometheus - node 14
HP XL730f/XL750f Gen9
• 2x Intel Xeon E5-2680 v3 (Haswell)
• 24 cores, 2100-3300 MHz
• 30 MB cache, 128 GB RAM DDR4
• Mellanox Connect-X3 IB FDR 56Gb/s
Prometheus - rack 15
HP Apollo 8000:
• 8 cells – 9 trays each – 18 CPU or 9 GPU nodes
• 8 IB FDR 36p 56 Gb/s switches
• Dry-disconnect and HEX water cooling
• HVDC 480V
HP Apollo 8000 CDU:
• Heat exchanger
• Vacuum pump
• Cooling controller
• IB FDR 36p 56 Gb/s (18+3) dist+core switches
Prometheus – compute island
Prometheus – IB network
service
nodes
Service island
I/O nodes
IB core network
576 CPU nodes
Compute island
576 CPU nodes
Compute island
576 CPU nodes
Compute island
432 CPU nodes
72 GPU nodes
Compute island
Over 250 Tb/s
aggregate throughput
• 30 km of cables
• 217 switches
• >10 000 ports
Monitoring 18
SLURM node states
IB network traffic
Monitoring of:
• CPU frequencies and temperatures
• Memory usage
• NFS and Lustre bandwith/IOPS/MDOPS
• Power and cooling
Linpack power draw
Linpack water temperatures
Top500 and Green500
3-rd level submision (Nov 2015)
#72, 2068 Mflops/W
#1 petascale x86 system in Europe
submission after expansion
(Nov 2015)
#38, 1,67 PFLOPS Rmax
GPUs not used for the run
Application & software
Academic workload
Lots of small/medium jobs
Few big jobs
330 projects
750 users
Main fields:
Chemistry
Biochemistry (farmaceuticals)
Astrophysics
22
thatmpi code
Institute of Nuclear Physics PAS in Krakow
Study of non-relativistic shock waves hosted by
supernova remnants that are believed to generate
most of the Galactic cosmic-rays
Thatmpi: Two-and-a-Half-Dimensional Astroparticle
Stanford code with MPI (PIC)
2.5-dimensional particle dynamics
fully relativistic with electro-magnetism
colliding plasma jets with perpendicular B-field
large simulations: up to 10k cores/run
Applications
thatmpi: left leptons jet animation
A. Dorobisz, M. Kotwica
thatmpi: joint development
low-level:
vectorization
flow control refactoring
register and cache-use optimization
high-level:
modernization from FORTRAN77 to Fortran2013
new particle sorting method
portable data dumping with HDF5
communication buffering
total time reduction: over 20%
>1.2 MWh less energy per run (350 nodes, 60h)
Future: 3D, domain partitioning, code refactoring
Lessons learned
There will be leaks!
Drycooler seems more simple than a chiller, but the
whole infrastructure is not
Sysadmins need to get a degree in plumbing
Traditional facilities people don’t understand HPC
SCADA systems are dumb and insecure
Monitor everyting, anytime, keep historic data
Keep data easy to correlate
Avoid SPOFs
Never settle for anything less than full load testing
Know your costs, calclulate TCO
Look at hardware, middleware and software
26
Thank you!

More Related Content

What's hot (9)

PDF
Provisioning Updates - Juno Edition
OpenStack Foundation
 
PDF
Towards Exascale Simulations of Stellar Explosions with FLASH
Ganesan Narayanasamy
 
PDF
SkyhookDM - Towards an Arrow-Native Storage System
JayjeetChakraborty
 
PDF
Quantifying Energy Consumption for Practical Fork-Join Parallelism on an Embe...
Tulipp. Eu
 
PDF
Simulating connected capacitor bank
Febrian Sasi Kirana
 
PPTX
Gas Turbine Nuclear Power Plants
Adam Doligalski
 
PPT
David Lovelace - Analysing, displaying and sharing historic landscapes from f...
Shaun Lewis
 
PDF
Push Technology's latest data distribution benchmark with Solarflare and Zing
Azul Systems Inc.
 
Provisioning Updates - Juno Edition
OpenStack Foundation
 
Towards Exascale Simulations of Stellar Explosions with FLASH
Ganesan Narayanasamy
 
SkyhookDM - Towards an Arrow-Native Storage System
JayjeetChakraborty
 
Quantifying Energy Consumption for Practical Fork-Join Parallelism on an Embe...
Tulipp. Eu
 
Simulating connected capacitor bank
Febrian Sasi Kirana
 
Gas Turbine Nuclear Power Plants
Adam Doligalski
 
David Lovelace - Analysing, displaying and sharing historic landscapes from f...
Shaun Lewis
 
Push Technology's latest data distribution benchmark with Solarflare and Zing
Azul Systems Inc.
 

Viewers also liked (20)

PDF
Model Simulation, Graphical Animation, and Omniscient Debugging with EcoreToo...
Benoit Combemale
 
PDF
Raspberry home server
Massimiliano Perrone
 
PDF
Latency tracing in distributed Java applications
Constantine Slisenka
 
PDF
Libnetwork updates
Moby Project
 
PDF
HPC DAY 2017 | Altair's PBS Pro: Your Gateway to HPC Computing
HPC DAY
 
PDF
HPC DAY 2017 | HPE Strategy And Portfolio for AI, BigData and HPC
HPC DAY
 
PDF
HPC DAY 2017 | Accelerating tomorrow's HPC and AI workflows with Intel Archit...
HPC DAY
 
PDF
HPC DAY 2017 | HPE Storage and Data Management for Big Data
HPC DAY
 
PDF
HPC DAY 2017 | NVIDIA Volta Architecture. Performance. Efficiency. Availability
HPC DAY
 
PPTX
Java on the GPU: Where are we now?
Dmitry Alexandrov
 
PDF
Database Security Threats - MariaDB Security Best Practices
MariaDB plc
 
PDF
LinuxKit and OpenOverlay
Moby Project
 
PDF
GPU databases - How to use them and what the future holds
Arnon Shimoni
 
PDF
Design patterns in Java - Monitis 2017
Arsen Gasparyan
 
PDF
Getting Started with Embedded Python: MicroPython and CircuitPython
Ayan Pahwa
 
PPTX
An Introduction to OMNeT++ 5.1
Alpen-Adria-Universität
 
PPT
Drive into calico architecture
Anirban Sen Chowdhary
 
PDF
세션1. block chain as a platform
Jay JH Park
 
PDF
Scylla Summit 2017: Repair, Backup, Restore: Last Thing Before You Go to Prod...
ScyllaDB
 
Model Simulation, Graphical Animation, and Omniscient Debugging with EcoreToo...
Benoit Combemale
 
Raspberry home server
Massimiliano Perrone
 
Latency tracing in distributed Java applications
Constantine Slisenka
 
Libnetwork updates
Moby Project
 
HPC DAY 2017 | Altair's PBS Pro: Your Gateway to HPC Computing
HPC DAY
 
HPC DAY 2017 | HPE Strategy And Portfolio for AI, BigData and HPC
HPC DAY
 
HPC DAY 2017 | Accelerating tomorrow's HPC and AI workflows with Intel Archit...
HPC DAY
 
HPC DAY 2017 | HPE Storage and Data Management for Big Data
HPC DAY
 
HPC DAY 2017 | NVIDIA Volta Architecture. Performance. Efficiency. Availability
HPC DAY
 
Java on the GPU: Where are we now?
Dmitry Alexandrov
 
Database Security Threats - MariaDB Security Best Practices
MariaDB plc
 
LinuxKit and OpenOverlay
Moby Project
 
GPU databases - How to use them and what the future holds
Arnon Shimoni
 
Design patterns in Java - Monitis 2017
Arsen Gasparyan
 
Getting Started with Embedded Python: MicroPython and CircuitPython
Ayan Pahwa
 
An Introduction to OMNeT++ 5.1
Alpen-Adria-Universität
 
Drive into calico architecture
Anirban Sen Chowdhary
 
세션1. block chain as a platform
Jay JH Park
 
Scylla Summit 2017: Repair, Backup, Restore: Last Thing Before You Go to Prod...
ScyllaDB
 
Ad

Similar to HPC DAY 2017 | Prometheus - energy efficient supercomputing (20)

PDF
European Exascale System Interconnect & Storage
inside-BigData.com
 
PPTX
¿Es posible construir el Airbus de la Supercomputación en Europa?
AMETIC
 
PDF
Architecting a 35 PB distributed parallel file system for science
Speck&Tech
 
PDF
From the Archives: Future of Supercomputing at Altparty 2009
Olli-Pekka Lehto
 
PPT
Necesidades de supercomputacion en las empresas españolas
Cein
 
PPT
NWU and HPC
Wilhelm van Belkum
 
PDF
HPC the new normal
Eurotech
 
PDF
HPC, the new normal‎: the Personal Computer is dead. Long live the Personal ...
Roberto Siagri
 
PDF
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
PDF
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
PDF
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
PDF
HiPEAC 2019 Tutorial - Maestro RTOS
Tulipp. Eu
 
PPT
Super Computer15 updated
shashthoughts
 
PDF
IBM and ASTRON 64-Bit Microserver Prototype Prepares for Big Bang's Big Data,...
IBM Research
 
PDF
Barcelona Supercomputing Center, Generador de Riqueza
Facultad de Informática UCM
 
PDF
Achitecture Aware Algorithms and Software for Peta and Exascale
inside-BigData.com
 
PDF
#VirtualDesignMaster 3 Challenge 2 - Harshvardhan Gupta
vdmchallenge
 
ODP
Systems Support for Many Task Computing
Eric Van Hensbergen
 
PPTX
HPC Top 5 Stories: January 12, 2018
NVIDIA
 
PPTX
Supercomputer
Systronic IT Group PTY LTD
 
European Exascale System Interconnect & Storage
inside-BigData.com
 
¿Es posible construir el Airbus de la Supercomputación en Europa?
AMETIC
 
Architecting a 35 PB distributed parallel file system for science
Speck&Tech
 
From the Archives: Future of Supercomputing at Altparty 2009
Olli-Pekka Lehto
 
Necesidades de supercomputacion en las empresas españolas
Cein
 
NWU and HPC
Wilhelm van Belkum
 
HPC the new normal
Eurotech
 
HPC, the new normal‎: the Personal Computer is dead. Long live the Personal ...
Roberto Siagri
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Larry Smarr
 
HiPEAC 2019 Tutorial - Maestro RTOS
Tulipp. Eu
 
Super Computer15 updated
shashthoughts
 
IBM and ASTRON 64-Bit Microserver Prototype Prepares for Big Bang's Big Data,...
IBM Research
 
Barcelona Supercomputing Center, Generador de Riqueza
Facultad de Informática UCM
 
Achitecture Aware Algorithms and Software for Peta and Exascale
inside-BigData.com
 
#VirtualDesignMaster 3 Challenge 2 - Harshvardhan Gupta
vdmchallenge
 
Systems Support for Many Task Computing
Eric Van Hensbergen
 
HPC Top 5 Stories: January 12, 2018
NVIDIA
 
Ad

Recently uploaded (20)

PDF
Smart Trailers 2025 Update with History and Overview
Paul Menig
 
PDF
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
PDF
Presentation - Vibe Coding The Future of Tech
yanuarsinggih1
 
PDF
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
PDF
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
PPTX
Q2 FY26 Tableau User Group Leader Quarterly Call
lward7
 
PPTX
From Sci-Fi to Reality: Exploring AI Evolution
Svetlana Meissner
 
PDF
Exolore The Essential AI Tools in 2025.pdf
Srinivasan M
 
PDF
Building Real-Time Digital Twins with IBM Maximo & ArcGIS Indoors
Safe Software
 
PDF
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
PDF
Transcript: New from BookNet Canada for 2025: BNC BiblioShare - Tech Forum 2025
BookNet Canada
 
PDF
Blockchain Transactions Explained For Everyone
CIFDAQ
 
PDF
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
PDF
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
PPTX
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
PDF
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
PDF
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
PDF
Newgen 2022-Forrester Newgen TEI_13 05 2022-The-Total-Economic-Impact-Newgen-...
darshakparmar
 
PPTX
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
PDF
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 
Smart Trailers 2025 Update with History and Overview
Paul Menig
 
Empower Inclusion Through Accessible Java Applications
Ana-Maria Mihalceanu
 
Presentation - Vibe Coding The Future of Tech
yanuarsinggih1
 
NewMind AI - Journal 100 Insights After The 100th Issue
NewMind AI
 
SWEBOK Guide and Software Services Engineering Education
Hironori Washizaki
 
Q2 FY26 Tableau User Group Leader Quarterly Call
lward7
 
From Sci-Fi to Reality: Exploring AI Evolution
Svetlana Meissner
 
Exolore The Essential AI Tools in 2025.pdf
Srinivasan M
 
Building Real-Time Digital Twins with IBM Maximo & ArcGIS Indoors
Safe Software
 
Chris Elwell Woburn, MA - Passionate About IT Innovation
Chris Elwell Woburn, MA
 
Transcript: New from BookNet Canada for 2025: BNC BiblioShare - Tech Forum 2025
BookNet Canada
 
Blockchain Transactions Explained For Everyone
CIFDAQ
 
DevBcn - Building 10x Organizations Using Modern Productivity Metrics
Justin Reock
 
The Builder’s Playbook - 2025 State of AI Report.pdf
jeroen339954
 
Webinar: Introduction to LF Energy EVerest
DanBrown980551
 
Using FME to Develop Self-Service CAD Applications for a Major UK Police Force
Safe Software
 
CIFDAQ Market Insights for July 7th 2025
CIFDAQ
 
Newgen 2022-Forrester Newgen TEI_13 05 2022-The-Total-Economic-Impact-Newgen-...
darshakparmar
 
"Autonomy of LLM Agents: Current State and Future Prospects", Oles` Petriv
Fwdays
 
Complete JavaScript Notes: From Basics to Advanced Concepts.pdf
haydendavispro
 

HPC DAY 2017 | Prometheus - energy efficient supercomputing

  • 2. ACC Cyfronet AGH-UST established in 1973 part of AGH University of Science and Technology in Krakow, Poland member of PIONIER consortium operator of Krakow MAN centre of competence in HPC and Grid Computing provides free computing resources for scientific institutions home for supercomputers
  • 4. Liquid Cooling Water: up to 1000x more efficient heat exchange than air Less energy needed to move the coolant Hardware (CPUs, DIMMs) can handle ~80C CPU/GPU vendors show TDPs up to 300W Challenge: cool 100% of HW with liquid network switches PSUs
  • 5. MTBF The less movement the better less pumps less fans less HDDs Example pump MTBF: 50 000 hrs fan MTBF: 50 000 hrs 1800 node system MTBF: 7 hrs
  • 6. Prometheus HP Apollo 8000 16 m2, 20 racks (4 CDU, 16 compute) 2.4 PFLOPS PUE <1.05, 800 kW peak power 2232 nodes: 2160 CPU nodes: 2x Intel Haswell E5-2680v3, 128 GB RAM, IB FDR 56 Gb/s, Ethernet 1 Gb/s 72 GPU nodes: +2 NVIDIA Tesla K40d 53568 cores, up to 13824 per island 279 TB DDR4 RAM CentOS 7 + SLURM
  • 7. Prometheus storage Diskless compute nodes Separate project for storage DDN SFA12kx hardware Lustre-based 2 file systems: Scratch: 120 GB/s, 5 PB usable space Archive: 60 GB/s, 5 PB usable space HSM-ready NFS for $HOME and software
  • 8. Why Apollo 8000? Most energy efficient The only solution with 100% warm water cooling Highest density Lowest TCO
  • 9. Even more Apollo Focuses also on ‘1’ in PUE! Power distribution Less fans Detailed monitoring ‘energy to solution’ Dry node maintenance Less cables Prefabricated piping Simplified management
  • 10. Deployment timeline Day 0 - Contract signed (20.10.2014) Day 23 - Installation of the primary loop starts Day 35 - First delivery (service island) Day 56 - Apollo piping arrives Day 98 - 1st and 2nd island delivered Day 101 - 3rd island delivered Day 111 - basic acceptance ends Official launch event on 27.04.2015
  • 11. Facility preparation Primary loop installation took 5 weeks Secondary (prefabricated) just 1 week Upgrade of the raised floor done „just in case” Additional pipes for leakage/condensation drain Water dam with emergency drain Lot of space needed for the hardware deliveries (over 100 pallets)
  • 14. Prometheus - node 14 HP XL730f/XL750f Gen9 • 2x Intel Xeon E5-2680 v3 (Haswell) • 24 cores, 2100-3300 MHz • 30 MB cache, 128 GB RAM DDR4 • Mellanox Connect-X3 IB FDR 56Gb/s
  • 15. Prometheus - rack 15 HP Apollo 8000: • 8 cells – 9 trays each – 18 CPU or 9 GPU nodes • 8 IB FDR 36p 56 Gb/s switches • Dry-disconnect and HEX water cooling • HVDC 480V HP Apollo 8000 CDU: • Heat exchanger • Vacuum pump • Cooling controller • IB FDR 36p 56 Gb/s (18+3) dist+core switches
  • 17. Prometheus – IB network service nodes Service island I/O nodes IB core network 576 CPU nodes Compute island 576 CPU nodes Compute island 576 CPU nodes Compute island 432 CPU nodes 72 GPU nodes Compute island Over 250 Tb/s aggregate throughput • 30 km of cables • 217 switches • >10 000 ports
  • 18. Monitoring 18 SLURM node states IB network traffic Monitoring of: • CPU frequencies and temperatures • Memory usage • NFS and Lustre bandwith/IOPS/MDOPS • Power and cooling
  • 21. Top500 and Green500 3-rd level submision (Nov 2015) #72, 2068 Mflops/W #1 petascale x86 system in Europe submission after expansion (Nov 2015) #38, 1,67 PFLOPS Rmax GPUs not used for the run
  • 22. Application & software Academic workload Lots of small/medium jobs Few big jobs 330 projects 750 users Main fields: Chemistry Biochemistry (farmaceuticals) Astrophysics 22
  • 23. thatmpi code Institute of Nuclear Physics PAS in Krakow Study of non-relativistic shock waves hosted by supernova remnants that are believed to generate most of the Galactic cosmic-rays Thatmpi: Two-and-a-Half-Dimensional Astroparticle Stanford code with MPI (PIC) 2.5-dimensional particle dynamics fully relativistic with electro-magnetism colliding plasma jets with perpendicular B-field large simulations: up to 10k cores/run Applications
  • 24. thatmpi: left leptons jet animation A. Dorobisz, M. Kotwica
  • 25. thatmpi: joint development low-level: vectorization flow control refactoring register and cache-use optimization high-level: modernization from FORTRAN77 to Fortran2013 new particle sorting method portable data dumping with HDF5 communication buffering total time reduction: over 20% >1.2 MWh less energy per run (350 nodes, 60h) Future: 3D, domain partitioning, code refactoring
  • 26. Lessons learned There will be leaks! Drycooler seems more simple than a chiller, but the whole infrastructure is not Sysadmins need to get a degree in plumbing Traditional facilities people don’t understand HPC SCADA systems are dumb and insecure Monitor everyting, anytime, keep historic data Keep data easy to correlate Avoid SPOFs Never settle for anything less than full load testing Know your costs, calclulate TCO Look at hardware, middleware and software 26