JUPITER is the first European exascale supercomputer being installed for 2024 at Juelich. Technical details are here: JUPITER Technical Overview (fz-juelich.de).
System overview
The system will be built on BullSequana XH3000 and will have two partitions:
- Booster Module:
- “roughly 6,000 compute nodes” - BullSequana XH3515 nodes
- Uses GH200 superchips
- 4C:4G (four superchips per node)
- 24,000 H200 GPUs
- 4 NDR200 InfiniBand HCAs per node
- “1 ExaFLOP/s (FP64, HPL)” which seems feasible given Meta’s H100 clusters are 24K H100s and estimated to be just a few FLOPS shy of 1 EF each.
- Cluster Module: SiPearl Rhea CPUs
- “more than 1300 nodes”
- “more than 5 PetaFLOP/s (FP64, HPL)”
- 1 NDR200 InfiniBand HCA per node
The nodes were on display at SC24:
Network architecture
NDR200 InfiniBand in dragonfly+ will be used throughout:
- Each dragonfly+ group is a nonblocking fat tree
- 25x GPU groups
- 2x CPU, storage, and management groups
- 867x switches
- with “25400 end points”
- At least 24,000 for GPU nodes
- Leaving 1350 for CPU nodes?
Storage subsystem
Storage: IBM ESS 3500
- 40 IBM ESS 3500 servers
- 29 PB raw, 21 PB formatted
Facility
JUPITER is being installed using a modular data center. FRZ Jülich has posted some great photos of what that looks like.