Béluga/en: Difference between revisions

Jump to navigation Jump to search
no edit summary
(Updating to match new version of source page)
No edit summary
Line 56: Line 56:
A Mellanox Infiniband EDR (100 Gb/s) network connects together all the nodes of the cluster. A central switch of 324 ports links the cluster's island topology with a maximum blocking factor of 5:1. The storage servers are networked with a non-blocking connection. The architecture permits multiple parallel jobs with up to 640 cores (or more) thanks to a non-blocking network. For jobs requiring greater parallelism, the blocking factor is 5:1 but even for jobs executed across several islands, the interconnection is high-performance.
A Mellanox Infiniband EDR (100 Gb/s) network connects together all the nodes of the cluster. A central switch of 324 ports links the cluster's island topology with a maximum blocking factor of 5:1. The storage servers are networked with a non-blocking connection. The architecture permits multiple parallel jobs with up to 640 cores (or more) thanks to a non-blocking network. For jobs requiring greater parallelism, the blocking factor is 5:1 but even for jobs executed across several islands, the interconnection is high-performance.


<div class="mw-translate-fuzzy">
=Node Characteristics=
The large memory nodes (752 GB) as well as the GPU nodes have Turbo Boost activated; it is turned off for all of the other nodes.
 
{| class="wikitable sortable"
{| class="wikitable sortable"
! nodes !! cores !! available memory !! CPU !! storage !! GPU
! nodes !! cores !! available memory !! CPU !! storage !! GPU
Line 68: Line 70:
| 172 || 40 || 186G or 191000M ||2 x Intel Gold 6148 Skylake @ 2.4 GHz || 1 x NVMe SSD 1.6T || 4 x NVidia V100SXM2 (16G memory), connected via NVLink
| 172 || 40 || 186G or 191000M ||2 x Intel Gold 6148 Skylake @ 2.4 GHz || 1 x NVMe SSD 1.6T || 4 x NVidia V100SXM2 (16G memory), connected via NVLink
|}
|}
</div>
Bureaucrats, cc_docs_admin, cc_staff
2,306

edits

Navigation menu