Niagara/fr: Difference between revisions

Jump to navigation Jump to search
Updating to match new version of source page
(Updating to match new version of source page)
(Updating to match new version of source page)
Line 39: Line 39:


* 1500 nodes, each with 40 Intel Skylake cores at 2.4GHz, for a total of 60,000 cores.
* 1500 nodes, each with 40 Intel Skylake cores at 2.4GHz, for a total of 60,000 cores.
* 192 GB of RAM per node.
* 202 GB (188 GiB) of RAM per node.
* EDR Infiniband network in a so-called 'Dragonfly+' topology.
* EDR Infiniband network in a so-called 'Dragonfly+' topology.
* 5PB of scratch, 5+2PB of project space (parallel file system: IBM Spectrum Scale, formerly known as GPFS).
* 5PB of scratch, 5+2PB of project space (parallel file system: IBM Spectrum Scale, formerly known as GPFS).
Line 90: Line 90:
* Computational perfomance: 3 TFlops (theoretical maximum)
* Computational perfomance: 3 TFlops (theoretical maximum)
* Network connection: 100Gb/s EDR  
* Network connection: 100Gb/s EDR  
* Memory: 192 GB of RAM, i.e., a bit over 4GB per core.
* Memory: 202 GB (188 GiB) GB of RAM, i.e., a bit over 4GiB per core.
* Local disk: none.
* Local disk: none.
* Operating system: Linux CentOS 7
* Operating system: Linux CentOS 7
Line 99: Line 99:


* Scheduling will be by node only. This means jobs will always need to use multiples of 40 cores per job.
* Scheduling will be by node only. This means jobs will always need to use multiples of 40 cores per job.
* Asking for specific amounts of memory will not be necessary and is discouraged; all nodes have the same amount of memory (192GB minus some operating system overhead).
* Asking for specific amounts of memory will not be necessary and is discouraged; all nodes have the same amount of memory (202GB/188GiB minus some operating system overhead).


Details, such as how to request burst buffer usage in jobs, are still being worked out.
Details, such as how to request burst buffer usage in jobs, are still being worked out.
38,760

edits

Navigation menu