Translations:Using GPUs with Slurm/2/en
Jump to navigation
Jump to search
Cluster | # of nodes | Slurm type specifier |
Per node | GPU model | Compute Capability(*) |
GPU mem (GiB) |
Notes | ||
---|---|---|---|---|---|---|---|---|---|
CPU cores | CPU memory | GPUs | |||||||
Béluga | 172 | v100 | 40 | 191000M | 4 | V100-16gb | 70 | 16 | All GPUs associated with the same CPU socket, connected via NVLink and SXM2 |
Cedar | 114 | p100 | 24 | 128000M | 4 | P100-12gb | 60 | 12 | Two GPUs per CPU socket, connected via PCIe |
32 | p100l | 24 | 257000M | 4 | P100-16gb | 60 | 16 | All GPUs associated with the same CPU socket, connected via PCIe | |
192 | v100l | 32 | 192000M | 4 | V100-32gb | 70 | 32 | Two GPUs per CPU socket; all GPUs connected via NVLink and SXM2 | |
Graham | 160 | p100 | 32 | 127518M | 2 | P100-12gb | 60 | 12 | One GPU per CPU socket, connected via PCIe |
7 | v100(**) | 28 | 183105M | 8 | V100-16gb | 70 | 16 | See Graham: Volta GPU nodes | |
2 | v100(***) | 28 | 183105M | 8 | V100-32gb | 70 | 32 | See Graham: Volta GPU nodes | |
30 | t4 | 44 | 192000M | 4 | T4-16gb | 75 | 16 | Two GPUs per CPU socket | |
6 | t4 | 16 | 192000M | 4 | T4-16gb | 75 | 16 | ||
Mist | 54 | (none) | 32 | 256GiB | 4 | V100-32gb | 70 | 32 | See Mist specifications |
Narval | 159 | a100 | 48 | 510000M | 4 | A100-40gb | 80 | 40 | Two GPUs per CPU socket; all GPUs connected via NVLink |
Arbutus | Cloud resources are not schedulable via Slurm. See Cloud resources for details of available hardware. |