Using GPUs with Slurm: Difference between revisions

Jump to navigation Jump to search
no edit summary
No edit summary
No edit summary
Line 81: Line 81:
<!--T:40-->
<!--T:40-->
If you do not supply a type specifier, Slurm may send your job to a node equipped with any type of GPU.   
If you do not supply a type specifier, Slurm may send your job to a node equipped with any type of GPU.   
For certain workflows this may be undesirable.
For certain workflows this may be undesirable; for example, molecular dynamics code requires high double-precision performance, for which T4 GPUs are not appropriate.
For example, molecular dynamics code requires high double-precision performance, for which T4 GPUs are not appropriate.
In such a case, make sure you include a type specifier.
In such a case, make sure you include a type specifier.


Line 178: Line 177:


<!--T:10-->
<!--T:10-->
There is a special group of GPU nodes on [[Cedar]] which have four Tesla P100 16GB cards each. (Other P100 GPUs on the cluster have 12GB and the V100 GPUs have 32G.) The GPUs in a P100L node all use the same PCI switch, so the inter-GPU communication latency is lower, but bandwidth between CPU and GPU is lower than on the regular GPU nodes. The nodes also have 256GB RAM. You may only request these nodes as whole nodes, therefore you must specify <code>--gres=gpu:p100l:4</code>. P100L GPU jobs up to 28 days can be run on Cedar.
There is a special group of GPU nodes on [[Cedar]] which have four Tesla P100 16GB cards each (Other P100 GPUs on the cluster have 12GB and the V100 GPUs have 32G). The GPUs in a P100L node all use the same PCI switch, so the inter-GPU communication latency is lower, but bandwidth between CPU and GPU is lower than on the regular GPU nodes. The nodes also have 256GB RAM. You may only request these nodes as whole nodes, therefore you must specify <code>--gres=gpu:p100l:4</code>. P100L GPU jobs up to 28 days can be run on Cedar.


<!--T:11-->
<!--T:11-->
rsnt_translations
56,420

edits

Navigation menu