Bureaucrats, cc_docs_admin, cc_staff
2,309
edits
No edit summary |
No edit summary |
||
Line 13: | Line 13: | ||
All of our clusters use Slurm for job submission, so many parts of a job submission script will work across clusters. However, you should note that the number of CPU cores per node may vary significantly across clusters, from 24 up to 64 cores, so check the page of the cluster you are using to verify how many cores can be used on a node. The amount of memory also varies so you may need to adapt your script to account for this as well, in addition to differences among GPUs that are available, if any. | All of our clusters use Slurm for job submission, so many parts of a job submission script will work across clusters. However, you should note that the number of CPU cores per node may vary significantly across clusters, from 24 up to 64 cores, so check the page of the cluster you are using to verify how many cores can be used on a node. The amount of memory also varies so you may need to adapt your script to account for this as well, in addition to differences among GPUs that are available, if any. | ||
On [[Cedar]], you may not submit jobs from your home directory and the compute nodes have direct Internet access; on [[Graham]], [[Béluga/en]] and [[Narval/en]], the compute nodes do not have Internet access. The maximum job duration is seven days on Béluga and Narval but 28 days on Cedar and Graham. All of the clusters except Cedar also restrict the number of jobs per user, both running and queued, to be no more than 1000. | On [[Cedar]], you may not submit jobs from your home directory and the compute nodes have direct Internet access; on [[Graham]], [[Béluga/en|Béluga]] and [[Narval/en|Narval]], the compute nodes do not have Internet access. The maximum job duration is seven days on Béluga and Narval but 28 days on Cedar and Graham. All of the clusters except Cedar also restrict the number of jobs per user, both running and queued, to be no more than 1000. |