Advanced MPI scheduling: Difference between revisions
No edit summary |
|||
Line 11: | Line 11: | ||
command and how its numerous options constrain the placement of processes. | command and how its numerous options constrain the placement of processes. | ||
=== | === Controlling the distribution of processes === | ||
To come | |||
=== Hybrid jobs: MPI and OpenMP, or MPI and threads === | === Hybrid jobs: MPI and OpenMP, or MPI and threads === | ||
Line 38: | Line 18: | ||
=== MPI and GPUs === | === MPI and GPUs === | ||
To come | |||
=== Troubleshooting and performance monitoring === | |||
To come | To come | ||
Revision as of 15:31, 5 April 2017
This is not a complete article: This is a draft, a work in progress that is intended to be published into an article, which may or may not be ready for inclusion in the main wiki. It should not necessarily be considered factual or authoritative.
Most users should submit MPI or distributed memory parallel jobs as illustrated
at Running jobs: MPI job. Simply request a number of
processes with --ntasks
or -n
and trust the scheduler
to allocate those processes in a way that balances the efficiency of your job
with the overall efficiency of the cluster.
If you need more detailed control over how your job is allocated, then read on
to learn about SLURM's sbatch
command and how its numerous options constrain the placement of processes.
Controlling the distribution of processes[edit]
To come
Hybrid jobs: MPI and OpenMP, or MPI and threads[edit]
To come
MPI and GPUs[edit]
To come
Troubleshooting and performance monitoring[edit]
To come
Why srun instead of mpiexec or mpirun?[edit]
To come