What's a job?
The requirement for
--account=project-name-cpu is still under discussion and may change before release.
- According to Kamil it is advisable for users to express their job in 1000s on MB instead of GB (i.e. 1000MB ~ 1GB). This will leave some RAM for the OS when reaching the memory limit on core nodes. e.g.
--ntasks=32 --mem-per-cpu=4000Mwill fit on a base node with 128G of RAM, while
--ntasks=32 --mem-per-cpu=4Grequires a Large-memory node.
- At least on Graham, one needs to use
--mem=xxxGto request a large-memory node and with using
--mem-per-cpu=yyyGone can only ever get base-nodes.
This slide deck from LLNL is old (2004), covers some architectural aspects probably not of interest to most users, and has some LLNL-local information that will not apply at CC (e.g. FIFO scheduling). Not recommended.
All the SchedMD videos I've looked at so far are just slide talks, mostly directed at administrators rather than users. Job submission commands come up in Introduction to SLURM, Part 3.
Doug P also suggests https://sites.google.com/a/case.edu/hpc-upgraded-cluster/slurm-cluster-commands from CWRU, specifically on moving from Torque to SLURM.
Discussion about the ordering of example job scripts
Would it make sense to put the array job information first in the list examples? It's likely one of the first techniques a novice HPC user would use to break up an embarrassing parallel computation where you might just break up the input data. It is not common in Bioinformatics and Genomics software stack to find any examples of MPI and even threaded applications are less common. If this segment of our user base is not that high, then the current order makes the most sense.