cc_staff
1,894
edits
No edit summary |
|||
Line 566: | Line 566: | ||
= Site-specific use = <!--T:20850--> | = Site-specific use = <!--T:20850--> | ||
== SHARCNET license == | == SHARCNET license == | ||
SHARCNET provides a small but free license consisting of 2 cae and 35 execute tokens where usage limits are imposed 10 tokens/user and 15 tokens/group. For groups that have purchased dedicated tokens, the free token usage limits are added to their reservation. The free tokens are available on a first come first serve basis and mainly intended for testing and light usage before deciding whether or not to purchase dedicated tokens. | SHARCNET provides a small but free license consisting of 2 cae and 35 execute tokens where usage limits are imposed 10 tokens/user and 15 tokens/group. For groups that have purchased dedicated tokens, the free token usage limits are added to their reservation. The free tokens are available on a first come first serve basis and mainly intended for testing and light usage before deciding whether or not to purchase dedicated tokens. Costs for dedicated tokens (in 2021) were approximately CAD$110 per compute token and CAD$400 per GUI token: submit a ticket to request an official quote. The license can be used by any Alliance researcher, but only on SHARCNET hardware. Groups that purchase dedicated tokens to run on the SHARCNET license server may likewise only use them on SHARCNET hardware including gra-vdi (for running Abaqus in full graphical mode) and Graham or Dusky clusters (for submitting compute batch jobs to the queue). Before you can use the license you must contact [[Technical support]] and request access. In your email 1) mention that it is for use on SHARCNET systems and 2) include a copy/paste of the following <code>License Agreement</code> statement with your full name and username entered in the indicated locations. Please note that every user must do this it cannot be done one time only for a group; this includes PIs who have purchased their own dedicated tokens. | ||
<!--T:20851--> | <!--T:20851--> | ||
=== License agreement === | |||
<pre>---------------------------------------------------------------------------------- | <pre>---------------------------------------------------------------------------------- | ||
Subject: Abaqus SHARCNET Academic License User Agreement | Subject: Abaqus SHARCNET Academic License User Agreement | ||
Line 589: | Line 589: | ||
<!--T:20854--> | <!--T:20854--> | ||
=== Configure license file === | |||
Configure your license file as follows, noting that it is only usable on SHARCNET systems: Graham, gra-vdi and Dusky. | Configure your license file as follows, noting that it is only usable on SHARCNET systems: Graham, gra-vdi and Dusky. | ||
</translate> | </translate> | ||
Line 601: | Line 601: | ||
<!--T:20855--> | <!--T:20855--> | ||
If your Abaqus jobs fail with the error message [*** ABAQUS/eliT_CheckLicense rank 0 terminated by signal 11 (Segmentation fault)] in the slurm output file, verify if your <code>abaqus.lic</code> file contains ABAQUSLM_LICENSE_FILE to use abaqus/2020. If your Abaqus jobs fail with an error message starting [License server machine is down or not responding, etc.] in the output file verify your <code>abaqus.lic</code> file contains LM_LICENSE_FILE to use abaqus/6.14.1 as shown. The <code>abaqus.lic</code> file shown contains both so you should not see this problem. | If your Abaqus jobs fail with the error message [*** ABAQUS/eliT_CheckLicense rank 0 terminated by signal 11 (Segmentation fault)] in the slurm output file, verify if your <code>abaqus.lic</code> file contains ABAQUSLM_LICENSE_FILE to use abaqus/2020. If your Abaqus jobs fail with an error message starting [License server machine is down or not responding, etc.] in the output file verify your <code>abaqus.lic</code> file contains LM_LICENSE_FILE to use abaqus/6.14.1 as shown. The <code>abaqus.lic</code> file shown contains both so you should not see this problem. | ||
< | |||
I) | <!--T:20861--> | ||
=== Query license server === | |||
Log into graham, load abaqus and then run one of the following: | |||
<source lang="bash"> | |||
ssh graham.alliancecan.ca | |||
module load StdEnv/2020 | |||
module load abaqus | |||
</source> | |||
I) Check the SHARCNET license server for started and queued jobs: | |||
</translate> | </translate> | ||
<source lang="bash"> | <source lang="bash"> | ||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | egrep "Users|start|queued" | |||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | | |||
</source> | </source> | ||
<translate> | <translate> | ||
<!--T:20856--> | <!--T:20856--> | ||
II) | II) Check the SHARCNET license server for started and queued jobs also showing reservations by purchasing groups: | ||
</translate> | </translate> | ||
<source lang="bash"> | <source lang="bash"> | ||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | egrep "Users|start|queued|RESERVATION" | |||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | | |||
</source> | </source> | ||
<translate> | <translate> | ||
<!--T:20857--> | <!--T:20857--> | ||
III) | III) Check the SHARCNET license server for only cae, standard and explicit product availability: | ||
</translate> | </translate> | ||
<source lang="bash"> | <source lang="bash"> | ||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | grep "Users of" | egrep "cae|standard|explicit" | |||
abaqus licensing lmstat -c $LM_LICENSE_FILE -a | grep "Users of" | | |||
</source> | </source> | ||
<translate> | <translate> | ||
Line 640: | Line 640: | ||
29802 roberpj def-roberpj scriptsp1.txt R 2:59:33 1 12 8G dus28 (None) | 29802 roberpj def-roberpj scriptsp1.txt R 2:59:33 1 12 8G dus28 (None) | ||
[roberpj@dus241:~] abaqus licensing lmstat -c $LM_LICENSE_FILE -a | | [roberpj@dus241:~] abaqus licensing lmstat -c $LM_LICENSE_FILE -a | egrep "Users|start|queued|RESERVATION" | ||
Users of abaqus: (Total of 78 licenses issued; Total of 71 licenses in use) | Users of abaqus: (Total of 78 licenses issued; Total of 71 licenses in use) | ||
roberpj dus47 /dev/tty (v62.2) (license3.sharcnet.ca/27050 275), start Thu 8/27 5:45, 14 licenses | roberpj dus47 /dev/tty (v62.2) (license3.sharcnet.ca/27050 275), start Thu 8/27 5:45, 14 licenses | ||
Line 647: | Line 647: | ||
<translate> | <translate> | ||
<!--T:20859--> | <!--T:20859--> | ||
=== Specify job resources === | |||
To ensure optimal usage of both your Abaqus tokens and our resources, it's important to carefully specify the required memory and ncpus in your Slurm script. The values can be determined by submitting a few short test jobs to the queue then checking their utilization. For <b>completed</b> jobs use <code>seff JobNumber</code> to show the total <i>Memory Utilized</i> and <i>Memory Efficiency</i>. If the <i>Memory Efficiency</i> is less than ~90%, decrease the value of the <code>#SBATCH --mem=</code> setting in your Slurm script accordingly. Notice that the <code>seff JobNumber</code> command also shows the total <i>CPU (time) Utilized</i> and <i>CPU Efficiency</i>. If the <i>CPU Efficiency</i> is less than ~90%, perform scaling tests to determine the optimal number of CPUs for optimal performance and then update the value of <code>#SBATCH --cpus-per-task=</code> in your Slurm script. For <b>running</b> jobs, use the <code>srun --jobid=29821580 --pty top -d 5 -u $USER</code> command to watch the %CPU, %MEM and RES for each Abaqus parent process on the compute node. The %CPU and %MEM columns display the percent usage relative to the total available on the node while the RES column shows the per process resident memory size (in human readable format for values over 1GB). Further information regarding how to [[Running jobs#Monitoring_jobs|monitor jobs]] is available on our documentation wiki | To ensure optimal usage of both your Abaqus tokens and our resources, it's important to carefully specify the required memory and ncpus in your Slurm script. The values can be determined by submitting a few short test jobs to the queue then checking their utilization. For <b>completed</b> jobs use <code>seff JobNumber</code> to show the total <i>Memory Utilized</i> and <i>Memory Efficiency</i>. If the <i>Memory Efficiency</i> is less than ~90%, decrease the value of the <code>#SBATCH --mem=</code> setting in your Slurm script accordingly. Notice that the <code>seff JobNumber</code> command also shows the total <i>CPU (time) Utilized</i> and <i>CPU Efficiency</i>. If the <i>CPU Efficiency</i> is less than ~90%, perform scaling tests to determine the optimal number of CPUs for optimal performance and then update the value of <code>#SBATCH --cpus-per-task=</code> in your Slurm script. For <b>running</b> jobs, use the <code>srun --jobid=29821580 --pty top -d 5 -u $USER</code> command to watch the %CPU, %MEM and RES for each Abaqus parent process on the compute node. The %CPU and %MEM columns display the percent usage relative to the total available on the node while the RES column shows the per process resident memory size (in human readable format for values over 1GB). Further information regarding how to [[Running jobs#Monitoring_jobs|monitor jobs]] is available on our documentation wiki | ||
<!--T:20860--> | <!--T:20860--> | ||
=== Core token mapping === | |||
</translate> | </translate> | ||
<pre> | <pre> |