COMSOL

Revision as of 05:55, 14 December 2020 by Roberpj (talk | contribs)
Other languages:

Introduction

COMSOL is a general-purpose platform software for modeling engineering applications. Compute Canada would like to thank COMSOL, Inc. for allowing its software to be hosted on Compute Canada clusters via a special agreement.

We recommend that Compute Canada users who wish to run COMSOL consult the documentation included with the software under File > Help > Documentation prior to attempting to use it on one of the clusters. Some of the basic manuals are available here.

Licensing

Compute Canada is a hosting provider for COMSOL. This means that we have COMSOL software installed on our clusters, but we do not provide a generic license accessible to everyone. Many institutions, faculties, and departments already have licenses that can be used on our clusters. Alternatively researchers can purchase a license from CMC for use anywhere in Canada, or purchase a dedicated Floating Network License directly from the company for use on Compute Canada systems to run on a Sharcnet license server.

Once the legal aspects are worked out for licensing, there will be remaining technical aspects. The license server on your end will need to be reachable by our compute nodes. This will require our technical team to get in touch with the technical people managing your license software. In some cases such as CMC, this has already been done. You should then be able to load the COMSOL modules, and it should find its license automatically. If this is not the case, please contact our Technical support, so that we can arrange this for you.

Configuring your own license file

Our module for COMSOL is designed to look for license information in a few places. One of those places is your home folder. If you have your own license server, you can write the information to access it in the following format:

File : comsol.lic

SERVER <server> ANY <port>
USE_SERVER


and put this file in the folder $HOME/.licenses/. Here <server> is your license server and <port> is the port number of the license server. Note that firewall changes will need to be done on both our side and your side. To arrange this, send an email containing the service port and ip address of your floating comsol license server to Technical support.

CMC License Setup

Researchers who purchase a comsol license subscription from CMC may use the following settings in their comsol.lic file:

  • Béluga: SERVER 132.219.136.89 ANY 6601
  • Graham: SERVER 199.241.162.97 ANY 6601
  • Cedar: SERVER 172.16.121.25 ANY 6601

If initial license checkout attempts fail contact <cmcsupport@cmc.ca> to verify they have your Compute Canada username on file.

Submit jobs

Single Compute Node

Sample submission script to run a COMSOL job with eight cores on a single cluster compute node:

File : mysub1.sh

#!/bin/bash
#SBATCH --time=0-03:00            # Specify (d-hh:mm)
#SBATCH --account=def-group       # Specify (some account)
#SBATCH --mem=3G                  # Specify (set to 0 to use all available node memory when using all cores)
#SBATCH --cpus-per-task=8         # Specify (set to 32or44 on graham, 32or48 on cedar, or 40 on beluga to use all cores)
#SBATCH --nodes=1                 # Do not change
#SBATCH --ntasks-per-node=1       # Do not change

FILENAME="MyModel.mph"            # Specify (inputfile filename)

module load StdEnv/2016.4
module load comsol/5.5

comsol batch -inputfile $FILENAME -outputfile solved_out.mph -np $SLURM_CPUS_ON_NODE


Note that depending on the complexity of the simulation, COMSOL may not be able to efficiently use very many cores. Therefore please test the scaling of your simulation by increasing the number of cores in #SBATCH --cpus-per-task=X from X=1 to the maximum number of cores on the compute node you are using. If you still get a good speed-up running on a full compute node then consider running the job over multiple full nodes by adjusting the following submission script.

Multiple Compute Nodes

Sample submission script to run a COMSOL job with eight cores distributed evenly over two cluster compute nodes. Ideal for very large simulations (that exceed the capabilities of a single compute node) this script supports restarting interupted jobs, allocation of large temporary files to scratch and utilizing the default comsolbatch.ini file settings. There is also an option to modify tha java heap memory as described below the script.

File : mysub2.sh

#!/bin/bash
#SBATCH --time=0-03:00            # Specify (d-hh:mm)
#SBATCH --account=def-account     # Specify (some account)
#SBATCH --mem=3G                  # Specify (set to 0 to use all available node memory when using all cores)
#SBATCH --cpus-per-task=4         # Specify (set to 32or44 on graham, 32or48 on cedar, or 40 on beluga to use all cores)
#SBATCH --nodes=2                 # Specify (the number of compute nodes)
#SBATCH --ntasks-per-node=1       # Do not change

FILENAME="MyModel.mph"            # Specify (inputfile filename)

module load StdEnv/2020
module load comsol/5.6

SCRTMP=/scratch/$USER/comsol/tmpdir
SCRREC=/scratch/$USER/comsol/recoverydir
mkdir -p $SCRTMP $SCRREC

cp -f ${EBROOTCOMSOL}/bin/glnxa64/comsolbatch.ini comsolbatch.ini
cp -f ${EBROOTCOMSOL}/mli/startup/java.opts java.opts

comsol batch -inputfile $FILENAME -outputfile solved_out.mph -recover \
 -nn $SLURM_NTASKS -nnhost $SLURM_NTASKS_PER_NODE -np $SLURM_CPUS_PER_TASK \
 -recoverydir $SCRREC -tmpdir $SCRTMP -comsolinifile comsolbatch.ini -alivetime 15


Add the following sed lines to the above slurm script (immediately after the cp commands) to increase the maximum java heap memory. See https://www.comsol.ch/support/knowledgebase/1243 for further information ...

sed -i 's/-Xmx2g/-Xmx4g/g' comsolbatch.ini
sed -i 's/-Xmx768m/-Xmx2g/g' java.opts

Graphical use

Comsol can be run interactively in full graphical mode using either of the following methods.

On a cluster

Suitable to interactively run computationally intensive test jobs requiring upto all cores or memory on a single cluster node.

  1. Connect to a compute node (3hr time limit) with TigerVNC
  2. module load StdEnv/2016.4
  3. module load comsol/5.5     (COMSOL Multiphysics 5.5 Build:292)
  4. comsol

On gra-vdi

Suitable to create or modify simulation input files, post process or visualize data, interactively run a single test job with upto 8cores.

  1. Connect to gra-vdi (no time limit) with TigerVNC
  2. module load CcEnv StdEnv/2016.4
  3. module load comsol/5.5     (COMSOL Multiphysics 5.5 Build:292)
  4. comsol

Users with complex visualization models may benefit from hardware level graphics acceleration.
This capability may only be obtained by using one of the local modules installed on gra-vdi.
If a specific version is required submit a problem ticket directed to the Sharcnet support team.

  1. Connect to gra-vdi (no time limit) with TigerVNC
  2. module load SnEnv
  3. module load comsol/5.3.1.348     (COMSOL Multiphysics 5.3a Build:348)
  4. comsol

Parameter Sweeps

Batch Sweep

When working interactively in the Comsol GUI parametric problems maybe solved using the Batch Sweep approach. Multiple parameter sweeps maybe carried out as shown in this Comsol video. Speedup due to Task Parallism maybe also be realized.

Cluster Sweep

To run a parameter sweep on a cluster, a job must be submitted to the schedular from the command line using "sbatch slurmscript". For a detailed discussion regarding additional required comsol arguments see a and b for details. Support for submitting parametric simulations to the cluster queue from the Comsol graphical interface using a Cluster Sweep node is not available at this time.