Vtune

From Alliance Doc
Jump to navigation Jump to search
This site replaces the former Compute Canada documentation site, and is now being managed by the Digital Research Alliance of Canada.

Ce site remplace l'ancien site de documentation de Calcul Canada et est maintenant géré par l'Alliance de recherche numérique du Canada.

Other languages:


Introduction

VTune is Intel's Performance Analysis tool for applications and systems. It is capable of Analyzing both OpenMP and MPI based applications.

Software module

To load the module on any Alliance cluster run:

Question.png
[name@server ~]$ module load vtune

Tool renaming

The content of this page is largely concerned with the legacy version named Intel® VTune™ Amplifier. Please note this tool has been renamed throughout Intel's documentation in latest versions (newer than the latest vtune module versions presently available on Alliance clusters) from Intel® VTune™ Amplifier to Intel® VTune™ Profiler. Likewise the application commands amplxe-cl and amplxe-gui have been renamed to vtune and vtune-gui for both the command line and graphical tools respectively. Further information can be found here.

Analysis types

To collect analysis information run:

Question.png
[name@server ~]$ vtune -collect <analysis-type> <target_exe> <exe_arguments>

where <analysis-type> should be replaced by one of the available analysis, e.g. hotspots, and <target_exe> is the path to the executable you would like to analyze. It is recommended to compile your executable with the "-g" option and to use the same optimization level as normal so as to obtain accurate results. A listing of version specific argument options and several usage examples maybe displayed on the command line by running vtune -help, after loading the vtune module. Complete downloadable documentation for Parallel Studio XE (including VTune) for all recent versions can be found here. The latest version of the Intel VTune Profiler User Guide may be found here.

Create reports

To create a report run this command:

Question.png
[name@server ~]$ vtune -report <report-type>

where <report-type> is the type of the report to generate, e.g. hotspots. See also:

Matrix example

Analyze and generate a summary report for the Intel Matrix Sample Project run from the command line with 4 cores:

salloc --time=1:00:00 --cpus-per-task=4 --ntasks=1 --mem=16G --account=def-yours
module load StdEnv/2020 vtune
cp -a $EBROOTVTUNE/vtune/$EBVERSIONVTUNE*/samples/en/C++/matrix . cd matrix/linux
make icc
vtune -collect hotspots ../matrix
vtune -report summary 

The latest version of matrix_multiply (uses cmake to build) can be found here.

Graphical mode

The Intel Matrix Sample Project can also be run using Vtune in GUI mode as explored here [1]. To run VTune over VNC follow the below directions depending on which system you wish to use. Running VTune graphically can be useful to generate command line configurations as discussed in [2].

Cluster nodes

  1. Connect to a cluster compute or login node with TigerVNC
  2. module load StdEnv/2020 vtune
  3. vtune-gui

VDI nodes

  1. Connect to gra-vdi.alliancecan.ca with TigerVNC
  2. module load CcEnv StdEnv/2020 vtune
  3. vtune-gui


MPI example

First, load the latest VTune module.

module load StdEnv/2020
module load vtune

Then compile your MPI program as you usually would and run it inside a job or in an interactive session started by a salloc command using:

srun aps your_mpi_program.x

After the program finishes, the profiling data will be stored in a directory called aps_result_YYYYMMDD where YYYYMMDD is the current date.

There is a lot of information you can extract from that data. To get the basic summary report of your program's performance, run:

aps-report  -D aps_result_YYYYMMDD

where you would replace YYYYMMDD to match the actual directory that has been created. This command creates an HTML file, which can be copied to your own computer and viewed in a browser. The report will clearly identify performance issues that are affecting your code.

Example report generated by aps for an MPI program