Development/Vampir and VampirServer: Difference between revisions
(First edition of Vampir and VampirServer) |
(Update description on running remote VampirServer) |
||
Line 26: | Line 26: | ||
<br> |
<br> |
||
= Introduction = |
= Introduction = |
||
Vampir and VampirServer are performance analysis tools developed Technical University of Dresden. With support from the Ministerium für Wissenschaft, Forschung und Kunst (MWK), |
Vampir and VampirServer are performance analysis tools developed at the Technical University of Dresden. With support from the Ministerium für Wissenschaft, Forschung und Kunst (MWK), |
||
all Universities participating in bwHPC (see [[bwUniCluster_2.0]]) have acquired a five year license. |
all Universities participating in bwHPC (see [[bwUniCluster_2.0]]) have acquired a five year license. |
||
<br> |
<br> |
||
= Versions and Availability = |
= Versions and Availability = |
||
A list of versions currently available on all bwHPC |
A list of versions currently available on all bwHPC Clusters can be obtained from the |
||
<br> |
<br> |
||
<big> |
<big> |
||
Line 43: | Line 43: | ||
On the command line please check for availability using <kbd>module avail devel/vampir</kbd>. |
On the command line please check for availability using <kbd>module avail devel/vampir</kbd>. |
||
Vampir provides the GUI and allows |
Vampir provides the GUI and allows analyzing traces of a few hundred Megabytes. For larger traces, you may want |
||
to revert to using a remote VampirServer |
to revert to using a remote VampirServer running in parallel on the compute nodes via a Batch script (see below). |
||
Application traces consist of information gathered on the clusters prior to running Vampir |
Application traces consist of information gathered on the clusters prior to running Vampir |
||
Line 66: | Line 66: | ||
For '''online documentation''' see the links section in the summary table at the top of this page. The local installation |
For '''online documentation''' see the links section in the summary table at the top of this page. The local installation |
||
provides Manuals in the <kbd>$VAMPIR_DOC_DIR</kbd> directory. |
provides Manuals in the <kbd>$VAMPIR_DOC_DIR</kbd> directory. |
||
Running Vampir GUI and VampirServer is possible in various way as highlighted in the following images: |
|||
<br> |
<br> |
||
= Running remote VampirServer = |
= Running remote VampirServer = |
||
The installation provides in <kbd>$VAMPIR_HOME</kbd> a SLURM batch script with which You may run parallel VampirServer instance on the compute nodes. |
|||
You may attach to your VampirServer node using the provided port (typically port 30000, please check in the SLURM output file, once started). |
|||
The SLURM script only supplies the queue name (default <kbd>multiple</kbd>); if You expect Your analysis to run for only 30 minutes or less, you |
|||
may want to use another Queue meant for short-running development purposes <kbd>dev_multiple</kbd> which on bwUniCluster allows specifying the maximum time: |
|||
<pre> |
|||
sbatch --partition=dev_multiple --time=30 $VAMPIR_HOME/vampirserver.slurm |
|||
</pre> |
|||
Meanwhile, you may want to start another job using the default <kbd>multiple</kbd> queue; so that it will be scheduled, once your first job runs out. |
|||
Please query using <kbd>squeue</kbd> on the current status of both jobs and check the relevant SLURM output files. |
|||
<br> |
<br> |
||
Revision as of 10:55, 5 November 2020
Description | Content |
---|---|
module load | devel/vampir |
Availability | bwUniCluster | BwForCluster_Chemistry |
License | [Vampir Professional License] |
Citing | n/a |
Links | Homepage | Tutorial | Use case |
Graphical Interface | Yes |
Included modules |
Introduction
Vampir and VampirServer are performance analysis tools developed at the Technical University of Dresden. With support from the Ministerium für Wissenschaft, Forschung und Kunst (MWK),
all Universities participating in bwHPC (see bwUniCluster_2.0) have acquired a five year license.
Versions and Availability
A list of versions currently available on all bwHPC Clusters can be obtained from the
Cluster Information System CIS
{{#widget:Iframe
| url=https://cis-hpc.uni-konstanz.de/prod.cis/bwUniCluster/devel/vampir
| width=99%
| height=120
}}
On the command line please check for availability using module avail devel/vampir. Vampir provides the GUI and allows analyzing traces of a few hundred Megabytes. For larger traces, you may want to revert to using a remote VampirServer running in parallel on the compute nodes via a Batch script (see below).
Application traces consist of information gathered on the clusters prior to running Vampir using VampirTrace or Score-P and include timing, MPI communication, MPI I/O, hardware performance counters and CUDA / OpenCL profiling (if enabled in the tracing library).
$ : bwUniCluster 2.0 $ module avail devel/vampir ------------------------ /opt/bwhpc/common/modulefiles/Core ------------------------- devel/vampir/9.9
Attention!
Do not run vampir on the head nodes with large traces with many analysis processes for a long period of time.
Please use on of the possibilities listed below.
Tutorial
For online documentation see the links section in the summary table at the top of this page. The local installation provides Manuals in the $VAMPIR_DOC_DIR directory.
Running Vampir GUI and VampirServer is possible in various way as highlighted in the following images:
Running remote VampirServer
The installation provides in $VAMPIR_HOME a SLURM batch script with which You may run parallel VampirServer instance on the compute nodes. You may attach to your VampirServer node using the provided port (typically port 30000, please check in the SLURM output file, once started). The SLURM script only supplies the queue name (default multiple); if You expect Your analysis to run for only 30 minutes or less, you may want to use another Queue meant for short-running development purposes dev_multiple which on bwUniCluster allows specifying the maximum time:
sbatch --partition=dev_multiple --time=30 $VAMPIR_HOME/vampirserver.slurm
Meanwhile, you may want to start another job using the default multiple queue; so that it will be scheduled, once your first job runs out.
Please query using squeue on the current status of both jobs and check the relevant SLURM output files.
VampirServer commands
If You want to run VampirServer as part of Your job-script, e.g. after finalizing Your application's run, add the following to your Batch script:
module load devel/vampir vampirserver start mpi
This shell scripts starts the MPI-parallel version of the VampirServer in the existing, already running SLURM job. The results of starting VampirServer is stored in $HOME/.vampir/server/list; you may check using the below commands, or by checking this file directly.
15 1604424211 mpi 20 uc2n001.localdomain 30000 2178460
Where the first column is the server number (incremented), the third column is the parallelisation mode VAMPIR_MODE, the next column is the number of tasks, followed by the name of the node (uc2n001) and the port (30000).
The commands available to the vampirserver shell script are
Command | Description |
---|---|
help | Show this help |
config | Interactively configure VampirServer for the given host system. |
list | List running servers including hostname and port (see file $HOME/.vampir/server/list). |
start [-t NUM] [LAUNCHER] | Starts a new VampirServer, using -t number of seconds with LAUNCHER being either smp (default), mpi and ap (Cray/HPE only) |
stop [SERVER_ID] | Stops the specified server again |
version | Print VampirServer's revision |