Running Calculations: Difference between revisions

From bwHPC Wiki
Jump to navigation Jump to search
mNo edit summary
Line 20: Line 20:
** [[Helix/Slurm | Slurm Helix]]
** [[Helix/Slurm | Slurm Helix]]
* Moab systems (legacy systems with deprecated queuing system)
* Moab systems (legacy systems with deprecated queuing system)
** [[NEMO/Moab|Moab NEMO specific information]]
** [[NEMO/Moab|Moab NEMO]]
** [[BinAC/Moab|Moab BinAC specific information]]
** [[BinAC/Moab|Moab BinAC]]


{|style="background:#deffee; width:100%;"
{|style="background:#deffee; width:100%;"
Line 38: Line 38:
How to find a reasonable number of how many compute cores to use for your calculation can be found under '''[[Scaling]]'''
How to find a reasonable number of how many compute cores to use for your calculation can be found under '''[[Scaling]]'''


Information regarding the supported parallel programming paradigms and specific hints on their usage are summarized at '''[[Parallel_Programming]]'''
Information regarding the supported parallel programming paradigms and specific hints on their usage are summarized at '''[[Parallel Programming]]'''


Running calculations on an HPC node consumes a lot of energy. To make the most of the available resources and keep cluster and energy use as efficient as possible please also see our advice for '''[[Energy Efficient Cluster Usage]]
Running calculations on an HPC node consumes a lot of energy. To make the most of the available resources and keep cluster and energy use as efficient as possible please also see our advice for '''[[Energy Efficient Cluster Usage]]

Revision as of 16:48, 8 January 2024

Description

Running calculations on cluster.svg

On your desktop computer, you start your calculations and they start immediately, run until they are finished, then your desktop does mostly nothing, until you start another calculation. A compute cluster has several hundred, maybe a thousand computers (compute nodes), all of them are busy most of the time and many people want to run a great number of calculations. So running your job has to include some extra steps:

  1. prepare a script (a set commands to run - usually as a shell script), with all the commands that are necessary to run your calculation from start to finish. In addition to the commands necessary to run the calculation, this batch script has a header section, in which you specify details like required compute cores (processing units witin a computer), estimated runtime, memory requirements, disk space needed, etc.
  2. Submit the script into a queue, where your job (calculation)
  3. is queued and waits in row with other compute jobs until the resources you requested in the header become available.
  4. Execution: Once your job reaches the front of the queue, your script is executed on a compute node. Your calculation runs on that node until it is finished or reaches the specified time limit.
  5. Save results: At the end of your script, include commands to save the calculation results back to your home directory.

There are two types of batch systems currently used on bwHPC clusters, called "Moab" (legacy installs) and "Slurm".

Link to Batch System per Cluster

Because of differences in configuration (partly due to different available hardware), each cluster has their own batch system documention:

Attention.svg

Scientific software installed on the bwHPC Clusters often comes with simple example jobs (job script and input files). See Software Modules on how to load examples.

How to Use Computing Ressources Efficiently

When you are running your calculations, you will have to decide on how many compute-cores your calculation will be simultaneously calculated. For this, your computational problem will have to be divided into pieces, which always causes some overhead.

How to find a reasonable number of how many compute cores to use for your calculation can be found under Scaling

Information regarding the supported parallel programming paradigms and specific hints on their usage are summarized at Parallel Programming

Running calculations on an HPC node consumes a lot of energy. To make the most of the available resources and keep cluster and energy use as efficient as possible please also see our advice for Energy Efficient Cluster Usage