JUSTUS2/Software/Dalton

From bwHPC Wiki
< JUSTUS2‎ | Software
Revision as of 11:37, 22 March 2021 by F Wagner (talk | contribs) (Disk Usage)
Jump to: navigation, search
Description Content
module load chem/dalton
Availability BwForCluster_JUSTUS_2
License Open-source software, distributed under the GNU Lesser General Public License (LGPL). More...
Citing Publications
Links Homepage | Documentation
Graphical Interface No

1 Description

Dalton (named after John Dalton) is an ab initio quantum chemistry computer program designed to to allow convenient, automated determination of a large number of molecular properties based on an HF, HF-srDFT, DFT, MP2, CC, CI, MCSCF, or MC-srDFT reference wave function. For additional information on features please visit the Description of the Dalton suite features web page.

2 Availability

Dalton is available on selected bwHPC-Clusters. A complete list of versions currently installed on the bwHPC-Clusters can be obtained from the Cluster Information System (CIS).

In order to check which versions of Dalton are installed on the compute cluster, run the following command:

$ module avail chem/dalton

3 License

Dalton is free, open-source software released under the GNU Lesser General Public License (LGPL). Anyone interested in using the Dalton program suite must read the conditions described in its license agreement.

4 Usage

4.1 Loading the module

You can load the default version of Dalton with the following command:

$ module load chem/dalton

The module will try to load all modules it needs to function (e.g., compiler, mpi, ...). If loading the module fails, check if you have already loaded one of those modules, but not in the version required by Dalton.

If you wish to load another (older) version of Dalton, you can do so using

$ module load chem/dalton/<version>

with <version> specifying the desired version.

Please cite Dalton in your publications according to the references.

4.2 Program Binaries

The binary dalton is main program of the Dalton package.

Usage: dalton [options] dalinp{.dal} [molinp{.mol} [potinp{.pot}] [pcmsolver{.pcm}]]

Options:
 -w dir                : change job directory to dir (default: $HOME)
 -b dir                : prepend dir to directory list for basis set searches (job directory and dalton basis library
                         are included automatically)
 -o file               : redirect output from program to job directory with "file" as file name
 -ow                   : redirect output from program to job directory with standard file name
 -dal file             : the dalton input file
 -mol file             : the molecule input file
 -pot file             : the potential input file (for .QM3, .QMMM and .PELIB)
 -pcm file             : the pcm input file
 -ext log              : change default output extension from ".out" to ".log"
 -nobackup             : do not backup files, simply overwrite outputs
 -f dal_mol[_pot]      : extract dal_mol[_pot].tar.gz archive from WRKDIR into DALTON_TMPDIR before calculation
                         starts
 -noarch               : do not create tar.gz archive
 -t dir                : set scratch directory DALTON_TMPDIR; this script will append '/DALTON_scratch_<USERNAME>' to
                         the path unless the path contains 'DALTON_scratch' or you explicitly set -noappend
 -d                    : delete job scratch directory before calculation starts
 -D                    : do not delete job scratch directory after calculation stops
 -noappend             : do not append anything to the scratch directory; be careful with this option since by
                         default scratch is wiped after calculation
 -get "file1 ..."      : get files back from DALTON_TMPDIR after calculation stops
 -put "file1 ..."      : put files to DALTON_TMPDIR before calculation starts
 -omp num              : set the number of OpenMP threads. Note that Dalton is not OpenMP parallelized, however, this
                         option can be used with e.g. threaded blas as MKL
 -N num |-np num       : use num MPI processes (defaults to 1, illegal if DALTON_LAUNCHER specified)
 -cpexe                : copy dalton.x to DALTON_TMPDIR before execution, either to global scratch (if
                         DALTON_USE_GLOBAL_SCRATCH is set) or to local scratch on all nodes
 -rsh                  : use rsh/rcp for communication with MPI nodes (default: ssh/scp)
 -nodelist "node1 ..." : set nodelist DALTON_NODELIST, dalton.x will be copied to DALTON_TMPDIR on each node unless
                         DALTON_USE_GLOBAL_SCRATCH is defined (the script uses PBS_NODEFILE or SLURM_NODELIST if
                         available)
 -x dalmol1 dalmol2    : calculate NEXAFS spectrum from ground and core hole states
 -exe exec             : change the executable from default ($DALTON_HOME/dalton/dalton.x) to exec
 -pg                   : do profiling with gprof
 -gb mem               : set dalton max usable work memory to mem Gigabytes (mem integer)
 -mb mem               : set dalton max usable work memory to mem Megabytes (mem integer)
 -ngb mem              : set node max usable work memory to mem Gigabytes (mem integer)
 -nmb mem              : set node max usable work memory to mem Megabytes (mem integer)

4.3 Hints for using Dalton

4.3.1 Input Files

For information about how to construct input files (dalinp{.dal} [molinp{.mol} [potinp{.pot}]]) for Dalton, please consult the documentation.

4.3.2 Environment Variables

Environment variables understood by Dalton:

DALTON_TMPDIR             : scratch directory
DALTON_USE_GLOBAL_SCRATCH : use global scratch directory, do not copy any files to worker nodes
DALTON_NODELIST           : list of nodes, dalton.x will be copied to DALTON_TMPDIR on each node unless
                            DALTON_USE_GLOBAL_SCRATCH is defined
DALTON_LAUNCHER           : launcher for the dalton.x binary (if defined, -N flag not allowed)

4.3.3 Disk Usage

Scratch files are written to $SCRATCH by default. This configuration option can be changed by setting the environment variable DALTON_TMPDIR (e.g., to a dedicated workspace) before starting your calculations with Dalton.

5 Examples

As with all processes that require more than a few minutes to run, non-trivial compute jobs must be submitted to the cluster queuing system.

Example scripts are available in the directory $DALTON_EXA_DIR:

$ module show chem/dalton                # show environment variables, which will be available after 'module load'
$ module load chem/dalton                # load module
$ ls $DALTON_EXA_DIR                     # show content of directory $DALTON_EXA_DIR
$ cat $DALTON_EXA_DIR/README             # show examples README

Run several example jobs on JUSTUS 2:

$ module load chem/dalton                           # load module
$ WORKSPACE=`ws_allocate dalton 3`                  # allocate workspace
$ cd $WORKSPACE                                     # change to workspace
$ cp -a $DALTON_HOME/bwhpc-examples .               # copy example files to workspace
$ cd bwhpc-examples                                 # change to test directory
$ sbatch dalton-2020.0.slurm                        # submit job
$ squeue                                            # obtain JOBID
$ scontrol show job <JOBID>                         # check state of job
$ ls                                                # when job finishes the results will be visible in this directory

6 Useful links