Skip to content

Quick reference

Useful hints and short information on issues that may vary between the different systems

Installed software software

To see the installed software available through the modules system, issue the command

module spider

To enquire about pre-requisites required for a specific package

module spider <module_name/version>

To see the currently loaded modules

module list

To load a module

module add <module_name>

To unload a module

module del <module_name>

Resource allocation

Number of cores

The number of cores for a job is specified in the batch script in the format

#SBATCH -N <number_of_nodes>
#SBATCH --ntasks-per-node=<number_of_cores_per_node>

COSMOS has 48 cores per node. In this system, 192 cores (4 nodes with 48 cores each) would be allocated through

# 192 cores on COSMOS
#SBACTH --ntasks-per-node=48

Memory per core

The amount of memory per core is specified in the format

#SBATCH --mem-per-cpu=<amount_of_memory_per_core_in_MB>

COSMOS has nodes with 256 GB of memory. The default allocation per core is therefore 5300 MB, allowing some memory for the operating system. Please note that if you increase your memory request beyond 5300 MB per core, some cores on the system will be idle due to the lack of memory. Your account gets charged for these cores as well.

Aurora is being decommissioned, so users are encouraged to shift existing projects to COSMOS and not start any new projects on Aurora. For legacy users' reference, Aurora has standard nodes with 64 GB of memory. The default allocation per core is therefore 3200 MB, allowing some memory for the operating system.

File systems

Home directory

Currently, all LUNARC systems have a home directory that is different for each system, i.e., the login directory for user xxxx is


This directory can be referenced as $HOME.

Global working directory

For job submission, we recommend using your home directory as the old /lunarc/nobackup/users directory is now decommisioned.

Local working directory

When a job is running, it has access to a temporary directory on the local disk of each allocated node. The directory can be referenced as $SNIC_TMP (or $TMPDIR). It will be deleted when the job finishes.

If a job is terminated prematurely, for example, if it exceeds the requested walltime, the files on the local disk will be lost. Files that would still be useful can be listed in a special file $SNIC_TMP/slurm_save_files. Filenames are assumed to be relative to $SNIC_TMP and should be separated by spaces or listed on separate lines. These files will be copied to the submission directory regardless of whether the job ends as planned or is deleted unless there is a problem with the disk or node itself.


To limit disk usage, quotas are set for each user and filesystem. The status can be seen at login. A quota report can also be obtained by issuing the command


The quota can be increased on request.

Test queues

On COSMOS, it is possible to request extra high priority to run short tests (maximum 1h) using at most 2 nodes using

#SBATCH --qos=test

It is not allowed to submit long series of jobs to a test queue.

Author: (LUNARC)

Last Updated: 2024-06-18