Job scheduler on BioHPC#
A SLURM cluster cbsueccosl01
is maintained by Lars on behalf of Econ, on several nodes. Some are dedicated to the SLURM scheduler, others “borrowed”; the latter might not always be available. There are between 48 and 144 “slots” (cpus) available for compute jobs (see Table).
Who can use#
Everybody in the ECCO group can submit jobs.
Current load#
The most current status (as per the date and time noted) is:
As of 2025-04-25 06:00:
CLUSTER: cbsueccosl01
For more details, see the SLURM Queue page.
Manually query the latest availability#
To see availability at any point in time, type
sinfo --cluster cbsueccosl01
in a terminal window on the head node,[1] to obtain a result such as
$ sinfo --cluster cbsueccosl01
CLUSTER: cbsueccosl01
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST
regular* up infinite 3 mix cbsuecco[07-08],cbsueccosl03
regular* up infinite 1 alloc cbsueccosl04
regular* up infinite 2 idle cbsuecco01,cbsueccosl01
which shows that currently, 6 nodes are available for jobs, of which 2 are idle, three have some jobs running on them, but can still accept smaller jobs (mix
means there are free CPUs), and one is completely used (alloc
).
List of nodes#
The following table shows the allocated nodes. Nodes marked flex
may not be available. Nodes marked slurm
are always available. HT
means “hyper-threading”, and effectively multiplies the number of cores by 2, but may not always lead to performance improvement. MATLAB ignores hyper-threading, and will only use the physical number of cores listed in the cores
column.
Nodename | allocation | cores | RAM | local storage in TB | model | cores per CPU | CPUs | HT | cpu benchmark (single thread) | vintage |
---|---|---|---|---|---|---|---|---|---|---|
Loading... (need help?) |
Size of the cluster#
Total cores possible across all SLURM nodes: 704
Total RAM possible across all SLURM nodes: 3712 GB
Total cores currently available across all SLURM nodes: 16
Total RAM currently available across all SLURM nodes: 128 GB