Descriptions of Slurm partitions
The following subsections sum up the parameters of the different partitions or queues available on Cholesky.
Default partition
No default partition.
You will have to select partition you want to submit your jobs. This can be done throught the command #SBATCH -p <partition name>
CPU partitions
These are the partitions for CPU computing on Cholesky.
Resources
| Partition Name | Nodes | Cores per node | Max. memory per node | Default memory per cpu | Max. Walltime |
|---|---|---|---|---|---|
| cpu_seq | [019-022] | 40 | 384 GB | 4 GB | 100:00:00 |
| cpu_shared | [001-020,023-032] | 40 | 188GB / 384 GB | 4 GB | 100:00:00 |
| cpu_dist | [023-068] | 40 | 188 GB | 4 GB | 24:00:00 |
| cpu_test* | [001-002,023-068] | 40 | 188 GB | 4 GB | 00:30:00 |
Limits / QOS
| Partition Name | Max Ncpus per user | Max jobs per user | Max. Walltime | Max Ncpus per job |
|---|---|---|---|---|
| cpu_seq | 40 | 40 | 100:00:00 | 1 |
| cpu_shared | 240 | - | 100:00:00 | 40 |
| cpu_dist | 400 | - | 24:00:00 | 400 |
| cpu_test* | 80 | 1 | 00:30:00 | 80 |
Note
cpu_test is a 'floating' queue used for debugging jobs with a QOS limit of 6 Maximum Nodes.
GPU partitions
These are the partitions for GPU computing on Cholesky.
Resources
| Partition Name | Nodes | Cores per node | Max. memory per node | Max. Walltime | GPU per node |
|---|---|---|---|---|---|
| gpu | [1-4] | 40 | 152 GB / 188 GB | 24:00:00 | 4 |
| gpu_v100 | [1-2] | 40 | 152 GB | 24:00:00 | 4 |
| gpu_a100 | [3-4] | 56 | 188 GB | 24:00:00 | 4 |
| gpu_a100_80g | 5 | 48 | 500 GB | 24:00:00 | 4 |
Limits / QOS
| Partition Name | Max Ngpus per user | Max jobs per user | Max. Walltime | Max Ngpus per job |
|---|---|---|---|---|
| gpu | 8 | - | 24:00:00 | 8 |
SLURM command
The sinfo SLURM command lists all available partitions.