The batch or job scheduling system on PALMA-II is called SLURM. If you are used to PBS/Maui and want to switch to SLURM, this document might help you. The job scheduler is used to start and manage computations on the cluster but also to distribute resources among all users depending on their needs. Computation jobs (but also interactive sessions) can be submitted to different queues (or partitions in the slurm language), which have different purposes:

Partitions

Available for everyone:

NamePurposeCPU Arch

# Nodes

# GPUs
/ node

max. CPUs (threads)
/ node
max. Mem
/ node
max. WalltimeSlurm Status
normalgeneral computationsSkylake
(Gold 5118)

143

160

-36 (72)

92 GB

192 GB

7 days

express

short running (test) jobs

compilation

Skylake
(Gold 5118)
5-36 (72)92 GB2 hours

bigsmpSMPSkylake
(Gold 5118)
3-72 (144)1.5 TB7 days

largesmpSMPSkylake
(Gold 5118)
2-72 (144)3 TB7 days

requeue*

This queue will use the free nodes from
the group exclusive nodes listed below.

Skylake

(Gold 5118)

68

50

3

-

36 (72)

36 (72)

72 (144)

92 GB

192 GB

1.5 TB

24 hours
gpuv100Nvidia V100 GPUsSkylake
(Gold 5118)
1424192 GB2 days

vis-gpu
Nvidia Titan XPSkylake
(Gold 5118)
1824192 GB2 days

visVisualization / GUIsSkylake
(Gold 5118)
1-36 (72)92 GB2 hours

broadwellLegacy Broadwell CPUs

Broadwell
(E5-2683 v4)

44-32 (64)118 GB7 days

gpuk20Nvidia K20 GPUsIvybridge
(E5-2640 v2)
4316 (32)125 GB2 days

gpu2080GeForce RTX 2080 TiIvybridge
(E5-2695 v2)
12424 (48)125 GB2 days

gputitanrtxNvidia Titan RTX

Zen3
(EPYC 7343)

1432(32)240 GB2 days

gpu3090GeForce RTX 3090

Zen3
(EPYC 7413)

1848(48)240 GB2 days


requeue*

If your jobs are running on one of the requeue nodes while they are requested by one of the exclusive group partitions, your job will be terminated and resubmitted, so use with care!

Group exclusive:

Name

# Nodes

max. CPUs (threads) / nodemax. Mem / nodemax. Walltime
p0fuchs936 (72)92 GB7 days
p0kulesz

6

3

36 (72)

92 GB

192 GB

7 days
p0kapp136 (72)92 GB7 days
p0klasen

1

1

36 (72)

92 GB

192 GB

7 days
hims

25

1

36 (72)

92 GB

192 GB

7 days
d0ow136 (72)92 GB7 days
q0heuer1536 (72)92 GB7 days
e0mi236 (72)192 GB7 days
e0bm136 (72)192 GB7 days
p0rohlfi

7

8

36 (72)

92 GB

192 GB

7 days
SFB858372 (144)1.5 TB21 days
The above listed partitions and their resources might vary from time to time. To receive further and up to date information about the partitions use: scontrol show partition. Even more details can be shown by using sinfo -p <partitionname> or just sinfo.
  • No labels