User Tools

Site Tools


slurm_queues

**This is an old revision of the document!**

Slurm Queues Pinnacle/Karpinski

See Selecting Resources for help on choosing the best node/queue for your work.
Updates:

tres288 queue added with 288 hour/12 day maximum
tres72 time limit changed to 288 hours, same as tres288, retained for existing scripts
csce-k2-72 queue added for new csce Pinnacle-2 nodes

Pinnacle queues or slurm “partitions” are:

pinnacle partitiondescriptiontime limitcores per nodenumber of nodesother
comp01192 GB nodes1 hr3248full node usage required
comp06192 GB nodes6 hr3244full node usage required
comp72192 GB nodes72 hr3240full node usage required
gpu06gpu nodes6 hr3219gpu usage required/full node usage required
gpu72gpu nodes72 hr3219gpu usage required/full node usage required
himem06768 GB nodes6 hr246>192 GB memory usage required/full node usage required
himem72768 GB nodes72 hr246>192 GB memory usage required/full node usage required
cloud72virtual machines/containers/single processor jobs72 hr323for non-intensive computing up to 4 cores
tres7264 GB nodes72hr3223Trestles nodes with Pinnacle operating system
tres28864 GB nodes288hr3223Trestles nodes with Pinnacle operating system
karpinski partitiondescriptiontime limitcores per nodenumber of nodes
csce7232 GB nodes72 hr818
csce-k2-72256 GB nodes72 hr646
cscloud72virtual machines/containers/single processor jobs72 hr818

Condo queues are:

pinnacle partitiondescriptiontime limitnumber of nodesother
condocondo nodesnone25authorization and appropriate properties required
pcon06public use of condo nodes6 hr25appropriate properties required

Condo nodes require specification of a sufficient set of slurm properties. Property choices available are:

condo/pcon06 jobs running on the wrong nodes through lack of specified properties will be canceled without notice

non-gpu jobs running on gpu nodes may be canceled without notice

gpu or not: 0gpu/1v100/2v100
processor: i6130/a7351/i6128
equivalently: 192gb/256gb/768gb
equivalently: 32c/32c/24c
local drive: nvme/no specification
research group: fwang equivalent to 0gpu/i6130|i6230/768gb/32c|40c/nvme
research group: tkaman equivalent to 2v100/i6130/192gb/32c
research group: aja equivalent to 0gpu/i6128/192gb|768gb/24c

examples:
#SBATCH –constraint=2v100
#SBATCH –constraint=fwang
#SBATCH –constraint=768gb&0gpu
#SBATCH –constraint=256gb

A script is available to show idle nodes like this (in this case 2 nodes idle in the1-hour comp queue, none in 6-hour or 72-hour comp queue, but nodes available in gpu, himem, csce, and csce cloud. Sufficient idle nodes in your queue of interest do not guarantee that your job will start immediately, but that is usually the case.

$ idle_pinnacle_nodes.sh
n01=2 n06=0 n72=0
g06=1 g72=1 h06=2 h72=2 c72=16 l72=16
condo aja=2 wang=0 mich=2 kama=0
$
slurm_queues.1662745850.txt.gz · Last modified: 2022/09/09 17:50 by root