User Tools

Site Tools


selecting_resources

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Next revision
Previous revision
selecting_resources [2023/03/01 20:21]
root created
selecting_resources [2024/03/10 17:18]
root
Line 1: Line 1:
 === Equipment/Selecting Resources/Slurm Parameters === === Equipment/Selecting Resources/Slurm Parameters ===
  
-There are five to six different Slurm parameters that must be specified to pick a computational resource and run a job. +There are six to seven different Slurm parameters that must be specified to pick a computational resource and run a job.  Additional Slurm parameters are optional.
-Additional Slurm parameters are optional.+
  
 <csv> <csv>
Line 12: Line 11:
 cpus-per-task , The number of hardware threads to allocate per process. cpus-per-task , The number of hardware threads to allocate per process.
 constraint , Sub-partition selection for partitions condo and pcon06. constraint , Sub-partition selection for partitions condo and pcon06.
 +qos, corresponding with partition, Quality of service class, generally the partition without the time, comp,gpu,tres,cloud,himem,condo
 </csv> </csv>
  
 Partitions are Partitions are
 <csv> <csv>
-Partition , Max hours , Max nodes, Max tasks , Max cpu/t , Num Nodes, Node type, Number of GPUs, Description +Partition , qos, Max hours , Max nodes, Max tasks , Max cpu/t , Num Nodes, Node type, Number of GPUs, Description 
-cloud72 ,  72, 1 , 1 , 2 , 3 , Xeon 6130/32c/192GB , 0 , shared queue for test jobs and low-effort tasks such as compiling and editing, usually available now for interactive +cloud72 ,  cloud, 72, 1 , 1 , 2 , 3 , Xeon 6130/32c/192GB , 0 , shared queue for test jobs and low-effort tasks such as compiling and editing, usually available now for interactive 
-comp72 , 72, n/a , 32* , 32* , 45, Dual Xeon 6130/32c/192GB , 0 , shared queue for long-term full-node computation +comp72 , comp, 72, n/a , 32* , 32* , 45, Dual Xeon 6130/32c/192GB , 0 , shared queue for long-term full-node computation 
-comp06 , 6, n/a , 32* , 32* , 45, Dual Xeon 6130/32c/192GB , 0 , same as comp72 except 6-hour limit and higher priority +comp06 , comp, 6, n/a , 32* , 32* , 45, Dual Xeon 6130/32c/192GB , 0 , same as comp72 except 6-hour limit and higher priority 
-comp01 , 1, n/a , 32* , 32* , 47, Dual Xeon 6130/32c/192GB , 0 , same as comp06 except 1-hour limit and higher priority +comp01 , comp, 1, n/a , 32* , 32* , 47, Dual Xeon 6130/32c/192GB , 0 , same as comp06 except 1-hour limit and higher priority 
-tres72 , 288, n/a , 32* , 32* , 111, Dual AMD 6136/32c/64GB , 0 , shared queue for very long-term full-node computation, uncrowded but slow nodes +tres72 , tres, 288, n/a , 32* , 32* , 111, Dual AMD 6136/32c/64GB , 0 , shared queue for very long-term full-node computation, uncrowded but slow nodes 
-tres288, 288, n/a , 32* , 32* , 111, Dual AMD 6136/32c/64GB , 0 , synonym for tres288 with expanded 288-hour limit +tres288, tres, 288, n/a , 32* , 32* , 111, Dual AMD 6136/32c/64GB , 0 , synonym for tres288 with expanded 288-hour limit 
-gpu72 , 72, n/a , 32* , 32* , 19, Dual Xeon 6130/32c/192GB , 1 V100 , shared queue for long-term full-node single GPU computation +gpu72 , gpu, 72, n/a , 32* , 32* , 19, Dual Xeon 6130/32c/192GB , 1 V100 , shared queue for long-term full-node single GPU computation 
-gpu06 , 6, n/a , 32* , 32* , 19, Dual Xeon 6130/32c/192GB , 1 V100 , same as gpu72 except 6-hour limit and higher priority +gpu06 , gpu, 6, n/a , 32* , 32* , 19, Dual Xeon 6130/32c/192GB , 1 V100 , same as gpu72 except 6-hour limit and higher priority 
-himem72 , 72, n/a , 24* , 24* , 6, Dual Xeon 6128/24c/768GB , 0 ,  shared queue for long-term full-node high-memory, same as comp72 except 24 cores, 768 GB +himem72 , himem, 72, n/a , 24* , 24* , 6, Dual Xeon 6128/24c/768GB , 0 ,  shared queue for long-term full-node high-memory, same as comp72 except 24 cores, 768 GB 
-himem06 , 6, n/a , 24* , 24* , 6, Dual Xeon 6128/24c/768GB , 0 ,  same as himem72 except 6-hour limit and higher priority +himem06 , himem, 6, n/a , 24* , 24* , 6, Dual Xeon 6128/24c/768GB , 0 ,  same as himem72 except 6-hour limit and higher priority 
-acomp06 , 6, n/a , 64* , 64* , 1, Dual AMD 7543/64c/1024GB , 0 ,  shared queue for medium-term full-node 64-core computation +acomp06 , comp, 6, n/a , 64* , 64* , 1, Dual AMD 7543/64c/1024GB , 0 ,  shared queue for medium-term full-node 64-core computation 
-agpu72 ,  72 , n/a , 64* , 64* , 16, Dual AMD 7543/64c/1024GB , 1 A100 ,  shared queue for medium-term full-node 64-core single GPU computation +agpu72 ,  gpu, 72 , n/a , 64* , 64* , 16, Dual AMD 7543/64c/1024GB , 1 A100 ,  shared queue for medium-term full-node 64-core single GPU computation 
-agpu06 , 6, n/a , 64* , 64* , 18, Dual AMD 7543/64c/1024GB , 1 A100 ,  same as agpu72 except 6-hour limit and higher priority +agpu06 , gpu, 6, n/a , 64* , 64* , 18, Dual AMD 7543/64c/1024GB , 1 A100 ,  same as agpu72 except 6-hour limit and higher priority 
-qgpu72 , 72 , n/a , 64* , 64* , 4, Dual AMD 7543/64c/1024GB , 4 A100 ,  shared queue for medium-term full-node 64-core quad GPU computation +qgpu72 , gpu, 72 , n/a , 64* , 64* , 4, Dual AMD 7543/64c/1024GB , 4 A100 ,  shared queue for medium-term full-node 64-core quad GPU computation 
-qgpu06 ,  6, n/a , 64* , 64* , 4, Dual AMD 7543/64c/1024GB , 4 A100 ,  same as qgpu72 except 6-hour limit and higher priority +qgpu06 ,  gpu, 6, n/a , 64* , 64* , 4, Dual AMD 7543/64c/1024GB , 4 A100 ,  same as qgpu72 except 6-hour limit and higher priority 
-csce72 , 72, n/a , 64* , 64* , 14, Dual AMD 7543/64c/1024GB , 0 , For CSCE users, long-term full-node 64-core computation +csce72 , csce, 72, n/a , 64* , 64* , 14, Dual AMD 7543/64c/1024GB , 0 , For CSCE users, long-term full-node 64-core computation 
-condo , n/a , n/a , n/a , n/a , 190, n/a , n/a, n/a , Contact HPCC for correct usage +condo , condo, n/a , n/a , n/a , n/a , 190, n/a , n/a, n/a , Contact HPCC for correct usage 
-pcon06 , 6, 1 , n/a , n/a , 190, n/a , n/a , n/a , Contact HPCC for correct usage+pcon06 , comp, 6, 1 , n/a , n/a , 190, n/a , n/a , n/a , Contact HPCC for correct usage
 </csv> </csv>
  
Line 44: Line 44:
 4. Single-CPU jobs should go to the ''cloud72'' partition 4. Single-CPU jobs should go to the ''cloud72'' partition
 5. Because of the expense of GPUs and large memory, the a)''gpu'' and b)''himem'' partitions are reserved for jobs that a) use the GPU  b) use more than 180 GB of memory respectively 5. Because of the expense of GPUs and large memory, the a)''gpu'' and b)''himem'' partitions are reserved for jobs that a) use the GPU  b) use more than 180 GB of memory respectively
-6. ''comp'', ''gpu'', and ''himem'' nodes are reserved for full-node jobs except for a few specific exceptions > 1 core and < full node.  Contact hpc-support@listserv.uark.edu +6. qos limits the total number of jobs per user and per group for each class of partitions 
-7. ''condo'' or ''pcon06'' and ''constraint'': Contact hpc-support@listserv.uark.edu+7. ''comp'', ''gpu'', and ''himem'' nodes are reserved for full-node jobs except for a few specific exceptions > 1 core and < full node.  Contact hpc-support@listserv.uark.edu 
 +8. ''condo'' or ''pcon06'' and ''constraint'': Contact hpc-support@listserv.uark.edu
  
  
selecting_resources.txt · Last modified: 2024/03/10 17:18 by root