User Tools

Site Tools


lsf_gpu

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
lsf_gpu [2014/02/07 10:42]
schlarbm
lsf_gpu [2018/05/08 08:36] (current)
meesters [GPU Queues] - fixed broken link
Line 1: Line 1:
 +====== GPU Queues ======
 +
 +There are three different [[partitions|partitions (SLURM lingo for '​queues'​)]] inside the cluster that support gpu usage: The titan-Queues (''​titanshort/​long''​) currently include hosts i0001-i0009,​ while the gpu-Queues (''​infogpu''​) include the hosts g0001-g0009 ((formally there have been the ''​gpushort/​long''​ queues on these nodes - access however is restricted.)). The titan-hosts carry 4 GeForce GTX TITAN, hence a usage request up to ''​cuda=4''​ can be selected (see below). In contrast the GeForce GTX 480 is installed on the gpu-hosts (for the ''​gpushort/​long''​ queues). Finally, for the tesla-Queues (''​teslashort/​long''​) 4 Tesla K20m cards are installed.
 +
 ====== GPU Usage ====== ====== GPU Usage ======
  
-To use a GPU you have to explicitely ​reserve it as a ressource ​in the bsub call:+To use a GPU you have to explicitly ​reserve it as a resource ​in the submission script:
  
 <code bash> <code bash>
-$ bsub -n 1 -R '​select[cuda=1],​rusage[cuda=1]' ​-q gpulong ./​my_program+#​!/​bin/​bash 
 +# ... other SBATCH statements 
 +#​SBATCH ​--gres=gpu:<​number>​ 
 +#​SBATCH ​-p <​appropriate partition>​
 </​code>​ </​code>​
  
 +Number can be anything from 1-4 on our GPU nodes. In order to use more than 1 GPU the application needs to support using this much, of course.
 +
 +===== Using multiple nodes and multiple GPUs =====
  
 +In order to use multiples nodes, you have to request more than one node.
lsf_gpu.1391766156.txt.gz · Last modified: 2014/02/07 10:42 by schlarbm