[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
slurm [2022/09/08 12:48]
vodrazka [Interactive mode]
slurm [2022/09/13 10:01]
vodrazka [ÚFAL Grid Engine (LRC)]
Line 1: Line 1:
 ====== ÚFAL Grid Engine (LRC) ====== ====== ÚFAL Grid Engine (LRC) ======
  
-LRC (Linguistic Research Cluster) is name of ÚFAL's computational grid/cluster.+LRC (Linguistic Research Cluster) is the name of ÚFAL's computational grid/cluster. The cluster is built on top of [[https://slurm.schedmd.com/|SLURM]] and is using [[https://www.lustre.org/|Lustre]] for [[internal:linux-network#directory-structure|data storage]].
  
 +Currently there are following partitions (queues) available for computing:
 +
 +| **Partition name** | **Nodes**  | **Note** |
 +| cpu-troja      | 7x CPU | default partition |
 +| gpu-troja      | 6x GPU | features: gpuram48G,gpuram40G |
 +| cpu-ms         | 7x GPU | features: gpuram48G,gpuram24G |
 +
 +In order to submit a job you need to login to one of the head nodes:
 +
 +   lrc1.ufal.hide.ms.mff.cuni.cz
 +   lrc2.ufal.hide.ms.mff.cuni.cz
 ===== Basic usage ===== ===== Basic usage =====
  
Line 17: Line 28:
 #!/bin/bash #!/bin/bash
 #SBATCH -J helloWorld   # name of job #SBATCH -J helloWorld   # name of job
-#SBATCH -p cpu-troja   # name of partition or queue+#SBATCH -p cpu-troja   # name of partition or queue (if not specified default partition is used)
 #SBATCH -o helloWorld.out   # name of output file for this submission script #SBATCH -o helloWorld.out   # name of output file for this submission script
 #SBATCH -e helloWorld.err   # name of error file for this submission script #SBATCH -e helloWorld.err   # name of error file for this submission script
Line 90: Line 101:
 <code> <code>
 sinfo sinfo
 +</code>
 +
 +List detailed info about nodes:
 +<code>
 +sinfo -l -N
 </code>  </code> 
  
-List types of available GPUs:+List nodes with some custom format info:
 <code> <code>
-sinfo -o %G+sinfo -N -o "%N %P %.11T %.15f"
 </code> </code>
  
Line 113: Line 129:
 <code>srun -p cpu-troja --mem=64G --pty bash</code> <code>srun -p cpu-troja --mem=64G --pty bash</code>
  
-  * ''-p cpu-troja'' explicitly requires partition ''cpu-troja''+  * ''-p cpu-troja'' explicitly requires partition ''cpu-troja''. If not specified slurm will use default partition.
   * ''--mem=64G'' requires 64G of memory for the job   * ''--mem=64G'' requires 64G of memory for the job
  
-<code>srun -p gpu-troja --nodelist=tdll-3gpu1 --mem=64G --gres=gpu:2 --pty bash</code> +<code>srun -p gpu-troja,gpu-ms --nodelist=tdll-3gpu1 --mem=64G --gres=gpu:2 --pty bash</code> 
 +  * ''-p gpu-troja,gpu-ms'' require only nodes from these two partitions
   * ''--nodelist=tdll-3gpu1'' explicitly requires one specific node   * ''--nodelist=tdll-3gpu1'' explicitly requires one specific node
   * ''--gres=gpu:2'' requires 2 GPUs   * ''--gres=gpu:2'' requires 2 GPUs
 +
 +<code>srun -p gpu-troja --constraint="gpuram44G|gpuram39G" --mem=64G --gres=gpu:2 --pty bash</code>
 +  * ''--constraint="gpuram44G|gpuram39G"'' only consider nodes that have either ''gpuram44G'' or ''gpuram39G'' feature defined
  
 To see all the available options type: To see all the available options type:

[ Back to the navigation ] [ Back to the content ]