Both sides previous revision
Previous revision
Next revision
|
Previous revision
Next revision
Both sides next revision
|
slurm [2022/09/08 13:25] vodrazka [Cluster info] |
slurm [2022/09/27 13:54] vodrazka [Batch mode] |
====== ÚFAL Grid Engine (LRC) ====== | ====== ÚFAL Grid Engine (LRC) ====== |
| |
LRC (Linguistic Research Cluster) is a name of ÚFAL's computational grid/cluster. | LRC (Linguistic Research Cluster) is the name of ÚFAL's computational grid/cluster. The cluster is built on top of [[https://slurm.schedmd.com/|SLURM]] and is using [[https://www.lustre.org/|Lustre]] for [[internal:linux-network#directory-structure|data storage]]. |
| |
| Currently there are following partitions (queues) available for computing: |
| |
| | **Partition name** | **Nodes** | **Note** | |
| | cpu-troja | 7x CPU | default partition | |
| | gpu-troja | 6x GPU | features: gpuram48G,gpuram40G | |
| | gpu-ms | 7x GPU | features: gpuram48G,gpuram24G | |
| |
| In order to submit a job you need to login to one of the head nodes: |
| |
| lrc1.ufal.hide.ms.mff.cuni.cz |
| lrc2.ufal.hide.ms.mff.cuni.cz |
===== Basic usage ===== | ===== Basic usage ===== |
| |
#!/bin/bash | #!/bin/bash |
#SBATCH -J helloWorld # name of job | #SBATCH -J helloWorld # name of job |
#SBATCH -p cpu-troja # name of partition or queue | #SBATCH -p cpu-troja # name of partition or queue (if not specified default partition is used) |
#SBATCH -o helloWorld.out # name of output file for this submission script | #SBATCH -o helloWorld.out # name of output file for this submission script |
#SBATCH -e helloWorld.err # name of error file for this submission script | #SBATCH -e helloWorld.err # name of error file for this submission script |
#SBATCH -N 2 # number of nodes (default 1) | #SBATCH -N 2 # number of nodes (default 1) |
#SBATCH --nodelist=node1,node2... # required node, or comma separated list of required nodes | #SBATCH --nodelist=node1,node2... # required node, or comma separated list of required nodes |
#SBATCH -c 4 # number of cores/threads per task (default 1) | #SBATCH --cpus-per-task=4 # number of cores/threads per task (default 1) |
#SBATCH --gres=gpu:1 # number of GPUs to request (default 0) | #SBATCH --gres=gpu:1 # number of GPUs to request (default 0) |
#SBATCH --mem=10G # request 10 gigabytes memory (per node, default depends on node) | #SBATCH --mem=10G # request 10 gigabytes memory (per node, default depends on node) |
<code>srun -p cpu-troja --mem=64G --pty bash</code> | <code>srun -p cpu-troja --mem=64G --pty bash</code> |
| |
* ''-p cpu-troja'' explicitly requires partition ''cpu-troja'' | * ''-p cpu-troja'' explicitly requires partition ''cpu-troja''. If not specified slurm will use default partition. |
* ''--mem=64G'' requires 64G of memory for the job | * ''--mem=64G'' requires 64G of memory for the job |
| |
<code>srun -p gpu-troja --nodelist=tdll-3gpu1 --mem=64G --gres=gpu:2 --pty bash</code> | <code>srun -p gpu-troja,gpu-ms --nodelist=tdll-3gpu1 --mem=64G --gres=gpu:2 --pty bash</code> |
| * ''-p gpu-troja,gpu-ms'' require only nodes from these two partitions |
* ''--nodelist=tdll-3gpu1'' explicitly requires one specific node | * ''--nodelist=tdll-3gpu1'' explicitly requires one specific node |
* ''--gres=gpu:2'' requires 2 GPUs | * ''--gres=gpu:2'' requires 2 GPUs |
| |
| <code>srun -p gpu-troja --constraint="gpuram44G|gpuram39G" --mem=64G --gres=gpu:2 --pty bash</code> |
| * ''--constraint="gpuram44G|gpuram39G"'' only consider nodes that have either ''gpuram44G'' or ''gpuram39G'' feature defined |
| |
To see all the available options type: | To see all the available options type: |