[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
grid [2020/01/07 15:07]
popel [Advanced usage]
grid [2024/10/02 15:21] (current)
popel
Line 5: Line 5:
 If you need GPU processing, see a special page about our [[:gpu|GPU cluster called DLL]] (which is actually a subsystem of LRC with an independent queue ''gpu-ms.q''). If you need GPU processing, see a special page about our [[:gpu|GPU cluster called DLL]] (which is actually a subsystem of LRC with an independent queue ''gpu-ms.q'').
 TODO: describe alternatives, e.g.: MetaCentrum / Cesnet cluster (all MFF students can use it), Amazon EC2, Microsoft Azure (some colleagues may have sometime free vouchers). TODO: describe alternatives, e.g.: MetaCentrum / Cesnet cluster (all MFF students can use it), Amazon EC2, Microsoft Azure (some colleagues may have sometime free vouchers).
 +
 +**TODO: IN 2022 MOVING FROM SGE TO SLURM** (see [[slurm|slurm guidelines]]) -- **commands like ''qsub'' and ''qstat'' will stop working!**
 +
 +**IN 2024: Newly, all the documentation is at a dedicated wiki https://ufal.mff.cuni.cz/lrc (you need to use ufal and [[internal:welcome-at-ufal#small-linguistic-password|small-linguistic password]] to access the wiki from outside of the UFAL network).***
  
 ===== List of Machines ===== ===== List of Machines =====
Line 23: Line 27:
  
 ^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^ ^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^
-| andromeda[1-13]     | AMD Opteron                     | 2.8 |    7 |   30  |      | 
-| lucifer[1-10]       | Intel(R) Xeon(R) CPU E5620      | 2.4 |   15 |  122  |      | 
 | hydra[1-4]          | AMD Opteron SSE4 AVX            | 2.6 |   15 |  122  |   AVX enabled   | | hydra[1-4]          | AMD Opteron SSE4 AVX            | 2.6 |   15 |  122  |   AVX enabled   |
 | orion[1-8]          | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |   39 |  122  |   AVX enabled   | | orion[1-8]          | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |   39 |  122  |   AVX enabled   |
-| cosmos              | Intel Xeon                      | 2.9 |    7 |  249  |      | 
 | belzebub            | Intel Xeon SSE4 AVX             | 2.9 |   31 |  249  |   AVX enabled   | | belzebub            | Intel Xeon SSE4 AVX             | 2.9 |   31 |  249  |   AVX enabled   |
 | iridium             | Intel Xeon SSE4                 | 1.9 |   15 |  501  |      | | iridium             | Intel Xeon SSE4                 | 1.9 |   15 |  501  |      |
- 
-Machines from old cluster (do not use!): 
- 
-^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^ 
-| fireball[1-10]      | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| hyperion[1-9]       | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| tauri[1-10]         | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| twister[1,2]        | Intel Xeon SSE4      | 2.4 |    8 |   48 | moved to GPU cluster| 
  
 <html><!-- <html><!--
Line 58: Line 51:
 === Submit hosts / test machines === === Submit hosts / test machines ===
 ^ Name                ^ CPU type                        ^ GHz ^cores ^ RAM(GB) ^ note  ^ ^ Name                ^ CPU type                        ^ GHz ^cores ^ RAM(GB) ^ note  ^
-| sol[1-10]           | Intel(R) Xeon(R) CPU E5345      | 2.|    7 |   31    | you can ssh here and compute or submit jobs |+| sol[1-8]           | Intel(R) Xeon(R) CPU E5-2630 v3 | 2.|    7 |   28    | you can ssh here and compute or submit jobs |
 | lrc[12]             | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |    4 |    4    | you can submit jobs here or monitor job execution - NO COMPUTATION IS ALLOWED HERE !!! | | lrc[12]             | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |    4 |    4    | you can submit jobs here or monitor job execution - NO COMPUTATION IS ALLOWED HERE !!! |
  
Line 168: Line 161:
  
 ''qsub **-q** cpu-troja.q'' ''qsub **-q** cpu-troja.q''
-This way your job is submitted to the Troja queue. The default is ''cpu-ms.q''. You can also use e.g. +This way your job is submitted to the Troja queue. The default is ''cpu-*.q''. You can also use e.g. 
-''-q '(troja*|ms*)''' to submit on any machine in those two queues (but **don't use ''-q '*'''** as this includes also [[:gpu|gpu.q]]),+''-q '(cpu-t*|cpu-m*)''' to submit on any machine in those two queues (but **don't use ''-q '*'''** as this includes also [[:gpu|gpu.q]]),
 ''-q '*@hector[14]''' to submit on hector1 or hector4, ''-q '*@hector[14]''' to submit on hector1 or hector4,
-''-q '[tm]*@!(hector*|iridium)''' to submit on any troja/ms machine except hectors and iridium.+''-q 'cpu-*@!(hector*|iridium)''' to submit on any troja/ms machine except hectors and iridium.
 However, usually you should specify just the queue (cpu-troja.q vs. cpu-ms.q), not a particular machine, and instead use ''-l'' to specify the needed resources in a general way. However, usually you should specify just the queue (cpu-troja.q vs. cpu-ms.q), not a particular machine, and instead use ''-l'' to specify the needed resources in a general way.
  
Line 223: Line 216:
 You can change some properties of already submitted jobs (both waiting in the queue and running). Changeable properties are listed in ''man qsub''. You can change some properties of already submitted jobs (both waiting in the queue and running). Changeable properties are listed in ''man qsub''.
  
-''**man** qsub qstat qalter qhold queue_conf sge_types complex''+''[[https://gridscheduler.sourceforge.net/htmlman/htmlman1/qsub.html|man qsub]] qstat qalter qhold queue_conf sge_types complex''
 Find out all the gory details which are missing here. You'll have to do it one day anyway:-). Find out all the gory details which are missing here. You'll have to do it one day anyway:-).
  
Line 329: Line 322:
 ===== Other ===== ===== Other =====
   * There is a **great course [[http://ufal.mff.cuni.cz/courses/npfl102|Data intensive computing]]**, see the 2016 handouts if you missed the course. It covers the usage of [[http://spark.apache.org/|Spark]] (MapReduce/Hadoop alternative, but better) and HDFS (Hadoop filesystem).   * There is a **great course [[http://ufal.mff.cuni.cz/courses/npfl102|Data intensive computing]]**, see the 2016 handouts if you missed the course. It covers the usage of [[http://spark.apache.org/|Spark]] (MapReduce/Hadoop alternative, but better) and HDFS (Hadoop filesystem).
-  * This course had used a special **DLRC (Demo LRC) cluster** (students had to login with ''ssh -p 11422 ufallab.ms.mff.cuni.cz'' and special NPFL102-only LDAP logins) with six virtual machines on one physical. During the years when NPFL102 is not taught (e.g. 2017), the DLRC cluster has just one virtual machine. +  * **Note:** some hadoop basics and a lot of NoSQL technologies are covered by [[https://is.cuni.cz/studium/predmety/index.php?do=predmet&kod=NDBI040|Big Data Management and NoSQL Databases]] 
-  * **Note:** soma hadoop basics and a lot of NoSQL technologies are covered by [[https://is.cuni.cz/studium/predmety/index.php?do=predmet&kod=NDBI040|Big Data Management and NoSQL Databases]]+  * There is a special cluster for Mgr (and Bc) students (but not for PhD and UFAL members): http://aic.ufal.mff.cuni.cz/
   * You can use environment variables ''$JOB_ID'', ''$JOB_NAME''.   * You can use environment variables ''$JOB_ID'', ''$JOB_NAME''.
   * One job can submit other jobs (but be careful with recursive:-)). A job submitted to the CPU cluster may submit GPU jobs (to the ''qpu.q'' queue).   * One job can submit other jobs (but be careful with recursive:-)). A job submitted to the CPU cluster may submit GPU jobs (to the ''qpu.q'' queue).
-  * It is important, that the files that are sourced during a login such as .bash_profile, .profile, .bashrc, .login etc. don't produce any output when a non-interactive login is done. If they do, changes are that your job will run, but that the batch system is unable to deliver to you the standard output and error files. In that case the status of your job will be 'E' after the job is finished. Here is an example how you can test in a .bash_profile or .bashrc if this is an interactive login:+  * It is important, that the files that are sourced during a login such as .bash_profile, .profile, .bashrc, .login etc. don't produce any output when a non-interactive login is done. If they do, chances are that your job will run, but that the batch system is unable to deliver to you the standard output and error files. In that case the status of your job will be 'E' after the job is finished. Here is an example how you can test in a .bash_profile or .bashrc if this is an interactive login:
 <code> <code>
 unset INTERACTIVE unset INTERACTIVE

[ Back to the navigation ] [ Back to the content ]