[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revision Both sides next revision
grid [2019/02/14 17:49]
dusek [Advanced usage]
grid [2021/04/13 11:43]
127.0.0.1 external edit
Line 23: Line 23:
  
 ^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^ ^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^
-| andromeda[1-13]     | AMD Opteron                     | 2.8 |    7 |   30  |      | 
-| lucifer[1-10]       | Intel(R) Xeon(R) CPU E5620      | 2.4 |   15 |  122  |      | 
 | hydra[1-4]          | AMD Opteron SSE4 AVX            | 2.6 |   15 |  122  |   AVX enabled   | | hydra[1-4]          | AMD Opteron SSE4 AVX            | 2.6 |   15 |  122  |   AVX enabled   |
 | orion[1-8]          | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |   39 |  122  |   AVX enabled   | | orion[1-8]          | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |   39 |  122  |   AVX enabled   |
-| cosmos              | Intel Xeon                      | 2.9 |    7 |  249  |      | 
 | belzebub            | Intel Xeon SSE4 AVX             | 2.9 |   31 |  249  |   AVX enabled   | | belzebub            | Intel Xeon SSE4 AVX             | 2.9 |   31 |  249  |   AVX enabled   |
 | iridium             | Intel Xeon SSE4                 | 1.9 |   15 |  501  |      | | iridium             | Intel Xeon SSE4                 | 1.9 |   15 |  501  |      |
- 
-Machines from old cluster (do not use!): 
- 
-^ Name                ^ CPU type and flags              ^ GHz ^cores ^RAM(GB)^ note ^ 
-| fireball[1-10]      | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| hyperion[1-9]       | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| tauri[1-10]         | Intel Xeon           | 3.0 |    4 |   32 | removed | 
-| twister[1,2]        | Intel Xeon SSE4      | 2.4 |    8 |   48 | moved to GPU cluster| 
  
 <html><!-- <html><!--
Line 58: Line 47:
 === Submit hosts / test machines === === Submit hosts / test machines ===
 ^ Name                ^ CPU type                        ^ GHz ^cores ^ RAM(GB) ^ note  ^ ^ Name                ^ CPU type                        ^ GHz ^cores ^ RAM(GB) ^ note  ^
-| sol[1-10]           | Intel(R) Xeon(R) CPU E5345      | 2.|    7 |   31    | you can ssh here and compute or submit jobs |+| sol[1-8]           | Intel(R) Xeon(R) CPU E5-2630 v3 | 2.|    7 |   28    | you can ssh here and compute or submit jobs |
 | lrc[12]             | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |    4 |    4    | you can submit jobs here or monitor job execution - NO COMPUTATION IS ALLOWED HERE !!! | | lrc[12]             | Intel(R) Xeon(R) CPU E5-2630 v4 | 2.2 |    4 |    4    | you can submit jobs here or monitor job execution - NO COMPUTATION IS ALLOWED HERE !!! |
  
Line 172: Line 161:
 ''-q '*@hector[14]''' to submit on hector1 or hector4, ''-q '*@hector[14]''' to submit on hector1 or hector4,
 ''-q '[tm]*@!(hector*|iridium)''' to submit on any troja/ms machine except hectors and iridium. ''-q '[tm]*@!(hector*|iridium)''' to submit on any troja/ms machine except hectors and iridium.
-However, usually you should specify just the queue (troja-all.q vs. ms-all.q), not a particular machine, and instead use ''-l'' to specify the needed resources in a general way.+However, usually you should specify just the queue (cpu-troja.q vs. cpu-ms.q), not a particular machine, and instead use ''-l'' to specify the needed resources in a general way.
  
 ''qsub **-l** ...'' ''qsub **-l** ...''
Line 190: Line 179:
  
 ''qsub **-b** y'' ''qsub **-b** y''
-Treat ''script.sh'' (or whatever is the name of the command you execute) as a binary, i.e. don't search for [[#in-script options]] within the file, don't transfer it to the qmaster and then to the execution node. This makes the execution a bit faster and it may prevent some rare but hard-to-detect errors caused SGE interpreting the script. The script must be available on the execution node via NFS, Lustre (which is our case), etc. With ''-b y'' (shortcut for ''-b yes''), ''script.sh'' can be script or a binary. With ''-b n'' (which is the default for ''qsub''), ''script.sh'' must be a script (text file).+Treat ''script.sh'' (or whatever is the name of the command you execute) as a binary, i.e. don't search for [[#in-script options]] within the file, don't transfer it to the qmaster and then to the execution node. This makes the execution a bit faster and it may prevent some rare but hard-to-detect errors caused SGE interpreting the script. The script must be available on the execution node via NFS, Lustre (which is our case), etc. With ''-b y'' (shortcut for ''-b yes''), ''script.sh'' can be an executable script or a binary (and you must provide full path, e.g. ''./script.sh''). With ''-b n'' (which is the default for ''qsub''), ''script.sh'' must be a script (text file).
  
 ''qsub **-M** popel@ufal.mff.cuni.cz,rosa@ufal.mff.cuni.cz **-m** beas'' ''qsub **-M** popel@ufal.mff.cuni.cz,rosa@ufal.mff.cuni.cz **-m** beas''
Line 311: Line 300:
   * ''cat /opt/LRC/REPORTER/LRC-UFAL/stats/userlist.weight'' -- all users sorted according to their activity (number of submitted jobs × their average duration), updated each night   * ''cat /opt/LRC/REPORTER/LRC-UFAL/stats/userlist.weight'' -- all users sorted according to their activity (number of submitted jobs × their average duration), updated each night
  
-  * [[http://ufaladm2/munin/ufal.hide.ms.mff.cuni.cz/lrc-headnode.ufal.hide.ms.mff.cuni.cz/index.html|Munin: graph of cluster usage by day and user]] and  [[http://ufaladm2/munin/ufal.hide.ms.mff.cuni.cz/apophis.ufal.hide.ms.mff.cuni.cz/index.html|Munin monitoring of Apophis disk server]] (both accessible only from ÚFAL network)+  * [[https://ufaladm2.ufal.hide.ms.mff.cuni.cz/munin/ufal.hide.ms.mff.cuni.cz/lrc-master.ufal.hide.ms.mff.cuni.cz/index.html|Munin: graph of cluster usage by day and user]] and  [[https://ufaladm2.ufal.hide.ms.mff.cuni.cz/munin/ufal.hide.ms.mff.cuni.cz/nfs-core.ufal.hide.ms.mff.cuni.cz/index.html|Munin monitoring of disk storage]] (both accessible only from ÚFAL network)
  
 ===== Profiling ===== ===== Profiling =====
Line 329: Line 318:
 ===== Other ===== ===== Other =====
   * There is a **great course [[http://ufal.mff.cuni.cz/courses/npfl102|Data intensive computing]]**, see the 2016 handouts if you missed the course. It covers the usage of [[http://spark.apache.org/|Spark]] (MapReduce/Hadoop alternative, but better) and HDFS (Hadoop filesystem).   * There is a **great course [[http://ufal.mff.cuni.cz/courses/npfl102|Data intensive computing]]**, see the 2016 handouts if you missed the course. It covers the usage of [[http://spark.apache.org/|Spark]] (MapReduce/Hadoop alternative, but better) and HDFS (Hadoop filesystem).
-  * This course had used a special **DLRC (Demo LRC) cluster** (students had to login with ''ssh -p 11422 ufallab.ms.mff.cuni.cz'' and special NPFL102-only LDAP logins) with six virtual machines on one physical. During the years when NPFL102 is not taught (e.g. 2017), the DLRC cluster has just one virtual machine. +  * **Note:** some hadoop basics and a lot of NoSQL technologies are covered by [[https://is.cuni.cz/studium/predmety/index.php?do=predmet&kod=NDBI040|Big Data Management and NoSQL Databases]] 
-  * **Note:** soma hadoop basics and a lot of NoSQL technologies are covered by [[https://is.cuni.cz/studium/predmety/index.php?do=predmet&kod=NDBI040|Big Data Management and NoSQL Databases]]+  * There is a special cluster for Mgr (and Bc) students (but not for PhD and UFAL members): http://aic.ufal.mff.cuni.cz/
   * You can use environment variables ''$JOB_ID'', ''$JOB_NAME''.   * You can use environment variables ''$JOB_ID'', ''$JOB_NAME''.
   * One job can submit other jobs (but be careful with recursive:-)). A job submitted to the CPU cluster may submit GPU jobs (to the ''qpu.q'' queue).   * One job can submit other jobs (but be careful with recursive:-)). A job submitted to the CPU cluster may submit GPU jobs (to the ''qpu.q'' queue).
-  * It is important, that the files that are sourced during a login such as .bash_profile, .profile, .bashrc, .login etc. don't produce any output when a non-interactive login is done. If they do, changes are that your job will run, but that the batch system is unable to deliver to you the standard output and error files. In that case the status of your job will be 'E' after the job is finished. Here is an example how you can test in a .bash_profile or .bashrc if this is an interactive login:+  * It is important, that the files that are sourced during a login such as .bash_profile, .profile, .bashrc, .login etc. don't produce any output when a non-interactive login is done. If they do, chances are that your job will run, but that the batch system is unable to deliver to you the standard output and error files. In that case the status of your job will be 'E' after the job is finished. Here is an example how you can test in a .bash_profile or .bashrc if this is an interactive login:
 <code> <code>
 unset INTERACTIVE unset INTERACTIVE

[ Back to the navigation ] [ Back to the content ]