[ Skip to the content ]

Institute of Formal and Applied Linguistics Wiki


[ Back to the navigation ]

This is an old revision of the document!


Table of Contents

Running Spark on Single Machine or on Cluster

In order to use Spark, environment has to bee set up according to Using Spark in UFAL Environment.

When Spark computation starts, it uses environment variable MASTER to determine the mode of computation. The following values are possible:

Running Spark on Single Machine

Spark computations can be started both on desktop machines and on cluster machines, either by specifying MASTER to one of local modes, or by not speficying MASTER at all (local[*] is used then).

Note that when you use qrsh or qsub, your job can usually use one core, so you should specify MASTER=local. If you do not, Spark will use all cores on the machine, even though SGE gave you onle one.

Starting Spark Cluster


[ Back to the navigation ] [ Back to the content ]