Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
courses:mapreduce-tutorial:managing-a-hadoop-cluster [2012/02/05 19:56] straka |
courses:mapreduce-tutorial:managing-a-hadoop-cluster [2013/02/08 15:25] (current) popel |
||
---|---|---|---|
Line 5: | Line 5: | ||
A Hadoop cluster can be created: | A Hadoop cluster can be created: | ||
* for a specific Hadoop job. This is done by executing the job with the '' | * for a specific Hadoop job. This is done by executing the job with the '' | ||
- | * manually using ''/ | + | * manually using ''/ |
When a Hadoop cluster is about to start, a job is submitted to SGE cluster. When the cluster starts successfully, | When a Hadoop cluster is about to start, a job is submitted to SGE cluster. When the cluster starts successfully, | ||
Line 13: | Line 13: | ||
A Hadoop cluster is stopped: | A Hadoop cluster is stopped: | ||
- | * after the timeout specified by '' | + | * after the timeout specified by '' |
* when the '' | * when the '' | ||
* using '' | * using '' | ||
+ | |||
+ | ===== Web interface ===== | ||
+ | |||
+ | The web interface provides a lot of useful information: | ||
+ | * running, failed and successfully completed jobs | ||
+ | * for running job, current progress and counters of the whole job and also of each mapper and reducer is available | ||
+ | * for any job, the counters and outputs of all mappers and reducers | ||
+ | * for any job, all Hadoop settings | ||
===== Killing running jobs ===== | ===== Killing running jobs ===== | ||
- | Jobs running | + | Jobs running |
+ | < | ||
+ | |||
+ | The jobs running on a cluster are present in the web interface, or can be printed using | ||
+ | < |