Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision | ||
courses:mapreduce-tutorial:running-jobs [2012/02/05 21:14] straka |
courses:mapreduce-tutorial:running-jobs [2013/02/08 14:33] (current) popel Milan improved our Hadoop |
||
---|---|---|---|
Line 5: | Line 5: | ||
The output of a Hadoop job must be a directory, which does not exist. | The output of a Hadoop job must be a directory, which does not exist. | ||
- | ===== Run Perl jobs ===== | + | ===== Running |
- | Choosing mode of operation: | + | |
- | | ^ Command ^ | + | |
- | ^ Run locally | '' | + | |
- | ^ Run using specified jobtracker | '' | + | |
- | ^ Run job in dedicated cluster | '' | + | |
- | ^ Run job in dedicated cluster and after it finishes, \\ wait for //W// seconds before stopping the cluster | '' | + | |
- | Specifying number of mappers and reducers: | ||
| ^ Command ^ | | ^ Command ^ | ||
- | ^ Run using //R// reducers \\ (//R//>1 not working when running locally)| '' | + | ^ Run Perl script '' |
- | ^ Run using //M// mappers | '' | + | ^ Run Java job '' |
+ | |||
+ | The options are the same for Perl and java: | ||
+ | |||
+ | | ^ Options ^ | ||
+ | ^ Run locally | '' | ||
+ | ^ Run using specified jobtracker | '' | ||
+ | ^ Run job in dedicated cluster | '' | ||
+ | ^ Run job in dedicated cluster and after it finishes, \\ wait for //W// seconds before stopping the cluster | '' | ||
+ | ^ Run using //R// reducers \\ (//R//>1 not working when running locally)| '' | ||
+ | ^ Run using //M// mappers | '' | ||
+ | |||
+ | From February 2012, using the parameter '' | ||
+ | |||
+ | ===== Running multiple jobs ===== | ||
+ | There are several ways of running multiple jobs: | ||
+ | * Java only: Create multiple '' | ||
+ | * Create a cluster using ''/ | ||
+ | * Create a shell script running multiple jobs using '' |