INSTALL 3.2 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889
  1. To compile Hadoop Mapreduce next following, do the following:
  2. Step 1) Install dependencies for yarn
  3. See http://svn.apache.org/repos/asf/hadoop/common/branches/MR-279/mapreduce/yarn/README
  4. Make sure protbuf library is in your library path or set: export LD_LIBRARY_PATH=/usr/local/lib
  5. Step 2) Checkout
  6. svn checkout http://svn.apache.org/repos/asf/hadoop/common/branches/MR-279/
  7. Step 3) Build common
  8. Go to common directory
  9. ant veryclean mvn-install
  10. Step 4) Build HDFS
  11. Go to hdfs directory
  12. ant veryclean mvn-install -Dresolvers=internal
  13. Step 5) Build yarn and mapreduce
  14. Go to mapreduce directory
  15. export MAVEN_OPTS=-Xmx512m
  16. mvn clean install assembly:assembly
  17. ant veryclean jar jar-test -Dresolvers=internal
  18. In case you want to skip the tests run:
  19. mvn clean install assembly:assembly -DskipTests
  20. ant veryclean jar jar-test -Dresolvers=internal
  21. You will see a tarball in
  22. ls target/hadoop-mapreduce-1.0-SNAPSHOT-bin.tar.gz
  23. Step 6) Untar the tarball in a clean and different directory.
  24. say HADOOP_YARN_INSTALL
  25. To run Hadoop Mapreduce next applications :
  26. Step 7) cd $HADOOP_YARN_INSTALL
  27. Step 8) export the following variables:
  28. HADOOP_MAPRED_HOME=
  29. HADOOP_COMMON_HOME=
  30. HADOOP_HDFS_HOME=
  31. YARN_HOME=directory where you untarred yarn
  32. HADOOP_CONF_DIR=
  33. YARN_CONF_DIR=$HADOOP_CONF_DIR
  34. Step 9) bin/yarn-daemon.sh start resourcemanager
  35. Step 10) bin/yarn-daemon.sh start nodemanager
  36. Step 11) bin/yarn-daemon.sh start historyserver
  37. Step 12) Create the following symlinks in hadoop-common/lib
  38. ln -s $HADOOP_YARN_INSTALL/modules/hadoop-mapreduce-client-app-1.0-SNAPSHOT.jar .
  39. ln -s $HADOOP_YARN_INSTALL/modules/yarn-api-1.0-SNAPSHOT.jar .
  40. ln -s $HADOOP_YARN_INSTALL/modules/hadoop-mapreduce-client-common-1.0-SNAPSHOT.jar .
  41. ln -s $HADOOP_YARN_INSTALL/modules/yarn-common-1.0-SNAPSHOT.jar .
  42. ln -s $HADOOP_YARN_INSTALL/modules/hadoop-mapreduce-client-core-1.0-SNAPSHOT.jar .
  43. ln -s $HADOOP_YARN_INSTALL/modules/yarn-server-common-1.0-SNAPSHOT.jar .
  44. ln -s $HADOOP_YARN_INSTALL/modules/hadoop-mapreduce-client-jobclient-1.0-SNAPSHOT.jar .
  45. ln -s $HADOOP_YARN_INSTALL/lib/protobuf-java-2.4.0a.jar .
  46. Step 13) Yarn daemons are up! But for running mapreduce applications, which now are in user land, you need to setup nodemanager with the following configuration in your yarn-site.xml before you start the nodemanager.
  47. <property>
  48. <name>nodemanager.auxiluary.services</name>
  49. <value>mapreduce.shuffle</value>
  50. </property>
  51. <property>
  52. <name>nodemanager.aux.service.mapreduce.shuffle.class</name>
  53. <value>org.apache.hadoop.mapred.ShuffleHandler</value>
  54. </property>
  55. Step 14) You are all set, an example on how to run a mapreduce job is:
  56. cd $HADOOP_MAPRED_HOME
  57. ant examples -Dresolvers=internal
  58. $HADOOP_COMMON_HOME/bin/hadoop jar $HADOOP_MAPRED_HOME/build/hadoop-mapred-examples-0.22.0-SNAPSHOT.jar randomwriter -Dmapreduce.job.user.name=$USER -Dmapreduce.clientfactory.class.name=org.apache.hadoop.mapred.YarnClientFactory -Dmapreduce.randomwriter.bytespermap=10000 -Ddfs.blocksize=536870912 -Ddfs.block.size=536870912 -libjars $HADOOP_YARN_INSTALL/hadoop-mapreduce-1.0-SNAPSHOT/modules/hadoop-mapreduce-client-jobclient-1.0-SNAPSHOT.jar output
  59. The output on the command line should be almost similar to what you see in the JT/TT setup (Hadoop 0.20/0.21)