hadoop 9.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273
  1. #!/usr/bin/env bash
  2. # Licensed to the Apache Software Foundation (ASF) under one or more
  3. # contributor license agreements. See the NOTICE file distributed with
  4. # this work for additional information regarding copyright ownership.
  5. # The ASF licenses this file to You under the Apache License, Version 2.0
  6. # (the "License"); you may not use this file except in compliance with
  7. # the License. You may obtain a copy of the License at
  8. #
  9. # http://www.apache.org/licenses/LICENSE-2.0
  10. #
  11. # Unless required by applicable law or agreed to in writing, software
  12. # distributed under the License is distributed on an "AS IS" BASIS,
  13. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  14. # See the License for the specific language governing permissions and
  15. # limitations under the License.
  16. # The Hadoop command script
  17. #
  18. # Environment Variables
  19. #
  20. # JAVA_HOME The java implementation to use. Overrides JAVA_HOME.
  21. #
  22. # HADOOP_CLASSPATH Extra Java CLASSPATH entries.
  23. #
  24. # HADOOP_HEAPSIZE The maximum amount of heap to use, in MB.
  25. # Default is 1000.
  26. #
  27. # HADOOP_OPTS Extra Java runtime options.
  28. #
  29. # HADOOP_NAMENODE_OPTS These options are added to HADOOP_OPTS
  30. # HADOOP_CLIENT_OPTS when the respective command is run.
  31. # HADOOP_{COMMAND}_OPTS etc HADOOP_JT_OPTS applies to JobTracker
  32. # for e.g. HADOOP_CLIENT_OPTS applies to
  33. # more than one command (fs, dfs, fsck,
  34. # dfsadmin etc)
  35. #
  36. # HADOOP_CONF_DIR Alternate conf dir. Default is ${HADOOP_HOME}/conf.
  37. #
  38. # HADOOP_ROOT_LOGGER The root appender. Default is INFO,console
  39. #
  40. bin=`dirname "$0"`
  41. bin=`cd "$bin"; pwd`
  42. . "$bin"/hadoop-config.sh
  43. cygwin=false
  44. case "`uname`" in
  45. CYGWIN*) cygwin=true;;
  46. esac
  47. # if no args specified, show usage
  48. if [ $# = 0 ]; then
  49. echo "Usage: hadoop [--config confdir] COMMAND"
  50. echo "where COMMAND is one of:"
  51. echo " namenode -format format the DFS filesystem"
  52. echo " secondarynamenode run the DFS secondary namenode"
  53. echo " namenode run the DFS namenode"
  54. echo " datanode run a DFS datanode"
  55. echo " dfsadmin run a DFS admin client"
  56. echo " fsck run a DFS filesystem checking utility"
  57. echo " fs run a generic filesystem user client"
  58. echo " balancer run a cluster balancing utility"
  59. echo " jobtracker run the MapReduce job Tracker node"
  60. echo " pipes run a Pipes job"
  61. echo " tasktracker run a MapReduce task Tracker node"
  62. echo " job manipulate MapReduce jobs"
  63. echo " queue get information regarding JobQueues"
  64. echo " version print the version"
  65. echo " jar <jar> run a jar file"
  66. echo " distcp <srcurl> <desturl> copy file or directories recursively"
  67. echo " archive -archiveName NAME <src>* <dest> create a hadoop archive"
  68. echo " daemonlog get/set the log level for each daemon"
  69. echo " or"
  70. echo " CLASSNAME run the class named CLASSNAME"
  71. echo "Most commands print help when invoked w/o parameters."
  72. exit 1
  73. fi
  74. # get arguments
  75. COMMAND=$1
  76. shift
  77. if [ -f "${HADOOP_CONF_DIR}/hadoop-env.sh" ]; then
  78. . "${HADOOP_CONF_DIR}/hadoop-env.sh"
  79. fi
  80. # some Java parameters
  81. if [ "$JAVA_HOME" != "" ]; then
  82. #echo "run java in $JAVA_HOME"
  83. JAVA_HOME=$JAVA_HOME
  84. fi
  85. if [ "$JAVA_HOME" = "" ]; then
  86. echo "Error: JAVA_HOME is not set."
  87. exit 1
  88. fi
  89. JAVA=$JAVA_HOME/bin/java
  90. JAVA_HEAP_MAX=-Xmx1000m
  91. # check envvars which might override default args
  92. if [ "$HADOOP_HEAPSIZE" != "" ]; then
  93. #echo "run with heapsize $HADOOP_HEAPSIZE"
  94. JAVA_HEAP_MAX="-Xmx""$HADOOP_HEAPSIZE""m"
  95. #echo $JAVA_HEAP_MAX
  96. fi
  97. # CLASSPATH initially contains $HADOOP_CONF_DIR
  98. CLASSPATH="${HADOOP_CONF_DIR}"
  99. CLASSPATH=${CLASSPATH}:$JAVA_HOME/lib/tools.jar
  100. # for developers, add Hadoop classes to CLASSPATH
  101. if [ -d "$HADOOP_HOME/build/classes" ]; then
  102. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/classes
  103. fi
  104. if [ -d "$HADOOP_HOME/build/webapps" ]; then
  105. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build
  106. fi
  107. if [ -d "$HADOOP_HOME/build/test/classes" ]; then
  108. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/test/classes
  109. fi
  110. if [ -d "$HADOOP_HOME/build/tools" ]; then
  111. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/tools
  112. fi
  113. # so that filenames w/ spaces are handled correctly in loops below
  114. IFS=
  115. # for releases, add core hadoop jar & webapps to CLASSPATH
  116. if [ -d "$HADOOP_HOME/webapps" ]; then
  117. CLASSPATH=${CLASSPATH}:$HADOOP_HOME
  118. fi
  119. for f in $HADOOP_HOME/hadoop-*-core.jar; do
  120. CLASSPATH=${CLASSPATH}:$f;
  121. done
  122. # add libs to CLASSPATH
  123. for f in $HADOOP_HOME/lib/*.jar; do
  124. CLASSPATH=${CLASSPATH}:$f;
  125. done
  126. for f in $HADOOP_HOME/lib/jetty-ext/*.jar; do
  127. CLASSPATH=${CLASSPATH}:$f;
  128. done
  129. for f in $HADOOP_HOME/hadoop-*-tools.jar; do
  130. TOOL_PATH=${TOOL_PATH}:$f;
  131. done
  132. for f in $HADOOP_HOME/build/hadoop-*-tools.jar; do
  133. TOOL_PATH=${TOOL_PATH}:$f;
  134. done
  135. # add user-specified CLASSPATH last
  136. if [ "$HADOOP_CLASSPATH" != "" ]; then
  137. CLASSPATH=${CLASSPATH}:${HADOOP_CLASSPATH}
  138. fi
  139. # default log directory & file
  140. if [ "$HADOOP_LOG_DIR" = "" ]; then
  141. HADOOP_LOG_DIR="$HADOOP_HOME/logs"
  142. fi
  143. if [ "$HADOOP_LOGFILE" = "" ]; then
  144. HADOOP_LOGFILE='hadoop.log'
  145. fi
  146. # restore ordinary behaviour
  147. unset IFS
  148. # figure out which class to run
  149. if [ "$COMMAND" = "namenode" ] ; then
  150. CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
  151. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_NAMENODE_OPTS"
  152. elif [ "$COMMAND" = "secondarynamenode" ] ; then
  153. CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
  154. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_SECONDARYNAMENODE_OPTS"
  155. elif [ "$COMMAND" = "datanode" ] ; then
  156. CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
  157. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_DATANODE_OPTS"
  158. elif [ "$COMMAND" = "fs" ] ; then
  159. CLASS=org.apache.hadoop.fs.FsShell
  160. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  161. elif [ "$COMMAND" = "dfs" ] ; then
  162. CLASS=org.apache.hadoop.fs.FsShell
  163. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  164. elif [ "$COMMAND" = "dfsadmin" ] ; then
  165. CLASS=org.apache.hadoop.hdfs.tools.DFSAdmin
  166. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  167. elif [ "$COMMAND" = "fsck" ] ; then
  168. CLASS=org.apache.hadoop.hdfs.tools.DFSck
  169. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  170. elif [ "$COMMAND" = "balancer" ] ; then
  171. CLASS=org.apache.hadoop.hdfs.server.balancer.Balancer
  172. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_BALANCER_OPTS"
  173. elif [ "$COMMAND" = "jobtracker" ] ; then
  174. CLASS=org.apache.hadoop.mapred.JobTracker
  175. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_JOBTRACKER_OPTS"
  176. elif [ "$COMMAND" = "tasktracker" ] ; then
  177. CLASS=org.apache.hadoop.mapred.TaskTracker
  178. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_TASKTRACKER_OPTS"
  179. elif [ "$COMMAND" = "job" ] ; then
  180. CLASS=org.apache.hadoop.mapred.JobClient
  181. elif [ "$COMMAND" = "queue" ] ; then
  182. CLASS=org.apache.hadoop.mapred.JobQueueClient
  183. elif [ "$COMMAND" = "pipes" ] ; then
  184. CLASS=org.apache.hadoop.mapred.pipes.Submitter
  185. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  186. elif [ "$COMMAND" = "version" ] ; then
  187. CLASS=org.apache.hadoop.util.VersionInfo
  188. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  189. elif [ "$COMMAND" = "jar" ] ; then
  190. CLASS=org.apache.hadoop.mapred.JobShell
  191. elif [ "$COMMAND" = "distcp" ] ; then
  192. CLASS=org.apache.hadoop.tools.DistCp
  193. CLASSPATH=${CLASSPATH}:${TOOL_PATH}
  194. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  195. elif [ "$COMMAND" = "daemonlog" ] ; then
  196. CLASS=org.apache.hadoop.log.LogLevel
  197. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  198. elif [ "$COMMAND" = "archive" ] ; then
  199. CLASS=org.apache.hadoop.tools.HadoopArchives
  200. CLASSPATH=${CLASSPATH}:${TOOL_PATH}
  201. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  202. elif [ "$COMMAND" = "sampler" ] ; then
  203. CLASS=org.apache.hadoop.mapred.lib.InputSampler
  204. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  205. else
  206. CLASS=$COMMAND
  207. fi
  208. # cygwin path translation
  209. if $cygwin; then
  210. CLASSPATH=`cygpath -p -w "$CLASSPATH"`
  211. HADOOP_HOME=`cygpath -d "$HADOOP_HOME"`
  212. HADOOP_LOG_DIR=`cygpath -d "$HADOOP_LOG_DIR"`
  213. TOOL_PATH=`cygpath -p -w "$TOOL_PATH"`
  214. fi
  215. # setup 'java.library.path' for native-hadoop code if necessary
  216. JAVA_LIBRARY_PATH=''
  217. if [ -d "${HADOOP_HOME}/build/native" -o -d "${HADOOP_HOME}/lib/native" ]; then
  218. JAVA_PLATFORM=`CLASSPATH=${CLASSPATH} ${JAVA} org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"`
  219. if [ -d "$HADOOP_HOME/build/native" ]; then
  220. JAVA_LIBRARY_PATH=${HADOOP_HOME}/build/native/${JAVA_PLATFORM}/lib
  221. fi
  222. if [ -d "${HADOOP_HOME}/lib/native" ]; then
  223. if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
  224. JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:${HADOOP_HOME}/lib/native/${JAVA_PLATFORM}
  225. else
  226. JAVA_LIBRARY_PATH=${HADOOP_HOME}/lib/native/${JAVA_PLATFORM}
  227. fi
  228. fi
  229. fi
  230. # cygwin path translation
  231. if $cygwin; then
  232. JAVA_LIBRARY_PATH=`cygpath -p "$JAVA_LIBRARY_PATH"`
  233. fi
  234. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.log.dir=$HADOOP_LOG_DIR"
  235. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.log.file=$HADOOP_LOGFILE"
  236. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.home.dir=$HADOOP_HOME"
  237. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.id.str=$HADOOP_IDENT_STRING"
  238. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.root.logger=${HADOOP_ROOT_LOGGER:-INFO,console}"
  239. if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
  240. HADOOP_OPTS="$HADOOP_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH"
  241. fi
  242. # run it
  243. exec "$JAVA" $JAVA_HEAP_MAX $HADOOP_OPTS -classpath "$CLASSPATH" $CLASS "$@"