hadoop 9.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289
  1. #!/usr/bin/env bash
  2. # Licensed to the Apache Software Foundation (ASF) under one or more
  3. # contributor license agreements. See the NOTICE file distributed with
  4. # this work for additional information regarding copyright ownership.
  5. # The ASF licenses this file to You under the Apache License, Version 2.0
  6. # (the "License"); you may not use this file except in compliance with
  7. # the License. You may obtain a copy of the License at
  8. #
  9. # http://www.apache.org/licenses/LICENSE-2.0
  10. #
  11. # Unless required by applicable law or agreed to in writing, software
  12. # distributed under the License is distributed on an "AS IS" BASIS,
  13. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  14. # See the License for the specific language governing permissions and
  15. # limitations under the License.
  16. # The Hadoop command script
  17. #
  18. # Environment Variables
  19. #
  20. # JAVA_HOME The java implementation to use. Overrides JAVA_HOME.
  21. #
  22. # HADOOP_CLASSPATH Extra Java CLASSPATH entries.
  23. #
  24. # HADOOP_HEAPSIZE The maximum amount of heap to use, in MB.
  25. # Default is 1000.
  26. #
  27. # HADOOP_OPTS Extra Java runtime options.
  28. #
  29. # HADOOP_NAMENODE_OPTS These options are added to HADOOP_OPTS
  30. # HADOOP_CLIENT_OPTS when the respective command is run.
  31. # HADOOP_{COMMAND}_OPTS etc HADOOP_JT_OPTS applies to JobTracker
  32. # for e.g. HADOOP_CLIENT_OPTS applies to
  33. # more than one command (fs, dfs, fsck,
  34. # dfsadmin etc)
  35. #
  36. # HADOOP_CONF_DIR Alternate conf dir. Default is ${HADOOP_HOME}/conf.
  37. #
  38. # HADOOP_ROOT_LOGGER The root appender. Default is INFO,console
  39. #
  40. bin=`dirname "$0"`
  41. bin=`cd "$bin"; pwd`
  42. . "$bin"/hadoop-config.sh
  43. cygwin=false
  44. case "`uname`" in
  45. CYGWIN*) cygwin=true;;
  46. esac
  47. # if no args specified, show usage
  48. if [ $# = 0 ]; then
  49. echo "Usage: hadoop [--config confdir] COMMAND"
  50. echo "where COMMAND is one of:"
  51. echo " namenode -format format the DFS filesystem"
  52. echo " secondarynamenode run the DFS secondary namenode"
  53. echo " namenode run the DFS namenode"
  54. echo " datanode run a DFS datanode"
  55. echo " dfsadmin run a DFS admin client"
  56. echo " mradmin run a Map-Reduce admin client"
  57. echo " fsck run a DFS filesystem checking utility"
  58. echo " fs run a generic filesystem user client"
  59. echo " balancer run a cluster balancing utility"
  60. echo " jobtracker run the MapReduce job Tracker node"
  61. echo " pipes run a Pipes job"
  62. echo " tasktracker run a MapReduce task Tracker node"
  63. echo " job manipulate MapReduce jobs"
  64. echo " queue get information regarding JobQueues"
  65. echo " version print the version"
  66. echo " jar <jar> run a jar file"
  67. echo " distcp <srcurl> <desturl> copy file or directories recursively"
  68. echo " archive -archiveName NAME <src>* <dest> create a hadoop archive"
  69. echo " daemonlog get/set the log level for each daemon"
  70. echo " or"
  71. echo " CLASSNAME run the class named CLASSNAME"
  72. echo "Most commands print help when invoked w/o parameters."
  73. exit 1
  74. fi
  75. # get arguments
  76. COMMAND=$1
  77. shift
  78. if [ -f "${HADOOP_CONF_DIR}/hadoop-env.sh" ]; then
  79. . "${HADOOP_CONF_DIR}/hadoop-env.sh"
  80. fi
  81. # some Java parameters
  82. if [ "$JAVA_HOME" != "" ]; then
  83. #echo "run java in $JAVA_HOME"
  84. JAVA_HOME=$JAVA_HOME
  85. fi
  86. if [ "$JAVA_HOME" = "" ]; then
  87. echo "Error: JAVA_HOME is not set."
  88. exit 1
  89. fi
  90. JAVA=$JAVA_HOME/bin/java
  91. JAVA_HEAP_MAX=-Xmx1000m
  92. # check envvars which might override default args
  93. if [ "$HADOOP_HEAPSIZE" != "" ]; then
  94. #echo "run with heapsize $HADOOP_HEAPSIZE"
  95. JAVA_HEAP_MAX="-Xmx""$HADOOP_HEAPSIZE""m"
  96. #echo $JAVA_HEAP_MAX
  97. fi
  98. # CLASSPATH initially contains $HADOOP_CONF_DIR
  99. CLASSPATH="${HADOOP_CONF_DIR}"
  100. CLASSPATH=${CLASSPATH}:$JAVA_HOME/lib/tools.jar
  101. # for developers, add Hadoop classes to CLASSPATH
  102. if [ -d "$HADOOP_HOME/build/classes" ]; then
  103. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/classes
  104. fi
  105. if [ -d "$HADOOP_HOME/build/webapps" ]; then
  106. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build
  107. fi
  108. if [ -d "$HADOOP_HOME/build/test/classes" ]; then
  109. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/test/classes
  110. fi
  111. if [ -d "$HADOOP_HOME/build/tools" ]; then
  112. CLASSPATH=${CLASSPATH}:$HADOOP_HOME/build/tools
  113. fi
  114. # so that filenames w/ spaces are handled correctly in loops below
  115. IFS=
  116. # for releases, add core hadoop jar & webapps to CLASSPATH
  117. if [ -d "$HADOOP_HOME/webapps" ]; then
  118. CLASSPATH=${CLASSPATH}:$HADOOP_HOME
  119. fi
  120. for f in $HADOOP_HOME/hadoop-*-core.jar; do
  121. CLASSPATH=${CLASSPATH}:$f;
  122. done
  123. # add libs to CLASSPATH
  124. for f in $HADOOP_HOME/lib/*.jar; do
  125. CLASSPATH=${CLASSPATH}:$f;
  126. done
  127. if [ -d "$HADOOP_HOME/build/ivy/lib/Hadoop/common" ]; then
  128. for f in $HADOOP_HOME/build/ivy/lib/Hadoop/common/*.jar; do
  129. CLASSPATH=${CLASSPATH}:$f;
  130. done
  131. fi
  132. for f in $HADOOP_HOME/lib/jsp-2.1/*.jar; do
  133. CLASSPATH=${CLASSPATH}:$f;
  134. done
  135. for f in $HADOOP_HOME/hadoop-*-tools.jar; do
  136. TOOL_PATH=${TOOL_PATH}:$f;
  137. done
  138. for f in $HADOOP_HOME/build/hadoop-*-tools.jar; do
  139. TOOL_PATH=${TOOL_PATH}:$f;
  140. done
  141. # add user-specified CLASSPATH last
  142. if [ "$HADOOP_CLASSPATH" != "" ]; then
  143. CLASSPATH=${CLASSPATH}:${HADOOP_CLASSPATH}
  144. fi
  145. # default log directory & file
  146. if [ "$HADOOP_LOG_DIR" = "" ]; then
  147. HADOOP_LOG_DIR="$HADOOP_HOME/logs"
  148. fi
  149. if [ "$HADOOP_LOGFILE" = "" ]; then
  150. HADOOP_LOGFILE='hadoop.log'
  151. fi
  152. # default policy file for service-level authorization
  153. if [ "$HADOOP_POLICYFILE" = "" ]; then
  154. HADOOP_POLICYFILE="hadoop-policy.xml"
  155. fi
  156. # restore ordinary behaviour
  157. unset IFS
  158. # figure out which class to run
  159. if [ "$COMMAND" = "namenode" ] ; then
  160. CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'
  161. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_NAMENODE_OPTS"
  162. elif [ "$COMMAND" = "secondarynamenode" ] ; then
  163. CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'
  164. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_SECONDARYNAMENODE_OPTS"
  165. elif [ "$COMMAND" = "datanode" ] ; then
  166. CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
  167. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_DATANODE_OPTS"
  168. elif [ "$COMMAND" = "fs" ] ; then
  169. CLASS=org.apache.hadoop.fs.FsShell
  170. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  171. elif [ "$COMMAND" = "dfs" ] ; then
  172. CLASS=org.apache.hadoop.fs.FsShell
  173. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  174. elif [ "$COMMAND" = "dfsadmin" ] ; then
  175. CLASS=org.apache.hadoop.hdfs.tools.DFSAdmin
  176. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  177. elif [ "$COMMAND" = "mradmin" ] ; then
  178. CLASS=org.apache.hadoop.mapred.tools.MRAdmin
  179. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  180. elif [ "$COMMAND" = "fsck" ] ; then
  181. CLASS=org.apache.hadoop.hdfs.tools.DFSck
  182. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  183. elif [ "$COMMAND" = "balancer" ] ; then
  184. CLASS=org.apache.hadoop.hdfs.server.balancer.Balancer
  185. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_BALANCER_OPTS"
  186. elif [ "$COMMAND" = "jobtracker" ] ; then
  187. CLASS=org.apache.hadoop.mapred.JobTracker
  188. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_JOBTRACKER_OPTS"
  189. elif [ "$COMMAND" = "tasktracker" ] ; then
  190. CLASS=org.apache.hadoop.mapred.TaskTracker
  191. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_TASKTRACKER_OPTS"
  192. elif [ "$COMMAND" = "job" ] ; then
  193. CLASS=org.apache.hadoop.mapred.JobClient
  194. elif [ "$COMMAND" = "queue" ] ; then
  195. CLASS=org.apache.hadoop.mapred.JobQueueClient
  196. elif [ "$COMMAND" = "pipes" ] ; then
  197. CLASS=org.apache.hadoop.mapred.pipes.Submitter
  198. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  199. elif [ "$COMMAND" = "version" ] ; then
  200. CLASS=org.apache.hadoop.util.VersionInfo
  201. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  202. elif [ "$COMMAND" = "jar" ] ; then
  203. CLASS=org.apache.hadoop.util.RunJar
  204. elif [ "$COMMAND" = "distcp" ] ; then
  205. CLASS=org.apache.hadoop.tools.DistCp
  206. CLASSPATH=${CLASSPATH}:${TOOL_PATH}
  207. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  208. elif [ "$COMMAND" = "daemonlog" ] ; then
  209. CLASS=org.apache.hadoop.log.LogLevel
  210. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  211. elif [ "$COMMAND" = "archive" ] ; then
  212. CLASS=org.apache.hadoop.tools.HadoopArchives
  213. CLASSPATH=${CLASSPATH}:${TOOL_PATH}
  214. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  215. elif [ "$COMMAND" = "sampler" ] ; then
  216. CLASS=org.apache.hadoop.mapred.lib.InputSampler
  217. HADOOP_OPTS="$HADOOP_OPTS $HADOOP_CLIENT_OPTS"
  218. else
  219. CLASS=$COMMAND
  220. fi
  221. # cygwin path translation
  222. if $cygwin; then
  223. CLASSPATH=`cygpath -p -w "$CLASSPATH"`
  224. HADOOP_HOME=`cygpath -w "$HADOOP_HOME"`
  225. HADOOP_LOG_DIR=`cygpath -w "$HADOOP_LOG_DIR"`
  226. TOOL_PATH=`cygpath -p -w "$TOOL_PATH"`
  227. fi
  228. # setup 'java.library.path' for native-hadoop code if necessary
  229. JAVA_LIBRARY_PATH=''
  230. if [ -d "${HADOOP_HOME}/build/native" -o -d "${HADOOP_HOME}/lib/native" ]; then
  231. JAVA_PLATFORM=`CLASSPATH=${CLASSPATH} ${JAVA} -Xmx32m org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"`
  232. if [ -d "$HADOOP_HOME/build/native" ]; then
  233. JAVA_LIBRARY_PATH=${HADOOP_HOME}/build/native/${JAVA_PLATFORM}/lib
  234. fi
  235. if [ -d "${HADOOP_HOME}/lib/native" ]; then
  236. if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
  237. JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:${HADOOP_HOME}/lib/native/${JAVA_PLATFORM}
  238. else
  239. JAVA_LIBRARY_PATH=${HADOOP_HOME}/lib/native/${JAVA_PLATFORM}
  240. fi
  241. fi
  242. fi
  243. # cygwin path translation
  244. if $cygwin; then
  245. JAVA_LIBRARY_PATH=`cygpath -p "$JAVA_LIBRARY_PATH"`
  246. fi
  247. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.log.dir=$HADOOP_LOG_DIR"
  248. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.log.file=$HADOOP_LOGFILE"
  249. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.home.dir=$HADOOP_HOME"
  250. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.id.str=$HADOOP_IDENT_STRING"
  251. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.root.logger=${HADOOP_ROOT_LOGGER:-INFO,console}"
  252. if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then
  253. HADOOP_OPTS="$HADOOP_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH"
  254. fi
  255. HADOOP_OPTS="$HADOOP_OPTS -Dhadoop.policy.file=$HADOOP_POLICYFILE"
  256. # run it
  257. exec "$JAVA" $JAVA_HEAP_MAX $HADOOP_OPTS -classpath "$CLASSPATH" $CLASS "$@"