~~ Licensed to the Apache Software Foundation (ASF) under one or more ~~ contributor license agreements. See the NOTICE file distributed with ~~ this work for additional information regarding copyright ownership. ~~ The ASF licenses this file to You under the Apache License, Version 2.0 ~~ (the "License"); you may not use this file except in compliance with ~~ the License. You may obtain a copy of the License at ~~ ~~ http://www.apache.org/licenses/LICENSE-2.0 ~~ ~~ Unless required by applicable law or agreed to in writing, software ~~ distributed under the License is distributed on an "AS IS" BASIS, ~~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ~~ See the License for the specific language governing permissions and ~~ limitations under the License. --- Hadoop Commands Guide --- --- ${maven.build.timestamp} %{toc} Overview All hadoop commands are invoked by the <<>> script. Running the hadoop script without any arguments prints the description for all commands. Usage: <<>> Hadoop has an option parsing framework that employs parsing generic options as well as running classes. *-----------------------+---------------+ || COMMAND_OPTION || Description *-----------------------+---------------+ | <<<--config confdir>>>| Overwrites the default Configuration directory. Default is <<<${HADOOP_HOME}/conf>>>. *-----------------------+---------------+ | GENERIC_OPTIONS | The common set of options supported by multiple commands. | COMMAND_OPTIONS | Various commands with their options are described in the following sections. The commands have been grouped into User Commands and Administration Commands. *-----------------------+---------------+ Generic Options The following options are supported by {{dfsadmin}}, {{fs}}, {{fsck}}, {{job}} and {{fetchdt}}. Applications should implement {{{../../api/org/apache/hadoop/util/Tool.html}Tool}} to support GenericOptions. *------------------------------------------------+-----------------------------+ || GENERIC_OPTION || Description *------------------------------------------------+-----------------------------+ |<<<-conf \ >>> | Specify an application | configuration file. *------------------------------------------------+-----------------------------+ |<<<-D \=\ >>> | Use value for given property. *------------------------------------------------+-----------------------------+ |<<<-jt \ or \>>> | Specify a ResourceManager. | Applies only to job. *------------------------------------------------+-----------------------------+ |<<<-files \ >>> | Specify comma separated files | to be copied to the map | reduce cluster. Applies only | to job. *------------------------------------------------+-----------------------------+ |<<<-libjars \ >>>| Specify comma separated jar | files to include in the | classpath. Applies only to | job. *------------------------------------------------+-----------------------------+ |<<<-archives \ >>> | Specify comma separated | archives to be unarchived on | the compute machines. Applies | only to job. *------------------------------------------------+-----------------------------+ User Commands Commands useful for users of a hadoop cluster. * <<>> Creates a hadoop archive. More information can be found at {{{../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/HadoopArchives.html} Hadoop Archives Guide}}. * <<>> Command to manage credentials, passwords and secrets within credential providers. The CredentialProvider API in Hadoop allows for the separation of applications and how they store their required passwords/secrets. In order to indicate a particular provider type and location, the user must provide the configuration element in core-site.xml or use the command line option <<<-provider>>> on each of the following commands. This provider path is a comma-separated list of URLs that indicates the type and location of a list of providers that should be consulted. For example, the following path: <<>> indicates that the current user's credentials file should be consulted through the User Provider, that the local file located at <<>> is a Java Keystore Provider and that the file located within HDFS at <<>> is also a store for a Java Keystore Provider. When utilizing the credential command it will often be for provisioning a password or secret to a particular credential store provider. In order to explicitly indicate which provider store to use the <<<-provider>>> option should be used. Otherwise, given a path of multiple providers, the first non-transient provider will be used. This may or may not be the one that you intended. Example: <<<-provider jceks://file/tmp/test.jceks>>> Usage: << [options]>>> *-------------------+-------------------------------------------------------+ ||COMMAND_OPTION || Description *-------------------+-------------------------------------------------------+ | create [-v ][-provider ]| Prompts the user for | a credential to be stored as the given alias when a value | is not provided via <<<-v>>>. The | within the | core-site.xml file will be used unless a <<<-provider>>> is | indicated. *-------------------+-------------------------------------------------------+ | delete [-i][-provider ] | Deletes the credential with | the provided alias and optionally warns the user when | <<<--interactive>>> is used. | The within the | core-site.xml file will be used unless a <<<-provider>>> is | indicated. *-------------------+-------------------------------------------------------+ | list [-provider ] | Lists all of the credential aliases | The within the | core-site.xml file will be used unless a <<<-provider>>> is | indicated. *-------------------+-------------------------------------------------------+ * <<>> Copy file or directories recursively. More information can be found at {{{../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/DistCp.html} Hadoop DistCp Guide}}. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#dfs}<<>>}} instead. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#fsck}<<>>}} instead. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#fetchdt} <<>>}} instead. * <<>> Runs a jar file. Users can bundle their Map Reduce code in a jar file and execute it using this command. Usage: << [mainClass] args...>>> The streaming jobs are run via this command. Examples can be referred from Streaming examples Word count example is also run using jar command. It can be referred from Wordcount example * <<>> Deprecated. Use {{{../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapredCommands.html#job} <<>>}} instead. * <<>> Deprecated. Use {{{../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapredCommands.html#pipes} <<>>}} instead. * <<>> Deprecated. Use {{{../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapredCommands.html#queue} <<>>}} instead. * <<>> Prints the version. Usage: <<>> * <<>> hadoop script can be used to invoke any class. Usage: <<>> Runs the class named <<>>. * <<>> Prints the class path needed to get the Hadoop jar and the required libraries. If called without arguments, then prints the classpath set up by the command scripts, which is likely to contain wildcards in the classpath entries. Additional options print the classpath after wildcard expansion or write the classpath into the manifest of a jar file. The latter is useful in environments where wildcards cannot be used and the expanded classpath exceeds the maximum supported command line length. Usage: <<|-h|--help]>>> *-----------------+-----------------------------------------------------------+ || COMMAND_OPTION || Description *-----------------+-----------------------------------------------------------+ | --glob | expand wildcards *-----------------+-----------------------------------------------------------+ | --jar | write classpath as manifest in jar named *-----------------+-----------------------------------------------------------+ | -h, --help | print help *-----------------+-----------------------------------------------------------+ Administration Commands Commands useful for administrators of a hadoop cluster. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#balancer} <<>>}} instead. * <<>> Get/Set the log level for each daemon. Usage: << >>> Usage: << >>> *------------------------------+-----------------------------------------------------------+ || COMMAND_OPTION || Description *------------------------------+-----------------------------------------------------------+ | -getlevel | Prints the log level of the daemon running at | . This command internally connects | to http:///logLevel?log= *------------------------------+-----------------------------------------------------------+ | -setlevel | Sets the log level of the daemon | running at . This command internally | connects to http:///logLevel?log= *------------------------------+-----------------------------------------------------------+ * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#datanode} <<>>}} instead. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#dfsadmin} <<>>}} instead. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#namenode} <<>>}} instead. * <<>> Deprecated, use {{{../hadoop-hdfs/HDFSCommands.html#secondarynamenode} <<>>}} instead.