#!/bin/bash # # Run a shell command on all slave hosts. # # Environment Variables # # HADOOP_SLAVES File naming remote hosts. # Default is ${HADOOP_CONF_DIR}/slaves. # HADOOP_CONF_DIR Alternate conf dir. Default is ${HADOOP_HOME}/conf. # HADOOP_SLAVE_SLEEP Seconds to sleep between spawning remote commands. # HADOOP_SSH_OPTS Options passed to ssh when running remote commands. ## usage="Usage: slaves.sh command..." # if no args specified, show usage if [ $# -le 0 ]; then echo $usage exit 1 fi # resolve links - $0 may be a softlink this="$0" while [ -h "$this" ]; do ls=`ls -ld "$this"` link=`expr "$ls" : '.*-> \(.*\)$'` if expr "$link" : '.*/.*' > /dev/null; then this="$link" else this=`dirname "$this"`/"$link" fi done # the root of the Hadoop installation HADOOP_HOME=`dirname "$this"`/.. # Allow alternate conf dir location. HADOOP_CONF_DIR="${HADOOP_CONF_DIR:=$HADOOP_HOME/conf}" if [ -f "${HADOOP_CONF_DIR}/hadoop-env.sh" ]; then source "${HADOOP_CONF_DIR}/hadoop-env.sh" fi if [ "$HADOOP_SLAVES" = "" ]; then export HADOOP_SLAVES="${HADOOP_CONF_DIR}/slaves" fi for slave in `cat "$HADOOP_SLAVES"`; do ssh $HADOOP_SSH_OPTS $slave $"${@// /\\ }" \ 2>&1 | sed "s/^/$slave: /" & if [ "$HADOOP_SLAVE_SLEEP" != "" ]; then sleep $HADOOP_SLAVE_SLEEP fi done wait