Forráskód Böngészése

HADOOP-8925. Remove the packaging. Contributed by Eli Collins

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/branches/branch-2@1399901 13f79535-47bb-0310-9956-ffa450edef68
Eli Collins 12 éve
szülő
commit
a38d2b8c78
64 módosított fájl, 2 hozzáadás és 6725 törlés
  1. 2 0
      hadoop-common-project/hadoop-common/CHANGES.txt
  2. 0 15
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/conffile
  3. 0 24
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/control
  4. 0 24
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/postinst
  5. 0 19
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/postrm
  6. 0 18
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/preinst
  7. 0 25
      hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/prerm
  8. 0 151
      hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-datanode
  9. 0 143
      hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-jobtracker
  10. 0 155
      hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-namenode
  11. 0 143
      hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-tasktracker
  12. 0 123
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-create-user.sh
  13. 0 142
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-applications.sh
  14. 0 707
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-conf.sh
  15. 0 157
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-hdfs.sh
  16. 0 219
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-single-node.sh
  17. 0 183
      hadoop-common-project/hadoop-common/src/main/packages/hadoop-validate-setup.sh
  18. 0 93
      hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-datanode
  19. 0 85
      hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-jobtracker
  20. 0 99
      hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-namenode
  21. 0 85
      hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-tasktracker
  22. 0 174
      hadoop-common-project/hadoop-common/src/main/packages/rpm/spec/hadoop.spec
  23. 0 191
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/capacity-scheduler.xml
  24. 0 22
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/commons-logging.properties
  25. 0 93
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/core-site.xml
  26. 0 75
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-env.sh
  27. 0 20
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-metrics2.properties
  28. 0 235
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-policy.xml
  29. 0 241
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hdfs-site.xml
  30. 0 212
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/log4j.properties
  31. 0 26
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/mapred-queue-acls.xml
  32. 0 300
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/mapred-site.xml
  33. 0 18
      hadoop-common-project/hadoop-common/src/main/packages/templates/conf/taskcontroller.cfg
  34. 0 168
      hadoop-common-project/hadoop-common/src/main/packages/update-hadoop-env.sh
  35. 0 15
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/conffile
  36. 0 24
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/control
  37. 0 24
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/postinst
  38. 0 19
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/postrm
  39. 0 18
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/preinst
  40. 0 25
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/prerm
  41. 0 142
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-datanode
  42. 0 154
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-namenode
  43. 0 142
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-secondarynamenode
  44. 0 84
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-datanode
  45. 0 98
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-namenode
  46. 0 92
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-secondarynamenode
  47. 0 176
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/spec/hadoop-hdfs.spec
  48. 0 45
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/templates/conf/hdfs-site.xml
  49. 0 128
      hadoop-hdfs-project/hadoop-hdfs/src/main/packages/update-hdfs-env.sh
  50. 0 15
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/conffile
  51. 0 24
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/control
  52. 0 24
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/postinst
  53. 0 19
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/postrm
  54. 0 18
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/preinst
  55. 0 25
      hadoop-mapreduce-project/src/packages/deb/hadoop.control/prerm
  56. 0 143
      hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-historyserver
  57. 0 142
      hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-jobtracker
  58. 0 142
      hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-tasktracker
  59. 0 85
      hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-historyserver
  60. 0 84
      hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-jobtracker
  61. 0 84
      hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-tasktracker
  62. 0 179
      hadoop-mapreduce-project/src/packages/rpm/spec/hadoop-mapred.spec
  63. 0 31
      hadoop-mapreduce-project/src/packages/templates/conf/mapred-site.xml
  64. 0 139
      hadoop-mapreduce-project/src/packages/update-mapred-env.sh

+ 2 - 0
hadoop-common-project/hadoop-common/CHANGES.txt

@@ -56,6 +56,8 @@ Release 2.0.3-alpha - Unreleased
 
     HADOOP-8931. Add Java version to startup message. (eli)
 
+    HADOOP-8925. Remove the packaging. (eli)
+
   OPTIMIZATIONS
 
     HADOOP-8866. SampleQuantiles#query is O(N^2) instead of O(N). (Andrew Wang

+ 0 - 15
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/conffile

@@ -1,15 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-/etc/hadoop

+ 0 - 24
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/control

@@ -1,24 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-Package: hadoop-common
-Version: @version@
-Section: misc
-Priority: optional
-Provides: hadoop-common
-Architecture: all
-Depends: openjdk-6-jre-headless
-Maintainer: Apache Software Foundation <general@hadoop.apache.org>
-Description: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing.
-Distribution: development

+ 0 - 24
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/postinst

@@ -1,24 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-hadoop-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop

+ 0 - 19
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/postrm

@@ -1,19 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-/usr/sbin/groupdel hadoop 2> /dev/null >dev/null
-exit 0

+ 0 - 18
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/preinst

@@ -1,18 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-getent group hadoop 2>/dev/null >/dev/null || /usr/sbin/groupadd -g 123 -r hadoop

+ 0 - 25
hadoop-common-project/hadoop-common/src/main/packages/deb/hadoop.control/prerm

@@ -1,25 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-hadoop-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop \
-  --uninstal

+ 0 - 151
hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-datanode

@@ -1,151 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-datanode	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Name Node server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-datanode: start and stop the Apache Hadoop Data Node daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-if [ -n "$HADOOP_SECURE_DN_USER" ]; then
-  DN_USER="root"
-  IDENT_USER=${HADOOP_SECURE_DN_USER}
-else
-  DN_USER="hdfs"
-  IDENT_USER=${DN_USER}
-fi
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-datanode_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-datanode_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Data Node server not in use (/etc/hadoop/hadoop-datanode_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Data Node server" "hadoop-datanode"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid -c ${DN_USER} -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Data Node server" "hadoop-datanode"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Data Node server" "hadoop-datanode"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid -c ${DN_USER} -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Data Node server" "hadoop-datanode"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid -c ${DN_USER} -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-${IDENT_USER}-datanode.pid ${JAVA_HOME}/bin/java hadoop-datanode && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-datanode {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 143
hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-jobtracker

@@ -1,143 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-jobtracker	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Job Tracker server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-jobtracker: start and stop the Apache Hadoop Job Tracker daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-jobtracker_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-jobtracker_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Job Tracker server not in use (/etc/hadoop/hadoop-jobtracker_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid ${JAVA_HOME}/bin/java hadoop-jobtracker && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-jobtracker {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 155
hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-namenode

@@ -1,155 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-namenode	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Name Node server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-namenode: start and stop the Apache Hadoop Name Node daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-namenode_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-namenode_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Name Node server not in use (/etc/hadoop/hadoop-namenode_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-format() {
-    sudo -u hdfs ${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} namenode -format
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Name Node server" "hadoop-namenode"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Name Node server" "hadoop-namenode"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  format)
-	log_daemon_msg "Formatting Apache Hadoop Name Node" "hadoop-namenode"
-	format
-	if [ $? -eq 0 ]; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Name Node server" "hadoop-namenode"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Name Node server" "hadoop-namenode"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid ${JAVA_HOME}/bin/java hadoop-namenode && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-namenode {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 143
hadoop-common-project/hadoop-common/src/main/packages/deb/init.d/hadoop-tasktracker

@@ -1,143 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-tasktracker	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Task Tracker server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-tasktracker: start and stop the Apache Hadoop Task Tracker daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-tasktracker_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-tasktracker_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Task Tracker server not in use (/etc/hadoop/hadoop-tasktracker_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid ${JAVA_HOME}/bin/java hadoop-tasktracker && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-tasktracker {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 123
hadoop-common-project/hadoop-common/src/main/packages/hadoop-create-user.sh

@@ -1,123 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-this="${BASH_SOURCE-$0}"
-bin=$(cd -P -- "$(dirname -- "$this")" && pwd -P)
-script="$(basename -- "$this")"
-this="$bin/$script"
-
-if [ "$HADOOP_HOME" != "" ]; then
-  echo "Warning: \$HADOOP_HOME is deprecated."
-  echo
-fi
-
-DEFAULT_LIBEXEC_DIR="$bin"/../libexec
-HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
-. $HADOOP_LIBEXEC_DIR/hadoop-config.sh
-
-usage() {
-  echo "
-usage: $0 <parameters>
-  Require parameter:
-     --config /etc/hadoop                                  Location of Hadoop configuration file
-     -u <username>                                         Create user on HDFS
-  Optional parameters:
-     -h                                                    Display this message
-     --kerberos-realm=KERBEROS.EXAMPLE.COM                 Set Kerberos realm
-     --super-user=hdfs                                     Set super user id
-     --super-user-keytab=/etc/security/keytabs/hdfs.keytab Set super user keytab location
-  "
-  exit 1
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'kerberos-realm:' \
-  -l 'super-user:' \
-  -l 'super-user-keytab:' \
-  -o 'h' \
-  -o 'u' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-    exit 1
-fi
-
-create_user() {
-  if [ "${SETUP_USER}" = "" ]; then
-    break
-  fi
-  HADOOP_HDFS_USER=${HADOOP_HDFS_USER:-hdfs}
-  export HADOOP_PREFIX
-  export HADOOP_CONF_DIR
-  export JAVA_HOME
-  export SETUP_USER=${SETUP_USER}
-  export SETUP_PATH=/user/${SETUP_USER}
-
-  if [ ! "${KERBEROS_REALM}" = "" ]; then
-    # locate kinit cmd
-    if [ -e /etc/lsb-release ]; then
-      KINIT_CMD="/usr/bin/kinit -kt ${HDFS_USER_KEYTAB} ${HADOOP_HDFS_USER}"
-    else
-      KINIT_CMD="/usr/kerberos/bin/kinit -kt ${HDFS_USER_KEYTAB} ${HADOOP_HDFS_USER}"
-    fi
-    su -c "${KINIT_CMD}" ${HADOOP_HDFS_USER} 
-  fi
-
-  su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} fs -mkdir ${SETUP_PATH}" ${HADOOP_HDFS_USER}
-  su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} fs -chown ${SETUP_USER}:${SETUP_USER} ${SETUP_PATH}" ${HADOOP_HDFS_USER}
-  su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} fs -chmod 711 ${SETUP_PATH}" ${HADOOP_HDFS_USER}
-
-  if [ "$?" == "0" ]; then
-    echo "User directory has been setup: ${SETUP_PATH}"
-  fi
-}
-
-eval set -- "${OPTS}"
-while true; do
-  case "$1" in
-    -u)
-      shift
-      ;;
-    --kerberos-realm)
-      KERBEROS_REALM=$2; shift 2
-      ;;
-    --super-user)
-      HADOOP_HDFS_USER=$2; shift 2
-      ;;
-    --super-user-keytab)
-      HDFS_USER_KEYTAB=$2; shift 2
-      ;;
-    -h)
-      usage
-      ;; 
-    --)
-      while shift; do
-        SETUP_USER=$1
-        create_user
-      done
-      break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1 
-      ;;
-  esac
-done 
-

+ 0 - 142
hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-applications.sh

@@ -1,142 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-this="${BASH_SOURCE-$0}"
-bin=$(cd -P -- "$(dirname -- "$this")" && pwd -P)
-script="$(basename -- "$this")"
-this="$bin/$script"
-
-DEFAULT_LIBEXEC_DIR="$bin"/../libexec
-HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
-. $HADOOP_LIBEXEC_DIR/hadoop-config.sh
-
-usage() {
-  echo "
-usage: $0 <parameters>
-  Require parameter:
-     --config /etc/hadoop                                  Location of Hadoop configuration file
-     --apps=<csl of apps:user hcat:hcat,hbase,hive:user>   Apps you want to setup on hdfs
-                                                           If user is not specified, app name
-                                                           will be used as the user name as well
-  Optional parameters:
-     -h                                                    Display this message
-     --kerberos-realm=KERBEROS.EXAMPLE.COM                 Set Kerberos realm
-     --super-user=hdfs                                     Set super user id
-     --super-user-keytab=/etc/security/keytabs/hdfs.keytab Set super user keytab location
-  "
-  exit 1
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'kerberos-realm:' \
-  -l 'super-user:' \
-  -l 'super-user-keytab:' \
-  -l 'apps:' \
-  -o 'h' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-    exit 1
-fi
-
-function setup_apps
-{
-  if [ -z $APPS ] 
-  then
-    usage
-    break
-  fi
-
-  #if super user is not set default to hdfs
-  HADOOP_HDFS_USER=${HADOOP_HDFS_USER:-hdfs}
-
-  if [ ! "${KERBEROS_REALM}" = "" ]; then
-    # locate kinit cmd
-    if [ -e /etc/lsb-release ]; then
-      KINIT_CMD="/usr/bin/kinit -kt ${HDFS_USER_KEYTAB} ${HADOOP_HDFS_USER}"
-    else
-      KINIT_CMD="/usr/kerberos/bin/kinit -kt ${HDFS_USER_KEYTAB} ${HADOOP_HDFS_USER}"
-    fi
-    su -c "${KINIT_CMD}" ${HADOOP_HDFS_USER}
-  fi
-  #process each app
-  oldIFS=$IFS 
-  IFS=','
-  for app in $APPS
-  do
-    IFS=":"
-    arr=($app)
-    app=${arr[0]}
-    user=${arr[1]}
-    IFS=','
-    #if user is empty, default it to app
-    if [ -z $user ]
-    then
-      user=$app
-    fi
-    
-    path="/apps/${app}"
-
-    #create the dir
-    cmd="su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir ${path}' ${HADOOP_HDFS_USER}"
-    echo $cmd
-    eval $cmd
-
-    #make owner to be the app
-    cmd="su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chown ${user} ${path}' ${HADOOP_HDFS_USER}"
-    echo $cmd
-    eval $cmd
-
-    if [ "$?" == "0" ]; then
-      echo "App directory has been setup: ${path}"
-    fi
-  done
-  IFS=$oldIFS
-}
-
-eval set -- "${OPTS}"
-while true; do
-  case "$1" in
-    --apps)
-      APPS=$2; shift 2
-      ;;
-    --kerberos-realm)
-      KERBEROS_REALM=$2; shift 2
-      ;;
-    --super-user)
-      HADOOP_HDFS_USER=$2; shift 2
-      ;;
-    --super-user-keytab)
-      HDFS_USER_KEYTAB=$2; shift 2
-      ;;
-    -h)
-      usage
-      ;; 
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1 
-      ;;
-  esac
-done
-
-setup_apps

+ 0 - 707
hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-conf.sh

@@ -1,707 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bin=`dirname "$0"`
-bin=`cd "$bin"; pwd`
-
-this="${BASH_SOURCE-$0}"
-HADOOP_DEFAULT_PREFIX=`dirname "$this"`/..
-HADOOP_PREFIX=${HADOOP_PREFIX:-$HADOOP_DEFAULT_PREFIX}
-export HADOOP_PREFIX
-
-usage() {
-  echo "
-usage: $0 <parameters>
-
-  Optional parameters:
-     --auto                                                          Setup path and configuration automatically
-     --default                                                       Setup configuration as default
-     --conf-dir=/etc/hadoop                                          Set configuration directory
-     --datanode-dir=/var/lib/hadoop/hdfs/datanode                    Set datanode directory
-     --group=hadoop                                                  Set Hadoop group name
-     -h                                                              Display this message
-     --hdfs-user=hdfs                                                Set HDFS user
-     --jobtracker-host=hostname                                      Set jobtracker host
-     --namenode-host=hostname                                        Set namenode host
-     --secondarynamenode-host=hostname                               Set secondary namenode host
-     --kerberos-realm=KERBEROS.EXAMPLE.COM                           Set Kerberos realm
-     --kinit-location=/usr/kerberos/bin/kinit                        Set kinit location
-     --keytab-dir=/etc/security/keytabs                              Set keytab directory
-     --log-dir=/var/log/hadoop                                       Set log directory
-     --pid-dir=/var/run/hadoop                                       Set pid directory
-     --hdfs-dir=/var/lib/hadoop/hdfs                                 Set HDFS directory
-     --hdfs-user-keytab=/home/hdfs/hdfs.keytab                       Set HDFS user key tab
-     --mapred-dir=/var/lib/hadoop/mapred                             Set mapreduce directory
-     --mapreduce-user=mr                                             Set mapreduce user
-     --mapreduce-user-keytab=/home/mr/hdfs.keytab                    Set mapreduce user key tab
-     --namenode-dir=/var/lib/hadoop/hdfs/namenode                    Set namenode directory
-     --replication=3                                                 Set replication factor
-     --taskscheduler=org.apache.hadoop.mapred.JobQueueTaskScheduler  Set task scheduler
-     --datanodes=hostname1,hostname2,...                             SET the datanodes
-     --tasktrackers=hostname1,hostname2,...                          SET the tasktrackers
-     --dfs-webhdfs-enabled=false|true                                Enable webhdfs
-     --dfs-support-append=false|true                                 Enable append
-     --hadoop-proxy-users='user1:groups:hosts;user2:groups:hosts'    Setup proxy users for hadoop
-     --hbase-user=hbase                                              User which hbase is running as. Defaults to hbase
-     --mapreduce-cluster-mapmemory-mb=memory                         Virtual memory of a map slot for the MR framework. Defaults to -1
-     --mapreduce-cluster-reducememory-mb=memory                      Virtual memory, of a reduce slot for the MR framework. Defaults to -1
-     --mapreduce-jobtracker-maxmapmemory-mb=memory                   Maximum virtual memory of a single map task. Defaults to -1
-                                                                     This value should be set to (mapreduce.cluster.mapmemory.mb * mapreduce.tasktracker.map.tasks.maximum)
-     --mapreduce-jobtracker-maxreducememory-mb=memory                Maximum virtual memory of a single reduce task. Defaults to -1
-                                                                     This value should be set to (mapreduce.cluster.reducememory.mb * mapreduce.tasktracker.reduce.tasks.maximum)
-     --mapreduce-map-memory-mb=memory                                Virtual memory of a single map slot for a job. Defaults to -1
-                                                                     This value should be <= mapred.cluster.max.map.memory.mb
-     --mapreduce-reduce-memory-mb=memory                             Virtual memory, of a single reduce slot for a job. Defaults to -1
-                                                                     This value should be <= mapred.cluster.max.reduce.memory.mb
-     --dfs-datanode-dir-perm=700                                     Set the permission for the datanode data directories. Defaults to 700
-     --dfs-block-local-path-access-user=user                         User for which you want to enable shortcircuit read.
-     --dfs-client-read-shortcircuit=true/false                       Enable shortcircuit read for the client. Will default to true if the shortcircuit user is set.
-     --dfs-client-read-shortcircuit-skip-checksum=false/true         Disable checking of checksum when shortcircuit read is taking place. Defaults to false.
-  "
-  exit 1
-}
-
-check_permission() {
-  TARGET=$1
-  OWNER="0"
-  RESULT=0
-  while [ "$TARGET" != "/" ]; do
-    if [ "`uname`" = "Darwin" ]; then
-      OWNER=`stat -f %u $TARGET`
-    else
-      OWNER=`stat -c %u $TARGET`
-    fi
-    if [ "$OWNER" != "0" ]; then
-      RESULT=1
-      break
-    fi
-    TARGET=`dirname $TARGET`
-  done
-  return $RESULT
-}
-
-template_generator() {
-  REGEX='(\$\{[a-zA-Z_][a-zA-Z_0-9]*\})'
-  if [ -e $2 ]; then
-    mv -f $2 "$2.bak"
-  fi
-  cat $1 |
-  while read line ; do
-    while [[ "$line" =~ $REGEX ]] ; do
-      LHS=${BASH_REMATCH[1]}
-      RHS="$(eval echo "\"$LHS\"")"
-      line=${line//$LHS/$RHS}
-    done
-    echo $line >> $2
-  done
-}
-
-#########################################
-# Function to modify a value of a field in an xml file
-# Params: $1 is the file with full path; $2 is the property, $3 is the new value
-#########################################
-function addPropertyToXMLConf
-{
-  #read the file name with full path
-  local file=$1
-  #get the property name
-  local property=$2
-  #get what value should be set for that
-  local propValue=$3
-  #get the description
-  local desc=$4
-  #get the value for the final tag
-  local finalVal=$5
-
-  #create the property text, make sure the / are escaped
-  propText="<property>\n<name>$property<\/name>\n<value>$propValue<\/value>\n"
-  #if description is not empty add it
-  if [ ! -z $desc ]
-  then
-    propText="${propText}<description>$desc<\/description>\n"
-  fi
-  
-  #if final is not empty add it
-  if [ ! -z $finalVal ]
-  then
-    propText="${propText}final>$finalVal<\/final>\n"
-  fi
-
-  #add the ending tag
-  propText="${propText}<\/property>\n"
-
-  #add the property to the file
-  endText="<\/configuration>"
-  #add the text using sed at the end of the file
-  sed -i "s|$endText|$propText$endText|" $file
-}
-
-##########################################
-# Function to setup up the short circuit read settings
-#########################################
-function setupShortCircuitRead
-{
-  local conf_file="${HADOOP_CONF_DIR}/hdfs-site.xml"
-  #if the shortcircuit user is not set then return
-  if [ -z $DFS_BLOCK_LOCAL_PATH_ACCESS_USER ]
-  then
-    return
-  fi
-  
-  #set the defaults if values not present
-  DFS_CLIENT_READ_SHORTCIRCUIT=${DFS_CLIENT_READ_SHORTCIRCUIT:-false}
-  DFS_CLIENT_READ_SHORTCIRCUIT_SKIP_CHECKSUM=${DFS_CLIENT_READ_SHORTCIRCUIT_SKIP_CHECKSUM:-false}
-
-  #add the user to the conf file
-  addPropertyToXMLConf "$conf_file" "dfs.block.local-path-access.user" "$DFS_BLOCK_LOCAL_PATH_ACCESS_USER"
-  addPropertyToXMLConf "$conf_file" "dfs.client.read.shortcircuit" "$DFS_CLIENT_READ_SHORTCIRCUIT"
-  addPropertyToXMLConf "$conf_file" "dfs.client.read.shortcircuit.skip.checksum" "$DFS_CLIENT_READ_SHORTCIRCUIT_SKIP_CHECKSUM"
-}
-
-##########################################
-# Function to setup up the proxy user settings
-#########################################
-function setupProxyUsers
-{
-  local conf_file="${HADOOP_CONF_DIR}/core-site.xml"
-  #if hadoop proxy users are sent, setup hadoop proxy
-  if [ ! -z $HADOOP_PROXY_USERS ]
-  then
-    oldIFS=$IFS
-    IFS=';'
-    #process each proxy config
-    for proxy in $HADOOP_PROXY_USERS
-    do
-      #get the user, group and hosts information for each proxy
-      IFS=':'
-      arr=($proxy)
-      user="${arr[0]}"
-      groups="${arr[1]}"
-      hosts="${arr[2]}"
-      #determine the property names and values
-      proxy_groups_property="hadoop.proxyuser.${user}.groups"
-      proxy_groups_val="$groups"
-      addPropertyToXMLConf "$conf_file" "$proxy_groups_property" "$proxy_groups_val"
-      proxy_hosts_property="hadoop.proxyuser.${user}.hosts"
-      proxy_hosts_val="$hosts"
-      addPropertyToXMLConf "$conf_file" "$proxy_hosts_property" "$proxy_hosts_val"
-      IFS=';'
-    done
-    IFS=$oldIFS
-  fi
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'auto' \
-  -l 'java-home:' \
-  -l 'conf-dir:' \
-  -l 'default' \
-  -l 'group:' \
-  -l 'hdfs-dir:' \
-  -l 'namenode-dir:' \
-  -l 'datanode-dir:' \
-  -l 'mapred-dir:' \
-  -l 'namenode-host:' \
-  -l 'secondarynamenode-host:' \
-  -l 'jobtracker-host:' \
-  -l 'log-dir:' \
-  -l 'pid-dir:' \
-  -l 'replication:' \
-  -l 'taskscheduler:' \
-  -l 'hdfs-user:' \
-  -l 'hdfs-user-keytab:' \
-  -l 'mapreduce-user:' \
-  -l 'mapreduce-user-keytab:' \
-  -l 'keytab-dir:' \
-  -l 'kerberos-realm:' \
-  -l 'kinit-location:' \
-  -l 'datanodes:' \
-  -l 'tasktrackers:' \
-  -l 'dfs-webhdfs-enabled:' \
-  -l 'hadoop-proxy-users:' \
-  -l 'dfs-support-append:' \
-  -l 'hbase-user:' \
-  -l 'mapreduce-cluster-mapmemory-mb:' \
-  -l 'mapreduce-cluster-reducememory-mb:' \
-  -l 'mapreduce-jobtracker-maxmapmemory-mb:' \
-  -l 'mapreduce-jobtracker-maxreducememory-mb:' \
-  -l 'mapreduce-map-memory-mb:' \
-  -l 'mapreduce-reduce-memory-mb:' \
-  -l 'dfs-datanode-dir-perm:' \
-  -l 'dfs-block-local-path-access-user:' \
-  -l 'dfs-client-read-shortcircuit:' \
-  -l 'dfs-client-read-shortcircuit-skip-checksum:' \
-  -o 'h' \
-  -- "$@") 
-  
-if [ $? != 0 ] ; then
-    usage
-fi
-
-# Make sure the HADOOP_LOG_DIR is not picked up from user environment.
-unset HADOOP_LOG_DIR
-  
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --auto)
-      AUTOSETUP=1
-      AUTOMATED=1
-      shift
-      ;; 
-    --java-home)
-      JAVA_HOME=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --conf-dir)
-      HADOOP_CONF_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --default)
-      AUTOMATED=1; shift
-      ;;
-    --group)
-      HADOOP_GROUP=$2; shift 2
-      AUTOMATED=1
-      ;;
-    -h)
-      usage
-      ;; 
-    --hdfs-dir)
-      HADOOP_HDFS_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --namenode-dir)
-      HADOOP_NN_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --datanode-dir)
-      HADOOP_DN_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --mapred-dir)
-      HADOOP_MAPRED_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --namenode-host)
-      HADOOP_NN_HOST=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --secondarynamenode-host)
-      HADOOP_SNN_HOST=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --jobtracker-host)
-      HADOOP_JT_HOST=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --log-dir)
-      HADOOP_LOG_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --pid-dir)
-      HADOOP_PID_DIR=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --replication)
-      HADOOP_REPLICATION=$2; shift 2
-      AUTOMATED=1
-      ;; 
-    --taskscheduler)
-      HADOOP_TASK_SCHEDULER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hdfs-user)
-      HADOOP_HDFS_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-user)
-      HADOOP_MR_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --keytab-dir)
-      KEYTAB_DIR=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hdfs-user-keytab)
-      HDFS_KEYTAB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-user-keytab)
-      MR_KEYTAB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --kerberos-realm)
-      KERBEROS_REALM=$2; shift 2
-      SECURITY_TYPE="kerberos"
-      AUTOMATED=1
-      ;;
-    --kinit-location)
-      KINIT=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --datanodes)
-      DATANODES=$2; shift 2
-      AUTOMATED=1
-      DATANODES=$(echo $DATANODES | tr ',' ' ')
-      ;;
-    --tasktrackers)
-      TASKTRACKERS=$2; shift 2
-      AUTOMATED=1
-      TASKTRACKERS=$(echo $TASKTRACKERS | tr ',' ' ')
-      ;;
-    --dfs-webhdfs-enabled)
-      DFS_WEBHDFS_ENABLED=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hadoop-proxy-users)
-      HADOOP_PROXY_USERS=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --dfs-support-append)
-      DFS_SUPPORT_APPEND=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hbase-user)
-      HBASE_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-cluster-mapmemory-mb)
-      MAPREDUCE_CLUSTER_MAPMEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-cluster-reducememory-mb)
-      MAPREDUCE_CLUSTER_REDUCEMEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-jobtracker-maxmapmemory-mb)
-      MAPREDUCE_JOBTRACKER_MAXMAPMEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-jobtracker-maxreducememory-mb)
-      MAPREDUCE_JOBTRACKER_MAXREDUCEMEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-map-memory-mb)
-      MAPREDUCE_MAP_MEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-reduce-memory-mb)
-      MAPREDUCE_REDUCE_MEMORY_MB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --dfs-datanode-dir-perm)
-      DFS_DATANODE_DIR_PERM=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --dfs-block-local-path-access-user)
-      DFS_BLOCK_LOCAL_PATH_ACCESS_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --dfs-client-read-shortcircuit)
-      DFS_CLIENT_READ_SHORTCIRCUIT=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --dfs-client-read-shortcircuit-skip-checksum)
-      DFS_CLIENT_READ_SHORTCIRCUIT_SKIP_CHECKSUM=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1 
-      ;;
-  esac
-done 
-
-AUTOSETUP=${AUTOSETUP:-1}
-JAVA_HOME=${JAVA_HOME:-/usr/java/default}
-HADOOP_GROUP=${HADOOP_GROUP:-hadoop}
-HADOOP_NN_HOST=${HADOOP_NN_HOST:-`hostname`}
-HADOOP_SNN_HOST=${HADOOP_SNN_HOST:-`hostname`}
-HADOOP_NN_DIR=${HADOOP_NN_DIR:-/var/lib/hadoop/hdfs/namenode}
-HADOOP_DN_DIR=${HADOOP_DN_DIR:-/var/lib/hadoop/hdfs/datanode}
-HADOOP_JT_HOST=${HADOOP_JT_HOST:-`hostname`}
-HADOOP_HDFS_DIR=${HADOOP_HDFS_DIR:-/var/lib/hadoop/hdfs}
-HADOOP_MAPRED_DIR=${HADOOP_MAPRED_DIR:-/var/lib/hadoop/mapred}
-HADOOP_LOG_DIR=${HADOOP_LOG_DIR:-/var/log/hadoop}
-HADOOP_PID_DIR=${HADOOP_PID_DIR:-/var/log/hadoop}
-HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-/etc/hadoop}
-HADOOP_REPLICATION=${HADOOP_RELICATION:-3}
-HADOOP_TASK_SCHEDULER=${HADOOP_TASK_SCHEDULER:-org.apache.hadoop.mapred.JobQueueTaskScheduler}
-HADOOP_HDFS_USER=${HADOOP_HDFS_USER:-hdfs}
-HADOOP_MR_USER=${HADOOP_MR_USER:-mr}
-DFS_WEBHDFS_ENABLED=${DFS_WEBHDFS_ENABLED:-false}
-DFS_SUPPORT_APPEND=${DFS_SUPPORT_APPEND:-false}
-HBASE_USER=${HBASE_USER:-hbase}
-MAPREDUCE_CLUSTER_MAPMEMORY_MB=${MAPREDUCE_CLUSTER_MAPMEMORY_MB:--1}
-MAPREDUCE_CLUSTER_REDUCEMEMORY_MB=${MAPREDUCE_CLUSTER_REDUCEMEMORY_MB:--1}
-MAPREDUCE_JOBTRACKER_MAXMAPMEMORY_MB=${MAPREDUCE_JOBTRACKER_MAXMAPMEMORY_MB:--1}
-MAPREDUCE_JOBTRACKER_MAXREDUCEMEMORY_MB=${MAPREDUCE_JOBTRACKER_MAXREDUCEMEMORY_MB:--1}
-MAPREDUCE_MAP_MEMORY_MB=${MAPREDUCE_MAP_MEMORY_MB:--1}
-MAPREDUCE_REDUCE_MEMORY_MB=${MAPREDUCE_REDUCE_MEMORY_MB:--1}
-KEYTAB_DIR=${KEYTAB_DIR:-/etc/security/keytabs}
-HDFS_KEYTAB=${HDFS_KEYTAB:-/home/hdfs/hdfs.keytab}
-MR_KEYTAB=${MR_KEYTAB:-/home/mr/mr.keytab}
-DFS_WEBHDFS_ENABLED=${DFS_WEBHDFS_ENABLED:-false}
-DFS_SUPPORT_APPEND=${DFS_SUPPORT_APPEND:-false}
-KERBEROS_REALM=${KERBEROS_REALM:-KERBEROS.EXAMPLE.COM}
-SECURITY_TYPE=${SECURITY_TYPE:-simple}
-KINIT=${KINIT:-/usr/kerberos/bin/kinit}
-#deault the data dir perm to 700
-DFS_DATANODE_DIR_PERM=${DFS_DATANODE_DIR_PERM:-700}
-if [ "${SECURITY_TYPE}" = "kerberos" ]; then
-  TASK_CONTROLLER="org.apache.hadoop.mapred.LinuxTaskController"
-  HADOOP_DN_ADDR="0.0.0.0:1019"
-  HADOOP_DN_HTTP_ADDR="0.0.0.0:1022"
-  SECURITY="true"
-  HADOOP_SECURE_DN_USER=${HADOOP_HDFS_USER}
-else
-  TASK_CONTROLLER="org.apache.hadoop.mapred.DefaultTaskController"
-  HADOOP_DN_ADDR="0.0.0.0:50010"
-  HADOOP_DN_HTTP_ADDR="0.0.0.0:50075"
-  SECURITY="false"
-  HADOOP_SECURE_DN_USER=""
-fi
-
-#unset env vars
-unset HADOOP_CLIENT_OPTS HADOOP_NAMENODE_OPTS HADOOP_DATANODE_OPTS HADOOP_SECONDARYNAMENODE_OPTS HADOOP_JAVA_PLATFORM_OPTS
-
-if [ "${AUTOMATED}" != "1" ]; then
-  echo "Setup Hadoop Configuration"
-  echo
-  echo -n "Where would you like to put config directory? (${HADOOP_CONF_DIR}) "
-  read USER_HADOOP_CONF_DIR
-  echo -n "Where would you like to put log directory? (${HADOOP_LOG_DIR}) "
-  read USER_HADOOP_LOG_DIR
-  echo -n "Where would you like to put pid directory? (${HADOOP_PID_DIR}) "
-  read USER_HADOOP_PID_DIR
-  echo -n "What is the host of the namenode? (${HADOOP_NN_HOST}) "
-  read USER_HADOOP_NN_HOST
-  echo -n "Where would you like to put namenode data directory? (${HADOOP_NN_DIR}) "
-  read USER_HADOOP_NN_DIR
-  echo -n "Where would you like to put datanode data directory? (${HADOOP_DN_DIR}) "
-  read USER_HADOOP_DN_DIR
-  echo -n "What is the host of the jobtracker? (${HADOOP_JT_HOST}) "
-  read USER_HADOOP_JT_HOST
-  echo -n "Where would you like to put jobtracker/tasktracker data directory? (${HADOOP_MAPRED_DIR}) "
-  read USER_HADOOP_MAPRED_DIR
-  echo -n "Where is JAVA_HOME directory? (${JAVA_HOME}) "
-  read USER_JAVA_HOME
-  echo -n "Would you like to create directories/copy conf files to localhost? (Y/n) "
-  read USER_AUTOSETUP
-  echo
-  JAVA_HOME=${USER_USER_JAVA_HOME:-$JAVA_HOME}
-  HADOOP_NN_HOST=${USER_HADOOP_NN_HOST:-$HADOOP_NN_HOST}
-  HADOOP_NN_DIR=${USER_HADOOP_NN_DIR:-$HADOOP_NN_DIR}
-  HADOOP_DN_DIR=${USER_HADOOP_DN_DIR:-$HADOOP_DN_DIR}
-  HADOOP_JT_HOST=${USER_HADOOP_JT_HOST:-$HADOOP_JT_HOST}
-  HADOOP_HDFS_DIR=${USER_HADOOP_HDFS_DIR:-$HADOOP_HDFS_DIR}
-  HADOOP_MAPRED_DIR=${USER_HADOOP_MAPRED_DIR:-$HADOOP_MAPRED_DIR}
-  HADOOP_TASK_SCHEDULER=${HADOOP_TASK_SCHEDULER:-org.apache.hadoop.mapred.JobQueueTaskScheduler}
-  HADOOP_LOG_DIR=${USER_HADOOP_LOG_DIR:-$HADOOP_LOG_DIR}
-  HADOOP_PID_DIR=${USER_HADOOP_PID_DIR:-$HADOOP_PID_DIR}
-  HADOOP_CONF_DIR=${USER_HADOOP_CONF_DIR:-$HADOOP_CONF_DIR}
-  AUTOSETUP=${USER_AUTOSETUP:-y}
-  echo "Review your choices:"
-  echo
-  echo "Config directory            : ${HADOOP_CONF_DIR}"
-  echo "Log directory               : ${HADOOP_LOG_DIR}"
-  echo "PID directory               : ${HADOOP_PID_DIR}"
-  echo "Namenode host               : ${HADOOP_NN_HOST}"
-  echo "Namenode directory          : ${HADOOP_NN_DIR}"
-  echo "Datanode directory          : ${HADOOP_DN_DIR}"
-  echo "Jobtracker host             : ${HADOOP_JT_HOST}"
-  echo "Mapreduce directory         : ${HADOOP_MAPRED_DIR}"
-  echo "Task scheduler              : ${HADOOP_TASK_SCHEDULER}"
-  echo "JAVA_HOME directory         : ${JAVA_HOME}"
-  echo "Create dirs/copy conf files : ${AUTOSETUP}"
-  echo
-  echo -n "Proceed with generate configuration? (y/N) "
-  read CONFIRM
-  if [ "${CONFIRM}" != "y" ]; then
-    echo "User aborted setup, exiting..."
-    exit 1
-  fi
-fi
-
-if [ "${AUTOSETUP}" == "1" -o "${AUTOSETUP}" == "y" ]; then
-  if [ -d ${KEYTAB_DIR} ]; then
-    chmod 700 ${KEYTAB_DIR}/*
-    chown ${HADOOP_MR_USER}:${HADOOP_GROUP} ${KEYTAB_DIR}/[jt]t.service.keytab
-    chown ${HADOOP_HDFS_USER}:${HADOOP_GROUP} ${KEYTAB_DIR}/[dns]n.service.keytab
-  fi
-  chmod 755 -R ${HADOOP_PREFIX}/sbin/*hadoop*
-  chmod 755 -R ${HADOOP_PREFIX}/bin/hadoop
-
-  HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-${HADOOP_PREFIX}/libexec}
-  chmod 755 -R ${HADOOP_LIBEXEC_DIR}/hadoop-config.sh
-
-  mkdir -p /home/${HADOOP_MR_USER}
-  chown ${HADOOP_MR_USER}:${HADOOP_GROUP} /home/${HADOOP_MR_USER}
-  HDFS_DIR=`echo ${HADOOP_HDFS_DIR} | sed -e 's/,/ /g'`
-  mkdir -p ${HDFS_DIR}
-  if [ -e ${HADOOP_NN_DIR} ]; then
-    rm -rf ${HADOOP_NN_DIR}
-  fi
-  DATANODE_DIR=`echo ${HADOOP_DN_DIR} | sed -e 's/,/ /g'`
-  mkdir -p ${DATANODE_DIR}
-  MAPRED_DIR=`echo ${HADOOP_MAPRED_DIR} | sed -e 's/,/ /g'`
-  mkdir -p ${MAPRED_DIR}
-  mkdir -p ${HADOOP_CONF_DIR}
-  check_permission ${HADOOP_CONF_DIR}
-  if [ $? == 1 ]; then
-    echo "Full path to ${HADOOP_CONF_DIR} should be owned by root."
-    exit 1
-  fi
-
-  mkdir -p ${HADOOP_LOG_DIR}
-  #create the log sub dir for diff users
-  mkdir -p ${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-  mkdir -p ${HADOOP_LOG_DIR}/${HADOOP_MR_USER}
-
-  mkdir -p ${HADOOP_PID_DIR}
-  chown ${HADOOP_HDFS_USER}:${HADOOP_GROUP} ${HDFS_DIR}
-  chown ${HADOOP_HDFS_USER}:${HADOOP_GROUP} ${DATANODE_DIR}
-  chmod 700 -R ${DATANODE_DIR}
-  chown ${HADOOP_MR_USER}:${HADOOP_GROUP} ${MAPRED_DIR}
-  chown ${HADOOP_HDFS_USER}:${HADOOP_GROUP} ${HADOOP_LOG_DIR}
-  chmod 775 ${HADOOP_LOG_DIR}
-  chmod 775 ${HADOOP_PID_DIR}
-  chown root:${HADOOP_GROUP} ${HADOOP_PID_DIR}
-
-  #change the permission and the owner
-  chmod 755 ${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-  chown ${HADOOP_HDFS_USER}:${HADOOP_GROUP} ${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-  chmod 755 ${HADOOP_LOG_DIR}/${HADOOP_MR_USER}
-  chown ${HADOOP_MR_USER}:${HADOOP_GROUP} ${HADOOP_LOG_DIR}/${HADOOP_MR_USER}
-
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/core-site.xml ${HADOOP_CONF_DIR}/core-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hdfs-site.xml ${HADOOP_CONF_DIR}/hdfs-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/mapred-site.xml ${HADOOP_CONF_DIR}/mapred-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-env.sh ${HADOOP_CONF_DIR}/hadoop-env.sh
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-policy.xml ${HADOOP_CONF_DIR}/hadoop-policy.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/commons-logging.properties ${HADOOP_CONF_DIR}/commons-logging.properties
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/mapred-queue-acls.xml ${HADOOP_CONF_DIR}/mapred-queue-acls.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/taskcontroller.cfg ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/capacity-scheduler.xml ${HADOOP_CONF_DIR}/capacity-scheduler.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/log4j.properties ${HADOOP_CONF_DIR}/log4j.properties
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-metrics2.properties ${HADOOP_CONF_DIR}/hadoop-metrics2.properties
-
-  #setup up the proxy users
-  setupProxyUsers
- 
-  #setup short circuit read
-  setupShortCircuitRead
-
-  #set the owner of the hadoop dir to root
-  chown root ${HADOOP_PREFIX}
-  chown root:${HADOOP_GROUP} ${HADOOP_CONF_DIR}/hadoop-env.sh
-  chmod 755 ${HADOOP_CONF_DIR}/hadoop-env.sh
-  
-  #set taskcontroller
-  chown root:${HADOOP_GROUP} ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  chmod 400 ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  chown root:${HADOOP_GROUP} ${HADOOP_PREFIX}/bin/task-controller
-  chmod 6050 ${HADOOP_PREFIX}/bin/task-controller
-
-
-  #generate the slaves file and include and exclude files for hdfs and mapred
-  echo '' > ${HADOOP_CONF_DIR}/slaves
-  echo '' > ${HADOOP_CONF_DIR}/dfs.include
-  echo '' > ${HADOOP_CONF_DIR}/dfs.exclude
-  echo '' > ${HADOOP_CONF_DIR}/mapred.include
-  echo '' > ${HADOOP_CONF_DIR}/mapred.exclude
-  for dn in $DATANODES
-  do
-    echo $dn >> ${HADOOP_CONF_DIR}/slaves
-    echo $dn >> ${HADOOP_CONF_DIR}/dfs.include
-  done
-  for tt in $TASKTRACKERS
-  do
-    echo $tt >> ${HADOOP_CONF_DIR}/mapred.include
-  done
-
-  echo "Configuration setup is completed."
-  if [[ "$HADOOP_NN_HOST" =~ "`hostname`" ]]; then
-    echo "Proceed to run hadoop-setup-hdfs.sh on namenode."
-  fi
-else
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/core-site.xml ${HADOOP_CONF_DIR}/core-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hdfs-site.xml ${HADOOP_CONF_DIR}/hdfs-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/mapred-site.xml ${HADOOP_CONF_DIR}/mapred-site.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-env.sh ${HADOOP_CONF_DIR}/hadoop-env.sh
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-policy.xml ${HADOOP_CONF_DIR}/hadoop-policy.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/commons-logging.properties ${HADOOP_CONF_DIR}/commons-logging.properties
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/mapred-queue-acls.xml ${HADOOP_CONF_DIR}/mapred-queue-acls.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/taskcontroller.cfg ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-metrics2.properties ${HADOOP_CONF_DIR}/hadoop-metrics2.properties
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/capacity-scheduler.xml ${HADOOP_CONF_DIR}/capacity-scheduler.xml
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/log4j.properties ${HADOOP_CONF_DIR}/log4j.properties
-  template_generator ${HADOOP_PREFIX}/share/hadoop/common/templates/conf/hadoop-metrics2.properties ${HADOOP_CONF_DIR}/hadoop-metrics2.properties
-  
-  #setup up the proxy users
-  setupProxyUsers
-  
-  #setup short circuit read
-  setupShortCircuitRead
-
-  chown root:${HADOOP_GROUP} ${HADOOP_CONF_DIR}/hadoop-env.sh
-  chmod 755 ${HADOOP_CONF_DIR}/hadoop-env.sh
-  #set taskcontroller
-  chown root:${HADOOP_GROUP} ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  chmod 400 ${HADOOP_CONF_DIR}/taskcontroller.cfg
-  chown root:${HADOOP_GROUP} ${HADOOP_PREFIX}/bin/task-controller
-  chmod 6050 ${HADOOP_PREFIX}/bin/task-controller
-  
-  #generate the slaves file and include and exclude files for hdfs and mapred
-  echo '' > ${HADOOP_CONF_DIR}/slaves
-  echo '' > ${HADOOP_CONF_DIR}/dfs.include
-  echo '' > ${HADOOP_CONF_DIR}/dfs.exclude
-  echo '' > ${HADOOP_CONF_DIR}/mapred.include
-  echo '' > ${HADOOP_CONF_DIR}/mapred.exclude
-  for dn in $DATANODES
-  do
-    echo $dn >> ${HADOOP_CONF_DIR}/slaves
-    echo $dn >> ${HADOOP_CONF_DIR}/dfs.include
-  done
-  for tt in $TASKTRACKERS
-  do
-    echo $tt >> ${HADOOP_CONF_DIR}/mapred.include
-  done
-  
-  echo
-  echo "Configuration file has been generated in:"
-  echo
-  echo "${HADOOP_CONF_DIR}/core-site.xml"
-  echo "${HADOOP_CONF_DIR}/hdfs-site.xml"
-  echo "${HADOOP_CONF_DIR}/mapred-site.xml"
-  echo "${HADOOP_CONF_DIR}/hadoop-env.sh"
-  echo "${HADOOP_CONF_DIR}/hadoop-policy.xml"
-  echo "${HADOOP_CONF_DIR}/commons-logging.properties"
-  echo "${HADOOP_CONF_DIR}/taskcontroller.cfg"
-  echo "${HADOOP_CONF_DIR}/capacity-scheduler.xml"
-  echo "${HADOOP_CONF_DIR}/log4j.properties"
-  echo "${HADOOP_CONF_DIR}/hadoop-metrics2.properties"
-  echo
-  echo " to ${HADOOP_CONF_DIR} on all nodes, and proceed to run hadoop-setup-hdfs.sh on namenode."
-fi

+ 0 - 157
hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-hdfs.sh

@@ -1,157 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bin=`dirname "$0"`
-bin=`cd "$bin"; pwd`
-
-DEFAULT_LIBEXEC_DIR="$bin"/../libexec
-HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
-. $HADOOP_LIBEXEC_DIR/hadoop-config.sh
-
-usage() {
-  echo "
-usage: $0 <parameters>
-
-  Optional parameters:
-     --format                                                        Force namenode format
-     --group=hadoop                                                  Set Hadoop group
-     -h                                                              Display this message
-     --hdfs-user=hdfs                                                Set HDFS user
-     --kerberos-realm=KERBEROS.EXAMPLE.COM                           Set Kerberos realm
-     --hdfs-user-keytab=/home/hdfs/hdfs.keytab                       Set HDFS user key tab
-     --mapreduce-user=mr                                             Set mapreduce user
-  "
-  exit 1
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'format' \
-  -l 'hdfs-user:' \
-  -l 'hdfs-user-keytab:' \
-  -l 'mapreduce-user:' \
-  -l 'kerberos-realm:' \
-  -o 'h' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --format)
-      FORMAT_NAMENODE=1; shift
-      AUTOMATED=1
-      ;;
-    --group)
-      HADOOP_GROUP=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hdfs-user)
-      HADOOP_HDFS_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --mapreduce-user)
-      HADOOP_MR_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --yarn-user)
-      HADOOP_YARN_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --hdfs-user-keytab)
-      HDFS_KEYTAB=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --kerberos-realm)
-      KERBEROS_REALM=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-HADOOP_GROUP=${HADOOP_GROUP:-hadoop}
-HADOOP_HDFS_USER=${HADOOP_HDFS_USER:-hdfs}
-HADOOP_YARN_USER=${HADOOP_YARN_USER:-yarn}
-HADOOP_MAPREDUCE_USER=${HADOOP_MR_USER:-mapred}
-
-if [ "${KERBEROS_REALM}" != "" ]; then
-  # Determine kerberos location base on Linux distro.
-  if [ -e /etc/lsb-release ]; then
-    KERBEROS_BIN=/usr/bin
-  else
-    KERBEROS_BIN=/usr/kerberos/bin
-  fi
-  kinit_cmd="${KERBEROS_BIN}/kinit -k -t ${HDFS_KEYTAB} ${HADOOP_HDFS_USER}"
-  su -c "${kinit_cmd}" ${HADOOP_HDFS_USER}
-fi
-
-echo "Setup Hadoop Distributed File System"
-echo
-
-# Format namenode
-if [ "${FORMAT_NAMENODE}" == "1" ]; then
-  echo "Formatting namenode"
-  echo
-  su -c "echo Y | ${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} namenode -format" ${HADOOP_HDFS_USER}
-  echo
-fi
-
-# Start namenode process
-echo "Starting namenode process"
-echo
-if [ -e ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh ]; then
-  DAEMON_PATH=${HADOOP_PREFIX}/sbin
-else
-  DAEMON_PATH=${HADOOP_PREFIX}/bin
-fi
-su -c "${DAEMON_PATH}/hadoop-daemon.sh --config ${HADOOP_CONF_DIR} start namenode" ${HADOOP_HDFS_USER}
-echo
-echo "Initialize HDFS file system: "
-echo
-
-#create the /user dir 
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir /user" ${HADOOP_HDFS_USER}
-
-#create /tmp and give it 777
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir /tmp" ${HADOOP_HDFS_USER}
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chmod 777 /tmp" ${HADOOP_HDFS_USER}
-
-#create /mapred
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir /mapred" ${HADOOP_HDFS_USER}
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chmod 700 /mapred" ${HADOOP_HDFS_USER}
-su -c "${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chown ${HADOOP_MAPREDUCE_USER}:system /mapred" ${HADOOP_HDFS_USER}
-
-if [ $? -eq 0 ]; then
-  echo "Completed."
-else
-  echo "Unknown error occurred, check hadoop logs for details."
-fi
-
-echo
-echo "Please startup datanode processes: /etc/init.d/hadoop-datanode start"

+ 0 - 219
hadoop-common-project/hadoop-common/src/main/packages/hadoop-setup-single-node.sh

@@ -1,219 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Script for setup HDFS file system for single node deployment
-
-bin=`dirname "$0"`
-bin=`cd "$bin"; pwd`
-
-if [ "$HADOOP_HOME" != "" ]; then
-  echo "Warning: \$HADOOP_HOME is deprecated."
-  echo
-fi
-
-DEFAULT_LIBEXEC_DIR="$bin"/../libexec
-HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
-. $HADOOP_LIBEXEC_DIR/hadoop-config.sh
-
-usage() {
-  echo "
-usage: $0 <parameters>
-
-  Optional parameters:
-     --default                   Setup system as default
-     -h                          Display this message
-  "
-  exit 1
-}
-
-template_generator() {
-  REGEX='(\$\{[a-zA-Z_][a-zA-Z_0-9]*\})'
-  cat $1 |
-  while read line ; do
-    while [[ "$line" =~ $REGEX ]] ; do
-      LHS=${BASH_REMATCH[1]}
-      RHS="$(eval echo "\"$LHS\"")"
-      line=${line//$LHS/$RHS}
-    done
-    echo $line >> $2
-  done
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'default' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-if [ -e /etc/hadoop/hadoop-env.sh ]; then
-  . /etc/hadoop/hadoop-env.sh
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --default)
-      AUTOMATED=1; shift
-      ;;
-    -h)
-      usage
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-if [ "${AUTOMATED}" != "1" ]; then
-  echo "Welcome to Hadoop single node setup wizard"
-  echo
-  echo -n "Would you like to use default single node configuration? (y/n) "
-  read SET_CONFIG
-  echo -n "Would you like to format name node? (y/n) "
-  read SET_FORMAT
-  echo -n "Would you like to setup default directory structure? (y/n) "
-  read SET_MKDIR
-  echo -n "Would you like to start up Hadoop? (y/n) "
-  read STARTUP
-  echo -n "Would you like to start up Hadoop on reboot? (y/n) "
-  read SET_REBOOT
-  echo
-  echo "Review your choices:"
-  echo
-  echo "Setup single node configuration    : ${SET_CONFIG}"
-  echo "Format namenode                    : ${SET_FORMAT}"
-  echo "Setup default file system structure: ${SET_MKDIR}"
-  echo "Start up Hadoop                    : ${STARTUP}"
-  echo "Start up Hadoop on reboot          : ${SET_REBOOT}"
-  echo
-  echo -n "Proceed with setup? (y/n) "
-  read CONFIRM
-  if [ "${CONFIRM}" != "y" ]; then
-    echo "User aborted setup, exiting..."
-    exit 1
-  fi
-else
-  SET_CONFIG="y"
-  SET_FORMAT="y"
-  SET_MKDIR="y"
-  STARTUP="y"
-  SET_REBOOT="y"
-fi
-
-AUTOMATED=${AUTOMATED:-0}
-SET_CONFIG=${SET_CONFIG:-y}
-SET_FORMAT=${SET_FORMAT:-n}
-SET_MKDIR=${SET_MKDIR:-y}
-STARTUP=${STARTUP:-y}
-SET_REBOOT=${SET_REBOOT:-y}
-
-# Make sure system is not already started
-/etc/init.d/hadoop-namenode stop 2>/dev/null >/dev/null
-/etc/init.d/hadoop-datanode stop 2>/dev/null >/dev/null
-/etc/init.d/hadoop-jobtracker stop 2>/dev/null >/dev/null
-/etc/init.d/hadoop-tasktracker stop 2>/dev/null >/dev/null
-
-if [ "${SET_CONFIG}" == "y" ]; then
-  JAVA_HOME=${JAVA_HOME:-/usr/java/default}
-  HADOOP_NN_HOST=${HADOOP_NN_HOST:-localhost}
-  HADOOP_NN_DIR=${HADOOP_NN_DIR:-/var/lib/hadoop/hdfs/namenode}
-  HADOOP_DN_DIR=${HADOOP_DN_DIR:-/var/lib/hadoop/hdfs/datanode}
-  HADOOP_JT_HOST=${HADOOP_JT_HOST:-localhost}
-  HADOOP_HDFS_DIR=${HADOOP_MAPRED_DIR:-/var/lib/hadoop/hdfs}
-  HADOOP_MAPRED_DIR=${HADOOP_MAPRED_DIR:-/var/lib/hadoop/mapred}
-  HADOOP_PID_DIR=${HADOOP_PID_DIR:-/var/run/hadoop}
-  HADOOP_LOG_DIR="/var/log/hadoop"
-  HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-/etc/hadoop}
-  HADOOP_REPLICATION=${HADOOP_RELICATION:-1}
-  ${HADOOP_PREFIX}/sbin/hadoop-setup-conf.sh --auto \
-    --hdfs-user=hdfs \
-    --mapreduce-user=mapred \
-    --conf-dir=${HADOOP_CONF_DIR} \
-    --datanode-dir=${HADOOP_DN_DIR} \
-    --hdfs-dir=${HADOOP_HDFS_DIR} \
-    --jobtracker-host=${HADOOP_JT_HOST} \
-    --log-dir=${HADOOP_LOG_DIR} \
-    --pid-dir=${HADOOP_PID_DIR} \
-    --mapred-dir=${HADOOP_MAPRED_DIR} \
-    --namenode-dir=${HADOOP_NN_DIR} \
-    --namenode-host=${HADOOP_NN_HOST} \
-    --replication=${HADOOP_REPLICATION}
-fi
-
-if [ ! -e ${HADOOP_NN_DIR} ]; then
-  rm -rf ${HADOOP_HDFS_DIR} 2>/dev/null >/dev/null
-  mkdir -p ${HADOOP_HDFS_DIR}
-  chmod 755 ${HADOOP_HDFS_DIR}
-  chown hdfs:hadoop ${HADOOP_HDFS_DIR}
-  /etc/init.d/hadoop-namenode format
-elif [ "${SET_FORMAT}" == "y" ]; then
-  rm -rf ${HADOOP_HDFS_DIR} 2>/dev/null >/dev/null
-  mkdir -p ${HADOOP_HDFS_DIR}
-  chmod 755 ${HADOOP_HDFS_DIR}
-  chown hdfs:hadoop ${HADOOP_HDFS_DIR}
-  rm -rf ${HADOOP_NN_DIR}
-  /etc/init.d/hadoop-namenode format
-fi
-
-/etc/init.d/hadoop-namenode start
-/etc/init.d/hadoop-datanode start
-
-su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir /user/mapred' hdfs
-su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chown mapred:mapred /user/mapred' hdfs
-su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -mkdir /tmp' hdfs
-su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} dfs -chmod 777 /tmp' hdfs
-
-/etc/init.d/hadoop-jobtracker start
-/etc/init.d/hadoop-tasktracker start
-
-if [ "${SET_REBOOT}" == "y" ]; then
-  if [ -e /etc/debian_version ]; then
-    ln -sf ../init.d/hadoop-namenode /etc/rc2.d/S90hadoop-namenode
-    ln -sf ../init.d/hadoop-datanode /etc/rc2.d/S91hadoop-datanode
-    ln -sf ../init.d/hadoop-jobtracker /etc/rc2.d/S92hadoop-jobtracker
-    ln -sf ../init.d/hadoop-tasktracker /etc/rc2.d/S93hadoop-tasktracker
-    ln -sf ../init.d/hadoop-namenode /etc/rc6.d/S10hadoop-namenode
-    ln -sf ../init.d/hadoop-datanode /etc/rc6.d/S11hadoop-datanode
-    ln -sf ../init.d/hadoop-jobtracker /etc/rc6.d/S12hadoop-jobtracker
-    ln -sf ../init.d/hadoop-tasktracker /etc/rc6.d/S13hadoop-tasktracker
-  elif [ -e /etc/redhat-release ]; then
-    /sbin/chkconfig hadoop-namenode --add
-    /sbin/chkconfig hadoop-datanode --add
-    /sbin/chkconfig hadoop-jobtracker --add
-    /sbin/chkconfig hadoop-tasktracker --add
-    /sbin/chkconfig hadoop-namenode on
-    /sbin/chkconfig hadoop-datanode on
-    /sbin/chkconfig hadoop-jobtracker on
-    /sbin/chkconfig hadoop-tasktracker on
-  fi
-fi
-
-if [ "${STARTUP}" != "y" ]; then
-  /etc/init.d/hadoop-namenode stop
-  /etc/init.d/hadoop-datanode stop
-  /etc/init.d/hadoop-jobtracker stop
-  /etc/init.d/hadoop-tasktracker stop
-fi

+ 0 - 183
hadoop-common-project/hadoop-common/src/main/packages/hadoop-validate-setup.sh

@@ -1,183 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-
-
-###############################################################################
-# Run the following jobs to validate a hadoop cluster
-## teragen
-## terasort
-## teravalidate
-# If they all pass 0 will be returned and 1 otherwise
-# The test will work for both secure and unsecure deploys. If the kerberos-realm
-# is passed we will assume that the deploy is secure and proceed with a kinit before
-# running the validation jobs.
-################################################################################
-
-bin=`dirname "$0"`
-bin=`cd "$bin"; pwd`
-
-DEFAULT_LIBEXEC_DIR="$bin"/../libexec
-HADOOP_LIBEXEC_DIR=${HADOOP_LIBEXEC_DIR:-$DEFAULT_LIBEXEC_DIR}
-. $HADOOP_LIBEXEC_DIR/hadoop-config.sh
-
-usage() {
-  echo "
-usage: $0 <parameters>
-
-  Optional parameters:
-     -h                                                              Display this message
-     --user=hdfs
-     --user_keytab=/home/hdfs/hdfs.keytab
-     --kerberos-realm=KERBEROS.EXAMPLE.COM                           Set Kerberos realm
-  "
-  exit 1
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'user:' \
-  -l 'user-keytab:' \
-  -l 'kerberos-realm:' \
-  -o 'h' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --user)
-      TEST_USER=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --user-keytab)
-      USER_KEYTAB_FILE=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --kerberos-realm)
-      KERBEROS_REALM=$2; shift 2
-      AUTOMATED=1
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-#set the hadoop command and the path to the hadoop examples jar
-HADOOP_CMD="${HADOOP_PREFIX}/bin/hadoop --config $HADOOP_CONF_DIR"
-
-#find the hadoop examples jar
-HADOOP_EXAMPLES_JAR=''
-
-#find under HADOOP_PREFIX (tar ball install)
-HADOOP_EXAMPLES_JAR=`find ${HADOOP_PREFIX} -name 'hadoop-mapreduce-examples-*.jar' | head -n1`
-
-#if its not found look under /usr/share/hadoop (rpm/deb installs)
-if [ "$HADOOP_EXAMPLES_JAR" == '' ]
-then
-  HADOOP_EXAMPLES_JAR=`find /usr/share/hadoop -name 'hadoop-mapreduce-examples-*.jar' | head -n1`
-fi
-
-#if it is still empty then dont run the tests
-if [ "$HADOOP_EXAMPLES_JAR" == '' ]
-then
-  echo "Did not find hadoop-examples-*.jar under '${HADOOP_PREFIX} or '/usr/share/hadoop'"
-  exit 1
-fi
-
-# do a kinit if secure
-if [ "${KERBEROS_REALM}" != "" ]; then
-  # Determine kerberos location base on Linux distro.
-  if [ -e /etc/lsb-release ]; then
-    KERBEROS_BIN=/usr/bin
-  else
-    KERBEROS_BIN=/usr/kerberos/bin
-  fi
-  kinit_cmd="su -c '${KERBEROS_BIN}/kinit -kt ${USER_KEYTAB_FILE} ${TEST_USER}' ${TEST_USER}"
-  echo $kinit_cmd
-  eval $kinit_cmd
-  if [ $? -ne 0 ]
-  then
-    echo "kinit command did not run successfully."
-    exit 1
-  fi
-fi
-
-
-#dir where to store the data on hdfs. The data is relative of the users home dir on hdfs.
-PARENT_DIR="validate_deploy_`date +%s`"
-TERA_GEN_OUTPUT_DIR="${PARENT_DIR}/tera_gen_data"
-TERA_SORT_OUTPUT_DIR="${PARENT_DIR}/tera_sort_data"
-TERA_VALIDATE_OUTPUT_DIR="${PARENT_DIR}/tera_validate_data"
-#tera gen cmd
-TERA_GEN_CMD="su -c '$HADOOP_CMD jar $HADOOP_EXAMPLES_JAR teragen 10000 $TERA_GEN_OUTPUT_DIR' $TEST_USER"
-
-#tera sort cmd
-TERA_SORT_CMD="su -c '$HADOOP_CMD jar $HADOOP_EXAMPLES_JAR terasort $TERA_GEN_OUTPUT_DIR $TERA_SORT_OUTPUT_DIR' $TEST_USER"
-
-#tera validate cmd
-TERA_VALIDATE_CMD="su -c '$HADOOP_CMD jar $HADOOP_EXAMPLES_JAR teravalidate $TERA_SORT_OUTPUT_DIR $TERA_VALIDATE_OUTPUT_DIR' $TEST_USER"
-
-echo "Starting teragen...."
-
-#run tera gen
-echo $TERA_GEN_CMD
-eval $TERA_GEN_CMD
-if [ $? -ne 0 ]; then
-  echo "tera gen failed."
-  exit 1
-fi
-
-echo "Teragen passed starting terasort...."
-
-
-#run tera sort
-echo $TERA_SORT_CMD
-eval $TERA_SORT_CMD
-if [ $? -ne 0 ]; then
-  echo "tera sort failed."
-  exit 1
-fi
-
-echo "Terasort passed starting teravalidate...."
-
-#run tera validate
-echo $TERA_VALIDATE_CMD
-eval $TERA_VALIDATE_CMD
-if [ $? -ne 0 ]; then
-  echo "tera validate failed."
-  exit 1
-fi
-
-echo "teragen, terasort, teravalidate passed."
-echo "Cleaning the data created by tests: $PARENT_DIR"
-
-CLEANUP_CMD="su -c '$HADOOP_CMD dfs -rmr -skipTrash $PARENT_DIR' $TEST_USER"
-echo $CLEANUP_CMD
-eval $CLEANUP_CMD
-
-exit 0

+ 0 - 93
hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-datanode

@@ -1,93 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop datanode
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop datanode
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid"
-desc="Hadoop datanode daemon"
-HADOOP_PREFIX="/usr"
-
-start() {
-  echo -n $"Starting $desc (hadoop-datanode): "
-  if [ -n "$HADOOP_SECURE_DN_USER" ]; then
-    daemon ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start datanode
-  else
-    daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start datanode
-  fi
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-datanode
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-datanode): "
-  if [ -n "$HADOOP_SECURE_DN_USER" ]; then
-    daemon ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop datanode
-  else
-    daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop datanode
-  fi
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-datanode $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-datanode ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 85
hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-jobtracker

@@ -1,85 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop jobtracker
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop jobtracker
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid"
-desc="Hadoop jobtracker daemon"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-start() {
-  echo -n $"Starting $desc (hadoop-jobtracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start jobtracker
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-jobtracker
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-jobtracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop jobtracker
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-jobtracker $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-jobtracker ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 99
hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-namenode

@@ -1,99 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop namenode
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop namenode
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid"
-desc="Hadoop namenode daemon"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-start() {
-  echo -n $"Starting $desc (hadoop-namenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start namenode $1
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-namenode
-  return $RETVAL
-}
-
-upgrade() {
-  start -upgrade
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-namenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop namenode
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-namenode $PIDFILE
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-restart() {
-  stop
-  start
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-namenode ] && restart || :
-}
-
-format() {
-  daemon --user hdfs ${HADOOP_PREFIX}/bin/hadoop namenode -format
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  upgrade)
-    upgrade
-    ;;
-  format)
-    format
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart|try-restart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|try-restart|upgrade}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 85
hadoop-common-project/hadoop-common/src/main/packages/rpm/init.d/hadoop-tasktracker

@@ -1,85 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop tasktracker
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop tasktracker
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid"
-desc="Hadoop tasktracker daemon"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-start() {
-  echo -n $"Starting $desc (hadoop-tasktracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start tasktracker
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-tasktracker
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-tasktracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop tasktracker
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-tasktracker $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-tasktracker ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 174
hadoop-common-project/hadoop-common/src/main/packages/rpm/spec/hadoop.spec

@@ -1,174 +0,0 @@
-#   Licensed to the Apache Software Foundation (ASF) under one or more
-#   contributor license agreements.  See the NOTICE file distributed with
-#   this work for additional information regarding copyright ownership.
-#   The ASF licenses this file to You under the Apache License, Version 2.0
-#   (the "License"); you may not use this file except in compliance with
-#   the License.  You may obtain a copy of the License at
-#
-#       http://www.apache.org/licenses/LICENSE-2.0
-#
-#   Unless required by applicable law or agreed to in writing, software
-#   distributed under the License is distributed on an "AS IS" BASIS,
-#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-#   See the License for the specific language governing permissions and
-#   limitations under the License.
-
-#
-# RPM Spec file for Hadoop version @version@
-#
-
-%define name         hadoop-common
-%define version      @version@
-%define release      @package.release@
-
-# Installation Locations
-%define _prefix      @package.prefix@
-%define _bin_dir     %{_prefix}/bin
-%define _conf_dir    @package.conf.dir@
-%define _lib_dir     %{_prefix}/lib
-%define _lib64_dir   %{_prefix}/lib64
-%define _libexec_dir %{_prefix}/libexec
-%define _log_dir     @package.log.dir@
-%define _pid_dir     @package.pid.dir@
-%define _sbin_dir    %{_prefix}/sbin
-%define _share_dir   %{_prefix}/share
-%define _var_dir     @package.var.dir@
-
-# Build time settings
-%define _build_dir  @package.build.dir@
-%define _final_name @final.name@
-%define debug_package %{nil}
-
-# Disable brp-java-repack-jars for aspect J
-%define __os_install_post    \
-    /usr/lib/rpm/redhat/brp-compress \
-    %{!?__debug_package:/usr/lib/rpm/redhat/brp-strip %{__strip}} \
-    /usr/lib/rpm/redhat/brp-strip-static-archive %{__strip} \
-    /usr/lib/rpm/redhat/brp-strip-comment-note %{__strip} %{__objdump} \
-    /usr/lib/rpm/brp-python-bytecompile %{nil}
-
-# RPM searches perl files for dependancies and this breaks for non packaged perl lib
-# like thrift so disable this
-%define _use_internal_dependency_generator 0
-
-%ifarch i386
-%global hadoop_arch Linux-i386-32
-%endif
-%ifarch amd64 x86_64
-%global hadoop_arch Linux-amd64-64
-%endif
-%ifarch noarch
-%global hadoop_arch ""
-%endif
-
-Summary: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing
-License: Apache License, Version 2.0
-URL: http://hadoop.apache.org/core/
-Vendor: Apache Software Foundation
-Group: Development/Libraries
-Name: %{name}
-Version: %{version}
-Release: %{release} 
-Source0: %{_final_name}-bin.tar.gz
-Prefix: %{_prefix}
-Prefix: %{_conf_dir}
-Prefix: %{_log_dir}
-Prefix: %{_pid_dir}
-Buildroot: %{_build_dir}
-Requires: sh-utils, textutils, /usr/sbin/useradd, /usr/sbin/usermod, /sbin/chkconfig, /sbin/service, jdk >= 1.6
-AutoReqProv: no
-Provides: hadoop
-
-%description
-The Apache Hadoop project develops open-source software for reliable, scalable, 
-distributed computing.  Hadoop includes these subprojects:
-
-Hadoop Common: The common utilities that support the other Hadoop subprojects.
-
-%prep
-%setup -n %{_final_name}
-
-%build
-if [ -d ${RPM_BUILD_DIR}%{_prefix} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_prefix}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_log_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_log_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_conf_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_conf_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_pid_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_pid_dir}
-fi
-
-mkdir -p ${RPM_BUILD_DIR}%{_prefix}
-mkdir -p ${RPM_BUILD_DIR}%{_bin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_lib_dir}
-%ifarch amd64 x86_64
-mkdir -p ${RPM_BUILD_DIR}%{_lib64_dir}
-%endif
-mkdir -p ${RPM_BUILD_DIR}%{_libexec_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_log_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_conf_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_pid_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_sbin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_share_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_var_dir}
-
-#########################
-#### INSTALL SECTION ####
-#########################
-%install
-mv ${RPM_BUILD_DIR}/%{_final_name}/bin/* ${RPM_BUILD_DIR}%{_bin_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/* ${RPM_BUILD_DIR}%{_conf_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/lib/* ${RPM_BUILD_DIR}%{_lib_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/libexec/* ${RPM_BUILD_DIR}%{_libexec_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/sbin/* ${RPM_BUILD_DIR}%{_sbin_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/share/* ${RPM_BUILD_DIR}%{_share_dir}
-rm -rf ${RPM_BUILD_DIR}/%{_final_name}/etc
-
-%pre
-getent group hadoop 2>/dev/null >/dev/null || /usr/sbin/groupadd -g 123 -r hadoop
-
-%post
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-hadoop-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3}
-
-%preun
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-hadoop-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3} \
-       --uninstall
-
-%files 
-%defattr(-,root,root)
-%attr(0755,root,hadoop) %{_log_dir}
-%attr(0775,root,hadoop) %{_pid_dir}
-%config(noreplace) %{_conf_dir}/configuration.xsl
-%config(noreplace) %{_conf_dir}/core-site.xml
-%config(noreplace) %{_conf_dir}/hadoop-env.sh
-%config(noreplace) %{_conf_dir}/hadoop-metrics.properties
-%config(noreplace) %{_conf_dir}/hadoop-metrics2.properties
-%config(noreplace) %{_conf_dir}/hadoop-policy.xml
-%config(noreplace) %{_conf_dir}/log4j.properties
-%config(noreplace) %{_conf_dir}/masters
-%config(noreplace) %{_conf_dir}/slaves
-%config{noreplace) %{_conf_dir}/fair-scheduler.xml
-%{_conf_dir}/hadoop-env.sh.template
-%{_conf_dir}/ssl-client.xml.example
-%{_conf_dir}/ssl-server.xml.example
-%{_prefix}
-

+ 0 - 191
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/capacity-scheduler.xml

@@ -1,191 +0,0 @@
-<?xml version="1.0"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- This is the configuration file for the resource manager in Hadoop. -->
-<!-- You can configure various scheduling parameters related to queues. -->
-<!-- The properties for a queue follow a naming convention,such as, -->
-<!-- mapred.capacity-scheduler.queue.<queue-name>.property-name. -->
-
-<configuration>
-
-  <property>
-    <name>mapred.capacity-scheduler.maximum-system-jobs</name>
-    <value>3000</value>
-    <description>Maximum number of jobs in the system which can be initialized,
-     concurrently, by the CapacityScheduler.
-    </description>    
-  </property>
-  
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.capacity</name>
-    <value>100</value>
-    <description>Percentage of the number of slots in the cluster that are
-      to be available for jobs in this queue.
-    </description>    
-  </property>
-  
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.maximum-capacity</name>
-    <value>-1</value>
-    <description>
-	maximum-capacity defines a limit beyond which a queue cannot use the capacity of the cluster.
-	This provides a means to limit how much excess capacity a queue can use. By default, there is no limit.
-	The maximum-capacity of a queue can only be greater than or equal to its minimum capacity.
-        Default value of -1 implies a queue can use complete capacity of the cluster.
-
-        This property could be to curtail certain jobs which are long running in nature from occupying more than a 
-        certain percentage of the cluster, which in the absence of pre-emption, could lead to capacity guarantees of 
-        other queues being affected.
-        
-        One important thing to note is that maximum-capacity is a percentage , so based on the cluster's capacity
-        the max capacity would change. So if large no of nodes or racks get added to the cluster , max Capacity in 
-        absolute terms would increase accordingly.
-    </description>    
-  </property>
-  
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.supports-priority</name>
-    <value>false</value>
-    <description>If true, priorities of jobs will be taken into 
-      account in scheduling decisions.
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.minimum-user-limit-percent</name>
-    <value>100</value>
-    <description> Each queue enforces a limit on the percentage of resources 
-    allocated to a user at any given time, if there is competition for them. 
-    This user limit can vary between a minimum and maximum value. The former
-    depends on the number of users who have submitted jobs, and the latter is
-    set to this property value. For example, suppose the value of this 
-    property is 25. If two users have submitted jobs to a queue, no single 
-    user can use more than 50% of the queue resources. If a third user submits
-    a job, no single user can use more than 33% of the queue resources. With 4 
-    or more users, no user can use more than 25% of the queue's resources. A 
-    value of 100 implies no user limits are imposed. 
-    </description>
-  </property>
-  
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.user-limit-factor</name>
-    <value>1</value>
-    <description>The multiple of the queue capacity which can be configured to 
-    allow a single user to acquire more slots. 
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks</name>
-    <value>200000</value>
-    <description>The maximum number of tasks, across all jobs in the queue, 
-    which can be initialized concurrently. Once the queue's jobs exceed this 
-    limit they will be queued on disk.  
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user</name>
-    <value>100000</value>
-    <description>The maximum number of tasks per-user, across all the of the 
-    user's jobs in the queue, which can be initialized concurrently. Once the 
-    user's jobs exceed this limit they will be queued on disk.  
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.queue.default.init-accept-jobs-factor</name>
-    <value>10</value>
-    <description>The multipe of (maximum-system-jobs * queue-capacity) used to 
-    determine the number of jobs which are accepted by the scheduler.  
-    </description>
-  </property>
-
-  <!-- The default configuration settings for the capacity task scheduler -->
-  <!-- The default values would be applied to all the queues which don't have -->
-  <!-- the appropriate property for the particular queue -->
-  <property>
-    <name>mapred.capacity-scheduler.default-supports-priority</name>
-    <value>false</value>
-    <description>If true, priorities of jobs will be taken into 
-      account in scheduling decisions by default in a job queue.
-    </description>
-  </property>
-  
-  <property>
-    <name>mapred.capacity-scheduler.default-minimum-user-limit-percent</name>
-    <value>100</value>
-    <description>The percentage of the resources limited to a particular user
-      for the job queue at any given point of time by default.
-    </description>
-  </property>
-
-
-  <property>
-    <name>mapred.capacity-scheduler.default-user-limit-factor</name>
-    <value>1</value>
-    <description>The default multiple of queue-capacity which is used to 
-    determine the amount of slots a single user can consume concurrently.
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.default-maximum-active-tasks-per-queue</name>
-    <value>200000</value>
-    <description>The default maximum number of tasks, across all jobs in the 
-    queue, which can be initialized concurrently. Once the queue's jobs exceed 
-    this limit they will be queued on disk.  
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.default-maximum-active-tasks-per-user</name>
-    <value>100000</value>
-    <description>The default maximum number of tasks per-user, across all the of 
-    the user's jobs in the queue, which can be initialized concurrently. Once 
-    the user's jobs exceed this limit they will be queued on disk.  
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.capacity-scheduler.default-init-accept-jobs-factor</name>
-    <value>10</value>
-    <description>The default multipe of (maximum-system-jobs * queue-capacity) 
-    used to determine the number of jobs which are accepted by the scheduler.  
-    </description>
-  </property>
-
-  <!-- Capacity scheduler Job Initialization configuration parameters -->
-  <property>
-    <name>mapred.capacity-scheduler.init-poll-interval</name>
-    <value>5000</value>
-    <description>The amount of time in miliseconds which is used to poll 
-    the job queues for jobs to initialize.
-    </description>
-  </property>
-  <property>
-    <name>mapred.capacity-scheduler.init-worker-threads</name>
-    <value>5</value>
-    <description>Number of worker threads which would be used by
-    Initialization poller to initialize jobs in a set of queue.
-    If number mentioned in property is equal to number of job queues
-    then a single thread would initialize jobs in a queue. If lesser
-    then a thread would get a set of queues assigned. If the number
-    is greater then number of threads would be equal to number of 
-    job queues.
-    </description>
-  </property>
-
-</configuration>

+ 0 - 22
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/commons-logging.properties

@@ -1,22 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-#Logging Implementation
-
-#Log4J
-org.apache.commons.logging.Log=org.apache.commons.logging.impl.Log4JLogger
-
-#JDK Logger
-#org.apache.commons.logging.Log=org.apache.commons.logging.impl.Jdk14Logger

+ 0 - 93
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/core-site.xml

@@ -1,93 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-
-  <property>
-    <name>local.realm</name>
-    <value>${KERBEROS_REALM}</value>
-  </property>
-
-  <!-- file system properties -->
-
-  <property>
-    <name>fs.default.name</name>
-    <value>hdfs://${HADOOP_NN_HOST}:8020</value>
-    <description>The name of the default file system.  Either the
-      literal string "local" or a host:port for NDFS.
-    </description>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>fs.trash.interval</name>
-    <value>360</value>
-    <description>Number of minutes between trash checkpoints.
-      If zero, the trash feature is disabled.
-    </description>
-  </property>
-  
-  <property>
-    <name>hadoop.security.auth_to_local</name>
-    <value>
-	RULE:[2:$1@$0]([jt]t@.*${KERBEROS_REALM})s/.*/${HADOOP_MR_USER}/
-        RULE:[2:$1@$0]([nd]n@.*${KERBEROS_REALM})s/.*/${HADOOP_HDFS_USER}/
-        RULE:[2:$1@$0](mapred@.*${KERBEROS_REALM})s/.*/${HADOOP_MR_USER}/
-        RULE:[2:$1@$0](hdfs@.*${KERBEROS_REALM})s/.*/${HADOOP_HDFS_USER}/
-        RULE:[2:$1@$0](mapredqa@.*${KERBEROS_REALM})s/.*/${HADOOP_MR_USER}/
-        RULE:[2:$1@$0](hdfsqa@.*${KERBEROS_REALM})s/.*/${HADOOP_HDFS_USER}/
-        RULE:[2:$1@$0](hm@.*${KERBEROS_REALM})s/.*/${HBASE_USER}/
-        RULE:[2:$1@$0](rs@.*${KERBEROS_REALM})s/.*/${HBASE_USER}/
-        DEFAULT
-    </value>
-    <description></description>
-  </property>
-
-  <property>
-    <name>hadoop.security.authentication</name>
-    <value>${SECURITY_TYPE}</value>
-    <description>
-      Set the authentication for the cluster. Valid values are: simple or
-      kerberos.
-    </description>
-  </property>
-
-  <property>
-    <name>hadoop.security.authorization</name>
-    <value>${SECURITY}</value>
-    <description>
-      Enable authorization for different protocols.
-    </description>
-  </property>
-
-  <property>
-    <name>hadoop.security.groups.cache.secs</name>
-    <value>14400</value>
-  </property>
-
-  <property>
-    <name>hadoop.kerberos.kinit.command</name>
-    <value>${KINIT}</value>
-  </property>
-
-  <property>
-    <name>hadoop.http.filter.initializers</name>
-    <value>org.apache.hadoop.http.lib.StaticUserWebFilter</value>
-  </property>
-
-</configuration>

+ 0 - 75
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-env.sh

@@ -1,75 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Set Hadoop-specific environment variables here.
-
-# The only required environment variable is JAVA_HOME.  All others are
-# optional.  When running a distributed configuration it is best to
-# set JAVA_HOME in this file, so that it is correctly defined on
-# remote nodes.
-
-# The java implementation to use.
-export JAVA_HOME=${JAVA_HOME}
-
-# The jsvc implementation to use. Jsvc is required to run secure datanodes.
-#export JSVC_HOME=${JSVC_HOME}
-
-export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
-
-# Extra Java CLASSPATH elements.  Automatically insert capacity-scheduler.
-for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
-  if [ "$HADOOP_CLASSPATH" ]; then
-    export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
-  else
-    export HADOOP_CLASSPATH=$f
-  fi
-done
-
-# The maximum amount of heap to use, in MB. Default is 1000.
-#export HADOOP_HEAPSIZE=
-#export HADOOP_NAMENODE_INIT_HEAPSIZE=""
-
-# Extra Java runtime options.  Empty by default.
-export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true $HADOOP_CLIENT_OPTS"
-
-# Command specific options appended to HADOOP_OPTS when specified
-export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS"
-export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS"
-
-export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS"
-
-# The ZKFC does not need a large heap, and keeping it small avoids
-# any potential for long GC pauses
-export HADOOP_ZKFC_OPTS="-Xmx256m $HADOOP_ZKFC_OPTS"
-
-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)
-export HADOOP_CLIENT_OPTS="-Xmx128m $HADOOP_CLIENT_OPTS"
-#HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS"
-
-# On secure datanodes, user to run the datanode as after dropping privileges
-export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
-
-# Where log files are stored.  $HADOOP_HOME/logs by default.
-#export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
-
-# Where log files are stored in the secure data environment.
-export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
-
-# The directory where pid files are stored. /tmp by default.
-export HADOOP_PID_DIR=${HADOOP_PID_DIR}
-export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
-
-# A string representing this instance of hadoop. $USER by default.
-export HADOOP_IDENT_STRING=$USER

+ 0 - 20
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-metrics2.properties

@@ -1,20 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# syntax: [prefix].[source|sink|jmx].[instance].[options]
-# See package.html for org.apache.hadoop.metrics2 for details
-
-*.period=60
-

+ 0 - 235
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hadoop-policy.xml

@@ -1,235 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-  <property>
-    <name>security.client.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for ClientProtocol, which is used by user code
-    via the DistributedFileSystem.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.client.datanode.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for ClientDatanodeProtocol, the client-to-datanode protocol
-    for block recovery.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.datanode.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for DatanodeProtocol, which is used by datanodes to
-    communicate with the namenode.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.inter.datanode.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for InterDatanodeProtocol, the inter-datanode protocol
-    for updating generation timestamp.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.namenode.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for NamenodeProtocol, the protocol used by the secondary
-    namenode to communicate with the namenode.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.inter.tracker.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for InterTrackerProtocol, used by the tasktrackers to
-    communicate with the jobtracker.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
- 
-  <property>
-    <name>security.job.submission.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for JobSubmissionProtocol, used by job clients to
-    communciate with the jobtracker for job submission, querying job status etc.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.task.umbilical.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for TaskUmbilicalProtocol, used by the map and reduce
-    tasks to communicate with the parent tasktracker.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
- <property>
-    <name>security.admin.operations.protocol.acl</name>
-    <value>${HADOOP_HDFS_USER}</value>
-    <description>ACL for AdminOperationsProtocol. Used for admin commands.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.refresh.usertogroups.mappings.protocol.acl</name>
-    <value>${HADOOP_HDFS_USER}</value>
-    <description>ACL for RefreshUserMappingsProtocol. Used to refresh
-    users mappings. The ACL is a comma-separated list of user and
-    group names. The user and group list is separated by a blank. For
-    e.g. "alice,bob users,wheel".  A special value of "*" means all
-    users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.refresh.policy.protocol.acl</name>
-    <value>${HADOOP_HDFS_USER}</value>
-    <description>ACL for RefreshAuthorizationPolicyProtocol, used by the
-    dfsadmin and mradmin commands to refresh the security policy in-effect.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-
-  <!-- YARN Protocols -->
-
-  <property>
-    <name>security.resourcetracker.protocol.acl</name>
-    <value>${HADOOP_YARN_USER}</value>
-    <description>ACL for ResourceTracker protocol, used by the
-    ResourceManager and NodeManager to communicate with each other.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.admin.protocol.acl</name>
-    <value>${HADOOP_YARN_USER}</value>
-    <description>ACL for RMAdminProtocol, for admin commands. 
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.client.resourcemanager.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for ClientRMProtocol, used by the ResourceManager 
-    and applications submission clients to communicate with each other.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.applicationmaster.resourcemanager.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for AMRMProtocol, used by the ResourceManager 
-    and ApplicationMasters to communicate with each other.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.containermanager.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for ContainerManager protocol, used by the NodeManager 
-    and ApplicationMasters to communicate with each other.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.resourcelocalizer.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for ResourceLocalizer protocol, used by the NodeManager 
-    and ResourceLocalizer to communicate with each other.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.job.task.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for TaskUmbilicalProtocol, used by the map and reduce
-    tasks to communicate with the parent tasktracker.
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-
-  <property>
-    <name>security.job.client.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for MRClientProtocol, used by job clients to
-    communciate with the MR ApplicationMaster to query job status etc. 
-    The ACL is a comma-separated list of user and group names. The user and
-    group list is separated by a blank. For e.g. "alice,bob users,wheel".
-    A special value of "*" means all users are allowed.</description>
-  </property>
-  
-  <property>
-    <name>security.ha.service.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for HAService protocol used by HAAdmin to manage the
-      active and stand-by states of namenode.</description>
-  </property>
-  <property>
-    <name>security.zkfc.protocol.acl</name>
-    <value>*</value>
-    <description>ACL for access to the ZK Failover Controller
-    </description>
-  </property>
-
-   <property>
-      <name>security.mrhs.client.protocol.acl</name>
-      <value>*</value>
-      <description>ACL for HSClientProtocol, used by job clients to
-       communciate with the MR History Server job status etc. 
-       The ACL is a comma-separated list of user and group names. The user and
-       group list is separated by a blank. For e.g. "alice,bob users,wheel".
-       A special value of "*" means all users are allowed.</description>
-    </property>
-
-</configuration>

+ 0 - 241
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/hdfs-site.xml

@@ -1,241 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<configuration>
-
-<!-- file system properties -->
-
-  <property>
-    <name>dfs.namenode.name.dir</name>
-    <value>${HADOOP_NN_DIR}</value>
-    <description>Determines where on the local filesystem the DFS name node
-      should store the name table.  If this is a comma-delimited list
-      of directories then the name table is replicated in all of the
-      directories, for redundancy. </description>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>dfs.datanode.data.dir</name>
-    <value>${HADOOP_DN_DIR}</value>
-    <description>Determines where on the local filesystem an DFS data node
-       should store its blocks.  If this is a comma-delimited
-       list of directories, then data will be stored in all named
-       directories, typically on different devices.
-       Directories that do not exist are ignored.
-    </description>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>dfs.namenode.safemode.threshold-pct</name>
-    <value>1.0f</value>
-    <description>
-        Specifies the percentage of blocks that should satisfy the minimal 
-        replication requirement defined by dfs.namenode.replication.min. 
-        Values less than or equal to 0 mean not to start in safe mode.
-        Values greater than 1 will make safe mode permanent.
-        </description>
-  </property>
-
-  <property>
-    <name>dfs.datanode.address</name>
-    <value>${HADOOP_DN_ADDR}</value>
-  </property>
-
-  <property>
-    <name>dfs.datanode.http.address</name>
-    <value>${HADOOP_DN_HTTP_ADDR}</value>
-  </property>
-
-  <property>
-    <name>dfs.namenode.http-address</name>
-    <value>${HADOOP_NN_HOST}:50070</value>
-    <description>The name of the default file system.  Either the
-       literal string "local" or a host:port for NDFS.
-    </description>
-    <final>true</final>
-  </property>
-
-  <!-- Permissions configuration -->
-  <property>
-    <name>dfs.umaskmode</name>
-    <value>077</value>
-    <description>
-      The octal umask used when creating files and directories.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.block.access.token.enable</name>
-    <value>${SECURITY}</value>
-    <description>
-      Are access tokens are used as capabilities for accessing datanodes.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.namenode.kerberos.principal</name>
-    <value>nn/_HOST@${local.realm}</value>
-    <description>
-      Kerberos principal name for the NameNode
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.secondary.namenode.kerberos.principal</name>
-    <value>nn/_HOST@${local.realm}</value>
-    <description>
-        Kerberos principal name for the secondary NameNode.
-    </description>
-  </property>
-
-
-  <property>
-    <name>dfs.namenode.kerberos.https.principal</name>
-    <value>host/_HOST@${local.realm}</value>
-    <description>
-       The Kerberos principal for the host that the NameNode runs on.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.secondary.namenode.kerberos.https.principal</name>
-    <value>host/_HOST@${local.realm}</value>
-    <description>
-      The Kerberos principal for the hostthat the secondary NameNode runs on.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.datanode.kerberos.principal</name>
-    <value>dn/_HOST@${local.realm}</value>
-    <description>
-      The Kerberos principal that the DataNode runs as. "_HOST" is replaced by 
-      the real host name.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.namenode.keytab.file</name>
-    <value>/etc/security/keytabs/nn.service.keytab</value>
-    <description>
-      Combined keytab file containing the namenode service and host principals.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.secondary.namenode.keytab.file</name>
-    <value>/etc/security/keytabs/nn.service.keytab</value>
-    <description>
-      Combined keytab file containing the namenode service and host principals.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.datanode.keytab.file</name>
-    <value>/etc/security/keytabs/dn.service.keytab</value>
-    <description>
-        The filename of the keytab file for the DataNode.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.https.port</name>
-    <value>50470</value>
-    <description>The https port where namenode binds</description>
-  </property>
-
-  <property>
-    <name>dfs.namenode.https-address</name>
-    <value>${HADOOP_NN_HOST}:50470</value>
-    <description>The https address where namenode binds</description>
-  </property>
-
-  <property>
-    <name>dfs.datanode.data.dir.perm</name>
-    <value>${DFS_DATANODE_DIR_PERM}</value>
-    <description>The permissions that should be there on dfs.datanode.data.dir
-      directories. The datanode will not come up if the permissions are
-      different on existing dfs.datanode.data.dir directories. If the directories
-      don't exist, they will be created with this permission.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.cluster.administrators</name>
-    <value>${HADOOP_HDFS_USER}</value>
-    <description>ACL for who all can view the default servlets in the HDFS</description>
-  </property>
-  
-  <property>
-    <name>dfs.permissions.superusergroup</name>
-    <value>${HADOOP_GROUP}</value>
-    <description>The name of the group of super-users.</description>
-  </property>
-  
-  <property>
-    <name>dfs.namenode.http-address</name>
-    <value>${HADOOP_NN_HOST}:50070</value>
-    <description>
-      The address and the base port where the dfs namenode web ui will listen on.
-      If the port is 0 then the server will start on a free port.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.namenode.https-address</name>
-    <value>${HADOOP_NN_HOST}:50470</value>
-  </property>
-
-  <property>
-    <name>dfs.namenode.secondary.http-address</name>
-    <value>${HADOOP_SNN_HOST}:50090</value>
-    <description>
-      The secondary namenode http server address and port.
-      If the port is 0 then the server will start on a free port.
-    </description>
-  </property>
-
-  <property>
-    <name>dfs.hosts</name>
-    <value>${HADOOP_CONF_DIR}/dfs.include</value>
-    <description>Names a file that contains a list of hosts that are
-      permitted to connect to the namenode. The full pathname of the file
-      must be specified.  If the value is empty, all hosts are
-      permitted.</description>
-  </property>
-
-  <property>
-    <name>dfs.hosts.exclude</name>
-    <value>${HADOOP_CONF_DIR}/dfs.exclude</value>
-    <description>Names a file that contains a list of hosts that are
-      not permitted to connect to the namenode.  The full pathname of the
-      file must be specified.  If the value is empty, no hosts are
-      excluded.
-    </description>
-  </property>
-  <property>
-    <name>dfs.webhdfs.enabled</name>
-    <value>${DFS_WEBHDFS_ENABLED}</value>
-    <description>Enable or disable webhdfs. Defaults to false</description>
-  </property>
-  <property>
-    <name>dfs.support.append</name>
-    <value>${DFS_SUPPORT_APPEND}</value>
-    <description>Enable or disable append. Defaults to false</description>
-  </property>
-</configuration>

+ 0 - 212
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/log4j.properties

@@ -1,212 +0,0 @@
-# Copyright 2011 The Apache Software Foundation
-# 
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# Define some default values that can be overridden by system properties
-hadoop.root.logger=INFO,console
-hadoop.log.dir=.
-hadoop.log.file=hadoop.log
-
-# Define the root logger to the system property "hadoop.root.logger".
-log4j.rootLogger=${hadoop.root.logger}, EventCounter
-
-# Logging Threshold
-log4j.threshold=ALL
-
-# Null Appender
-log4j.appender.NullAppender=org.apache.log4j.varia.NullAppender
-
-#
-# Rolling File Appender - cap space usage at 5gb.
-#
-hadoop.log.maxfilesize=256MB
-hadoop.log.maxbackupindex=20
-log4j.appender.RFA=org.apache.log4j.RollingFileAppender
-log4j.appender.RFA.File=${hadoop.log.dir}/${hadoop.log.file}
-
-log4j.appender.RFA.MaxFileSize=${hadoop.log.maxfilesize}
-log4j.appender.RFA.MaxBackupIndex=${hadoop.log.maxbackupindex}
-
-log4j.appender.RFA.layout=org.apache.log4j.PatternLayout
-
-# Pattern format: Date LogLevel LoggerName LogMessage
-log4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
-# Debugging Pattern format
-#log4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %-5p %c{2} (%F:%M(%L)) - %m%n
-
-
-#
-# Daily Rolling File Appender
-#
-
-log4j.appender.DRFA=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.DRFA.File=${hadoop.log.dir}/${hadoop.log.file}
-
-# Rollver at midnight
-log4j.appender.DRFA.DatePattern=.yyyy-MM-dd
-
-# 30-day backup
-#log4j.appender.DRFA.MaxBackupIndex=30
-log4j.appender.DRFA.layout=org.apache.log4j.PatternLayout
-
-# Pattern format: Date LogLevel LoggerName LogMessage
-log4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
-# Debugging Pattern format
-#log4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %-5p %c{2} (%F:%M(%L)) - %m%n
-
-
-#
-# console
-# Add "console" to rootlogger above if you want to use this 
-#
-
-log4j.appender.console=org.apache.log4j.ConsoleAppender
-log4j.appender.console.target=System.err
-log4j.appender.console.layout=org.apache.log4j.PatternLayout
-log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%n
-
-#
-# TaskLog Appender
-#
-
-#Default values
-hadoop.tasklog.taskid=null
-hadoop.tasklog.iscleanup=false
-hadoop.tasklog.noKeepSplits=4
-hadoop.tasklog.totalLogFileSize=100
-hadoop.tasklog.purgeLogSplits=true
-hadoop.tasklog.logsRetainHours=12
-
-log4j.appender.TLA=org.apache.hadoop.mapred.TaskLogAppender
-log4j.appender.TLA.taskId=${hadoop.tasklog.taskid}
-log4j.appender.TLA.isCleanup=${hadoop.tasklog.iscleanup}
-log4j.appender.TLA.totalLogFileSize=${hadoop.tasklog.totalLogFileSize}
-
-log4j.appender.TLA.layout=org.apache.log4j.PatternLayout
-log4j.appender.TLA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
-
-#
-#Security appender
-#
-hadoop.security.logger=INFO,NullAppender
-hadoop.security.log.maxfilesize=256MB
-hadoop.security.log.maxbackupindex=20
-log4j.category.SecurityLogger=${hadoop.security.logger}
-hadoop.security.log.file=SecurityAuth-${user.name}.audit
-log4j.appender.RFAS=org.apache.log4j.RollingFileAppender 
-log4j.appender.RFAS.File=${hadoop.log.dir}/${hadoop.security.log.file}
-log4j.appender.RFAS.layout=org.apache.log4j.PatternLayout
-log4j.appender.RFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
-log4j.appender.RFAS.MaxFileSize=${hadoop.security.log.maxfilesize}
-log4j.appender.RFAS.MaxBackupIndex=${hadoop.security.log.maxbackupindex}
-
-#
-# Daily Rolling Security appender
-#
-log4j.appender.DRFAS=org.apache.log4j.DailyRollingFileAppender 
-log4j.appender.DRFAS.File=${hadoop.log.dir}/${hadoop.security.log.file}
-log4j.appender.DRFAS.layout=org.apache.log4j.PatternLayout
-log4j.appender.DRFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n
-log4j.appender.DRFAS.DatePattern=.yyyy-MM-dd
-
-#
-# hdfs audit logging
-#
-hdfs.audit.logger=INFO,NullAppender
-hdfs.audit.log.maxfilesize=256MB
-hdfs.audit.log.maxbackupindex=20
-log4j.logger.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=${hdfs.audit.logger}
-log4j.additivity.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=false
-log4j.appender.RFAAUDIT=org.apache.log4j.RollingFileAppender
-log4j.appender.RFAAUDIT.File=${hadoop.log.dir}/hdfs-audit.log
-log4j.appender.RFAAUDIT.layout=org.apache.log4j.PatternLayout
-log4j.appender.RFAAUDIT.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n
-log4j.appender.RFAAUDIT.MaxFileSize=${hdfs.audit.log.maxfilesize}
-log4j.appender.RFAAUDIT.MaxBackupIndex=${hdfs.audit.log.maxbackupindex}
-
-#
-# mapred audit logging
-#
-mapred.audit.logger=INFO,NullAppender
-mapred.audit.log.maxfilesize=256MB
-mapred.audit.log.maxbackupindex=20
-log4j.logger.org.apache.hadoop.mapred.AuditLogger=${mapred.audit.logger}
-log4j.additivity.org.apache.hadoop.mapred.AuditLogger=false
-log4j.appender.MRAUDIT=org.apache.log4j.RollingFileAppender
-log4j.appender.MRAUDIT.File=${hadoop.log.dir}/mapred-audit.log
-log4j.appender.MRAUDIT.layout=org.apache.log4j.PatternLayout
-log4j.appender.MRAUDIT.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n
-log4j.appender.MRAUDIT.MaxFileSize=${mapred.audit.log.maxfilesize}
-log4j.appender.MRAUDIT.MaxBackupIndex=${mapred.audit.log.maxbackupindex}
-
-# Custom Logging levels
-
-#log4j.logger.org.apache.hadoop.mapred.JobTracker=DEBUG
-#log4j.logger.org.apache.hadoop.mapred.TaskTracker=DEBUG
-#log4j.logger.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=DEBUG
-
-# Jets3t library
-log4j.logger.org.jets3t.service.impl.rest.httpclient.RestS3Service=ERROR
-
-#
-# Event Counter Appender
-# Sends counts of logging messages at different severity levels to Hadoop Metrics.
-#
-log4j.appender.EventCounter=org.apache.hadoop.log.metrics.EventCounter
-
-#
-# Job Summary Appender 
-#
-# Use following logger to send summary to separate file defined by 
-# hadoop.mapreduce.jobsummary.log.file :
-# hadoop.mapreduce.jobsummary.logger=INFO,JSA
-# 
-hadoop.mapreduce.jobsummary.logger=${hadoop.root.logger}
-hadoop.mapreduce.jobsummary.log.file=hadoop-mapreduce.jobsummary.log
-hadoop.mapreduce.jobsummary.log.maxfilesize=256MB
-hadoop.mapreduce.jobsummary.log.maxbackupindex=20
-log4j.appender.JSA=org.apache.log4j.RollingFileAppender
-log4j.appender.JSA.File=${hadoop.log.dir}/${hadoop.mapreduce.jobsummary.log.file}
-log4j.appender.JSA.MaxFileSize=${hadoop.mapreduce.jobsummary.log.maxfilesize}
-log4j.appender.JSA.MaxBackupIndex=${hadoop.mapreduce.jobsummary.log.maxbackupindex}
-log4j.appender.JSA.layout=org.apache.log4j.PatternLayout
-log4j.appender.JSA.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%n
-log4j.logger.org.apache.hadoop.mapred.JobInProgress$JobSummary=${hadoop.mapreduce.jobsummary.logger}
-log4j.additivity.org.apache.hadoop.mapred.JobInProgress$JobSummary=false
-
-#
-# Yarn ResourceManager Application Summary Log 
-#
-# Set the ResourceManager summary log filename
-#yarn.server.resourcemanager.appsummary.log.file=rm-appsummary.log
-# Set the ResourceManager summary log level and appender
-#yarn.server.resourcemanager.appsummary.logger=INFO,RMSUMMARY
-
-# Appender for ResourceManager Application Summary Log
-# Requires the following properties to be set
-#    - hadoop.log.dir (Hadoop Log directory)
-#    - yarn.server.resourcemanager.appsummary.log.file (resource manager app summary log filename)
-#    - yarn.server.resourcemanager.appsummary.logger (resource manager app summary log level and appender)
-
-#log4j.logger.org.apache.hadoop.yarn.server.resourcemanager.RMAppManager$ApplicationSummary=${yarn.server.resourcemanager.appsummary.logger}
-#log4j.additivity.org.apache.hadoop.yarn.server.resourcemanager.RMAppManager$ApplicationSummary=false
-#log4j.appender.RMSUMMARY=org.apache.log4j.RollingFileAppender
-#log4j.appender.RMSUMMARY.File=${hadoop.log.dir}/${yarn.server.resourcemanager.appsummary.log.file}
-#log4j.appender.RMSUMMARY.MaxFileSize=256MB
-#log4j.appender.RMSUMMARY.MaxBackupIndex=20
-#log4j.appender.RMSUMMARY.layout=org.apache.log4j.PatternLayout
-#log4j.appender.RMSUMMARY.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n

+ 0 - 26
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/mapred-queue-acls.xml

@@ -1,26 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<configuration>
-<property>
-<name>mapred.queue.default.acl-submit-job</name>
-<value>*</value>
-</property>
-<property>
-<name>mapred.queue.default.acl-administer-jobs</name>
-<value>*</value>
-</property>
-</configuration>

+ 0 - 300
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/mapred-site.xml

@@ -1,300 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
- 
-  <property>
-    <name>mapred.tasktracker.tasks.sleeptime-before-sigkill</name>
-    <value>250</value>
-    <description>Normally, this is the amount of time before killing
-      processes, and the recommended-default is 5.000 seconds - a value of
-      5000 here.  In this case, we are using it solely to blast tasks before
-      killing them, and killing them very quickly (1/4 second) to guarantee
-      that we do not leave VMs around for later jobs.
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.system.dir</name>
-    <value>/mapred/mapredsystem</value>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>mapred.job.tracker</name>
-    <value>${HADOOP_JT_HOST}:9000</value>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>mapred.job.tracker.http.address</name>
-    <value>${HADOOP_JT_HOST}:50030</value>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>mapred.local.dir</name>
-    <value>${HADOOP_MAPRED_DIR}</value>
-    <final>true</final>
-  </property>
-
-  <property>
-    <name>mapreduce.cluster.administrators</name>
-    <value>${HADOOP_MR_USER}</value>
-  </property>
-
-  <property>
-    <name>mapred.map.tasks.speculative.execution</name>
-    <value>false</value>
-    <description>If true, then multiple instances of some map tasks
-               may be executed in parallel.</description>
-  </property>
-
-  <property>
-    <name>mapred.reduce.tasks.speculative.execution</name>
-    <value>false</value>
-    <description>If true, then multiple instances of some reduce tasks
-               may be executed in parallel.</description>
-  </property>
-
-  <property>
-    <name>mapred.output.compression.type</name>
-    <value>BLOCK</value>
-    <description>If the job outputs are to compressed as SequenceFiles, how 
-       should they be compressed? Should be one of NONE, RECORD or BLOCK.
-    </description>
-  </property>
-
-  <property>
-    <name>jetty.connector</name>
-    <value>org.mortbay.jetty.nio.SelectChannelConnector</value>
-  </property>
-
-  <property>
-    <name>mapred.task.tracker.task-controller</name>
-    <value>${TASK_CONTROLLER}</value>
-  </property>
-
-  <property>
-    <name>mapred.child.root.logger</name>
-    <value>INFO,TLA</value>
-  </property>
-
-  <property>
-    <name>mapred.child.java.opts</name>
-    <value>-server -Xmx640m -Djava.net.preferIPv4Stack=true</value>
-  </property>
-
-  <property>
-    <name>mapred.job.tracker.persist.jobstatus.active</name>
-    <value>true</value>
-    <description>Indicates if persistency of job status information is
-      active or not.
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.job.tracker.persist.jobstatus.dir</name>
-    <value>file:///${HADOOP_LOG_DIR}/${HADOOP_MR_USER}/jobstatus</value>
-    <description>The directory where the job status information is persisted
-      in a file system to be available after it drops of the memory queue and
-      between jobtracker restarts.
-    </description>
-  </property>
-
-  <property>
-    <name>mapred.job.tracker.history.completed.location</name>
-    <value>/mapred/history/done</value>
-  </property>
-
-  <property>
-    <name>mapred.heartbeats.in.second</name>
-    <value>200</value>
-    <description>to enable HADOOP:5784</description>
-  </property>
-
-  <property>
-    <name>mapreduce.tasktracker.outofband.heartbeat</name>
-    <value>true</value>
-    <description>to enable MAPREDUCE:270</description>
-  </property>
-
-  <property>
-    <name>mapred.jobtracker.maxtasks.per.job</name>
-    <value>200000</value>
-    <final>true</final>
-    <description>The maximum number of tasks for a single job.
-      A value of -1 indicates that there is no maximum.  
-    </description>
-  </property>
-
-  <property>
-    <name>mapreduce.jobtracker.kerberos.principal</name>
-    <value>jt/_HOST@${local.realm}</value>   
-    <description>
-       JT principal
-    </description>
-  </property>
-
-  <property>
-    <name>mapreduce.tasktracker.kerberos.principal</name>
-    <value>tt/_HOST@${local.realm}</value>   
-    <description>       
-       TT principal.
-    </description>
-  </property>
-
-
-  <property>
-    <name>hadoop.job.history.user.location</name>
-    <value>none</value>
-  </property>
-
-  <property>
-    <name>mapreduce.jobtracker.keytab.file</name>
-    <value>/etc/security/keytabs/jt.service.keytab</value>
-    <description>
-        The keytab for the jobtracker principal.
-    </description>
-  </property>
-
-  <property>
-    <name>mapreduce.tasktracker.keytab.file</name>
-    <value>/etc/security/keytabs/tt.service.keytab</value>
-    <description>The filename of the keytab for the task tracker</description>
-  </property>
-
-  <property>
-    <name>mapreduce.jobtracker.staging.root.dir</name>
-    <value>/user</value>
-    <description>The Path prefix for where the staging directories should be 
-      placed. The next level is always the user's
-      name. It is a path in the default file system.
-    </description>
-  </property>
-
-
-  <property>
-    <name>mapreduce.job.acl-modify-job</name>
-    <value></value>
-  </property>
-
-  <property>
-    <name>mapreduce.job.acl-view-job</name>
-    <value>Dr.Who</value>
-  </property>
-
-  <property>
-    <name>mapreduce.tasktracker.group</name>
-    <value>${HADOOP_GROUP}</value>
-    <description>The group that the task controller uses for accessing the
-      task controller. The mapred user must be a member and users should *not*
-      be members.
-    </description> 
-  </property>
-
-  <property>
-    <name>mapred.acls.enabled</name>
-    <value>true</value>
-  </property>
-  
-  <property>
-    <name>mapred.jobtracker.taskScheduler</name>
-    <value>org.apache.hadoop.mapred.CapacityTaskScheduler</value>
-  </property>
-  <property>
-    <name>mapred.queue.names</name>
-    <value>default</value>
-  </property>
-
-  <!-- settings for the history server -->
-  <property>
-    <name>mapreduce.history.server.embedded</name>
-    <value>false</value>
-  </property>
-  <property>
-    <name>mapreduce.history.server.http.address</name>
-    <value>${HADOOP_JT_HOST}:51111</value>
-  </property>
-  <property>
-    <name>mapreduce.jobhistory.kerberos.principal</name>
-    <value>jt/_HOST@${local.realm}</value>   
-    <description>history server principal</description>
-  </property>
-  <property>
-    <name>mapreduce.jobhistory.keytab.file</name>
-    <value>/etc/security/keytabs/jt.service.keytab</value>
-    <description>
-        The keytab for the jobtracker principal.
-    </description>
-  </property>
-  
-  <property>
-    <name>mapred.hosts</name>
-    <value>${HADOOP_CONF_DIR}/mapred.include</value>
-    <description>Names a file that contains the list of nodes that may
-    connect to the jobtracker.  If the value is empty, all hosts are
-    permitted.</description>
-  </property>
-          
-  <property>
-    <name>mapred.hosts.exclude</name>
-    <value>${HADOOP_CONF_DIR}/mapred.exclude</value>
-    <description>Names a file that contains the list of hosts that
-      should be excluded by the jobtracker.  If the value is empty, no
-      hosts are excluded.</description>
-  </property>
-  <property>
-    <name>mapred.jobtracker.retirejob.check</name>
-    <value>10000</value>
-  </property>
-  <property>
-    <name>mapred.jobtracker.retirejob.interval</name>
-    <value>0</value>
-  </property>
-
-  <property>
-    <name>mapreduce.cluster.mapmemory.mb</name>
-    <value>${MAPREDUCE_CLUSTER_MAPMEMORY_MB}</value>
-  </property>
-
-  <property>
-    <name>mapreduce.cluster.reducememory.mb</name>
-    <value>${MAPREDUCE_CLUSTER_REDUCEMEMORY_MB}</value>
-  </property>
-
-  <property>
-    <name>mapreduce.jobtracker.maxmapmemory.mb</name>
-    <value>${MAPREDUCE_JOBTRACKER_MAXMAPMEMORY_MB}</value>
-  </property>
-
-  <property>
-    <name>mapreduce.jobtracker.maxreducememory.mb</name>
-    <value>${MAPREDUCE_JOBTRACKER_MAXREDUCEMEMORY_MB}</value>
-  </property>
-
-  <property>
-    <name>mapreduce.map.memory.mb</name>
-    <value>${MAPREDUCE_MAP_MEMORY_MB}</value>
-  </property>
-
-  <property>
-    <name>mapreduce.reduce.memory.mb</name>
-    <value>${MAPREDUCE_REDUCE_MEMORY_MB}</value>
-  </property>
-</configuration>

+ 0 - 18
hadoop-common-project/hadoop-common/src/main/packages/templates/conf/taskcontroller.cfg

@@ -1,18 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-mapreduce.cluster.local.dir=${HADOOP_MAPRED_DIR}
-mapreduce.tasktracker.group=${HADOOP_GROUP}
-hadoop.log.dir=${HADOOP_LOG_DIR}/${HADOOP_MR_USER}

+ 0 - 168
hadoop-common-project/hadoop-common/src/main/packages/update-hadoop-env.sh

@@ -1,168 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# This script configures hadoop-env.sh and symlinkis directories for 
-# relocating RPM locations.
-
-usage() {
-  echo "
-usage: $0 <parameters>
-  Required parameters:
-     --prefix=PREFIX             path to install into
-
-  Optional parameters:
-     --arch=i386                 OS Architecture
-     --bin-dir=PREFIX/bin        Executable directory
-     --conf-dir=/etc/hadoop      Configuration directory
-     --log-dir=/var/log/hadoop   Log directory
-     --pid-dir=/var/run          PID file location
-     --sbin-dir=PREFIX/sbin      System executable directory
-  "
-  exit 1
-}
-
-template_generator() {
-  REGEX='(\$\{[a-zA-Z_][a-zA-Z_0-9]*\})'
-  cat $1 |
-  while read line ; do
-    while [[ "$line" =~ $REGEX ]] ; do
-      LHS=${BASH_REMATCH[1]}
-      RHS="$(eval echo "\"$LHS\"")"
-      line=${line//$LHS/$RHS}
-    done
-    echo $line >> $2
-  done
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'arch:' \
-  -l 'prefix:' \
-  -l 'bin-dir:' \
-  -l 'conf-dir:' \
-  -l 'lib-dir:' \
-  -l 'log-dir:' \
-  -l 'pid-dir:' \
-  -l 'sbin-dir:' \
-  -l 'uninstall' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --arch)
-      ARCH=$2 ; shift 2
-      ;;
-    --prefix)
-      PREFIX=$2 ; shift 2
-      ;;
-    --bin-dir)
-      BIN_DIR=$2 ; shift 2
-      ;;
-    --log-dir)
-      LOG_DIR=$2 ; shift 2
-      ;;
-    --lib-dir)
-      LIB_DIR=$2 ; shift 2
-      ;;
-    --conf-dir)
-      CONF_DIR=$2 ; shift 2
-      ;;
-    --pid-dir)
-      PID_DIR=$2 ; shift 2
-      ;;
-    --sbin-dir)
-      SBIN_DIR=$2 ; shift 2
-      ;;
-    --uninstall)
-      UNINSTALL=1; shift
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-for var in PREFIX; do
-  if [ -z "$(eval "echo \$$var")" ]; then
-    echo Missing param: $var
-    usage
-  fi
-done
-
-ARCH=${ARCH:-i386}
-HADOOP_PREFIX=$PREFIX
-HADOOP_BIN_DIR=${BIN_DIR:-$PREFIX/bin}
-HADOOP_CONF_DIR=${CONF_DIR:-$PREFIX/etc/hadoop}
-HADOOP_LIB_DIR=${LIB_DIR:-$PREFIX/lib}
-HADOOP_LOG_DIR=${LOG_DIR:-$PREFIX/var/log}
-HADOOP_PID_DIR=${PID_DIR:-$PREFIX/var/run}
-HADOOP_SBIN_DIR=${SBIN_DIR:-$PREFIX/sbin}
-UNINSTALL=${UNINSTALL:-0}
-
-if [ "${ARCH}" != "i386" ]; then
-  HADOOP_LIB_DIR=${HADOOP_LIB_DIR}64
-fi
-
-if [ "${UNINSTALL}" -eq "1" ]; then
-  # Remove symlinks
-  if [ "${HADOOP_CONF_DIR}" != "${HADOOP_PREFIX}/etc/hadoop" ]; then
-    rm -rf ${HADOOP_PREFIX}/etc/hadoop
-  fi
-  rm -f /etc/default/hadoop-env.sh
-  rm -f /etc/profile.d/hadoop-env.sh
-else
-  # Create symlinks
-  if [ "${HADOOP_CONF_DIR}" != "${HADOOP_PREFIX}/etc/hadoop" ]; then
-    mkdir -p ${HADOOP_PREFIX}/etc
-    ln -sf ${HADOOP_CONF_DIR} ${HADOOP_PREFIX}/etc/hadoop
-  fi
-  ln -sf ${HADOOP_CONF_DIR}/hadoop-env.sh /etc/default/hadoop-env.sh
-  ln -sf ${HADOOP_CONF_DIR}/hadoop-env.sh /etc/profile.d/hadoop-env.sh
-
-  mkdir -p ${HADOOP_LOG_DIR}
-  chown root:hadoop ${HADOOP_LOG_DIR}
-  chmod 775 ${HADOOP_LOG_DIR}
-
-  if [ ! -d ${HADOOP_PID_DIR} ]; then
-    mkdir -p ${HADOOP_PID_DIR}
-    chown root:hadoop ${HADOOP_PID_DIR}
-    chmod 775 ${HADOOP_PID_DIR}
-  fi
-
-  TFILE="/tmp/$(basename $0).$$.tmp"
-  if [ -z "${JAVA_HOME}" ]; then
-    if [ -e /etc/debian_version ]; then
-      JAVA_HOME=`update-alternatives --config java | grep java | cut -f2 -d':' | cut -f2 -d' ' | sed -e 's/\/bin\/java//'`
-    else
-      JAVA_HOME=/usr/java/default
-    fi
-  fi
-  template_generator ${HADOOP_CONF_DIR}/hadoop-env.sh.template $TFILE
-  cp ${TFILE} ${CONF_DIR}/hadoop-env.sh
-  rm -f ${TFILE}
-fi

+ 0 - 15
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/conffile

@@ -1,15 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-/etc/hadoop

+ 0 - 24
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/control

@@ -1,24 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-Package: hadoop-hdfs
-Version: @version@
-Section: misc
-Priority: optional
-Provides: hadoop-hdfs
-Architecture: all
-Depends: openjdk-6-jre-headless, hadoop-common
-Maintainer: Apache Software Foundation <general@hadoop.apache.org>
-Description: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing.
-Distribution: development

+ 0 - 24
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/postinst

@@ -1,24 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-hdfs-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop

+ 0 - 19
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/postrm

@@ -1,19 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-/usr/sbin/groupdel hadoop 2> /dev/null >dev/null
-exit 0

+ 0 - 18
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/preinst

@@ -1,18 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-/usr/sbin/useradd --comment "Hadoop HDFS" -u 201 --shell /bin/bash -M -r --groups hadoop --home /var/lib/hadoop/hdfs hdfs 2> /dev/null || :

+ 0 - 25
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/hadoop.control/prerm

@@ -1,25 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-hdfs-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop \
-  --uninstal

+ 0 - 142
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-datanode

@@ -1,142 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-datanode	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Name Node server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-datanode: start and stop the Apache Hadoop Data Node daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-datanode_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-datanode_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Data Node server not in use (/etc/hadoop/hadoop-datanode_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Data Node server" "hadoop-datanode"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Data Node server" "hadoop-datanode"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Data Node server" "hadoop-datanode"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Data Node server" "hadoop-datanode"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start datanode; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid ${JAVA_HOME}/bin/java hadoop-datanode && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-datanode {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 154
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-namenode

@@ -1,154 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-namenode	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Name Node server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-namenode: start and stop the Apache Hadoop Name Node daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-namenode_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-namenode_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Name Node server not in use (/etc/hadoop/hadoop-namenode_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-format() {
-    su -c '${HADOOP_PREFIX}/bin/hadoop --config ${HADOOP_CONF_DIR} namenode -format' hdfs
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Name Node server" "hadoop-namenode"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Name Node server" "hadoop-namenode"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  format)
-	log_daemon_msg "Formatting Apache Hadoop Name Node" "hadoop-namenode"
-	format
-	if [ $? -eq 0 ]; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Name Node server" "hadoop-namenode"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Name Node server" "hadoop-namenode"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start namenode; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid ${JAVA_HOME}/bin/java hadoop-namenode && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-namenode {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 142
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/deb/init.d/hadoop-secondarynamenode

@@ -1,142 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-secondarynamenode	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Name Node server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-secondarynamenode: start and stop the Apache Hadoop Secondary Name Node daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-secondarynamenode_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-secondarynamenode_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Name Node server not in use (/etc/hadoop/hadoop-secondarynamenode_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Secondary Name Node server" "hadoop-secondarynamenode"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start secondarynamenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Secondary Name Node server" "hadoop-secondarynamenode"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Secondary Name Node server" "hadoop-secondarynamenode"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start secondarynamenode; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Secondary Name Node server" "hadoop-secondarynamenode"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid -c hdfs -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start secondarynamenode; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid ${JAVA_HOME}/bin/java hadoop-secondarynamenode && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-secondarynamenode {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 84
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-datanode

@@ -1,84 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop datanode
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop datanode
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-hdfs-datanode.pid"
-desc="Hadoop datanode daemon"
-
-start() {
-  echo -n $"Starting $desc (hadoop-datanode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start datanode
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-datanode
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-datanode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop datanode
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-datanode $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-datanode ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 98
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-namenode

@@ -1,98 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop namenode
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop namenode
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-hdfs-namenode.pid"
-desc="Hadoop namenode daemon"
-
-start() {
-  echo -n $"Starting $desc (hadoop-namenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start namenode $1
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-namenode
-  return $RETVAL
-}
-
-upgrade() {
-  start -upgrade
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-namenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop namenode
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-namenode $PIDFILE
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-restart() {
-  stop
-  start
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-namenode ] && restart || :
-}
-
-format() {
-  daemon --user hdfs ${HADOOP_PREFIX}/bin/hdfs namenode -format
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  upgrade)
-    upgrade
-    ;;
-  format)
-    format
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart|try-restart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|try-restart|upgrade}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 92
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/init.d/hadoop-secondarynamenode

@@ -1,92 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop secondarynamenode
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop secondarynamenode
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-hdfs-secondarynamenode.pid"
-desc="Hadoop secondary namenode daemon"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-start() {
-  echo -n $"Starting $desc (hadoop-secondarynamenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start secondarynamenode $1
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-secondarynamenode
-  return $RETVAL
-}
-
-upgrade() {
-  start -upgrade
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-secondarynamenode): "
-  daemon --user hdfs ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop secondarynamenode
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-secondarynamenode $PIDFILE
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-restart() {
-  stop
-  start
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-secondarynamenode ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  upgrade)
-    upgrade
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart|try-restart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|try-restart|upgrade}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 176
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/rpm/spec/hadoop-hdfs.spec

@@ -1,176 +0,0 @@
-#   Licensed to the Apache Software Foundation (ASF) under one or more
-#   contributor license agreements.  See the NOTICE file distributed with
-#   this work for additional information regarding copyright ownership.
-#   The ASF licenses this file to You under the Apache License, Version 2.0
-#   (the "License"); you may not use this file except in compliance with
-#   the License.  You may obtain a copy of the License at
-#
-#       http://www.apache.org/licenses/LICENSE-2.0
-#
-#   Unless required by applicable law or agreed to in writing, software
-#   distributed under the License is distributed on an "AS IS" BASIS,
-#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-#   See the License for the specific language governing permissions and
-#   limitations under the License.
-
-#
-# RPM Spec file for Hadoop version @version@
-#
-
-%define name          hadoop-hdfs
-%define version       @version@
-%define release       @package.release@
-%define major_version %(echo %{version} | cut -d. -f -2)
-
-# Installation Locations
-%define _prefix      @package.prefix@
-%define _bin_dir     %{_prefix}/bin
-%define _conf_dir    @package.conf.dir@
-%define _lib_dir     %{_prefix}/lib
-%define _lib64_dir   %{_prefix}/lib64
-%define _libexec_dir %{_prefix}/libexec
-%define _log_dir     @package.log.dir@
-%define _pid_dir     @package.pid.dir@
-%define _sbin_dir    %{_prefix}/sbin
-%define _share_dir   %{_prefix}/share
-%define _var_dir     @package.var.dir@
-
-# Build time settings
-%define _build_dir  @package.build.dir@
-%define _final_name @final.name@
-%define debug_package %{nil}
-
-# Disable brp-java-repack-jars for aspect J
-%define __os_install_post    \
-    /usr/lib/rpm/redhat/brp-compress \
-    %{!?__debug_package:/usr/lib/rpm/redhat/brp-strip %{__strip}} \
-    /usr/lib/rpm/redhat/brp-strip-static-archive %{__strip} \
-    /usr/lib/rpm/redhat/brp-strip-comment-note %{__strip} %{__objdump} \
-    /usr/lib/rpm/brp-python-bytecompile %{nil}
-
-# RPM searches perl files for dependancies and this breaks for non packaged perl lib
-# like thrift so disable this
-%define _use_internal_dependency_generator 0
-
-%ifarch i386
-%global hadoop_arch Linux-i386-32
-%endif
-%ifarch amd64 x86_64
-%global hadoop_arch Linux-amd64-64
-%endif
-%ifarch noarch
-%global hadoop_arch ""
-%endif
-
-Summary: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing
-License: Apache License, Version 2.0
-URL: http://hadoop.apache.org/core/
-Vendor: Apache Software Foundation
-Group: Development/Libraries
-Name: %{name}
-Version: %{version}
-Release: %{release} 
-Source0: %{_final_name}-bin.tar.gz
-Prefix: %{_prefix}
-Prefix: %{_conf_dir}
-Prefix: %{_log_dir}
-Prefix: %{_pid_dir}
-Buildroot: %{_build_dir}
-Requires: sh-utils, textutils, /usr/sbin/useradd, /usr/sbin/usermod, /sbin/chkconfig, /sbin/service, hadoop-common >= %{major_version}.0, hadoop-common <= %{major_version}.9999
-AutoReqProv: no
-Provides: hadoop-hdfs
-
-%description
-The Apache Hadoop project develops open-source software for reliable, scalable, 
-distributed computing.  Hadoop includes these subprojects:
-
-HDFS: A distributed file system that provides high throughput access to application data.
-
-%prep
-%setup -n %{_final_name}
-
-%build
-if [ -d ${RPM_BUILD_DIR}%{_prefix} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_prefix}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_log_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_log_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_conf_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_conf_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_pid_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_pid_dir}
-fi
-
-mkdir -p ${RPM_BUILD_DIR}%{_prefix}
-mkdir -p ${RPM_BUILD_DIR}%{_bin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_lib_dir}
-%ifarch amd64 x86_64
-mkdir -p ${RPM_BUILD_DIR}%{_lib64_dir}
-%endif
-mkdir -p ${RPM_BUILD_DIR}%{_libexec_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_log_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_conf_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_pid_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_sbin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_share_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_var_dir}
-mkdir -p ${RPM_BUILD_DIR}/etc/init.d
-
-cp ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-namenode.redhat ${RPM_BUILD_DIR}/etc/init.d/hadoop-namenode
-cp ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-datanode.redhat ${RPM_BUILD_DIR}/etc/init.d/hadoop-datanode
-rm -f ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-namenode.*
-rm -f ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-namenode.*
-
-chmod 0755 ${RPM_BUILD_DIR}/etc/init.d/hadoop-*
-
-
-#########################
-#### INSTALL SECTION ####
-#########################
-%install
-mv ${RPM_BUILD_DIR}/%{_final_name}/bin/* ${RPM_BUILD_DIR}%{_bin_dir}
-rm ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/configuration.xsl
-rm ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/hadoop-metrics2.properties
-mv ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/* ${RPM_BUILD_DIR}%{_conf_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/lib/* ${RPM_BUILD_DIR}%{_lib_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/libexec/* ${RPM_BUILD_DIR}%{_libexec_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/sbin/* ${RPM_BUILD_DIR}%{_sbin_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/share/* ${RPM_BUILD_DIR}%{_share_dir}
-rm -rf ${RPM_BUILD_DIR}/%{_final_name}/etc
-
-%pre
-getent group hadoop 2>/dev/null >/dev/null || /usr/sbin/groupadd -g 123 -r hadoop
-/usr/sbin/useradd --comment "Hadoop HDFS" -u 201 --shell /bin/bash -M -r --groups hadoop --home %{_var_dir}/hdfs hdfs 2> /dev/null || :
-
-%post
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-hdfs-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3}
-
-%preun
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-hdfs-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3} \
-       --uninstall
-
-%files 
-%defattr(-,root,root)
-%attr(0755,root,hadoop) %{_log_dir}
-%attr(0775,root,hadoop) %{_pid_dir}
-%config(noreplace) %{_conf_dir}/hdfs-site.xml
-%{_prefix}
-%attr(0775,root,root) /etc/init.d/hadoop-namenode
-%attr(0775,root,root) /etc/init.d/hadoop-datanode

+ 0 - 45
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/templates/conf/hdfs-site.xml

@@ -1,45 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<!--
-  Licensed under the Apache License, Version 2.0 (the "License");
-  you may not use this file except in compliance with the License.
-  You may obtain a copy of the License at
-
-    http://www.apache.org/licenses/LICENSE-2.0
-
-  Unless required by applicable law or agreed to in writing, software
-  distributed under the License is distributed on an "AS IS" BASIS,
-  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-  See the License for the specific language governing permissions and
-  limitations under the License. See accompanying LICENSE file.
--->
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-  <property>
-    <name>dfs.replication</name>
-    <value>${HADOOP_REPLICATION}</value>
-  </property>
-  <property>
-    <name>dfs.namenode.name.dir</name>
-    <value>${HADOOP_NN_DIR}</value>
-  </property>
-  <property>
-    <name>dfs.datanode.data.dir</name>
-    <value>${HADOOP_DN_DIR}</value>
-  </property>
-  <property>
-    <name>hadoop.tmp.dir</name>
-    <value>/tmp</value>
-  </property>
-  <property>
-    <name>dfs.namenode.safemode.threshold-pct</name>
-    <value>1.0f</value>
-  </property>
-  <property>
-    <name>dfs.namenode.safemode.extension</name>
-    <value>3</value>
-  </property>
-</configuration>
-

+ 0 - 128
hadoop-hdfs-project/hadoop-hdfs/src/main/packages/update-hdfs-env.sh

@@ -1,128 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# This script configures hdfs-env.sh and symlinkis directories for 
-# relocating RPM locations.
-
-usage() {
-  echo "
-usage: $0 <parameters>
-  Required parameters:
-     --prefix=PREFIX             path to install into
-
-  Optional parameters:
-     --arch=i386                 OS Architecture
-     --bin-dir=PREFIX/bin        Executable directory
-     --conf-dir=/etc/hadoop      Configuration directory
-     --log-dir=/var/log/hadoop   Log directory
-     --pid-dir=/var/run          PID file location
-     --sbin-dir=PREFIX/sbin      System executable directory
-  "
-  exit 1
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'arch:' \
-  -l 'prefix:' \
-  -l 'bin-dir:' \
-  -l 'conf-dir:' \
-  -l 'lib-dir:' \
-  -l 'log-dir:' \
-  -l 'pid-dir:' \
-  -l 'sbin-dir:' \
-  -l 'uninstall' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --arch)
-      ARCH=$2 ; shift 2
-      ;;
-    --prefix)
-      PREFIX=$2 ; shift 2
-      ;;
-    --bin-dir)
-      BIN_DIR=$2 ; shift 2
-      ;;
-    --log-dir)
-      LOG_DIR=$2 ; shift 2
-      ;;
-    --lib-dir)
-      LIB_DIR=$2 ; shift 2
-      ;;
-    --conf-dir)
-      CONF_DIR=$2 ; shift 2
-      ;;
-    --pid-dir)
-      PID_DIR=$2 ; shift 2
-      ;;
-    --sbin-dir)
-      SBIN_DIR=$2 ; shift 2
-      ;;
-    --uninstall)
-      UNINSTALL=1; shift
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-for var in PREFIX; do
-  if [ -z "$(eval "echo \$$var")" ]; then
-    echo Missing param: $var
-    usage
-  fi
-done
-
-ARCH=${ARCH:-i386}
-BIN_DIR=${BIN_DIR:-$PREFIX/bin}
-CONF_DIR=${CONF_DIR:-$PREFIX/etc/hadoop}
-LIB_DIR=${LIB_DIR:-$PREFIX/lib}
-LOG_DIR=${LOG_DIR:-$PREFIX/var/log}
-PID_DIR=${PID_DIR:-$PREFIX/var/run}
-SBIN_DIR=${SBIN_DIR:-$PREFIX/sbin}
-UNINSTALL=${UNINSTALL:-0}
-
-if [ "${ARCH}" != "i386" ]; then
-  LIB_DIR=${LIB_DIR}64
-fi
-
-if [ "${UNINSTALL}" -ne "1" ]; then
-  mkdir -p ${LOG_DIR}
-  chown hdfs:hadoop ${LOG_DIR}
-  chmod 755 ${LOG_DIR}
-
-  if [ ! -d ${PID_DIR} ]; then
-    mkdir -p ${PID_DIR}
-    chown root:hadoop ${PID_DIR}
-    chmod 775 ${PID_DIR}
-  fi
-
-fi

+ 0 - 15
hadoop-mapreduce-project/src/packages/deb/hadoop.control/conffile

@@ -1,15 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-/etc/hadoop

+ 0 - 24
hadoop-mapreduce-project/src/packages/deb/hadoop.control/control

@@ -1,24 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-Package: hadoop-mapreduce
-Version: @version@
-Section: misc
-Priority: optional
-Provides: hadoop-mapreduce
-Architecture: all
-Depends: openjdk-6-jre-headless, hadoop-common, hadoop-hdfs
-Maintainer: Apache Software Foundation <general@hadoop.apache.org>
-Description: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing.
-Distribution: development

+ 0 - 24
hadoop-mapreduce-project/src/packages/deb/hadoop.control/postinst

@@ -1,24 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-mapred-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop

+ 0 - 19
hadoop-mapreduce-project/src/packages/deb/hadoop.control/postrm

@@ -1,19 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-/usr/sbin/groupdel hadoop 2> /dev/null >dev/null
-exit 0

+ 0 - 18
hadoop-mapreduce-project/src/packages/deb/hadoop.control/preinst

@@ -1,18 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-/usr/sbin/useradd --comment "Hadoop MapReduce" -u 202 --shell /bin/bash -M -r --groups hadoop --home /var/lib/hadoop/mapred mapred 2> /dev/null || :

+ 0 - 25
hadoop-mapreduce-project/src/packages/deb/hadoop.control/prerm

@@ -1,25 +0,0 @@
-#!/bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-bash /usr/sbin/update-mapred-env.sh \
-  --prefix=/usr \
-  --bin-dir=/usr/bin \
-  --sbin-dir=/usr/sbin \
-  --conf-dir=/etc/hadoop \
-  --log-dir=/var/log/hadoop \
-  --pid-dir=/var/run/hadoop \
-  --uninstal

+ 0 - 143
hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-historyserver

@@ -1,143 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-historyserver	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Job Tracker server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-historyserver: start and stop the Apache Hadoop Job History daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-historyserver_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-historyserver_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Job History server not in use (/etc/hadoop/hadoop-historyserver_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Job History server" "hadoop-historyserver"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start historyserver; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Job History server" "hadoop-historyserver"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job History server" "hadoop-historyserver"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start historyserver; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job History server" "hadoop-historyserver"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start historyserver; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid ${JAVA_HOME}/bin/java hadoop-historyserver && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-historyserver {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 142
hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-jobtracker

@@ -1,142 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-jobtracker	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Job Tracker server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-jobtracker: start and stop the Apache Hadoop Job Tracker daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-jobtracker_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-jobtracker_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Job Tracker server not in use (/etc/hadoop/hadoop-jobtracker_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Job Tracker server" "hadoop-jobtracker"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start jobtracker; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid ${JAVA_HOME}/bin/java hadoop-jobtracker && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-jobtracker {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 142
hadoop-mapreduce-project/src/packages/deb/init.d/hadoop-tasktracker

@@ -1,142 +0,0 @@
-#! /bin/sh
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-### BEGIN INIT INFO
-# Provides:		hadoop-tasktracker	
-# Required-Start:	$remote_fs $syslog
-# Required-Stop:	$remote_fs $syslog
-# Default-Start:	2 3 4 5
-# Default-Stop:		
-# Short-Description:	Apache Hadoop Task Tracker server
-### END INIT INFO
-
-set -e
-
-# /etc/init.d/hadoop-tasktracker: start and stop the Apache Hadoop Task Tracker daemon
-
-test -x /usr/bin/hadoop || exit 0
-( /usr/bin/hadoop 2>&1 | grep -q hadoop ) 2>/dev/null || exit 0
-
-umask 022
-
-if test -f /etc/default/hadoop-env.sh; then
-    . /etc/default/hadoop-env.sh
-fi
-
-. /lib/lsb/init-functions
-
-# Are we running from init?
-run_by_init() {
-    ([ "$previous" ] && [ "$runlevel" ]) || [ "$runlevel" = S ]
-}
-
-check_for_no_start() {
-    # forget it if we're trying to start, and /etc/hadoop/hadoop-tasktracker_not_to_be_run exists
-    if [ -e /etc/hadoop/hadoop-tasktracker_not_to_be_run ]; then 
-	if [ "$1" = log_end_msg ]; then
-	    log_end_msg 0
-	fi
-	if ! run_by_init; then
-	    log_action_msg "Apache Hadoop Task Tracker server not in use (/etc/hadoop/hadoop-tasktracker_not_to_be_run)"
-	fi
-	exit 0
-    fi
-}
-
-check_privsep_dir() {
-    # Create the PrivSep empty dir if necessary
-    if [ ! -d ${HADOOP_PID_DIR} ]; then
-	mkdir -p ${HADOOP_PID_DIR}
-        chown root:hadoop ${HADOOP_PID_DIR}
-	chmod 0775 ${HADOOP_PID_DIR} 
-    fi
-}
-
-export PATH="${PATH:+$PATH:}/usr/sbin:/usr/bin"
-
-case "$1" in
-  start)
-	check_privsep_dir
-	check_for_no_start
-	log_daemon_msg "Starting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-  stop)
-	log_daemon_msg "Stopping Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	if start-stop-daemon --stop --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	start-stop-daemon --stop --quiet --oknodo --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid
-	check_for_no_start log_end_msg
-	if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-	    log_end_msg 0
-	else
-	    log_end_msg 1
-	fi
-	;;
-
-  try-restart)
-	check_privsep_dir
-	log_daemon_msg "Restarting Apache Hadoop Task Tracker server" "hadoop-tasktracker"
-	set +e
-	start-stop-daemon --stop --quiet --retry 30 --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid
-	RET="$?"
-	set -e
-	case $RET in
-	    0)
-		# old daemon stopped
-		check_for_no_start log_end_msg
-		if start-stop-daemon --start --quiet --oknodo --pidfile ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid -c mapred -x ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh -- --config ${HADOOP_CONF_DIR} start tasktracker; then
-		    log_end_msg 0
-		else
-		    log_end_msg 1
-		fi
-		;;
-	    1)
-		# daemon not running
-		log_progress_msg "(not running)"
-		log_end_msg 0
-		;;
-	    *)
-		# failed to stop
-		log_progress_msg "(failed to stop)"
-		log_end_msg 1
-		;;
-	esac
-	;;
-
-  status)
-	status_of_proc -p ${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid ${JAVA_HOME}/bin/java hadoop-tasktracker && exit 0 || exit $?
-	;;
-
-  *)
-	log_action_msg "Usage: /etc/init.d/hadoop-tasktracker {start|stop|restart|try-restart|status}"
-	exit 1
-esac
-
-exit 0

+ 0 - 85
hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-historyserver

@@ -1,85 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop historyserver
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop historyserver
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-mapred-historyserver.pid"
-desc="Hadoop historyserver daemon"
-export HADOOP_PREFIX=${HADOOP_PREFIX:-/usr}
-
-start() {
-  echo -n $"Starting $desc (hadoop-historyserver): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start historyserver
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-historyserver
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-historyserver): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop historyserver
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-historyserver $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-historyserver ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 84
hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-jobtracker

@@ -1,84 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop jobtracker
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop jobtracker
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-mapred-jobtracker.pid"
-desc="Hadoop jobtracker daemon"
-
-start() {
-  echo -n $"Starting $desc (hadoop-jobtracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start jobtracker
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-jobtracker
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-jobtracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop jobtracker
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-jobtracker $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-jobtracker ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 84
hadoop-mapreduce-project/src/packages/rpm/init.d/hadoop-tasktracker

@@ -1,84 +0,0 @@
-#!/bin/bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# 
-# Starts a Hadoop tasktracker
-# 
-# chkconfig: 2345 90 10
-# description: Hadoop tasktracker
-
-source /etc/rc.d/init.d/functions
-source /etc/default/hadoop-env.sh
-
-RETVAL=0
-PIDFILE="${HADOOP_PID_DIR}/hadoop-mapred-tasktracker.pid"
-desc="Hadoop tasktracker daemon"
-
-start() {
-  echo -n $"Starting $desc (hadoop-tasktracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" start tasktracker
-  RETVAL=$?
-  echo
-  [ $RETVAL -eq 0 ] && touch /var/lock/subsys/hadoop-tasktracker
-  return $RETVAL
-}
-
-stop() {
-  echo -n $"Stopping $desc (hadoop-tasktracker): "
-  daemon --user mapred ${HADOOP_PREFIX}/sbin/hadoop-daemon.sh --config "${HADOOP_CONF_DIR}" stop tasktracker
-  RETVAL=$?
-  sleep 5
-  echo
-  [ $RETVAL -eq 0 ] && rm -f /var/lock/subsys/hadoop-tasktracker $PIDFILE
-}
-
-restart() {
-  stop
-  start
-}
-
-checkstatus(){
-  status -p $PIDFILE ${JAVA_HOME}/bin/java
-  RETVAL=$?
-}
-
-condrestart(){
-  [ -e /var/lock/subsys/hadoop-tasktracker ] && restart || :
-}
-
-case "$1" in
-  start)
-    start
-    ;;
-  stop)
-    stop
-    ;;
-  status)
-    checkstatus
-    ;;
-  restart)
-    restart
-    ;;
-  condrestart)
-    condrestart
-    ;;
-  *)
-    echo $"Usage: $0 {start|stop|status|restart|condrestart}"
-    exit 1
-esac
-
-exit $RETVAL

+ 0 - 179
hadoop-mapreduce-project/src/packages/rpm/spec/hadoop-mapred.spec

@@ -1,179 +0,0 @@
-#   Licensed to the Apache Software Foundation (ASF) under one or more
-#   contributor license agreements.  See the NOTICE file distributed with
-#   this work for additional information regarding copyright ownership.
-#   The ASF licenses this file to You under the Apache License, Version 2.0
-#   (the "License"); you may not use this file except in compliance with
-#   the License.  You may obtain a copy of the License at
-#
-#       http://www.apache.org/licenses/LICENSE-2.0
-#
-#   Unless required by applicable law or agreed to in writing, software
-#   distributed under the License is distributed on an "AS IS" BASIS,
-#   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-#   See the License for the specific language governing permissions and
-#   limitations under the License.
-
-#
-# RPM Spec file for Hadoop version @version@
-#
-
-%define name          hadoop-mapreduce
-%define version       @version@
-%define release       @package.release@
-%define major_version %(echo %{version} | cut -d. -f -2)
-
-# Installation Locations
-%define _prefix      @package.prefix@
-%define _bin_dir     %{_prefix}/bin
-%define _conf_dir    @package.conf.dir@
-%define _include_dir %{_prefix}/include
-%define _lib_dir     %{_prefix}/lib
-%define _lib64_dir   %{_prefix}/lib64
-%define _libexec_dir %{_prefix}/libexec
-%define _log_dir     @package.log.dir@
-%define _pid_dir     @package.pid.dir@
-%define _sbin_dir    %{_prefix}/sbin
-%define _share_dir   %{_prefix}/share
-%define _var_dir     @package.var.dir@
-
-# Build time settings
-%define _build_dir  @package.build.dir@
-%define _final_name @final.name@
-%define debug_package %{nil}
-
-# Disable brp-java-repack-jars for aspect J
-%define __os_install_post    \
-    /usr/lib/rpm/redhat/brp-compress \
-    %{!?__debug_package:/usr/lib/rpm/redhat/brp-strip %{__strip}} \
-    /usr/lib/rpm/redhat/brp-strip-static-archive %{__strip} \
-    /usr/lib/rpm/redhat/brp-strip-comment-note %{__strip} %{__objdump} \
-    /usr/lib/rpm/brp-python-bytecompile %{nil}
-
-# RPM searches perl files for dependancies and this breaks for non packaged perl lib
-# like thrift so disable this
-%define _use_internal_dependency_generator 0
-
-%ifarch i386
-%global hadoop_arch Linux-i386-32
-%endif
-%ifarch amd64 x86_64
-%global hadoop_arch Linux-amd64-64
-%endif
-%ifarch noarch
-%global hadoop_arch ""
-%endif
-
-Summary: The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing
-License: Apache License, Version 2.0
-URL: http://hadoop.apache.org/core/
-Vendor: Apache Software Foundation
-Group: Development/Libraries
-Name: %{name}
-Version: %{version}
-Release: %{release} 
-Source0: %{_final_name}-bin.tar.gz
-Prefix: %{_prefix}
-Prefix: %{_conf_dir}
-Prefix: %{_log_dir}
-Prefix: %{_pid_dir}
-Buildroot: %{_build_dir}
-Requires: sh-utils, textutils, /usr/sbin/useradd, /usr/sbin/usermod, /sbin/chkconfig, /sbin/service, hadoop-common >= %{major_version}, , hadoop-common <= %{major_version}.9999, hadoop-hdfs >= %{major_version}, hadoop-hdfs <= %{major_version}.9999
-AutoReqProv: no
-Provides: hadoop-mapreduce
-
-%description
-The Apache Hadoop project develops open-source software for reliable, scalable, 
-distributed computing.  Hadoop includes these subprojects:
-
-MapReduce: A software framework for distributed processing of large data sets on compute clusters.
-
-%prep
-%setup -n %{_final_name}
-
-%build
-if [ -d ${RPM_BUILD_DIR}%{_prefix} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_prefix}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_log_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_log_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_conf_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_conf_dir}
-fi
-
-if [ -d ${RPM_BUILD_DIR}%{_pid_dir} ]; then
-  rm -rf ${RPM_BUILD_DIR}%{_pid_dir}
-fi
-
-mkdir -p ${RPM_BUILD_DIR}%{_prefix}
-mkdir -p ${RPM_BUILD_DIR}%{_bin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_lib_dir}
-%ifarch amd64 x86_64
-mkdir -p ${RPM_BUILD_DIR}%{_lib64_dir}
-%endif
-mkdir -p ${RPM_BUILD_DIR}%{_libexec_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_log_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_conf_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_pid_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_sbin_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_share_dir}
-mkdir -p ${RPM_BUILD_DIR}%{_var_dir}
-mkdir -p ${RPM_BUILD_DIR}/etc/init.d
-
-cp ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-jobtracker.redhat ${RPM_BUILD_DIR}/etc/init.d/hadoop-jobtracker
-cp ${RPM_BUILD_DIR}/%{_final_name}/sbin/hadoop-tasktracker.redhat ${RPM_BUILD_DIR}/etc/init.d/hadoop-tasktracker
-chmod 0755 ${RPM_BUILD_DIR}/etc/init.d/hadoop-*
-
-#########################
-#### INSTALL SECTION ####
-#########################
-%install
-mv ${RPM_BUILD_DIR}/%{_final_name}/bin/* ${RPM_BUILD_DIR}%{_bin_dir}
-rm -f ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/configuration.xsl
-rm -f ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/hadoop-metrics2.properties
-mv ${RPM_BUILD_DIR}/%{_final_name}/etc/hadoop/* ${RPM_BUILD_DIR}%{_conf_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/include/* ${RPM_BUILD_DIR}%{_include_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/lib/* ${RPM_BUILD_DIR}%{_lib_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/libexec/* ${RPM_BUILD_DIR}%{_libexec_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/sbin/* ${RPM_BUILD_DIR}%{_sbin_dir}
-mv ${RPM_BUILD_DIR}/%{_final_name}/share/* ${RPM_BUILD_DIR}%{_share_dir}
-rm -rf ${RPM_BUILD_DIR}/%{_final_name}/etc
-
-%pre
-getent group hadoop 2>/dev/null >/dev/null || /usr/sbin/groupadd -g 123 -r hadoop
-/usr/sbin/useradd --comment "Hadoop MapReduce" -u 202 --shell /bin/bash -M -r --groups hadoop --home %{_var_dir}/mapred mapred 2> /dev/null || :
-
-%post
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-mapred-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3}
-
-%preun
-bash ${RPM_INSTALL_PREFIX0}/sbin/update-mapred-env.sh \
-       --prefix=${RPM_INSTALL_PREFIX0} \
-       --bin-dir=${RPM_INSTALL_PREFIX0}/bin \
-       --sbin-dir=${RPM_INSTALL_PREFIX0}/sbin \
-       --conf-dir=${RPM_INSTALL_PREFIX1} \
-       --log-dir=${RPM_INSTALL_PREFIX2} \
-       --pid-dir=${RPM_INSTALL_PREFIX3} \
-       --uninstall
-
-%files 
-%defattr(-,root,root)
-%attr(0755,root,hadoop) %{_log_dir}
-%attr(0775,root,hadoop) %{_pid_dir}
-%config(noreplace) %{_conf_dir}/mapred-site.xml
-%config(noreplace) %{_conf_dir}/capacity-scheduler.xml
-%config(noreplace) %{_conf_dir}/fair-scheduler.xml
-%config(noreplace) %{_conf_dir}/mapred-queues.xml
-%config(noreplace) %{_conf_dir}/taskcontroller.cfg
-%{_prefix}
-%attr(0775,root,root) /etc/init.d/hadoop-jobtracker
-%attr(0775,root,root) /etc/init.d/hadoop-tasktracker
-

+ 0 - 31
hadoop-mapreduce-project/src/packages/templates/conf/mapred-site.xml

@@ -1,31 +0,0 @@
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-
-<!-- Put site-specific property overrides in this file. -->
-
-<configuration>
-  <property>
-    <name>mapred.job.tracker</name>
-    <value>${HADOOP_JT_HOST}</value>
-  </property>
-
-  <property>
-    <name>mapred.system.dir</name>
-    <value>/user/mapred/system</value>
-  </property>
-
-  <property>
-    <name>mapred.local.dir</name>
-    <value>${HADOOP_MAPRED_DIR}</value>
-  </property>
-
-  <property>
-    <name>hadoop.tmp.dir</name>
-    <value>/tmp</value>
-  </property>
-
-  <property>
-    <name>mapred.jobtracker.taskScheduler</name>
-    <value>${HADOOP_TASK_SCHEDULER}</value>
-  </property>
-</configuration>

+ 0 - 139
hadoop-mapreduce-project/src/packages/update-mapred-env.sh

@@ -1,139 +0,0 @@
-#!/usr/bin/env bash
-
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# This script configures mapred-env.sh and symlinkis directories for 
-# relocating RPM locations.
-
-usage() {
-  echo "
-usage: $0 <parameters>
-  Required parameters:
-     --prefix=PREFIX             path to install into
-
-  Optional parameters:
-     --arch=i386                 OS Architecture
-     --bin-dir=PREFIX/bin        Executable directory
-     --conf-dir=/etc/hadoop      Configuration directory
-     --log-dir=/var/log/hadoop   Log directory
-     --pid-dir=/var/run          PID file location
-     --sbin-dir=PREFIX/sbin      System executable directory
-  "
-  exit 1
-}
-
-template_generator() {
-  cat $1 |
-  while read line ; do
-    while [[ "$line" =~ '(\$\{[a-zA-Z_][a-zA-Z_0-9]*\})' ]] ; do
-      LHS=${BASH_REMATCH[1]}
-      RHS="$(eval echo "\"$LHS\"")"
-      line=${line//$LHS/$RHS}
-    done
-    echo $line >> $2
-  done
-}
-
-OPTS=$(getopt \
-  -n $0 \
-  -o '' \
-  -l 'arch:' \
-  -l 'prefix:' \
-  -l 'bin-dir:' \
-  -l 'conf-dir:' \
-  -l 'lib-dir:' \
-  -l 'log-dir:' \
-  -l 'pid-dir:' \
-  -l 'sbin-dir:' \
-  -l 'uninstall' \
-  -- "$@")
-
-if [ $? != 0 ] ; then
-    usage
-fi
-
-eval set -- "${OPTS}"
-while true ; do
-  case "$1" in
-    --arch)
-      ARCH=$2 ; shift 2
-      ;;
-    --prefix)
-      PREFIX=$2 ; shift 2
-      ;;
-    --bin-dir)
-      BIN_DIR=$2 ; shift 2
-      ;;
-    --log-dir)
-      LOG_DIR=$2 ; shift 2
-      ;;
-    --lib-dir)
-      LIB_DIR=$2 ; shift 2
-      ;;
-    --conf-dir)
-      CONF_DIR=$2 ; shift 2
-      ;;
-    --pid-dir)
-      PID_DIR=$2 ; shift 2
-      ;;
-    --sbin-dir)
-      SBIN_DIR=$2 ; shift 2
-      ;;
-    --uninstall)
-      UNINSTALL=1; shift
-      ;;
-    --)
-      shift ; break
-      ;;
-    *)
-      echo "Unknown option: $1"
-      usage
-      exit 1
-      ;;
-  esac
-done
-
-for var in PREFIX; do
-  if [ -z "$(eval "echo \$$var")" ]; then
-    echo Missing param: $var
-    usage
-  fi
-done
-
-ARCH=${ARCH:-i386}
-BIN_DIR=${BIN_DIR:-$PREFIX/bin}
-CONF_DIR=${CONF_DIR:-$PREFIX/etc/hadoop}
-LIB_DIR=${LIB_DIR:-$PREFIX/lib}
-LOG_DIR=${LOG_DIR:-$PREFIX/var/log}
-PID_DIR=${PID_DIR:-$PREFIX/var/run}
-SBIN_DIR=${SBIN_DIR:-$PREFIX/sbin}
-UNINSTALL=${UNINSTALL:-0}
-
-if [ "${ARCH}" != "i386" ]; then
-  LIB_DIR=${LIB_DIR}64
-fi
-
-if [ "${UNINSTALL}" -ne "1" ]; then
-  mkdir -p ${LOG_DIR}
-  chown mapred:hadoop ${LOG_DIR}
-  chmod 755 ${LOG_DIR}
-
-  if [ ! -d ${PID_DIR} ]; then
-    mkdir -p ${PID_DIR}
-    chown root:hadoop ${PID_DIR}
-    chmod 775 ${PID_DIR}
-  fi
-fi