123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677 |
- #!/usr/bin/env bash
- # Licensed to the Apache Software Foundation (ASF) under one or more
- # contributor license agreements. See the NOTICE file distributed with
- # this work for additional information regarding copyright ownership.
- # The ASF licenses this file to You under the Apache License, Version 2.0
- # (the "License"); you may not use this file except in compliance with
- # the License. You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- if [[ -z "${BASH_VERSINFO[0]}" ]] \
- || [[ "${BASH_VERSINFO[0]}" -lt 3 ]] \
- || [[ "${BASH_VERSINFO[0]}" -eq 3 && "${BASH_VERSINFO[1]}" -lt 2 ]]; then
- echo "bash v3.2+ is required. Sorry."
- exit 1
- fi
- function centered_text
- {
- local text="$*"
- local spacing=$(( (75+${#text}) /2 ))
- printf "%*s\n" ${spacing} "${text}"
- }
- function big_console_header
- {
- printf "\n\n"
- echo "****************************************************************************"
- centered_text "${@}"
- echo "****************************************************************************"
- printf "\n\n"
- }
- ## @description Given a filename or dir, return the absolute version of it
- ## @audience public
- ## @stability stable
- ## @param directory
- ## @replaceable no
- ## @return 0 success
- ## @return 1 failure
- ## @return stdout abspath
- function hadoop_abs
- {
- declare obj=$1
- declare dir
- declare fn
- if [[ ! -e ${obj} ]]; then
- return 1
- elif [[ -d ${obj} ]]; then
- dir=${obj}
- else
- dir=$(dirname -- "${obj}")
- fn=$(basename -- "${obj}")
- fn="/${fn}"
- fi
- dir=$(cd -P -- "${dir}" >/dev/null 2>/dev/null && pwd -P)
- if [[ $? = 0 ]]; then
- echo "${dir}${fn}"
- return 0
- fi
- return 1
- }
- ## @description Print a message to stderr
- ## @audience public
- ## @stability stable
- ## @replaceable no
- ## @param string
- function hadoop_error
- {
- echo "$*" 1>&2
- }
- function run_and_redirect
- {
- declare logfile=$1
- shift
- declare res
- echo "\$ ${*} > ${logfile} 2>&1"
- # to the log
- {
- date
- echo "cd $(pwd)"
- echo "${*}"
- } > "${logfile}"
- # run the actual command
- "${@}" >> "${logfile}" 2>&1
- res=$?
- if [[ ${res} != 0 ]]; then
- echo
- echo "Failed!"
- echo
- exit "${res}"
- fi
- }
- function hadoop_native_flags
- {
- # modified version of the Yetus personality
- if [[ ${NATIVE} != true ]]; then
- return
- fi
- # Based upon HADOOP-11937
- #
- # Some notes:
- #
- # - getting fuse to compile on anything but Linux
- # is always tricky.
- # - Darwin assumes homebrew is in use.
- # - HADOOP-12027 required for bzip2 on OS X.
- # - bzip2 is broken in lots of places.
- # e.g, HADOOP-12027 for OS X. so no -Drequire.bzip2
- #
- case "${OSNAME}" in
- Linux)
- # shellcheck disable=SC2086
- echo -Pnative -Drequire.snappy -Drequire.openssl -Drequire.fuse
- ;;
- Darwin)
- echo \
- -Pnative -Drequire.snappy \
- -Drequire.openssl \
- -Dopenssl.prefix=/usr/local/opt/openssl/ \
- -Dopenssl.include=/usr/local/opt/openssl/include \
- -Dopenssl.lib=/usr/local/opt/openssl/lib
- ;;
- *)
- # shellcheck disable=SC2086
- echo \
- -Pnative \
- -Drequire.snappy -Drequire.openssl \
- -Drequire.test.libhadoop
- ;;
- esac
- }
- # Function to probe the exit code of the script commands,
- # and stop in the case of failure with an contextual error
- # message.
- function run()
- {
- declare res
- declare logfile
- echo "\$ ${*}"
- "${@}"
- res=$?
- if [[ ${res} != 0 ]]; then
- echo
- echo "Failed!"
- echo
- exit "${res}"
- fi
- }
- function domd5()
- {
- run "${MD5SUM}" "${1}" > "${1}.md5"
- }
- ## @description set JAVA_HOME properly
- ## @audience public
- ## @stability unstable
- function locate_jvm()
- {
- JAVA_HOME="$(ls -d /usr/lib/jvm/*${JVM_VERSION}* | grep "${JVM_HINT}" | head -1 )"
- export JAVA_HOME
- }
- function header()
- {
- echo
- printf "\n\n"
- echo "============================================================================"
- echo "============================================================================"
- centered_text "Hadoop Release Creator"
- echo "============================================================================"
- echo "============================================================================"
- printf "\n\n"
- echo "Version to create : ${HADOOP_VERSION}"
- echo "Release Candidate Label: ${RC_LABEL##-}"
- echo "Source Version : ${DEFAULT_HADOOP_VERSION}"
- echo "Using JDK : ${JAVA_HOME}"
- printf "\n\n"
- }
- function set_defaults
- {
- BINDIR=$(dirname "${BIN}")
- BASEDIR=$(hadoop_abs "${BINDIR}/../..")
- ARTIFACTS_DIR="${BASEDIR}/target/artifacts"
- # Extract Hadoop version from ${BASEDIR}/pom.xml
- DEFAULT_HADOOP_VERSION=$(grep "<version>" "${BASEDIR}/pom.xml" \
- | head -1 \
- | sed -e 's|^ *<version>||' -e 's|</version>.*$||')
- DOCKER=false
- DOCKERCACHE=false
- DOCKERFILE="${BASEDIR}/dev-support/docker/Dockerfile"
- DOCKERRAN=false
- # Extract Java version from ${BASEDIR}/pom.xml
- # doing this outside of maven means we can do this before
- # the docker container comes up...
- JVM_VERSION=$(grep "<javac.version>" "${BASEDIR}/hadoop-project/pom.xml" \
- | head -1 \
- | sed -e 's|^ *<javac.version>||' -e 's|</javac.version>.*$||' -e 's|..||')
- GIT=$(command -v git)
- GPG=$(command -v gpg)
- GPGAGENT=$(command -v gpg-agent)
- HADOOP_VERSION="${DEFAULT_HADOOP_VERSION}"
- INDOCKER=false
- LOGDIR="${BASEDIR}/patchprocess"
- if [[ -z "${MVN}" ]]; then
- if [[ -n "${MAVEN_HOME}" ]]; then
- MVN=${MAVEN_HOME}/bin/mvn
- else
- MVN=$(command -v mvn)
- fi
- fi
- MD5SUM=$(command -v md5sum)
- if [[ -z "${MD5SUM}" ]]; then
- MD5SUM=$(command -v md5)
- fi
- NATIVE=false
- OSNAME=$(uname -s)
- PUBKEYFILE="https://dist.apache.org/repos/dist/release/hadoop/common/KEYS"
- SIGN=false
- }
- function startgpgagent
- {
- if [[ "${SIGN}" = true ]]; then
- if [[ -n "${GPGAGENT}" && -z "${GPG_AGENT_INFO}" ]]; then
- echo "starting gpg agent"
- echo "default-cache-ttl 14400" > "${LOGDIR}/gpgagent.conf"
- # shellcheck disable=2046
- eval $("${GPGAGENT}" --daemon \
- --options "${LOGDIR}/gpgagent.conf" \
- --log-file="${LOGDIR}/create-release-gpgagent.log")
- GPGAGENTPID=$(echo "${GPG_AGENT_INFO}" | cut -f 2 -d:)
- fi
- if [[ -n "${GPG_AGENT_INFO}" ]]; then
- echo "Warming the gpg-agent cache prior to calling maven"
- # warm the agent's cache:
- touch "${LOGDIR}/warm"
- ${GPG} --use-agent --armor --output "${LOGDIR}/warm.asc" --detach-sig "${LOGDIR}/warm"
- rm "${LOGDIR}/warm.asc" "${LOGDIR}/warm"
- else
- SIGN=false
- hadoop_error "ERROR: Unable to launch or acquire gpg-agent. Disable signing."
- fi
- fi
- }
- function stopgpgagent
- {
- if [[ -n "${GPGAGENTPID}" ]]; then
- kill "${GPGAGENTPID}"
- fi
- }
- function usage
- {
- echo "--artifactsdir=[path] Path to use to store release bits"
- echo "--asfrelease Make an ASF release"
- echo "--docker Use Hadoop's Dockerfile for guaranteed environment"
- echo "--dockercache Use a Docker-private maven cache"
- echo "--jvmhint=[filter] Simple filter to pick a JVM to use"
- echo "--logdir=[path] Path to store logs"
- echo "--mvncache=[path] Path to the maven cache to use"
- echo "--native Also build the native components"
- echo "--rc-label=[label] Add this label to the builds"
- echo "--sign Use .gnupg dir to sign the artifacts and jars"
- echo "--version=[version] Use an alternative version string"
- }
- function option_parse
- {
- declare i
- for i in "$@"; do
- case ${i} in
- --asfrelease)
- ASFRELEASE=true
- NATIVE=true
- SIGN=true
- ;;
- --artifactsdir=*)
- ARTIFACTS_DIR=${i#*=}
- ;;
- --docker)
- DOCKER=true
- ;;
- --dockercache)
- DOCKERCACHE=true
- ;;
- --help)
- usage
- exit
- ;;
- --indocker)
- INDOCKER=true
- ;;
- --jvmhint=*)
- JVM_HINT=${i#*=}
- ;;
- --logdir=*)
- LOGDIR=${i#*=}
- ;;
- --mvncache=*)
- MVNCACHE=${i#*=}
- ;;
- --native)
- NATIVE=true
- ;;
- --rc-label=*)
- RC_LABEL=${i#*=}
- ;;
- --sign)
- SIGN=true
- ;;
- --version=*)
- HADOOP_VERSION=${i#*=}
- ;;
- esac
- done
- if [[ ! -d "${HOME}/.gnupg" ]]; then
- hadoop_error "ERROR: No .gnupg dir. Disabling signing capability."
- SIGN=false
- fi
- if [[ "${SIGN}" = true ]]; then
- if [[ -n "${GPG_AGENT_INFO}" ]]; then
- echo "NOTE: Using existing gpg-agent. If the default-cache-ttl"
- echo "is set to less than ~20 mins, maven commands will fail."
- elif [[ -z "${GPGAGENT}" ]]; then
- hadoop_error "ERROR: No gpg-agent. Disabling signing capability."
- SIGN=false
- fi
- fi
- DOCKERCMD=$(command -v docker)
- if [[ "${DOCKER}" = true && -z "${DOCKERCMD}" ]]; then
- hadoop_error "ERROR: docker binary not found. Disabling docker mode."
- DOCKER=false
- fi
- if [[ "${DOCKERCACHE}" = true && "${DOCKER}" = false ]]; then
- if [[ "${INDOCKER}" = false ]]; then
- hadoop_error "ERROR: docker mode not enabled. Disabling dockercache."
- fi
- DOCKERCACHE=false
- fi
- if [[ "${DOCKERCACHE}" = true && -n "${MVNCACHE}" ]]; then
- hadoop_error "ERROR: Cannot set --mvncache and --dockercache simultaneously."
- exit 1
- else
- MVNCACHE=${MVNCACHE:-"${HOME}/.m2"}
- fi
- if [[ "${ASFRELEASE}" = true ]]; then
- if [[ "${SIGN}" = false ]]; then
- hadoop_error "ERROR: --asfrelease requires --sign. Exiting."
- exit 1
- fi
- if [[ "${OSNAME}" = Linux ]]; then
- if [[ "${DOCKER}" = false && "${INDOCKER}" = false ]]; then
- hadoop_error "ERROR: --asfrelease requires --docker on Linux. Exiting."
- exit 1
- elif [[ "${DOCKERCACHE}" = false && "${INDOCKER}" = false ]]; then
- hadoop_error "ERROR: --asfrelease on Linux requires --dockercache. Exiting."
- exit 1
- fi
- fi
- fi
- if [[ -n "${MVNCACHE}" ]]; then
- mkdir -p "${MVNCACHE}"
- if [[ -d "${MVNCACHE}" ]]; then
- MVN_ARGS=("-Dmaven.repo.local=${MVNCACHE}")
- fi
- fi
- }
- function dockermode
- {
- declare lines
- declare -a modp
- declare imgname
- declare -a extrad
- declare user_name
- declare group_id
- if [[ "${DOCKER}" != true ]]; then
- return
- fi
- user_name=${SUDO_USER:=$USER}
- user_id=$(id -u "${user_name}")
- group_id=$(id -g "${user_name}")
- imgname="hadoop/createrelease:${HADOOP_VERSION}_${RANDOM}"
- if [[ -d "${HOME}/.gnupg" ]]; then
- extrad+=("-v" "${HOME}/.gnupg:/home/${user_name}/.gnupg")
- fi
- if [[ -n "${LOGDIR}" ]]; then
- if [[ ! -d "${LOGDIR}" ]]; then
- mkdir -p "${LOGDIR}"
- fi
- lines=$(hadoop_abs "${LOGDIR}")
- extrad+=("-v" "${lines}:${lines}")
- fi
- if [[ -n "${ARTIFACTS_DIR}" ]]; then
- if [[ ! -d "${ARTIFACTS_DIR}" ]]; then
- mkdir -p "${ARTIFACTS_DIR}"
- fi
- lines=$(hadoop_abs "${ARTIFACTS_DIR}")
- extrad+=("-v" "${lines}:${lines}")
- fi
- if [[ "${DOCKERCACHE}" = true ]]; then
- modp+=("--mvncache=/maven")
- else
- lines=$(hadoop_abs "${MVNCACHE}")
- extrad+=("-v" "${lines}:${lines}")
- fi
- for lines in "${PARAMS[@]}"; do
- if [[ "${lines}" != "--docker" ]]; then
- modp+=("$lines")
- fi
- done
- modp+=("--indocker")
- (
- lines=$(grep -n 'YETUS CUT HERE' "${DOCKERFILE}" | cut -f1 -d:)
- if [[ -z "${lines}" ]]; then
- cat "${DOCKERFILE}"
- else
- head -n "${lines}" "${DOCKERFILE}"
- fi
- # make sure we put some space between, just in case last
- # line isn't an empty line or whatever
- printf "\n\n"
- # force a new image for every run to make it easier to remove later
- echo "LABEL org.apache.hadoop.create-release=\"cr-${RANDOM}\""
- # setup ownerships, etc
- echo "RUN groupadd --non-unique -g ${group_id} ${user_name}"
- echo "RUN useradd -g ${group_id} -u ${user_id} -m ${user_name}"
- echo "RUN chown -R ${user_name} /home/${user_name}"
- echo "ENV HOME /home/${user_name}"
- echo "RUN mkdir -p /maven"
- echo "RUN chown -R ${user_name} /maven"
- echo "USER ${user_name}"
- printf "\n\n"
- ) | docker build -t "${imgname}" -
- run docker run -i -t \
- --privileged \
- "${extrad[@]}" \
- -v "${BASEDIR}:/build/source" \
- -u "${user_name}" \
- -w "/build/source" \
- "${imgname}" \
- "/build/source/dev-support/bin/create-release" "${modp[@]}"
- DOCKERRAN=true
- }
- function makearelease
- {
- # let's start at the root
- run cd "${BASEDIR}"
- big_console_header "Cleaning the Source Tree"
- # git clean to clear any remnants from previous build
- run "${GIT}" clean -xdf
- mkdir -p "${LOGDIR}"
- # Install the Hadoop maven plugins first
- run_and_redirect "${LOGDIR}/mvn_install_maven_plugins.log" "${MVN}" "${MVN_ARGS[@]}" -pl hadoop-maven-plugins -am clean install
- # mvn clean for sanity
- run_and_redirect "${LOGDIR}/mvn_clean.log" "${MVN}" "${MVN_ARGS[@]}" clean
- # Create staging dir for release artifacts
- run mkdir -p "${ARTIFACTS_DIR}"
- big_console_header "Apache RAT Check"
- # Create RAT report
- run_and_redirect "${LOGDIR}/mvn_apache_rat.log" "${MVN}" "${MVN_ARGS[@]}" apache-rat:check
- big_console_header "Maven Build and Install"
- if [[ "${SIGN}" = true ]]; then
- signflags=("-Psign" "-Dgpg.useagent=true" -Dgpg.executable="${GPG}")
- fi
- # Create SRC and BIN tarballs for release,
- # shellcheck disable=SC2046
- run_and_redirect "${LOGDIR}/mvn_install.log" \
- "${MVN}" "${MVN_ARGS[@]}" install \
- -Pdist,src,yarn-ui \
- "${signflags[@]}" \
- -DskipTests -Dtar $(hadoop_native_flags)
- # Create site for release
- # we need to do install again so that jdiff and
- # a few other things get registered in the maven
- # universe correctly
- run_and_redirect "${LOGDIR}/mvn_site.log" \
- "${MVN}" "${MVN_ARGS[@]}" install \
- site site:stage \
- -DskipTests \
- -Pdist,src,releasedocs,docs
- big_console_header "Staging the release"
- run mv "${BASEDIR}/target/staging/hadoop-project" "${BASEDIR}/target/r${HADOOP_VERSION}/"
- run cd "${BASEDIR}/target/"
- run tar czpf "hadoop-site-${HADOOP_VERSION}.tar.gz" "r${HADOOP_VERSION}"/*
- run cd "${BASEDIR}"
- # Stage RAT report
- #shellcheck disable=SC2038
- find . -name rat.txt | xargs -I% cat % > "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-rat.txt"
- # Stage CHANGES and RELEASENOTES files
- for i in CHANGES RELEASENOTES; do
- run cp -p \
- "${BASEDIR}/hadoop-common-project/hadoop-common/src/site/markdown/release/${HADOOP_VERSION}"/${i}*.md \
- "${ARTIFACTS_DIR}/${i}.md"
- done
- # Prepare and stage BIN tarball
- run cd "${BASEDIR}/hadoop-dist/target/"
- run tar -xzpf "hadoop-${HADOOP_VERSION}.tar.gz"
- run mkdir -p "hadoop-${HADOOP_VERSION}/share/doc/hadoop/"
- run cp -r "${BASEDIR}/target/r${HADOOP_VERSION}"/* "hadoop-${HADOOP_VERSION}/share/doc/hadoop/"
- run tar -czpf "hadoop-${HADOOP_VERSION}.tar.gz" "hadoop-${HADOOP_VERSION}"
- run cd "${BASEDIR}"
- run mv \
- "${BASEDIR}/hadoop-dist/target/hadoop-${HADOOP_VERSION}.tar.gz" \
- "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}.tar.gz"
- # Stage SRC tarball
- run mv \
- "${BASEDIR}/hadoop-dist/target/hadoop-${HADOOP_VERSION}-src.tar.gz" \
- "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-src.tar.gz"
- # Stage SITE tarball
- run mv \
- "${BASEDIR}/target/hadoop-site-${HADOOP_VERSION}.tar.gz" \
- "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}-site.tar.gz"
- }
- function signartifacts
- {
- declare i
- if [[ "${SIGN}" = false ]]; then
- for i in ${ARTIFACTS_DIR}/*; do
- domd5 "${i}"
- done
- echo ""
- echo "Remember to sign the artifacts before staging them on the open"
- echo ""
- return
- fi
- big_console_header "Signing the release"
- for i in ${ARTIFACTS_DIR}/*; do
- ${GPG} --use-agent --armor --output "${i}.asc" --detach-sig "${i}"
- ${GPG} --print-mds "${i}" > "${i}.mds"
- domd5 "${i}"
- done
- if [[ "${ASFRELEASE}" = true ]]; then
- echo "Fetching the Apache Hadoop KEYS file..."
- curl -L "${PUBKEYFILE}" -o "${BASEDIR}/target/KEYS"
- ${GPG} --import --trustdb "${BASEDIR}/target/testkeysdb" "${BASEDIR}/target/KEYS"
- ${GPG} --verify --trustdb "${BASEDIR}/target/testkeysdb" \
- "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}.tar.gz.asc" \
- "${ARTIFACTS_DIR}/hadoop-${HADOOP_VERSION}${RC_LABEL}.tar.gz"
- if [[ $? != 0 ]]; then
- hadoop_error "ERROR: GPG key is not present in ${PUBKEYFILE}."
- hadoop_error "ERROR: This MUST be fixed. Exiting."
- exit 1
- fi
- fi
- }
- # find root of the source tree
- BIN=$(hadoop_abs "${BASH_SOURCE:-$0}")
- PARAMS=("$@")
- set_defaults
- option_parse "${PARAMS[@]}"
- dockermode
- locate_jvm
- header
- if [[ -n ${RC_LABEL} ]]; then
- RC_LABEL="-${RC_LABEL}"
- fi
- if [[ "${INDOCKER}" = true || "${DOCKERRAN}" = false ]]; then
- startgpgagent
- makearelease
- signartifacts
- stopgpgagent
- fi
- if [[ "${INDOCKER}" = true ]]; then
- exit $?
- fi
- if [[ $? == 0 ]]; then
- echo
- echo "Congratulations, you have successfully built the release"
- echo "artifacts for Apache Hadoop ${HADOOP_VERSION}${RC_LABEL}"
- echo
- echo "The artifacts for this run are available at ${ARTIFACTS_DIR}:"
- run ls -1 "${ARTIFACTS_DIR}"
- echo
- fi
|