Explorar o código

Fix jdiff targets and build.xml cleanup

git-svn-id: https://svn.apache.org/repos/asf/hadoop/core/branches/HADOOP-4687/hdfs@780681 13f79535-47bb-0310-9956-ffa450edef68
Giridharan Kesavan %!s(int64=16) %!d(string=hai) anos
pai
achega
ef1f5804e7
Modificáronse 2 ficheiros con 33 adicións e 648 borrados
  1. 22 648
      build.xml
  2. 11 0
      lib/jdiff/hadoop-hdfs_0.20.0.xml

+ 22 - 648
build.xml

@@ -31,7 +31,7 @@
   <property name="final.name" value="${name}-${version}"/>
   <property name="test.hdfs.final.name" value="${name}-test-${version}"/>
   <property name="test.hdfswithmr.final.name" value="${name}-hdsfwithmr-test-${version}"/>
-  <property name="test.final.name" value="${name}-test-${version}"/>
+  <property name="ant.final.name" value="${name}-ant-${version}"/>
   <property name="year" value="2009"/>
 
   <property name="src.dir" value="${basedir}/src"/>  	
@@ -44,7 +44,6 @@
   <property name="src.docs.cn" value="${basedir}/src/docs/cn"/>
   <property name="changes.src" value="${docs.src}/changes"/>
 
-  <property name="xercescroot" value=""/> 
   <property name="build.dir" value="${basedir}/build"/>
   <property name="build.classes" value="${build.dir}/classes"/>
   <property name="build.src" value="${build.dir}/src"/>
@@ -119,7 +118,7 @@
   <property name="jdiff.xml.dir" value="${lib.dir}/jdiff"/>
   <property name="jdiff.stable" value="0.20.0"/>
   <property name="jdiff.stable.javadoc" 
-            value="http://hadoop.apache.org/core/docs/r${jdiff.stable}/api/"/>
+            value="http://hadoop.apache.org/hdfs/docs/r${jdiff.stable}/api/"/>
 
   <property name="scratch.dir" value="${user.home}/tmp"/>
   <property name="svn.cmd" value="svn"/>
@@ -141,14 +140,14 @@
   <property name="common.ivy.lib.dir" location="${build.ivy.lib.dir}/${ant.project.name}/common"/>
   <property name="build.ivy.report.dir" location="${build.ivy.dir}/report" />
   <property name="build.ivy.maven.dir" location="${build.ivy.dir}/maven" />
-  <property name="build.ivy.maven.pom" location="${build.ivy.maven.dir}/hadoop-core-${version}.pom" />
-  <property name="build.ivy.maven.jar" location="${build.ivy.maven.dir}/hadoop-core-${version}.jar" />
+  <property name="build.ivy.maven.pom" location="${build.ivy.maven.dir}/hadoop-hdfs-${version}.pom" />
+  <property name="build.ivy.maven.jar" location="${build.ivy.maven.dir}/hadoop-hdfs-${version}.jar" />
 
   <!--this is the naming policy for artifacts we want pulled down-->
   <property name="ivy.artifact.retrieve.pattern" value="${ant.project.name}/[conf]/[artifact]-[revision].[ext]"/>
 
   <!--this is how artifacts that get built are named-->
-  <property name="ivy.publish.pattern" value="hadoop-[revision]-core.[ext]"/>
+  <property name="ivy.publish.pattern" value="hadoop-hdfs-[revision].[ext]"/>
   <property name="hadoop-hdfs.jar" location="${build.dir}/${final.name}.jar" />
 
   <!-- jdiff.home property set -->
@@ -186,7 +185,7 @@
     <pathelement location="${build.examples}"/>
     <pathelement location="${build.tools}"/>
     <pathelement path="${clover.jar}"/>
-    <path refid="ivy.test.classpath"/>
+    <path refid="ivy-test.classpath"/>
     <fileset dir="${lib.dir}">
       <include name="hadoop-core-test-${version}.jar" />
       <exclude name="**/excluded/" />
@@ -200,6 +199,7 @@
     <pathelement location="${lib.dir}/hadoop-mapred-test-${version}.jar" />
     <pathelement location="${lib.dir}/hadoop-mapred-${version}.jar" />
     <pathelement location="${lib.dir}/hadoop-mapred-tools-${version}.jar" />
+    <pathelement location="${lib.dir}/hadoop-mapred-examples-${version}.jar" />
   </path>
 
   <!-- the cluster test classpath: uses conf.dir for configuration -->
@@ -210,9 +210,6 @@
     <pathelement location="${build.dir}"/>
   </path>
 
-  <!-- properties dependent on the items defined above. -->
-  <!--<available classname="${rat.reporting.classname}" classpathref="classpath" property="rat.present" value="true"/> -->
-
   <!-- ====================================================== -->
   <!-- Macro definitions                                      -->
   <!-- ====================================================== -->
@@ -244,7 +241,6 @@
     <mkdir dir="${test.build.extraconf}"/>
     <tempfile property="touch.temp.file" destDir="${java.io.tmpdir}"/>
     <touch millis="0" file="${touch.temp.file}">
-      <fileset dir="${conf.dir}" includes="**/*.template"/>
       <fileset dir="${contrib.dir}" includes="**/*.template"/>
     </touch>
     <delete file="${touch.temp.file}"/>
@@ -259,96 +255,8 @@
   <!-- ====================================================== -->
   <!-- Compile the Java files                                 -->
   <!-- ====================================================== -->
-  <target name="record-parser" depends="init" if="javacc.home">
-      <javacc
-          target="${core.src.dir}/org/apache/hadoop/record/compiler/generated/rcc.jj"
-          outputdirectory="${core.src.dir}/org/apache/hadoop/record/compiler/generated"
-          javacchome="${javacc.home}" />
-  </target>
-  
-  <target name="compile-rcc-compiler" depends="init, record-parser">
-    <javac 
-        encoding="${build.encoding}" 
-        srcdir="${hdfs.src.dir}"
-        includes="org/apache/hadoop/record/compiler/**/*.java"
-        destdir="${build.classes}"
-        debug="${javac.debug}"
-        optimize="${javac.optimize}"
-        target="${javac.version}"
-        source="${javac.version}"
-        deprecation="${javac.deprecation}">
-        <compilerarg line="${javac.args}"/>
-        <classpath refid="classpath"/>
-    </javac>
-    
-    <taskdef name="recordcc" classname="org.apache.hadoop.record.compiler.ant.RccTask">
-      <classpath refid="classpath" />
-    </taskdef>
-  </target>
-  
-  <target name="compile-core-classes" depends="init, compile-rcc-compiler">
-     <taskdef classname="org.apache.jasper.JspC" name="jsp-compile" >
-        <classpath refid="test.classpath"/>
-     </taskdef>
-    <!-- Compile Java files (excluding JSPs) checking warnings -->
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${core.src.dir}"	
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${build.classes}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="classpath"/>
-    </javac>
 
-    <copy todir="${build.classes}">
-      <fileset dir="${core.src.dir}" includes="**/*.properties"/>
-      <fileset dir="${core.src.dir}" includes="core-default.xml"/>
-    </copy>
-     
-  </target>
-<!--
-  <target name="compile-mapred-classes" depends="compile-core-classes">
-    <jsp-compile
-     uriroot="${src.webapps}/task"
-     outputdir="${build.src}"
-     package="org.apache.hadoop.mapred"
-     webxml="${build.webapps}/task/WEB-INF/web.xml">
-    </jsp-compile>
-
-    <jsp-compile
-     uriroot="${src.webapps}/job"
-     outputdir="${build.src}"
-     package="org.apache.hadoop.mapred"
-     webxml="${build.webapps}/job/WEB-INF/web.xml">
-    </jsp-compile>
-
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${mapred.src.dir};${build.src}" 
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${build.classes}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="classpath"/>
-    </javac>   
-    
-    <copy todir="${build.classes}">
-      <fileset dir="${mapred.src.dir}" includes="**/*.properties"/>
-      <fileset dir="${mapred.src.dir}" includes="mapred-default.xml"/>
-    </copy>
-  </target>
--->
-  <!--<target name="compile-hdfs-classes" depends="compile-core-classes">-->
-  <target name="compile-hdfs-classes" depends="init, compile-rcc-compiler">
+  <target name="compile-hdfs-classes" depends="init">
     <taskdef classname="org.apache.jasper.JspC" name="jsp-compile" >
        <classpath refid="classpath"/>
     </taskdef>
@@ -393,78 +301,8 @@
      <fileset dir="${hdfs.src.dir}" includes="hdfs-default.xml"/>
     </copy>
   </target>
-<!--
-  <target name="compile-tools" depends="init">
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${tools.src}"
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${build.tools}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="classpath"/>
-    </javac>   
-  	
-    <copy todir="${build.tools}">
-      <fileset 
-        dir="${tools.src}" 
-        includes="**/*.properties"
-      />
-    </copy>
-  </target>
 
-  <target name="compile-native">
-    <antcall target="compile-core-native">
-      <param name="compile.native" value="true"/>
-    </antcall> 
-  </target>
-
-  <target name="compile-core-native" depends="compile-core-classes"
-          if="compile.native">
-  	
-    <mkdir dir="${build.native}/lib"/>
-    <mkdir dir="${build.native}/src/org/apache/hadoop/io/compress/zlib"/>
-
-  	<javah
-  	  classpath="${build.classes}"
-  	  destdir="${build.native}/src/org/apache/hadoop/io/compress/zlib"
-      force="yes"
-  	  verbose="yes"
-  	  >
-  	  <class name="org.apache.hadoop.io.compress.zlib.ZlibCompressor" />
-      <class name="org.apache.hadoop.io.compress.zlib.ZlibDecompressor" />
-  	</javah>
-
-	<exec dir="${build.native}" executable="sh" failonerror="true">
-	  <env key="OS_NAME" value="${os.name}"/>
-	  <env key="OS_ARCH" value="${os.arch}"/>
-	  <env key="JVM_DATA_MODEL" value="${sun.arch.data.model}"/>
-	  <env key="HADOOP_NATIVE_SRCDIR" value="${native.src.dir}"/>
-	  <arg line="${native.src.dir}/configure"/>
-    </exec>
-
-    <exec dir="${build.native}" executable="${make.cmd}" failonerror="true">
-      <env key="OS_NAME" value="${os.name}"/>
-      <env key="OS_ARCH" value="${os.arch}"/>
-  	  <env key="JVM_DATA_MODEL" value="${sun.arch.data.model}"/>
-  	  <env key="HADOOP_NATIVE_SRCDIR" value="${native.src.dir}"/>
-    </exec>
-
-	<exec dir="${build.native}" executable="sh" failonerror="true">
-	  <arg line="${build.native}/libtool mode=install cp ${build.native}/lib/libhadoop.la ${build.native}/lib"/>
-    </exec>
-
-  </target>
--->
-  <target name="compile-core"
-          depends="clover,
-  	compile-hdfs-classes" 
-  	description="Compile core only">
-  </target>
+  <target name="compile-core" depends="clover, compile-hdfs-classes" description="Compile"/> 
 
   <target name="compile-contrib" depends="compile-core">
      <subant target="compile">
@@ -473,38 +311,14 @@
      </subant>  	
   </target>
   
-  <target name="compile" depends="compile-core, compile-contrib, compile-ant-tasks" description="Compile core, contrib">
-  </target>
-
-  <target name="compile-examples" 
-          depends="compile-core">
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${examples.dir}"
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${build.examples}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath>
-        <path refid="classpath"/>
-        <pathelement location="${build.tools}"/>
-      </classpath>
-    </javac>    
-  </target>
+  <target name="compile" depends="compile-core, compile-contrib, compile-ant-tasks" description="Compile hdfs and contrib" />
 
   <!-- ================================================================== -->
-  <!-- Make hadoop.jar                                                     -->
+  <!-- Make hadoop-hdfs.jar                                               -->
   <!-- ================================================================== -->
   <!--                                                                    -->
   <!-- ================================================================== -->
   <target name="jar" depends="compile-core" description="Make hadoop.jar">
-   <!-- <tar compression="gzip" destfile="${build.classes}/bin.tgz">
-      <tarfileset dir="bin" mode="755"/>
-    </tar> -->
     <jar jarfile="${hadoop-hdfs.jar}"
          basedir="${build.classes}">
       <manifest>
@@ -521,80 +335,10 @@
     </jar>
   </target>
 
-  <!-- ================================================================== -->
-  <!-- Make the Hadoop examples jar.                                      -->
-  <!-- ================================================================== -->
-  <!--                                                                    -->
-  <!-- ================================================================== -->
-<!--
-  <target name="examples" depends="jar, compile-examples" description="Make the Hadoop examples jar.">
-    <jar jarfile="${build.dir}/${final.name}-examples.jar"
-         basedir="${build.examples}">
-      <manifest>
-        <attribute name="Main-Class" 
-                   value="org/apache/hadoop/examples/ExampleDriver"/>
-      </manifest>
-    </jar>
-  </target>
-
-  <target name="tools-jar" depends="jar, compile-tools" 
-          description="Make the Hadoop tools jar.">
-    <jar jarfile="${build.dir}/${final.name}-tools.jar"
-         basedir="${build.tools}">
-      <manifest>
-        <attribute name="Main-Class" 
-                   value="org/apache/hadoop/examples/ExampleDriver"/>
-      </manifest>
-    </jar>
-  </target>
-
-  <target name="generate-test-records" depends="compile-rcc-compiler">
-    <recordcc destdir="${test.generated.dir}">
-      <fileset dir="${test.src.dir}"
-	         includes="**/*.jr" />
-    </recordcc>
-  </target>
--->
   <!-- ================================================================== -->
   <!-- Compile test code                                                  --> 
   <!-- ================================================================== -->
 
-<!-- 
-  <target name="compile-core-test" depends="compile-core-classes, generate-test-records">
-    <mkdir dir="${test.core.build.classes}"/>
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${test.generated.dir}"
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${test.core.build.classes}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args}" />
-      <classpath refid="test.classpath"/>
-    </javac>
-    <javac 
-     encoding="${build.encoding}" 
-     srcdir="${test.src.dir}/core"
-     includes="org/apache/hadoop/**/*.java"
-     destdir="${test.core.build.classes}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="test.classpath"/>
-     </javac>
-
-    <delete dir="${test.cache.data}"/>
-    <mkdir dir="${test.cache.data}"/>
-    <copy file="${test.src.dir}/core/org/apache/hadoop/cli/testConf.xml" todir="${test.cache.data}"/>
-  </target>
--->
-
   <target name="compile-hdfs-test" depends="compile-hdfs-classes, ivy-retrieve-test">
     <mkdir dir="${test.hdfs.build.classes}"/>
     <javac 
@@ -623,79 +367,7 @@
     <copy file="${test.src.dir}/hdfs/org/apache/hadoop/hdfs/tools/offlineImageViewer/fsimageV18" todir="${test.cache.data}"/>
     <copy file="${test.src.dir}/hdfs/org/apache/hadoop/hdfs/tools/offlineImageViewer/fsimageV19" todir="${test.cache.data}"/>
   </target>
-<!--
-  <target name="compile-mapred-test" depends="compile-examples, compile-hdfs-test">
-
-    <mkdir dir="${test.mapred.build.classes}"/>
-    <mkdir dir="${test.mapred.build.testjar}"/>
-    <mkdir dir="${test.mapred.build.testshell}"/>
-
-    <javac 
-      encoding="${build.encoding}" 
-      srcdir="${test.src.dir}/mapred"
-      includes="org/apache/hadoop/**/*.java"
-      destdir="${test.mapred.build.classes}"
-      debug="${javac.debug}"
-      optimize="${javac.optimize}"
-      target="${javac.version}"
-      source="${javac.version}"
-      deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="test.mapred.classpath"/>
-    </javac>
-
-    <javac
-      encoding="${build.encoding}"
-      srcdir="${test.src.dir}/mapred/testjar"
-      includes="*.java"
-      destdir="${test.mapred.build.testjar}"
-      debug="${javac.debug}"
-      optimize="${javac.optimize}"
-      target="${javac.version}"
-      source="${javac.version}"
-      deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}" />
-      <classpath refid="test.mapred.classpath"/>
-    </javac>
-
-    <delete file="${test.mapred.build.testjar}/testjob.jar"/> 
-    <jar jarfile="${test.mapred.build.testjar}/testjob.jar"
-       basedir="${test.mapred.build.testjar}">
-    </jar>
 
-    <javac 
-     encoding="${build.encoding}"
-     srcdir="${test.src.dir}/mapred/testshell"
-     includes="*.java"
-     destdir="${test.mapred.build.testshell}"
-     debug="${javac.debug}"
-     optimize="${javac.optimize}"
-     target="${javac.version}"
-     source="${javac.version}"
-     deprecation="${javac.deprecation}">
-      <compilerarg line="${javac.args} ${javac.args.warnings}"/>
-      <classpath refid="test.mapred.classpath"/>
-    </javac>
-    <delete file="${test.mapred.build.testshell}/testshell.jar"/>
-    <jar jarfile="${test.mapred.build.testshell}/testshell.jar"
-      basedir="${test.mapred.build.testshell}">
-    </jar>
-
-     <delete dir="${test.cache.data}"/>
-     <mkdir dir="${test.cache.data}"/>
-     <delete dir="${test.debug.data}"/>
-     <mkdir dir="${test.debug.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/testscript.txt" todir="${test.debug.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.txt" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.jar" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.zip" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.tar" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.tgz" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/mapred/test.tar.gz" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/mapred/org/apache/hadoop/cli/testMRConf.xml" todir="${test.cache.data}"/>
-     <copy file="${test.src.dir}/hdfs/org/apache/hadoop/cli/clitest_data/data60bytes" todir="${test.cache.data}"/>
-  </target>
--->
   <target name="compile-hdfs-with-mr-test" depends="compile-hdfs-test">
     <mkdir dir="${test.hdfs.with.mr.build.classes}"/>
     <javac 
@@ -784,11 +456,6 @@
       <sysproperty key="hadoop.policy.file" value="hadoop-policy.xml"/>
       <sysproperty key="java.library.path"
        value="${build.native}/lib:${lib.dir}/native/${build.platform}"/>
-      <sysproperty key="install.c++.examples" value="${install.c++.examples}"/>
-      <!-- set compile.c++ in the child jvm only if it is set -->
-      <syspropertyset dynamic="no">
-         <propertyref name="compile.c++"/>
-      </syspropertyset>
       <classpath refid="test.classpath"/>
       <formatter type="${test.junit.output.format}" />
       <batchtest todir="${test.build.dir}" unless="testcase">
@@ -804,7 +471,6 @@
   </target>  
 
   <target name="run-test-hdfs-with-mr" depends="compile-hdfs-with-mr-test" description="Run hdfs unit tests that require mapred">
-
     <delete dir="${test.build.data}"/>
     <mkdir dir="${test.build.data}"/>
     <delete dir="${test.log.dir}"/>
@@ -828,11 +494,6 @@
       <sysproperty key="hadoop.policy.file" value="hadoop-policy.xml"/>
       <sysproperty key="java.library.path"
        value="${build.native}/lib:${lib.dir}/native/${build.platform}"/>
-      <sysproperty key="install.c++.examples" value="${install.c++.examples}"/>
-      <!-- set compile.c++ in the child jvm only if it is set -->
-      <syspropertyset dynamic="no">
-         <propertyref name="compile.c++"/>
-      </syspropertyset>
       <classpath refid="test.hdfs.with.mr.classpath"/>
       <formatter type="${test.junit.output.format}" />
       <batchtest todir="${test.build.dir}" unless="testcase">
@@ -860,7 +521,7 @@
     </subant> 
   </target>
 
-  <target name="test-core" description="Run core, hdfs and mapred unit tests">
+  <target name="test-core" description="Run hdfs, hdfs with mapred unit tests">
     <delete file="${test.build.dir}/testsfailed"/>
     <property name="continueOnFailure" value="true"/>
     <antcall target="run-test-hdfs"/>
@@ -1059,7 +720,7 @@
        <doclet name="jdiff.JDiff"
                path="${jdiff.jar}:${xerces.jar}">
          <param name="-apidir" value="${jdiff.xml.dir}"/>
-         <param name="-apiname" value="${ant.project.name} ${version}"/>
+         <param name="-apiname" value="hadoop-hdfs ${version}"/>
        </doclet>
        <packageset dir="src/java"/>
        <classpath >
@@ -1084,8 +745,8 @@
 	     maxmemory="${javadoc.maxmemory}">
        <doclet name="jdiff.JDiff"
                path="${jdiff.jar}:${xerces.jar}">
-         <param name="-oldapi" value="${ant.project.name} ${jdiff.stable}"/>
-         <param name="-newapi" value="${ant.project.name} ${version}"/>
+         <param name="-oldapi" value="hadoop-hdfs ${jdiff.stable}"/>
+         <param name="-newapi" value="hadoop-hdfs ${version}"/>
          <param name="-oldapidir" value="${jdiff.xml.dir}"/>
          <param name="-newapidir" value="${jdiff.xml.dir}"/>
          <param name="-javadocold" value="${jdiff.stable.javadoc}"/>
@@ -1135,13 +796,6 @@
       </fileset>
     </copy>
 
-  	<exec dir="${dist.dir}" executable="sh" failonerror="true">
-	  <env key="BASE_NATIVE_LIB_DIR" value="${lib.dir}/native"/>
-	  <env key="BUILD_NATIVE_DIR" value="${build.dir}/native"/>
-	  <env key="DIST_LIB_DIR" value="${dist.dir}/lib/native"/>
-	  <arg line="${native.src.dir}/packageNativeHadoop.sh"/>
-    </exec>
-
     <subant target="package">
       <!--Pass down the version in case its needed again and the target
       distribution directory so contribs know where to install to.-->
@@ -1158,10 +812,6 @@
       <fileset file="${build.dir}/${final.name}-*.jar"/>
     </copy>
     
-    <copy todir="${dist.dir}/bin">
-      <fileset dir="bin"/>
-    </copy>
-
     <copy todir="${dist.dir}/conf">
       <fileset dir="${conf.dir}" excludes="**/*.template"/>
     </copy>
@@ -1188,13 +838,6 @@
 
     <copy todir="${dist.dir}/" file="build.xml"/>
 
-    <chmod perm="ugo+x" type="file" parallel="false">
-        <fileset dir="${dist.dir}/src/contrib/">
-          <include name="*/bin/*" />
-        </fileset>
-        <fileset dir="${dist.dir}/src/contrib/ec2/bin/image"/>
-    </chmod>
-
   </target>
 
   <!-- ================================================================== -->
@@ -1237,13 +880,6 @@
       </fileset>
     </copy>
 
-  	<exec dir="${dist.dir}" executable="sh" failonerror="true">
-	  <env key="BASE_NATIVE_LIB_DIR" value="${lib.dir}/native"/>
-	  <env key="BUILD_NATIVE_DIR" value="${build.dir}/native"/>
-	  <env key="DIST_LIB_DIR" value="${dist.dir}/lib/native"/>
-	  <arg line="${native.src.dir}/packageNativeHadoop.sh"/>
-    </exec>
-
     <subant target="package">
       <!--Pass down the version in case its needed again and the target
       distribution directory so contribs know where to install to.-->
@@ -1259,10 +895,6 @@
     <copy todir="${dist.dir}"> 
       <fileset file="${build.dir}/${final.name}-*.jar"/>
     </copy>
-    
-    <copy todir="${dist.dir}/bin">
-      <fileset dir="bin"/>
-    </copy>
 
     <copy todir="${dist.dir}/conf">
       <fileset dir="${conf.dir}" excludes="**/*.template"/>
@@ -1280,15 +912,8 @@
       </fileset>
     </copy>
   	
-    <copy todir="${dist.dir}/c++" includeEmptyDirs="false">
-      <fileset dir="${build.dir}/c++"/>
-    </copy>
-
     <copy todir="${dist.dir}/" file="build.xml"/>
 
-    <chmod perm="ugo+x" type="file" parallel="false">
-        <fileset dir="${dist.dir}/bin"/>
-    </chmod>
   </target>
 
   <target name="binary" depends="bin-package" description="Make tarball without source and documentation">
@@ -1300,9 +925,6 @@
           <exclude name="${final.name}/docs/**" />
           <include name="${final.name}/**" />
         </tarfileset>
-        <tarfileset dir="${build.dir}" mode="755">
-          <include name="${final.name}/bin/*" />
-        </tarfileset>
       </param.listofitems>
     </macro_tar>
   </target>
@@ -1337,204 +959,6 @@
      </subant>  	
   </target>
 	
- <target name="test-c++-libhdfs" depends="compile-c++-libhdfs, compile-core" if="islibhdfs">
-    <delete dir="${test.libhdfs.dir}"/>
-    <mkdir dir="${test.libhdfs.dir}"/>
-    <mkdir dir="${test.libhdfs.dir}/logs"/>
-    <mkdir dir="${test.libhdfs.dir}/hdfs/name"/>
-
-    <exec dir="${build.c++.libhdfs}" executable="${make.cmd}" failonerror="true">
-        <env key="OS_NAME" value="${os.name}"/>
-        <env key="OS_ARCH" value="${os.arch}"/>
-        <env key="JVM_ARCH" value="${jvm.arch}"/>
-        <env key="LIBHDFS_BUILD_DIR" value="${build.c++.libhdfs}"/>
-        <env key="HADOOP_HOME" value="${basedir}"/>
-        <env key="HADOOP_CONF_DIR" value="${test.libhdfs.conf.dir}"/>
-        <env key="HADOOP_LOG_DIR" value="${test.libhdfs.dir}/logs"/>
-        <env key="LIBHDFS_SRC_DIR" value="${c++.libhdfs.src}"/>
-        <env key="LIBHDFS_INSTALL_DIR" value="${install.c++}/lib"/>  
-        <env key="LIB_DIR" value="${common.ivy.lib.dir}"/>
-		<arg value="test"/>
-    </exec>
-  </target>
-
-<!-- ================================================================== -->
-<!-- librecordio targets.                                               -->
-<!-- ================================================================== -->		
-
-  <target name="compile-librecordio" depends="init" if="librecordio" >
-     <mkdir dir="${build.librecordio}"/>
-     <exec dir="${librecordio.src}" executable="${make.cmd}" failonerror="true">
-        <env key="XERCESCROOT" value="${xercescroot}"/>
-        <env key="LIBRECORDIO_BUILD_DIR" value="${build.librecordio}"/>
-     </exec>
-  </target>
-  	
-  <target name="test-librecordio" depends="compile-librecordio, compile-core" if="librecordio">
-    <delete dir="${librecordio.test.dir}"/>
-    <mkdir dir="${librecordio.test.dir}"/>
-    <exec dir="${librecordio.src}/test" executable="${make.cmd}" failonerror="true">
-        <env key="HADOOP_HOME" value="${basedir}"/>
-	<env key="XERCESCROOT" value="${xercescroot}"/>
-        <env key="LIBRECORDIO_BUILD_DIR" value="${build.librecordio}"/> 	
-        <env key="LIBRECORDIO_TEST_DIR" value="${librecordio.test.dir}"/>
-      		<arg value="all"/>		
-    </exec>
-  </target>
-
-  <target name="package-librecordio" depends="compile-librecordio" if="librecordio">
-    <mkdir dir="${dist.dir}/librecordio"/> 
-    <copy todir="${dist.dir}/librecordio">
-       <fileset dir="${build.librecordio}" casesensitive="yes" followsymlinks="false">
-          <exclude name="**/tests/**"/>
-          <exclude name="*.so"/> 
-          <exclude name="*.o"/>
-       </fileset>
-    </copy>
-    <chmod perm="ugo+x" type="file">
-       <fileset dir="${dist.dir}/librecordio"/>
-    </chmod>
-  </target>
- 
-  <target name="create-c++-configure" depends="init" if="compile.c++">
-    <exec executable="autoreconf" dir="${c++.utils.src}" searchpath="yes" 
-          failonerror="yes">
-       <arg value="-if"/>
-    </exec>
-    <exec executable="autoreconf" dir="${c++.pipes.src}" searchpath="yes" 
-          failonerror="yes">
-       <arg value="-if"/>
-    </exec>
-    <exec executable="autoreconf" dir="${c++.examples.pipes.src}" 
-          searchpath="yes" failonerror="yes">
-       <arg value="-if"/>
-    </exec>
-    <antcall target="create-c++-configure-libhdfs"/>
-  </target>
-   
-  <target name="create-c++-configure-libhdfs" depends="check-c++-libhdfs" if="islibhdfs">
-    <exec executable="autoreconf" dir="${c++.libhdfs.src}" 
-          searchpath="yes" failonerror="yes">
-       <arg value="-if"/>
-    </exec>
-  </target>
-
-  <target name="check-c++-makefiles" depends="init" if="compile.c++">
-    <condition property="need.c++.utils.makefile">
-       <not> <available file="${build.c++.utils}/Makefile"/> </not>
-    </condition>
-    <condition property="need.c++.pipes.makefile">
-       <not> <available file="${build.c++.pipes}/Makefile"/> </not>
-    </condition>
-    <condition property="need.c++.examples.pipes.makefile">
-       <not> <available file="${build.c++.examples.pipes}/Makefile"/> </not>
-    </condition>
-  </target>
-
-  <target name="check-c++-libhdfs">
-    <condition property="islibhdfs">
-      <and>
-        <isset property="compile.c++"/>
-        <isset property="libhdfs"/>
-      </and>
-    </condition>
-  </target>
-
-  <target name="check-c++-makefile-libhdfs" depends="init,check-c++-libhdfs" if="islibhdfs">
-    <condition property="need.c++.libhdfs.makefile">
-       <not> <available file="${build.c++.libhdfs}/Makefile"/> </not>
-    </condition>
-  </target>
-
-  <target name="create-c++-libhdfs-makefile" depends="check-c++-makefile-libhdfs" 
-                                           if="need.c++.libhdfs.makefile">
-    <mkdir dir="${build.c++.libhdfs}"/>
-    <chmod file="${c++.libhdfs.src}/configure" perm="ugo+x"/>
-    <exec executable="${c++.libhdfs.src}/configure" dir="${build.c++.libhdfs}"
-          failonerror="yes">
-      <env key="ac_cv_func_malloc_0_nonnull" value="yes"/>
-      <env key="JVM_ARCH" value="${jvm.arch}"/>
-      <arg value="--prefix=${install.c++}"/>
-    </exec>
-  </target>
-
-  <target name="create-c++-utils-makefile" depends="check-c++-makefiles" 
-                                           if="need.c++.utils.makefile">
-    <mkdir dir="${build.c++.utils}"/>
-    <chmod file="${c++.utils.src}/configure" perm="ugo+x"/>
-    <exec executable="${c++.utils.src}/configure" dir="${build.c++.utils}"
-          failonerror="yes">
-      <arg value="--prefix=${install.c++}"/>
-    </exec>
-  </target>
-
-  <target name="compile-c++-utils" depends="create-c++-utils-makefile"
-                                   if="compile.c++">
-    <exec executable="${make.cmd}" dir="${build.c++.utils}" searchpath="yes" 
-          failonerror="yes">
-      <arg  value="install"/>
-    </exec>
-  </target>
-
-  <target name="create-c++-pipes-makefile" depends="check-c++-makefiles" 
-                                           if="need.c++.pipes.makefile">
-    <mkdir dir="${build.c++.pipes}"/>
-    <chmod file="${c++.pipes.src}/configure" perm="ugo+x"/>
-    <exec executable="${c++.pipes.src}/configure" dir="${build.c++.pipes}"
-          failonerror="yes">
-      <arg value="--prefix=${install.c++}"/>
-    </exec>
-  </target>
-
-  <target name="compile-c++-pipes" 
-          depends="create-c++-pipes-makefile,compile-c++-utils"
-          if="compile.c++">
-    <exec executable="${make.cmd}" dir="${build.c++.pipes}" searchpath="yes" 
-          failonerror="yes">
-      <arg value="install"/>
-    </exec>
-  </target>
-
-  <target name="compile-c++" 
-          depends="compile-c++-pipes"/>
-
-  <target name="create-c++-examples-pipes-makefile" 
-          depends="check-c++-makefiles" 
-          if="need.c++.examples.pipes.makefile">
-    <mkdir dir="${build.c++.examples.pipes}"/>
-    <chmod file="${c++.examples.pipes.src}/configure" perm="ugo+x"/>
-    <exec executable="${c++.examples.pipes.src}/configure" 
-          dir="${build.c++.examples.pipes}"
-          failonerror="yes">
-      <arg value="--prefix=${install.c++.examples}"/>
-      <arg value="--with-hadoop-utils=${install.c++}"/>
-      <arg value="--with-hadoop-pipes=${install.c++}"/>
-    </exec>
-  </target>
-
-  <target name="compile-c++-examples-pipes" 
-          depends="create-c++-examples-pipes-makefile,compile-c++-pipes"
-          if="compile.c++">
-    <exec executable="${make.cmd}" dir="${build.c++.examples.pipes}" searchpath="yes" 
-          failonerror="yes">
-      <arg  value="install"/>
-    </exec>
-  </target>
-
-  <target name="compile-c++-examples" 
-          depends="compile-c++-examples-pipes"/>
-
-  <target name="compile-c++-libhdfs" depends="create-c++-libhdfs-makefile" if="islibhdfs">
-    <exec executable="${make.cmd}" dir="${build.c++.libhdfs}" searchpath="yes"
-          failonerror="yes">
-      <env key="ac_cv_func_malloc_0_nonnull" value="yes"/>
-      <env key="JVM_ARCH" value="${jvm.arch}"/>
-      <arg value="install"/>
-    </exec>
-  </target>
-
-
-
   <target name="compile-ant-tasks" depends="compile-core">
     <javac
         encoding="${build.encoding}"
@@ -1554,20 +978,18 @@
   <target name="ant-tasks" depends="jar, compile-ant-tasks">
     <copy file="${anttasks.dir}/org/apache/hadoop/ant/antlib.xml"
           todir="${build.anttasks}/org/apache/hadoop/ant"/>
-    <jar destfile="${build.dir}/${final.name}-ant.jar">
+    <jar destfile="${build.dir}/${ant.final.name}.jar">
       <fileset dir="${build.anttasks}"/>
     </jar>
   </target>
 
-
-
  <target name="clover" depends="clover.setup, clover.info" description="Instrument the Unit tests using Clover.  To use, specify -Dclover.home=&lt;base of clover installation&gt; -Drun.clover=true on the command line."/>
 
 <target name="clover.setup" if="clover.enabled">
    <taskdef resource="cloverlib.xml" classpath="${clover.jar}"/>
    <mkdir dir="${clover.db.dir}"/>
    <clover-setup initString="${clover.db.dir}/hadoop_coverage.db">
-     <fileset dir="src" includes="core/**/* tools/**/* hdfs/**/* mapred/**/*"/>
+     <fileset dir="src" includes="java/**/*"/>
    </clover-setup>
 </target>
 
@@ -1773,7 +1195,7 @@
     description="Retrieve Ivy-managed artifacts for the test configurations">
     <ivy:retrieve settingsRef="${ant.project.name}.ivy.settings"
       pattern="${build.ivy.lib.dir}/${ivy.artifact.retrieve.pattern}"/>
-    <ivy:cachepath pathid="ivy.test.classpath" conf="test"/>
+    <ivy:cachepath pathid="ivy-test.classpath" conf="test"/>
   </target>
 
   <target name="ivy-retrieve-common" depends="ivy-resolve-common"
@@ -1804,10 +1226,10 @@
     <fail>
       <condition >
         <not>
-          <available file="${hadoop.jar}" />
+          <available file="${hadoop-hdfs.jar}" />
         </not>
       </condition>
-      Not found: ${hadoop.jar}
+      Not found: ${hadoop-hdfs.jar}
       Please run the target "jar" in the main build file
     </fail>
 
@@ -1840,7 +1262,7 @@
 
 
   <target name="copy-jar-to-maven" depends="ready-to-publish">
-    <copy file="${hadoop.jar}"
+    <copy file="${hadoop-hdfs.jar}"
       tofile="${build.ivy.maven.jar}"/>
     <checksum file="${build.ivy.maven.jar}" algorithm="md5"/>
   </target>
@@ -1855,7 +1277,7 @@
     </copy>
    </presetdef>
 
-   <expandingcopy file="ivy/hadoop-core.pom"
+   <expandingcopy file="ivy/hadoop-hdfs.pom"
       tofile="${build.ivy.maven.pom}"/>
    <checksum file="${build.ivy.maven.pom}" algorithm="md5"/>
   </target>
@@ -1866,52 +1288,4 @@
 
   </target>
 
-  <!-- taskcontroller targets -->
-  <target name="init-task-controller-build">
-    <mkdir dir="${build.c++.task-controller}" />
-    <copy todir="${build.c++.task-controller}">
-      <fileset dir="${c++.task-controller.src}" includes="*.c">
-      </fileset>
-      <fileset dir="${c++.task-controller.src}" includes="*.h">
-      </fileset>
-    </copy>
-    <chmod file="${c++.task-controller.src}/configure" perm="ugo+x"/> 
-    <condition property="task-controller.conf.dir.passed">
-      <not>
-        <equals arg1="${hadoop.conf.dir}" arg2="$${hadoop.conf.dir}"/>
-      </not>
-    </condition>
-  </target>
-  <target name="configure-task-controller" depends="init,
-          init-task-controller-build,
-          task-controller-configuration-with-confdir,
-          task-controller-configuration-with-no-confdir">
-  </target>
-  <target name="task-controller-configuration-with-confdir" 
-          if="task-controller.conf.dir.passed" >
-    <exec executable="${c++.task-controller.src}/configure" 
-          dir="${build.c++.task-controller}" failonerror="yes">
-      <arg value="--prefix=${task-controller.install.dir}" />
-      <arg value="--with-confdir=${hadoop.conf.dir}" />
-    </exec>
-  </target>
-  <target name="task-controller-configuration-with-no-confdir" 
-          unless="task-controller.conf.dir.passed">
-    <exec executable="${c++.task-controller.src}/configure" 
-          dir="${build.c++.task-controller}" failonerror="yes">
-      <arg value="--prefix=${task-controller.install.dir}" />
-    </exec>
-  </target>
-  <!--
-    * Create the installation directory.
-    * Do a make install.
-   -->
-  <target name="task-controller" depends="configure-task-controller">
-    <mkdir dir="${task-controller.install.dir}" />
-    <exec executable="${make.cmd}" dir="${build.c++.task-controller}" 
-        searchpath="yes" failonerror="yes">
-      <arg value="install" />
-    </exec>
-  </target>
-  <!-- end of task-controller target -->
 </project>

A diferenza do arquivo foi suprimida porque é demasiado grande
+ 11 - 0
lib/jdiff/hadoop-hdfs_0.20.0.xml


Algúns arquivos non se mostraron porque demasiados arquivos cambiaron neste cambio