Browse Source

HDFS-3537. Move libhdfs and fuse-dfs source to native subdirectories. Contributed by Colin Patrick McCabe

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1362119 13f79535-47bb-0310-9956-ffa450edef68
Eli Collins 13 năm trước cách đây
mục cha
commit
69f7a329ba
62 tập tin đã thay đổi với 22 bổ sung567 xóa
  1. 3 0
      hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
  2. 16 20
      hadoop-hdfs-project/hadoop-hdfs/src/CMakeLists.txt
  3. 0 312
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/build-contrib.xml
  4. 0 87
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/build.xml
  5. 0 18
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/global_footer.mk
  6. 0 51
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/global_header.mk
  7. 0 71
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/ivy.xml
  8. 0 5
      hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/ivy/libraries.properties
  9. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/CMakeLists.txt
  10. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/doc/README
  11. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_connect.c
  12. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_connect.h
  13. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_context_handle.h
  14. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs.c
  15. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs.h
  16. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs_wrapper.sh
  17. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_file_handle.h
  18. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls.h
  19. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_access.c
  20. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_chmod.c
  21. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_chown.c
  22. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_create.c
  23. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_flush.c
  24. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_getattr.c
  25. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_mkdir.c
  26. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_mknod.c
  27. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_open.c
  28. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_read.c
  29. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_readdir.c
  30. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_release.c
  31. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_rename.c
  32. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_rmdir.c
  33. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_statfs.c
  34. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_symlink.c
  35. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_truncate.c
  36. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_unlink.c
  37. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_utimens.c
  38. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_write.c
  39. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_init.c
  40. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_init.h
  41. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_options.c
  42. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_options.h
  43. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_stat_struct.c
  44. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_stat_struct.h
  45. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_trash.c
  46. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_trash.h
  47. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_users.c
  48. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_users.h
  49. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/test/TestFuseDFS.java
  50. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/expect.h
  51. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs.c
  52. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs.h
  53. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs_test.h
  54. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/jni_helper.c
  55. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/jni_helper.h
  56. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.c
  57. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.h
  58. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_ops.c
  59. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_read.c
  60. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_write.c
  61. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test_libhdfs_threaded.c
  62. 0 0
      hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test_native_mini_dfs.c

+ 3 - 0
hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt

@@ -333,6 +333,9 @@ Branch-2 ( Unreleased changes )
     HDFS-3659. Add missing @Override to methods across the hadoop-hdfs
     project. (Brandon Li via harsh)
 
+    HDFS-3537. Move libhdfs and fuse-dfs source to native subdirectories.
+    (Colin Patrick McCabe via eli)
+
   OPTIMIZATIONS
 
     HDFS-2982. Startup performance suffers when there are many edit log

+ 16 - 20
hadoop-hdfs-project/hadoop-hdfs/src/CMakeLists.txt

@@ -87,15 +87,15 @@ include_directories(
     ${CMAKE_CURRENT_SOURCE_DIR}
     ${CMAKE_BINARY_DIR}
     ${JNI_INCLUDE_DIRS}
-    main/native/
+    main/native/libhdfs
 )
 
 set(_FUSE_DFS_VERSION 0.1.0)
 CONFIGURE_FILE(${CMAKE_SOURCE_DIR}/config.h.cmake ${CMAKE_BINARY_DIR}/config.h)
 
 add_dual_library(hdfs
-    main/native/hdfs.c
-    main/native/hdfsJniHelper.c
+    main/native/libhdfs/hdfs.c
+    main/native/libhdfs/jni_helper.c
 )
 target_link_dual_libraries(hdfs
     ${JAVA_JVM_LIBRARY}
@@ -105,49 +105,46 @@ set(LIBHDFS_VERSION "0.0.0")
 set_target_properties(hdfs PROPERTIES
     SOVERSION ${LIBHDFS_VERSION})
 
-add_executable(hdfs_test
-    main/native/hdfs_test.c
+add_executable(test_libhdfs_ops
+    main/native/libhdfs/test/test_libhdfs_ops.c
 )
-target_link_libraries(hdfs_test
+target_link_libraries(test_libhdfs_ops
     hdfs
     ${JAVA_JVM_LIBRARY}
 )
-output_directory(hdfs_test target/usr/local/bin)
 
-add_executable(hdfs_read
-    main/native/hdfs_read.c
+add_executable(test_libhdfs_read
+    main/native/libhdfs/test/test_libhdfs_read.c
 )
-target_link_libraries(hdfs_read
+target_link_libraries(test_libhdfs_read
     hdfs
     ${JAVA_JVM_LIBRARY}
 )
-output_directory(hdfs_read target/usr/local/bin)
 
-add_executable(hdfs_write
-    main/native/hdfs_write.c
+add_executable(test_libhdfs_write
+    main/native/libhdfs/test/test_libhdfs_write.c
 )
-target_link_libraries(hdfs_write
+target_link_libraries(test_libhdfs_write
     hdfs
     ${JAVA_JVM_LIBRARY}
 )
-output_directory(hdfs_write target/usr/local/bin)
 
 add_library(native_mini_dfs
-    main/native/native_mini_dfs.c
+    main/native/libhdfs/native_mini_dfs.c
 )
 target_link_libraries(native_mini_dfs
     hdfs
 )
 
 add_executable(test_native_mini_dfs
-    main/native/test_native_mini_dfs.c
+    main/native/libhdfs/test_native_mini_dfs.c
 )
 target_link_libraries(test_native_mini_dfs
     native_mini_dfs
 )
 
 add_executable(test_libhdfs_threaded
-    main/native/test_libhdfs_threaded.c
+    main/native/libhdfs/test_libhdfs_threaded.c
 )
 target_link_libraries(test_libhdfs_threaded
     hdfs
@@ -155,5 +152,4 @@ target_link_libraries(test_libhdfs_threaded
     pthread
 )
 
-add_subdirectory(contrib/fuse-dfs/src)
-
+add_subdirectory(main/native/fuse-dfs)

+ 0 - 312
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/build-contrib.xml

@@ -1,312 +0,0 @@
-<?xml version="1.0"?>
-
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-
-<!-- Imported by contrib/*/build.xml files to share generic targets. -->
-
-<project name="hadoopbuildcontrib" xmlns:ivy="antlib:org.apache.ivy.ant">
-
-  <property name="name" value="${ant.project.name}"/>
-  <property name="root" value="${basedir}"/>
-  <property name="hadoop.root" location="${root}/../../../"/>
-
-  <!-- Load all the default properties, and any the user wants    -->
-  <!-- to contribute (without having to type -D or edit this file -->
-  <property file="${user.home}/${name}.build.properties" />
-  <property file="${root}/build.properties" />
-  <property file="${hadoop.root}/build.properties" />
-
-  <property name="src.dir"  location="${root}/src/java"/>
-  <property name="src.test" location="${root}/src/test"/>
-  <property name="src.examples" location="${root}/src/examples"/>
-
-  <available file="${src.examples}" type="dir" property="examples.available"/>
-  <available file="${src.test}" type="dir" property="test.available"/>
-
-  <property name="conf.dir" location="${hadoop.root}/conf"/>
-  <property name="test.junit.output.format" value="plain"/>
-  <property name="test.output" value="no"/>
-  <property name="test.timeout" value="900000"/>
-  <property name="build.dir" location="${hadoop.root}/build/contrib/${name}"/>
-  <property name="build.webapps.root.dir" value="${hadoop.root}/build/web"/>
-  <property name="build.webapps" value="${build.webapps.root.dir}/webapps"/>
-  <property name="build.classes" location="${build.dir}/classes"/>
-  <!-- NB: sun.arch.data.model is not supported on all platforms -->
-  <property name="build.platform"
-            value="${os.name}-${os.arch}-${sun.arch.data.model}"/>
-  <property name="build.c++.libhdfs" value="${build.dir}/../../c++/${build.platform}/lib"/>
-  <property name="build.test" location="${build.dir}/test"/>
-  <property name="build.examples" location="${build.dir}/examples"/>
-  <property name="hadoop.log.dir" location="${build.dir}/test/logs"/>
-  <!-- all jars together -->
-  <property name="javac.deprecation" value="off"/>
-  <property name="javac.debug" value="on"/>
-  <property name="build.ivy.lib.dir" value="${hadoop.root}/build/ivy/lib"/> 
-
-  <property name="javadoc.link"
-            value="http://java.sun.com/j2se/1.4/docs/api/"/>
-
-  <property name="build.encoding" value="ISO-8859-1"/>
-
-  <fileset id="lib.jars" dir="${root}" includes="lib/*.jar"/>
-
-
-   <!-- IVY properties set here -->
-  <property name="ivy.dir" location="ivy" />
-  <property name="ivysettings.xml" location="${hadoop.root}/ivy/ivysettings.xml"/>
-  <loadproperties srcfile="${ivy.dir}/libraries.properties"/>
-  <loadproperties srcfile="ivy/libraries.properties"/>
-  <property name="ivy.jar" location="${hadoop.root}/ivy/ivy-${ivy.version}.jar"/>
-  <property name="ivy_repo_url" 
-	value="http://repo2.maven.org/maven2/org/apache/ivy/ivy/${ivy.version}/ivy-${ivy.version}.jar" />
-  <property name="build.dir" location="build" />
-  <property name="build.ivy.dir" location="${build.dir}/ivy" />
-  <property name="build.ivy.lib.dir" location="${build.ivy.dir}/lib" />
-  <property name="build.ivy.report.dir" location="${build.ivy.dir}/report" />
-  <property name="common.ivy.lib.dir" location="${build.ivy.lib.dir}/${ant.project.name}/common"/> 
-
-  <!--this is the naming policy for artifacts we want pulled down-->
-  <property name="ivy.artifact.retrieve.pattern"
-    			value="${ant.project.name}/[conf]/[artifact]-[revision](-[classifier]).[ext]"/>
-
-  <!-- the normal classpath -->
-  <path id="contrib-classpath">
-    <pathelement location="${build.classes}"/>
-    <fileset refid="lib.jars"/>
-    <pathelement location="${hadoop.root}/build/classes"/>
-    <fileset dir="${hadoop.root}/lib">
-      <include name="**/*.jar" />
-    </fileset>
-    <path refid="${ant.project.name}.common-classpath"/>
-    <pathelement path="${clover.jar}"/>
-  </path>
-
-  <!-- the unit test classpath -->
-  <path id="test.classpath">
-    <pathelement location="${build.test}"/>
-    <pathelement location="${build.webapps.root.dir}"/>
-    <pathelement location="${hadoop.root}/build/test/core/classes"/>
-    <pathelement location="${hadoop.root}/build/test/hdfs/classes"/>
-    <pathelement location="${hadoop.root}/build/test/mapred/classes"/>
-    <pathelement location="${hadoop.root}/src/contrib/test"/>
-    <pathelement location="${conf.dir}"/>
-    <pathelement location="${hadoop.root}/build"/>
-    <pathelement location="${build.examples}"/>
-    <path refid="contrib-classpath"/>
-  </path>
-
-
-  <!-- to be overridden by sub-projects -->
-  <target name="check-contrib"/>
-  <target name="init-contrib"/>
-
-  <!-- ====================================================== -->
-  <!-- Stuff needed by all targets                            -->
-  <!-- ====================================================== -->
-  <target name="init" depends="check-contrib" unless="skip.contrib">
-    <echo message="contrib: ${name}"/>
-    <mkdir dir="${build.dir}"/>
-    <mkdir dir="${build.classes}"/>
-    <mkdir dir="${build.test}"/>
-    <mkdir dir="${build.examples}"/>
-    <mkdir dir="${hadoop.log.dir}"/>
-    <antcall target="init-contrib"/>
-  </target>
-
-
-  <!-- ====================================================== -->
-  <!-- Compile a Hadoop contrib's files                       -->
-  <!-- ====================================================== -->
-  <target name="compile" depends="init, ivy-retrieve-common" unless="skip.contrib">
-    <echo message="contrib: ${name}"/>
-    <javac
-     encoding="${build.encoding}"
-     srcdir="${src.dir}"
-     includes="**/*.java"
-     destdir="${build.classes}"
-     debug="${javac.debug}"
-     deprecation="${javac.deprecation}">
-     <classpath refid="contrib-classpath"/>
-    </javac>
-  </target>
-
-
-  <!-- ======================================================= -->
-  <!-- Compile a Hadoop contrib's example files (if available) -->
-  <!-- ======================================================= -->
-  <target name="compile-examples" depends="compile" if="examples.available">
-    <echo message="contrib: ${name}"/>
-    <javac
-     encoding="${build.encoding}"
-     srcdir="${src.examples}"
-     includes="**/*.java"
-     destdir="${build.examples}"
-     debug="${javac.debug}">
-     <classpath refid="contrib-classpath"/>
-    </javac>
-  </target>
-
-
-  <!-- ================================================================== -->
-  <!-- Compile test code                                                  -->
-  <!-- ================================================================== -->
-  <target name="compile-test" depends="compile-examples" if="test.available">
-    <echo message="contrib: ${name}"/>
-    <javac
-     encoding="${build.encoding}"
-     srcdir="${src.test}"
-     includes="**/*.java"
-     destdir="${build.test}"
-     debug="${javac.debug}">
-    <classpath refid="test.classpath"/>
-    </javac>
-  </target>
-  
-
-  <!-- ====================================================== -->
-  <!-- Make a Hadoop contrib's jar                            -->
-  <!-- ====================================================== -->
-  <target name="jar" depends="compile" unless="skip.contrib">
-    <echo message="contrib: ${name}"/>
-    <jar
-      jarfile="${build.dir}/hadoop-${version}-${name}.jar"
-      basedir="${build.classes}"      
-    />
-  </target>
-
-  
-  <!-- ====================================================== -->
-  <!-- Make a Hadoop contrib's examples jar                   -->
-  <!-- ====================================================== -->
-  <target name="jar-examples" depends="compile-examples"
-          if="examples.available" unless="skip.contrib">
-    <echo message="contrib: ${name}"/>
-    <jar jarfile="${build.dir}/hadoop-${version}-${name}-examples.jar">
-      <fileset dir="${build.classes}">
-      </fileset>
-      <fileset dir="${build.examples}">
-      </fileset>
-    </jar>
-  </target>
-  
-  <!-- ====================================================== -->
-  <!-- Package a Hadoop contrib                               -->
-  <!-- ====================================================== -->
-  <target name="package" depends="jar, jar-examples" unless="skip.contrib"> 
-    <mkdir dir="${dist.dir}/contrib/${name}"/>
-    <copy todir="${dist.dir}/contrib/${name}" includeEmptyDirs="false" flatten="true">
-      <fileset dir="${build.dir}">
-        <include name="hadoop-${version}-${name}.jar" />
-      </fileset>
-    </copy>
-  </target>
-  
-  <!-- ================================================================== -->
-  <!-- Run unit tests                                                     -->
-  <!-- ================================================================== -->
-  <target name="test" depends="compile-test, compile" if="test.available">
-    <echo message="contrib: ${name}"/>
-    <delete dir="${hadoop.log.dir}"/>
-    <mkdir dir="${hadoop.log.dir}"/>
-    <junit
-      printsummary="yes" showoutput="${test.output}" 
-      haltonfailure="no" fork="yes" maxmemory="256m"
-      errorProperty="tests.failed" failureProperty="tests.failed"
-      timeout="${test.timeout}">
-      
-      <sysproperty key="test.build.data" value="${build.test}/data"/>
-      <sysproperty key="build.test" value="${build.test}"/>
-      <sysproperty key="contrib.name" value="${name}"/>
-      
-      <!-- requires fork=yes for: 
-        relative File paths to use the specified user.dir 
-        classpath to use build/contrib/*.jar
-      -->
-      <sysproperty key="java.net.preferIPv4Stack" value="true"/>
-      <sysproperty key="user.dir" value="${build.test}/data"/>
-      
-      <sysproperty key="fs.default.name" value="${fs.default.name}"/>
-      <sysproperty key="hadoop.test.localoutputfile" value="${hadoop.test.localoutputfile}"/>
-      <sysproperty key="hadoop.log.dir" value="${hadoop.log.dir}"/> 
-      <sysproperty key="taskcontroller-path" value="${taskcontroller-path}"/>
-      <sysproperty key="taskcontroller-user" value="${taskcontroller-user}"/>
-      <classpath refid="test.classpath"/>
-      <formatter type="${test.junit.output.format}" />
-      <batchtest todir="${build.test}" unless="testcase">
-        <fileset dir="${src.test}"
-                 includes="**/Test*.java" excludes="**/${test.exclude}.java" />
-      </batchtest>
-      <batchtest todir="${build.test}" if="testcase">
-        <fileset dir="${src.test}" includes="**/${testcase}.java"/>
-      </batchtest>
-    </junit>
-    <fail if="tests.failed">Tests failed!</fail>
-  </target>
-
-  <!-- ================================================================== -->
-  <!-- Clean.  Delete the build files, and their directories              -->
-  <!-- ================================================================== -->
-  <target name="clean">
-    <echo message="contrib: ${name}"/>
-    <delete dir="${build.dir}"/>
-  </target>
-
-  <target name="ivy-probe-antlib" >
-    <condition property="ivy.found">
-      <typefound uri="antlib:org.apache.ivy.ant" name="cleancache"/>
-    </condition>
-  </target>
-
-
-  <target name="ivy-download" description="To download ivy " unless="offline">
-    <get src="${ivy_repo_url}" dest="${ivy.jar}" usetimestamp="true"/>
-  </target>
-
-  <target name="ivy-init-antlib" depends="ivy-download,ivy-probe-antlib" unless="ivy.found">
-    <typedef uri="antlib:org.apache.ivy.ant" onerror="fail"
-      loaderRef="ivyLoader">
-      <classpath>
-        <pathelement location="${ivy.jar}"/>
-      </classpath>
-    </typedef>
-    <fail >
-      <condition >
-        <not>
-          <typefound uri="antlib:org.apache.ivy.ant" name="cleancache"/>
-        </not>
-      </condition>
-      You need Apache Ivy 2.0 or later from http://ant.apache.org/
-      It could not be loaded from ${ivy_repo_url}
-    </fail>
-  </target>
-
-  <target name="ivy-init" depends="ivy-init-antlib">
-    <ivy:configure settingsid="${ant.project.name}.ivy.settings" file="${ivysettings.xml}"/>
-  </target>
-
-  <target name="ivy-resolve-common" depends="ivy-init">
-    <ivy:resolve settingsRef="${ant.project.name}.ivy.settings" conf="common" />
-  </target>
-
-  <target name="ivy-retrieve-common" depends="ivy-resolve-common"
-    description="Retrieve Ivy-managed artifacts for the compile/test configurations">
-    <ivy:retrieve settingsRef="${ant.project.name}.ivy.settings" 
-      pattern="${build.ivy.lib.dir}/${ivy.artifact.retrieve.pattern}" sync="true" />
-    <ivy:cachepath pathid="${ant.project.name}.common-classpath" conf="common" />
-  </target>
-</project>

+ 0 - 87
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/build.xml

@@ -1,87 +0,0 @@
-<?xml version="1.0"?>
-
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-
-<project name="fuse-dfs" default="compile" xmlns:ivy="antlib:org.apache.ivy.ant">
-
-  <import file="build-contrib.xml"/>
-
-  <target name="check-libhdfs-exists">
-    <property name="libhdfs.lib" value="${build.c++.libhdfs}/libhdfs.so"/>
-    <available file="${libhdfs.lib}" property="libhdfs-exists"/>
-    <fail message="libhdfs.so does not exist: ${libhdfs.lib}.">
-      <condition><not><isset property="libhdfs-exists"/></not></condition>
-    </fail>
-  </target>
-
-  <target name="compile">
-    <exec executable="autoreconf" dir="${basedir}" 
-          searchpath="yes" failonerror="yes">
-       <arg value="-if"/>
-    </exec>
-
-    <exec executable="${basedir}/configure" dir="${basedir}"
-          failonerror="yes">
-    </exec>
-
-    <exec executable="make" failonerror="true">
-      <env key="OS_NAME" value="${os.name}"/>
-      <env key="OS_ARCH" value="${os.arch}"/>
-      <env key="HADOOP_PREFIX" value="${hadoop.root}"/>
-      <env key="PACKAGE_VERSION" value="0.1.0"/>
-      <env key="BUILD_PLATFORM" value="${build.platform}" />
-    </exec>
-  </target>
-
-  <target name="jar" />
-  <target name="package" />
-
-  <target name="compile-test" depends="ivy-retrieve-common, check-libhdfs-exists">
-    <javac encoding="${build.encoding}"
-	   srcdir="${src.test}"
-	   includes="**/*.java"
-	   destdir="${build.test}"
-	   debug="${javac.debug}">
-      <classpath refid="test.classpath"/>
-    </javac>
-  </target>
-
-  <target name="test" depends="compile-test,check-libhdfs-exists">
-    <junit showoutput="${test.output}" fork="yes" printsummary="yes"
-           errorProperty="tests.failed" haltonfailure="no" failureProperty="tests.failed">
-      <classpath refid="test.classpath"/>
-      <sysproperty key="test.build.data" value="${build.test}/data"/>
-      <sysproperty key="build.test" value="${build.test}"/>
-      <sysproperty key="user.dir" value="${build.test}/data"/>
-      <sysproperty key="hadoop.log.dir" value="${hadoop.log.dir}"/>
-      <sysproperty key="test.src.dir" value="${test.src.dir}"/>
-      <formatter type="${test.junit.output.format}" />
-      <batchtest todir="${build.test}" unless="testcase">
-        <fileset dir="${src.test}">
-          <include name="**/Test*.java"/>
-        </fileset>
-      </batchtest>
-      <batchtest todir="${build.test}" if="testcase">
-        <fileset dir="${src.test}">
-          <include name="**/${testcase}.java"/>
-        </fileset>
-      </batchtest>
-    </junit>
-    <fail if="tests.failed">Tests failed!</fail>
- </target>
-</project>

+ 0 - 18
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/global_footer.mk

@@ -1,18 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-thriftstyle : $(XBUILT_SOURCES)
-

+ 0 - 51
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/global_header.mk

@@ -1,51 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#     http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-ifneq ($$(XBUILT_SOURCES),)
-    XBUILT_SOURCES := $$(XBUILT_SOURCES) $$(XTARGET)
-else
-    XBUILT_SOURCES := $$(XTARGET)
-endif
-
-showvars:
-	@echo BUILD_SOURCES = $(BUILT_SOURCES)
-	@echo XBUILTSOURCES = $(XBUILT_SOURCES)
-	@echo DEFS = $(DEFS)
-	@echo CXXFLAGS = $(CXXFLAGS)
-	@echo AM_CXXFLAGS = $(AM_CXXFLAGS)
-	@echo CPPFLAGS = $(CPPFLAGS)
-	@echo AM_CPPFLAGS = $(AM_CPPFLAGS)
-	@echo LDFLAGS = $(LDFLAGS)
-	@echo AM_LDFLAGS = $(AM_LDFLAGS)
-	@echo LDADD = $(LDADD)
-	@echo LIBS = $(LIBS)
-	@echo EXTERNAL_LIBS = $(EXTERNAL_LIBS)
-	@echo EXTERNAL_PATH = $(EXTERNAL_PATH)
-	@echo MAKE = $(MAKE)
-	@echo MAKE_FLAGS = $(MAKE_FLAGS)
-	@echo AM_MAKEFLAGS = $(AM_MAKEFLAGS)
-	@echo top_builddir = $(top_builddir)
-	@echo top_srcdir = $(top_srcdir)
-	@echo srcdir = $(srcdir)
-	@echo PHPVAL = $(PHPVAL)
-	@echo PHPCONFIGDIR  = $(PHPCONFIGDIR)
-	@echo PHPCONFIGINCLUDEDIR = $(PHPCONFIGINCLUDEDIR)
-	@echo PHPCONFIGINCLUDES  = $(PHPCONFIGINCLUDES)
-	@echo PHPCONFIGLDFLAGS  = $(PHPCONFIGLDFLAGS)
-	@echo PHPCONFIGLIBS  = $(PHPCONFIGLIBS)
-
-clean-common:
-	rm -rf gen-*

+ 0 - 71
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/ivy.xml

@@ -1,71 +0,0 @@
-<?xml version="1.0" ?>
-<!--
-   Licensed to the Apache Software Foundation (ASF) under one or more
-   contributor license agreements.  See the NOTICE file distributed with
-   this work for additional information regarding copyright ownership.
-   The ASF licenses this file to You under the Apache License, Version 2.0
-   (the "License"); you may not use this file except in compliance with
-   the License.  You may obtain a copy of the License at
-
-       http://www.apache.org/licenses/LICENSE-2.0
-
-   Unless required by applicable law or agreed to in writing, software
-   distributed under the License is distributed on an "AS IS" BASIS,
-   WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-   See the License for the specific language governing permissions and
-   limitations under the License.
--->
-<ivy-module version="1.0" xmlns:m="http://ant.apache.org/ivy/maven">
-  <info organisation="org.apache.hadoop" module="${ant.project.name}">
-    <license name="Apache 2.0"/>
-    <ivyauthor name="Apache Hadoop Team" url="http://hadoop.apache.org"/>
-    <description>
-        FUSE plugin for HDFS
-    </description>
-  </info>
-  <configurations defaultconfmapping="default">
-    <!--these match the Maven configurations-->
-    <conf name="default" extends="master,runtime"/>
-    <conf name="master" description="contains the artifact but no dependencies"/>
-    <conf name="runtime" description="runtime but not the artifact" />
-
-    <conf name="common" visibility="private" 
-      extends="runtime"
-      description="artifacts needed to compile/test the application"/>
-    <conf name="test" visibility="private" extends="runtime"/>
-  </configurations>
-
-  <publications>
-    <!--get the artifact from our module name-->
-    <artifact conf="master"/>
-  </publications>
-  <dependencies>
-    <dependency org="org.apache.hadoop"
-      name="hadoop-common"
-      rev="${hadoop-common.version}"
-      conf="common->default"/>
-    <dependency org="org.apache.hadoop"
-      name="hadoop-common"
-      rev="${hadoop-common.version}"
-      conf="common->default">
-      <artifact name="hadoop-common" type="tests" ext="jar" m:classifier="tests"/>
-    </dependency>
-    <dependency org="log4j"
-      name="log4j"
-      rev="${log4j.version}"
-      conf="common->master">
-      <exclude org="com.sun.jdmk"/>
-      <exclude org="com.sun.jmx"/>
-      <exclude org="javax.jms"/> 
-    </dependency>
-    
-    <dependency org="commons-logging"
-      name="commons-logging"
-      rev="${commons-logging.version}"
-      conf="common->master"/>
-    <dependency org="junit"
-      name="junit"
-      rev="${junit.version}"
-      conf="common->master"/>
-  </dependencies>
-</ivy-module>

+ 0 - 5
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/ivy/libraries.properties

@@ -1,5 +0,0 @@
-#This properties file lists the versions of the various artifacts used by streaming.
-#It drives ivy and the generation of a maven POM
-
-#Please list the dependencies name with version if they are different from the ones 
-#listed in the global libraries.properties file (in alphabetical order)

+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/CMakeLists.txt → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/CMakeLists.txt


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/README → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/doc/README


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_connect.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_connect.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_connect.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_connect.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_context_handle.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_context_handle.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_dfs.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_dfs.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_dfs_wrapper.sh → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_dfs_wrapper.sh


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_file_handle.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_file_handle.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_access.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_access.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_chmod.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_chmod.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_chown.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_chown.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_create.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_create.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_flush.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_flush.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_getattr.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_getattr.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_mkdir.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_mkdir.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_mknod.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_mknod.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_open.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_open.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_read.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_read.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_readdir.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_readdir.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_release.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_release.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_rename.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_rename.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_rmdir.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_rmdir.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_statfs.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_statfs.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_symlink.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_symlink.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_truncate.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_truncate.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_unlink.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_unlink.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_utimens.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_utimens.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_impls_write.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_impls_write.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_init.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_init.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_init.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_init.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_options.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_options.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_options.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_options.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_stat_struct.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_stat_struct.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_stat_struct.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_stat_struct.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_trash.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_trash.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_trash.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_trash.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_users.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_users.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/fuse_users.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/fuse_users.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/contrib/fuse-dfs/src/test/TestFuseDFS.java → hadoop-hdfs-project/hadoop-hdfs/src/main/native/fuse-dfs/test/TestFuseDFS.java


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/expect.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/expect.h


+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs.c

@@ -17,7 +17,7 @@
  */
 
 #include "hdfs.h"
-#include "hdfsJniHelper.h"
+#include "jni_helper.h"
 
 #include <stdio.h>
 #include <string.h>

+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs_test.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/hdfs_test.h


+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfsJniHelper.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/jni_helper.c

@@ -17,7 +17,7 @@
  */
 
 #include "config.h"
-#include "hdfsJniHelper.h"
+#include "jni_helper.h"
 
 #include <stdio.h> 
 #include <string.h> 

+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfsJniHelper.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/jni_helper.h


+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/native/native_mini_dfs.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.c

@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-#include "hdfsJniHelper.h"
+#include "jni_helper.h"
 #include "native_mini_dfs.h"
 
 #include <errno.h>

+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/native_mini_dfs.h → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/native_mini_dfs.h


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs_test.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_ops.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs_read.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_read.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/hdfs_write.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test/test_libhdfs_write.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/test_libhdfs_threaded.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test_libhdfs_threaded.c


+ 0 - 0
hadoop-hdfs-project/hadoop-hdfs/src/main/native/test_native_mini_dfs.c → hadoop-hdfs-project/hadoop-hdfs/src/main/native/libhdfs/test_native_mini_dfs.c