소스 검색

HADOOP-16928. Make javadoc work on Java 17 (#6976)

Contributed by Cheng Pan
Cheng Pan 9 달 전
부모
커밋
9486844610
72개의 변경된 파일317개의 추가작업 그리고 336개의 파일을 삭제
  1. 23 23
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java
  2. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java
  3. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFileSystem.java
  4. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java
  5. 16 16
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java
  6. 14 14
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java
  7. 2 2
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RemoteIterator.java
  8. 9 9
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/EnumSetWritable.java
  9. 2 2
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/ObjectWritable.java
  10. 5 5
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java
  11. 4 4
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/bzip2/CBZip2InputStream.java
  12. 41 41
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/bzip2/CBZip2OutputStream.java
  13. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/zlib/BuiltInZlibDeflater.java
  14. 2 2
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/Chunk.java
  15. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcClientException.java
  16. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcException.java
  17. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcServerException.java
  18. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/UnexpectedServerException.java
  19. 13 13
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/package-info.java
  20. 3 3
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/net/NetUtils.java
  21. 3 3
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/AccessControlException.java
  22. 3 3
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/AuthorizationException.java
  23. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java
  24. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedReadLock.java
  25. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedReadWriteLock.java
  26. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedWriteLock.java
  27. 8 8
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ShutdownThreadsHelper.java
  28. 1 1
      hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java
  29. 6 6
      hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/MiniRPCBenchmark.java
  30. 1 1
      hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/client/impl/BlockReaderLocalLegacy.java
  31. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java
  32. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EncryptionZoneManager.java
  33. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java
  34. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/DiffList.java
  35. 9 9
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FileDistributionCalculator.java
  36. 8 8
      hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FileDistributionVisitor.java
  37. 1 1
      hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeMode.java
  38. 3 3
      hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestReadOnlySharedStorage.java
  39. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/preemption/AMPreemptionPolicy.java
  40. 8 8
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java
  41. 2 2
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java
  42. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/MapRunnable.java
  43. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/jobcontrol/Job.java
  44. 6 6
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/CompositeInputFormat.java
  45. 2 2
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/CompositeRecordReader.java
  46. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/OverrideRecordReader.java
  47. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/Parser.java
  48. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TotalOrderPartitioner.java
  49. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/jobcontrol/ControlledJob.java
  50. 6 6
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/CompositeInputFormat.java
  51. 2 2
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/CompositeRecordReader.java
  52. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/OverrideRecordReader.java
  53. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/Parser.java
  54. 1 1
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java
  55. 4 4
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/output/FileOutputFormat.java
  56. 5 5
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.java
  57. 4 4
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/AccumulatingReducer.java
  58. 2 2
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/IOMapperBase.java
  59. 21 21
      hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/JHLogAnalyzer.java
  60. 37 34
      hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/pi/package.html
  61. 0 23
      hadoop-project/pom.xml
  62. 0 1
      hadoop-tools/hadoop-aws/pom.xml
  63. 1 1
      hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/FilePool.java
  64. 1 1
      hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/io/IdentifierResolver.java
  65. 1 1
      hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/package.html
  66. 5 3
      hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/typedbytes/package.html
  67. 1 1
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/SignalContainerRequest.java
  68. 1 1
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceMetricsSink.java
  69. 2 2
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/AdminACLsManager.java
  70. 1 1
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BoundedAppender.java
  71. 1 1
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/utils/LeveldbIterator.java
  72. 1 1
      hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/src/main/java/org/apache/hadoop/yarn/server/timelineservice/storage/common/BaseTable.java

+ 23 - 23
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java

@@ -119,7 +119,7 @@ import static org.apache.commons.lang3.StringUtils.isNotBlank;
 /**
 /**
  * Provides access to configuration parameters.
  * Provides access to configuration parameters.
  *
  *
- * <h3 id="Resources">Resources</h3>
+ * <h2 id="Resources">Resources</h2>
  *
  *
  * <p>Configurations are specified by resources. A resource contains a set of
  * <p>Configurations are specified by resources. A resource contains a set of
  * name/value pairs as XML data. Each resource is named by either a 
  * name/value pairs as XML data. Each resource is named by either a 
@@ -130,16 +130,16 @@ import static org.apache.commons.lang3.StringUtils.isNotBlank;
  *
  *
  * <p>Unless explicitly turned off, Hadoop by default specifies two 
  * <p>Unless explicitly turned off, Hadoop by default specifies two 
  * resources, loaded in-order from the classpath: <ol>
  * resources, loaded in-order from the classpath: <ol>
- * <li><tt>
+ * <li><code>
  * <a href="{@docRoot}/../hadoop-project-dist/hadoop-common/core-default.xml">
  * <a href="{@docRoot}/../hadoop-project-dist/hadoop-common/core-default.xml">
- * core-default.xml</a></tt>: Read-only defaults for hadoop.</li>
- * <li><tt>core-site.xml</tt>: Site-specific configuration for a given hadoop
+ * core-default.xml</a></code>: Read-only defaults for hadoop.</li>
+ * <li><code>core-site.xml</code>: Site-specific configuration for a given hadoop
  * installation.</li>
  * installation.</li>
  * </ol>
  * </ol>
  * Applications may add additional resources, which are loaded
  * Applications may add additional resources, which are loaded
  * subsequent to these resources in the order they are added.
  * subsequent to these resources in the order they are added.
  * 
  * 
- * <h4 id="FinalParams">Final Parameters</h4>
+ * <h3 id="FinalParams">Final Parameters</h3>
  *
  *
  * <p>Configuration parameters may be declared <i>final</i>. 
  * <p>Configuration parameters may be declared <i>final</i>. 
  * Once a resource declares a value final, no subsequently-loaded 
  * Once a resource declares a value final, no subsequently-loaded 
@@ -153,9 +153,9 @@ import static org.apache.commons.lang3.StringUtils.isNotBlank;
  *  &lt;/property&gt;</code></pre>
  *  &lt;/property&gt;</code></pre>
  *
  *
  * Administrators typically define parameters as final in 
  * Administrators typically define parameters as final in 
- * <tt>core-site.xml</tt> for values that user applications may not alter.
+ * <code>core-site.xml</code> for values that user applications may not alter.
  *
  *
- * <h4 id="VariableExpansion">Variable Expansion</h4>
+ * <h3 id="VariableExpansion">Variable Expansion</h3>
  *
  *
  * <p>Value strings are first processed for <i>variable expansion</i>. The
  * <p>Value strings are first processed for <i>variable expansion</i>. The
  * available properties are:<ol>
  * available properties are:<ol>
@@ -185,22 +185,22 @@ import static org.apache.commons.lang3.StringUtils.isNotBlank;
  *  &lt;/property&gt;
  *  &lt;/property&gt;
  *  </code></pre>
  *  </code></pre>
  *
  *
- * <p>When <tt>conf.get("tempdir")</tt> is called, then <tt>${<i>basedir</i>}</tt>
+ * <p>When <code>conf.get("tempdir")</code> is called, then <code>${<i>basedir</i>}</code>
  * will be resolved to another property in this Configuration, while
  * will be resolved to another property in this Configuration, while
- * <tt>${<i>user.name</i>}</tt> would then ordinarily be resolved to the value
+ * <code>${<i>user.name</i>}</code> would then ordinarily be resolved to the value
  * of the System property with that name.
  * of the System property with that name.
- * <p>When <tt>conf.get("otherdir")</tt> is called, then <tt>${<i>env.BASE_DIR</i>}</tt>
- * will be resolved to the value of the <tt>${<i>BASE_DIR</i>}</tt> environment variable.
- * It supports <tt>${<i>env.NAME:-default</i>}</tt> and <tt>${<i>env.NAME-default</i>}</tt> notations.
- * The former is resolved to "default" if <tt>${<i>NAME</i>}</tt> environment variable is undefined
+ * <p>When <code>conf.get("otherdir")</code> is called, then <code>${<i>env.BASE_DIR</i>}</code>
+ * will be resolved to the value of the <code>${<i>BASE_DIR</i>}</code> environment variable.
+ * It supports <code>${<i>env.NAME:-default</i>}</code> and <code>${<i>env.NAME-default</i>}</code> notations.
+ * The former is resolved to "default" if <code>${<i>NAME</i>}</code> environment variable is undefined
  * or its value is empty.
  * or its value is empty.
- * The latter behaves the same way only if <tt>${<i>NAME</i>}</tt> is undefined.
+ * The latter behaves the same way only if <code>${<i>NAME</i>}</code> is undefined.
  * <p>By default, warnings will be given to any deprecated configuration 
  * <p>By default, warnings will be given to any deprecated configuration 
  * parameters and these are suppressible by configuring
  * parameters and these are suppressible by configuring
- * <tt>log4j.logger.org.apache.hadoop.conf.Configuration.deprecation</tt> in
+ * <code>log4j.logger.org.apache.hadoop.conf.Configuration.deprecation</code> in
  * log4j.properties file.
  * log4j.properties file.
  *
  *
- * <h4 id="Tags">Tags</h4>
+ * <h3 id="Tags">Tags</h3>
  *
  *
  * <p>Optionally we can tag related properties together by using tag
  * <p>Optionally we can tag related properties together by using tag
  * attributes. System tags are defined by hadoop.tags.system property. Users
  * attributes. System tags are defined by hadoop.tags.system property. Users
@@ -220,9 +220,9 @@ import static org.apache.commons.lang3.StringUtils.isNotBlank;
  *    &lt;tag&gt;HDFS,SECURITY&lt;/tag&gt;
  *    &lt;tag&gt;HDFS,SECURITY&lt;/tag&gt;
  *  &lt;/property&gt;
  *  &lt;/property&gt;
  * </code></pre>
  * </code></pre>
- * <p> Properties marked with tags can be retrieved with <tt>conf
- * .getAllPropertiesByTag("HDFS")</tt> or <tt>conf.getAllPropertiesByTags
- * (Arrays.asList("YARN","SECURITY"))</tt>.</p>
+ * <p> Properties marked with tags can be retrieved with <code>conf
+ * .getAllPropertiesByTag("HDFS")</code> or <code>conf.getAllPropertiesByTags
+ * (Arrays.asList("YARN","SECURITY"))</code>.</p>
  */
  */
 @InterfaceAudience.Public
 @InterfaceAudience.Public
 @InterfaceStability.Stable
 @InterfaceStability.Stable
@@ -576,7 +576,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
    * It does not override any existing entries in the deprecation map.
    * It does not override any existing entries in the deprecation map.
    * This is to be used only by the developers in order to add deprecation of
    * This is to be used only by the developers in order to add deprecation of
    * keys, and attempts to call this method after loading resources once,
    * keys, and attempts to call this method after loading resources once,
-   * would lead to <tt>UnsupportedOperationException</tt>
+   * would lead to <code>UnsupportedOperationException</code>
    * 
    * 
    * If a key is deprecated in favor of multiple keys, they are all treated as 
    * If a key is deprecated in favor of multiple keys, they are all treated as 
    * aliases of each other, and setting any one of them resets all the others 
    * aliases of each other, and setting any one of them resets all the others 
@@ -604,7 +604,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
    * It does not override any existing entries in the deprecation map.
    * It does not override any existing entries in the deprecation map.
    * This is to be used only by the developers in order to add deprecation of
    * This is to be used only by the developers in order to add deprecation of
    * keys, and attempts to call this method after loading resources once,
    * keys, and attempts to call this method after loading resources once,
-   * would lead to <tt>UnsupportedOperationException</tt>
+   * would lead to <code>UnsupportedOperationException</code>
    * 
    * 
    * If you have multiple deprecation entries to add, it is more efficient to
    * If you have multiple deprecation entries to add, it is more efficient to
    * use #addDeprecations(DeprecationDelta[] deltas) instead.
    * use #addDeprecations(DeprecationDelta[] deltas) instead.
@@ -624,7 +624,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
    * It does not override any existing entries in the deprecation map.
    * It does not override any existing entries in the deprecation map.
    * This is to be used only by the developers in order to add deprecation of
    * This is to be used only by the developers in order to add deprecation of
    * keys, and attempts to call this method after loading resources once,
    * keys, and attempts to call this method after loading resources once,
-   * would lead to <tt>UnsupportedOperationException</tt>
+   * would lead to <code>UnsupportedOperationException</code>
    * 
    * 
    * If a key is deprecated in favor of multiple keys, they are all treated as 
    * If a key is deprecated in favor of multiple keys, they are all treated as 
    * aliases of each other, and setting any one of them resets all the others 
    * aliases of each other, and setting any one of them resets all the others 
@@ -648,7 +648,7 @@ public class Configuration implements Iterable<Map.Entry<String,String>>,
    * It does not override any existing entries in the deprecation map.
    * It does not override any existing entries in the deprecation map.
    * This is to be used only by the developers in order to add deprecation of
    * This is to be used only by the developers in order to add deprecation of
    * keys, and attempts to call this method after loading resources once,
    * keys, and attempts to call this method after loading resources once,
-   * would lead to <tt>UnsupportedOperationException</tt>
+   * would lead to <code>UnsupportedOperationException</code>
    * 
    * 
    * If you have multiple deprecation entries to add, it is more efficient to
    * If you have multiple deprecation entries to add, it is more efficient to
    * use #addDeprecations(DeprecationDelta[] deltas) instead.
    * use #addDeprecations(DeprecationDelta[] deltas) instead.

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/AbstractFileSystem.java

@@ -247,7 +247,7 @@ public abstract class AbstractFileSystem implements PathCapabilities {
    * The main factory method for creating a file system. Get a file system for
    * The main factory method for creating a file system. Get a file system for
    * the URI's scheme and authority. The scheme of the <code>uri</code>
    * the URI's scheme and authority. The scheme of the <code>uri</code>
    * determines a configuration property name,
    * determines a configuration property name,
-   * <tt>fs.AbstractFileSystem.<i>scheme</i>.impl</tt> whose value names the
+   * <code>fs.AbstractFileSystem.<i>scheme</i>.impl</code> whose value names the
    * AbstractFileSystem class.
    * AbstractFileSystem class.
    * 
    * 
    * The entire URI and conf is passed to the AbstractFileSystem factory method.
    * The entire URI and conf is passed to the AbstractFileSystem factory method.

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFileSystem.java

@@ -866,7 +866,7 @@ public abstract class ChecksumFileSystem extends FilterFileSystem {
 
 
   /**
   /**
    * Set replication for an existing file.
    * Set replication for an existing file.
-   * Implement the abstract <tt>setReplication</tt> of <tt>FileSystem</tt>
+   * Implement the abstract <code>setReplication</code> of <code>FileSystem</code>
    * @param src file name
    * @param src file name
    * @param replication new replication
    * @param replication new replication
    * @throws IOException if an I/O error occurs.
    * @throws IOException if an I/O error occurs.

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ChecksumFs.java

@@ -453,7 +453,7 @@ public abstract class ChecksumFs extends FilterFs {
   }
   }
   /**
   /**
    * Set replication for an existing file.
    * Set replication for an existing file.
-   * Implement the abstract <tt>setReplication</tt> of <tt>FileSystem</tt>
+   * Implement the abstract <code>setReplication</code> of <code>FileSystem</code>
    * @param src file name
    * @param src file name
    * @param replication new replication
    * @param replication new replication
    * @throws IOException if an I/O error occurs.
    * @throws IOException if an I/O error occurs.

+ 16 - 16
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java

@@ -1977,9 +1977,9 @@ public class FileContext implements PathCapabilities {
         LocatedFileStatus curFile;
         LocatedFileStatus curFile;
 
 
         /**
         /**
-         * Returns <tt>true</tt> if the iterator has more files.
+         * Returns <code>true</code> if the iterator has more files.
          *
          *
-         * @return <tt>true</tt> if the iterator has more files.
+         * @return <code>true</code> if the iterator has more files.
          * @throws AccessControlException if not allowed to access next
          * @throws AccessControlException if not allowed to access next
          *                                file's status or locations
          *                                file's status or locations
          * @throws FileNotFoundException if next file does not exist any more
          * @throws FileNotFoundException if next file does not exist any more
@@ -2071,34 +2071,34 @@ public class FileContext implements PathCapabilities {
      * <dl>
      * <dl>
      *  <dd>
      *  <dd>
      *   <dl>
      *   <dl>
-     *    <dt> <tt> ? </tt>
+     *    <dt> <code> ? </code>
      *    <dd> Matches any single character.
      *    <dd> Matches any single character.
      *
      *
-     *    <dt> <tt> * </tt>
+     *    <dt> <code> * </code>
      *    <dd> Matches zero or more characters.
      *    <dd> Matches zero or more characters.
      *
      *
-     *    <dt> <tt> [<i>abc</i>] </tt>
+     *    <dt> <code> [<i>abc</i>] </code>
      *    <dd> Matches a single character from character set
      *    <dd> Matches a single character from character set
-     *     <tt>{<i>a,b,c</i>}</tt>.
+     *     <code>{<i>a,b,c</i>}</code>.
      *
      *
-     *    <dt> <tt> [<i>a</i>-<i>b</i>] </tt>
+     *    <dt> <code> [<i>a</i>-<i>b</i>] </code>
      *    <dd> Matches a single character from the character range
      *    <dd> Matches a single character from the character range
-     *     <tt>{<i>a...b</i>}</tt>. Note: character <tt><i>a</i></tt> must be
-     *     lexicographically less than or equal to character <tt><i>b</i></tt>.
+     *     <code>{<i>a...b</i>}</code>. Note: character <code><i>a</i></code> must be
+     *     lexicographically less than or equal to character <code><i>b</i></code>.
      *
      *
-     *    <dt> <tt> [^<i>a</i>] </tt>
+     *    <dt> <code> [^<i>a</i>] </code>
      *    <dd> Matches a single char that is not from character set or range
      *    <dd> Matches a single char that is not from character set or range
-     *     <tt>{<i>a</i>}</tt>.  Note that the <tt>^</tt> character must occur
+     *     <code>{<i>a</i>}</code>.  Note that the <code>^</code> character must occur
      *     immediately to the right of the opening bracket.
      *     immediately to the right of the opening bracket.
      *
      *
-     *    <dt> <tt> \<i>c</i> </tt>
+     *    <dt> <code> \<i>c</i> </code>
      *    <dd> Removes (escapes) any special meaning of character <i>c</i>.
      *    <dd> Removes (escapes) any special meaning of character <i>c</i>.
      *
      *
-     *    <dt> <tt> {ab,cd} </tt>
-     *    <dd> Matches a string from the string set <tt>{<i>ab, cd</i>} </tt>
+     *    <dt> <code> {ab,cd} </code>
+     *    <dd> Matches a string from the string set <code>{<i>ab, cd</i>} </code>
      *
      *
-     *    <dt> <tt> {ab,c{de,fh}} </tt>
-     *    <dd> Matches a string from string set <tt>{<i>ab, cde, cfh</i>}</tt>
+     *    <dt> <code> {ab,c{de,fh}} </code>
+     *    <dd> Matches a string from string set <code>{<i>ab, cde, cfh</i>}</code>
      *
      *
      *   </dl>
      *   </dl>
      *  </dd>
      *  </dd>

+ 14 - 14
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java

@@ -2178,34 +2178,34 @@ public abstract class FileSystem extends Configured
    * <dl>
    * <dl>
    *  <dd>
    *  <dd>
    *   <dl>
    *   <dl>
-   *    <dt> <tt> ? </tt>
+   *    <dt> <code> ? </code>
    *    <dd> Matches any single character.
    *    <dd> Matches any single character.
    *
    *
-   *    <dt> <tt> * </tt>
+   *    <dt> <code> * </code>
    *    <dd> Matches zero or more characters.
    *    <dd> Matches zero or more characters.
    *
    *
-   *    <dt> <tt> [<i>abc</i>] </tt>
+   *    <dt> <code> [<i>abc</i>] </code>
    *    <dd> Matches a single character from character set
    *    <dd> Matches a single character from character set
-   *     <tt>{<i>a,b,c</i>}</tt>.
+   *     <code>{<i>a,b,c</i>}</code>.
    *
    *
-   *    <dt> <tt> [<i>a</i>-<i>b</i>] </tt>
+   *    <dt> <code> [<i>a</i>-<i>b</i>] </code>
    *    <dd> Matches a single character from the character range
    *    <dd> Matches a single character from the character range
-   *     <tt>{<i>a...b</i>}</tt>.  Note that character <tt><i>a</i></tt> must be
-   *     lexicographically less than or equal to character <tt><i>b</i></tt>.
+   *     <code>{<i>a...b</i>}</code>.  Note that character <code><i>a</i></code> must be
+   *     lexicographically less than or equal to character <code><i>b</i></code>.
    *
    *
-   *    <dt> <tt> [^<i>a</i>] </tt>
+   *    <dt> <code> [^<i>a</i>] </code>
    *    <dd> Matches a single character that is not from character set or range
    *    <dd> Matches a single character that is not from character set or range
-   *     <tt>{<i>a</i>}</tt>.  Note that the <tt>^</tt> character must occur
+   *     <code>{<i>a</i>}</code>.  Note that the <code>^</code> character must occur
    *     immediately to the right of the opening bracket.
    *     immediately to the right of the opening bracket.
    *
    *
-   *    <dt> <tt> \<i>c</i> </tt>
+   *    <dt> <code> \<i>c</i> </code>
    *    <dd> Removes (escapes) any special meaning of character <i>c</i>.
    *    <dd> Removes (escapes) any special meaning of character <i>c</i>.
    *
    *
-   *    <dt> <tt> {ab,cd} </tt>
-   *    <dd> Matches a string from the string set <tt>{<i>ab, cd</i>} </tt>
+   *    <dt> <code> {ab,cd} </code>
+   *    <dd> Matches a string from the string set <code>{<i>ab, cd</i>} </code>
    *
    *
-   *    <dt> <tt> {ab,c{de,fh}} </tt>
-   *    <dd> Matches a string from the string set <tt>{<i>ab, cde, cfh</i>}</tt>
+   *    <dt> <code> {ab,c{de,fh}} </code>
+   *    <dd> Matches a string from the string set <code>{<i>ab, cde, cfh</i>}</code>
    *
    *
    *   </dl>
    *   </dl>
    *  </dd>
    *  </dd>

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/RemoteIterator.java

@@ -24,9 +24,9 @@ import java.util.NoSuchElementException;
  */
  */
 public interface RemoteIterator<E> {
 public interface RemoteIterator<E> {
   /**
   /**
-   * Returns <tt>true</tt> if the iteration has more elements.
+   * Returns <code>true</code> if the iteration has more elements.
    *
    *
-   * @return <tt>true</tt> if the iterator has more elements.
+   * @return <code>true</code> if the iterator has more elements.
    * @throws IOException if any IO error occurs
    * @throws IOException if any IO error occurs
    */
    */
   boolean hasNext() throws IOException;
   boolean hasNext() throws IOException;

+ 9 - 9
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/EnumSetWritable.java

@@ -59,10 +59,10 @@ public class EnumSetWritable<E extends Enum<E>> extends AbstractCollection<E>
   }
   }
 
 
   /**
   /**
-   * Construct a new EnumSetWritable. If the <tt>value</tt> argument is null or
-   * its size is zero, the <tt>elementType</tt> argument must not be null. If
-   * the argument <tt>value</tt>'s size is bigger than zero, the argument
-   * <tt>elementType</tt> is not be used.
+   * Construct a new EnumSetWritable. If the <code>value</code> argument is null or
+   * its size is zero, the <code>elementType</code> argument must not be null. If
+   * the argument <code>value</code>'s size is bigger than zero, the argument
+   * <code>elementType</code> is not be used.
    * 
    * 
    * @param value enumSet value.
    * @param value enumSet value.
    * @param elementType elementType.
    * @param elementType elementType.
@@ -72,7 +72,7 @@ public class EnumSetWritable<E extends Enum<E>> extends AbstractCollection<E>
   }
   }
 
 
   /**
   /**
-   * Construct a new EnumSetWritable. Argument <tt>value</tt> should not be null
+   * Construct a new EnumSetWritable. Argument <code>value</code> should not be null
    * or empty.
    * or empty.
    * 
    * 
    * @param value enumSet value.
    * @param value enumSet value.
@@ -83,10 +83,10 @@ public class EnumSetWritable<E extends Enum<E>> extends AbstractCollection<E>
 
 
   /**
   /**
    * reset the EnumSetWritable with specified
    * reset the EnumSetWritable with specified
-   * <tt>value</tt> and <tt>elementType</tt>. If the <tt>value</tt> argument
-   * is null or its size is zero, the <tt>elementType</tt> argument must not be
-   * null. If the argument <tt>value</tt>'s size is bigger than zero, the
-   * argument <tt>elementType</tt> is not be used.
+   * <code>value</code> and <code>elementType</code>. If the <code>value</code> argument
+   * is null or its size is zero, the <code>elementType</code> argument must not be
+   * null. If the argument <code>value</code>'s size is bigger than zero, the
+   * argument <code>elementType</code> is not be used.
    * 
    * 
    * @param value enumSet Value.
    * @param value enumSet Value.
    * @param elementType elementType.
    * @param elementType elementType.

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/ObjectWritable.java

@@ -401,8 +401,8 @@ public class ObjectWritable implements Writable, Configurable {
   }
   }
 
 
   /**
   /**
-   * Find and load the class with given name <tt>className</tt> by first finding
-   * it in the specified <tt>conf</tt>. If the specified <tt>conf</tt> is null,
+   * Find and load the class with given name <code>className</code> by first finding
+   * it in the specified <code>conf</code>. If the specified <code>conf</code> is null,
    * try load it directly.
    * try load it directly.
    *
    *
    * @param conf configuration.
    * @param conf configuration.

+ 5 - 5
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/SequenceFile.java

@@ -91,19 +91,19 @@ import static org.apache.hadoop.util.functional.FutureIO.awaitFuture;
  * <p>The actual compression algorithm used to compress key and/or values can be
  * <p>The actual compression algorithm used to compress key and/or values can be
  * specified by using the appropriate {@link CompressionCodec}.</p>
  * specified by using the appropriate {@link CompressionCodec}.</p>
  * 
  * 
- * <p>The recommended way is to use the static <tt>createWriter</tt> methods
+ * <p>The recommended way is to use the static <code>createWriter</code> methods
  * provided by the <code>SequenceFile</code> to chose the preferred format.</p>
  * provided by the <code>SequenceFile</code> to chose the preferred format.</p>
  *
  *
  * <p>The {@link SequenceFile.Reader} acts as the bridge and can read any of the
  * <p>The {@link SequenceFile.Reader} acts as the bridge and can read any of the
  * above <code>SequenceFile</code> formats.</p>
  * above <code>SequenceFile</code> formats.</p>
  *
  *
- * <h3 id="Formats">SequenceFile Formats</h3>
+ * <h2 id="Formats">SequenceFile Formats</h2>
  * 
  * 
  * <p>Essentially there are 3 different formats for <code>SequenceFile</code>s
  * <p>Essentially there are 3 different formats for <code>SequenceFile</code>s
  * depending on the <code>CompressionType</code> specified. All of them share a
  * depending on the <code>CompressionType</code> specified. All of them share a
  * <a href="#Header">common header</a> described below.
  * <a href="#Header">common header</a> described below.
  * 
  * 
- * <h4 id="Header">SequenceFile Header</h4>
+ * <h3 id="Header">SequenceFile Header</h3>
  * <ul>
  * <ul>
  *   <li>
  *   <li>
  *   version - 3 bytes of magic header <b>SEQ</b>, followed by 1 byte of actual 
  *   version - 3 bytes of magic header <b>SEQ</b>, followed by 1 byte of actual 
@@ -136,7 +136,7 @@ import static org.apache.hadoop.util.functional.FutureIO.awaitFuture;
  *   </li>
  *   </li>
  * </ul>
  * </ul>
  * 
  * 
- * <h5>Uncompressed SequenceFile Format</h5>
+ * <h4>Uncompressed SequenceFile Format</h4>
  * <ul>
  * <ul>
  * <li>
  * <li>
  * <a href="#Header">Header</a>
  * <a href="#Header">Header</a>
@@ -155,7 +155,7 @@ import static org.apache.hadoop.util.functional.FutureIO.awaitFuture;
  * </li>
  * </li>
  * </ul>
  * </ul>
  *
  *
- * <h5>Record-Compressed SequenceFile Format</h5>
+ * <h4>Record-Compressed SequenceFile Format</h4>
  * <ul>
  * <ul>
  * <li>
  * <li>
  * <a href="#Header">Header</a>
  * <a href="#Header">Header</a>

+ 4 - 4
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/bzip2/CBZip2InputStream.java

@@ -38,13 +38,13 @@ import org.apache.hadoop.io.compress.SplittableCompressionCodec.READ_MODE;
  * <p>
  * <p>
  * The decompression requires large amounts of memory. Thus you should call the
  * The decompression requires large amounts of memory. Thus you should call the
  * {@link #close() close()} method as soon as possible, to force
  * {@link #close() close()} method as soon as possible, to force
- * <tt>CBZip2InputStream</tt> to release the allocated memory. See
+ * <code>CBZip2InputStream</code> to release the allocated memory. See
  * {@link CBZip2OutputStream CBZip2OutputStream} for information about memory
  * {@link CBZip2OutputStream CBZip2OutputStream} for information about memory
  * usage.
  * usage.
  * </p>
  * </p>
  *
  *
  * <p>
  * <p>
- * <tt>CBZip2InputStream</tt> reads bytes from the compressed source stream via
+ * <code>CBZip2InputStream</code> reads bytes from the compressed source stream via
  * the single byte {@link java.io.InputStream#read() read()} method exclusively.
  * the single byte {@link java.io.InputStream#read() read()} method exclusively.
  * Thus you should consider to use a buffered source stream.
  * Thus you should consider to use a buffered source stream.
  * </p>
  * </p>
@@ -279,7 +279,7 @@ public class CBZip2InputStream extends InputStream implements BZip2Constants {
   * specified stream.
   * specified stream.
   *
   *
   * <p>
   * <p>
-  * Although BZip2 headers are marked with the magic <tt>"Bz"</tt> this
+  * Although BZip2 headers are marked with the magic <code>"Bz"</code> this
   * constructor expects the next byte in the stream to be the first one after
   * constructor expects the next byte in the stream to be the first one after
   * the magic. Thus callers have to skip the first two bytes. Otherwise this
   * the magic. Thus callers have to skip the first two bytes. Otherwise this
   * constructor will throw an exception.
   * constructor will throw an exception.
@@ -289,7 +289,7 @@ public class CBZip2InputStream extends InputStream implements BZip2Constants {
   * @throws IOException
   * @throws IOException
   *             if the stream content is malformed or an I/O error occurs.
   *             if the stream content is malformed or an I/O error occurs.
   * @throws NullPointerException
   * @throws NullPointerException
-  *             if <tt>in == null</tt>
+  *             if <code>in == null</code>
   */
   */
   public CBZip2InputStream(final InputStream in, READ_MODE readMode)
   public CBZip2InputStream(final InputStream in, READ_MODE readMode)
       throws IOException {
       throws IOException {

+ 41 - 41
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/bzip2/CBZip2OutputStream.java

@@ -37,7 +37,7 @@ import org.apache.hadoop.io.IOUtils;
  * <p>
  * <p>
  * The compression requires large amounts of memory. Thus you should call the
  * The compression requires large amounts of memory. Thus you should call the
  * {@link #close() close()} method as soon as possible, to force
  * {@link #close() close()} method as soon as possible, to force
- * <tt>CBZip2OutputStream</tt> to release the allocated memory.
+ * <code>CBZip2OutputStream</code> to release the allocated memory.
  * </p>
  * </p>
  *
  *
  * <p>
  * <p>
@@ -64,64 +64,64 @@ import org.apache.hadoop.io.IOUtils;
  * &lt;code&gt;65k + (5 * blocksize)&lt;/code&gt;.
  * &lt;code&gt;65k + (5 * blocksize)&lt;/code&gt;.
  * </pre>
  * </pre>
  *
  *
- * <table width="100%" border="1">
+ * <table border="1">
  * <caption>Memory usage by blocksize</caption>
  * <caption>Memory usage by blocksize</caption>
- * <colgroup> <col width="33%" > <col width="33%" > <col width="33%" >
+ * <colgroup> <col> <col> <col>
  * </colgroup>
  * </colgroup>
  * <tr>
  * <tr>
- * <th align="right">Blocksize</th> <th align="right">Compression<br>
- * memory usage</th> <th align="right">Decompression<br>
+ * <th>Blocksize</th> <th>Compression<br>
+ * memory usage</th> <th>Decompression<br>
  * memory usage</th>
  * memory usage</th>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">100k</td>
- * <td align="right">1300k</td>
- * <td align="right">565k</td>
+ * <td>100k</td>
+ * <td>1300k</td>
+ * <td>565k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">200k</td>
- * <td align="right">2200k</td>
- * <td align="right">1065k</td>
+ * <td>200k</td>
+ * <td>2200k</td>
+ * <td>1065k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">300k</td>
- * <td align="right">3100k</td>
- * <td align="right">1565k</td>
+ * <td>300k</td>
+ * <td>3100k</td>
+ * <td>1565k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">400k</td>
- * <td align="right">4000k</td>
- * <td align="right">2065k</td>
+ * <td>400k</td>
+ * <td>4000k</td>
+ * <td>2065k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">500k</td>
- * <td align="right">4900k</td>
- * <td align="right">2565k</td>
+ * <td>500k</td>
+ * <td>4900k</td>
+ * <td>2565k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">600k</td>
- * <td align="right">5800k</td>
- * <td align="right">3065k</td>
+ * <td>600k</td>
+ * <td>5800k</td>
+ * <td>3065k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">700k</td>
- * <td align="right">6700k</td>
- * <td align="right">3565k</td>
+ * <td>700k</td>
+ * <td>6700k</td>
+ * <td>3565k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">800k</td>
- * <td align="right">7600k</td>
- * <td align="right">4065k</td>
+ * <td>800k</td>
+ * <td>7600k</td>
+ * <td>4065k</td>
  * </tr>
  * </tr>
  * <tr>
  * <tr>
- * <td align="right">900k</td>
- * <td align="right">8500k</td>
- * <td align="right">4565k</td>
+ * <td>900k</td>
+ * <td>8500k</td>
+ * <td>4565k</td>
  * </tr>
  * </tr>
  * </table>
  * </table>
  *
  *
  * <p>
  * <p>
- * For decompression <tt>CBZip2InputStream</tt> allocates less memory if the
+ * For decompression <code>CBZip2InputStream</code> allocates less memory if the
  * bzipped input is smaller than one block.
  * bzipped input is smaller than one block.
  * </p>
  * </p>
  *
  *
@@ -137,12 +137,12 @@ import org.apache.hadoop.io.IOUtils;
 public class CBZip2OutputStream extends OutputStream implements BZip2Constants {
 public class CBZip2OutputStream extends OutputStream implements BZip2Constants {
 
 
   /**
   /**
-  * The minimum supported blocksize <tt> == 1</tt>.
+  * The minimum supported blocksize <code> == 1</code>.
   */
   */
   public static final int MIN_BLOCKSIZE = 1;
   public static final int MIN_BLOCKSIZE = 1;
 
 
   /**
   /**
-  * The maximum supported blocksize <tt> == 9</tt>.
+  * The maximum supported blocksize <code> == 9</code>.
   */
   */
   public static final int MAX_BLOCKSIZE = 9;
   public static final int MAX_BLOCKSIZE = 9;
 
 
@@ -566,12 +566,12 @@ public class CBZip2OutputStream extends OutputStream implements BZip2Constants {
   *
   *
   * @return The blocksize, between {@link #MIN_BLOCKSIZE} and
   * @return The blocksize, between {@link #MIN_BLOCKSIZE} and
   *         {@link #MAX_BLOCKSIZE} both inclusive. For a negative
   *         {@link #MAX_BLOCKSIZE} both inclusive. For a negative
-  *         <tt>inputLength</tt> this method returns <tt>MAX_BLOCKSIZE</tt>
+  *         <code>inputLength</code> this method returns <code>MAX_BLOCKSIZE</code>
   *         always.
   *         always.
   *
   *
   * @param inputLength
   * @param inputLength
   *            The length of the data which will be compressed by
   *            The length of the data which will be compressed by
-  *            <tt>CBZip2OutputStream</tt>.
+  *            <code>CBZip2OutputStream</code>.
   */
   */
   public static int chooseBlockSize(long inputLength) {
   public static int chooseBlockSize(long inputLength) {
     return (inputLength > 0) ? (int) Math
     return (inputLength > 0) ? (int) Math
@@ -579,11 +579,11 @@ public class CBZip2OutputStream extends OutputStream implements BZip2Constants {
   }
   }
 
 
   /**
   /**
-  * Constructs a new <tt>CBZip2OutputStream</tt> with a blocksize of 900k.
+  * Constructs a new <code>CBZip2OutputStream</code> with a blocksize of 900k.
   *
   *
   * <p>
   * <p>
   * <b>Attention: </b>The caller is resonsible to write the two BZip2 magic
   * <b>Attention: </b>The caller is resonsible to write the two BZip2 magic
-  * bytes <tt>"BZ"</tt> to the specified stream prior to calling this
+  * bytes <code>"BZ"</code> to the specified stream prior to calling this
   * constructor.
   * constructor.
   * </p>
   * </p>
   *
   *
@@ -600,11 +600,11 @@ public class CBZip2OutputStream extends OutputStream implements BZip2Constants {
   }
   }
 
 
   /**
   /**
-  * Constructs a new <tt>CBZip2OutputStream</tt> with specified blocksize.
+  * Constructs a new <code>CBZip2OutputStream</code> with specified blocksize.
   *
   *
   * <p>
   * <p>
   * <b>Attention: </b>The caller is resonsible to write the two BZip2 magic
   * <b>Attention: </b>The caller is resonsible to write the two BZip2 magic
-  * bytes <tt>"BZ"</tt> to the specified stream prior to calling this
+  * bytes <code>"BZ"</code> to the specified stream prior to calling this
   * constructor.
   * constructor.
   * </p>
   * </p>
   *
   *

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/zlib/BuiltInZlibDeflater.java

@@ -57,7 +57,7 @@ public class BuiltInZlibDeflater extends Deflater implements Compressor {
   /**
   /**
    * reinit the compressor with the given configuration. It will reset the
    * reinit the compressor with the given configuration. It will reset the
    * compressor's compression level and compression strategy. Different from
    * compressor's compression level and compression strategy. Different from
-   * <tt>ZlibCompressor</tt>, <tt>BuiltInZlibDeflater</tt> only support three
+   * <code>ZlibCompressor</code>, <code>BuiltInZlibDeflater</code> only support three
    * kind of compression strategy: FILTERED, HUFFMAN_ONLY and DEFAULT_STRATEGY.
    * kind of compression strategy: FILTERED, HUFFMAN_ONLY and DEFAULT_STRATEGY.
    * It will use DEFAULT_STRATEGY as default if the configured compression
    * It will use DEFAULT_STRATEGY as default if the configured compression
    * strategy is not supported.
    * strategy is not supported.

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/Chunk.java

@@ -219,8 +219,8 @@ final class Chunk {
 
 
     /**
     /**
      * The number of valid bytes in the buffer. This value is always in the
      * The number of valid bytes in the buffer. This value is always in the
-     * range <tt>0</tt> through <tt>buf.length</tt>; elements <tt>buf[0]</tt>
-     * through <tt>buf[count-1]</tt> contain valid byte data.
+     * range <code>0</code> through <code>buf.length</code>; elements <code>buf[0]</code>
+     * through <code>buf[count-1]</code> contain valid byte data.
      */
      */
     private int count;
     private int count;
 
 

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcClientException.java

@@ -38,7 +38,7 @@ public class RpcClientException extends RpcException {
    * @param message message.
    * @param message message.
    * @param cause that cause this exception
    * @param cause that cause this exception
    * @param cause the cause (can be retried by the {@link #getCause()} method).
    * @param cause the cause (can be retried by the {@link #getCause()} method).
-   *          (A <tt>null</tt> value is permitted, and indicates that the cause
+   *          (A <code>null</code> value is permitted, and indicates that the cause
    *          is nonexistent or unknown.)
    *          is nonexistent or unknown.)
    */
    */
   RpcClientException(final String message, final Throwable cause) {
   RpcClientException(final String message, final Throwable cause) {

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcException.java

@@ -40,7 +40,7 @@ public class RpcException extends IOException {
    * @param message message.
    * @param message message.
    * @param cause that cause this exception
    * @param cause that cause this exception
    * @param cause the cause (can be retried by the {@link #getCause()} method).
    * @param cause the cause (can be retried by the {@link #getCause()} method).
-   *          (A <tt>null</tt> value is permitted, and indicates that the cause
+   *          (A <code>null</code> value is permitted, and indicates that the cause
    *          is nonexistent or unknown.)
    *          is nonexistent or unknown.)
    */
    */
   RpcException(final String message, final Throwable cause) {
   RpcException(final String message, final Throwable cause) {

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcServerException.java

@@ -39,7 +39,7 @@ public class RpcServerException extends RpcException {
    * 
    * 
    * @param message message.
    * @param message message.
    * @param cause the cause (can be retried by the {@link #getCause()} method).
    * @param cause the cause (can be retried by the {@link #getCause()} method).
-   *          (A <tt>null</tt> value is permitted, and indicates that the cause
+   *          (A <code>null</code> value is permitted, and indicates that the cause
    *          is nonexistent or unknown.)
    *          is nonexistent or unknown.)
    */
    */
   public RpcServerException(final String message, final Throwable cause) {
   public RpcServerException(final String message, final Throwable cause) {

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/UnexpectedServerException.java

@@ -39,7 +39,7 @@ public class UnexpectedServerException extends RpcException {
    * @param message message.
    * @param message message.
    * @param cause that cause this exception
    * @param cause that cause this exception
    * @param cause the cause (can be retried by the {@link #getCause()} method).
    * @param cause the cause (can be retried by the {@link #getCause()} method).
-   *          (A <tt>null</tt> value is permitted, and indicates that the cause
+   *          (A <code>null</code> value is permitted, and indicates that the cause
    *          is nonexistent or unknown.)
    *          is nonexistent or unknown.)
    */
    */
   UnexpectedServerException(final String message, final Throwable cause) {
   UnexpectedServerException(final String message, final Throwable cause) {

+ 13 - 13
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/package-info.java

@@ -17,7 +17,7 @@
  */
  */
 
 
 /**
 /**
-  <h1>Metrics 2.0</h1>
+  <h2>Metrics 2.0</h2>
   <ul id="toc">
   <ul id="toc">
     <li><a href="#overview">Overview</a></li>
     <li><a href="#overview">Overview</a></li>
     <li><a href="#gettingstarted">Getting Started</a></li>
     <li><a href="#gettingstarted">Getting Started</a></li>
@@ -26,7 +26,7 @@
     <li><a href="#instrumentation">Metrics Instrumentation Strategy</a></li>
     <li><a href="#instrumentation">Metrics Instrumentation Strategy</a></li>
     <li><a href="#migration">Migration from previous system</a></li>
     <li><a href="#migration">Migration from previous system</a></li>
   </ul>
   </ul>
-  <h2><a name="overview">Overview</a></h2>
+  <h3><a>Overview</a></h3>
   <p>This package provides a framework for metrics instrumentation
   <p>This package provides a framework for metrics instrumentation
     and publication.
     and publication.
   </p>
   </p>
@@ -46,7 +46,7 @@
     <a href="http://wiki.apache.org/hadoop/HADOOP-6728-MetricsV2">design
     <a href="http://wiki.apache.org/hadoop/HADOOP-6728-MetricsV2">design
     document</a> for architecture and implementation notes.
     document</a> for architecture and implementation notes.
   </p>
   </p>
-  <h3>Sub-packages</h3>
+  <h4>Sub-packages</h4>
   <dl>
   <dl>
     <dt><code>org.apache.hadoop.metrics2.annotation</code></dt>
     <dt><code>org.apache.hadoop.metrics2.annotation</code></dt>
     <dd>Public annotation interfaces for simpler metrics instrumentation.
     <dd>Public annotation interfaces for simpler metrics instrumentation.
@@ -84,9 +84,9 @@
     </dd>
     </dd>
   </dl>
   </dl>
 
 
-  <h2><a name="gettingstarted">Getting started</a></h2>
-  <h3>Implementing metrics sources</h3>
-  <table width="99%" border="1" cellspacing="0" cellpadding="4">
+  <h3><a>Getting started</a></h3>
+  <h4>Implementing metrics sources</h4>
+  <table border="1">
     <caption>Implementing metrics sources</caption>
     <caption>Implementing metrics sources</caption>
     <tbody>
     <tbody>
       <tr>
       <tr>
@@ -153,7 +153,7 @@
     allowing generated metrics names and multiple records. In fact, the
     allowing generated metrics names and multiple records. In fact, the
     annotation interface is implemented with the MetricsSource interface
     annotation interface is implemented with the MetricsSource interface
     internally.</p>
     internally.</p>
-  <h3>Implementing metrics sinks</h3>
+  <h4>Implementing metrics sinks</h4>
   <pre>
   <pre>
   public class MySink implements MetricsSink {
   public class MySink implements MetricsSink {
     public void putMetrics(MetricsRecord record) {
     public void putMetrics(MetricsRecord record) {
@@ -187,7 +187,7 @@
   <pre>
   <pre>
   DefaultMetricsSystem.initialize("test"); // called once per application
   DefaultMetricsSystem.initialize("test"); // called once per application
   DefaultMetricsSystem.register(new MyStat());</pre>
   DefaultMetricsSystem.register(new MyStat());</pre>
-  <h2><a name="config">Metrics system configuration</a></h2>
+  <h2><a>Metrics system configuration</a></h2>
   <p>Sinks are usually specified in a configuration file, say,
   <p>Sinks are usually specified in a configuration file, say,
   "hadoop-metrics2-test.properties", as:
   "hadoop-metrics2-test.properties", as:
   </p>
   </p>
@@ -209,7 +209,7 @@
     for more examples.
     for more examples.
   </p>
   </p>
 
 
-  <h2><a name="filtering">Metrics Filtering</a></h2>
+  <h3><a>Metrics Filtering</a></h3>
   <p>One of the features of the default metrics system is metrics filtering
   <p>One of the features of the default metrics system is metrics filtering
     configuration by source, context, record/tags and metrics. The least
     configuration by source, context, record/tags and metrics. The least
     expensive way to filter out metrics would be at the source level, e.g.,
     expensive way to filter out metrics would be at the source level, e.g.,
@@ -241,7 +241,7 @@
     level, respectively. Filters can be combined to optimize
     level, respectively. Filters can be combined to optimize
     the filtering efficiency.</p>
     the filtering efficiency.</p>
 
 
-  <h2><a name="instrumentation">Metrics instrumentation strategy</a></h2>
+  <h3><a>Metrics instrumentation strategy</a></h3>
 
 
   In previous examples, we showed a minimal example to use the
   In previous examples, we showed a minimal example to use the
   metrics framework. In a larger system (like Hadoop) that allows
   metrics framework. In a larger system (like Hadoop) that allows
@@ -279,7 +279,7 @@
     </dd>
     </dd>
   </dl>
   </dl>
 
 
-  <h2><a name="migration">Migration from previous system</a></h2>
+  <h3><a>Migration from previous system</a></h3>
   <p>Users of the previous metrics system would notice the lack of
   <p>Users of the previous metrics system would notice the lack of
     <code>context</code> prefix in the configuration examples. The new
     <code>context</code> prefix in the configuration examples. The new
     metrics system decouples the concept for context (for grouping) with the
     metrics system decouples the concept for context (for grouping) with the
@@ -289,7 +289,7 @@
     configure an implementation instance per context, even if you have a
     configure an implementation instance per context, even if you have a
     backend that can handle multiple contexts (file, gangalia etc.):
     backend that can handle multiple contexts (file, gangalia etc.):
   </p>
   </p>
-  <table width="99%" border="1" cellspacing="0" cellpadding="4">
+  <table border="1">
     <caption>Migration from previous system</caption>
     <caption>Migration from previous system</caption>
     <tbody>
     <tbody>
       <tr>
       <tr>
@@ -311,7 +311,7 @@
   <p>In the new metrics system, you can simulate the previous behavior by
   <p>In the new metrics system, you can simulate the previous behavior by
     using the context option in the sink options like the following:
     using the context option in the sink options like the following:
   </p>
   </p>
-  <table width="99%" border="1" cellspacing="0" cellpadding="4">
+  <table border="1">
     <caption>Metrics2</caption>
     <caption>Metrics2</caption>
     <tbody>
     <tbody>
       <tr>
       <tr>

+ 3 - 3
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/net/NetUtils.java

@@ -83,9 +83,9 @@ public class NetUtils {
   /**
   /**
    * Get the socket factory for the given class according to its
    * Get the socket factory for the given class according to its
    * configuration parameter
    * configuration parameter
-   * <tt>hadoop.rpc.socket.factory.class.&lt;ClassName&gt;</tt>. When no
+   * <code>hadoop.rpc.socket.factory.class.&lt;ClassName&gt;</code>. When no
    * such parameter exists then fall back on the default socket factory as
    * such parameter exists then fall back on the default socket factory as
-   * configured by <tt>hadoop.rpc.socket.factory.class.default</tt>. If
+   * configured by <code>hadoop.rpc.socket.factory.class.default</code>. If
    * this default socket factory is not configured, then fall back on the JVM
    * this default socket factory is not configured, then fall back on the JVM
    * default socket factory.
    * default socket factory.
    * 
    * 
@@ -111,7 +111,7 @@ public class NetUtils {
 
 
   /**
   /**
    * Get the default socket factory as specified by the configuration
    * Get the default socket factory as specified by the configuration
-   * parameter <tt>hadoop.rpc.socket.factory.default</tt>
+   * parameter <code>hadoop.rpc.socket.factory.default</code>
    * 
    * 
    * @param conf the configuration
    * @param conf the configuration
    * @return the default socket factory as specified in the configuration or
    * @return the default socket factory as specified in the configuration or

+ 3 - 3
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/AccessControlException.java

@@ -48,10 +48,10 @@ public class AccessControlException extends IOException {
   
   
   /**
   /**
    * Constructs a new exception with the specified cause and a detail
    * Constructs a new exception with the specified cause and a detail
-   * message of <tt>(cause==null ? null : cause.toString())</tt> (which
-   * typically contains the class and detail message of <tt>cause</tt>).
+   * message of <code>(cause==null ? null : cause.toString())</code> (which
+   * typically contains the class and detail message of <code>cause</code>).
    * @param  cause the cause (which is saved for later retrieval by the
    * @param  cause the cause (which is saved for later retrieval by the
-   *         {@link #getCause()} method).  (A <tt>null</tt> value is
+   *         {@link #getCause()} method).  (A <code>null</code> value is
    *         permitted, and indicates that the cause is nonexistent or
    *         permitted, and indicates that the cause is nonexistent or
    *         unknown.)
    *         unknown.)
    */
    */

+ 3 - 3
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/authorize/AuthorizationException.java

@@ -44,10 +44,10 @@ public class AuthorizationException extends AccessControlException {
   
   
   /**
   /**
    * Constructs a new exception with the specified cause and a detail
    * Constructs a new exception with the specified cause and a detail
-   * message of <tt>(cause==null ? null : cause.toString())</tt> (which
-   * typically contains the class and detail message of <tt>cause</tt>).
+   * message of <code>(cause==null ? null : cause.toString())</code> (which
+   * typically contains the class and detail message of <code>cause</code>).
    * @param  cause the cause (which is saved for later retrieval by the
    * @param  cause the cause (which is saved for later retrieval by the
-   *         {@link #getCause()} method).  (A <tt>null</tt> value is
+   *         {@link #getCause()} method).  (A <code>null</code> value is
    *         permitted, and indicates that the cause is nonexistent or
    *         permitted, and indicates that the cause is nonexistent or
    *         unknown.)
    *         unknown.)
    */
    */

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/GenericOptionsParser.java

@@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
  * line arguments, enabling applications to easily specify a namenode, a 
  * line arguments, enabling applications to easily specify a namenode, a 
  * ResourceManager, additional configuration resources etc.
  * ResourceManager, additional configuration resources etc.
  * 
  * 
- * <h3 id="GenericOptions">Generic Options</h3>
+ * <h2 id="GenericOptions">Generic Options</h2>
  * 
  * 
  * <p>The supported generic options are:</p>
  * <p>The supported generic options are:</p>
  * <blockquote>
  * <blockquote>

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedReadLock.java

@@ -26,7 +26,7 @@ import org.apache.hadoop.classification.VisibleForTesting;
 import org.slf4j.Logger;
 import org.slf4j.Logger;
 
 
 /**
 /**
- * This is a wrap class of a <tt>ReadLock</tt>.
+ * This is a wrap class of a <code>ReadLock</code>.
  * It extends the class {@link InstrumentedLock}, and can be used to track
  * It extends the class {@link InstrumentedLock}, and can be used to track
  * whether a specific read lock is being held for too long and log
  * whether a specific read lock is being held for too long and log
  * warnings if so.
  * warnings if so.

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedReadWriteLock.java

@@ -28,7 +28,7 @@ import org.slf4j.Logger;
 /**
 /**
  * This is a wrap class of a {@link ReentrantReadWriteLock}.
  * This is a wrap class of a {@link ReentrantReadWriteLock}.
  * It implements the interface {@link ReadWriteLock}, and can be used to
  * It implements the interface {@link ReadWriteLock}, and can be used to
- * create instrumented <tt>ReadLock</tt> and <tt>WriteLock</tt>.
+ * create instrumented <code>ReadLock</code> and <code>WriteLock</code>.
  */
  */
 @InterfaceAudience.Private
 @InterfaceAudience.Private
 @InterfaceStability.Unstable
 @InterfaceStability.Unstable

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/InstrumentedWriteLock.java

@@ -26,7 +26,7 @@ import org.apache.hadoop.classification.VisibleForTesting;
 import org.slf4j.Logger;
 import org.slf4j.Logger;
 
 
 /**
 /**
- * This is a wrap class of a <tt>WriteLock</tt>.
+ * This is a wrap class of a <code>WriteLock</code>.
  * It extends the class {@link InstrumentedLock}, and can be used to track
  * It extends the class {@link InstrumentedLock}, and can be used to track
  * whether a specific write lock is being held for too long and log
  * whether a specific write lock is being held for too long and log
  * warnings if so.
  * warnings if so.

+ 8 - 8
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ShutdownThreadsHelper.java

@@ -37,8 +37,8 @@ public class ShutdownThreadsHelper {
 
 
   /**
   /**
    * @param thread {@link Thread to be shutdown}
    * @param thread {@link Thread to be shutdown}
-   * @return <tt>true</tt> if the thread is successfully interrupted,
-   * <tt>false</tt> otherwise
+   * @return <code>true</code> if the thread is successfully interrupted,
+   * <code>false</code> otherwise
    */
    */
   public static boolean shutdownThread(Thread thread) {
   public static boolean shutdownThread(Thread thread) {
     return shutdownThread(thread, SHUTDOWN_WAIT_MS);
     return shutdownThread(thread, SHUTDOWN_WAIT_MS);
@@ -48,8 +48,8 @@ public class ShutdownThreadsHelper {
    * @param thread {@link Thread to be shutdown}
    * @param thread {@link Thread to be shutdown}
    * @param timeoutInMilliSeconds time to wait for thread to join after being
    * @param timeoutInMilliSeconds time to wait for thread to join after being
    *                              interrupted
    *                              interrupted
-   * @return <tt>true</tt> if the thread is successfully interrupted,
-   * <tt>false</tt> otherwise
+   * @return <code>true</code> if the thread is successfully interrupted,
+   * <code>false</code> otherwise
    */
    */
   public static boolean shutdownThread(Thread thread,
   public static boolean shutdownThread(Thread thread,
                                     long timeoutInMilliSeconds) {
                                     long timeoutInMilliSeconds) {
@@ -71,8 +71,8 @@ public class ShutdownThreadsHelper {
    * shutdownExecutorService.
    * shutdownExecutorService.
    *
    *
    * @param service {@link ExecutorService to be shutdown}
    * @param service {@link ExecutorService to be shutdown}
-   * @return <tt>true</tt> if the service is terminated,
-   * <tt>false</tt> otherwise
+   * @return <code>true</code> if the service is terminated,
+   * <code>false</code> otherwise
    * @throws InterruptedException if the thread is interrupted.
    * @throws InterruptedException if the thread is interrupted.
    */
    */
   public static boolean shutdownExecutorService(ExecutorService service)
   public static boolean shutdownExecutorService(ExecutorService service)
@@ -87,8 +87,8 @@ public class ShutdownThreadsHelper {
    * @param timeoutInMs time to wait for {@link
    * @param timeoutInMs time to wait for {@link
    * ExecutorService#awaitTermination(long, java.util.concurrent.TimeUnit)}
    * ExecutorService#awaitTermination(long, java.util.concurrent.TimeUnit)}
    *                    calls in milli seconds.
    *                    calls in milli seconds.
-   * @return <tt>true</tt> if the service is terminated,
-   * <tt>false</tt> otherwise
+   * @return <code>true</code> if the service is terminated,
+   * <code>false</code> otherwise
    * @throws InterruptedException if the thread is interrupted.
    * @throws InterruptedException if the thread is interrupted.
    */
    */
   public static boolean shutdownExecutorService(ExecutorService service,
   public static boolean shutdownExecutorService(ExecutorService service,

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/StringUtils.java

@@ -245,7 +245,7 @@ public class StringUtils {
   /**
   /**
    * @param str
    * @param str
    *          The string array to be parsed into an URI array.
    *          The string array to be parsed into an URI array.
-   * @return <tt>null</tt> if str is <tt>null</tt>, else the URI array
+   * @return <code>null</code> if str is <code>null</code>, else the URI array
    *         equivalent to str.
    *         equivalent to str.
    * @throws IllegalArgumentException
    * @throws IllegalArgumentException
    *           If any string in str violates RFC&nbsp;2396.
    *           If any string in str violates RFC&nbsp;2396.

+ 6 - 6
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/MiniRPCBenchmark.java

@@ -53,18 +53,18 @@ import org.slf4j.event.Level;
  * The benchmark supports three authentication methods:
  * The benchmark supports three authentication methods:
  * <ol>
  * <ol>
  * <li>simple - no authentication. In order to enter this mode 
  * <li>simple - no authentication. In order to enter this mode 
- * the configuration file <tt>core-site.xml</tt> should specify
- * <tt>hadoop.security.authentication = simple</tt>.
+ * the configuration file <code>core-site.xml</code> should specify
+ * <code>hadoop.security.authentication = simple</code>.
  * This is the default mode.</li>
  * This is the default mode.</li>
  * <li>kerberos - kerberos authentication. In order to enter this mode 
  * <li>kerberos - kerberos authentication. In order to enter this mode 
- * the configuration file <tt>core-site.xml</tt> should specify
- * <tt>hadoop.security.authentication = kerberos</tt> and 
+ * the configuration file <code>core-site.xml</code> should specify
+ * <code>hadoop.security.authentication = kerberos</code> and
  * the argument string should provide qualifying
  * the argument string should provide qualifying
- * <tt>keytabFile</tt> and <tt>userName</tt> parameters.
+ * <code>keytabFile</code> and <code>userName</code> parameters.
  * <li>delegation token - authentication using delegation token.
  * <li>delegation token - authentication using delegation token.
  * In order to enter this mode the benchmark should provide all the
  * In order to enter this mode the benchmark should provide all the
  * mentioned parameters for kerberos authentication plus the
  * mentioned parameters for kerberos authentication plus the
- * <tt>useToken</tt> argument option.
+ * <code>useToken</code> argument option.
  * </ol>
  * </ol>
  * Input arguments:
  * Input arguments:
  * <ul>
  * <ul>

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/hdfs/client/impl/BlockReaderLocalLegacy.java

@@ -503,7 +503,7 @@ class BlockReaderLocalLegacy implements BlockReader {
    *          byte buffer to write bytes to. If checksums are not required, buf
    *          byte buffer to write bytes to. If checksums are not required, buf
    *          can have any number of bytes remaining, otherwise there must be a
    *          can have any number of bytes remaining, otherwise there must be a
    *          multiple of the checksum chunk size remaining.
    *          multiple of the checksum chunk size remaining.
-   * @return <tt>max(min(totalBytesRead, len) - offsetFromChunkBoundary, 0)</tt>
+   * @return <code>max(min(totalBytesRead, len) - offsetFromChunkBoundary, 0)</code>
    *         that is, the the number of useful bytes (up to the amount
    *         that is, the the number of useful bytes (up to the amount
    *         requested) readable from the buffer by the client.
    *         requested) readable from the buffer by the client.
    */
    */

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeDescriptor.java

@@ -107,7 +107,7 @@ public class DatanodeDescriptor extends DatanodeInfo {
     }
     }
 
 
     /**
     /**
-     * Returns <tt>true</tt> if the queue contains the specified element.
+     * Returns <code>true</code> if the queue contains the specified element.
      */
      */
     synchronized boolean contains(E e) {
     synchronized boolean contains(E e) {
       return blockq.contains(e);
       return blockq.contains(e);

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EncryptionZoneManager.java

@@ -369,7 +369,7 @@ public class EncryptionZoneManager {
   }
   }
 
 
   /**
   /**
-   * Get the key name for an encryption zone. Returns null if <tt>iip</tt> is
+   * Get the key name for an encryption zone. Returns null if <code>iip</code> is
    * not within an encryption zone.
    * not within an encryption zone.
    * <p>
    * <p>
    * Called while holding the FSDirectory lock.
    * Called while holding the FSDirectory lock.

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/NameNode.java

@@ -1120,7 +1120,7 @@ public class NameNode extends ReconfigurableBase implements
    * <li>{@link StartupOption#IMPORT IMPORT} - import checkpoint</li>
    * <li>{@link StartupOption#IMPORT IMPORT} - import checkpoint</li>
    * </ul>
    * </ul>
    * The option is passed via configuration field: 
    * The option is passed via configuration field: 
-   * <tt>dfs.namenode.startup</tt>
+   * <code>dfs.namenode.startup</code>
    * 
    * 
    * The conf will be modified to reflect the actual ports on which 
    * The conf will be modified to reflect the actual ports on which 
    * the NameNode is up and running if the user passes the port as
    * the NameNode is up and running if the user passes the port as

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/snapshot/DiffList.java

@@ -100,7 +100,7 @@ public interface DiffList<T extends Comparable<Integer>> extends Iterable<T> {
    * @param index index of the element to return
    * @param index index of the element to return
    * @return the element at the specified position in this list
    * @return the element at the specified position in this list
    * @throws IndexOutOfBoundsException if the index is out of range
    * @throws IndexOutOfBoundsException if the index is out of range
-   *         (<tt>index &lt; 0 || index &gt;= size()</tt>)
+   *         (<code>index &lt; 0 || index &gt;= size()</code>)
    */
    */
   T get(int index);
   T get(int index);
 
 

+ 9 - 9
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FileDistributionCalculator.java

@@ -37,20 +37,20 @@ import org.apache.hadoop.util.Preconditions;
 
 
 /**
 /**
  * This is the tool for analyzing file sizes in the namespace image. In order to
  * This is the tool for analyzing file sizes in the namespace image. In order to
- * run the tool one should define a range of integers <tt>[0, maxSize]</tt> by
- * specifying <tt>maxSize</tt> and a <tt>step</tt>. The range of integers is
- * divided into segments of size <tt>step</tt>:
- * <tt>[0, s<sub>1</sub>, ..., s<sub>n-1</sub>, maxSize]</tt>, and the visitor
+ * run the tool one should define a range of integers <code>[0, maxSize]</code> by
+ * specifying <code>maxSize</code> and a <code>step</code>. The range of integers is
+ * divided into segments of size <code>step</code>:
+ * <code>[0, s<sub>1</sub>, ..., s<sub>n-1</sub>, maxSize]</code>, and the visitor
  * calculates how many files in the system fall into each segment
  * calculates how many files in the system fall into each segment
- * <tt>[s<sub>i-1</sub>, s<sub>i</sub>)</tt>. Note that files larger than
- * <tt>maxSize</tt> always fall into the very last segment.
+ * <code>[s<sub>i-1</sub>, s<sub>i</sub>)</code>. Note that files larger than
+ * <code>maxSize</code> always fall into the very last segment.
  *
  *
  * <h3>Input.</h3>
  * <h3>Input.</h3>
  * <ul>
  * <ul>
- * <li><tt>filename</tt> specifies the location of the image file;</li>
- * <li><tt>maxSize</tt> determines the range <tt>[0, maxSize]</tt> of files
+ * <li><code>filename</code> specifies the location of the image file;</li>
+ * <li><code>maxSize</code> determines the range <code>[0, maxSize]</code> of files
  * sizes considered by the visitor;</li>
  * sizes considered by the visitor;</li>
- * <li><tt>step</tt> the range is divided into segments of size step.</li>
+ * <li><code>step</code> the range is divided into segments of size step.</li>
  * </ul>
  * </ul>
  *
  *
  * <h3>Output.</h3> The output file is formatted as a tab separated two column
  * <h3>Output.</h3> The output file is formatted as a tab separated two column

+ 8 - 8
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/tools/offlineImageViewer/FileDistributionVisitor.java

@@ -28,20 +28,20 @@ import org.apache.hadoop.util.StringUtils;
  * <h3>Description.</h3>
  * <h3>Description.</h3>
  * This is the tool for analyzing file sizes in the namespace image.
  * This is the tool for analyzing file sizes in the namespace image.
  * In order to run the tool one should define a range of integers
  * In order to run the tool one should define a range of integers
- * <tt>[0, maxSize]</tt> by specifying <tt>maxSize</tt> and a <tt>step</tt>.
- * The range of integers is divided into segments of size <tt>step</tt>: 
- * <tt>[0, s<sub>1</sub>, ..., s<sub>n-1</sub>, maxSize]</tt>,
+ * <code>[0, maxSize]</code> by specifying <code>maxSize</code> and a <code>step</code>.
+ * The range of integers is divided into segments of size <code>step</code>: 
+ * <code>[0, s<sub>1</sub>, ..., s<sub>n-1</sub>, maxSize]</code>,
  * and the visitor calculates how many files in the system fall into 
  * and the visitor calculates how many files in the system fall into 
- * each segment <tt>[s<sub>i-1</sub>, s<sub>i</sub>)</tt>. 
- * Note that files larger than <tt>maxSize</tt> always fall into 
+ * each segment <code>[s<sub>i-1</sub>, s<sub>i</sub>)</code>. 
+ * Note that files larger than <code>maxSize</code> always fall into 
  * the very last segment.
  * the very last segment.
  * 
  * 
  * <h3>Input.</h3>
  * <h3>Input.</h3>
  * <ul>
  * <ul>
- * <li><tt>filename</tt> specifies the location of the image file;</li>
- * <li><tt>maxSize</tt> determines the range <tt>[0, maxSize]</tt> of files
+ * <li><code>filename</code> specifies the location of the image file;</li>
+ * <li><code>maxSize</code> determines the range <code>[0, maxSize]</code> of files
  * sizes considered by the visitor;</li>
  * sizes considered by the visitor;</li>
- * <li><tt>step</tt> the range is divided into segments of size step.</li>
+ * <li><code>step</code> the range is divided into segments of size step.</li>
  * </ul>
  * </ul>
  *
  *
  * <h3>Output.</h3>
  * <h3>Output.</h3>

+ 1 - 1
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSafeMode.java

@@ -110,7 +110,7 @@ public class TestSafeMode {
    * Name-node should stay in automatic safe-mode.</li>
    * Name-node should stay in automatic safe-mode.</li>
    * <li>Enter safe mode manually.</li>
    * <li>Enter safe mode manually.</li>
    * <li>Start the data-node.</li>
    * <li>Start the data-node.</li>
-   * <li>Wait longer than <tt>dfs.namenode.safemode.extension</tt> and 
+   * <li>Wait longer than <code>dfs.namenode.safemode.extension</code> and
    * verify that the name-node is still in safe mode.</li>
    * verify that the name-node is still in safe mode.</li>
    * </ol>
    * </ol>
    *  
    *  

+ 3 - 3
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestReadOnlySharedStorage.java

@@ -205,7 +205,7 @@ public class TestReadOnlySharedStorage {
   }
   }
   
   
   /**
   /**
-   * Verify that <tt>READ_ONLY_SHARED</tt> replicas are <i>not</i> counted towards the overall 
+   * Verify that <code>READ_ONLY_SHARED</code> replicas are <i>not</i> counted towards the overall 
    * replication count, but <i>are</i> included as replica locations returned to clients for reads.
    * replication count, but <i>are</i> included as replica locations returned to clients for reads.
    */
    */
   @Test
   @Test
@@ -221,7 +221,7 @@ public class TestReadOnlySharedStorage {
   }
   }
 
 
   /**
   /**
-   * Verify that the NameNode is able to still use <tt>READ_ONLY_SHARED</tt> replicas even 
+   * Verify that the NameNode is able to still use <code>READ_ONLY_SHARED</code> replicas even 
    * when the single NORMAL replica is offline (and the effective replication count is 0).
    * when the single NORMAL replica is offline (and the effective replication count is 0).
    */
    */
   @Test
   @Test
@@ -253,7 +253,7 @@ public class TestReadOnlySharedStorage {
   }
   }
   
   
   /**
   /**
-   * Verify that corrupt <tt>READ_ONLY_SHARED</tt> replicas aren't counted 
+   * Verify that corrupt <code>READ_ONLY_SHARED</code> replicas aren't counted 
    * towards the corrupt replicas total.
    * towards the corrupt replicas total.
    */
    */
   @Test
   @Test

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/preemption/AMPreemptionPolicy.java

@@ -109,7 +109,7 @@ public interface AMPreemptionPolicy {
    * TaskId}. Assigning a null is akin to remove all previous checkpoints for
    * TaskId}. Assigning a null is akin to remove all previous checkpoints for
    * this task.
    * this task.
    * @param taskId TaskID
    * @param taskId TaskID
-   * @param cid Checkpoint to assign or <tt>null</tt> to remove it.
+   * @param cid Checkpoint to assign or <code>null</code> to remove it.
    */
    */
   public void setCheckpointID(TaskId taskId, TaskCheckpointID cid);
   public void setCheckpointID(TaskId taskId, TaskCheckpointID cid);
 
 

+ 8 - 8
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java

@@ -185,7 +185,7 @@ public abstract class FileOutputFormat<K, V> implements OutputFormat<K, V> {
    *  is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not 
    *  is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not 
    *  a <code>FileOutputCommitter</code>, the task's temporary output
    *  a <code>FileOutputCommitter</code>, the task's temporary output
    *  directory is same as {@link #getOutputPath(JobConf)} i.e.
    *  directory is same as {@link #getOutputPath(JobConf)} i.e.
-   *  <tt>${mapreduce.output.fileoutputformat.outputdir}$</tt></p>
+   *  <code>${mapreduce.output.fileoutputformat.outputdir}$</code></p>
    *  
    *  
    * <p>Some applications need to create/write-to side-files, which differ from
    * <p>Some applications need to create/write-to side-files, which differ from
    * the actual job-outputs.
    * the actual job-outputs.
@@ -194,27 +194,27 @@ public abstract class FileOutputFormat<K, V> implements OutputFormat<K, V> {
    * (running simultaneously e.g. speculative tasks) trying to open/write-to the
    * (running simultaneously e.g. speculative tasks) trying to open/write-to the
    * same file (path) on HDFS. Hence the application-writer will have to pick 
    * same file (path) on HDFS. Hence the application-writer will have to pick 
    * unique names per task-attempt (e.g. using the attemptid, say 
    * unique names per task-attempt (e.g. using the attemptid, say 
-   * <tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p> 
+   * <code>attempt_200709221812_0001_m_000000_0</code>), not just per TIP.</p>
    * 
    * 
    * <p>To get around this the Map-Reduce framework helps the application-writer 
    * <p>To get around this the Map-Reduce framework helps the application-writer 
    * out by maintaining a special 
    * out by maintaining a special 
-   * <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> 
+   * <code>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</code>
    * sub-directory for each task-attempt on HDFS where the output of the 
    * sub-directory for each task-attempt on HDFS where the output of the 
    * task-attempt goes. On successful completion of the task-attempt the files 
    * task-attempt goes. On successful completion of the task-attempt the files 
-   * in the <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> (only) 
-   * are <i>promoted</i> to <tt>${mapreduce.output.fileoutputformat.outputdir}</tt>. Of course, the 
+   * in the <code>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</code> (only)
+   * are <i>promoted</i> to <code>${mapreduce.output.fileoutputformat.outputdir}</code>. Of course, the
    * framework discards the sub-directory of unsuccessful task-attempts. This 
    * framework discards the sub-directory of unsuccessful task-attempts. This 
    * is completely transparent to the application.</p>
    * is completely transparent to the application.</p>
    * 
    * 
    * <p>The application-writer can take advantage of this by creating any 
    * <p>The application-writer can take advantage of this by creating any 
-   * side-files required in <tt>${mapreduce.task.output.dir}</tt> during execution 
+   * side-files required in <code>${mapreduce.task.output.dir}</code> during execution
    * of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the 
    * of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the 
    * framework will move them out similarly - thus she doesn't have to pick 
    * framework will move them out similarly - thus she doesn't have to pick 
    * unique paths per task-attempt.</p>
    * unique paths per task-attempt.</p>
    * 
    * 
-   * <p><i>Note</i>: the value of <tt>${mapreduce.task.output.dir}</tt> during 
+   * <p><i>Note</i>: the value of <code>${mapreduce.task.output.dir}</code> during
    * execution of a particular task-attempt is actually 
    * execution of a particular task-attempt is actually 
-   * <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_{$taskid}</tt>, and this value is 
+   * <code>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_{$taskid}</code>, and this value is
    * set by the map-reduce framework. So, just create any side-files in the 
    * set by the map-reduce framework. So, just create any side-files in the 
    * path  returned by {@link #getWorkOutputPath(JobConf)} from map/reduce 
    * path  returned by {@link #getWorkOutputPath(JobConf)} from map/reduce 
    * task to take advantage of this feature.</p>
    * task to take advantage of this feature.</p>

+ 2 - 2
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java

@@ -1873,8 +1873,8 @@ public class JobConf extends Configuration {
    * Set the uri to be invoked in-order to send a notification after the job
    * Set the uri to be invoked in-order to send a notification after the job
    * has completed (success/failure).
    * has completed (success/failure).
    * 
    * 
-   * <p>The uri can contain 2 special parameters: <tt>$jobId</tt> and 
-   * <tt>$jobStatus</tt>. Those, if present, are replaced by the job's 
+   * <p>The uri can contain 2 special parameters: <code>$jobId</code> and 
+   * <code>$jobStatus</code>. Those, if present, are replaced by the job's 
    * identifier and completion-status respectively.</p>
    * identifier and completion-status respectively.</p>
    * 
    * 
    * <p>This is typically used by application-writers to implement chaining of 
    * <p>This is typically used by application-writers to implement chaining of 

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/MapRunnable.java

@@ -37,7 +37,7 @@ public interface MapRunnable<K1, V1, K2, V2>
     extends JobConfigurable {
     extends JobConfigurable {
   
   
   /** 
   /** 
-   * Start mapping input <tt>&lt;key, value&gt;</tt> pairs.
+   * Start mapping input <code>&lt;key, value&gt;</code> pairs.
    *  
    *  
    * <p>Mapping of input records to output records is complete when this method 
    * <p>Mapping of input records to output records is complete when this method 
    * returns.</p>
    * returns.</p>

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/jobcontrol/Job.java

@@ -143,7 +143,7 @@ public class Job extends ControlledJob {
    * is waiting to run, not during or afterwards.
    * is waiting to run, not during or afterwards.
    * 
    * 
    * @param dependingJob Job that this Job depends on.
    * @param dependingJob Job that this Job depends on.
-   * @return <tt>true</tt> if the Job was added.
+   * @return <code>true</code> if the Job was added.
    */
    */
   public synchronized boolean addDependingJob(Job dependingJob) {
   public synchronized boolean addDependingJob(Job dependingJob) {
     return super.addDependingJob(dependingJob);
     return super.addDependingJob(dependingJob);

+ 6 - 6
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/CompositeInputFormat.java

@@ -38,10 +38,10 @@ import org.apache.hadoop.mapred.Reporter;
  * and partitioned the same way.
  * and partitioned the same way.
  *
  *
  * A user may define new join types by setting the property
  * A user may define new join types by setting the property
- * <tt>mapred.join.define.&lt;ident&gt;</tt> to a classname. In the expression
- * <tt>mapred.join.expr</tt>, the identifier will be assumed to be a
+ * <code>mapred.join.define.&lt;ident&gt;</code> to a classname. In the expression
+ * <code>mapred.join.expr</code>, the identifier will be assumed to be a
  * ComposableRecordReader.
  * ComposableRecordReader.
- * <tt>mapred.join.keycomparator</tt> can be a classname used to compare keys
+ * <code>mapred.join.keycomparator</code> can be a classname used to compare keys
  * in the join.
  * in the join.
  * @see #setFormat
  * @see #setFormat
  * @see JoinRecordReader
  * @see JoinRecordReader
@@ -66,9 +66,9 @@ public class CompositeInputFormat<K extends WritableComparable>
    *   class ::= @see java.lang.Class#forName(java.lang.String)
    *   class ::= @see java.lang.Class#forName(java.lang.String)
    *   path  ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
    *   path  ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
    * }
    * }
-   * Reads expression from the <tt>mapred.join.expr</tt> property and
-   * user-supplied join types from <tt>mapred.join.define.&lt;ident&gt;</tt>
-   *  types. Paths supplied to <tt>tbl</tt> are given as input paths to the
+   * Reads expression from the <code>mapred.join.expr</code> property and
+   * user-supplied join types from <code>mapred.join.define.&lt;ident&gt;</code>
+   *  types. Paths supplied to <code>tbl</code> are given as input paths to the
    * InputFormat class listed.
    * InputFormat class listed.
    * @see #compose(java.lang.String, java.lang.Class, java.lang.String...)
    * @see #compose(java.lang.String, java.lang.Class, java.lang.String...)
    */
    */

+ 2 - 2
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/CompositeRecordReader.java

@@ -61,8 +61,8 @@ public abstract class CompositeRecordReader<
   protected abstract boolean combine(Object[] srcs, TupleWritable value);
   protected abstract boolean combine(Object[] srcs, TupleWritable value);
 
 
   /**
   /**
-   * Create a RecordReader with <tt>capacity</tt> children to position
-   * <tt>id</tt> in the parent reader.
+   * Create a RecordReader with <code>capacity</code> children to position
+   * <code>id</code> in the parent reader.
    * The id of a root CompositeRecordReader is -1 by convention, but relying
    * The id of a root CompositeRecordReader is -1 by convention, but relying
    * on this is not recommended.
    * on this is not recommended.
    */
    */

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/OverrideRecordReader.java

@@ -31,7 +31,7 @@ import org.apache.hadoop.mapred.JobConf;
 
 
 /**
 /**
  * Prefer the &quot;rightmost&quot; data source for this key.
  * Prefer the &quot;rightmost&quot; data source for this key.
- * For example, <tt>override(S1,S2,S3)</tt> will prefer values
+ * For example, <code>override(S1,S2,S3)</code> will prefer values
  * from S3 over S2, and values from S2 over S1 for all keys
  * from S3 over S2, and values from S2 over S1 for all keys
  * emitted from all sources.
  * emitted from all sources.
  */
  */

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/join/Parser.java

@@ -275,7 +275,7 @@ public class Parser {
 
 
     /**
     /**
      * Let the first actual define the InputFormat and the second define
      * Let the first actual define the InputFormat and the second define
-     * the <tt>mapred.input.dir</tt> property.
+     * the <code>mapred.input.dir</code> property.
      */
      */
     public void parse(List<Token> ll, JobConf job) throws IOException {
     public void parse(List<Token> ll, JobConf job) throws IOException {
       StringBuilder sb = new StringBuilder();
       StringBuilder sb = new StringBuilder();

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TotalOrderPartitioner.java

@@ -43,7 +43,7 @@ public class TotalOrderPartitioner<K ,V>
 
 
   /**
   /**
    * Set the path to the SequenceFile storing the sorted partition keyset.
    * Set the path to the SequenceFile storing the sorted partition keyset.
-   * It must be the case that for <tt>R</tt> reduces, there are <tt>R-1</tt>
+   * It must be the case that for <code>R</code> reduces, there are <code>R-1</code>
    * keys in the SequenceFile.
    * keys in the SequenceFile.
    * @deprecated Use 
    * @deprecated Use 
    * {@link #setPartitionFile(Configuration, Path)}
    * {@link #setPartitionFile(Configuration, Path)}

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/jobcontrol/ControlledJob.java

@@ -205,7 +205,7 @@ public class ControlledJob {
    * is waiting to run, not during or afterwards.
    * is waiting to run, not during or afterwards.
    * 
    * 
    * @param dependingJob Job that this Job depends on.
    * @param dependingJob Job that this Job depends on.
-   * @return <tt>true</tt> if the Job was added.
+   * @return <code>true</code> if the Job was added.
    */
    */
   public synchronized boolean addDependingJob(ControlledJob dependingJob) {
   public synchronized boolean addDependingJob(ControlledJob dependingJob) {
     if (this.state == State.WAITING) { //only allowed to add jobs when waiting
     if (this.state == State.WAITING) { //only allowed to add jobs when waiting

+ 6 - 6
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/CompositeInputFormat.java

@@ -41,10 +41,10 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
  * and partitioned the same way.
  * and partitioned the same way.
  *
  *
  * A user may define new join types by setting the property
  * A user may define new join types by setting the property
- * <tt>mapreduce.join.define.&lt;ident&gt;</tt> to a classname. 
- * In the expression <tt>mapreduce.join.expr</tt>, the identifier will be
+ * <code>mapreduce.join.define.&lt;ident&gt;</code> to a classname.
+ * In the expression <code>mapreduce.join.expr</code>, the identifier will be
  * assumed to be a ComposableRecordReader.
  * assumed to be a ComposableRecordReader.
- * <tt>mapreduce.join.keycomparator</tt> can be a classname used to compare 
+ * <code>mapreduce.join.keycomparator</code> can be a classname used to compare
  * keys in the join.
  * keys in the join.
  * @see #setFormat
  * @see #setFormat
  * @see JoinRecordReader
  * @see JoinRecordReader
@@ -73,9 +73,9 @@ public class CompositeInputFormat<K extends WritableComparable>
    *   class ::= @see java.lang.Class#forName(java.lang.String)
    *   class ::= @see java.lang.Class#forName(java.lang.String)
    *   path  ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
    *   path  ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
    * }
    * }
-   * Reads expression from the <tt>mapreduce.join.expr</tt> property and
-   * user-supplied join types from <tt>mapreduce.join.define.&lt;ident&gt;</tt>
-   *  types. Paths supplied to <tt>tbl</tt> are given as input paths to the
+   * Reads expression from the <code>mapreduce.join.expr</code> property and
+   * user-supplied join types from <code>mapreduce.join.define.&lt;ident&gt;</code>
+   *  types. Paths supplied to <code>tbl</code> are given as input paths to the
    * InputFormat class listed.
    * InputFormat class listed.
    * @see #compose(java.lang.String, java.lang.Class, java.lang.String...)
    * @see #compose(java.lang.String, java.lang.Class, java.lang.String...)
    */
    */

+ 2 - 2
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/CompositeRecordReader.java

@@ -67,8 +67,8 @@ public abstract class CompositeRecordReader<
   protected X value;
   protected X value;
 
 
   /**
   /**
-   * Create a RecordReader with <tt>capacity</tt> children to position
-   * <tt>id</tt> in the parent reader.
+   * Create a RecordReader with <code>capacity</code> children to position
+   * <code>id</code> in the parent reader.
    * The id of a root CompositeRecordReader is -1 by convention, but relying
    * The id of a root CompositeRecordReader is -1 by convention, but relying
    * on this is not recommended.
    * on this is not recommended.
    */
    */

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/OverrideRecordReader.java

@@ -33,7 +33,7 @@ import org.apache.hadoop.util.ReflectionUtils;
 
 
 /**
 /**
  * Prefer the &quot;rightmost&quot; data source for this key.
  * Prefer the &quot;rightmost&quot; data source for this key.
- * For example, <tt>override(S1,S2,S3)</tt> will prefer values
+ * For example, <code>override(S1,S2,S3)</code> will prefer values
  * from S3 over S2, and values from S2 over S1 for all keys
  * from S3 over S2, and values from S2 over S1 for all keys
  * emitted from all sources.
  * emitted from all sources.
  */
  */

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/Parser.java

@@ -290,7 +290,7 @@ public abstract static class Node extends ComposableInputFormat {
 
 
     /**
     /**
      * Let the first actual define the InputFormat and the second define
      * Let the first actual define the InputFormat and the second define
-     * the <tt>mapred.input.dir</tt> property.
+     * the <code>mapred.input.dir</code> property.
      */
      */
     @Override
     @Override
     public void parse(List<Token> ll, Configuration conf) throws IOException {
     public void parse(List<Token> ll, Configuration conf) throws IOException {

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java

@@ -144,7 +144,7 @@ public class TupleWritable implements Writable, Iterable<Writable> {
 
 
   /**
   /**
    * Convert Tuple to String as in the following.
    * Convert Tuple to String as in the following.
-   * <tt>[&lt;child1&gt;,&lt;child2&gt;,...,&lt;childn&gt;]</tt>
+   * <code>[&lt;child1&gt;,&lt;child2&gt;,...,&lt;childn&gt;]</code>
    */
    */
   public String toString() {
   public String toString() {
     StringBuilder buf = new StringBuilder("[");
     StringBuilder buf = new StringBuilder("[");

+ 4 - 4
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/output/FileOutputFormat.java

@@ -208,15 +208,15 @@ public abstract class FileOutputFormat<K, V> extends OutputFormat<K, V> {
    * (running simultaneously e.g. speculative tasks) trying to open/write-to the
    * (running simultaneously e.g. speculative tasks) trying to open/write-to the
    * same file (path) on HDFS. Hence the application-writer will have to pick 
    * same file (path) on HDFS. Hence the application-writer will have to pick 
    * unique names per task-attempt (e.g. using the attemptid, say 
    * unique names per task-attempt (e.g. using the attemptid, say 
-   * <tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p> 
+   * <code>attempt_200709221812_0001_m_000000_0</code>), not just per TIP.</p>
    * 
    * 
    * <p>To get around this the Map-Reduce framework helps the application-writer 
    * <p>To get around this the Map-Reduce framework helps the application-writer 
    * out by maintaining a special 
    * out by maintaining a special 
-   * <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> 
+   * <code>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</code>
    * sub-directory for each task-attempt on HDFS where the output of the 
    * sub-directory for each task-attempt on HDFS where the output of the 
    * task-attempt goes. On successful completion of the task-attempt the files 
    * task-attempt goes. On successful completion of the task-attempt the files 
-   * in the <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> (only) 
-   * are <i>promoted</i> to <tt>${mapreduce.output.fileoutputformat.outputdir}</tt>. Of course, the 
+   * in the <code>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</code> (only)
+   * are <i>promoted</i> to <code>${mapreduce.output.fileoutputformat.outputdir}</code>. Of course, the
    * framework discards the sub-directory of unsuccessful task-attempts. This 
    * framework discards the sub-directory of unsuccessful task-attempts. This 
    * is completely transparent to the application.</p>
    * is completely transparent to the application.</p>
    * 
    * 

+ 5 - 5
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.java

@@ -65,8 +65,8 @@ public class TotalOrderPartitioner<K,V>
   /**
   /**
    * Read in the partition file and build indexing data structures.
    * Read in the partition file and build indexing data structures.
    * If the keytype is {@link org.apache.hadoop.io.BinaryComparable} and
    * If the keytype is {@link org.apache.hadoop.io.BinaryComparable} and
-   * <tt>total.order.partitioner.natural.order</tt> is not false, a trie
-   * of the first <tt>total.order.partitioner.max.trie.depth</tt>(2) + 1 bytes
+   * <code>total.order.partitioner.natural.order</code> is not false, a trie
+   * of the first <code>total.order.partitioner.max.trie.depth</code>(2) + 1 bytes
    * will be built. Otherwise, keys will be located using a binary search of
    * will be built. Otherwise, keys will be located using a binary search of
    * the partition keyset using the {@link org.apache.hadoop.io.RawComparator}
    * the partition keyset using the {@link org.apache.hadoop.io.RawComparator}
    * defined for this job. The input file must be sorted with the same
    * defined for this job. The input file must be sorted with the same
@@ -128,7 +128,7 @@ public class TotalOrderPartitioner<K,V>
 
 
   /**
   /**
    * Set the path to the SequenceFile storing the sorted partition keyset.
    * Set the path to the SequenceFile storing the sorted partition keyset.
-   * It must be the case that for <tt>R</tt> reduces, there are <tt>R-1</tt>
+   * It must be the case that for <code>R</code> reduces, there are <code>R-1</code>
    * keys in the SequenceFile.
    * keys in the SequenceFile.
    */
    */
   public static void setPartitionFile(Configuration conf, Path p) {
   public static void setPartitionFile(Configuration conf, Path p) {
@@ -156,7 +156,7 @@ public class TotalOrderPartitioner<K,V>
 
 
   /**
   /**
    * Base class for trie nodes. If the keytype is memcomp-able, this builds
    * Base class for trie nodes. If the keytype is memcomp-able, this builds
-   * tries of the first <tt>total.order.partitioner.max.trie.depth</tt>
+   * tries of the first <code>total.order.partitioner.max.trie.depth</code>
    * bytes.
    * bytes.
    */
    */
   static abstract class TrieNode implements Node<BinaryComparable> {
   static abstract class TrieNode implements Node<BinaryComparable> {
@@ -171,7 +171,7 @@ public class TotalOrderPartitioner<K,V>
 
 
   /**
   /**
    * For types that are not {@link org.apache.hadoop.io.BinaryComparable} or
    * For types that are not {@link org.apache.hadoop.io.BinaryComparable} or
-   * where disabled by <tt>total.order.partitioner.natural.order</tt>,
+   * where disabled by <code>total.order.partitioner.natural.order</code>,
    * search the partition keyset with a binary search.
    * search the partition keyset with a binary search.
    */
    */
   class BinarySearchNode implements Node<K> {
   class BinarySearchNode implements Node<K> {

+ 4 - 4
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/AccumulatingReducer.java

@@ -31,13 +31,13 @@ import org.slf4j.LoggerFactory;
  * The type is specified in the key part of the key-value pair 
  * The type is specified in the key part of the key-value pair 
  * as a prefix to the key in the following way
  * as a prefix to the key in the following way
  * <p>
  * <p>
- * <tt>type:key</tt>
+ * <code>type:key</code>
  * <p>
  * <p>
  * The values are accumulated according to the types:
  * The values are accumulated according to the types:
  * <ul>
  * <ul>
- * <li><tt>s:</tt> - string, concatenate</li>
- * <li><tt>f:</tt> - float, summ</li>
- * <li><tt>l:</tt> - long, summ</li>
+ * <li><code>s:</code> - string, concatenate</li>
+ * <li><code>f:</code> - float, summ</li>
+ * <li><code>l:</code> - long, summ</li>
  * </ul>
  * </ul>
  * 
  * 
  */
  */

+ 2 - 2
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/IOMapperBase.java

@@ -109,8 +109,8 @@ public abstract class IOMapperBase<T> extends Configured
    * Map file name and offset into statistical data.
    * Map file name and offset into statistical data.
    * <p>
    * <p>
    * The map task is to get the 
    * The map task is to get the 
-   * <tt>key</tt>, which contains the file name, and the 
-   * <tt>value</tt>, which is the offset within the file.
+   * <code>key</code>, which contains the file name, and the
+   * <code>value</code>, which is the offset within the file.
    * 
    * 
    * The parameters are passed to the abstract method 
    * The parameters are passed to the abstract method 
    * {@link #doIO(Reporter,String,long)}, which performs the io operation, 
    * {@link #doIO(Reporter,String,long)}, which performs the io operation, 

+ 21 - 21
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/fs/JHLogAnalyzer.java

@@ -76,7 +76,7 @@ import org.slf4j.LoggerFactory;
  * specific attempt A during hour h.
  * specific attempt A during hour h.
  * The tool then sums all slots for all attempts for every hour.
  * The tool then sums all slots for all attempts for every hour.
  * The result is the slot hour utilization of the cluster:
  * The result is the slot hour utilization of the cluster:
- * <tt>slotTime(h) = SUM<sub>A</sub> slotTime(A,h)</tt>.
+ * <code>slotTime(h) = SUM<sub>A</sub> slotTime(A,h)</code>.
  * <p>
  * <p>
  * Log analyzer calculates slot hours for <em>MAP</em> and <em>REDUCE</em> 
  * Log analyzer calculates slot hours for <em>MAP</em> and <em>REDUCE</em> 
  * attempts separately.
  * attempts separately.
@@ -88,8 +88,8 @@ import org.slf4j.LoggerFactory;
  * <p>
  * <p>
  * Map-reduce clusters are usually configured to have a fixed number of MAP 
  * Map-reduce clusters are usually configured to have a fixed number of MAP 
  * and REDUCE slots per node. Thus the maximal possible number of slots on
  * and REDUCE slots per node. Thus the maximal possible number of slots on
- * the cluster is <tt>total_slots = total_nodes * slots_per_node</tt>.
- * Effective slot hour cannot exceed <tt>total_slots</tt> for successful
+ * the cluster is <code>total_slots = total_nodes * slots_per_node</code>.
+ * Effective slot hour cannot exceed <code>total_slots</code> for successful
  * attempts.
  * attempts.
  * <p>
  * <p>
  * <em>Pending time</em> characterizes the wait time of attempts.
  * <em>Pending time</em> characterizes the wait time of attempts.
@@ -106,39 +106,39 @@ import org.slf4j.LoggerFactory;
  * The following input parameters can be specified in the argument string
  * The following input parameters can be specified in the argument string
  * to the job log analyzer:
  * to the job log analyzer:
  * <ul>
  * <ul>
- * <li><tt>-historyDir inputDir</tt> specifies the location of the directory
+ * <li><code>-historyDir inputDir</code> specifies the location of the directory
  * where analyzer will be looking for job history log files.</li>
  * where analyzer will be looking for job history log files.</li>
- * <li><tt>-resFile resultFile</tt> the name of the result file.</li>
- * <li><tt>-usersIncluded | -usersExcluded userList</tt> slot utilization and 
+ * <li><code>-resFile resultFile</code> the name of the result file.</li>
+ * <li><code>-usersIncluded | -usersExcluded userList</code> slot utilization and
  * pending time can be calculated for all or for all but the specified users.
  * pending time can be calculated for all or for all but the specified users.
  * <br>
  * <br>
- * <tt>userList</tt> is a comma or semicolon separated list of users.</li>
- * <li><tt>-gzip</tt> is used if history log files are compressed.
+ * <code>userList</code> is a comma or semicolon separated list of users.</li>
+ * <li><code>-gzip</code> is used if history log files are compressed.
  * Only {@link GzipCodec} is currently supported.</li>
  * Only {@link GzipCodec} is currently supported.</li>
- * <li><tt>-jobDelimiter pattern</tt> one can concatenate original log files into 
+ * <li><code>-jobDelimiter pattern</code> one can concatenate original log files into
  * larger file(s) with the specified delimiter to recognize the end of the log
  * larger file(s) with the specified delimiter to recognize the end of the log
  * for one job from the next one.<br>
  * for one job from the next one.<br>
- * <tt>pattern</tt> is a java regular expression
+ * <code>pattern</code> is a java regular expression
  * {@link java.util.regex.Pattern}, which should match only the log delimiters.
  * {@link java.util.regex.Pattern}, which should match only the log delimiters.
  * <br>
  * <br>
- * E.g. pattern <tt>".!!FILE=.*!!"</tt> matches delimiters, which contain
+ * E.g. pattern <code>".!!FILE=.*!!"</code> matches delimiters, which contain
  * the original history log file names in the following form:<br>
  * the original history log file names in the following form:<br>
- * <tt>"$!!FILE=my.job.tracker.com_myJobId_user_wordcount.log!!"</tt></li>
- * <li><tt>-clean</tt> cleans up default directories used by the analyzer.</li>
- * <li><tt>-test</tt> test one file locally and exit;
+ * <code>"$!!FILE=my.job.tracker.com_myJobId_user_wordcount.log!!"</code></li>
+ * <li><code>-clean</code> cleans up default directories used by the analyzer.</li>
+ * <li><code>-test</code> test one file locally and exit;
  * does not require map-reduce.</li>
  * does not require map-reduce.</li>
- * <li><tt>-help</tt> print usage.</li>
+ * <li><code>-help</code> print usage.</li>
  * </ul>
  * </ul>
  * 
  * 
  * <h3>Output.</h3>
  * <h3>Output.</h3>
  * The output file is formatted as a tab separated table consisting of four
  * The output file is formatted as a tab separated table consisting of four
- * columns: <tt>SERIES, PERIOD, TYPE, SLOT_HOUR</tt>.
+ * columns: <code>SERIES, PERIOD, TYPE, SLOT_HOUR</code>.
  * <ul>
  * <ul>
- * <li><tt>SERIES</tt> one of the four statistical series;</li>
- * <li><tt>PERIOD</tt> the start of the time interval in the following format:
- * <tt>"yyyy-mm-dd hh:mm:ss"</tt>;</li>
- * <li><tt>TYPE</tt> the slot type, e.g. MAP or REDUCE;</li>
- * <li><tt>SLOT_HOUR</tt> the value of the slot usage during this 
+ * <li><code>SERIES</code> one of the four statistical series;</li>
+ * <li><code>PERIOD</code> the start of the time interval in the following format:
+ * <code>"yyyy-mm-dd hh:mm:ss"</code>;</li>
+ * <li><code>TYPE</code> the slot type, e.g. MAP or REDUCE;</li>
+ * <li><code>SLOT_HOUR</code> the value of the slot usage during this
  * time interval.</li>
  * time interval.</li>
  * </ul>
  * </ul>
  */
  */

+ 37 - 34
hadoop-mapreduce-project/hadoop-mapreduce-examples/src/main/java/org/apache/hadoop/examples/pi/package.html

@@ -23,7 +23,7 @@
   for large n, say n &gt; 100,000,000.
   for large n, say n &gt; 100,000,000.
   For computing the lower bits of &pi;, consider using <i>bbp</i>.
   For computing the lower bits of &pi;, consider using <i>bbp</i>.
 
 
-  <h3>The distbbp Program</h3>
+  <h2>The distbbp Program</h2>
   The main class is DistBbp
   The main class is DistBbp
   and the actually computation is done by DistSum jobs.
   and the actually computation is done by DistSum jobs.
   The steps for launching the jobs are:
   The steps for launching the jobs are:
@@ -39,8 +39,10 @@
     <li>Combine the job outputs and print the &pi; bits.</li>
     <li>Combine the job outputs and print the &pi; bits.</li>
   </ol>
   </ol>
 
 
-<table summary="The Bits of Pi"><tr valign=top><td width=420>
-  <h3>The Bits of &pi;</h3>
+<table>
+<caption>"The Bits of Pi"</caption>
+<tr><td>
+  <h2>The Bits of &pi;</h2>
   <p>
   <p>
 The table on the right are the results computed by distbbp.
 The table on the right are the results computed by distbbp.
 </p>
 </p>
@@ -56,7 +58,7 @@ The table on the right are the results computed by distbbp.
     <li>The computations in Row 13 and Row 14 were completed on May 20, 2009. 
     <li>The computations in Row 13 and Row 14 were completed on May 20, 2009. 
         It seems that the corresponding bits were never computed before.</li>
         It seems that the corresponding bits were never computed before.</li>
 </ul></li>
 </ul></li>
-<li>The first part of Row 15 (<tt>6216B06</tt>)
+<li>The first part of Row 15 (<code>6216B06</code>)
 
 
 <ul><li>The first 30% of the computation was done in idle cycles of some 
 <ul><li>The first 30% of the computation was done in idle cycles of some 
         clusters spread over 20 days.</li>
         clusters spread over 20 days.</li>
@@ -69,7 +71,7 @@ The table on the right are the results computed by distbbp.
         <a href="http://yahoohadoop.tumblr.com/post/98338598026/hadoop-computes-the-10-15-1st-bit-of-%CF%80">this YDN blog</a>.</li>
         <a href="http://yahoohadoop.tumblr.com/post/98338598026/hadoop-computes-the-10-15-1st-bit-of-%CF%80">this YDN blog</a>.</li>
 
 
 </ul></li>
 </ul></li>
-<li>The second part of Row 15 (<tt>D3611</tt>)
+<li>The second part of Row 15 (<code>D3611</code>)
 <ul><li>The starting position is 1,000,000,000,000,053, totally 20 bits.</li>
 <ul><li>The starting position is 1,000,000,000,000,053, totally 20 bits.</li>
     <li>Two computations, at positions <i>n</i> and <i>n</i>+4, were performed.
     <li>Two computations, at positions <i>n</i> and <i>n</i>+4, were performed.
     <li>A single computation was divided into 14,000 jobs
     <li>A single computation was divided into 14,000 jobs
@@ -85,42 +87,42 @@ The table on the right are the results computed by distbbp.
         computed ever in the history.</li>
         computed ever in the history.</li>
 </ul></li>
 </ul></li>
 </ul>
 </ul>
-</td><td width=20></td><td>
-<table border=1 width=400 cellpadding=5 summary="Pi in hex">
-<tr><th width=30></th><th>Position <i>n</i></th><th>&pi; bits (in hex) starting at <i>n</i></th></tr>
-
-<tr><td align=right>0</td><td align=right>1</td><td><tt>243F6A8885A3</tt><sup>*</sup></td></tr>
-<tr><td align=right>1</td><td align=right>11</td><td><tt>FDAA22168C23</tt></td></tr>
-<tr><td align=right>2</td><td align=right>101</td><td><tt>3707344A409</tt></td></tr>
-<tr><td align=right>3</td><td align=right>1,001</td><td><tt>574E69A458F</tt></td></tr>
-
-<tr><td align=right>4</td><td align=right>10,001</td><td><tt>44EC5716F2B</tt></td></tr>
-<tr><td align=right>5</td><td align=right>100,001</td><td><tt>944F7A204</tt></td></tr>
-<tr><td align=right>6</td><td align=right>1,000,001</td><td><tt>6FFFA4103</tt></td></tr>
-<tr><td align=right>7</td><td align=right>10,000,001</td><td><tt>6CFDD54E3</tt></td></tr>
-<tr><td align=right>8</td><td align=right>100,000,001</td><td><tt>A306CFA7</tt></td></tr>
-
-<tr><td align=right>9</td><td align=right>1,000,000,001</td><td><tt>3E08FF2B</tt></td></tr>
-<tr><td align=right>10</td><td align=right>10,000,000,001</td><td><tt>0A8BD8C0</tt></td></tr>
-<tr><td align=right>11</td><td align=right>100,000,000,001</td><td><tt>B2238C1</tt></td></tr>
-<tr><td align=right>12</td><td align=right>1,000,000,000,001</td><td><tt>0FEE563</tt></td></tr>
-<tr><td align=right>13</td><td align=right>10,000,000,000,001</td><td><tt>896DC3</tt></td></tr>
-
-<tr><td align=right>14</td><td align=right>100,000,000,000,001</td><td><tt>C216EC</tt></td></tr>
-<tr><td align=right>15</td><td align=right>1,000,000,000,000,001</td><td><tt>6216B06</tt> ... <tt>D3611</tt></td></tr>
+</td><td></td><td>
+<table border=1><caption>"Pi in hex"</caption>
+<tr><th></th><th>Position <i>n</i></th><th>&pi; bits (in hex) starting at <i>n</i></th></tr>
+
+<tr><td>0</td><td>1</td><td><code>243F6A8885A3</code><sup>*</sup></td></tr>
+<tr><td>1</td><td>11</td><td><code>FDAA22168C23</code></td></tr>
+<tr><td>2</td><td>101</td><td><code>3707344A409</code></td></tr>
+<tr><td>3</td><td>1,001</td><td><code>574E69A458F</code></td></tr>
+
+<tr><td>4</td><td>10,001</td><td><code>44EC5716F2B</code></td></tr>
+<tr><td>5</td><td>100,001</td><td><code>944F7A204</code></td></tr>
+<tr><td>6</td><td>1,000,001</td><td><code>6FFFA4103</code></td></tr>
+<tr><td>7</td><td>10,000,001</td><td><code>6CFDD54E3</code></td></tr>
+<tr><td>8</td><td>100,000,001</td><td><code>A306CFA7</code></td></tr>
+
+<tr><td>9</td><td>1,000,000,001</td><td><code>3E08FF2B</code></td></tr>
+<tr><td>10</td><td>10,000,000,001</td><td><code>0A8BD8C0</code></td></tr>
+<tr><td>11</td><td>100,000,000,001</td><td><code>B2238C1</code></td></tr>
+<tr><td>12</td><td>1,000,000,000,001</td><td><code>0FEE563</code></td></tr>
+<tr><td>13</td><td>10,000,000,000,001</td><td><code>896DC3</code></td></tr>
+
+<tr><td>14</td><td>100,000,000,000,001</td><td><code>C216EC</code></td></tr>
+<tr><td>15</td><td>1,000,000,000,000,001</td><td><code>6216B06</code> ... <code>D3611</code></td></tr>
 </table>
 </table>
 
 
 <sup>*</sup>
 <sup>*</sup>
 By representing &pi; in decimal, hexadecimal and binary, we have
 By representing &pi; in decimal, hexadecimal and binary, we have
 
 
-<table summary="Pi in various formats"><tr>
-  <td>&pi;</td><td>=</td><td><tt>3.1415926535 8979323846 2643383279</tt> ...</td>
+<table><caption>"Pi in various formats"</caption><tr>
+  <td>&pi;</td><td>=</td><td><code>3.1415926535 8979323846 2643383279</code> ...</td>
 </tr><tr>
 </tr><tr>
-  <td></td><td>=</td><td><tt>3.243F6A8885 A308D31319 8A2E037073</tt> ...</td>
+  <td></td><td>=</td><td><code>3.243F6A8885 A308D31319 8A2E037073</code> ...</td>
 </tr><tr>
 </tr><tr>
-  <td></td><td>=</td><td><tt>11.0010010000 1111110110 1010100010</tt> ...</td>
+  <td></td><td>=</td><td><code>11.0010010000 1111110110 1010100010</code> ...</td>
 </tr></table>
 </tr></table>
-The first ten bits of &pi; are <tt>0010010000</tt>.
+The first ten bits of &pi; are <code>0010010000</code>.
 </td></tr></table>
 </td></tr></table>
 
 
 
 
@@ -130,7 +132,8 @@ The first ten bits of &pi; are <tt>0010010000</tt>.
 $ hadoop org.apache.hadoop.examples.pi.DistBbp \
 $ hadoop org.apache.hadoop.examples.pi.DistBbp \
          &lt;b&gt; &lt;nThreads&gt; &lt;nJobs&gt; &lt;type&gt; &lt;nPart&gt; &lt;remoteDir&gt; &lt;localDir&gt;</pre>
          &lt;b&gt; &lt;nThreads&gt; &lt;nJobs&gt; &lt;type&gt; &lt;nPart&gt; &lt;remoteDir&gt; &lt;localDir&gt;</pre>
   And the parameters are:
   And the parameters are:
-  <table summary="command line option">
+  <table>
+    <caption>"command line option"</caption>
     <tr>
     <tr>
       <td>&lt;b&gt;</td>
       <td>&lt;b&gt;</td>
       <td>The number of bits to skip, i.e. compute the (b+1)th position.</td>
       <td>The number of bits to skip, i.e. compute the (b+1)th position.</td>

+ 0 - 23
hadoop-project/pom.xml

@@ -158,7 +158,6 @@
 
 
     <!-- define the Java language version used by the compiler -->
     <!-- define the Java language version used by the compiler -->
     <javac.version>1.8</javac.version>
     <javac.version>1.8</javac.version>
-    <javadoc.skip.jdk11>false</javadoc.skip.jdk11>
 
 
     <!-- The java version enforced by the maven enforcer -->
     <!-- The java version enforced by the maven enforcer -->
     <!-- more complex patterns can be used here, such as
     <!-- more complex patterns can be used here, such as
@@ -2719,28 +2718,6 @@
         </dependencies>
         </dependencies>
       </dependencyManagement>
       </dependencyManagement>
     </profile>
     </profile>
-    <profile>
-      <id>jdk11</id>
-      <activation>
-        <jdk>[11,)</jdk>
-      </activation>
-      <build>
-        <plugins>
-          <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-javadoc-plugin</artifactId>
-            <configuration>
-              <skip>${javadoc.skip.jdk11}</skip>
-              <detectJavaApiLink>false</detectJavaApiLink>
-              <additionalOptions>
-                <!-- TODO: remove -html4 option to generate html5 docs when we stop supporting JDK8 -->
-                <additionalOption>-html4</additionalOption>
-              </additionalOptions>
-            </configuration>
-          </plugin>
-        </plugins>
-      </build>
-    </profile>
   </profiles>
   </profiles>
 
 
   <repositories>
   <repositories>

+ 0 - 1
hadoop-tools/hadoop-aws/pom.xml

@@ -35,7 +35,6 @@
     <file.encoding>UTF-8</file.encoding>
     <file.encoding>UTF-8</file.encoding>
     <downloadSources>true</downloadSources>
     <downloadSources>true</downloadSources>
     <hadoop.tmp.dir>${project.build.directory}/test</hadoop.tmp.dir>
     <hadoop.tmp.dir>${project.build.directory}/test</hadoop.tmp.dir>
-    <javadoc.skip.jdk11>true</javadoc.skip.jdk11>
 
 
     <!-- are scale tests enabled ? -->
     <!-- are scale tests enabled ? -->
     <fs.s3a.scale.test.enabled>unset</fs.s3a.scale.test.enabled>
     <fs.s3a.scale.test.enabled>unset</fs.s3a.scale.test.enabled>

+ 1 - 1
hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/FilePool.java

@@ -128,7 +128,7 @@ class FilePool {
 
 
     /**
     /**
      * Return a set of files whose cumulative size is at least
      * Return a set of files whose cumulative size is at least
-     * <tt>targetSize</tt>.
+     * <code>targetSize</code>.
      * TODO Clearly size is not the only criterion, e.g. refresh from
      * TODO Clearly size is not the only criterion, e.g. refresh from
      * generated data without including running task output, tolerance
      * generated data without including running task output, tolerance
      * for permission issues, etc.
      * for permission issues, etc.

+ 1 - 1
hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/io/IdentifierResolver.java

@@ -26,7 +26,7 @@ import org.apache.hadoop.typedbytes.TypedBytesWritable;
 /**
 /**
  * This class is used to resolve a string identifier into the required IO
  * This class is used to resolve a string identifier into the required IO
  * classes. By extending this class and pointing the property
  * classes. By extending this class and pointing the property
- * <tt>stream.io.identifier.resolver.class</tt> to this extension, additional
+ * <code>stream.io.identifier.resolver.class</code> to this extension, additional
  * IO classes can be added by external code.
  * IO classes can be added by external code.
  */
  */
 public class IdentifierResolver {
 public class IdentifierResolver {

+ 1 - 1
hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/streaming/package.html

@@ -19,7 +19,7 @@
 
 
 <body>
 <body>
 
 
-<tt>Hadoop Streaming</tt> is a utility which allows users to create and run 
+<code>Hadoop Streaming</code> is a utility which allows users to create and run
 Map-Reduce jobs with any executables (e.g. Unix shell utilities) as the mapper 
 Map-Reduce jobs with any executables (e.g. Unix shell utilities) as the mapper 
 and/or the reducer.
 and/or the reducer.
 
 

+ 5 - 3
hadoop-tools/hadoop-streaming/src/main/java/org/apache/hadoop/typedbytes/package.html

@@ -22,11 +22,12 @@
 Typed bytes are sequences of bytes in which the first byte is a type code. They are especially useful as a 
 Typed bytes are sequences of bytes in which the first byte is a type code. They are especially useful as a 
 (simple and very straightforward) binary format for transferring data to and from Hadoop Streaming programs.
 (simple and very straightforward) binary format for transferring data to and from Hadoop Streaming programs.
 
 
-<h3>Type Codes</h3>
+<h2>Type Codes</h2>
 
 
 Each typed bytes sequence starts with an unsigned byte that contains the type code. Possible values are:
 Each typed bytes sequence starts with an unsigned byte that contains the type code. Possible values are:
 
 
-<table border="1" cellpadding="2" summary="Type Codes">
+<table border="1">
+<caption>"Type Codes"</caption>
 <tr><th>Code</th><th>Type</th></tr>
 <tr><th>Code</th><th>Type</th></tr>
 <tr><td><i>0</i></td><td>A sequence of bytes.</td></tr>
 <tr><td><i>0</i></td><td>A sequence of bytes.</td></tr>
 <tr><td><i>1</i></td><td>A byte.</td></tr>
 <tr><td><i>1</i></td><td>A byte.</td></tr>
@@ -48,7 +49,8 @@ application-specific serialization.
 
 
 These are the subsequent bytes for the different type codes (everything is big-endian and unpadded):
 These are the subsequent bytes for the different type codes (everything is big-endian and unpadded):
 
 
-<table border="1" cellpadding="2" summary="Subsequent Bytes">
+<table border="1">
+<caption>"Subsequent Bytes"</caption>
 <tr><th>Code</th><th>Subsequent Bytes</th></tr>
 <tr><th>Code</th><th>Subsequent Bytes</th></tr>
 <tr><td><i>0</i></td><td>&lt;32-bit signed integer&gt; &lt;as many bytes as indicated by the integer&gt;</td></tr>
 <tr><td><i>0</i></td><td>&lt;32-bit signed integer&gt; &lt;as many bytes as indicated by the integer&gt;</td></tr>
 <tr><td><i>1</i></td><td>&lt;signed byte&gt;</td></tr>
 <tr><td><i>1</i></td><td>&lt;signed byte&gt;</td></tr>

+ 1 - 1
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/protocolrecords/SignalContainerRequest.java

@@ -29,7 +29,7 @@ import org.apache.hadoop.yarn.util.Records;
  * <p>The request sent by the client to the <code>ResourceManager</code>
  * <p>The request sent by the client to the <code>ResourceManager</code>
  * or by the <code>ApplicationMaster</code> to the <code>NodeManager</code>
  * or by the <code>ApplicationMaster</code> to the <code>NodeManager</code>
  * to signal a container.
  * to signal a container.
- * @see SignalContainerCommand </p>
+ * @see SignalContainerCommand
  */
  */
 @Public
 @Public
 @Evolving
 @Evolving

+ 1 - 1
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceMetricsSink.java

@@ -31,7 +31,7 @@ import org.slf4j.LoggerFactory;
  * adding the following to by This would actually be set as: <code>
  * adding the following to by This would actually be set as: <code>
  * [prefix].sink.[some instance name].class
  * [prefix].sink.[some instance name].class
  * =org.apache.hadoop.yarn.service.timelineservice.ServiceMetricsSink
  * =org.apache.hadoop.yarn.service.timelineservice.ServiceMetricsSink
- * </code>, where <tt>prefix</tt> is "atsv2": and <tt>some instance name</tt> is
+ * </code>, where <code>prefix</code> is "atsv2": and <code>some instance name</code> is
  * just any unique name, so properties can be differentiated if there are
  * just any unique name, so properties can be differentiated if there are
  * multiple sinks of the same type created
  * multiple sinks of the same type created
  */
  */

+ 2 - 2
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/AdminACLsManager.java

@@ -93,7 +93,7 @@ public class AdminACLsManager {
    *
    *
    * @see YarnConfiguration#YARN_ACL_ENABLE
    * @see YarnConfiguration#YARN_ACL_ENABLE
    * @see YarnConfiguration#DEFAULT_YARN_ACL_ENABLE
    * @see YarnConfiguration#DEFAULT_YARN_ACL_ENABLE
-   * @return <tt>true</tt> if ACLs are enabled
+   * @return <code>true</code> if ACLs are enabled
    */
    */
   public boolean areACLsEnabled() {
   public boolean areACLsEnabled() {
     return aclsEnabled;
     return aclsEnabled;
@@ -103,7 +103,7 @@ public class AdminACLsManager {
    * Returns whether the specified user/group is an administrator
    * Returns whether the specified user/group is an administrator
    *
    *
    * @param callerUGI user/group to to check
    * @param callerUGI user/group to to check
-   * @return <tt>true</tt> if the UserGroupInformation specified
+   * @return <code>true</code> if the UserGroupInformation specified
    *         is a member of the access control list for administrators
    *         is a member of the access control list for administrators
    */
    */
   public boolean isAdmin(UserGroupInformation callerUGI) {
   public boolean isAdmin(UserGroupInformation callerUGI) {

+ 1 - 1
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/BoundedAppender.java

@@ -56,7 +56,7 @@ import org.apache.hadoop.classification.InterfaceStability;
  * }
  * }
  * </pre>
  * </pre>
  * <p>
  * <p>
- * Note that <tt>null</tt> values are {@link #append(CharSequence) append}ed
+ * Note that <code>null</code> values are {@link #append(CharSequence) append}ed
  * just like in {@link StringBuilder#append(CharSequence) original
  * just like in {@link StringBuilder#append(CharSequence) original
  * implementation}.
  * implementation}.
  * <p>
  * <p>

+ 1 - 1
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common/src/main/java/org/apache/hadoop/yarn/server/utils/LeveldbIterator.java

@@ -112,7 +112,7 @@ public class LeveldbIterator implements Iterator<Map.Entry<byte[], byte[]>>,
   }
   }
 
 
   /**
   /**
-   * Returns <tt>true</tt> if the iteration has more elements.
+   * Returns <code>true</code> if the iteration has more elements.
    */
    */
   public boolean hasNext() throws DBException {
   public boolean hasNext() throws DBException {
     try {
     try {

+ 1 - 1
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/src/main/java/org/apache/hadoop/yarn/server/timelineservice/storage/common/BaseTable.java

@@ -21,7 +21,7 @@ package org.apache.hadoop.yarn.server.timelineservice.storage.common;
 
 
 /**
 /**
  * The base type of tables.
  * The base type of tables.
- * @param T table type
+ * @param <T> table type
  */
  */
 public abstract class BaseTable<T> {
 public abstract class BaseTable<T> {
 }
 }