Explorar o código

MAPREDUCE-6655. Fix a typo (STRICT_IE6) in Encrypted Shuffle. Contributed by Wei-Chiu Chuang.
(cherry picked from commit 90fcb16c0fc7cded2593948f1cc08f3d1ae88bda)

Junping Du %!s(int64=9) %!d(string=hai) anos
pai
achega
ae29b1a17b

+ 1 - 1
hadoop-common-project/hadoop-common/src/main/resources/core-default.xml

@@ -1579,7 +1579,7 @@
   <value>DEFAULT</value>
   <value>DEFAULT</value>
   <description>
   <description>
     The hostname verifier to provide for HttpsURLConnections.
     The hostname verifier to provide for HttpsURLConnections.
-    Valid values are: DEFAULT, STRICT, STRICT_I6, DEFAULT_AND_LOCALHOST and
+    Valid values are: DEFAULT, STRICT, STRICT_IE6, DEFAULT_AND_LOCALHOST and
     ALLOW_ALL
     ALLOW_ALL
   </description>
   </description>
 </property>
 </property>

+ 2 - 2
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/site/markdown/EncryptedShuffle.md

@@ -40,7 +40,7 @@ To enable encrypted shuffle, set the following properties in core-site.xml of al
 | **Property** | **Default Value** | **Explanation** |
 | **Property** | **Default Value** | **Explanation** |
 |:---- |:---- |:---- |
 |:---- |:---- |:---- |
 | `hadoop.ssl.require.client.cert` | `false` | Whether client certificates are required |
 | `hadoop.ssl.require.client.cert` | `false` | Whether client certificates are required |
-| `hadoop.ssl.hostname.verifier` | `DEFAULT` | The hostname verifier to provide for HttpsURLConnections. Valid values are: **DEFAULT**, **STRICT**, **STRICT\_I6**, **DEFAULT\_AND\_LOCALHOST** and **ALLOW\_ALL** |
+| `hadoop.ssl.hostname.verifier` | `DEFAULT` | The hostname verifier to provide for HttpsURLConnections. Valid values are: **DEFAULT**, **STRICT**, **STRICT\_IE6**, **DEFAULT\_AND\_LOCALHOST** and **ALLOW\_ALL** |
 | `hadoop.ssl.keystores.factory.class` | `org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory` | The KeyStoresFactory implementation to use |
 | `hadoop.ssl.keystores.factory.class` | `org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory` | The KeyStoresFactory implementation to use |
 | `hadoop.ssl.server.conf` | `ssl-server.xml` | Resource file from which ssl server keystore information will be extracted. This file is looked up in the classpath, typically it should be in Hadoop conf/ directory |
 | `hadoop.ssl.server.conf` | `ssl-server.xml` | Resource file from which ssl server keystore information will be extracted. This file is looked up in the classpath, typically it should be in Hadoop conf/ directory |
 | `hadoop.ssl.client.conf` | `ssl-client.xml` | Resource file from which ssl server keystore information will be extracted. This file is looked up in the classpath, typically it should be in Hadoop conf/ directory |
 | `hadoop.ssl.client.conf` | `ssl-client.xml` | Resource file from which ssl server keystore information will be extracted. This file is looked up in the classpath, typically it should be in Hadoop conf/ directory |
@@ -260,4 +260,4 @@ Encrypted Intermediate Data Spill files
 This capability allows encryption of the intermediate files generated during the merge and shuffle phases.
 This capability allows encryption of the intermediate files generated during the merge and shuffle phases.
 It can be enabled by setting the `mapreduce.job.encrypted-intermediate-data` job property to `true`.
 It can be enabled by setting the `mapreduce.job.encrypted-intermediate-data` job property to `true`.
 
 
-**NOTE:** Currently, enabling encrypted intermediate data spills would restrict the number of attempts of the job to 1.
+**NOTE:** Currently, enabling encrypted intermediate data spills would restrict the number of attempts of the job to 1.