Browse Source

MAPREDUCE-7132. JobSplitWriter prints unnecessary warnings if EC(RS10,4) is used. Contributed by Peter Bacsko.

(cherry picked from commit 25f8fcb06476938826cdc92858a61124b18cd98d)
Xiao Chen 6 years ago
parent
commit
424c99681b

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRConfig.java

@@ -78,7 +78,7 @@ public interface MRConfig {
     "mapreduce.task.max.status.length";
     "mapreduce.task.max.status.length";
   public static final int PROGRESS_STATUS_LEN_LIMIT_DEFAULT = 512;
   public static final int PROGRESS_STATUS_LEN_LIMIT_DEFAULT = 512;
 
 
-  public static final int MAX_BLOCK_LOCATIONS_DEFAULT = 10;
+  public static final int MAX_BLOCK_LOCATIONS_DEFAULT = 15;
   public static final String MAX_BLOCK_LOCATIONS_KEY =
   public static final String MAX_BLOCK_LOCATIONS_KEY =
     "mapreduce.job.max.split.locations";
     "mapreduce.job.max.split.locations";
 
 

+ 1 - 1
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/resources/mapred-default.xml

@@ -149,7 +149,7 @@
 
 
   <property>
   <property>
     <name>mapreduce.job.max.split.locations</name>
     <name>mapreduce.job.max.split.locations</name>
-    <value>10</value>
+    <value>15</value>
     <description>The max number of block locations to store for each split for
     <description>The max number of block locations to store for each split for
     locality calculation.
     locality calculation.
     </description>
     </description>

+ 128 - 0
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/test/java/org/apache/hadoop/mapreduce/split/TestJobSplitWriterWithEC.java

@@ -0,0 +1,128 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.mapreduce.split;
+
+import static org.junit.Assert.assertEquals;
+
+import java.io.File;
+import java.io.IOException;
+import java.util.List;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.DFSConfigKeys;
+import org.apache.hadoop.hdfs.DFSTestUtil;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
+import org.apache.hadoop.hdfs.HdfsConfiguration;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.StripedFileTestUtil;
+import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
+import org.apache.hadoop.hdfs.protocol.SystemErasureCodingPolicies;
+import org.apache.hadoop.mapred.JobConf;
+import org.apache.hadoop.mapreduce.InputSplit;
+import org.apache.hadoop.mapreduce.Job;
+import org.apache.hadoop.mapreduce.JobID;
+import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
+import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Test;
+
+
+/**
+ * Tests that maxBlockLocations default value is sufficient for RS-10-4.
+ */
+public class TestJobSplitWriterWithEC {
+  // This will ensure 14 block locations
+  private ErasureCodingPolicy ecPolicy = SystemErasureCodingPolicies
+      .getByID(SystemErasureCodingPolicies.RS_10_4_POLICY_ID);
+  private static final int BLOCKSIZE = 1024 * 1024 * 10;
+
+  private MiniDFSCluster cluster;
+  private DistributedFileSystem fs;
+  private Configuration conf;
+  private Path submitDir;
+  private Path testFile;
+
+  @Before
+  public void setup() throws Exception {
+    Configuration hdfsConf = new HdfsConfiguration();
+    hdfsConf.setLong(DFSConfigKeys.DFS_NAMENODE_MIN_BLOCK_SIZE_KEY, 0);
+    String namenodeDir = new File(MiniDFSCluster.getBaseDirectory(), "name").
+        getAbsolutePath();
+    hdfsConf.set(DFSConfigKeys.DFS_NAMENODE_NAME_DIR_KEY, namenodeDir);
+    hdfsConf.set(DFSConfigKeys.DFS_NAMENODE_EDITS_DIR_KEY, namenodeDir);
+    hdfsConf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, BLOCKSIZE);
+    cluster = new MiniDFSCluster.Builder(hdfsConf).numDataNodes(15).build();
+    fs = cluster.getFileSystem();
+    fs.enableErasureCodingPolicy(ecPolicy.getName());
+    fs.setErasureCodingPolicy(new Path("/"), ecPolicy.getName());
+    cluster.waitActive();
+
+    conf = new Configuration();
+    submitDir = new Path("/");
+    testFile = new Path("/testfile");
+    DFSTestUtil.writeFile(fs, testFile,
+        StripedFileTestUtil.generateBytes(BLOCKSIZE));
+    conf.set(FileInputFormat.INPUT_DIR,
+        fs.getUri().toString() + testFile.toString());
+  }
+
+  @After
+  public void after() {
+    cluster.close();
+  }
+
+  @Test
+  public void testMaxBlockLocationsNewSplitsWithErasureCoding()
+      throws Exception {
+    Job job = Job.getInstance(conf);
+    final FileInputFormat<?, ?> fileInputFormat = new TextInputFormat();
+    final List<InputSplit> splits = fileInputFormat.getSplits(job);
+
+    JobSplitWriter.createSplitFiles(submitDir, conf, fs, splits);
+
+    validateSplitMetaInfo();
+  }
+
+  @Test
+  public void testMaxBlockLocationsOldSplitsWithErasureCoding()
+      throws Exception {
+    JobConf jobConf = new JobConf(conf);
+    org.apache.hadoop.mapred.TextInputFormat fileInputFormat
+        = new org.apache.hadoop.mapred.TextInputFormat();
+    fileInputFormat.configure(jobConf);
+    final org.apache.hadoop.mapred.InputSplit[] splits =
+        fileInputFormat.getSplits(jobConf, 1);
+
+    JobSplitWriter.createSplitFiles(submitDir, conf, fs, splits);
+
+    validateSplitMetaInfo();
+  }
+
+  private void validateSplitMetaInfo() throws IOException {
+    JobSplit.TaskSplitMetaInfo[] splitInfo =
+        SplitMetaInfoReader.readSplitMetaInfo(new JobID(), fs, conf,
+            submitDir);
+
+    assertEquals("Number of splits", 1, splitInfo.length);
+    assertEquals("Number of block locations", 14,
+        splitInfo[0].getLocations().length);
+  }
+}