|
@@ -21,6 +21,8 @@ package org.apache.hadoop.mapreduce.split;
|
|
|
import java.io.IOException;
|
|
|
import java.util.Arrays;
|
|
|
|
|
|
+import org.apache.hadoop.classification.InterfaceAudience;
|
|
|
+import org.apache.hadoop.classification.InterfaceStability;
|
|
|
import org.apache.hadoop.conf.Configuration;
|
|
|
import org.apache.hadoop.fs.FSDataInputStream;
|
|
|
import org.apache.hadoop.fs.FileStatus;
|
|
@@ -29,9 +31,7 @@ import org.apache.hadoop.fs.Path;
|
|
|
import org.apache.hadoop.io.WritableUtils;
|
|
|
import org.apache.hadoop.mapreduce.JobID;
|
|
|
import org.apache.hadoop.mapreduce.JobSubmissionFiles;
|
|
|
-import org.apache.hadoop.mapreduce.server.jobtracker.JTConfig;
|
|
|
-import org.apache.hadoop.classification.InterfaceAudience;
|
|
|
-import org.apache.hadoop.classification.InterfaceStability;
|
|
|
+import org.apache.hadoop.mapreduce.MRJobConfig;
|
|
|
|
|
|
/**
|
|
|
* A utility that reads the split meta info and creates
|
|
@@ -44,8 +44,8 @@ public class SplitMetaInfoReader {
|
|
|
public static JobSplit.TaskSplitMetaInfo[] readSplitMetaInfo(
|
|
|
JobID jobId, FileSystem fs, Configuration conf, Path jobSubmitDir)
|
|
|
throws IOException {
|
|
|
- long maxMetaInfoSize = conf.getLong(JTConfig.JT_MAX_JOB_SPLIT_METAINFO_SIZE,
|
|
|
- 10000000L);
|
|
|
+ long maxMetaInfoSize = conf.getLong(MRJobConfig.SPLIT_METAINFO_MAXSIZE,
|
|
|
+ MRJobConfig.DEFAULT_SPLIT_METAINFO_MAXSIZE);
|
|
|
Path metaSplitFile = JobSubmissionFiles.getJobSplitMetaFile(jobSubmitDir);
|
|
|
String jobSplitFile = JobSubmissionFiles.getJobSplitFile(jobSubmitDir).toString();
|
|
|
FileStatus fStatus = fs.getFileStatus(metaSplitFile);
|