|
@@ -34,7 +34,6 @@ import org.apache.hadoop.mapreduce.test.system.TaskInfo;
|
|
|
import org.apache.hadoop.mapreduce.test.system.TTClient;
|
|
|
import org.apache.hadoop.mapreduce.test.system.JTClient;
|
|
|
import org.apache.hadoop.mapreduce.test.system.FinishTaskControlAction;
|
|
|
-import org.apache.hadoop.mapred.JobClient.NetworkedJob;
|
|
|
import org.apache.hadoop.io.NullWritable;
|
|
|
import org.apache.hadoop.io.IntWritable;
|
|
|
import org.apache.hadoop.io.Text;
|
|
@@ -100,7 +99,7 @@ public class TestTaskKilling {
|
|
|
jtClient.isTaskStarted(taskInfo));
|
|
|
|
|
|
// Fail the running task.
|
|
|
- NetworkedJob networkJob = jobClient.new NetworkedJob(jInfo.getStatus());
|
|
|
+ RunningJob networkJob = jobClient.getJob(jobId);
|
|
|
TaskID tID = TaskID.downgrade(taskInfo.getTaskID());
|
|
|
TaskAttemptID taskAttID = new TaskAttemptID(tID , 0);
|
|
|
networkJob.killTask(taskAttID, true);
|
|
@@ -206,7 +205,7 @@ public class TestTaskKilling {
|
|
|
taskAttID + " has not been found while task was running.",
|
|
|
isTempFolderExists);
|
|
|
|
|
|
- NetworkedJob networkJob = jobClient.new NetworkedJob(jInfo.getStatus());
|
|
|
+ RunningJob networkJob = jobClient.getJob(id);
|
|
|
networkJob.killTask(taskAttID, false);
|
|
|
ttClient.getProxy().sendAction(action);
|
|
|
taskInfo = remoteJTClient.getTaskInfo(tID);
|
|
@@ -353,8 +352,7 @@ public class TestTaskKilling {
|
|
|
TaskAttemptID tAttID = new TaskAttemptID(taskId,
|
|
|
taskInfo.numFailedAttempts());
|
|
|
while(taskInfo.numFailedAttempts() < MAX_MAP_TASK_ATTEMPTS) {
|
|
|
- NetworkedJob networkJob = jtClient.getClient().
|
|
|
- new NetworkedJob(jobInfo.getStatus());
|
|
|
+ RunningJob networkJob = jobClient.getJob(id);
|
|
|
networkJob.killTask(taskAttID, true);
|
|
|
taskInfo = remoteJTClient.getTaskInfo(taskInfo.getTaskID());
|
|
|
taskAttID = new TaskAttemptID(taskId, taskInfo.numFailedAttempts());
|
|
@@ -484,8 +482,7 @@ public class TestTaskKilling {
|
|
|
taskIdKilled = taskid.toString();
|
|
|
taskAttemptID = new TaskAttemptID(taskid, i);
|
|
|
LOG.info("taskAttemptid going to be killed is : " + taskAttemptID);
|
|
|
- (jobClient.new NetworkedJob(jInfo.getStatus())).
|
|
|
- killTask(taskAttemptID,true);
|
|
|
+ rJob.killTask(taskAttemptID,true);
|
|
|
checkTaskCompletionEvent(taskAttemptID, jInfo);
|
|
|
break;
|
|
|
} else {
|
|
@@ -495,8 +492,7 @@ public class TestTaskKilling {
|
|
|
UtilsForTests.waitFor(20000);
|
|
|
LOG.info("taskAttemptid going to be killed is : " +
|
|
|
taskAttemptID);
|
|
|
- (jobClient.new NetworkedJob(jInfo.getStatus())).
|
|
|
- killTask(taskAttemptID,true);
|
|
|
+ rJob.killTask(taskAttemptID,true);
|
|
|
checkTaskCompletionEvent(taskAttemptID,jInfo);
|
|
|
break;
|
|
|
}
|
|
@@ -536,8 +532,10 @@ public class TestTaskKilling {
|
|
|
boolean match = false;
|
|
|
int count = 0;
|
|
|
while (!match) {
|
|
|
- TaskCompletionEvent[] taskCompletionEvents = jobClient.new
|
|
|
- NetworkedJob(jInfo.getStatus()).getTaskCompletionEvents(0);
|
|
|
+ org.apache.hadoop.mapreduce.JobID temp = jInfo.getID();
|
|
|
+ RunningJob rJob = jobClient.getJob(new JobID(temp.getJtIdentifier(), temp.getId()));
|
|
|
+
|
|
|
+ TaskCompletionEvent[] taskCompletionEvents = rJob.getTaskCompletionEvents(0);
|
|
|
for (TaskCompletionEvent taskCompletionEvent : taskCompletionEvents) {
|
|
|
LOG.info("taskCompletionEvent.getTaskAttemptId().toString() is : " +
|
|
|
taskCompletionEvent.getTaskAttemptId().toString());
|