Modifier and Type | Class and Description |
---|---|
static class |
JobImpl.InitTransition |
static class |
JobImpl.StartTransition |
Modifier and Type | Field and Description |
---|---|
org.apache.hadoop.mapred.JobConf |
conf |
static String |
JOB_KILLED_DIAG |
org.apache.hadoop.fs.Path |
remoteJobConfFile |
protected static org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> |
stateMachineFactory |
Constructor and Description |
---|
JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId,
org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.yarn.event.EventHandler eventHandler,
TaskAttemptListener taskAttemptListener,
org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager,
org.apache.hadoop.security.Credentials jobCredentials,
org.apache.hadoop.yarn.util.Clock clock,
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun,
MRAppMetrics metrics,
org.apache.hadoop.mapreduce.OutputCommitter committer,
boolean newApiCommitter,
String userName,
long appSubmitTime,
List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos,
AppContext appContext,
JobStateInternal forcedState,
String forcedDiagnostic) |
Modifier and Type | Method and Description |
---|---|
protected void |
addDiagnostic(String diag) |
protected void |
addTask(Task task) |
boolean |
checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI,
org.apache.hadoop.mapreduce.JobACL jobOperation) |
protected JobStateInternal |
checkReadyForCommit() |
void |
constructFinalFullcounters() |
static String |
escapeString(String data) |
org.apache.hadoop.mapreduce.Counters |
getAllCounters()
Get all the counters of this job.
|
List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> |
getAMInfos() |
int |
getCompletedMaps() |
int |
getCompletedReduces() |
org.apache.hadoop.fs.Path |
getConfFile() |
List<String> |
getDiagnostics() |
int |
getFailedMaps() |
int |
getFailedReduces() |
protected org.apache.hadoop.fs.FileSystem |
getFileSystem(org.apache.hadoop.conf.Configuration conf)
Create the default file System for this job.
|
org.apache.hadoop.mapreduce.v2.api.records.JobId |
getID() |
JobStateInternal |
getInternalState() |
Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> |
getJobACLs() |
int |
getKilledMaps() |
int |
getKilledReduces() |
org.apache.hadoop.mapred.TaskCompletionEvent[] |
getMapAttemptCompletionEvents(int startIndex,
int maxEvents) |
float |
getMaxAllowedFetchFailuresFraction() |
int |
getMaxFetchFailuresNotifications() |
String |
getName() |
float |
getProgress() |
String |
getQueueName() |
org.apache.hadoop.mapreduce.v2.api.records.JobReport |
getReport() |
org.apache.hadoop.mapreduce.v2.api.records.JobState |
getState() |
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> |
getStateMachine() |
Task |
getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID) |
org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] |
getTaskAttemptCompletionEvents(int fromEventId,
int maxEvents) |
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks() |
Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> |
getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType) |
int |
getTotalMaps() |
int |
getTotalReduces() |
String |
getUserName() |
void |
handle(JobEvent event) |
static org.apache.hadoop.mapreduce.Counters |
incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters,
Collection<Task> tasks) |
boolean |
isUber() |
org.apache.hadoop.conf.Configuration |
loadConfFile() |
protected void |
scheduleTasks(Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs,
boolean recoverTaskOutput) |
void |
setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority) |
void |
setQueueName(String queueName) |
public static final String JOB_KILLED_DIAG
public org.apache.hadoop.mapred.JobConf conf
public org.apache.hadoop.fs.Path remoteJobConfFile
protected static final org.apache.hadoop.yarn.state.StateMachineFactory<JobImpl,JobStateInternal,JobEventType,JobEvent> stateMachineFactory
public JobImpl(org.apache.hadoop.mapreduce.v2.api.records.JobId jobId, org.apache.hadoop.yarn.api.records.ApplicationAttemptId applicationAttemptId, org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.yarn.event.EventHandler eventHandler, TaskAttemptListener taskAttemptListener, org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager jobTokenSecretManager, org.apache.hadoop.security.Credentials jobCredentials, org.apache.hadoop.yarn.util.Clock clock, Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo> completedTasksFromPreviousRun, MRAppMetrics metrics, org.apache.hadoop.mapreduce.OutputCommitter committer, boolean newApiCommitter, String userName, long appSubmitTime, List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> amInfos, AppContext appContext, JobStateInternal forcedState, String forcedDiagnostic)
protected org.apache.hadoop.yarn.state.StateMachine<JobStateInternal,JobEventType,JobEvent> getStateMachine()
public org.apache.hadoop.mapreduce.v2.api.records.JobId getID()
public boolean checkAccess(org.apache.hadoop.security.UserGroupInformation callerUGI, org.apache.hadoop.mapreduce.JobACL jobOperation)
checkAccess
in interface Job
public Task getTask(org.apache.hadoop.mapreduce.v2.api.records.TaskId taskID)
public int getCompletedMaps()
getCompletedMaps
in interface Job
public int getCompletedReduces()
getCompletedReduces
in interface Job
public org.apache.hadoop.mapreduce.Counters getAllCounters()
Job
getAllCounters
in interface Job
public static org.apache.hadoop.mapreduce.Counters incrTaskCounters(org.apache.hadoop.mapreduce.Counters counters, Collection<Task> tasks)
public org.apache.hadoop.mapreduce.v2.api.records.TaskAttemptCompletionEvent[] getTaskAttemptCompletionEvents(int fromEventId, int maxEvents)
getTaskAttemptCompletionEvents
in interface Job
public org.apache.hadoop.mapred.TaskCompletionEvent[] getMapAttemptCompletionEvents(int startIndex, int maxEvents)
getMapAttemptCompletionEvents
in interface Job
public List<String> getDiagnostics()
getDiagnostics
in interface Job
public org.apache.hadoop.mapreduce.v2.api.records.JobReport getReport()
public float getProgress()
getProgress
in interface Job
public Map<org.apache.hadoop.mapreduce.v2.api.records.TaskId,Task> getTasks(org.apache.hadoop.mapreduce.v2.api.records.TaskType taskType)
public org.apache.hadoop.mapreduce.v2.api.records.JobState getState()
protected void scheduleTasks(Set<org.apache.hadoop.mapreduce.v2.api.records.TaskId> taskIDs, boolean recoverTaskOutput)
public void handle(JobEvent event)
handle
in interface org.apache.hadoop.yarn.event.EventHandler<JobEvent>
@InterfaceAudience.Private public JobStateInternal getInternalState()
protected void addTask(Task task)
protected org.apache.hadoop.fs.FileSystem getFileSystem(org.apache.hadoop.conf.Configuration conf) throws IOException
conf
- the conf objectIOException
protected JobStateInternal checkReadyForCommit()
public String getUserName()
getUserName
in interface Job
public String getQueueName()
getQueueName
in interface Job
public void setQueueName(String queueName)
setQueueName
in interface Job
public org.apache.hadoop.fs.Path getConfFile()
getConfFile
in interface Job
public int getTotalMaps()
getTotalMaps
in interface Job
public int getTotalReduces()
getTotalReduces
in interface Job
public Map<org.apache.hadoop.mapreduce.JobACL,org.apache.hadoop.security.authorize.AccessControlList> getJobACLs()
getJobACLs
in interface Job
public List<org.apache.hadoop.mapreduce.v2.api.records.AMInfo> getAMInfos()
getAMInfos
in interface Job
@InterfaceAudience.Private public void constructFinalFullcounters()
protected void addDiagnostic(String diag)
public org.apache.hadoop.conf.Configuration loadConfFile() throws IOException
loadConfFile
in interface Job
Job.getConfFile()
.IOException
- on any error trying to load the conf file.public float getMaxAllowedFetchFailuresFraction()
public int getMaxFetchFailuresNotifications()
public void setJobPriority(org.apache.hadoop.yarn.api.records.Priority priority)
setJobPriority
in interface Job
public int getFailedMaps()
getFailedMaps
in interface Job
public int getFailedReduces()
getFailedReduces
in interface Job
public int getKilledMaps()
getKilledMaps
in interface Job
public int getKilledReduces()
getKilledReduces
in interface Job
Copyright © 2008–2024 Apache Software Foundation. All rights reserved.