| Modifier and Type | Interface and Description | 
|---|---|
| interface  | JobContext | 
| Modifier and Type | Method and Description | 
|---|---|
| void | OutputCommitter. abortJob(JobContext context,
                org.apache.hadoop.mapreduce.JobStatus.State runState)This method implements the new interface by calling the old method. | 
| void | OutputCommitter. cleanupJob(JobContext context) | 
| void | OutputCommitter. commitJob(JobContext context)This method implements the new interface by calling the old method. | 
| boolean | OutputCommitter. isRecoverySupported(JobContext context)This method implements the new interface by calling the old method. | 
| void | OutputCommitter. setupJob(JobContext jobContext)This method implements the new interface by calling the old method. | 
| Modifier and Type | Method and Description | 
|---|---|
| protected boolean | CombineFileInputFormat. isSplitable(JobContext context,
                      Path file)Subclasses should avoid overriding this method and should instead only
 override  CombineFileInputFormat.isSplitable(FileSystem, Path). | 
| Modifier and Type | Interface and Description | 
|---|---|
| interface  | MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>The context that is given to the  Mapper. | 
| interface  | ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>The context passed to the  Reducer. | 
| interface  | TaskAttemptContextThe context for task attempts. | 
| interface  | TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>A context object that allows input and output from the task. | 
| Modifier and Type | Class and Description | 
|---|---|
| class  | JobThe job submitter's view of the Job. | 
| Modifier and Type | Method and Description | 
|---|---|
| void | OutputCommitter. abortJob(JobContext jobContext,
                org.apache.hadoop.mapreduce.JobStatus.State state)For aborting an unsuccessful job's output. | 
| abstract void | OutputFormat. checkOutputSpecs(JobContext context)Check for validity of the output-specification for the job. | 
| void | OutputCommitter. cleanupJob(JobContext jobContext)Deprecated. 
 | 
| void | OutputCommitter. commitJob(JobContext jobContext)For committing job's output after successful job completion. | 
| abstract List<InputSplit> | InputFormat. getSplits(JobContext context)Logically split the set of input files for the job. | 
| boolean | OutputCommitter. isRecoverySupported(JobContext jobContext)Is task output recovery supported for restarting jobs?
 
 If task output recovery is supported, job restart can be done more
 efficiently. | 
| abstract void | OutputCommitter. setupJob(JobContext jobContext)For the framework to setup the job output during initialization. | 
| Modifier and Type | Method and Description | 
|---|---|
| void | DBOutputFormat. checkOutputSpecs(JobContext context) | 
| List<InputSplit> | DBInputFormat. getSplits(JobContext job)Logically split the set of input files for the job. | 
| List<InputSplit> | DataDrivenDBInputFormat. getSplits(JobContext job)Logically split the set of input files for the job. | 
| Modifier and Type | Method and Description | 
|---|---|
| static boolean | FileInputFormat. getInputDirRecursive(JobContext job) | 
| static PathFilter | FileInputFormat. getInputPathFilter(JobContext context)Get a PathFilter instance of the filter set for the input paths. | 
| static Path[] | FileInputFormat. getInputPaths(JobContext context)Get the list of input  Paths for the map-reduce job. | 
| static long | FileInputFormat. getMaxSplitSize(JobContext context)Get the maximum split size. | 
| static long | FileInputFormat. getMinSplitSize(JobContext job)Get the minimum split size | 
| static int | NLineInputFormat. getNumLinesPerSplit(JobContext job)Get the number of lines per split | 
| List<InputSplit> | FileInputFormat. getSplits(JobContext job)Generate the list of files and make them into FileSplits. | 
| List<InputSplit> | CombineFileInputFormat. getSplits(JobContext job) | 
| List<InputSplit> | NLineInputFormat. getSplits(JobContext job)Logically splits the set of input files for the job, splits N lines
 of the input as one split. | 
| protected boolean | TextInputFormat. isSplitable(JobContext context,
                      Path file) | 
| protected boolean | FileInputFormat. isSplitable(JobContext context,
                      Path filename)Is the given filename splitable? Usually, true, but if the file is
 stream compressed, it will not be. | 
| protected boolean | CombineFileInputFormat. isSplitable(JobContext context,
                      Path file) | 
| protected boolean | KeyValueTextInputFormat. isSplitable(JobContext context,
                      Path file) | 
| protected boolean | FixedLengthInputFormat. isSplitable(JobContext context,
                      Path file) | 
| protected List<FileStatus> | FileInputFormat. listStatus(JobContext job)List input directories. | 
| protected List<FileStatus> | SequenceFileInputFormat. listStatus(JobContext job) | 
| Modifier and Type | Method and Description | 
|---|---|
| List<InputSplit> | CompositeInputFormat. getSplits(JobContext job)Build a CompositeInputSplit from the child InputFormats by assigning the
 ith split from each child to the ith composite split. | 
| Modifier and Type | Method and Description | 
|---|---|
| static <K1,V1,K2,V2>  | MultithreadedMapper. getMapperClass(JobContext job)Get the application's mapper class. | 
| static int | MultithreadedMapper. getNumberOfThreads(JobContext job)The number of threads in the thread pool that will run the map function. | 
| Modifier and Type | Method and Description | 
|---|---|
| void | FileOutputCommitter. abortJob(JobContext context,
                org.apache.hadoop.mapreduce.JobStatus.State state)Delete the temporary directory, including all of the work directories. | 
| void | FilterOutputFormat. checkOutputSpecs(JobContext context) | 
| void | LazyOutputFormat. checkOutputSpecs(JobContext context) | 
| void | NullOutputFormat. checkOutputSpecs(JobContext context) | 
| void | FileOutputFormat. checkOutputSpecs(JobContext job) | 
| void | SequenceFileAsBinaryOutputFormat. checkOutputSpecs(JobContext job) | 
| void | FileOutputCommitter. cleanupJob(JobContext context)Deprecated.  | 
| void | FileOutputCommitter. commitJob(JobContext context)The job has completed so move all committed tasks to the final output dir. | 
| static boolean | FileOutputFormat. getCompressOutput(JobContext job)Is the job output compressed? | 
| static boolean | MultipleOutputs. getCountersEnabled(JobContext job)Returns if the counters for the named outputs are enabled or not. | 
| Path | FileOutputCommitter. getJobAttemptPath(JobContext context)Compute the path where the output of a given job attempt will be placed. | 
| static Path | FileOutputCommitter. getJobAttemptPath(JobContext context,
                                  Path out)Compute the path where the output of a given job attempt will be placed. | 
| static org.apache.hadoop.io.SequenceFile.CompressionType | SequenceFileOutputFormat. getOutputCompressionType(JobContext job)Get the  SequenceFile.CompressionTypefor the outputSequenceFile. | 
| static Class<? extends CompressionCodec> | FileOutputFormat. getOutputCompressorClass(JobContext job,
                                                Class<? extends CompressionCodec> defaultValue)Get the  CompressionCodecfor compressing the job outputs. | 
| protected static String | FileOutputFormat. getOutputName(JobContext job)Get the base output name for the output file. | 
| static Path | FileOutputFormat. getOutputPath(JobContext job)Get the  Pathto the output directory for the map-reduce job. | 
| static Class<? extends WritableComparable> | SequenceFileAsBinaryOutputFormat. getSequenceFileOutputKeyClass(JobContext job)Get the key class for the  SequenceFile | 
| static Class<? extends Writable> | SequenceFileAsBinaryOutputFormat. getSequenceFileOutputValueClass(JobContext job)Get the value class for the  SequenceFile | 
| protected static void | FileOutputFormat. setOutputName(JobContext job,
                          String name)Set the base output name for output file to be created. | 
| void | FileOutputCommitter. setupJob(JobContext context)Create the temporary directory that is the root of all of the task 
 work directories. | 
| Constructor and Description | 
|---|
| FileOutputCommitter(Path outputPath,
                                      JobContext context)Create a file output committer | 
| PartialFileOutputCommitter(Path outputPath,
                                                    JobContext context) | 
| Modifier and Type | Method and Description | 
|---|---|
| static String | KeyFieldBasedComparator. getKeyFieldComparatorOption(JobContext job)Get the  KeyFieldBasedComparatoroptions | 
| String | KeyFieldBasedPartitioner. getKeyFieldPartitionerOption(JobContext job)Get the  KeyFieldBasedPartitioneroptions | 
| Modifier and Type | Class and Description | 
|---|---|
| class  | org.apache.hadoop.mapreduce.task.JobContextImplA read-only view of the job that is provided to the tasks while they
 are running. | 
Copyright © 2015 Apache Software Foundation. All rights reserved.