Uses of Interface
org.apache.hadoop.mapreduce.JobContext

Packages that use JobContext
org.apache.hadoop.mapred   
org.apache.hadoop.mapred.lib   
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.db   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.join   
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.output   
org.apache.hadoop.mapreduce.lib.partition   
org.apache.hadoop.mapreduce.task   
 

Uses of JobContext in org.apache.hadoop.mapred
 

Subinterfaces of JobContext in org.apache.hadoop.mapred
 interface JobContext
           
 

Methods in org.apache.hadoop.mapred with parameters of type JobContext
 void OutputCommitter.abortJob(JobContext context, org.apache.hadoop.mapreduce.JobStatus.State runState)
          This method implements the new interface by calling the old method.
 void OutputCommitter.cleanupJob(JobContext context)
          Deprecated. Use OutputCommitter.commitJob(org.apache.hadoop.mapreduce.JobContext) or OutputCommitter.abortJob(org.apache.hadoop.mapreduce.JobContext, org.apache.hadoop.mapreduce.JobStatus.State) instead.
 void OutputCommitter.commitJob(JobContext context)
          This method implements the new interface by calling the old method.
 void OutputCommitter.setupJob(JobContext jobContext)
          This method implements the new interface by calling the old method.
 

Uses of JobContext in org.apache.hadoop.mapred.lib
 

Methods in org.apache.hadoop.mapred.lib with parameters of type JobContext
protected  boolean CombineFileInputFormat.isSplitable(JobContext context, Path file)
          Subclasses should avoid overriding this method and should instead only override CombineFileInputFormat.isSplitable(FileSystem, Path).
 

Uses of JobContext in org.apache.hadoop.mapreduce
 

Subinterfaces of JobContext in org.apache.hadoop.mapreduce
 interface MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context that is given to the Mapper.
 interface ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context passed to the Reducer.
 interface TaskAttemptContext
          The context for task attempts.
 interface TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce that implement JobContext
 class Job
          The job submitter's view of the Job.
 

Methods in org.apache.hadoop.mapreduce with parameters of type JobContext
 void OutputCommitter.abortJob(JobContext jobContext, org.apache.hadoop.mapreduce.JobStatus.State state)
          For aborting an unsuccessful job's output.
abstract  void OutputFormat.checkOutputSpecs(JobContext context)
          Check for validity of the output-specification for the job.
 void OutputCommitter.cleanupJob(JobContext jobContext)
          Deprecated. Use OutputCommitter.commitJob(JobContext) and OutputCommitter.abortJob(JobContext, JobStatus.State) instead.
 void OutputCommitter.commitJob(JobContext jobContext)
          For committing job's output after successful job completion.
abstract  List<InputSplit> InputFormat.getSplits(JobContext context)
          Logically split the set of input files for the job.
abstract  void OutputCommitter.setupJob(JobContext jobContext)
          For the framework to setup the job output during initialization.
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type JobContext
 void DBOutputFormat.checkOutputSpecs(JobContext context)
           
 List<InputSplit> DBInputFormat.getSplits(JobContext job)
          Logically split the set of input files for the job.
 List<InputSplit> DataDrivenDBInputFormat.getSplits(JobContext job)
          Logically split the set of input files for the job.
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.input
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type JobContext
static boolean FileInputFormat.getInputDirRecursive(JobContext job)
           
static PathFilter FileInputFormat.getInputPathFilter(JobContext context)
          Get a PathFilter instance of the filter set for the input paths.
static Path[] FileInputFormat.getInputPaths(JobContext context)
          Get the list of input Paths for the map-reduce job.
static long FileInputFormat.getMaxSplitSize(JobContext context)
          Get the maximum split size.
static long FileInputFormat.getMinSplitSize(JobContext job)
          Get the minimum split size
static int NLineInputFormat.getNumLinesPerSplit(JobContext job)
          Get the number of lines per split
 List<InputSplit> FileInputFormat.getSplits(JobContext job)
          Generate the list of files and make them into FileSplits.
 List<InputSplit> NLineInputFormat.getSplits(JobContext job)
          Logically splits the set of input files for the job, splits N lines of the input as one split.
 List<InputSplit> CombineFileInputFormat.getSplits(JobContext job)
           
protected  boolean FileInputFormat.isSplitable(JobContext context, Path filename)
          Is the given filename splitable? Usually, true, but if the file is stream compressed, it will not be.
protected  boolean KeyValueTextInputFormat.isSplitable(JobContext context, Path file)
           
protected  boolean TextInputFormat.isSplitable(JobContext context, Path file)
           
protected  boolean CombineFileInputFormat.isSplitable(JobContext context, Path file)
           
protected  List<FileStatus> FileInputFormat.listStatus(JobContext job)
          List input directories.
protected  List<FileStatus> SequenceFileInputFormat.listStatus(JobContext job)
           
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.join
 

Methods in org.apache.hadoop.mapreduce.lib.join with parameters of type JobContext
 List<InputSplit> CompositeInputFormat.getSplits(JobContext job)
          Build a CompositeInputSplit from the child InputFormats by assigning the ith split from each child to the ith composite split.
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.map
 

Methods in org.apache.hadoop.mapreduce.lib.map with parameters of type JobContext
static
<K1,V1,K2,V2>
Class<Mapper<K1,V1,K2,V2>>
MultithreadedMapper.getMapperClass(JobContext job)
          Get the application's mapper class.
static int MultithreadedMapper.getNumberOfThreads(JobContext job)
          The number of threads in the thread pool that will run the map function.
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type JobContext
 void FileOutputCommitter.abortJob(JobContext context, org.apache.hadoop.mapreduce.JobStatus.State state)
          Delete the temporary directory, including all of the work directories.
 void FileOutputFormat.checkOutputSpecs(JobContext job)
           
 void FilterOutputFormat.checkOutputSpecs(JobContext context)
           
 void NullOutputFormat.checkOutputSpecs(JobContext context)
           
 void SequenceFileAsBinaryOutputFormat.checkOutputSpecs(JobContext job)
           
 void LazyOutputFormat.checkOutputSpecs(JobContext context)
           
 void FileOutputCommitter.cleanupJob(JobContext context)
          Deprecated. 
 void FileOutputCommitter.commitJob(JobContext context)
          The job has completed so move all committed tasks to the final output dir.
static boolean FileOutputFormat.getCompressOutput(JobContext job)
          Is the job output compressed?
static boolean MultipleOutputs.getCountersEnabled(JobContext job)
          Returns if the counters for the named outputs are enabled or not.
 Path FileOutputCommitter.getJobAttemptPath(JobContext context)
          Compute the path where the output of a given job attempt will be placed.
static Path FileOutputCommitter.getJobAttemptPath(JobContext context, Path out)
          Compute the path where the output of a given job attempt will be placed.
static org.apache.hadoop.io.SequenceFile.CompressionType SequenceFileOutputFormat.getOutputCompressionType(JobContext job)
          Get the SequenceFile.CompressionType for the output SequenceFile.
static Class<? extends CompressionCodec> FileOutputFormat.getOutputCompressorClass(JobContext job, Class<? extends CompressionCodec> defaultValue)
          Get the CompressionCodec for compressing the job outputs.
protected static String FileOutputFormat.getOutputName(JobContext job)
          Get the base output name for the output file.
static Path FileOutputFormat.getOutputPath(JobContext job)
          Get the Path to the output directory for the map-reduce job.
static Class<? extends WritableComparable> SequenceFileAsBinaryOutputFormat.getSequenceFileOutputKeyClass(JobContext job)
          Get the key class for the SequenceFile
static Class<? extends Writable> SequenceFileAsBinaryOutputFormat.getSequenceFileOutputValueClass(JobContext job)
          Get the value class for the SequenceFile
protected static void FileOutputFormat.setOutputName(JobContext job, String name)
          Set the base output name for output file to be created.
 void FileOutputCommitter.setupJob(JobContext context)
          Create the temporary directory that is the root of all of the task work directories.
 

Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type JobContext
FileOutputCommitter(Path outputPath, JobContext context)
          Create a file output committer
PartialFileOutputCommitter(Path outputPath, JobContext context)
           
 

Uses of JobContext in org.apache.hadoop.mapreduce.lib.partition
 

Methods in org.apache.hadoop.mapreduce.lib.partition with parameters of type JobContext
static String KeyFieldBasedComparator.getKeyFieldComparatorOption(JobContext job)
          Get the KeyFieldBasedComparator options
 String KeyFieldBasedPartitioner.getKeyFieldPartitionerOption(JobContext job)
          Get the KeyFieldBasedPartitioner options
 

Uses of JobContext in org.apache.hadoop.mapreduce.task
 

Classes in org.apache.hadoop.mapreduce.task that implement JobContext
 class org.apache.hadoop.mapreduce.task.JobContextImpl
          A read-only view of the job that is provided to the tasks while they are running.
 



Copyright © 2013 Apache Software Foundation. All Rights Reserved.