Uses of Class
org.apache.hadoop.mapreduce.Job

Packages that use Job
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.aggregate   
org.apache.hadoop.mapreduce.lib.chain   
org.apache.hadoop.mapreduce.lib.db   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.jobcontrol   
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.output   
org.apache.hadoop.mapreduce.lib.partition   
org.apache.hadoop.mapreduce.tools   
 

Uses of Job in org.apache.hadoop.mapreduce
 

Methods in org.apache.hadoop.mapreduce that return Job
 Job[] Cluster.getAllJobs()
          Deprecated. Use Cluster.getAllJobStatuses() instead.
static Job Job.getInstance()
          Creates a new Job with no particular Cluster .
static Job Job.getInstance(Cluster ignored)
          Deprecated. Use getInstance()
static Job Job.getInstance(Cluster ignored, Configuration conf)
          Deprecated. Use getInstance(Configuration)
static Job Job.getInstance(Cluster cluster, JobStatus status, Configuration conf)
          Creates a new Job with no particular Cluster and given Configuration and JobStatus.
static Job Job.getInstance(Configuration conf)
          Creates a new Job with no particular Cluster and a given Configuration.
static Job Job.getInstance(Configuration conf, String jobName)
          Creates a new Job with no particular Cluster and a given jobName.
static Job Job.getInstance(JobStatus status, Configuration conf)
          Creates a new Job with no particular Cluster and given Configuration and JobStatus.
 Job Cluster.getJob(JobID jobId)
          Get job corresponding to jobid.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.aggregate
 

Methods in org.apache.hadoop.mapreduce.lib.aggregate that return Job
static Job ValueAggregatorJob.createValueAggregatorJob(Configuration conf, String[] args)
          Create an Aggregate based map/reduce job.
static Job ValueAggregatorJob.createValueAggregatorJob(String[] args, Class<? extends ValueAggregatorDescriptor>[] descriptors)
           
 

Uses of Job in org.apache.hadoop.mapreduce.lib.chain
 

Methods in org.apache.hadoop.mapreduce.lib.chain with parameters of type Job
static void ChainMapper.addMapper(Job job, Class<? extends Mapper> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, Configuration mapperConf)
          Adds a Mapper class to the chain mapper.
static void ChainReducer.addMapper(Job job, Class<? extends Mapper> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, Configuration mapperConf)
          Adds a Mapper class to the chain reducer.
static void ChainReducer.setReducer(Job job, Class<? extends Reducer> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, Configuration reducerConf)
          Sets the Reducer class to the chain job.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type Job
static void DataDrivenDBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String inputQuery, String inputBoundingQuery)
          setInput() takes a custom query and a separate "bounding query" to use instead of the custom "count query" used by DBInputFormat.
static void DBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String inputQuery, String inputCountQuery)
          Initializes the map-part of the job with the appropriate input settings.
static void DataDrivenDBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String tableName, String conditions, String splitBy, String... fieldNames)
          Note that the "orderBy" column is called the "splitBy" in this version.
static void DBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String tableName, String conditions, String orderBy, String... fieldNames)
          Initializes the map-part of the job with the appropriate input settings.
static void DBOutputFormat.setOutput(Job job, String tableName, int fieldCount)
          Initializes the reduce-part of the job with the appropriate output settings
static void DBOutputFormat.setOutput(Job job, String tableName, String... fieldNames)
          Initializes the reduce-part of the job with the appropriate output settings
 

Uses of Job in org.apache.hadoop.mapreduce.lib.input
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type Job
static void FileInputFormat.addInputPath(Job job, Path path)
          Add a Path to the list of inputs for the map-reduce job.
static void MultipleInputs.addInputPath(Job job, Path path, Class<? extends InputFormat> inputFormatClass)
          Add a Path with a custom InputFormat to the list of inputs for the map-reduce job.
static void MultipleInputs.addInputPath(Job job, Path path, Class<? extends InputFormat> inputFormatClass, Class<? extends Mapper> mapperClass)
          Add a Path with a custom InputFormat and Mapper to the list of inputs for the map-reduce job.
static void FileInputFormat.addInputPaths(Job job, String commaSeparatedPaths)
          Add the given comma separated paths to the list of inputs for the map-reduce job.
static void SequenceFileInputFilter.setFilterClass(Job job, Class<?> filterClass)
          set the filter class
static void FileInputFormat.setInputDirRecursive(Job job, boolean inputDirRecursive)
           
static void FileInputFormat.setInputPathFilter(Job job, Class<? extends PathFilter> filter)
          Set a PathFilter to be applied to the input paths for the map-reduce job.
static void FileInputFormat.setInputPaths(Job job, Path... inputPaths)
          Set the array of Paths as the list of inputs for the map-reduce job.
static void FileInputFormat.setInputPaths(Job job, String commaSeparatedPaths)
          Sets the given comma separated paths as the list of inputs for the map-reduce job.
static void FileInputFormat.setMaxInputSplitSize(Job job, long size)
          Set the maximum split size
static void FileInputFormat.setMinInputSplitSize(Job job, long size)
          Set the minimum input split size
static void NLineInputFormat.setNumLinesPerSplit(Job job, int numLines)
          Set the number of lines per split
 

Uses of Job in org.apache.hadoop.mapreduce.lib.jobcontrol
 

Methods in org.apache.hadoop.mapreduce.lib.jobcontrol that return Job
 Job ControlledJob.getJob()
           
 

Methods in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job
 void ControlledJob.setJob(Job job)
          Set the mapreduce job
 

Constructors in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job
ControlledJob(Job job, List<ControlledJob> dependingJobs)
          Construct a job.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.map
 

Methods in org.apache.hadoop.mapreduce.lib.map with parameters of type Job
static
<K1,V1,K2,V2>
void
MultithreadedMapper.setMapperClass(Job job, Class<? extends Mapper<K1,V1,K2,V2>> cls)
          Set the application's mapper class.
static void MultithreadedMapper.setNumberOfThreads(Job job, int threads)
          Set the number of threads in the pool for running maps.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type Job
static void MultipleOutputs.addNamedOutput(Job job, String namedOutput, Class<? extends OutputFormat> outputFormatClass, Class<?> keyClass, Class<?> valueClass)
          Adds a named output for the job.
static void FileOutputFormat.setCompressOutput(Job job, boolean compress)
          Set whether the output of the job is compressed.
static void MultipleOutputs.setCountersEnabled(Job job, boolean enabled)
          Enables or disables counters for the named outputs.
static void SequenceFileOutputFormat.setOutputCompressionType(Job job, org.apache.hadoop.io.SequenceFile.CompressionType style)
          Set the SequenceFile.CompressionType for the output SequenceFile.
static void FileOutputFormat.setOutputCompressorClass(Job job, Class<? extends CompressionCodec> codecClass)
          Set the CompressionCodec to be used to compress job outputs.
static void LazyOutputFormat.setOutputFormatClass(Job job, Class<? extends OutputFormat> theClass)
          Set the underlying output format for LazyOutputFormat.
static void FileOutputFormat.setOutputPath(Job job, Path outputDir)
          Set the Path of the output directory for the map-reduce job.
static void SequenceFileAsBinaryOutputFormat.setSequenceFileOutputKeyClass(Job job, Class<?> theClass)
          Set the key class for the SequenceFile
static void SequenceFileAsBinaryOutputFormat.setSequenceFileOutputValueClass(Job job, Class<?> theClass)
          Set the value class for the SequenceFile
 

Uses of Job in org.apache.hadoop.mapreduce.lib.partition
 

Methods in org.apache.hadoop.mapreduce.lib.partition with parameters of type Job
static void KeyFieldBasedComparator.setKeyFieldComparatorOptions(Job job, String keySpec)
          Set the KeyFieldBasedComparator options used to compare keys.
 void KeyFieldBasedPartitioner.setKeyFieldPartitionerOptions(Job job, String keySpec)
          Set the KeyFieldBasedPartitioner options used for Partitioner
static
<K,V> void
InputSampler.writePartitionFile(Job job, org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V> sampler)
          Write a partition file for the given job, using the Sampler provided.
 

Uses of Job in org.apache.hadoop.mapreduce.tools
 

Methods in org.apache.hadoop.mapreduce.tools with parameters of type Job
protected  void CLI.displayTasks(Job job, String type, String state)
          Display the information about a job's tasks, of a particular type and in a particular state
 



Copyright © 2014 Apache Software Foundation. All Rights Reserved.