|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Uses of Job in org.apache.hadoop.mapreduce |
---|
Methods in org.apache.hadoop.mapreduce that return Job | |
---|---|
Job[] |
Cluster.getAllJobs()
Deprecated. Use Cluster.getAllJobStatuses() instead. |
static Job |
Job.getInstance()
Creates a new Job with no particular Cluster . |
static Job |
Job.getInstance(Cluster ignored)
Deprecated. Use getInstance() |
static Job |
Job.getInstance(Cluster ignored,
Configuration conf)
Deprecated. Use getInstance(Configuration) |
static Job |
Job.getInstance(Cluster cluster,
JobStatus status,
Configuration conf)
Creates a new Job with no particular Cluster and given
Configuration and JobStatus . |
static Job |
Job.getInstance(Configuration conf)
Creates a new Job with no particular Cluster and a
given Configuration . |
static Job |
Job.getInstance(Configuration conf,
String jobName)
Creates a new Job with no particular Cluster and a given jobName. |
static Job |
Job.getInstance(JobStatus status,
Configuration conf)
Creates a new Job with no particular Cluster and given
Configuration and JobStatus . |
Job |
Cluster.getJob(JobID jobId)
Get job corresponding to jobid. |
Uses of Job in org.apache.hadoop.mapreduce.lib.aggregate |
---|
Methods in org.apache.hadoop.mapreduce.lib.aggregate that return Job | |
---|---|
static Job |
ValueAggregatorJob.createValueAggregatorJob(Configuration conf,
String[] args)
Create an Aggregate based map/reduce job. |
static Job |
ValueAggregatorJob.createValueAggregatorJob(String[] args,
Class<? extends ValueAggregatorDescriptor>[] descriptors)
|
Uses of Job in org.apache.hadoop.mapreduce.lib.chain |
---|
Methods in org.apache.hadoop.mapreduce.lib.chain with parameters of type Job | |
---|---|
static void |
ChainMapper.addMapper(Job job,
Class<? extends Mapper> klass,
Class<?> inputKeyClass,
Class<?> inputValueClass,
Class<?> outputKeyClass,
Class<?> outputValueClass,
Configuration mapperConf)
Adds a Mapper class to the chain mapper. |
static void |
ChainReducer.addMapper(Job job,
Class<? extends Mapper> klass,
Class<?> inputKeyClass,
Class<?> inputValueClass,
Class<?> outputKeyClass,
Class<?> outputValueClass,
Configuration mapperConf)
Adds a Mapper class to the chain reducer. |
static void |
ChainReducer.setReducer(Job job,
Class<? extends Reducer> klass,
Class<?> inputKeyClass,
Class<?> inputValueClass,
Class<?> outputKeyClass,
Class<?> outputValueClass,
Configuration reducerConf)
Sets the Reducer class to the chain job. |
Uses of Job in org.apache.hadoop.mapreduce.lib.db |
---|
Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type Job | |
---|---|
static void |
DataDrivenDBInputFormat.setInput(Job job,
Class<? extends DBWritable> inputClass,
String inputQuery,
String inputBoundingQuery)
setInput() takes a custom query and a separate "bounding query" to use instead of the custom "count query" used by DBInputFormat. |
static void |
DBInputFormat.setInput(Job job,
Class<? extends DBWritable> inputClass,
String inputQuery,
String inputCountQuery)
Initializes the map-part of the job with the appropriate input settings. |
static void |
DataDrivenDBInputFormat.setInput(Job job,
Class<? extends DBWritable> inputClass,
String tableName,
String conditions,
String splitBy,
String... fieldNames)
Note that the "orderBy" column is called the "splitBy" in this version. |
static void |
DBInputFormat.setInput(Job job,
Class<? extends DBWritable> inputClass,
String tableName,
String conditions,
String orderBy,
String... fieldNames)
Initializes the map-part of the job with the appropriate input settings. |
static void |
DBOutputFormat.setOutput(Job job,
String tableName,
int fieldCount)
Initializes the reduce-part of the job with the appropriate output settings |
static void |
DBOutputFormat.setOutput(Job job,
String tableName,
String... fieldNames)
Initializes the reduce-part of the job with the appropriate output settings |
Uses of Job in org.apache.hadoop.mapreduce.lib.input |
---|
Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type Job | |
---|---|
static void |
FileInputFormat.addInputPath(Job job,
Path path)
Add a Path to the list of inputs for the map-reduce job. |
static void |
MultipleInputs.addInputPath(Job job,
Path path,
Class<? extends InputFormat> inputFormatClass)
Add a Path with a custom InputFormat to the list of
inputs for the map-reduce job. |
static void |
MultipleInputs.addInputPath(Job job,
Path path,
Class<? extends InputFormat> inputFormatClass,
Class<? extends Mapper> mapperClass)
Add a Path with a custom InputFormat and
Mapper to the list of inputs for the map-reduce job. |
static void |
FileInputFormat.addInputPaths(Job job,
String commaSeparatedPaths)
Add the given comma separated paths to the list of inputs for the map-reduce job. |
static void |
SequenceFileInputFilter.setFilterClass(Job job,
Class<?> filterClass)
set the filter class |
static void |
FileInputFormat.setInputDirRecursive(Job job,
boolean inputDirRecursive)
|
static void |
FileInputFormat.setInputPathFilter(Job job,
Class<? extends PathFilter> filter)
Set a PathFilter to be applied to the input paths for the map-reduce job. |
static void |
FileInputFormat.setInputPaths(Job job,
Path... inputPaths)
Set the array of Path s as the list of inputs
for the map-reduce job. |
static void |
FileInputFormat.setInputPaths(Job job,
String commaSeparatedPaths)
Sets the given comma separated paths as the list of inputs for the map-reduce job. |
static void |
FileInputFormat.setMaxInputSplitSize(Job job,
long size)
Set the maximum split size |
static void |
FileInputFormat.setMinInputSplitSize(Job job,
long size)
Set the minimum input split size |
static void |
NLineInputFormat.setNumLinesPerSplit(Job job,
int numLines)
Set the number of lines per split |
Uses of Job in org.apache.hadoop.mapreduce.lib.jobcontrol |
---|
Methods in org.apache.hadoop.mapreduce.lib.jobcontrol that return Job | |
---|---|
Job |
ControlledJob.getJob()
|
Methods in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job | |
---|---|
void |
ControlledJob.setJob(Job job)
Set the mapreduce job |
Constructors in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job | |
---|---|
ControlledJob(Job job,
List<ControlledJob> dependingJobs)
Construct a job. |
Uses of Job in org.apache.hadoop.mapreduce.lib.map |
---|
Methods in org.apache.hadoop.mapreduce.lib.map with parameters of type Job | ||
---|---|---|
static
|
MultithreadedMapper.setMapperClass(Job job,
Class<? extends Mapper<K1,V1,K2,V2>> cls)
Set the application's mapper class. |
|
static void |
MultithreadedMapper.setNumberOfThreads(Job job,
int threads)
Set the number of threads in the pool for running maps. |
Uses of Job in org.apache.hadoop.mapreduce.lib.output |
---|
Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type Job | |
---|---|
static void |
MultipleOutputs.addNamedOutput(Job job,
String namedOutput,
Class<? extends OutputFormat> outputFormatClass,
Class<?> keyClass,
Class<?> valueClass)
Adds a named output for the job. |
static void |
FileOutputFormat.setCompressOutput(Job job,
boolean compress)
Set whether the output of the job is compressed. |
static void |
MultipleOutputs.setCountersEnabled(Job job,
boolean enabled)
Enables or disables counters for the named outputs. |
static void |
SequenceFileOutputFormat.setOutputCompressionType(Job job,
org.apache.hadoop.io.SequenceFile.CompressionType style)
Set the SequenceFile.CompressionType for the output SequenceFile . |
static void |
FileOutputFormat.setOutputCompressorClass(Job job,
Class<? extends CompressionCodec> codecClass)
Set the CompressionCodec to be used to compress job outputs. |
static void |
LazyOutputFormat.setOutputFormatClass(Job job,
Class<? extends OutputFormat> theClass)
Set the underlying output format for LazyOutputFormat. |
static void |
FileOutputFormat.setOutputPath(Job job,
Path outputDir)
Set the Path of the output directory for the map-reduce job. |
static void |
SequenceFileAsBinaryOutputFormat.setSequenceFileOutputKeyClass(Job job,
Class<?> theClass)
Set the key class for the SequenceFile |
static void |
SequenceFileAsBinaryOutputFormat.setSequenceFileOutputValueClass(Job job,
Class<?> theClass)
Set the value class for the SequenceFile |
Uses of Job in org.apache.hadoop.mapreduce.lib.partition |
---|
Methods in org.apache.hadoop.mapreduce.lib.partition with parameters of type Job | ||
---|---|---|
static void |
KeyFieldBasedComparator.setKeyFieldComparatorOptions(Job job,
String keySpec)
Set the KeyFieldBasedComparator options used to compare keys. |
|
void |
KeyFieldBasedPartitioner.setKeyFieldPartitionerOptions(Job job,
String keySpec)
Set the KeyFieldBasedPartitioner options used for
Partitioner |
|
static
|
InputSampler.writePartitionFile(Job job,
org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V> sampler)
Write a partition file for the given job, using the Sampler provided. |
Uses of Job in org.apache.hadoop.mapreduce.tools |
---|
Methods in org.apache.hadoop.mapreduce.tools with parameters of type Job | |
---|---|
protected void |
CLI.displayTasks(Job job,
String type,
String state)
Display the information about a job's tasks, of a particular type and in a particular state |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |