|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use TaskAttemptContext | |
---|---|
org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. |
org.apache.hadoop.mapreduce | |
org.apache.hadoop.mapreduce.lib.db | |
org.apache.hadoop.mapreduce.lib.input | |
org.apache.hadoop.mapreduce.lib.output |
Uses of TaskAttemptContext in org.apache.hadoop.mapred |
---|
Subclasses of TaskAttemptContext in org.apache.hadoop.mapred | |
---|---|
class |
TaskAttemptContext
|
Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptContext | |
---|---|
void |
OutputCommitter.abortTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method. |
void |
OutputCommitter.commitTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method. |
boolean |
OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method. |
void |
OutputCommitter.setupTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method. |
Uses of TaskAttemptContext in org.apache.hadoop.mapreduce |
---|
Subclasses of TaskAttemptContext in org.apache.hadoop.mapreduce | |
---|---|
class |
MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context that is given to the Mapper . |
class |
Mapper.Context
|
class |
ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context passed to the Reducer . |
class |
Reducer.Context
|
class |
TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
A context object that allows input and output from the task. |
Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptContext | |
---|---|
abstract void |
OutputCommitter.abortTask(TaskAttemptContext taskContext)
Discard the task output |
abstract void |
RecordWriter.close(TaskAttemptContext context)
Close this RecordWriter to future operations. |
abstract void |
OutputCommitter.commitTask(TaskAttemptContext taskContext)
To promote the task's temporary output to final output location The task's output is moved to the job's output directory. |
abstract RecordReader<K,V> |
InputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for a given split. |
abstract OutputCommitter |
OutputFormat.getOutputCommitter(TaskAttemptContext context)
Get the output committer for this output format. |
abstract RecordWriter<K,V> |
OutputFormat.getRecordWriter(TaskAttemptContext context)
Get the RecordWriter for the given task. |
abstract void |
RecordReader.initialize(InputSplit split,
TaskAttemptContext context)
Called once at initialization. |
abstract boolean |
OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
Check whether task needs a commit |
abstract void |
OutputCommitter.setupTask(TaskAttemptContext taskContext)
Sets up output for the task. |
Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.db |
---|
Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type TaskAttemptContext | |
---|---|
void |
DBOutputFormat.DBRecordWriter.close(TaskAttemptContext context)
Close this RecordWriter to future operations. |
RecordReader<LongWritable,T> |
DBInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for a given split. |
OutputCommitter |
DBOutputFormat.getOutputCommitter(TaskAttemptContext context)
|
RecordWriter<K,V> |
DBOutputFormat.getRecordWriter(TaskAttemptContext context)
Get the RecordWriter for the given task. |
void |
DBRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.input |
---|
Fields in org.apache.hadoop.mapreduce.lib.input declared as TaskAttemptContext | |
---|---|
protected TaskAttemptContext |
CombineFileRecordReader.context
|
Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext | |
---|---|
RecordReader<K,V> |
SequenceFileInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<LongWritable,Text> |
TextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<K,V> |
SequenceFileInputFilter.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for the given split |
RecordReader<BytesWritable,BytesWritable> |
SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<LongWritable,Text> |
NLineInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context)
|
abstract RecordReader<K,V> |
CombineFileInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
This is not implemented yet. |
RecordReader<K,V> |
DelegatingInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
void |
CombineFileRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
void |
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
void |
KeyValueLineRecordReader.initialize(InputSplit genericSplit,
TaskAttemptContext context)
|
void |
SequenceFileAsTextRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
void |
SequenceFileRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
void |
LineRecordReader.initialize(InputSplit genericSplit,
TaskAttemptContext context)
|
void |
DelegatingRecordReader.initialize(InputSplit split,
TaskAttemptContext context)
|
Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext | |
---|---|
CombineFileRecordReader(CombineFileSplit split,
TaskAttemptContext context,
Class<? extends RecordReader<K,V>> rrClass)
A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit. |
|
DelegatingRecordReader(InputSplit split,
TaskAttemptContext context)
Constructs the DelegatingRecordReader. |
Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.output |
---|
Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext | |
---|---|
void |
FileOutputCommitter.abortTask(TaskAttemptContext context)
Delete the work directory |
void |
FilterOutputFormat.FilterRecordWriter.close(TaskAttemptContext context)
|
void |
TextOutputFormat.LineRecordWriter.close(TaskAttemptContext context)
|
void |
FileOutputCommitter.commitTask(TaskAttemptContext context)
Move the files from the work directory to the job output directory |
Path |
FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context,
String extension)
Get the default path and filename for the output format. |
OutputCommitter |
NullOutputFormat.getOutputCommitter(TaskAttemptContext context)
|
OutputCommitter |
LazyOutputFormat.getOutputCommitter(TaskAttemptContext context)
|
OutputCommitter |
FileOutputFormat.getOutputCommitter(TaskAttemptContext context)
|
OutputCommitter |
FilterOutputFormat.getOutputCommitter(TaskAttemptContext context)
|
RecordWriter<K,V> |
NullOutputFormat.getRecordWriter(TaskAttemptContext context)
|
RecordWriter<K,V> |
LazyOutputFormat.getRecordWriter(TaskAttemptContext context)
|
abstract RecordWriter<K,V> |
FileOutputFormat.getRecordWriter(TaskAttemptContext job)
|
RecordWriter<K,V> |
FilterOutputFormat.getRecordWriter(TaskAttemptContext context)
|
RecordWriter<BytesWritable,BytesWritable> |
SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context)
|
RecordWriter<K,V> |
SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context)
|
RecordWriter<K,V> |
TextOutputFormat.getRecordWriter(TaskAttemptContext job)
|
protected SequenceFile.Writer |
SequenceFileAsBinaryOutputFormat.getSequenceWriter(TaskAttemptContext context,
Class<?> keyClass,
Class<?> valueClass)
|
static String |
FileOutputFormat.getUniqueFile(TaskAttemptContext context,
String name,
String extension)
Generate a unique filename, based on the task id, name, and extension |
boolean |
FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)
Did this task write any files in the work directory? |
void |
FileOutputCommitter.setupTask(TaskAttemptContext context)
No task setup required. |
Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext | |
---|---|
FileOutputCommitter(Path outputPath,
TaskAttemptContext context)
Create a file output committer |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |