Uses of Class
org.apache.hadoop.mapreduce.TaskAttemptContext

Packages that use TaskAttemptContext
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.db   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.output   
 

Uses of TaskAttemptContext in org.apache.hadoop.mapred
 

Subclasses of TaskAttemptContext in org.apache.hadoop.mapred
 class TaskAttemptContext
           
 

Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptContext
 void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce
 

Subclasses of TaskAttemptContext in org.apache.hadoop.mapreduce
 class MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context that is given to the Mapper.
 class Mapper.Context
           
 class ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context passed to the Reducer.
 class Reducer.Context
           
 class TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          A context object that allows input and output from the task.
 

Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptContext
abstract  void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          Discard the task output
abstract  void RecordWriter.close(TaskAttemptContext context)
          Close this RecordWriter to future operations.
abstract  void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          To promote the task's temporary output to final output location The task's output is moved to the job's output directory.
abstract  RecordReader<K,V> InputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
abstract  OutputCommitter OutputFormat.getOutputCommitter(TaskAttemptContext context)
          Get the output committer for this output format.
abstract  RecordWriter<K,V> OutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
abstract  void RecordReader.initialize(InputSplit split, TaskAttemptContext context)
          Called once at initialization.
abstract  boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          Check whether task needs a commit
abstract  void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          Sets up output for the task.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type TaskAttemptContext
 void DBOutputFormat.DBRecordWriter.close(TaskAttemptContext context)
          Close this RecordWriter to future operations.
 RecordReader<LongWritable,T> DBInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
 OutputCommitter DBOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 RecordWriter<K,V> DBOutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
 void DBRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.input
 

Fields in org.apache.hadoop.mapreduce.lib.input declared as TaskAttemptContext
protected  TaskAttemptContext CombineFileRecordReader.context
           
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
 RecordReader<K,V> DelegatingInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
abstract  RecordReader<K,V> CombineFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          This is not implemented yet.
 RecordReader<K,V> SequenceFileInputFilter.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for the given split
 RecordReader<LongWritable,Text> NLineInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 RecordReader<K,V> SequenceFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<LongWritable,Text> TextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<Text,Text> KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 void SequenceFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void SequenceFileAsTextRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void DelegatingRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void LineRecordReader.initialize(InputSplit genericSplit, TaskAttemptContext context)
           
 void CombineFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void KeyValueLineRecordReader.initialize(InputSplit genericSplit, TaskAttemptContext context)
           
 

Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
CombineFileRecordReader(CombineFileSplit split, TaskAttemptContext context, Class<? extends RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
DelegatingRecordReader(InputSplit split, TaskAttemptContext context)
          Constructs the DelegatingRecordReader.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
 void FileOutputCommitter.abortTask(TaskAttemptContext context)
          Delete the work directory
 void FilterOutputFormat.FilterRecordWriter.close(TaskAttemptContext context)
           
 void TextOutputFormat.LineRecordWriter.close(TaskAttemptContext context)
           
 void FileOutputCommitter.commitTask(TaskAttemptContext context)
          Move the files from the work directory to the job output directory
 Path FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context, String extension)
          Get the default path and filename for the output format.
 OutputCommitter FileOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter FilterOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter LazyOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter NullOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
abstract  RecordWriter<K,V> FileOutputFormat.getRecordWriter(TaskAttemptContext job)
           
 RecordWriter<BytesWritable,BytesWritable> SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> FilterOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> LazyOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> TextOutputFormat.getRecordWriter(TaskAttemptContext job)
           
 RecordWriter<K,V> NullOutputFormat.getRecordWriter(TaskAttemptContext context)
           
protected  SequenceFile.Writer SequenceFileAsBinaryOutputFormat.getSequenceWriter(TaskAttemptContext context, Class<?> keyClass, Class<?> valueClass)
           
static String FileOutputFormat.getUniqueFile(TaskAttemptContext context, String name, String extension)
          Generate a unique filename, based on the task id, name, and extension
 boolean FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)
          Did this task write any files in the work directory?
 void FileOutputCommitter.setupTask(TaskAttemptContext context)
          No task setup required.
 

Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
FileOutputCommitter(Path outputPath, TaskAttemptContext context)
          Create a file output committer
 



Copyright © 2009 The Apache Software Foundation