Uses of Interface
org.apache.hadoop.mapreduce.TaskAttemptContext

Packages that use TaskAttemptContext
org.apache.hadoop.mapred   
org.apache.hadoop.mapred.lib   
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.db   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.join   
org.apache.hadoop.mapreduce.lib.output   
 

Uses of TaskAttemptContext in org.apache.hadoop.mapred
 

Subinterfaces of TaskAttemptContext in org.apache.hadoop.mapred
 interface TaskAttemptContext
           
 

Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptContext
 void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.recoverTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapred.lib
 

Methods in org.apache.hadoop.mapred.lib with parameters of type TaskAttemptContext
 RecordReader<K,V> CombineFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce
 

Subinterfaces of TaskAttemptContext in org.apache.hadoop.mapreduce
 interface MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context that is given to the Mapper.
 interface ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context passed to the Reducer.
 interface TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          A context object that allows input and output from the task.
 

Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptContext
abstract  void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          Discard the task output.
abstract  void RecordWriter.close(TaskAttemptContext context)
          Close this RecordWriter to future operations.
abstract  void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          To promote the task's temporary output to final output location.
abstract  RecordReader<K,V> InputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
abstract  OutputCommitter OutputFormat.getOutputCommitter(TaskAttemptContext context)
          Get the output committer for this output format.
abstract  RecordWriter<K,V> OutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
abstract  void RecordReader.initialize(InputSplit split, TaskAttemptContext context)
          Called once at initialization.
abstract  boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          Check whether task needs a commit.
 void OutputCommitter.recoverTask(TaskAttemptContext taskContext)
          Recover the task output.
abstract  void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          Sets up output for the task.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type TaskAttemptContext
 RecordReader<LongWritable,T> DBInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
 OutputCommitter DBOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 RecordWriter<K,V> DBOutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
 void DBRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.input
 

Fields in org.apache.hadoop.mapreduce.lib.input declared as TaskAttemptContext
protected  TaskAttemptContext CombineFileRecordReader.context
           
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<LongWritable,Text> TextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
abstract  RecordReader<K,V> CombineFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          This is not implemented yet.
 RecordReader<K,V> SequenceFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<K,V> CombineSequenceFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<LongWritable,BytesWritable> FixedLengthInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<K,V> SequenceFileInputFilter.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for the given split
 RecordReader<LongWritable,Text> NLineInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 RecordReader<Text,Text> KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 RecordReader<LongWritable,Text> CombineTextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 void KeyValueLineRecordReader.initialize(InputSplit genericSplit, TaskAttemptContext context)
           
 void CombineFileRecordReaderWrapper.initialize(InputSplit split, TaskAttemptContext context)
           
 void CombineFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void SequenceFileAsTextRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void SequenceFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
CombineFileRecordReader(CombineFileSplit split, TaskAttemptContext context, Class<? extends RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
CombineFileRecordReaderWrapper(FileInputFormat<K,V> inputFormat, CombineFileSplit split, TaskAttemptContext context, Integer idx)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.join
 

Methods in org.apache.hadoop.mapreduce.lib.join with parameters of type TaskAttemptContext
abstract  ComposableRecordReader<K,V> ComposableInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<K,TupleWritable> CompositeInputFormat.createRecordReader(InputSplit split, TaskAttemptContext taskContext)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 void CompositeRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void WrappedRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void MultiFilterRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
 void FileOutputCommitter.abortTask(TaskAttemptContext context)
          Delete the work directory
 void FileOutputCommitter.abortTask(TaskAttemptContext context, Path taskAttemptPath)
           
 void PartialFileOutputCommitter.cleanUpPartialOutputForTask(TaskAttemptContext context)
           
 void PartialOutputCommitter.cleanUpPartialOutputForTask(TaskAttemptContext context)
          Remove all previously committed outputs from prior executions of this task.
 void FileOutputCommitter.commitTask(TaskAttemptContext context)
          Move the files from the work directory to the job output directory
 void FileOutputCommitter.commitTask(TaskAttemptContext context, Path taskAttemptPath)
           
 Path PartialFileOutputCommitter.getCommittedTaskPath(int appAttemptId, TaskAttemptContext context)
           
protected  Path FileOutputCommitter.getCommittedTaskPath(int appAttemptId, TaskAttemptContext context)
          Compute the path where the output of a committed task is stored until the entire job is committed for a specific application attempt.
 Path FileOutputCommitter.getCommittedTaskPath(TaskAttemptContext context)
          Compute the path where the output of a committed task is stored until the entire job is committed.
static Path FileOutputCommitter.getCommittedTaskPath(TaskAttemptContext context, Path out)
           
 Path FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context, String extension)
          Get the default path and filename for the output format.
 OutputCommitter NullOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter FilterOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter LazyOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter FileOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 RecordWriter<K,V> SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> NullOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> FilterOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> TextOutputFormat.getRecordWriter(TaskAttemptContext job)
           
 RecordWriter<WritableComparable<?>,Writable> MapFileOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> LazyOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<BytesWritable,BytesWritable> SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context)
           
abstract  RecordWriter<K,V> FileOutputFormat.getRecordWriter(TaskAttemptContext job)
           
protected  org.apache.hadoop.io.SequenceFile.Writer SequenceFileOutputFormat.getSequenceWriter(TaskAttemptContext context, Class<?> keyClass, Class<?> valueClass)
           
 Path FileOutputCommitter.getTaskAttemptPath(TaskAttemptContext context)
          Compute the path where the output of a task attempt is stored until that task is committed.
static Path FileOutputCommitter.getTaskAttemptPath(TaskAttemptContext context, Path out)
          Compute the path where the output of a task attempt is stored until that task is committed.
static String FileOutputFormat.getUniqueFile(TaskAttemptContext context, String name, String extension)
          Generate a unique filename, based on the task id, name, and extension
 boolean FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)
          Did this task write any files in the work directory?
 boolean FileOutputCommitter.needsTaskCommit(TaskAttemptContext context, Path taskAttemptPath)
           
 void FileOutputCommitter.recoverTask(TaskAttemptContext context)
           
 void FileOutputCommitter.setupTask(TaskAttemptContext context)
          No task setup required.
 

Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
FileOutputCommitter(Path outputPath, TaskAttemptContext context)
          Create a file output committer
PartialFileOutputCommitter(Path outputPath, TaskAttemptContext context)
           
 



Copyright © 2014 Apache Software Foundation. All Rights Reserved.