| 
 | ||||||||||
| PREV NEXT | FRAMES NO FRAMES | |||||||||
| Packages that use TaskAttemptContext | |
|---|---|
| org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. | 
| org.apache.hadoop.mapreduce | |
| org.apache.hadoop.mapreduce.lib.db | |
| org.apache.hadoop.mapreduce.lib.input | |
| org.apache.hadoop.mapreduce.lib.output | |
| Uses of TaskAttemptContext in org.apache.hadoop.mapred | 
|---|
| Subclasses of TaskAttemptContext in org.apache.hadoop.mapred | |
|---|---|
|  class | TaskAttemptContext | 
| Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptContext | |
|---|---|
|  void | OutputCommitter.abortTask(TaskAttemptContext taskContext)This method implements the new interface by calling the old method. | 
|  void | OutputCommitter.commitTask(TaskAttemptContext taskContext)This method implements the new interface by calling the old method. | 
|  boolean | OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)This method implements the new interface by calling the old method. | 
|  void | OutputCommitter.setupTask(TaskAttemptContext taskContext)This method implements the new interface by calling the old method. | 
| Uses of TaskAttemptContext in org.apache.hadoop.mapreduce | 
|---|
| Subclasses of TaskAttemptContext in org.apache.hadoop.mapreduce | |
|---|---|
|  class | MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>The context that is given to the Mapper. | 
|  class | Mapper.Context | 
|  class | ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>The context passed to the Reducer. | 
|  class | Reducer.Context | 
|  class | TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>A context object that allows input and output from the task. | 
| Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptContext | |
|---|---|
| abstract  void | OutputCommitter.abortTask(TaskAttemptContext taskContext)Discard the task output | 
| abstract  void | RecordWriter.close(TaskAttemptContext context)Close this RecordWriterto future operations. | 
| abstract  void | OutputCommitter.commitTask(TaskAttemptContext taskContext)To promote the task's temporary output to final output location The task's output is moved to the job's output directory. | 
| abstract  RecordReader<K,V> | InputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context)Create a record reader for a given split. | 
| abstract  OutputCommitter | OutputFormat.getOutputCommitter(TaskAttemptContext context)Get the output committer for this output format. | 
| abstract  RecordWriter<K,V> | OutputFormat.getRecordWriter(TaskAttemptContext context)Get the RecordWriterfor the given task. | 
| abstract  void | RecordReader.initialize(InputSplit split,
           TaskAttemptContext context)Called once at initialization. | 
| abstract  boolean | OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)Check whether task needs a commit | 
| abstract  void | OutputCommitter.setupTask(TaskAttemptContext taskContext)Sets up output for the task. | 
| Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.db | 
|---|
| Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type TaskAttemptContext | |
|---|---|
|  void | DBOutputFormat.DBRecordWriter.close(TaskAttemptContext context)Close this RecordWriterto future operations. | 
|  RecordReader<LongWritable,T> | DBInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context)Create a record reader for a given split. | 
|  OutputCommitter | DBOutputFormat.getOutputCommitter(TaskAttemptContext context) | 
|  RecordWriter<K,V> | DBOutputFormat.getRecordWriter(TaskAttemptContext context)Get the RecordWriterfor the given task. | 
|  void | DBRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
| Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.input | 
|---|
| Fields in org.apache.hadoop.mapreduce.lib.input declared as TaskAttemptContext | |
|---|---|
| protected  TaskAttemptContext | CombineFileRecordReader.context | 
| Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext | |
|---|---|
|  RecordReader<K,V> | DelegatingInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context) | 
|  RecordReader<BytesWritable,BytesWritable> | SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context) | 
| abstract  RecordReader<K,V> | CombineFileInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context)This is not implemented yet. | 
|  RecordReader<K,V> | SequenceFileInputFilter.createRecordReader(InputSplit split,
                   TaskAttemptContext context)Create a record reader for the given split | 
|  RecordReader<LongWritable,Text> | NLineInputFormat.createRecordReader(InputSplit genericSplit,
                   TaskAttemptContext context) | 
|  RecordReader<K,V> | SequenceFileInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context) | 
|  RecordReader<LongWritable,Text> | TextInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context) | 
|  RecordReader<Text,Text> | SequenceFileAsTextInputFormat.createRecordReader(InputSplit split,
                   TaskAttemptContext context) | 
|  RecordReader<Text,Text> | KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit,
                   TaskAttemptContext context) | 
|  void | SequenceFileRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
|  void | SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
|  void | SequenceFileAsTextRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
|  void | DelegatingRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
|  void | LineRecordReader.initialize(InputSplit genericSplit,
           TaskAttemptContext context) | 
|  void | CombineFileRecordReader.initialize(InputSplit split,
           TaskAttemptContext context) | 
|  void | KeyValueLineRecordReader.initialize(InputSplit genericSplit,
           TaskAttemptContext context) | 
| Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext | |
|---|---|
| CombineFileRecordReader(CombineFileSplit split,
                        TaskAttemptContext context,
                        Class<? extends RecordReader<K,V>> rrClass)A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit. | |
| DelegatingRecordReader(InputSplit split,
                       TaskAttemptContext context)Constructs the DelegatingRecordReader. | |
| Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.output | 
|---|
| Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext | |
|---|---|
|  void | FileOutputCommitter.abortTask(TaskAttemptContext context)Delete the work directory | 
|  void | FilterOutputFormat.FilterRecordWriter.close(TaskAttemptContext context) | 
|  void | TextOutputFormat.LineRecordWriter.close(TaskAttemptContext context) | 
|  void | FileOutputCommitter.commitTask(TaskAttemptContext context)Move the files from the work directory to the job output directory | 
|  Path | FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context,
                   String extension)Get the default path and filename for the output format. | 
|  OutputCommitter | FileOutputFormat.getOutputCommitter(TaskAttemptContext context) | 
|  OutputCommitter | FilterOutputFormat.getOutputCommitter(TaskAttemptContext context) | 
|  OutputCommitter | LazyOutputFormat.getOutputCommitter(TaskAttemptContext context) | 
|  OutputCommitter | NullOutputFormat.getOutputCommitter(TaskAttemptContext context) | 
| abstract  RecordWriter<K,V> | FileOutputFormat.getRecordWriter(TaskAttemptContext job) | 
|  RecordWriter<BytesWritable,BytesWritable> | SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context) | 
|  RecordWriter<K,V> | FilterOutputFormat.getRecordWriter(TaskAttemptContext context) | 
|  RecordWriter<K,V> | LazyOutputFormat.getRecordWriter(TaskAttemptContext context) | 
|  RecordWriter<K,V> | SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context) | 
|  RecordWriter<K,V> | TextOutputFormat.getRecordWriter(TaskAttemptContext job) | 
|  RecordWriter<K,V> | NullOutputFormat.getRecordWriter(TaskAttemptContext context) | 
| protected  SequenceFile.Writer | SequenceFileAsBinaryOutputFormat.getSequenceWriter(TaskAttemptContext context,
                  Class<?> keyClass,
                  Class<?> valueClass) | 
| static String | FileOutputFormat.getUniqueFile(TaskAttemptContext context,
              String name,
              String extension)Generate a unique filename, based on the task id, name, and extension | 
|  boolean | FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)Did this task write any files in the work directory? | 
|  void | FileOutputCommitter.setupTask(TaskAttemptContext context)No task setup required. | 
| Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext | |
|---|---|
| FileOutputCommitter(Path outputPath,
                    TaskAttemptContext context)Create a file output committer | |
| 
 | ||||||||||
| PREV NEXT | FRAMES NO FRAMES | |||||||||