Uses of Interface
org.apache.hadoop.mapred.Reporter

Packages that use Reporter
org.apache.hadoop.contrib.index.example   
org.apache.hadoop.contrib.index.mapred   
org.apache.hadoop.contrib.utils.join   
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.examples.dancing This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. 
org.apache.hadoop.examples.terasort This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition. 
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapred.join Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. 
org.apache.hadoop.mapred.lib Library of generally useful mappers, reducers, and partitioners. 
org.apache.hadoop.mapred.lib.aggregate Classes for performing various counting and aggregations. 
org.apache.hadoop.mapred.lib.db org.apache.hadoop.mapred.lib.db Package 
org.apache.hadoop.streaming Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. 
 

Uses of Reporter in org.apache.hadoop.contrib.index.example
 

Methods in org.apache.hadoop.contrib.index.example with parameters of type Reporter
 RecordReader<DocumentID,LineDocTextAndOp> LineDocInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void IdentityLocalAnalysis.map(DocumentID key, DocumentAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)
           
 void LineDocLocalAnalysis.map(DocumentID key, LineDocTextAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.contrib.index.mapred
 

Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Reporter
 void IndexUpdateMapper.map(K key, V value, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
          Map a key-value pair to a shard-and-intermediate form pair.
 void IndexUpdateCombiner.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
           
 void IndexUpdateReducer.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,Text> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.contrib.utils.join
 

Fields in org.apache.hadoop.contrib.utils.join declared as Reporter
protected  Reporter DataJoinMapperBase.reporter
           
protected  Reporter DataJoinReducerBase.reporter
           
 

Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter
protected  void DataJoinReducerBase.collect(Object key, TaggedMapOutput aRecord, OutputCollector output, Reporter reporter)
          The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected.
 void DataJoinMapperBase.map(Object key, Object value, OutputCollector output, Reporter reporter)
           
 void DataJoinReducerBase.map(Object arg0, Object arg1, OutputCollector arg2, Reporter arg3)
           
 void DataJoinMapperBase.reduce(Object arg0, Iterator arg1, OutputCollector arg2, Reporter arg3)
           
 void DataJoinReducerBase.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.examples
 

Methods in org.apache.hadoop.examples with parameters of type Reporter
 RecordReader<IntWritable,IntWritable> SleepJob.SleepInputFormat.getRecordReader(InputSplit ignored, JobConf conf, Reporter reporter)
           
 RecordReader<MultiFileWordCount.WordOffset,Text> MultiFileWordCount.MyInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void SleepJob.map(IntWritable key, IntWritable value, OutputCollector<IntWritable,NullWritable> output, Reporter reporter)
           
 void PiEstimator.PiMapper.map(LongWritable offset, LongWritable size, OutputCollector<BooleanWritable,LongWritable> out, Reporter reporter)
          Map method.
 void MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void PiEstimator.PiReducer.reduce(BooleanWritable isInside, Iterator<LongWritable> values, OutputCollector<WritableComparable<?>,Writable> output, Reporter reporter)
          Accumulate number of points inside/outside results from the mappers.
 void SleepJob.reduce(IntWritable key, Iterator<NullWritable> values, OutputCollector<NullWritable,NullWritable> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.examples.dancing
 

Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter
 void DistributedPentomino.PentMap.map(WritableComparable key, Text value, OutputCollector<Text,Text> output, Reporter reporter)
          Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order).
 

Uses of Reporter in org.apache.hadoop.examples.terasort
 

Methods in org.apache.hadoop.examples.terasort with parameters of type Reporter
 RecordReader<Text,Text> TeraInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void TeraGen.SortGenMapper.map(LongWritable row, NullWritable ignored, OutputCollector<Text,Text> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred
 

Classes in org.apache.hadoop.mapred that implement Reporter
protected  class Task.TaskReporter
           
 

Fields in org.apache.hadoop.mapred declared as Reporter
static Reporter Reporter.NULL
          A constant of Reporter type that does nothing.
 

Methods in org.apache.hadoop.mapred with parameters of type Reporter
 void RecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 void TextOutputFormat.LineRecordWriter.close(Reporter reporter)
           
 RecordReader<K,V> InputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFilter.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Create a record reader for the given split
 RecordReader<K,V> SequenceFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<LongWritable,Text> TextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> FileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<Text,Text> KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> MultiFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Deprecated.  
 void Task.initialize(JobConf job, JobID id, Reporter reporter, boolean useNewApi)
           
 void Mapper.map(K1 key, V1 value, OutputCollector<K2,V2> output, Reporter reporter)
          Maps a single input key/value pair into an intermediate key/value pair.
 void Reducer.reduce(K2 key, Iterator<V2> values, OutputCollector<K3,V3> output, Reporter reporter)
          Reduces values for a given key.
 void MapRunnable.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
          Start mapping input <key, value> pairs.
 void MapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.mapred with parameters of type Reporter
Task.CombineValuesIterator(RawKeyValueIterator in, RawComparator<KEY> comparator, Class<KEY> keyClass, Class<VALUE> valClass, Configuration conf, Reporter reporter, Counters.Counter combineInputCounter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.join
 

Methods in org.apache.hadoop.mapred.join with parameters of type Reporter
 ComposableRecordReader<K,V> ComposableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 ComposableRecordReader<K,TupleWritable> CompositeInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 

Uses of Reporter in org.apache.hadoop.mapred.lib
 

Fields in org.apache.hadoop.mapred.lib declared as Reporter
protected  Reporter CombineFileRecordReader.reporter
           
 

Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter
 OutputCollector MultipleOutputs.getCollector(String namedOutput, Reporter reporter)
          Gets the output collector for a named output.
 OutputCollector MultipleOutputs.getCollector(String namedOutput, String multiName, Reporter reporter)
          Gets the output collector for a multi named output.
 RecordReader<K,V> DelegatingInputFormat.getRecordReader(InputSplit split, JobConf conf, Reporter reporter)
           
abstract  RecordReader<K,V> CombineFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          This is not implemented yet.
 RecordReader<LongWritable,Text> NLineInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 void DelegatingMapper.map(K1 key, V1 value, OutputCollector<K2,V2> outputCollector, Reporter reporter)
           
 void TokenCountMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void RegexMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void IdentityMapper.map(K key, V val, OutputCollector<K,V> output, Reporter reporter)
          The identify function.
 void FieldSelectionMapReduce.map(K key, V val, OutputCollector<Text,Text> output, Reporter reporter)
          The identify function.
 void InverseMapper.map(K key, V value, OutputCollector<V,K> output, Reporter reporter)
          The inverse function.
 void ChainMapper.map(Object key, Object value, OutputCollector output, Reporter reporter)
          Chains the map(...) methods of the Mappers in the chain.
 void LongSumReducer.reduce(K key, Iterator<LongWritable> values, OutputCollector<K,LongWritable> output, Reporter reporter)
           
 void IdentityReducer.reduce(K key, Iterator<V> values, OutputCollector<K,V> output, Reporter reporter)
          Writes all keys and values directly to output.
 void ChainReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
          Chains the reduce(...) method of the Reducer with the map(...) methods of the Mappers in the chain.
 void FieldSelectionMapReduce.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 void MultithreadedMapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.mapred.lib with parameters of type Reporter
CombineFileRecordReader(JobConf job, CombineFileSplit split, Reporter reporter, Class<RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
 

Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate
 

Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter
 void ValueAggregatorCombiner.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorMapper.map(K1 key, V1 value, OutputCollector<Text,Text> output, Reporter reporter)
          the map function.
 void ValueAggregatorReducer.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorCombiner.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
          Combines values for a given key.
 void ValueAggregatorMapper.reduce(Text arg0, Iterator<Text> arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorReducer.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.lib.db
 

Methods in org.apache.hadoop.mapred.lib.db with parameters of type Reporter
 void DBOutputFormat.DBRecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<LongWritable,T> DBInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 

Uses of Reporter in org.apache.hadoop.streaming
 

Methods in org.apache.hadoop.streaming with parameters of type Reporter
 RecordReader<Text,Text> StreamInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 RecordReader AutoInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void PipeMapper.map(Object key, Object value, OutputCollector output, Reporter reporter)
           
 void PipeReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
           
 void PipeMapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.streaming with parameters of type Reporter
StreamBaseRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
StreamXmlRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
 



Copyright © 2009 The Apache Software Foundation