Uses of Interface

Packages that use Reporter
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.examples.dancing This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. 
org.apache.hadoop.examples.terasort This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition. 
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapred.join Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. 
org.apache.hadoop.mapred.lib Library of generally useful mappers, reducers, and partitioners. 
org.apache.hadoop.mapred.lib.aggregate Classes for performing various counting and aggregations. 
org.apache.hadoop.mapred.lib.db org.apache.hadoop.mapred.lib.db Package 
org.apache.hadoop.streaming Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. 

Uses of Reporter in org.apache.hadoop.contrib.index.example

Methods in org.apache.hadoop.contrib.index.example with parameters of type Reporter
 RecordReader<DocumentID,LineDocTextAndOp> LineDocInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 void key, DocumentAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)
 void key, LineDocTextAndOp value, OutputCollector<DocumentID,DocumentAndOp> output, Reporter reporter)

Uses of Reporter in org.apache.hadoop.contrib.index.mapred

Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Reporter
 void key, V value, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
          Map a key-value pair to a shard-and-intermediate form pair.
 void IndexUpdateCombiner.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,IntermediateForm> output, Reporter reporter)
 void IndexUpdateReducer.reduce(Shard key, Iterator<IntermediateForm> values, OutputCollector<Shard,Text> output, Reporter reporter)

Uses of Reporter in org.apache.hadoop.contrib.utils.join

Fields in org.apache.hadoop.contrib.utils.join declared as Reporter
protected  Reporter DataJoinMapperBase.reporter
protected  Reporter DataJoinReducerBase.reporter

Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter
protected  void DataJoinReducerBase.collect(Object key, TaggedMapOutput aRecord, OutputCollector output, Reporter reporter)
          The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected.
 void key, Object value, OutputCollector output, Reporter reporter)
 void arg0, Object arg1, OutputCollector arg2, Reporter arg3)
 void DataJoinMapperBase.reduce(Object arg0, Iterator arg1, OutputCollector arg2, Reporter arg3)
 void DataJoinReducerBase.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)

Uses of Reporter in org.apache.hadoop.examples

Methods in org.apache.hadoop.examples with parameters of type Reporter
 RecordReader<IntWritable,IntWritable> SleepJob.SleepInputFormat.getRecordReader(InputSplit ignored, JobConf conf, Reporter reporter)
 RecordReader<MultiFileWordCount.WordOffset,Text> MultiFileWordCount.MyInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 void key, IntWritable value, OutputCollector<IntWritable,NullWritable> output, Reporter reporter)
 void offset, LongWritable size, OutputCollector<BooleanWritable,LongWritable> out, Reporter reporter)
          Map method.
 void key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
 void PiEstimator.PiReducer.reduce(BooleanWritable isInside, Iterator<LongWritable> values, OutputCollector<WritableComparable<?>,Writable> output, Reporter reporter)
          Accumulate number of points inside/outside results from the mappers.
 void SleepJob.reduce(IntWritable key, Iterator<NullWritable> values, OutputCollector<NullWritable,NullWritable> output, Reporter reporter)

Uses of Reporter in org.apache.hadoop.examples.dancing

Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter
 void key, Text value, OutputCollector<Text,Text> output, Reporter reporter)
          Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order).

Uses of Reporter in org.apache.hadoop.examples.terasort

Methods in org.apache.hadoop.examples.terasort with parameters of type Reporter
 RecordReader<Text,Text> TeraInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 void row, NullWritable ignored, OutputCollector<Text,Text> output, Reporter reporter)

Uses of Reporter in org.apache.hadoop.mapred

Classes in org.apache.hadoop.mapred that implement Reporter
protected  class Task.TaskReporter

Fields in org.apache.hadoop.mapred declared as Reporter
static Reporter Reporter.NULL
          A constant of Reporter type that does nothing.

Methods in org.apache.hadoop.mapred with parameters of type Reporter
 void RecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 void TextOutputFormat.LineRecordWriter.close(Reporter reporter)
 RecordReader<K,V> SequenceFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 RecordReader<LongWritable,Text> TextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
 RecordReader<K,V> SequenceFileInputFilter.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Create a record reader for the given split
abstract  RecordReader<K,V> MultiFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 RecordReader<K,V> InputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 RecordReader<Text,Text> KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
abstract  RecordReader<K,V> FileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 void Task.initialize(JobConf job, JobID id, Reporter reporter, boolean useNewApi)
 void key, V1 value, OutputCollector<K2,V2> output, Reporter reporter)
          Maps a single input key/value pair into an intermediate key/value pair.
 void Reducer.reduce(K2 key, Iterator<V2> values, OutputCollector<K3,V3> output, Reporter reporter)
          Reduces values for a given key.
 void<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
          Start mapping input <key, value> pairs.
 void<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)

Constructors in org.apache.hadoop.mapred with parameters of type Reporter
Task.CombineValuesIterator(RawKeyValueIterator in, RawComparator<KEY> comparator, Class<KEY> keyClass, Class<VALUE> valClass, Configuration conf, Reporter reporter, Counters.Counter combineInputCounter)

Uses of Reporter in org.apache.hadoop.mapred.join

Methods in org.apache.hadoop.mapred.join with parameters of type Reporter
 ComposableRecordReader<K,V> ComposableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 ComposableRecordReader<K,TupleWritable> CompositeInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.

Uses of Reporter in org.apache.hadoop.mapred.lib

Fields in org.apache.hadoop.mapred.lib declared as Reporter
protected  Reporter CombineFileRecordReader.reporter

Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter
 OutputCollector MultipleOutputs.getCollector(String namedOutput, Reporter reporter)
          Gets the output collector for a named output.
 OutputCollector MultipleOutputs.getCollector(String namedOutput, String multiName, Reporter reporter)
          Gets the output collector for a multi named output.
 RecordReader<LongWritable,Text> NLineInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
abstract  RecordReader<K,V> CombineFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          This is not implemented yet.
 RecordReader<K,V> DelegatingInputFormat.getRecordReader(InputSplit split, JobConf conf, Reporter reporter)
 void key, V1 value, OutputCollector<K2,V2> outputCollector, Reporter reporter)
 void key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
 void key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
 void key, V val, OutputCollector<K,V> output, Reporter reporter)
          The identify function.
 void key, V val, OutputCollector<Text,Text> output, Reporter reporter)
          The identify function.
 void key, V value, OutputCollector<V,K> output, Reporter reporter)
          The inverse function.
 void key, Object value, OutputCollector output, Reporter reporter)
          Chains the map(...) methods of the Mappers in the chain.
 void LongSumReducer.reduce(K key, Iterator<LongWritable> values, OutputCollector<K,LongWritable> output, Reporter reporter)
 void IdentityReducer.reduce(K key, Iterator<V> values, OutputCollector<K,V> output, Reporter reporter)
          Writes all keys and values directly to output.
 void ChainReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
          Chains the reduce(...) method of the Reducer with the map(...) methods of the Mappers in the chain.
 void FieldSelectionMapReduce.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
 void<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)

Constructors in org.apache.hadoop.mapred.lib with parameters of type Reporter
CombineFileRecordReader(JobConf job, CombineFileSplit split, Reporter reporter, Class<RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.

Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate

Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter
 void arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void key, V1 value, OutputCollector<Text,Text> output, Reporter reporter)
          the map function.
 void ValueAggregatorReducer.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
 void ValueAggregatorCombiner.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
          Combines values for a given key.
 void ValueAggregatorMapper.reduce(Text arg0, Iterator<Text> arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.

Uses of Reporter in org.apache.hadoop.mapred.lib.db

Methods in org.apache.hadoop.mapred.lib.db with parameters of type Reporter
 void DBOutputFormat.DBRecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<LongWritable,T> DBInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.

Uses of Reporter in org.apache.hadoop.streaming

Methods in org.apache.hadoop.streaming with parameters of type Reporter
 RecordReader AutoInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
 RecordReader<Text,Text> StreamInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
 void key, Object value, OutputCollector output, Reporter reporter)
 void PipeReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
 void<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)

Constructors in org.apache.hadoop.streaming with parameters of type Reporter
StreamBaseRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
StreamXmlRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)

Copyright © 2009 The Apache Software Foundation