|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Reporter | |
---|---|
org.apache.hadoop.contrib.index.example | |
org.apache.hadoop.contrib.index.mapred | |
org.apache.hadoop.contrib.utils.join | |
org.apache.hadoop.examples | Hadoop example code. |
org.apache.hadoop.examples.dancing | This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. |
org.apache.hadoop.examples.terasort | This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition. |
org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. |
org.apache.hadoop.mapred.join | Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. |
org.apache.hadoop.mapred.lib | Library of generally useful mappers, reducers, and partitioners. |
org.apache.hadoop.mapred.lib.aggregate | Classes for performing various counting and aggregations. |
org.apache.hadoop.mapred.lib.db | org.apache.hadoop.mapred.lib.db Package |
org.apache.hadoop.streaming | Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. |
Uses of Reporter in org.apache.hadoop.contrib.index.example |
---|
Methods in org.apache.hadoop.contrib.index.example with parameters of type Reporter | |
---|---|
RecordReader<DocumentID,LineDocTextAndOp> |
LineDocInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
void |
IdentityLocalAnalysis.map(DocumentID key,
DocumentAndOp value,
OutputCollector<DocumentID,DocumentAndOp> output,
Reporter reporter)
|
void |
LineDocLocalAnalysis.map(DocumentID key,
LineDocTextAndOp value,
OutputCollector<DocumentID,DocumentAndOp> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.contrib.index.mapred |
---|
Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Reporter | |
---|---|
void |
IndexUpdateMapper.map(K key,
V value,
OutputCollector<Shard,IntermediateForm> output,
Reporter reporter)
Map a key-value pair to a shard-and-intermediate form pair. |
void |
IndexUpdateCombiner.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,IntermediateForm> output,
Reporter reporter)
|
void |
IndexUpdateReducer.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,Text> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.contrib.utils.join |
---|
Fields in org.apache.hadoop.contrib.utils.join declared as Reporter | |
---|---|
protected Reporter |
DataJoinMapperBase.reporter
|
protected Reporter |
DataJoinReducerBase.reporter
|
Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter | |
---|---|
protected void |
DataJoinReducerBase.collect(Object key,
TaggedMapOutput aRecord,
OutputCollector output,
Reporter reporter)
The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected. |
void |
DataJoinMapperBase.map(Object key,
Object value,
OutputCollector output,
Reporter reporter)
|
void |
DataJoinReducerBase.map(Object arg0,
Object arg1,
OutputCollector arg2,
Reporter arg3)
|
void |
DataJoinMapperBase.reduce(Object arg0,
Iterator arg1,
OutputCollector arg2,
Reporter arg3)
|
void |
DataJoinReducerBase.reduce(Object key,
Iterator values,
OutputCollector output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.examples |
---|
Uses of Reporter in org.apache.hadoop.examples.dancing |
---|
Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Uses of Reporter in org.apache.hadoop.examples.terasort |
---|
Methods in org.apache.hadoop.examples.terasort with parameters of type Reporter | |
---|---|
RecordReader<Text,Text> |
TeraInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
void |
TeraGen.SortGenMapper.map(LongWritable row,
NullWritable ignored,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.mapred |
---|
Classes in org.apache.hadoop.mapred that implement Reporter | |
---|---|
protected class |
Task.TaskReporter
|
Fields in org.apache.hadoop.mapred declared as Reporter | |
---|---|
static Reporter |
Reporter.NULL
A constant of Reporter type that does nothing. |
Methods in org.apache.hadoop.mapred with parameters of type Reporter | |
---|---|
void |
RecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
void |
TextOutputFormat.LineRecordWriter.close(Reporter reporter)
|
RecordReader<K,V> |
InputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Get the RecordReader for the given InputSplit . |
RecordReader<BytesWritable,BytesWritable> |
SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<K,V> |
SequenceFileInputFilter.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Create a record reader for the given split |
RecordReader<K,V> |
SequenceFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<LongWritable,Text> |
TextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
FileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
MultiFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Deprecated. |
void |
Task.initialize(JobConf job,
JobID id,
Reporter reporter,
boolean useNewApi)
|
void |
Mapper.map(K1 key,
V1 value,
OutputCollector<K2,V2> output,
Reporter reporter)
Maps a single input key/value pair into an intermediate key/value pair. |
void |
Reducer.reduce(K2 key,
Iterator<V2> values,
OutputCollector<K3,V3> output,
Reporter reporter)
Reduces values for a given key. |
void |
MapRunnable.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
Start mapping input <key, value> pairs. |
void |
MapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.mapred with parameters of type Reporter | |
---|---|
Task.CombineValuesIterator(RawKeyValueIterator in,
RawComparator<KEY> comparator,
Class<KEY> keyClass,
Class<VALUE> valClass,
Configuration conf,
Reporter reporter,
Counters.Counter combineInputCounter)
|
Uses of Reporter in org.apache.hadoop.mapred.join |
---|
Methods in org.apache.hadoop.mapred.join with parameters of type Reporter | |
---|---|
ComposableRecordReader<K,V> |
ComposableInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
ComposableRecordReader<K,TupleWritable> |
CompositeInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression. |
Uses of Reporter in org.apache.hadoop.mapred.lib |
---|
Fields in org.apache.hadoop.mapred.lib declared as Reporter | |
---|---|
protected Reporter |
CombineFileRecordReader.reporter
|
Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter | |
---|---|
OutputCollector |
MultipleOutputs.getCollector(String namedOutput,
Reporter reporter)
Gets the output collector for a named output. |
OutputCollector |
MultipleOutputs.getCollector(String namedOutput,
String multiName,
Reporter reporter)
Gets the output collector for a multi named output. |
RecordReader<K,V> |
DelegatingInputFormat.getRecordReader(InputSplit split,
JobConf conf,
Reporter reporter)
|
abstract RecordReader<K,V> |
CombineFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
This is not implemented yet. |
RecordReader<LongWritable,Text> |
NLineInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
void |
DelegatingMapper.map(K1 key,
V1 value,
OutputCollector<K2,V2> outputCollector,
Reporter reporter)
|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
IdentityMapper.map(K key,
V val,
OutputCollector<K,V> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
InverseMapper.map(K key,
V value,
OutputCollector<V,K> output,
Reporter reporter)
The inverse function. |
void |
ChainMapper.map(Object key,
Object value,
OutputCollector output,
Reporter reporter)
Chains the map(...) methods of the Mappers in the chain. |
void |
LongSumReducer.reduce(K key,
Iterator<LongWritable> values,
OutputCollector<K,LongWritable> output,
Reporter reporter)
|
void |
IdentityReducer.reduce(K key,
Iterator<V> values,
OutputCollector<K,V> output,
Reporter reporter)
Writes all keys and values directly to output. |
void |
ChainReducer.reduce(Object key,
Iterator values,
OutputCollector output,
Reporter reporter)
Chains the reduce(...) method of the Reducer with the
map(...) methods of the Mappers in the chain. |
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
MultithreadedMapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.mapred.lib with parameters of type Reporter | |
---|---|
CombineFileRecordReader(JobConf job,
CombineFileSplit split,
Reporter reporter,
Class<RecordReader<K,V>> rrClass)
A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit. |
Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate |
---|
Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter | |
---|---|
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.mapred.lib.db |
---|
Methods in org.apache.hadoop.mapred.lib.db with parameters of type Reporter | |
---|---|
void |
DBOutputFormat.DBRecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
RecordReader<LongWritable,T> |
DBInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Get the RecordReader for the given InputSplit . |
Uses of Reporter in org.apache.hadoop.streaming |
---|
Methods in org.apache.hadoop.streaming with parameters of type Reporter | |
---|---|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader |
AutoInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
void |
PipeMapper.map(Object key,
Object value,
OutputCollector output,
Reporter reporter)
|
void |
PipeReducer.reduce(Object key,
Iterator values,
OutputCollector output,
Reporter reporter)
|
void |
PipeMapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.streaming with parameters of type Reporter | |
---|---|
StreamBaseRecordReader(FSDataInputStream in,
FileSplit split,
Reporter reporter,
JobConf job,
FileSystem fs)
|
|
StreamXmlRecordReader(FSDataInputStream in,
FileSplit split,
Reporter reporter,
JobConf job,
FileSystem fs)
|
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |