| Package | Description | 
|---|---|
| org.apache.hadoop.io | Generic i/o code for use when reading and writing data to the network,
to databases, and to files. | 
| org.apache.hadoop.mapred | |
| org.apache.hadoop.mapred.lib | |
| org.apache.hadoop.mapred.lib.db | |
| org.apache.hadoop.mapreduce.lib.db | |
| org.apache.hadoop.mapreduce.lib.input | |
| org.apache.hadoop.mapreduce.lib.reduce | 
| Modifier and Type | Method and Description | 
|---|---|
| int | LongWritable. compareTo(LongWritable o)Compares two LongWritables. | 
| Modifier and Type | Method and Description | 
|---|---|
| RecordReader<LongWritable,BytesWritable> | FixedLengthInputFormat. getRecordReader(InputSplit genericSplit,
               JobConf job,
               Reporter reporter) | 
| RecordReader<LongWritable,Text> | TextInputFormat. getRecordReader(InputSplit genericSplit,
               JobConf job,
               Reporter reporter) | 
| Modifier and Type | Method and Description | 
|---|---|
| RecordReader<LongWritable,Text> | CombineTextInputFormat. getRecordReader(InputSplit split,
               JobConf conf,
               Reporter reporter) | 
| RecordReader<LongWritable,Text> | NLineInputFormat. getRecordReader(InputSplit genericSplit,
               JobConf job,
               Reporter reporter) | 
| Modifier and Type | Method and Description | 
|---|---|
| void | TokenCountMapper. map(K key,
   Text value,
   OutputCollector<Text,LongWritable> output,
   Reporter reporter) | 
| void | RegexMapper. map(K key,
   Text value,
   OutputCollector<Text,LongWritable> output,
   Reporter reporter) | 
| void | LongSumReducer. reduce(K key,
      Iterator<LongWritable> values,
      OutputCollector<K,LongWritable> output,
      Reporter reporter) | 
| void | LongSumReducer. reduce(K key,
      Iterator<LongWritable> values,
      OutputCollector<K,LongWritable> output,
      Reporter reporter) | 
| Modifier and Type | Method and Description | 
|---|---|
| RecordReader<LongWritable,T> | DBInputFormat. getRecordReader(InputSplit split,
               JobConf job,
               Reporter reporter)Get the  RecordReaderfor the givenInputSplit. | 
| Modifier and Type | Method and Description | 
|---|---|
| LongWritable | DBRecordReader. getCurrentKey()Get the current key | 
| Modifier and Type | Method and Description | 
|---|---|
| protected RecordReader<LongWritable,T> | DataDrivenDBInputFormat. createDBRecordReader(org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit split,
                    Configuration conf) | 
| protected RecordReader<LongWritable,T> | DBInputFormat. createDBRecordReader(org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit split,
                    Configuration conf) | 
| protected RecordReader<LongWritable,T> | OracleDataDrivenDBInputFormat. createDBRecordReader(org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit split,
                    Configuration conf) | 
| RecordReader<LongWritable,T> | DBInputFormat. createRecordReader(InputSplit split,
                  TaskAttemptContext context)Create a record reader for a given split. | 
| Modifier and Type | Method and Description | 
|---|---|
| boolean | DBRecordReader. next(LongWritable key,
    T value)Deprecated. 
 | 
| Modifier and Type | Method and Description | 
|---|---|
| RecordReader<LongWritable,Text> | CombineTextInputFormat. createRecordReader(InputSplit split,
                  TaskAttemptContext context) | 
| RecordReader<LongWritable,BytesWritable> | FixedLengthInputFormat. createRecordReader(InputSplit split,
                  TaskAttemptContext context) | 
| RecordReader<LongWritable,Text> | TextInputFormat. createRecordReader(InputSplit split,
                  TaskAttemptContext context) | 
| RecordReader<LongWritable,Text> | NLineInputFormat. createRecordReader(InputSplit genericSplit,
                  TaskAttemptContext context) | 
| Modifier and Type | Method and Description | 
|---|---|
| void | LongSumReducer. reduce(KEY key,
      Iterable<LongWritable> values,
      org.apache.hadoop.mapreduce.Reducer.Context context) | 
Copyright © 2018 Apache Software Foundation. All rights reserved.