|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Shard | |
---|---|
org.apache.hadoop.contrib.index.example | |
org.apache.hadoop.contrib.index.lucene | |
org.apache.hadoop.contrib.index.mapred |
Uses of Shard in org.apache.hadoop.contrib.index.example |
---|
Methods in org.apache.hadoop.contrib.index.example with parameters of type Shard | |
---|---|
void |
RoundRobinDistributionPolicy.init(Shard[] shards)
|
void |
HashingDistributionPolicy.init(Shard[] shards)
|
Uses of Shard in org.apache.hadoop.contrib.index.lucene |
---|
Constructors in org.apache.hadoop.contrib.index.lucene with parameters of type Shard | |
---|---|
ShardWriter(FileSystem fs,
Shard shard,
String tempDir,
IndexUpdateConfiguration iconf)
Constructor |
Uses of Shard in org.apache.hadoop.contrib.index.mapred |
---|
Methods in org.apache.hadoop.contrib.index.mapred that return Shard | |
---|---|
static Shard[] |
Shard.getIndexShards(IndexUpdateConfiguration conf)
|
Methods in org.apache.hadoop.contrib.index.mapred that return types with arguments of type Shard | |
---|---|
RecordWriter<Shard,Text> |
IndexUpdateOutputFormat.getRecordWriter(FileSystem fs,
JobConf job,
String name,
Progressable progress)
|
Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Shard | |
---|---|
int |
Shard.compareTo(Shard other)
Compare to another shard. |
int |
IndexUpdatePartitioner.getPartition(Shard key,
IntermediateForm value,
int numPartitions)
|
void |
IDistributionPolicy.init(Shard[] shards)
Initialization. |
void |
IndexUpdateCombiner.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,IntermediateForm> output,
Reporter reporter)
|
void |
IndexUpdateReducer.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,Text> output,
Reporter reporter)
|
void |
IndexUpdater.run(Configuration conf,
Path[] inputPaths,
Path outputPath,
int numMapTasks,
Shard[] shards)
|
void |
IIndexUpdater.run(Configuration conf,
Path[] inputPaths,
Path outputPath,
int numMapTasks,
Shard[] shards)
Create a Map/Reduce job configuration and run the Map/Reduce job to analyze documents and update Lucene instances in parallel. |
static void |
Shard.setIndexShards(IndexUpdateConfiguration conf,
Shard[] shards)
|
Method parameters in org.apache.hadoop.contrib.index.mapred with type arguments of type Shard | |
---|---|
void |
IndexUpdateMapper.map(K key,
V value,
OutputCollector<Shard,IntermediateForm> output,
Reporter reporter)
Map a key-value pair to a shard-and-intermediate form pair. |
void |
IndexUpdateCombiner.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,IntermediateForm> output,
Reporter reporter)
|
void |
IndexUpdateReducer.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,Text> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.contrib.index.mapred with parameters of type Shard | |
---|---|
Shard(Shard shard)
Construct using a shard object. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |