|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Text | |
---|---|
org.apache.hadoop.contrib.index.example | |
org.apache.hadoop.contrib.index.mapred | |
org.apache.hadoop.contrib.utils.join | |
org.apache.hadoop.examples | Hadoop example code. |
org.apache.hadoop.examples.dancing | This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. |
org.apache.hadoop.examples.terasort | This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition. |
org.apache.hadoop.io | Generic i/o code for use when reading and writing data to the network, to databases, and to files. |
org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. |
org.apache.hadoop.mapred.lib | Library of generally useful mappers, reducers, and partitioners. |
org.apache.hadoop.mapred.lib.aggregate | Classes for performing various counting and aggregations. |
org.apache.hadoop.mapreduce.lib.fieldsel | |
org.apache.hadoop.mapreduce.lib.input | |
org.apache.hadoop.mapreduce.lib.map | |
org.apache.hadoop.mapreduce.security | |
org.apache.hadoop.mapreduce.security.token | |
org.apache.hadoop.mapreduce.security.token.delegation | |
org.apache.hadoop.security | |
org.apache.hadoop.security.token | |
org.apache.hadoop.security.token.delegation | |
org.apache.hadoop.streaming | Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. |
org.apache.hadoop.streaming.io | |
org.apache.hadoop.typedbytes | Typed bytes are sequences of bytes in which the first byte is a type code. |
org.apache.hadoop.util | Common utilities. |
Uses of Text in org.apache.hadoop.contrib.index.example |
---|
Methods in org.apache.hadoop.contrib.index.example that return Text | |
---|---|
Text |
LineDocTextAndOp.getText()
Get the text that represents a document. |
Uses of Text in org.apache.hadoop.contrib.index.mapred |
---|
Methods in org.apache.hadoop.contrib.index.mapred that return Text | |
---|---|
Text |
DocumentID.getText()
The text of the document id. |
Methods in org.apache.hadoop.contrib.index.mapred that return types with arguments of type Text | |
---|---|
RecordWriter<Shard,Text> |
IndexUpdateOutputFormat.getRecordWriter(FileSystem fs,
JobConf job,
String name,
Progressable progress)
|
Method parameters in org.apache.hadoop.contrib.index.mapred with type arguments of type Text | |
---|---|
void |
IndexUpdateReducer.reduce(Shard key,
Iterator<IntermediateForm> values,
OutputCollector<Shard,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.contrib.utils.join |
---|
Fields in org.apache.hadoop.contrib.utils.join declared as Text | |
---|---|
protected Text |
DataJoinMapperBase.inputTag
|
static Text |
DataJoinReducerBase.NUM_OF_VALUES_FIELD
|
static Text |
DataJoinReducerBase.SOURCE_TAGS_FIELD
|
protected Text |
TaggedMapOutput.tag
|
Methods in org.apache.hadoop.contrib.utils.join that return Text | |
---|---|
protected abstract Text |
DataJoinMapperBase.generateGroupKey(TaggedMapOutput aRecord)
Generate a map output key. |
protected abstract Text |
DataJoinMapperBase.generateInputTag(String inputFile)
Determine the source tag based on the input file name. |
Text |
TaggedMapOutput.getTag()
|
Methods in org.apache.hadoop.contrib.utils.join with parameters of type Text | |
---|---|
void |
TaggedMapOutput.setTag(Text tag)
|
Uses of Text in org.apache.hadoop.examples |
---|
Methods in org.apache.hadoop.examples that return Text | |
---|---|
Text |
MultiFileWordCount.MultiFileLineRecordReader.createValue()
|
Methods in org.apache.hadoop.examples that return types with arguments of type Text | |
---|---|
ArrayList<Map.Entry<Text,Text>> |
AggregateWordHistogram.AggregateWordHistogramPlugin.generateKeyValPairs(Object key,
Object val)
Parse the given value, generate an aggregation-id/value pair per word. |
ArrayList<Map.Entry<Text,Text>> |
AggregateWordHistogram.AggregateWordHistogramPlugin.generateKeyValPairs(Object key,
Object val)
Parse the given value, generate an aggregation-id/value pair per word. |
ArrayList<Map.Entry<Text,Text>> |
AggregateWordCount.WordCountPlugInClass.generateKeyValPairs(Object key,
Object val)
|
ArrayList<Map.Entry<Text,Text>> |
AggregateWordCount.WordCountPlugInClass.generateKeyValPairs(Object key,
Object val)
|
RecordReader<MultiFileWordCount.WordOffset,Text> |
MultiFileWordCount.MyInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.examples with parameters of type Text | |
---|---|
void |
SecondarySort.MapClass.map(LongWritable inKey,
Text inValue,
Mapper.Context context)
|
void |
MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
WordCount.TokenizerMapper.map(Object key,
Text value,
Mapper.Context context)
|
boolean |
MultiFileWordCount.MultiFileLineRecordReader.next(MultiFileWordCount.WordOffset key,
Text value)
|
void |
WordCount.IntSumReducer.reduce(Text key,
Iterable<IntWritable> values,
Reducer.Context context)
|
Method parameters in org.apache.hadoop.examples with type arguments of type Text | |
---|---|
void |
MultiFileWordCount.MapClass.map(MultiFileWordCount.WordOffset key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.examples.dancing |
---|
Methods in org.apache.hadoop.examples.dancing with parameters of type Text | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Method parameters in org.apache.hadoop.examples.dancing with type arguments of type Text | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Uses of Text in org.apache.hadoop.examples.terasort |
---|
Methods in org.apache.hadoop.examples.terasort that return types with arguments of type Text | |
---|---|
RecordReader<Text,Text> |
TeraInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
TeraInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordWriter<Text,Text> |
TeraOutputFormat.getRecordWriter(FileSystem ignored,
JobConf job,
String name,
Progressable progress)
|
RecordWriter<Text,Text> |
TeraOutputFormat.getRecordWriter(FileSystem ignored,
JobConf job,
String name,
Progressable progress)
|
Method parameters in org.apache.hadoop.examples.terasort with type arguments of type Text | |
---|---|
void |
TeraGen.SortGenMapper.map(LongWritable row,
NullWritable ignored,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
TeraGen.SortGenMapper.map(LongWritable row,
NullWritable ignored,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.io |
---|
Methods in org.apache.hadoop.io that return Text | |
---|---|
Text |
SequenceFile.Metadata.get(Text name)
|
Methods in org.apache.hadoop.io that return types with arguments of type Text | |
---|---|
TreeMap<Text,Text> |
SequenceFile.Metadata.getMetadata()
|
TreeMap<Text,Text> |
SequenceFile.Metadata.getMetadata()
|
Methods in org.apache.hadoop.io with parameters of type Text | |
---|---|
Text |
SequenceFile.Metadata.get(Text name)
|
void |
Text.set(Text other)
copy a text. |
void |
SequenceFile.Metadata.set(Text name,
Text value)
|
Constructors in org.apache.hadoop.io with parameters of type Text | |
---|---|
Text(Text utf8)
Construct from another text. |
Constructor parameters in org.apache.hadoop.io with type arguments of type Text | |
---|---|
SequenceFile.Metadata(TreeMap<Text,Text> arg)
|
|
SequenceFile.Metadata(TreeMap<Text,Text> arg)
|
Uses of Text in org.apache.hadoop.mapred |
---|
Methods in org.apache.hadoop.mapred that return Text | |
---|---|
Text |
SequenceFileAsTextRecordReader.createKey()
|
Text |
KeyValueLineRecordReader.createKey()
|
Text |
SequenceFileAsTextRecordReader.createValue()
|
Text |
LineRecordReader.createValue()
|
Text |
KeyValueLineRecordReader.createValue()
|
Methods in org.apache.hadoop.mapred that return types with arguments of type Text | |
---|---|
RecordReader<LongWritable,Text> |
TextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.mapred with parameters of type Text | |
---|---|
Token<DelegationTokenIdentifier> |
JobClient.getDelegationToken(Text renewer)
|
Token<DelegationTokenIdentifier> |
JobTracker.getDelegationToken(Text renewer)
Get a new delegation token. |
Token<DelegationTokenIdentifier> |
LocalJobRunner.getDelegationToken(Text renewer)
|
boolean |
JobClient.Renewer.handleKind(Text kind)
|
boolean |
LineRecordReader.next(LongWritable key,
Text value)
Read a line. |
boolean |
SequenceFileAsTextRecordReader.next(Text key,
Text value)
Read key/value pair in a line. |
boolean |
KeyValueLineRecordReader.next(Text key,
Text value)
Read key/value pair in a line. |
Uses of Text in org.apache.hadoop.mapred.lib |
---|
Methods in org.apache.hadoop.mapred.lib that return types with arguments of type Text | |
---|---|
RecordReader<LongWritable,Text> |
NLineInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.mapred.lib with parameters of type Text | |
---|---|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Method parameters in org.apache.hadoop.mapred.lib with type arguments of type Text | |
---|---|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.mapred.lib.aggregate |
---|
Fields in org.apache.hadoop.mapred.lib.aggregate declared as Text | |
---|---|
static Text |
ValueAggregatorDescriptor.ONE
|
Methods in org.apache.hadoop.mapred.lib.aggregate that return types with arguments of type Text | |
---|---|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
ArrayList<Map.Entry<Text,Text>> |
UserDefinedValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pairs by delegating the invocation to the real object. |
ArrayList<Map.Entry<Text,Text>> |
UserDefinedValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pairs by delegating the invocation to the real object. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorDescriptor.generateKeyValPairs(Object key,
Object val)
Generate a list of aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorBaseDescriptor.generateKeyValPairs(Object key,
Object val)
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair. |
ArrayList<Map.Entry<Text,Text>> |
ValueAggregatorBaseDescriptor.generateKeyValPairs(Object key,
Object val)
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair. |
Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Text | |
---|---|
static Map.Entry<Text,Text> |
ValueAggregatorBaseDescriptor.generateEntry(String type,
String id,
Text val)
|
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Method parameters in org.apache.hadoop.mapred.lib.aggregate with type arguments of type Text | |
---|---|
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
Uses of Text in org.apache.hadoop.mapreduce.lib.fieldsel |
---|
Fields in org.apache.hadoop.mapreduce.lib.fieldsel declared as Text | |
---|---|
static Text |
FieldSelectionHelper.emptyText
|
Methods in org.apache.hadoop.mapreduce.lib.fieldsel that return Text | |
---|---|
Text |
FieldSelectionHelper.getKey()
|
Text |
FieldSelectionHelper.getValue()
|
Methods in org.apache.hadoop.mapreduce.lib.fieldsel with parameters of type Text | |
---|---|
void |
FieldSelectionReducer.reduce(Text key,
Iterable<Text> values,
Reducer.Context context)
|
Method parameters in org.apache.hadoop.mapreduce.lib.fieldsel with type arguments of type Text | |
---|---|
void |
FieldSelectionReducer.reduce(Text key,
Iterable<Text> values,
Reducer.Context context)
|
Constructors in org.apache.hadoop.mapreduce.lib.fieldsel with parameters of type Text | |
---|---|
FieldSelectionHelper(Text key,
Text val)
|
Uses of Text in org.apache.hadoop.mapreduce.lib.input |
---|
Methods in org.apache.hadoop.mapreduce.lib.input that return Text | |
---|---|
Text |
SequenceFileAsTextRecordReader.getCurrentKey()
|
Text |
KeyValueLineRecordReader.getCurrentKey()
|
Text |
SequenceFileAsTextRecordReader.getCurrentValue()
|
Text |
LineRecordReader.getCurrentValue()
|
Text |
KeyValueLineRecordReader.getCurrentValue()
|
Methods in org.apache.hadoop.mapreduce.lib.input that return types with arguments of type Text | |
---|---|
RecordReader<LongWritable,Text> |
NLineInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context)
|
RecordReader<LongWritable,Text> |
TextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context)
|
Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type Text | |
---|---|
static void |
KeyValueLineRecordReader.setKeyValue(Text key,
Text value,
byte[] line,
int lineLen,
int pos)
|
Uses of Text in org.apache.hadoop.mapreduce.lib.map |
---|
Methods in org.apache.hadoop.mapreduce.lib.map with parameters of type Text | |
---|---|
void |
TokenCounterMapper.map(Object key,
Text value,
Mapper.Context context)
|
Uses of Text in org.apache.hadoop.mapreduce.security |
---|
Methods in org.apache.hadoop.mapreduce.security with parameters of type Text | |
---|---|
static byte[] |
TokenCache.getSecretKey(Credentials credentials,
Text alias)
auxiliary method to get user's secret keys.. |
Uses of Text in org.apache.hadoop.mapreduce.security.token |
---|
Fields in org.apache.hadoop.mapreduce.security.token declared as Text | |
---|---|
static Text |
JobTokenIdentifier.KIND_NAME
|
Methods in org.apache.hadoop.mapreduce.security.token that return Text | |
---|---|
Text |
JobTokenIdentifier.getJobId()
Get the jobid |
Text |
JobTokenIdentifier.getKind()
Get the token kind |
protected Text |
JobTokenIdentifier.Renewer.getKind()
|
Methods in org.apache.hadoop.mapreduce.security.token with parameters of type Text | |
---|---|
Token<JobTokenIdentifier> |
JobTokenSelector.selectToken(Text service,
Collection<Token<? extends TokenIdentifier>> tokens)
|
Constructors in org.apache.hadoop.mapreduce.security.token with parameters of type Text | |
---|---|
JobTokenIdentifier(Text jobid)
Create a job token identifier from a jobid |
Uses of Text in org.apache.hadoop.mapreduce.security.token.delegation |
---|
Fields in org.apache.hadoop.mapreduce.security.token.delegation declared as Text | |
---|---|
static Text |
DelegationTokenIdentifier.MAPREDUCE_DELEGATION_KIND
|
Methods in org.apache.hadoop.mapreduce.security.token.delegation that return Text | |
---|---|
Text |
DelegationTokenIdentifier.getKind()
|
Constructors in org.apache.hadoop.mapreduce.security.token.delegation with parameters of type Text | |
---|---|
DelegationTokenIdentifier(Text owner,
Text renewer,
Text realUser)
Create a new delegation token identifier |
Uses of Text in org.apache.hadoop.security |
---|
Methods in org.apache.hadoop.security that return Text | |
---|---|
static Text |
SecurityUtil.buildTokenService(InetSocketAddress addr)
Construct the service key for a token |
Methods in org.apache.hadoop.security with parameters of type Text | |
---|---|
void |
Credentials.addSecretKey(Text alias,
byte[] key)
Set the key for an alias |
void |
Credentials.addToken(Text alias,
Token<? extends TokenIdentifier> t)
Add a token in the storage (in memory) |
byte[] |
Credentials.getSecretKey(Text alias)
Returns the key bytes for the alias |
Token<? extends TokenIdentifier> |
Credentials.getToken(Text alias)
Returns the Token object for the alias |
Uses of Text in org.apache.hadoop.security.token |
---|
Methods in org.apache.hadoop.security.token that return Text | |
---|---|
abstract Text |
TokenIdentifier.getKind()
Get the token kind |
Text |
Token.getKind()
Get the token kind |
protected Text |
Token.TrivialRenewer.getKind()
|
Text |
Token.getService()
Get the service on which the token is supposed to be used |
Methods in org.apache.hadoop.security.token with parameters of type Text | |
---|---|
boolean |
Token.TrivialRenewer.handleKind(Text kind)
|
abstract boolean |
TokenRenewer.handleKind(Text kind)
Does this renewer handle this kind of token? |
Token<T> |
TokenSelector.selectToken(Text service,
Collection<Token<? extends TokenIdentifier>> tokens)
|
void |
Token.setKind(Text newKind)
Set the token kind. |
void |
Token.setService(Text newService)
Set the service on which the token is supposed to be used |
Constructors in org.apache.hadoop.security.token with parameters of type Text | |
---|---|
Token(byte[] identifier,
byte[] password,
Text kind,
Text service)
Construct a token from the components. |
Uses of Text in org.apache.hadoop.security.token.delegation |
---|
Methods in org.apache.hadoop.security.token.delegation that return Text | |
---|---|
abstract Text |
AbstractDelegationTokenIdentifier.getKind()
|
Text |
AbstractDelegationTokenIdentifier.getRenewer()
|
Methods in org.apache.hadoop.security.token.delegation with parameters of type Text | |
---|---|
Token<TokenIdent> |
AbstractDelegationTokenSelector.selectToken(Text service,
Collection<Token<? extends TokenIdentifier>> tokens)
|
Constructors in org.apache.hadoop.security.token.delegation with parameters of type Text | |
---|---|
AbstractDelegationTokenIdentifier(Text owner,
Text renewer,
Text realUser)
|
|
AbstractDelegationTokenSelector(Text kindName)
|
Uses of Text in org.apache.hadoop.streaming |
---|
Methods in org.apache.hadoop.streaming that return Text | |
---|---|
Text |
StreamBaseRecordReader.createKey()
|
Text |
StreamBaseRecordReader.createValue()
|
Methods in org.apache.hadoop.streaming that return types with arguments of type Text | |
---|---|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
Methods in org.apache.hadoop.streaming with parameters of type Text | |
---|---|
boolean |
StreamXmlRecordReader.next(Text key,
Text value)
|
abstract boolean |
StreamBaseRecordReader.next(Text key,
Text value)
Read a record. |
static int |
UTF8ByteArrayUtils.readLine(LineReader lineReader,
Text out)
Deprecated. use StreamKeyValUtil.readLine(LineReader, Text) |
static int |
StreamKeyValUtil.readLine(LineReader lineReader,
Text out)
Read a utf8 encoded line from a data input stream. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
int start,
int length,
Text key,
Text val,
int splitPos)
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int) |
static void |
StreamKeyValUtil.splitKeyVal(byte[] utf,
int start,
int length,
Text key,
Text val,
int splitPos)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
int start,
int length,
Text key,
Text val,
int splitPos,
int separatorLength)
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text,
int, int) |
static void |
StreamKeyValUtil.splitKeyVal(byte[] utf,
int start,
int length,
Text key,
Text val,
int splitPos,
int separatorLength)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
Text key,
Text val,
int splitPos)
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int) |
static void |
StreamKeyValUtil.splitKeyVal(byte[] utf,
Text key,
Text val,
int splitPos)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
static void |
UTF8ByteArrayUtils.splitKeyVal(byte[] utf,
Text key,
Text val,
int splitPos,
int separatorLength)
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int, int) |
static void |
StreamKeyValUtil.splitKeyVal(byte[] utf,
Text key,
Text val,
int splitPos,
int separatorLength)
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos. |
Uses of Text in org.apache.hadoop.streaming.io |
---|
Methods in org.apache.hadoop.streaming.io that return Text | |
---|---|
Text |
TextOutputReader.getCurrentKey()
|
Text |
TextOutputReader.getCurrentValue()
|
Uses of Text in org.apache.hadoop.typedbytes |
---|
Methods in org.apache.hadoop.typedbytes that return Text | |
---|---|
Text |
TypedBytesWritableInput.readText()
|
Text |
TypedBytesWritableInput.readText(Text t)
|
Methods in org.apache.hadoop.typedbytes with parameters of type Text | |
---|---|
Text |
TypedBytesWritableInput.readText(Text t)
|
void |
TypedBytesWritableOutput.writeText(Text t)
|
Uses of Text in org.apache.hadoop.util |
---|
Methods in org.apache.hadoop.util with parameters of type Text | |
---|---|
int |
LineReader.readLine(Text str)
Read from the InputStream into the given Text. |
int |
LineReader.readLine(Text str,
int maxLineLength)
Read from the InputStream into the given Text. |
int |
LineReader.readLine(Text str,
int maxLineLength,
int maxBytesToConsume)
Read one line from the InputStream into the given Text. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |