|
||||||||||
PREV NEXT | FRAMES NO FRAMES All Classes |
org.apache.hadoop.metrics2
usage.Serialization
supports the given class.
AccessControlException
instead.RemoteException
.
AccessControlException
with the specified detail message.
RemoteException
.
AccessControlException
with the specified detail message.
DistributedCache.addArchiveToClassPath(Path, Configuration, FileSystem)
instead. The FileSystem
should be obtained within an
appropriate doAs
.
DistributedCache.addFileToClassPath(Path, Configuration, FileSystem)
instead. The FileSystem
should be obtained within an
appropriate doAs
.
Path
to the list of inputs for the map-reduce job.
Path
with a custom InputFormat
to the list of
inputs for the map-reduce job.
Path
with a custom InputFormat
and
Mapper
to the list of inputs for the map-reduce job.
Path
to the list of inputs for the map-reduce job.
Path
with a custom InputFormat
to the list of
inputs for the map-reduce job.
Path
with a custom InputFormat
and
Mapper
to the list of inputs for the map-reduce job.
UserLogEvent
for processing.
HttpServer.addSslListener(InetSocketAddress, Configuration, boolean)
AuthenticatedURL.Token
to be
used by subsequent requests.
AuthenticationToken
only
after the Kerberos SPNEGO sequence has completed successfully.
AuthenticatedURL
class enables the use of the JDK URL
class
against HTTP endpoints protected with the AuthenticationFilter
.AuthenticatedURL
.
AuthenticatedURL
.
AuthenticationException
.
AuthenticationException
.
AuthenticationException
.
AuthenticationFilter
enables protecting web application resources with different (pluggable)
authentication mechanisms.AuthenticationToken
contains information about an authenticated
HTTP client and doubles as the Principal
to be returned by
authenticated HttpServletRequest
s
The token can be serialized/deserialized to and from a string as it is sent
and received in HTTP client responses and requests as a HTTP cookie (this is
done by the AuthenticationFilter
).
InputFormat
that tries to deduce the types of the input files
automatically.WritableComparable
types supporting ordering/permutation by a representative set of bytes.BinaryComparable
keys using a configurable part of
the bytes array returned by BinaryComparable.getBytes()
.BinaryComparable
keys using a configurable part of
the bytes array returned by BinaryComparable.getBytes()
.FileSystemStore
.CompressorStream
which works
with 'block-based' based compression algorithms, as opposed to
'stream-based' compression algorithms.BlockCompressorStream
.
BlockCompressorStream
with given output-stream and
compressor.
DecompressorStream
which works
with 'block-based' based compression algorithms, as opposed to
'stream-based' compression algorithms.BlockDecompressorStream
.
BlockDecompressorStream
.
MapFile
and provides very much the same
functionality.BufferedFSInputStream
with the specified buffer size,
and saves its argument, the input stream
in
, for later use.
Decompressor
based on the popular gzip compressed file format.BytesWritable
.
StringUtils.byteDesc(long)
instead.
DistributedCache.CACHE_ARCHIVES
is not a *public* constant.
DistributedCache.CACHE_ARCHIVES_SIZES
is not a *public* constant.
DistributedCache.CACHE_ARCHIVES_TIMESTAMPS
is not a *public* constant.
DistributedCache.CACHE_FILES
is not a *public* constant.
DistributedCache.CACHE_FILES_SIZES
is not a *public* constant.
DistributedCache.CACHE_FILES_TIMESTAMPS
is not a *public* constant.
DistributedCache.CACHE_LOCALARCHIVES
is not a *public* constant.
DistributedCache.CACHE_LOCALFILES
is not a *public* constant.
DistributedCache.CACHE_SYMLINK
is not a *public* constant.
Client.call(Writable, ConnectionId)
instead
Client.call(Writable, ConnectionId)
instead
Client.call(Writable, ConnectionId)
instead
param
, to the IPC server running at
address
which is servicing the protocol
protocol,
with the ticket
credentials, rpcTimeout
as timeout
and conf
as configuration for this connection, returning the
value.
param
, to the IPC server defined by
remoteId
, returning the value.
Client.call(Writable[], InetSocketAddress[],
Class, UserGroupInformation, Configuration)
instead
Client.call(Writable[], InetSocketAddress[],
Class, UserGroupInformation, Configuration)
instead
RPC.call(Method, Object[][], InetSocketAddress[], UserGroupInformation, Configuration)
instead
Server.call(Class, Writable, long)
instead
TaskTracker
is declared as 'lost/blacklisted'
by the JobTracker.
position
.
IOException
or
null pointers.
OutputCommitter.commitJob(JobContext)
or
OutputCommitter.abortJob(JobContext, int)
instead
OutputCommitter.commitJob(JobContext)
or
OutputCommitter.abortJob(JobContext, JobStatus.State)
instead
Writable
class.
JobClient
.
InputSplit
to future operations.
RecordWriter
to future operations.
InputSplit
to future operations.
RecordWriter
to future operations.
RecordWriter
to future operations.
RecordWriter
to future operations.
IOException
IOException
.
MultiFilterRecordReader.emit(org.apache.hadoop.mapred.join.TupleWritable)
every Tuple from the
collector (the outer join of child RRs).
InputFormat
that returns CombineFileSplit
's
in InputFormat.getSplits(JobConf, int)
method.InputFormat
that returns CombineFileSplit
's in
InputFormat.getSplits(JobContext)
method.CombineFileSplit
.CombineFileSplit
.Comparator.compare(Object, Object)
.
org.apache.hadoop.metrics2
usage.CompressionOutputStream
to compress data.Configuration
.JobConf
.
JobConf
.
JobConf
.
Configuration
.Configuration
.
Socket.connect(SocketAddress, int)
.
NetUtils.connect(Socket, SocketAddress, int)
but
also takes a local address and port to bind the socket to.
org.apache.hadoop.metrics2
usage.Counter
s that logically belong together.Group
of counters, comprising of counters from a particular
counter Enum
class.Compressor
for use by this CompressionCodec
.
Compressor
for use by this CompressionCodec
.
SequenceFile.Reader
returned.
Decompressor
for use by this CompressionCodec
.
Decompressor
for use by this CompressionCodec
.
FsPermission
object.
PermissionStatus
object.
CompressionInputStream
that will read from the given
InputStream
with the given Decompressor
.
CompressionInputStream
that will read from the given
input stream.
CompressionInputStream
that will read from the given
InputStream
with the given Decompressor
.
IOException
.
CompressionOutputStream
that will write to the given
OutputStream
.
CompressionOutputStream
that will write to the given
OutputStream
with the given Compressor
.
CompressionOutputStream
that will write to the given
OutputStream
.
CompressionOutputStream
that will write to the given
OutputStream
with the given Compressor
.
recordName
.
recordName
.
TFile.Reader.createScannerByKey(byte[], byte[])
instead.
TFile.Reader.createScannerByKey(RawComparable, RawComparable)
instead.
Iterator
for a NavigableMap
.Iterable
object,
so that an Iterator
can be created
for iterating the given NavigableMap
.
Thread.setDaemon(boolean)
with true.DataInput
implementation that reads from an in-memory
buffer.DataOutput
implementation that writes to an in-memory
buffer.DBWritable
.DBWritable
.CompressionInputStream
to compress data.PolicyProvider
without any defined services.
Stringifier
interface which stringifies the objects using base64 encoding of the
serialized version of the objects.WritableComparable
implementation.
Record
implementation.
InputFormat
that delegates behaviour of paths to multiple other
InputFormats.InputFormat
that delegates behavior of paths to multiple other
InputFormats.Mapper
that delegates behaviour of paths to multiple other
mappers.Mapper
that delegates behavior of paths to multiple other
mappers.TaggedInputSplit
UserLogEvent
sent when job logs should be deleted.InputStream
.RawComparator
that uses a Deserializer
to deserialize
the objects to be compared so that the standard Comparator
can
be used to compare them.AuthenticationHandler
.
TryIterator.tryNext()
must call this method
when there are no more elements left in the iteration.
Writer
The format of the output would be
{ "properties" : [ {key1,value1,key1.isFinal,key1.resource}, {key2,value2,
key2.isFinal,key2.resource}...
o
is a ByteWritable with the same value.
o
is a DoubleWritable with the same value.
o
is a FloatWritable with the same value.
o
is a IntWritable with the same value.
o
is a LongWritable with the same value.
o
is an MD5Hash whose digest contains the
same values.
o
is a Text with the same contents.
o
is a UTF8 with the same contents.
o
is a VIntWritable with the same value.
o
is a VLongWritable with the same value.
charToEscape
in the string
with the escape char escapeChar
EventCounter
insteadEventCounter
insteadFileSink
instead.InputFormat
.InputFormat
s.OutputCommitter
that commits files specified
in job output directory i.e.OutputCommitter
that commits files specified
in job output directory i.e.OutputFormat
.OutputFormat
s that read from FileSystem
s.INode
s and Block
s.FilterFileSystem
contains
some other file system, which it uses as
its basic file system, possibly transforming
the data along the way or providing additional
functionality.FilterRecordWriter
is a convenience wrapper
class that extends the RecordWriter
.what
in the backing
buffer, starting as position start
.
UTF8ByteArrayUtils.findByte(byte[], int,
int, byte)
UTF8ByteArrayUtils.findBytes(byte[], int,
int, byte[])
UTF8ByteArrayUtils.findNthByte(byte[], int,
int, byte, int)
UTF8ByteArrayUtils.findNthByte(byte[],
byte, int)
StreamKeyValUtil.findTab(byte[], int, int)
StreamKeyValUtil.findTab(byte[])
FSInputStream
in a DataInputStream
and buffers input through a BufferedInputStream
.OutputStream
in a DataOutputStream
,
buffers output through a BufferedOutputStream
and creates a checksum
file.FsAction
.
Throwable
into a Runtime Exception.FileSystem
backed by an FTP client provided by Apache Commons Net.FileSystem.delete(Path, boolean)
org.apache.hadoop.metrics2
usage.GenericOptionsParser
is a utility to parse command line
arguments generic to the Hadoop framework.GenericOptionsParser to parse only the generic Hadoop
arguments.
GenericOptionsParser(Configuration, Options, String[]) -
Constructor for class org.apache.hadoop.util.GenericOptionsParser
Create a GenericOptionsParser
to parse given options as well
as generic Hadoop options.
GenericsUtil - Class in org.apache.hadoop.util
Contains utility methods for dealing with Java Generics.
GenericsUtil() -
Constructor for class org.apache.hadoop.util.GenericsUtil
GenericWritable - Class in org.apache.hadoop.io
A wrapper for Writable instances.
GenericWritable() -
Constructor for class org.apache.hadoop.io.GenericWritable
get(String) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property, null
if
no such property exists.
get(String, String) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property.
get(String) -
Method in class org.apache.hadoop.contrib.failmon.EventRecord
Get the value of a property of the EventRecord.
get(String) -
Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
Get the value of a property of the EventRecord.
get(URI, Configuration, String) -
Static method in class org.apache.hadoop.fs.FileSystem
get(Configuration) -
Static method in class org.apache.hadoop.fs.FileSystem
Returns the configured filesystem implementation.
get(URI, Configuration) -
Static method in class org.apache.hadoop.fs.FileSystem
Returns the FileSystem for this URI's scheme and authority.
get(long, Writable) -
Method in class org.apache.hadoop.io.ArrayFile.Reader
Return the n
th value in the file.
get() -
Method in class org.apache.hadoop.io.ArrayWritable
get(WritableComparable, Writable) -
Method in class org.apache.hadoop.io.BloomMapFile.Reader
Fast version of the
MapFile.Reader.get(WritableComparable, Writable)
method.
get() -
Method in class org.apache.hadoop.io.BooleanWritable
Returns the value of the BooleanWritable
get() -
Method in class org.apache.hadoop.io.BytesWritable
Deprecated. Use BytesWritable.getBytes()
instead.
get() -
Method in class org.apache.hadoop.io.ByteWritable
Return the value of this ByteWritable.
get() -
Method in class org.apache.hadoop.io.DoubleWritable
get(BytesWritable, BytesWritable) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy the key and value in one shot into BytesWritables.
get() -
Method in class org.apache.hadoop.io.FloatWritable
Return the value of this FloatWritable.
get() -
Method in class org.apache.hadoop.io.GenericWritable
Return the wrapped instance.
get() -
Method in class org.apache.hadoop.io.IntWritable
Return the value of this IntWritable.
get() -
Method in class org.apache.hadoop.io.LongWritable
Return the value of this LongWritable.
get(WritableComparable, Writable) -
Method in class org.apache.hadoop.io.MapFile.Reader
Return the value for the named key, or null if none exists.
get(Object) -
Method in class org.apache.hadoop.io.MapWritable
get() -
Static method in class org.apache.hadoop.io.NullWritable
Returns the single instance of this class.
get() -
Method in class org.apache.hadoop.io.ObjectWritable
Return the instance, or null if none.
get(Text) -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
get(WritableComparable) -
Method in class org.apache.hadoop.io.SetFile.Reader
Read the matching key from a set into key
.
get(Object) -
Method in class org.apache.hadoop.io.SortedMapWritable
get() -
Method in class org.apache.hadoop.io.TwoDArrayWritable
get() -
Method in class org.apache.hadoop.io.VIntWritable
Return the value of this VIntWritable.
get() -
Method in class org.apache.hadoop.io.VLongWritable
Return the value of this LongWritable.
get(Class<? extends WritableComparable>) -
Static method in class org.apache.hadoop.io.WritableComparator
Get a comparator for a WritableComparable
implementation.
get() -
Static method in class org.apache.hadoop.ipc.Server
Returns the server instance called under or null.
get(int) -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Get ith child InputSplit.
get(int) -
Method in class org.apache.hadoop.mapred.join.TupleWritable
Get ith Writable from Tuple.
get() -
Method in class org.apache.hadoop.metrics.util.MetricsIntValue
Deprecated. Get value
get() -
Method in class org.apache.hadoop.metrics.util.MetricsLongValue
Deprecated. Get value
get(String) -
Method in class org.apache.hadoop.metrics.util.MetricsRegistry
Deprecated.
get(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
Get a metric by name
get(String, Collection<MetricsTag>) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache
Get the cached record
get(DataInput) -
Static method in class org.apache.hadoop.record.BinaryRecordInput
Get a thread-local record input for the supplied DataInput.
get(DataOutput) -
Static method in class org.apache.hadoop.record.BinaryRecordOutput
Get a thread-local record output for the supplied DataOutput.
get() -
Method in class org.apache.hadoop.record.Buffer
Get the data from the Buffer.
get(DataInput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesInput
Get a thread-local typed bytes input for the supplied DataInput
.
get(DataOutput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Get a thread-local typed bytes output for the supplied DataOutput
.
get(TypedBytesInput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Get a thread-local typed bytes record input for the supplied
TypedBytesInput
.
get(DataInput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Get a thread-local typed bytes record input for the supplied
DataInput
.
get(TypedBytesOutput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Get a thread-local typed bytes record input for the supplied
TypedBytesOutput
.
get(DataOutput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Get a thread-local typed bytes record output for the supplied
DataOutput
.
get(TypedBytesInput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Get a thread-local typed bytes writable input for the supplied
TypedBytesInput
.
get(DataInput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Get a thread-local typed bytes writable input for the supplied
DataInput
.
get(TypedBytesOutput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Get a thread-local typed bytes writable input for the supplied
TypedBytesOutput
.
get(DataOutput) -
Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Get a thread-local typed bytes writable output for the supplied
DataOutput
.
get() -
Method in class org.apache.hadoop.util.Progress
Returns the overall progress of the root.
getAbsolutePath(String) -
Method in class org.apache.hadoop.streaming.PathFinder
Returns the full path name of this file if it is listed in the
path
getAcceptAnonymous() -
Method in class org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler
Returns if the handler is configured to support anonymous users.
getAccessKey() -
Method in class org.apache.hadoop.fs.s3.S3Credentials
getAccessTime() -
Method in class org.apache.hadoop.fs.FileStatus
Get the access time of the file.
getAclName() -
Method in enum org.apache.hadoop.mapreduce.JobACL
Get the name of the ACL.
getACLString() -
Method in class org.apache.hadoop.security.authorize.AccessControlList
Returns the String representation of this ACL.
getActiveTrackerNames() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the names of active task trackers in the cluster.
getAddress(Configuration) -
Static method in class org.apache.hadoop.mapred.JobTracker
getAdjustedEnd() -
Method in class org.apache.hadoop.io.compress.SplitCompressionInputStream
After calling createInputStream, the values of start or end
might change.
getAdjustedStart() -
Method in class org.apache.hadoop.io.compress.SplitCompressionInputStream
After calling createInputStream, the values of start or end
might change.
getAdminAcls(Configuration, String) -
Static method in class org.apache.hadoop.security.SecurityUtil
Get the ACL object representing the cluster administrators
The user who starts the daemon is automatically added as an admin
getAlgorithmName() -
Method in class org.apache.hadoop.fs.FileChecksum
The checksum algorithm name
getAlgorithmName() -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
The checksum algorithm name
getAliveNodesInfoJson() -
Method in class org.apache.hadoop.mapred.JobTracker
getAliveNodesInfoJson() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getAllAttempts() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.JVMInfo
getAllContexts() -
Method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns all MetricsContexts built by this factory.
getAllJobs() -
Method in class org.apache.hadoop.mapred.JobClient
Get the jobs that are submitted.
getAllJobs() -
Method in class org.apache.hadoop.mapred.JobTracker
getAllJobs() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getAllKeys() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
getAllLocalPathsToRead(String, Configuration) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
Get all of the paths that currently exist in the working directories.
getAllRecords() -
Method in interface org.apache.hadoop.metrics.MetricsContext
Deprecated. Retrieves all the records managed by this MetricsContext.
getAllRecords() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Retrieves all the records managed by this MetricsContext.
getAllStaticResolutions() -
Static method in class org.apache.hadoop.net.NetUtils
This is used to get all the resolutions that were added using
NetUtils.addStaticResolution(String, String)
.
getAllStatistics() -
Static method in class org.apache.hadoop.fs.FileSystem
Return the FileSystem classes that have Statistics
getAllTasks() -
Method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Returns all map and reduce tasks .
getAllTokens() -
Method in class org.apache.hadoop.security.Credentials
Return all the tokens in the in-memory map
getApproxChkSumLength(long) -
Static method in class org.apache.hadoop.fs.ChecksumFileSystem
getArchiveClassPaths(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get the archive entries in classpath as an array of Path.
getArchiveTimestamps(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get the timestamps of the archives.
getArchiveVisibilities(Configuration) -
Static method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
Get the booleans on whether the archives are public or not.
getAssignedJobID() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getAssignedTracker(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.JobTracker
Get tracker name for a given task id.
getAttemptsToStartSkipping(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of Task attempts AFTER which skip mode
will be kicked off.
getAttribute(String) -
Method in class org.apache.hadoop.http.HttpServer
Get the value in the webapp context.
getAttribute(String) -
Method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns the value of the named attribute, or null if there is no
attribute of that name.
getAttribute(String) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Convenience method for subclasses to access factory attributes.
getAttribute(String) -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
Deprecated.
getAttributeNames() -
Method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns the names of all the factory's attributes.
getAttributes(String[]) -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
Deprecated.
getAttributeTable(String) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Returns an attribute-value map derived from the factory attributes
by finding all factory attributes that begin with
contextName.tableName.
getAuthenticationHandler() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the authentication handler being used.
getAuthenticationMethod() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the authentication method from the subject
getAutoIncrMapperProcCount(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the flag which if set to true,
SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS
is incremented
by MapRunner after invoking the map function.
getAutoIncrReducerProcCount(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the flag which if set to true,
SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS
is incremented
by framework after invoking the reduce function.
getAvailable() -
Method in class org.apache.hadoop.fs.DF
getAvailableMapSlots() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get available map slots.
getAvailablePhysicalMemorySize() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the total size of the available physical memory present
in the system.
getAvailablePhysicalMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the total size of the available physical memory present
in the system.
getAvailableReduceSlots() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get available reduce slots.
getAvailableSlots(TaskType) -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
Get the number of currently available slots on this tasktracker for the
given type of the task.
getAvailableVirtualMemorySize() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the total size of the available virtual memory present
in the system.
getAvailableVirtualMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the total size of the available virtual memory present
in the system.
getBaseLogDir() -
Static method in class org.apache.hadoop.mapred.TaskLog
getBasePathInJarOut(String) -
Method in class org.apache.hadoop.streaming.JarBuilder
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat
getBeginColumn() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getBeginLine() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getBlacklistedNodesInfoJson() -
Method in class org.apache.hadoop.mapred.JobTracker
getBlacklistedNodesInfoJson() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getBlackListedTaskTrackerCount() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of blacklisted trackers in the cluster.
getBlacklistedTrackerNames() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the names of blacklisted task trackers in the cluster.
getBlacklistedTrackers() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of blacklisted task trackers in the cluster.
getBlockIndex(BlockLocation[], long) -
Method in class org.apache.hadoop.mapred.FileInputFormat
getBlockIndex(BlockLocation[], long) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
getBlocks() -
Method in class org.apache.hadoop.fs.s3.INode
getBlockSize() -
Method in class org.apache.hadoop.fs.FileStatus
Get the block size of the file.
getBlockSize(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. Use getFileStatus() instead
getBlockSize() -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
Returns the blocksize parameter specified at construction time.
getBloomFilter() -
Method in class org.apache.hadoop.io.BloomMapFile.Reader
Retrieve the Bloom filter used by this instance of the Reader.
getBoolean(String, boolean) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a boolean
.
getBoundAntProperty(String, String) -
Static method in class org.apache.hadoop.streaming.StreamUtil
getBoundingValsQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
getBuildVersion() -
Method in class org.apache.hadoop.mapred.JobTracker
getBuildVersion() -
Static method in class org.apache.hadoop.util.VersionInfo
Returns the full version string containing version,
revision, user and source checksum.
getByName(String) -
Static method in class org.apache.hadoop.security.SecurityUtil
Resolves a host subject to the security requirements determined by
hadoop.security.token.service.use_ip.
getByName(String) -
Method in class org.apache.hadoop.security.SecurityUtil.QualifiedHostResolver
Create an InetAddress with a fully qualified hostname of the given
hostname.
getBytes() -
Method in class org.apache.hadoop.fs.FileChecksum
The value of the checksum in bytes
getBytes() -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
The value of the checksum in bytes
getBytes() -
Method in class org.apache.hadoop.io.BinaryComparable
Return representative byte array for this instance.
getBytes() -
Method in class org.apache.hadoop.io.BytesWritable
Get the data from the BytesWritable.
getBytes() -
Method in class org.apache.hadoop.io.Text
Returns the raw bytes; however, only data up to Text.getLength()
is
valid.
getBytes() -
Method in class org.apache.hadoop.io.UTF8
Deprecated. The raw bytes.
getBytes(String) -
Static method in class org.apache.hadoop.io.UTF8
Deprecated. Convert a string to a UTF-8 encoded byte array.
getBytes() -
Method in class org.apache.hadoop.security.token.TokenIdentifier
Get the bytes for the token identifier
getBytes() -
Method in class org.apache.hadoop.util.bloom.Key
getBytesPerChecksum() -
Method in class org.apache.hadoop.util.DataChecksum
getBytesPerSum() -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
Return the bytes Per Checksum
getBytesRead() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
Get the total number of bytes read
getBytesRead() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
getBytesRead() -
Method in interface org.apache.hadoop.io.compress.Compressor
Return number of uncompressed bytes input so far.
getBytesRead() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
Return number of bytes given to this compressor since last reset.
getBytesRead() -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
Returns the total number of compressed bytes input so far, including
gzip header/trailer bytes.
getBytesRead() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
Returns the total number of uncompressed bytes input so far.
getBytesRead() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
Returns the total number of compressed bytes input so far.
getBytesWritten() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
Get the total number of bytes written
getBytesWritten() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
getBytesWritten() -
Method in interface org.apache.hadoop.io.compress.Compressor
Return number of compressed bytes output so far.
getBytesWritten() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
Return number of bytes consumed by callers of compress since last reset.
getBytesWritten() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
Returns the total number of compressed bytes output so far.
getBytesWritten() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
Returns the total number of uncompressed bytes output so far.
getCacheArchives(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get cache archives set in the Configuration.
getCacheFiles(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get cache files set in the Configuration.
getCacheSize() -
Static method in class org.apache.hadoop.fs.FileSystem
Get the number of entries in the filesystem cache
getCallQueueLen() -
Method in class org.apache.hadoop.ipc.Server
The number of rpc calls in the queue.
getCanonicalServiceName() -
Method in class org.apache.hadoop.fs.FileSystem
Get a canonical service name for this file system.
getCanonicalServiceName() -
Method in class org.apache.hadoop.fs.FilterFileSystem
getCanonicalServiceName() -
Method in class org.apache.hadoop.fs.HarFileSystem
getCanonicalUri() -
Method in class org.apache.hadoop.fs.FileSystem
Resolve the uri's hostname and add the default port if not in the uri
getCanonicalUri(URI, int) -
Static method in class org.apache.hadoop.net.NetUtils
Resolve the uri's hostname and add the default port if not in the uri
getCapacity() -
Method in class org.apache.hadoop.fs.DF
getCapacity() -
Method in class org.apache.hadoop.io.BytesWritable
Get the capacity, which is the maximum size that could handled without
resizing the backing storage.
getCapacity() -
Method in class org.apache.hadoop.record.Buffer
Get the capacity, which is the maximum count that could handled without
resizing the backing storage.
getCategory(List<List<Pentomino.ColumnName>>) -
Method in class org.apache.hadoop.examples.dancing.Pentomino
Find whether the solution has the x in the upper left quadrant, the
x-midline, the y-midline or in the center.
getChannel() -
Method in class org.apache.hadoop.net.SocketInputStream
Returns underlying channel used by inputstream.
getChannel() -
Method in class org.apache.hadoop.net.SocketOutputStream
Returns underlying channel used by this stream.
getChecksumFile(Path) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
Return the name of the checksum file associated with a file.
getChecksumFileLength(Path, long) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
Return the length of the checksum file given the size of the
actual file.
getChecksumHeaderSize() -
Static method in class org.apache.hadoop.util.DataChecksum
getChecksumLength(long, int) -
Static method in class org.apache.hadoop.fs.ChecksumFileSystem
Calculated the length of the checksum file in bytes.
getChecksumSize() -
Method in class org.apache.hadoop.util.DataChecksum
getChecksumType() -
Method in class org.apache.hadoop.util.DataChecksum
getChunkPosition(long) -
Method in class org.apache.hadoop.fs.FSInputChecker
Return position of beginning of chunk containing pos.
getClass(String, Class<?>) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a Class
.
getClass(String, Class<? extends U>, Class<U>) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a Class
implementing the interface specified by xface
.
getClass(byte) -
Method in class org.apache.hadoop.io.AbstractMapWritable
getClass(String, Configuration) -
Static method in class org.apache.hadoop.io.WritableName
Return the class for a name.
getClass(T) -
Static method in class org.apache.hadoop.util.GenericsUtil
Returns the Class object (of type Class<T>
) of the
argument of type T
.
getClass(T) -
Static method in class org.apache.hadoop.util.ReflectionUtils
Return the correctly-typed Class
of the given object.
getClassByName(String) -
Method in class org.apache.hadoop.conf.Configuration
Load a class by name.
getClassByName(String) -
Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
getClasses(String, Class<?>...) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property
as an array of Class
.
getClassLoader() -
Method in class org.apache.hadoop.conf.Configuration
Get the ClassLoader
for this job.
getClassName() -
Method in exception org.apache.hadoop.ipc.RemoteException
getClassPaths() -
Method in class org.apache.hadoop.filecache.TaskDistributedCacheManager
Retrieves class paths (as local references) to add.
getCleanupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the cleanup tasks of a job.
getCleanupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getCleanupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getClientInput() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the DataInput from which the client output is read.
getClientOutput() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the DataOutput to which the client input is written.
getClientVersion() -
Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
Get the client's preferred version
getClock() -
Method in class org.apache.hadoop.mapred.JobTracker
getClosest(WritableComparable, Writable) -
Method in class org.apache.hadoop.io.MapFile.Reader
Finds the record that is the closest match to the specified key.
getClosest(WritableComparable, Writable, boolean) -
Method in class org.apache.hadoop.io.MapFile.Reader
Finds the record that is the closest match to the specified key.
getClusterMetrics() -
Method in class org.apache.hadoop.mapred.JobTracker
getClusterNick() -
Method in class org.apache.hadoop.streaming.StreamJob
Deprecated.
getClusterStatus() -
Method in class org.apache.hadoop.mapred.JobClient
Get status information about the Map-Reduce cluster.
getClusterStatus(boolean) -
Method in class org.apache.hadoop.mapred.JobClient
Get status information about the Map-Reduce cluster.
getClusterStatus() -
Method in class org.apache.hadoop.mapred.JobTracker
Deprecated. use JobTracker.getClusterStatus(boolean)
getClusterStatus(boolean) -
Method in class org.apache.hadoop.mapred.JobTracker
getClusterStatus(boolean) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getCodec(Path) -
Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
Find the relevant compression codec for the given file based on its
filename suffix.
getCodecByClassName(String) -
Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
Find the relevant compression codec for the codec's canonical class name.
getCodecByName(String) -
Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
Find the relevant compression codec for the codec's canonical class name
or by codec alias.
getCodecClassByName(String) -
Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
Find the relevant compression codec for the codec's canonical class name
or by codec alias and returns its implemetation class.
getCodecClasses(Configuration) -
Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
Get the list of codecs listed in the configuration
getCollector(String, Reporter) -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Gets the output collector for a named output.
getCollector(String, String, Reporter) -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Gets the output collector for a multi named output.
getColumnName(int) -
Method in class org.apache.hadoop.examples.dancing.DancingLinks
Get the name of a given column as a string
getCombinerClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the user-defined combiner class used to combine map-outputs
before being sent to the reducers.
getCombinerClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the combiner class for the job.
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
getCombinerOutput() -
Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
getCombinerOutput() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
getCommandLine() -
Method in class org.apache.hadoop.util.GenericOptionsParser
Returns the commons-cli CommandLine
object
to process the parsed arguments.
getCommandName() -
Method in class org.apache.hadoop.fs.shell.Command
Return the command's name excluding the leading character -
getCommandName() -
Method in class org.apache.hadoop.fs.shell.Count
getComparator() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get an instance of the RawComparator that is constructed based on the
string comparator representation.
getComparator() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return comparator defining the ordering for RecordReaders in this
composite.
getComparatorName() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get the string representation of the comparator.
getCompressedData() -
Method in class org.apache.hadoop.io.compress.BlockDecompressorStream
getCompressedData() -
Method in class org.apache.hadoop.io.compress.DecompressorStream
getCompressionCodec() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the compression codec of data in this file.
getCompressionCodec() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
Returns the compression codec of data in this file.
getCompressionLevel(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
getCompressionStrategy(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
getCompressionType(Configuration) -
Static method in class org.apache.hadoop.io.SequenceFile
Deprecated. Use
SequenceFileOutputFormat.getOutputCompressionType(org.apache.hadoop.mapred.JobConf)
to get SequenceFile.CompressionType
for job-outputs.
getCompressMapOutput() -
Method in class org.apache.hadoop.mapred.JobConf
Are the outputs of the maps be compressed?
getCompressor(CompressionCodec, Configuration) -
Static method in class org.apache.hadoop.io.compress.CodecPool
Get a Compressor
for the given CompressionCodec
from the
pool or a new one.
getCompressor(CompressionCodec) -
Static method in class org.apache.hadoop.io.compress.CodecPool
getCompressorType() -
Method in class org.apache.hadoop.io.compress.BZip2Codec
This functionality is currently not supported.
getCompressorType() -
Method in interface org.apache.hadoop.io.compress.CompressionCodec
Get the type of Compressor
needed by this CompressionCodec
.
getCompressorType() -
Method in class org.apache.hadoop.io.compress.DefaultCodec
getCompressorType() -
Method in class org.apache.hadoop.io.compress.GzipCodec
getCompressorType() -
Method in class org.apache.hadoop.io.compress.SnappyCodec
Get the type of Compressor
needed by this CompressionCodec
.
getCompressOutput(JobConf) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Is the job output compressed?
getCompressOutput(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Is the job output compressed?
getConditions() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getConf() -
Method in interface org.apache.hadoop.conf.Configurable
Return the configuration used by this object.
getConf() -
Method in class org.apache.hadoop.conf.Configured
getConf() -
Method in class org.apache.hadoop.fs.FilterFileSystem
getConf() -
Method in class org.apache.hadoop.io.AbstractMapWritable
getConf() -
Method in class org.apache.hadoop.io.compress.DefaultCodec
getConf() -
Method in class org.apache.hadoop.io.compress.SnappyCodec
Return the configuration used by this object.
getConf() -
Method in class org.apache.hadoop.io.GenericWritable
getConf() -
Method in class org.apache.hadoop.io.ObjectWritable
getConf() -
Method in class org.apache.hadoop.mapred.JobTracker
Returns a handle to the JobTracker's Configuration
getConf() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return the configuration used by this object.
getConf() -
Method in class org.apache.hadoop.mapred.lib.InputSampler
getConf() -
Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase
getConf() -
Method in class org.apache.hadoop.mapred.Task
getConf() -
Method in class org.apache.hadoop.mapred.TaskController
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
getConf() -
Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
getConf() -
Method in class org.apache.hadoop.net.ScriptBasedMapping
getConf() -
Method in class org.apache.hadoop.net.SocksSocketFactory
getConf() -
Method in class org.apache.hadoop.streaming.DumpTypedBytes
getConf() -
Method in class org.apache.hadoop.streaming.LoadTypedBytes
getConf() -
Method in class org.apache.hadoop.streaming.StreamJob
getConf() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
getConfiguration() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the underlying configuration object.
getConfiguration() -
Method in class org.apache.hadoop.mapreduce.JobContext
Return the configuration for the job.
getConfiguration(String, FilterConfig) -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the filtered configuration (only properties starting with the specified prefix).
getConfiguration() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the Configuration.
getConfiguration() -
Method in class org.apache.hadoop.util.GenericOptionsParser
Get the modified configuration
getConfigVersion() -
Method in class org.apache.hadoop.mapred.JobTracker
getConfigVersion() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getConfigVersion() -
Method in class org.apache.hadoop.mapred.TaskTracker
getConfigVersion() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getConfResourceAsInputStream(String) -
Method in class org.apache.hadoop.conf.Configuration
Get an input stream attached to the configuration resource with the
given name
.
getConfResourceAsReader(String) -
Method in class org.apache.hadoop.conf.Configuration
Get a Reader
attached to the configuration resource with the
given name
.
getConnectAddress(Server) -
Static method in class org.apache.hadoop.net.NetUtils
Returns InetSocketAddress that a client can use to
connect to the server.
getConnection() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Returns a connection object o the DB
getConnection() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
getConnection() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
getConnection() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getContentSummary(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Return the ContentSummary
of a given Path
.
getContext(String, String) -
Method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns the named MetricsContext instance, constructing it if necessary
using the factory's current configuration attributes.
getContext(String) -
Method in class org.apache.hadoop.metrics.ContextFactory
Deprecated.
getContext(String) -
Static method in class org.apache.hadoop.metrics.MetricsUtil
Deprecated.
getContext(String, String) -
Static method in class org.apache.hadoop.metrics.MetricsUtil
Deprecated. Utility method to return the named context.
getContext() -
Method in class org.apache.hadoop.streaming.PipeMapRed
getContextFactory() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Returns the factory by which this context was created.
getContextName() -
Method in interface org.apache.hadoop.metrics.MetricsContext
Deprecated. Returns the context name.
getContextName() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Returns the context name.
getCookieDomain() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the cookie domain to use for the HTTP cookie.
getCookiePath() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the cookie path to use for the HTTP cookie.
getCount() -
Method in class org.apache.hadoop.record.Buffer
Get the current count of the buffer.
getCounter() -
Method in class org.apache.hadoop.mapred.Counters.Counter
What is the current value of this counter?
getCounter(Enum) -
Method in class org.apache.hadoop.mapred.Counters
Returns current value of the specified counter, or 0 if the counter
does not exist.
getCounter(String) -
Method in class org.apache.hadoop.mapred.Counters.Group
Returns the value of the specified counter, or 0 if the counter does
not exist.
getCounter(int, String) -
Method in class org.apache.hadoop.mapred.Counters.Group
Deprecated. use Counters.Group.getCounter(String)
instead
getCounter(Enum<?>) -
Method in interface org.apache.hadoop.mapred.Reporter
Get the Counters.Counter
of the given group with the given name.
getCounter(String, String) -
Method in interface org.apache.hadoop.mapred.Reporter
Get the Counters.Counter
of the given group with the given name.
getCounter(String, String) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
getCounter(Enum<?>) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
getCounter(Enum<?>) -
Method in class org.apache.hadoop.mapreduce.StatusReporter
getCounter(String, String) -
Method in class org.apache.hadoop.mapreduce.StatusReporter
getCounter(Enum<?>) -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
getCounter(String, String) -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
getCounterForName(String) -
Method in class org.apache.hadoop.mapred.Counters.Group
Get the counter for the given name and create it if it doesn't exist.
getCounters(Counters) -
Method in class org.apache.hadoop.mapred.JobInProgress
Returns the total job counters, by adding together the job,
the map and the reduce counters.
getCounters() -
Method in interface org.apache.hadoop.mapred.RunningJob
Gets the counters for this job.
getCounters() -
Method in class org.apache.hadoop.mapred.TaskReport
A table of counters.
getCounters() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get task's counters.
getCounters() -
Method in class org.apache.hadoop.mapreduce.Job
Gets the counters for this job.
getCountersEnabled(JobConf) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns if the counters for the named outputs are enabled or not.
getCountersEnabled(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Returns if the counters for the named outputs are enabled or not.
getCountQuery() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Returns the query for getting the total number of rows,
subclasses can override this for custom behaviour.
getCountQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Returns the query for getting the total number of rows,
subclasses can override this for custom behaviour.
getCpuFrequency() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the CPU frequency of on the system.
getCpuFrequency() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the CPU frequency of on the system.
getCpuUsage() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the CPU usage % of the machine.
getCpuUsage() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the CPU usage % of the machine.
getCredentials() -
Method in class org.apache.hadoop.mapred.JobConf
Get credentials for the job.
getCredentials() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get credentials for the job.
getCumulativeCpuTime() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the cumulative CPU time since the system is on.
getCumulativeCpuTime() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the CPU time in millisecond used by all the processes in the
process-tree since the process-tree created
getCumulativeCpuTime() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the cumulative CPU time since the system is on.
getCumulativeCpuTime() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin.ProcResourceValues
Obtain the cumulative CPU time used by a current process tree.
getCumulativeRssmem() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the cumulative resident set size (rss) memory used by all the processes
in the process-tree.
getCumulativeRssmem(int) -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the cumulative resident set size (rss) memory used by all the processes
in the process-tree that are older than the passed in age.
getCumulativeVmem() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the cumulative virtual memory used by all the processes in the
process-tree.
getCumulativeVmem(int) -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the cumulative virtual memory used by all the processes in the
process-tree that are older than the passed in age.
getCurrentIntervalValue() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
Deprecated. The Value at the current interval
getCurrentIntervalValue() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
Deprecated. The Value at the current interval
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Get the current key
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.MapContext
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.RecordReader
Get the current key
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.ReduceContext
getCurrentKey() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the current key.
getCurrentKey() -
Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the current key.
getCurrentKey() -
Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
getCurrentKey() -
Method in class org.apache.hadoop.streaming.io.TextOutputReader
getCurrentKey() -
Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
getCurrentSegmentGeneration(Directory) -
Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Get the generation (N) of the current segments_N file in the directory.
getCurrentSegmentGeneration(String[]) -
Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Get the generation (N) of the current segments_N file from a list of
files.
getCurrentSplit(JobConf) -
Static method in class org.apache.hadoop.streaming.StreamUtil
getCurrentStatus() -
Method in class org.apache.hadoop.mapred.TaskReport
The current status
getCurrentTrashDir() -
Method in class org.apache.hadoop.fs.FsShell
Returns the Trash object associated with this shell.
getCurrentUser() -
Static method in class org.apache.hadoop.security.UserGroupInformation
Return the current user, including any doAs in the current stack.
getCurrentValue(Writable) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Get the 'value' corresponding to the last read 'key'.
getCurrentValue(Object) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Get the 'value' corresponding to the last read 'key'.
getCurrentValue(V) -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Get the current value.
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.MapContext
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.RecordReader
Get the current value.
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.ReduceContext
getCurrentValue() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the current value.
getCurrentValue() -
Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the current value.
getCurrentValue() -
Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
getCurrentValue() -
Method in class org.apache.hadoop.streaming.io.TextOutputReader
getCurrentValue() -
Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
getData() -
Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
getData() -
Method in class org.apache.hadoop.io.DataInputBuffer
getData() -
Method in class org.apache.hadoop.io.DataOutputBuffer
Returns the current contents of the buffer.
getData() -
Method in class org.apache.hadoop.io.OutputBuffer
Returns the current contents of the buffer.
getDate() -
Static method in class org.apache.hadoop.util.VersionInfo
The date that Hadoop was compiled.
getDBConf() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
getDBConf() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getDBProductName() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
getDeclaredClass() -
Method in class org.apache.hadoop.io.ObjectWritable
Return the class this is meant to be.
getDecommissionedTaskTrackerCount() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of decommissioned trackers in the cluster.
getDecompressor(CompressionCodec) -
Static method in class org.apache.hadoop.io.compress.CodecPool
Get a Decompressor
for the given CompressionCodec
from the
pool or a new one.
getDecompressorType() -
Method in class org.apache.hadoop.io.compress.BZip2Codec
This functionality is currently not supported.
getDecompressorType() -
Method in interface org.apache.hadoop.io.compress.CompressionCodec
Get the type of Decompressor
needed by this CompressionCodec
.
getDecompressorType() -
Method in class org.apache.hadoop.io.compress.DefaultCodec
getDecompressorType() -
Method in class org.apache.hadoop.io.compress.GzipCodec
getDecompressorType() -
Method in class org.apache.hadoop.io.compress.SnappyCodec
Get the type of Decompressor
needed by this CompressionCodec
.
getDefault() -
Static method in class org.apache.hadoop.fs.permission.FsPermission
Get the default permission.
getDefaultAuthenticator() -
Static method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL
Returns the default Authenticator
class to use when an AuthenticatedURL
instance
is created without specifying an authenticator.
getDefaultBlockSize() -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. use FileSystem.getDefaultBlockSize(Path)
instead
getDefaultBlockSize(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.
getDefaultBlockSize() -
Method in class org.apache.hadoop.fs.FilterFileSystem
Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.
getDefaultBlockSize() -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
getDefaultExtension() -
Method in class org.apache.hadoop.io.compress.BZip2Codec
.bz2 is recognized as the default extension for compressed BZip2 files
getDefaultExtension() -
Method in interface org.apache.hadoop.io.compress.CompressionCodec
Get the default filename extension for this kind of compression.
getDefaultExtension() -
Method in class org.apache.hadoop.io.compress.DefaultCodec
getDefaultExtension() -
Method in class org.apache.hadoop.io.compress.GzipCodec
getDefaultExtension() -
Method in class org.apache.hadoop.io.compress.SnappyCodec
Get the default filename extension for this kind of compression.
getDefaultHost(String, String) -
Static method in class org.apache.hadoop.net.DNS
Returns the default (first) host name associated by the provided
nameserver with the address bound to the specified network interface
getDefaultHost(String) -
Static method in class org.apache.hadoop.net.DNS
Returns the default (first) host name associated by the default
nameserver with the address bound to the specified network interface
getDefaultIP(String) -
Static method in class org.apache.hadoop.net.DNS
Returns the first available IP address associated with the provided
network interface
getDefaultMaps() -
Method in class org.apache.hadoop.mapred.JobClient
Get status information about the max available Maps in the cluster.
getDefaultPort() -
Method in class org.apache.hadoop.fs.FileSystem
Get the default port for this file system.
getDefaultRealm() -
Static method in class org.apache.hadoop.security.authentication.util.KerberosUtil
getDefaultRealm() -
Method in class org.apache.hadoop.security.KerberosName
Get the configured default realm.
getDefaultReduces() -
Method in class org.apache.hadoop.mapred.JobClient
Get status information about the max available Reduces in the cluster.
getDefaultReplication() -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. use FileSystem.getDefaultReplication(Path)
instead
getDefaultReplication(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Get the default replication.
getDefaultReplication() -
Method in class org.apache.hadoop.fs.FilterFileSystem
Get the default replication.
getDefaultReplication() -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
getDefaultRetryPolicy(Configuration, String, boolean, String, String, Class<? extends Exception>...) -
Static method in class org.apache.hadoop.io.retry.RetryUtils
Return the default retry policy set in conf.
getDefaultSocketFactory(Configuration) -
Static method in class org.apache.hadoop.net.NetUtils
Get the default socket factory as specified by the configuration
parameter hadoop.rpc.socket.factory.default
getDefaultUri(Configuration) -
Static method in class org.apache.hadoop.fs.FileSystem
Get the default filesystem URI from a configuration.
getDefaultWorkFile(TaskAttemptContext, String) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the default path and filename for the output format.
getDelegate() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Obtain an iterator over the child RRs apropos of the value type
ultimately emitted from this join.
getDelegate() -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader
Return an iterator wrapping the JoinCollector.
getDelegate() -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
Return an iterator returning a single value from the tuple.
getDelegationToken(String) -
Method in class org.apache.hadoop.fs.FileSystem
Get a new delegation token for this file system.
getDelegationToken(Text) -
Method in class org.apache.hadoop.mapred.JobClient
getDelegationToken(Text) -
Method in class org.apache.hadoop.mapred.JobTracker
Get a new delegation token.
getDelegationToken(Text) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getDelegationToken(Credentials, String) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
getDelegationTokens(Configuration, Credentials) -
Static method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
For each archive or cache file - get the corresponding delegation token
getDelegationTokenSecretManager() -
Method in class org.apache.hadoop.mapred.JobTracker
getDependentJobs() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getDependingJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getDescription() -
Method in class org.apache.hadoop.metrics.util.MetricsBase
Deprecated.
getDeserializer(Class<Serializable>) -
Method in class org.apache.hadoop.io.serializer.JavaSerialization
getDeserializer(Class<T>) -
Method in interface org.apache.hadoop.io.serializer.Serialization
getDeserializer(Class<T>) -
Method in class org.apache.hadoop.io.serializer.SerializationFactory
getDeserializer(Class<Writable>) -
Method in class org.apache.hadoop.io.serializer.WritableSerialization
getDiagnosticInfo() -
Method in class org.apache.hadoop.mapred.TaskStatus
getDiagnostics() -
Method in class org.apache.hadoop.mapred.TaskReport
A list of error messages.
getDigest() -
Method in class org.apache.hadoop.io.MD5Hash
Returns the digest bytes.
getDigester() -
Static method in class org.apache.hadoop.io.MD5Hash
Create a thread local MD5 digester
getDirectory() -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Get the ram directory of the intermediate form.
getDirectory() -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the directory where this shard resides.
getDirectoryCount() -
Method in class org.apache.hadoop.fs.ContentSummary
getDirFailures() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the number of local directories that have failed on this tracker.
getDirPath() -
Method in class org.apache.hadoop.fs.DF
getDirPath() -
Method in class org.apache.hadoop.fs.DU
getDisplayName() -
Method in class org.apache.hadoop.mapred.Counters.Group
Returns localized name of the group.
getDisplayName() -
Method in class org.apache.hadoop.mapreduce.Counter
Get the name of the counter.
getDisplayName() -
Method in class org.apache.hadoop.mapreduce.CounterGroup
Get the display name of the group.
getDistance(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopology
Return the distance between two nodes
It is assumed that the distance from one node to its parent is 1
The distance between two nodes is calculated by summing up their distances
to their closest common ancestor.
getDistributionPolicyClass() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the distribution policy class.
getDmax(String) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
Deprecated.
getDocument() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the document.
getDocumentAnalyzerClass() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the analyzer class.
getDoubleValue(Object) -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
getDU(File) -
Static method in class org.apache.hadoop.fs.FileUtil
Takes an input dir and returns the du on that local directory.
getElementTypeID() -
Method in class org.apache.hadoop.record.meta.VectorTypeID
getEmptier() -
Method in class org.apache.hadoop.fs.Trash
Return a Runnable
that periodically empties the trash of all
users, intended to be run by the superuser.
getEnd() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
getEnd() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
getEndColumn() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getEndLine() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getEntry(MapFile.Reader[], Partitioner<K, V>, K, V) -
Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
Get an entry from output generated by this class.
getEntryComparator() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get a Comparator object to compare Entries.
getEntryCount() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get the number of key-value pair entries in TFile.
getEnum(String, T) -
Method in class org.apache.hadoop.conf.Configuration
Return value matching this enumerated type.
getErrno() -
Method in exception org.apache.hadoop.io.nativeio.NativeIOException
getError() -
Static method in class org.apache.hadoop.log.metrics.EventCounter
getEventId() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns event Id.
getEventType() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogEvent
Return the UserLogEvent.EventType
.
getExceptions() -
Method in exception org.apache.hadoop.io.MultipleIOException
getExcludedHosts() -
Method in class org.apache.hadoop.util.HostsFileReader
getExecString() -
Method in class org.apache.hadoop.fs.DF
getExecString() -
Method in class org.apache.hadoop.fs.DU
getExecString() -
Method in class org.apache.hadoop.util.Shell
return an array containing the command name & its parameters
getExecString() -
Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
getExecutable(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Get the URI of the application's executable.
getExitCode() -
Method in exception org.apache.hadoop.util.Shell.ExitCodeException
getExitCode() -
Method in class org.apache.hadoop.util.Shell
get the exit code
getExpires() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
Returns the expiration time of the token.
getExpiryDate() -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
getFactor() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
Get the number of streams to merge at once.
getFactory(Class) -
Static method in class org.apache.hadoop.io.WritableFactories
Define a factory for a class.
getFactory() -
Static method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns the singleton ContextFactory instance, constructing it if
necessary.
getFailedJobList() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getFailedJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getFailureInfo() -
Method in class org.apache.hadoop.mapred.JobStatus
gets any available info on the reason of failure of the job.
getFailureInfo() -
Method in interface org.apache.hadoop.mapred.RunningJob
Get failure info for the job.
getFallBackAuthenticator() -
Method in class org.apache.hadoop.security.authentication.client.KerberosAuthenticator
If the specified URL does not support SPNEGO authentication, a fallback Authenticator
will be used.
getFatal() -
Static method in class org.apache.hadoop.log.metrics.EventCounter
getFetchFailedMaps() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get the list of maps from which output-fetches failed.
getFieldID() -
Method in class org.apache.hadoop.record.meta.FieldTypeInfo
get the field's id (name)
getFieldNames() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getFieldSeparator() -
Method in class org.apache.hadoop.streaming.PipeMapper
getFieldSeparator() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the field separator to be used.
getFieldSeparator() -
Method in class org.apache.hadoop.streaming.PipeReducer
getFieldTypeInfos() -
Method in class org.apache.hadoop.record.meta.RecordTypeInfo
Return a collection of field type infos
getFieldTypeInfos() -
Method in class org.apache.hadoop.record.meta.StructTypeID
getFile(String, String) -
Method in class org.apache.hadoop.conf.Configuration
Get a local file name under a directory named in dirsProp with
the given path.
getFileBlockLocations(FileStatus, long, long) -
Method in class org.apache.hadoop.fs.FileSystem
Return an array containing hostnames, offset and size of
portions of the given file.
getFileBlockLocations(FileStatus, long, long) -
Method in class org.apache.hadoop.fs.FilterFileSystem
getFileBlockLocations(FileStatus, long, long) -
Method in class org.apache.hadoop.fs.HarFileSystem
Get block locations from the underlying fs and fix their
offsets and lengths.
getFileBlockLocations(FileStatus, long, long) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
Return null if the file doesn't exist; otherwise, get the
locations of the various chunks of the file file from KFS.
getFileBlockLocations(FileSystem, FileStatus) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
getFileChecksum(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Get the checksum of a file.
getFileChecksum(Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
Get the checksum of a file.
getFileChecksum(Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
getFileClassPaths(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get the file entries in classpath as an array of Path.
getFileCount() -
Method in class org.apache.hadoop.fs.ContentSummary
getFileDescriptor() -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
getFileDescriptor() -
Method in class org.apache.hadoop.fs.FSDataInputStream
getFileDescriptor() -
Method in interface org.apache.hadoop.fs.HasFileDescriptor
getFileName() -
Method in class org.apache.hadoop.metrics.file.FileContext
Deprecated. Returns the configured file name, or null.
getFiles(PathFilter) -
Method in class org.apache.hadoop.fs.InMemoryFileSystem
Deprecated.
getFileStatus(Configuration, URI) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Returns FileStatus
of a given cache file on hdfs.
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Return a file status object that represents the path.
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
Get file status.
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
return the filestatus of files in har archive.
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
FileStatus for S3 file systems.
getFileStatus(Path) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
getFilesystem() -
Method in class org.apache.hadoop.fs.DF
getFileSystem(Configuration) -
Method in class org.apache.hadoop.fs.Path
Return the FileSystem that owns this Path.
getFileSystemCounterNames(String) -
Static method in class org.apache.hadoop.mapred.Task
Counters to measure the usage of the different file systems.
getFilesystemName() -
Method in class org.apache.hadoop.mapred.JobTracker
Grab the local fs name
getFilesystemName() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getFileTimestamps(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Get the timestamps of the files.
getFileType() -
Method in class org.apache.hadoop.fs.s3.INode
getFileVisibilities(Configuration) -
Static method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
Get the booleans on whether the files are public or not.
getFinalSync(JobConf) -
Static method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
Does the user want a final sync at close?
getFinishTime() -
Method in class org.apache.hadoop.mapred.JobInProgress
getFinishTime() -
Method in class org.apache.hadoop.mapred.TaskReport
Get finish time of task.
getFinishTime() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get task finish time.
getFirst() -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
getFirstHalf(String) -
Static method in class org.apache.hadoop.net.NetworkTopology
Divide networklocation string into two parts by last separator, and get the
first part here.
getFirstKey() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get the first key in the TFile.
getFlippable() -
Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
getFloat(String, float) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a float
.
getFormatMinSplitSize() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the lower bound on split size imposed by the format.
getFormatMinSplitSize() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
getFormattedTimeWithDiff(DateFormat, long, long) -
Static method in class org.apache.hadoop.util.StringUtils
Formats time in ms and appends difference (finishTime - startTime)
as returned by formatTimeDiff().
getFS() -
Method in class org.apache.hadoop.fs.FsShell
getFs() -
Method in class org.apache.hadoop.mapred.JobClient
Get a filesystem handle.
getFSSize() -
Method in class org.apache.hadoop.fs.InMemoryFileSystem
Deprecated.
getFsStatistics(Path, Configuration) -
Static method in class org.apache.hadoop.mapred.Task
Gets a handle to the Statistics instance based on the scheme associated
with path.
getGangliaConfForMetric(String) -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
Lookup GangliaConf from cache.
getGeneration() -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the generation of the Lucene instance.
getGET_PERMISSION_COMMAND() -
Static method in class org.apache.hadoop.util.Shell
Return a Unix command to get permission information.
getGraylistedNodesInfoJson() -
Method in class org.apache.hadoop.mapred.JobTracker
getGraylistedNodesInfoJson() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getGrayListedTaskTrackerCount() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of graylisted trackers in the cluster.
getGraylistedTrackerNames() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the names of graylisted task trackers in the cluster.
getGraylistedTrackers() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of graylisted task trackers in the cluster.
getGroup() -
Method in class org.apache.hadoop.fs.FileStatus
Get the group associated with the file.
getGroup(String) -
Method in class org.apache.hadoop.mapred.Counters
Returns the named counter group, or an empty group if there is none
with the specified name.
getGroup(String) -
Method in class org.apache.hadoop.mapreduce.Counters
Returns the named counter group, or an empty group if there is none
with the specified name.
getGroupAction() -
Method in class org.apache.hadoop.fs.permission.FsPermission
Return group FsAction
.
getGroupingComparator() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the user defined RawComparator
comparator for
grouping keys of inputs to the reduce.
getGroupName() -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
Return group name
getGroupNames() -
Method in class org.apache.hadoop.mapred.Counters
Returns the names of all counter classes.
getGroupNames() -
Method in class org.apache.hadoop.mapreduce.Counters
Returns the names of all counter classes.
getGroupNames() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the group names for this user.
getGroups(String) -
Method in class org.apache.hadoop.security.Groups
Get the group memberships of a given user.
getGroups(String) -
Method in class org.apache.hadoop.security.JniBasedUnixGroupsMapping
getGroups(String) -
Method in class org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback
getGroups(String) -
Method in class org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMapping
Gets unix groups and netgroups for the user.
getGroups(String) -
Method in class org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMappingWithFallback
getGroups(String) -
Method in class org.apache.hadoop.security.ShellBasedUnixGroupsMapping
getGroups(String) -
Method in class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
getGroupsCommand() -
Static method in class org.apache.hadoop.util.Shell
a Unix command to get the current user's groups list
getGroupsForUserCommand(String) -
Static method in class org.apache.hadoop.util.Shell
a Unix command to get a given user's groups list
getHadoopClientHome() -
Method in class org.apache.hadoop.streaming.StreamJob
getHarHash(Path) -
Static method in class org.apache.hadoop.fs.HarFileSystem
the hash of the path p inside iniside
the filesystem
getHarVersion() -
Method in class org.apache.hadoop.fs.HarFileSystem
getHashType(Configuration) -
Static method in class org.apache.hadoop.util.hash.Hash
This utility method converts the name of the configured
hash type to a symbolic constant.
getHeader(boolean) -
Static method in class org.apache.hadoop.fs.ContentSummary
Return the header of the output.
getHeader() -
Method in class org.apache.hadoop.util.DataChecksum
getHealthStatus() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Returns health status of the task tracker.
getHistoryFilePath(JobID) -
Static method in class org.apache.hadoop.mapred.JobHistory
Given the job id, return the history file path from the cache
getHomeDirectory() -
Method in class org.apache.hadoop.fs.FileSystem
Return the current user's home directory in this filesystem.
getHomeDirectory() -
Method in class org.apache.hadoop.fs.FilterFileSystem
getHomeDirectory() -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
getHomeDirectory() -
Method in class org.apache.hadoop.fs.HarFileSystem
return the top level archive path.
getHomeDirectory() -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
getHost() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
getHost() -
Method in class org.apache.hadoop.streaming.Environment
getHostAddress() -
Method in class org.apache.hadoop.ipc.Server.Connection
getHostFromPrincipal(String) -
Static method in class org.apache.hadoop.security.SecurityUtil
Get the host name from the principal name of format /host@realm.
getHostInetAddress() -
Method in class org.apache.hadoop.ipc.Server.Connection
getHostname() -
Method in class org.apache.hadoop.mapred.JobTracker
getHostname() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getHostname() -
Method in class org.apache.hadoop.mapred.TaskTracker
getHostname() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getHostName() -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
getHostName() -
Method in class org.apache.hadoop.security.KerberosName
Get the second component of the name.
getHostname() -
Static method in class org.apache.hadoop.util.StringUtils
Return hostname without throwing exception.
getHosts() -
Method in class org.apache.hadoop.fs.BlockLocation
Get the list of hosts (hostname) hosting this block
getHosts(String, String) -
Static method in class org.apache.hadoop.net.DNS
Returns all the host names associated by the provided nameserver with the
address bound to the specified network interface
getHosts(String) -
Static method in class org.apache.hadoop.net.DNS
Returns all the host names associated by the default nameserver with the
address bound to the specified network interface
getHosts() -
Method in class org.apache.hadoop.util.HostsFileReader
getHttpPort() -
Method in class org.apache.hadoop.mapred.TaskTracker
getHttpPort() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getHttpPort() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the port that this task tracker is serving http requests on.
getId() -
Method in class org.apache.hadoop.fs.s3.Block
getId(Class) -
Method in class org.apache.hadoop.io.AbstractMapWritable
getID() -
Method in interface org.apache.hadoop.mapred.RunningJob
Get the job identifier.
getId() -
Method in class org.apache.hadoop.mapreduce.ID
returns the int which represents the identifier
getIdentifier(String, SecretManager<T>) -
Static method in class org.apache.hadoop.security.SaslRpcServer
getIdentifier() -
Method in class org.apache.hadoop.security.token.Token
Get the token identifier
GetImage() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getIncludeCounters() -
Method in class org.apache.hadoop.mapred.TaskStatus
getIndexInputFormatClass() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the index input format class.
getIndexInterval() -
Method in class org.apache.hadoop.io.MapFile.Writer
The number of entries that are added before an index entry is added.
getIndexMaxFieldLength() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the max field length for a Lucene instance.
getIndexMaxNumSegments() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the max number of segments for a Lucene instance.
getIndexShards() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the string representation of a number of shards.
getIndexShards(IndexUpdateConfiguration) -
Static method in class org.apache.hadoop.contrib.index.mapred.Shard
getIndexUpdaterClass() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the index updater class.
getIndexUseCompoundFile() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Check whether to use the compound file format for a Lucene instance.
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.CPUParser
Return a String with information about this class
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.HadoopLogParser
Return a String with information about this class
getInfo() -
Method in interface org.apache.hadoop.contrib.failmon.Monitored
Return a String with information about the implementing
class
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.NICParser
Return a String with information about this class
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.SensorsParser
Return a String with information about this class
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.SMARTParser
Return a String with information about this class
getInfo() -
Method in class org.apache.hadoop.contrib.failmon.SystemLogParser
Return a String with information about this class
getInfo() -
Static method in class org.apache.hadoop.log.metrics.EventCounter
getInfoPort() -
Method in class org.apache.hadoop.mapred.JobTracker
getInputBoundingQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputClass() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputConditions() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputCountQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputDataLength() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
getInputDataLength() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
getInputFieldNames() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputFileBasedOutputFileName(JobConf, String) -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the outfile name based on a given anme and the input file name.
getInputFormat() -
Method in class org.apache.hadoop.mapred.JobConf
Get the InputFormat
implementation for the map-reduce job,
defaults to TextInputFormat
if not specified explicity.
getInputFormatClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the InputFormat
class for the job.
getInputOrderBy() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputPathFilter(JobConf) -
Static method in class org.apache.hadoop.mapred.FileInputFormat
Get a PathFilter instance of the filter set for the input paths.
getInputPathFilter(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get a PathFilter instance of the filter set for the input paths.
getInputPaths(JobConf) -
Static method in class org.apache.hadoop.mapred.FileInputFormat
Get the list of input Path
s for the map-reduce job.
getInputPaths(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the list of input Path
s for the map-reduce job.
getInputQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputSeparator() -
Method in class org.apache.hadoop.streaming.PipeMapper
getInputSeparator() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the input separator to be used.
getInputSeparator() -
Method in class org.apache.hadoop.streaming.PipeReducer
getInputSplit() -
Method in interface org.apache.hadoop.mapred.Reporter
Get the InputSplit
object for a map.
getInputSplit() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
getInputSplit() -
Method in class org.apache.hadoop.mapreduce.MapContext
Get the input split for this map.
getInputStream(Socket) -
Static method in class org.apache.hadoop.net.NetUtils
Same as getInputStream(socket, socket.getSoTimeout()).
From documentation for NetUtils.getInputStream(Socket, long)
:
Returns InputStream for the socket.
getInputStream(Socket, long) -
Static method in class org.apache.hadoop.net.NetUtils
Returns InputStream for the socket.
getInputStream(InputStream) -
Method in class org.apache.hadoop.security.SaslRpcClient
Get a SASL wrapped InputStream.
getInputTableName() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getInputWriterClass() -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved InputWriter
class.
getInstance() -
Static method in class org.apache.hadoop.io.ReadaheadPool
Return the singleton instance for the current process.
getInstance() -
Static method in class org.apache.hadoop.mapred.CleanupQueue
getInstance() -
Static method in class org.apache.hadoop.mapreduce.Job
Creates a new Job
A Job will be created with a generic Configuration
.
getInstance(Configuration) -
Static method in class org.apache.hadoop.mapreduce.Job
Creates a new Job
with a given Configuration
.
getInstance(Configuration, String) -
Static method in class org.apache.hadoop.mapreduce.Job
Creates a new Job
with a given Configuration
and a given jobName.
getInstance(Configuration) -
Static method in class org.apache.hadoop.net.NetworkTopology
Get an instance of NetworkTopology based on the value of the configuration
parameter net.topology.impl.
getInstance(int) -
Static method in class org.apache.hadoop.util.hash.Hash
Get a singleton instance of hash function of a given type.
getInstance(Configuration) -
Static method in class org.apache.hadoop.util.hash.Hash
Get a singleton instance of hash function of a type
defined in the configuration.
getInstance() -
Static method in class org.apache.hadoop.util.hash.JenkinsHash
getInstance() -
Static method in class org.apache.hadoop.util.hash.MurmurHash
getInstances(String, Class<U>) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a List
of objects implementing the interface specified by xface
.
getInt(String, int) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as an int
.
getInterfaceName() -
Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
Get the interface name
getInterval(ArrayList<MonitorJob>) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
Determines the minimum interval at which the executor thread
needs to wake upto execute jobs.
getIOSortMB() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the IO sort space in MB.
getIPs(String) -
Static method in class org.apache.hadoop.net.DNS
Like DNS.getIPs(String, boolean)
, but returns all
IPs associated with the given interface and its subinterfaces.
getIPs(String, boolean) -
Static method in class org.apache.hadoop.net.DNS
Returns all the IPs associated with the provided interface, if any, in
textual form.
getIPs(String, boolean) -
Static method in class org.apache.hadoop.net.NetUtils
Return an InetAddress for each interface that matches the
given subnet specified using CIDR notation.
getIsCleanup() -
Method in class org.apache.hadoop.mapred.TaskLogAppender
Get whether task is cleanup attempt or not.
getIsJavaMapper(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java Mapper.
getIsJavaRecordReader(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java RecordReader
getIsJavaRecordWriter(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Will the reduce use a Java RecordWriter?
getIsJavaReducer(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java Reducer.
getIsMap() -
Method in class org.apache.hadoop.mapred.TaskStatus
getIssueDate() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getJar() -
Method in class org.apache.hadoop.mapred.JobConf
Get the user jar for the map-reduce job.
getJar() -
Method in class org.apache.hadoop.mapreduce.Job
Get the pathname of the job's jar.
getJar() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the pathname of the job's jar.
getJob(JobID) -
Method in class org.apache.hadoop.mapred.JobClient
Get an RunningJob
object to track an ongoing job.
getJob(String) -
Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getJob(JobID)
.
getJob(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getJob() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
getJob() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getJobACLs() -
Method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the job acls.
getJobACLs() -
Method in class org.apache.hadoop.mapred.JobStatus
Get the acls for Job.
getJobCacheSubdir(String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getJobClient() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getJobClient() -
Method in class org.apache.hadoop.mapred.TaskTracker
The connection to the JobTracker, used by the TaskRunner
for locating remote files.
getJobCompletionTime() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobCompletedEvent
Get the job completion time-stamp in milli-seconds.
getJobConf() -
Method in class org.apache.hadoop.mapred.JobContext
Get the job Configuration
getJobConf() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getJobConf() -
Method in class org.apache.hadoop.mapred.TaskAttemptContext
getJobConfPath(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Get the job conf path.
getJobCounters() -
Method in class org.apache.hadoop.mapred.JobInProgress
Returns the job-level counters.
getJobCounters(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getJobCounters(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getJobDir(String) -
Static method in class org.apache.hadoop.mapred.TaskLog
Get the user log directory for the job jobid.
getJobDir(JobID) -
Static method in class org.apache.hadoop.mapred.TaskLog
Get the user log directory for the job jobid.
getJobDistCacheArchives(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Get the job distributed cache archives path.
getJobDistCacheFiles(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Get the job distributed cache files path.
getJobDistCacheLibjars(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Get the job distributed cache libjars path.
getJobEndNotificationURI() -
Method in class org.apache.hadoop.mapred.JobConf
Get the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
getJobFile() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the configuration file for the job.
getJobFile() -
Method in interface org.apache.hadoop.mapred.RunningJob
Get the path of the submitted job configuration.
getJobFile() -
Method in class org.apache.hadoop.mapred.Task
getJobForFallowSlot(TaskType) -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
Get the JobInProgress
for which the fallow slot(s) are held.
getJobHistoryFileName(JobConf, JobID) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Recover the job history filename from the history folder.
getJobHistoryLogLocation(String) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the job history file path given the history filename
getJobHistoryLogLocationForUser(String, JobConf) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the user job history file path
getJobID() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getJobID() -
Method in class org.apache.hadoop.mapred.JobInProgress
getJobID() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the job id.
getJobId() -
Method in class org.apache.hadoop.mapred.JobProfile
Deprecated. use getJobID() instead
getJobId() -
Method in class org.apache.hadoop.mapred.JobStatus
Deprecated. use getJobID instead
getJobID() -
Method in class org.apache.hadoop.mapred.JobStatus
getJobID() -
Method in interface org.apache.hadoop.mapred.RunningJob
Deprecated. This method is deprecated and will be removed. Applications should
rather use RunningJob.getID()
.
getJobID() -
Method in class org.apache.hadoop.mapred.Task
Get the job name for this task.
getJobID() -
Method in class org.apache.hadoop.mapred.TaskAttemptID
getJobID() -
Method in class org.apache.hadoop.mapred.TaskID
getJobID() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the unique ID for the job.
getJobID() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getJobId() -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
Get the jobid
getJobID() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.DeleteJobEvent
Get the jobid.
getJobID() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobCompletedEvent
Get the job id.
getJobID() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobStartedEvent
Get the job id.
getJobID() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Returns the JobID
object that this task attempt belongs to
getJobID() -
Method in class org.apache.hadoop.mapreduce.TaskID
Returns the JobID
object that this tip belongs to
getJobIDsPattern(String, Integer) -
Static method in class org.apache.hadoop.mapred.JobID
Deprecated.
getJobJar(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Get the job jar path.
getJobJarFile(String, String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getJobLocalDir() -
Method in class org.apache.hadoop.mapred.JobConf
Get job-specific shared directory for use as scratch space
getJobName() -
Method in class org.apache.hadoop.mapred.JobConf
Get the user-specified job name.
getJobName() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getJobName() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the user-specified job name.
getJobName() -
Method in interface org.apache.hadoop.mapred.RunningJob
Get the name of the job.
getJobName() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the user-specified job name.
getJobName() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getJobPriority() -
Method in class org.apache.hadoop.mapred.JobConf
Get the JobPriority
for this job.
getJobPriority() -
Method in class org.apache.hadoop.mapred.JobStatus
Return the priority of the job
getJobProfile(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getJobProfile(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getJobRunState(int) -
Static method in class org.apache.hadoop.mapred.JobStatus
Helper method to get human-readable state of the job.
getJobs() -
Static method in class org.apache.hadoop.contrib.failmon.Environment
Scans the configuration file to determine which monitoring
utilities are available in the system.
getJobsFromQueue(String) -
Method in class org.apache.hadoop.mapred.JobClient
Gets all the jobs which were added to particular Job Queue
getJobsFromQueue(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getJobsFromQueue(String) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getJobSplitFile(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
getJobSplitMetaFile(Path) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
getJobState() -
Method in interface org.apache.hadoop.mapred.RunningJob
Returns the current state of the Job.
getJobState() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getJobStatus(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getJobStatus(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getJobStatus() -
Method in interface org.apache.hadoop.mapred.RunningJob
Returns a snapshot of the current status, JobStatus
, of the Job.
getJobSubmitHostAddress() -
Method in class org.apache.hadoop.mapred.JobInProgress
getJobSubmitHostName() -
Method in class org.apache.hadoop.mapred.JobInProgress
getJobToken(Credentials) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
getJobTokenSecret() -
Method in class org.apache.hadoop.mapred.Task
Get the job token secret
getJobTrackerHostPort() -
Method in class org.apache.hadoop.streaming.StreamJob
getJobTrackerMachine() -
Method in class org.apache.hadoop.mapred.JobTracker
getJobTrackerState() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the current state of the JobTracker
,
as JobTracker.State
getJobTrackerUrl() -
Method in class org.apache.hadoop.mapred.TaskTracker
getJobTrackerUrl() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getJtIdentifier() -
Method in class org.apache.hadoop.mapreduce.JobID
getJvmContext() -
Method in class org.apache.hadoop.mapred.Task
Gets the task JvmContext
getJvmInfo() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JvmFinishedEvent
Get the jvm info.
getJvmManagerInstance() -
Method in class org.apache.hadoop.mapred.TaskTracker
getKeepCommandFile(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
Does the user want to keep the command file for debugging? If this is
true, pipes will write a copy of the command data to a file in the
task directory named "downlink.data", which may be used to run the C++
program under the debugger.
getKeepFailedTaskFiles() -
Method in class org.apache.hadoop.mapred.JobConf
Should the temporary files for failed tasks be kept?
getKeepTaskFilesPattern() -
Method in class org.apache.hadoop.mapred.JobConf
Get the regular expression that is matched against the task names
to see if we need to keep the files.
getKey(BytesWritable) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy the key into BytesWritable.
getKey(byte[]) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy the key into user supplied buffer.
getKey(byte[], int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy the key into user supplied buffer.
getKey() -
Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
Gets the current raw key
getKey() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
Returns the stored rawKey
getKey() -
Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
Gets the current raw key.
getKey() -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
getKey() -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
getKeyClass() -
Method in class org.apache.hadoop.io.MapFile.Reader
Returns the class of keys in this file.
getKeyClass() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the class of keys in this file.
getKeyClass() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
Returns the class of keys in this file.
getKeyClass() -
Method in class org.apache.hadoop.io.WritableComparator
Returns the WritableComparable implementation class.
getKeyClass() -
Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
getKeyClass() -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
The class of key that must be passed to SequenceFileRecordReader.next(Object, Object)
..
getKeyClass() -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
getKeyClassName() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the name of the key class.
getKeyClassName() -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the key class for this SequenceFile.
getKeyClassName() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the key class for this SequenceFile.
getKeyFieldComparatorOption() -
Method in class org.apache.hadoop.mapred.JobConf
Get the KeyFieldBasedComparator
options
getKeyFieldComparatorOption(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
Get the KeyFieldBasedComparator
options
getKeyFieldPartitionerOption() -
Method in class org.apache.hadoop.mapred.JobConf
Get the KeyFieldBasedPartitioner
options
getKeyFieldPartitionerOption(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
Get the KeyFieldBasedPartitioner
options
getKeyId() -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
getKeyLength() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Get the length of the key.
getKeyList() -
Method in class org.apache.hadoop.metrics.util.MetricsRegistry
Deprecated.
getKeyNear(long) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get a sample key that is within a block whose starting offset is greater
than or equal to the specified offset.
getKeyStream() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Streaming access to the key.
getKeytab() -
Method in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
Returns the keytab used by the authentication handler.
getKeyTypeID() -
Method in class org.apache.hadoop.record.meta.MapTypeID
get the TypeID of the map's key element
getKind() -
Method in class org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier
getKind() -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
Get the token kind
getKind() -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier.Renewer
getKind() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getKind() -
Method in class org.apache.hadoop.security.token.Token
Get the token kind
getKind() -
Method in class org.apache.hadoop.security.token.Token.TrivialRenewer
getKind() -
Method in class org.apache.hadoop.security.token.TokenIdentifier
Get the token kind
getKrb5LoginModuleName() -
Static method in class org.apache.hadoop.security.authentication.util.KerberosUtil
getLargeReadOps() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
Get the number of large file system read operations such as list files
under a large directory
getLastContact() -
Method in class org.apache.hadoop.ipc.Server.Connection
getLastHalf(String) -
Static method in class org.apache.hadoop.net.NetworkTopology
Divide networklocation string into two parts by last separator, and get the
second part here.
getLastKey() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get the last key in the TFile.
getLastOutput() -
Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the last output from the client as a String.
getLastOutput() -
Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
getLastOutput() -
Method in class org.apache.hadoop.streaming.io.TextOutputReader
getLastOutput() -
Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
getLastSeen() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
getLaunchTime() -
Method in class org.apache.hadoop.mapred.JobInProgress
getLeaves(String) -
Method in class org.apache.hadoop.net.NetworkTopology
return leaves in scope
getLen() -
Method in class org.apache.hadoop.fs.FileStatus
getLength() -
Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
getLength() -
Method in class org.apache.hadoop.fs.BlockLocation
Get the length of the block
getLength() -
Method in class org.apache.hadoop.fs.ContentSummary
getLength() -
Method in class org.apache.hadoop.fs.FileChecksum
The length of the checksum in bytes
getLength(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. Use getFileStatus() instead
getLength(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
Deprecated.
getLength() -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
The length of the checksum in bytes
getLength() -
Method in class org.apache.hadoop.fs.s3.Block
getLength() -
Method in class org.apache.hadoop.io.BinaryComparable
Return n st bytes 0..n-1 from {#getBytes()} are valid.
getLength() -
Method in class org.apache.hadoop.io.BytesWritable
Get the current size of the buffer.
getLength() -
Method in class org.apache.hadoop.io.DataInputBuffer
Returns the length of the input.
getLength() -
Method in class org.apache.hadoop.io.DataOutputBuffer
Returns the length of the valid data currently in the buffer.
getLength() -
Method in class org.apache.hadoop.io.InputBuffer
Returns the length of the input.
getLength() -
Method in class org.apache.hadoop.io.OutputBuffer
Returns the length of the valid data currently in the buffer.
getLength() -
Method in interface org.apache.hadoop.io.ReadaheadPool.ReadaheadRequest
getLength() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
Returns the current length of the output file.
getLength() -
Method in class org.apache.hadoop.io.Text
Returns the number of bytes in the byte array
getLength() -
Method in class org.apache.hadoop.io.UTF8
Deprecated. The number of bytes in the encoded string.
getLength() -
Method in class org.apache.hadoop.mapred.FileSplit
The number of bytes in the file to process.
getLength() -
Method in interface org.apache.hadoop.mapred.InputSplit
Get the total number of bytes in the data of the InputSplit
.
getLength() -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Return the aggregate length of all child InputSplits currently added.
getLength(int) -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Get the length of ith child InputSplit.
getLength() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
getLength(int) -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns the length of the ith Path
getLength() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
getLength() -
Method in class org.apache.hadoop.mapreduce.InputSplit
Get the size of the split, so that the input splits can be sorted by size.
getLength() -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
getLength() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
getLength() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
getLength(int) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the length of the ith Path
getLength() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The number of bytes in the file to process.
getLengths() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns an array containing the lengths of the files in the split
getLengths() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns an array containing the lengths of the files in the split
getLevel() -
Method in interface org.apache.hadoop.net.Node
Return this node's level in the tree.
getLevel() -
Method in class org.apache.hadoop.net.NodeBase
Return this node's level in the tree.
getLibJars(Configuration) -
Static method in class org.apache.hadoop.util.GenericOptionsParser
If libjars are set in the conf, parse the libjars.
getLinkCount(File) -
Static method in class org.apache.hadoop.fs.HardLink
Retrieves the number of links to the specified file.
getLinkMultArgLength(File, String[], File) -
Static method in class org.apache.hadoop.fs.HardLink
Calculate the nominal length of all contributors to the total
commandstring length, including fixed overhead of the OS-dependent
command.
getListenerAddress() -
Method in class org.apache.hadoop.ipc.Server
Return the socket (ip+port) on which the RPC server is listening to.
getLoadNativeLibraries(Configuration) -
Method in class org.apache.hadoop.util.NativeCodeLoader
Return if native hadoop libraries, if present, can be used for this job.
getLocal(Configuration) -
Static method in class org.apache.hadoop.fs.FileSystem
Get the local file syste
getLocalAnalysisClass() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the local analysis class.
getLocalCacheArchives(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Return the path array of the localized caches.
getLocalCacheFiles(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Return the path array of the localized files.
getLocalDirs() -
Method in class org.apache.hadoop.mapred.JobConf
getLocalDirs() -
Method in class org.apache.hadoop.mapred.TaskController
getLocalHostName() -
Static method in class org.apache.hadoop.security.SecurityUtil
Get the fqdn for the current host.
getLocalInetAddress(String) -
Static method in class org.apache.hadoop.net.NetUtils
Checks if host
is a local host name and return InetAddress
corresponding to that address.
getLocalJobDir(String, String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getLocalJobFilePath(JobID) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the path of the locally stored job file
getLocalJobFilePath(JobID) -
Static method in class org.apache.hadoop.mapred.JobTracker
Get the localized job file path on the job trackers local file system
getLocalMaxRunningMaps(JobContext) -
Static method in class org.apache.hadoop.mapred.LocalJobRunner
getLocalPath(String, String) -
Method in class org.apache.hadoop.conf.Configuration
Get a local file under a directory named by dirsProp with
the given path.
getLocalPath(String) -
Method in class org.apache.hadoop.mapred.JobConf
Constructs a local file name.
getLocalPathForWrite(String, Configuration) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
Get a path from the local FS.
getLocalPathForWrite(String, long, Configuration) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
Get a path from the local FS.
getLocalPathForWrite(String, long, Configuration, boolean) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
Get a path from the local FS.
getLocalPathToRead(String, Configuration) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
Get a path from the local FS for reading.
getLocalTaskDir(String, String, String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getLocalTaskDir(String, String, String, boolean) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getLocation(int) -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
getLocations from ith InputSplit.
getLocations() -
Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
getLocations() -
Method in class org.apache.hadoop.mapred.FileSplit
getLocations() -
Method in interface org.apache.hadoop.mapred.InputSplit
Get the list of hostnames where the input split is located.
getLocations() -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Collect a set of hosts from all child InputSplits.
getLocations() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns all the Paths where this input-split resides
getLocations() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
Get the list of hostnames where the input split is located.
getLocations() -
Method in class org.apache.hadoop.mapred.MultiFileSplit
Deprecated.
getLocations() -
Method in class org.apache.hadoop.mapreduce.InputSplit
Get the list of nodes by name where the data for the split would be local.
getLocations() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
Get the list of nodes by name where the data for the split would be local.
getLocations() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns all the Paths where this input-split resides
getLocations() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
getLocations() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
getLocations() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
getLoginUser() -
Static method in class org.apache.hadoop.security.UserGroupInformation
Get the currently logged in user.
getLogLocation() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.JVMInfo
getLong(String, long) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property as a long
.
getLongValue(Object) -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
getLowerClause() -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
getMajor() -
Method in class org.apache.hadoop.io.file.tfile.Utils.Version
Get the major version.
getMap() -
Method in class org.apache.hadoop.contrib.failmon.EventRecord
Return the HashMap of properties of the EventRecord.
getMapCompletionEvents(JobID, int, int, TaskAttemptID, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
getMapCompletionEvents(JobID, int, int, TaskAttemptID, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
Called by a reduce task to get the map output locations for finished maps.
getMapCounters(Counters) -
Method in class org.apache.hadoop.mapred.JobInProgress
Returns map phase counters by summing over all map tasks in progress.
getMapDebugScript() -
Method in class org.apache.hadoop.mapred.JobConf
Get the map task's debug script.
getMapOutputCompressorClass(Class<? extends CompressionCodec>) -
Method in class org.apache.hadoop.mapred.JobConf
Get the CompressionCodec
for compressing the map outputs.
getMapOutputKeyClass() -
Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
Get the map output key class.
getMapOutputKeyClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the key class for the map output data.
getMapOutputKeyClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the key class for the map output data.
getMapOutputValueClass() -
Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
Get the map output value class.
getMapOutputValueClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the value class for the map output data.
getMapOutputValueClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the value class for the map output data.
getMapper() -
Method in class org.apache.hadoop.mapred.MapRunner
getMapperClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the Mapper
class for the job.
getMapperClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the Mapper
class for the job.
getMapperClass(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
Get the application's mapper class.
getMapperMaxSkipRecords(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of acceptable skip records surrounding the bad record PER
bad record in mapper.
getMapredJobID() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
Deprecated. use Job.getAssignedJobID()
instead
getMapredJobID() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getMapredTempDir() -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the Map/Reduce temp directory.
getMapRunnerClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the MapRunnable
class for the job.
getMapSlotCapacity() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of map slots in the cluster.
getMapSpeculativeExecution() -
Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job for map tasks?
Defaults to true
.
getMapTaskCompletionEvents() -
Method in class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
getMapTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the map tasks of a job.
getMapTaskReports(String) -
Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getMapTaskReports(JobID)
getMapTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getMapTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getMapTasks() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of currently running map tasks in the cluster.
getMasterKeyId() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getMaxAllowedCmdArgLength() -
Static method in class org.apache.hadoop.fs.HardLink
Return this private value for use by unit tests.
getMaxDate() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getMaxDepth(int) -
Static method in class org.apache.hadoop.util.QuickSort
Deepest recursion before giving up and doing a heapsort.
getMaxMapAttempts() -
Method in class org.apache.hadoop.mapred.JobConf
Get the configured number of maximum attempts that will be made to run a
map task, as specified by the mapred.map.max.attempts
property.
getMaxMapSlots() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the maximum map slots for this node.
getMaxMapTaskFailuresPercent() -
Method in class org.apache.hadoop.mapred.JobConf
Get the maximum percentage of map tasks that can fail without
the job being aborted.
getMaxMapTasks() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the maximum capacity for running map tasks in the cluster.
getMaxMemory() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Deprecated.
getMaxPhysicalMemoryForTask() -
Method in class org.apache.hadoop.mapred.JobConf
Deprecated. this variable is deprecated and nolonger in use.
getMaxReduceAttempts() -
Method in class org.apache.hadoop.mapred.JobConf
Get the configured number of maximum attempts that will be made to run a
reduce task, as specified by the mapred.reduce.max.attempts
property.
getMaxReduceSlots() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the maximum reduce slots for this node.
getMaxReduceTaskFailuresPercent() -
Method in class org.apache.hadoop.mapred.JobConf
Get the maximum percentage of reduce tasks that can fail without
the job being aborted.
getMaxReduceTasks() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the maximum capacity for running reduce tasks in the cluster.
getMaxSplitSize(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the maximum split size.
getMaxTaskFailuresPerTracker() -
Method in class org.apache.hadoop.mapred.JobConf
Expert: Get the maximum no.
getMaxTime() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
Deprecated. The max time for a single operation since the last reset
MetricsTimeVaryingRate.resetMinMax()
getMaxVirtualMemoryForTask() -
Method in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.getMemoryForMapTask()
and
JobConf.getMemoryForReduceTask()
getMBeanInfo() -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
Deprecated.
getMD5Hash(String) -
Static method in class org.apache.hadoop.contrib.failmon.Anonymizer
Create the MD5 digest of an input text.
getMechanismName() -
Method in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
Return the SASL mechanism name
getMemory() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
Get the total amount of buffer memory, in bytes.
getMemoryCalculatorPlugin(Class<? extends MemoryCalculatorPlugin>, Configuration) -
Static method in class org.apache.hadoop.util.MemoryCalculatorPlugin
Deprecated. Get the MemoryCalculatorPlugin from the class name and configure it.
getMemoryForMapTask() -
Method in class org.apache.hadoop.mapred.JobConf
Get memory required to run a map task of the job, in MB.
getMemoryForReduceTask() -
Method in class org.apache.hadoop.mapred.JobConf
Get memory required to run a reduce task of the job, in MB.
getMessage() -
Method in exception org.apache.hadoop.mapred.InvalidInputException
Get a summary message of the problems found.
getMessage() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getMessage() -
Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
Get a summary message of the problems found.
getMessage() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
getMessage() -
Method in exception org.apache.hadoop.record.compiler.generated.ParseException
This method has the standard behavior when this object has been
created using the standard constructors.
getMessage() -
Method in error org.apache.hadoop.record.compiler.generated.TokenMgrError
You can also modify the body of this method to customize your error messages.
getMetaBlock(String) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Stream access to a meta block.``
getMetadata() -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
getMetadata() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the metadata object of the file
getMetric(String) -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns the metric object which can be a Float, Integer, Short or Byte.
getMetric(String) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache.Record
Get the metric value
getMetricInstance(String) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache.Record
Get the metric value
getMetricNames() -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns the set of metric names.
getMetrics(MetricsBuilder, boolean) -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
getMetrics(MetricsBuilder, boolean) -
Method in class org.apache.hadoop.mapred.TaskTrackerMetricsSource
getMetrics(MetricsBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.AbstractMetricsSource
getMetrics(MetricsBuilder, boolean) -
Method in interface org.apache.hadoop.metrics2.MetricsSource
Get metrics from the source
getMetrics(MetricsBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.source.JvmMetricsSource
getMetricsCopy() -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns a copy of this record's metrics.
getMetricsList() -
Method in class org.apache.hadoop.metrics.util.MetricsRegistry
Deprecated.
getMinor() -
Method in class org.apache.hadoop.io.file.tfile.Utils.Version
Get the minor version.
getMinSplitSize(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the minimum split size
getMinTime() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
Deprecated. The min time for a single operation since the last reset
MetricsTimeVaryingRate.resetMinMax()
getMode() -
Method in class org.apache.hadoop.io.nativeio.NativeIO.Stat
getModificationTime() -
Method in class org.apache.hadoop.fs.FileStatus
Get the modification time of the file.
getMount() -
Method in class org.apache.hadoop.fs.DF
getMultipleLinearRandomRetry(Configuration, String, boolean, String, String) -
Static method in class org.apache.hadoop.io.retry.RetryUtils
Return the MultipleLinearRandomRetry policy specified in the conf,
or null if the feature is disabled.
getName() -
Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
getName() -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. call #getUri() instead.
getName() -
Method in class org.apache.hadoop.fs.FilterFileSystem
Deprecated. call #getUri() instead.
getName() -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
Deprecated.
getName() -
Method in class org.apache.hadoop.fs.Path
Returns the final component of this path.
getName() -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
getName(Class) -
Static method in class org.apache.hadoop.io.WritableName
Return the name for a class.
getName() -
Method in class org.apache.hadoop.mapred.Counters.Group
Returns raw name of the group.
getName() -
Method in class org.apache.hadoop.mapreduce.Counter
getName() -
Method in class org.apache.hadoop.mapreduce.CounterGroup
Get the internal name of the group
getName() -
Method in class org.apache.hadoop.metrics.util.MetricsBase
Deprecated.
getName() -
Method in interface org.apache.hadoop.net.Node
Return this node's name
getName() -
Method in class org.apache.hadoop.net.NodeBase
Return this node's name
getName() -
Method in class org.apache.hadoop.record.meta.RecordTypeInfo
return the name of the record
getName() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
Returns the principal name (this method name comes from the JDK Principal
interface).
getNamed(String, Configuration) -
Static method in class org.apache.hadoop.fs.FileSystem
Deprecated. call #get(URI,Configuration) instead.
getNamedOutputFormatClass(JobConf, String) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the named output OutputFormat.
getNamedOutputKeyClass(JobConf, String) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the key class for a named output.
getNamedOutputs() -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns iterator with the defined name outputs.
getNamedOutputsList(JobConf) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns list of channel names.
getNamedOutputValueClass(JobConf, String) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the value class for a named output.
getNames() -
Method in class org.apache.hadoop.fs.BlockLocation
Get the list of names (hostname:port) hosting this block
getNestedStructTypeInfo(String) -
Method in class org.apache.hadoop.record.meta.RecordTypeInfo
Return the type info of a nested record.
getNetgroupNames() -
Method in class org.apache.hadoop.security.NetgroupCache
getNetgroups(String, List<String>) -
Method in class org.apache.hadoop.security.NetgroupCache
getNetgroups(String, List<String>) -
Method in class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
getNetworkLocation() -
Method in interface org.apache.hadoop.net.Node
Return the string representation of this node's network location
getNetworkLocation() -
Method in class org.apache.hadoop.net.NodeBase
Return this node's network location
getNewJobId() -
Method in class org.apache.hadoop.mapred.JobTracker
Allocates a new JobId string.
getNewJobId() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getNext() -
Method in class org.apache.hadoop.contrib.failmon.LogParser
Continue parsing the log file until a valid log entry is identified.
getNextHeartbeatInterval() -
Method in class org.apache.hadoop.mapred.JobTracker
Calculates next heartbeat interval using cluster size.
getNextRecordRange() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get the next record range which is going to be processed by Task.
getNextToken() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
getNextToken() -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
getNode(String) -
Method in class org.apache.hadoop.mapred.JobTracker
Return the Node in the network topology that corresponds to the hostname
getNode() -
Method in class org.apache.hadoop.mapred.join.Parser.NodeToken
getNode() -
Method in class org.apache.hadoop.mapred.join.Parser.Token
getNode(String) -
Method in class org.apache.hadoop.net.NetworkTopology
Given a string representation of a node, return its reference
getNodeForNetworkLocation(Node) -
Method in class org.apache.hadoop.net.NetworkTopology
Return a reference to the node given its string representation.
getNodeForNetworkLocation(Node) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
getNodeGroup(String) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
Given a string representation of a node group for a specific network
location
getNodesAtMaxLevel() -
Method in class org.apache.hadoop.mapred.JobTracker
Returns a collection of nodes at the max level
getNullContext(String) -
Static method in class org.apache.hadoop.metrics.ContextFactory
Deprecated. Returns a "null" context - one which does nothing.
getNum() -
Method in class org.apache.hadoop.mapred.join.Parser.NumToken
getNum() -
Method in class org.apache.hadoop.mapred.join.Parser.Token
getNumber() -
Method in class org.apache.hadoop.metrics.spi.MetricValue
Deprecated.
getNumberColumns() -
Method in class org.apache.hadoop.examples.dancing.DancingLinks
Get the number of columns.
getNumberOfThreads(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
The number of threads in the thread pool that will run the map function.
getNumberOfUniqueHosts() -
Method in class org.apache.hadoop.mapred.JobTracker
getNumBytesInSum() -
Method in class org.apache.hadoop.util.DataChecksum
getNumExcludedNodes() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of excluded hosts in the cluster.
getNumFiles(PathFilter) -
Method in class org.apache.hadoop.fs.InMemoryFileSystem
Deprecated.
getNumLinesPerSplit(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
Get the number of lines per split
getNumMapTasks() -
Method in class org.apache.hadoop.mapred.JobConf
Get configured the number of reduce tasks for this job.
getNumOfKeyFields() -
Method in class org.apache.hadoop.streaming.PipeMapper
getNumOfKeyFields() -
Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the number of key fields.
getNumOfKeyFields() -
Method in class org.apache.hadoop.streaming.PipeReducer
getNumOfLeaves() -
Method in class org.apache.hadoop.net.NetworkTopology
Return the total number of nodes
getNumOfRacks() -
Method in class org.apache.hadoop.net.NetworkTopology
Return the total number of racks
getNumOpenConnections() -
Method in class org.apache.hadoop.ipc.Server
The number of open RPC conections
getNumPaths() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns the number of Paths in the split
getNumPaths() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the number of Paths in the split
getNumProcessors() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the total number of processors present on the system.
getNumProcessors() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the total number of processors present on the system.
getNumReduceTasks() -
Method in class org.apache.hadoop.mapred.JobConf
Get configured the number of reduce tasks for this job.
getNumReduceTasks() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get configured the number of reduce tasks for this job.
getNumRequests() -
Method in class org.apache.hadoop.mapred.ShuffleExceptionTracker
Gets the number of requests we are tracking
getNumReservedTaskTrackersForMaps() -
Method in class org.apache.hadoop.mapred.JobInProgress
getNumReservedTaskTrackersForReduces() -
Method in class org.apache.hadoop.mapred.JobInProgress
getNumResolvedTaskTrackers() -
Method in class org.apache.hadoop.mapred.JobTracker
getNumSchedulingOpportunities() -
Method in class org.apache.hadoop.mapred.JobInProgress
getNumSlots() -
Method in class org.apache.hadoop.mapred.TaskStatus
getNumSlotsPerTask(TaskType) -
Method in class org.apache.hadoop.mapred.JobInProgress
getNumSlotsRequired() -
Method in class org.apache.hadoop.mapred.Task
getNumTaskCacheLevels() -
Method in class org.apache.hadoop.mapred.JobTracker
getNumTasksToExecutePerJvm() -
Method in class org.apache.hadoop.mapred.JobConf
Get the number of tasks that a spawned JVM should execute
getOccupiedMapSlots() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get number of occupied map slots in the cluster.
getOccupiedReduceSlots() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of occupied reduce slots in the cluster.
getOffset() -
Method in class org.apache.hadoop.fs.BlockLocation
Get the start offset of file associated with this block
getOffset() -
Method in interface org.apache.hadoop.io.ReadaheadPool.ReadaheadRequest
getOffset(int) -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns the start offset of the ith Path
getOffset(int) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the start offset of the ith Path
getOidInstance(String) -
Static method in class org.apache.hadoop.security.authentication.util.KerberosUtil
getOp() -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Get the type of the operation.
getOp() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the type of operation.
getOperations() -
Method in class org.apache.hadoop.mapred.QueueAclsInfo
getOpt(String) -
Method in class org.apache.hadoop.fs.shell.CommandFormat
Return if the option is set or not
getOtherAction() -
Method in class org.apache.hadoop.fs.permission.FsPermission
Return other FsAction
.
getOutput() -
Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
Get the output of the shell command.
getOutputCommitter() -
Method in class org.apache.hadoop.mapred.JobConf
Get the OutputCommitter
implementation for the map-reduce job,
defaults to FileOutputCommitter
if not specified explicitly.
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
getOutputCommitter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.OutputFormat
Get the output committer for this output format.
getOutputCommitter() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
getOutputCompressionType(JobConf) -
Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
Get the SequenceFile.CompressionType
for the output SequenceFile
.
getOutputCompressionType(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
Get the SequenceFile.CompressionType
for the output SequenceFile
.
getOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the CompressionCodec
for compressing the job outputs.
getOutputCompressorClass(JobContext, Class<? extends CompressionCodec>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the CompressionCodec
for compressing the job outputs.
getOutputFieldCount() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getOutputFieldNames() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getOutputFormat() -
Method in class org.apache.hadoop.mapred.JobConf
Get the OutputFormat
implementation for the map-reduce job,
defaults to TextOutputFormat
if not specified explicity.
getOutputFormatClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the OutputFormat
class for the job.
getOutputKeyClass() -
Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
Get the reduce output key class.
getOutputKeyClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the key class for the job output data.
getOutputKeyClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the key class for the job output data.
getOutputKeyClass() -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved output key class.
getOutputKeyComparator() -
Method in class org.apache.hadoop.mapred.JobConf
Get the RawComparator
comparator used to compare keys.
getOutputName(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the base output name for the output file.
getOutputPath(JobConf) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the Path
to the output directory for the map-reduce job.
getOutputPath(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the Path
to the output directory for the map-reduce job.
getOutputReaderClass() -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved OutputReader
class.
getOutputSize() -
Method in class org.apache.hadoop.mapred.TaskStatus
Returns the number of bytes of output from this map.
getOutputStream(Socket) -
Static method in class org.apache.hadoop.net.NetUtils
Same as getOutputStream(socket, 0).
getOutputStream(Socket, long) -
Static method in class org.apache.hadoop.net.NetUtils
Returns OutputStream for the socket.
getOutputStream(OutputStream) -
Method in class org.apache.hadoop.security.SaslRpcClient
Get a SASL wrapped OutputStream.
getOutputTableName() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
getOutputValueClass() -
Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
Get the reduce output value class.
getOutputValueClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the value class for job outputs.
getOutputValueClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the value class for job outputs.
getOutputValueClass() -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved output value class.
getOutputValueGroupingComparator() -
Method in class org.apache.hadoop.mapred.JobConf
Get the user defined WritableComparable
comparator for
grouping keys of inputs to the reduce.
getOwner() -
Method in class org.apache.hadoop.fs.FileStatus
Get the owner of the file.
getOwner(FileDescriptor) -
Static method in class org.apache.hadoop.io.nativeio.NativeIO
getOwner() -
Method in class org.apache.hadoop.io.nativeio.NativeIO.Stat
getParameter(String) -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
Unquote the name and quote the value.
getParameter(ServletRequest, String) -
Static method in class org.apache.hadoop.util.ServletUtil
Get a parameter from a ServletRequest.
getParameterMap() -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
getParameterNames() -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
Return the set of parameter names, quoting each name.
getParameterValues(String) -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
getParent() -
Method in class org.apache.hadoop.fs.Path
Returns the parent of a path or null if at root.
getParent() -
Method in interface org.apache.hadoop.net.Node
Return this node's parent
getParent() -
Method in class org.apache.hadoop.net.NodeBase
Return this node's parent
getParentNode(Node, int) -
Static method in class org.apache.hadoop.mapred.JobTracker
getPartition(Shard, IntermediateForm, int) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
getPartition(SecondarySort.IntPair, IntWritable, int) -
Method in class org.apache.hadoop.examples.SecondarySort.FirstPartitioner
getPartition(IntWritable, NullWritable, int) -
Method in class org.apache.hadoop.examples.SleepJob
getPartition(K2, V2, int) -
Method in class org.apache.hadoop.mapred.lib.HashPartitioner
Use Object.hashCode()
to partition.
getPartition(K2, V2, int) -
Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
getPartition(int, int) -
Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
getPartition(K, V, int) -
Method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
getPartition(K2, V2, int) -
Method in interface org.apache.hadoop.mapred.Partitioner
Get the paritition number for a given key (hence record) given the total
number of partitions i.e.
getPartition() -
Method in class org.apache.hadoop.mapred.Task
Get the index of this task within the job.
getPartition(BinaryComparable, V, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
Use (the specified slice of the array returned by)
BinaryComparable.getBytes()
to partition.
getPartition(K, V, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner
Use Object.hashCode()
to partition.
getPartition(K2, V2, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
getPartition(int, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
getPartition(K, V, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
getPartition(KEY, VALUE, int) -
Method in class org.apache.hadoop.mapreduce.Partitioner
Get the partition number for a given key (hence record) given the total
number of partitions i.e.
getPartitionerClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the Partitioner
used to partition Mapper
-outputs
to be sent to the Reducer
s.
getPartitionerClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the Partitioner
class for the job.
getPartitionFile(JobConf) -
Static method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
Get the path to the SequenceFile storing the sorted partition keyset.
getPartitionFile(Configuration) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
Get the path to the SequenceFile storing the sorted partition keyset.
getPassword() -
Method in class org.apache.hadoop.security.token.Token
Get the token password/secret
getPath() -
Method in class org.apache.hadoop.fs.FileStatus
getPath() -
Method in class org.apache.hadoop.mapred.FileSplit
The file containing this split's data.
getPath(int) -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns the ith Path
getPath(int) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the ith Path
getPath() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The file containing this split's data.
getPath(Node) -
Static method in class org.apache.hadoop.net.NodeBase
Return this node's path
getPathForCustomFile(JobConf, String) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to generate a Path
for a file that is unique for
the task within the job output directory.
getPathForWorkFile(TaskInputOutputContext<?, ?, ?, ?>, String, String) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Helper function to generate a Path
for a file that is unique for
the task within the job output directory.
getPaths() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns all the Paths in the split
getPaths() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns all the Paths in the split
getPercentExceptions() -
Method in class org.apache.hadoop.mapred.ShuffleExceptionTracker
Gets the percent of the requests that had exceptions occur.
getPercentUsed() -
Method in class org.apache.hadoop.fs.DF
getPercentUsed() -
Method in class org.apache.hadoop.fs.InMemoryFileSystem
Deprecated.
getPeriod() -
Method in interface org.apache.hadoop.metrics.MetricsContext
Deprecated. Returns the timer period.
getPeriod() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
Deprecated. Returns the timer period.
getPermission() -
Method in class org.apache.hadoop.fs.FileStatus
Get FsPermission associated with the file.
getPermission() -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
Return permission
getPhase() -
Method in class org.apache.hadoop.mapred.Task
Return current phase of the task.
getPhase() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get current phase of this task.
getPhysicalMemorySize() -
Method in class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin
Deprecated. Obtain the total size of the physical memory present in the system.
getPhysicalMemorySize() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the total size of the physical memory present in the system.
getPhysicalMemorySize() -
Method in class org.apache.hadoop.util.MemoryCalculatorPlugin
Deprecated. Obtain the total size of the physical memory present in the system.
getPhysicalMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the total size of the physical memory present in the system.
getPhysicalMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin.ProcResourceValues
Obtain the physical memory size used by current process tree.
getPlatformName() -
Static method in class org.apache.hadoop.util.PlatformName
Get the complete platform as per the java-vm.
getPort() -
Method in class org.apache.hadoop.http.HttpServer
Get the port that the server is on
getPos() -
Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
getPos() -
Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
getPos() -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
getPos() -
Method in exception org.apache.hadoop.fs.ChecksumException
getPos() -
Method in class org.apache.hadoop.fs.FSDataInputStream
getPos() -
Method in class org.apache.hadoop.fs.FSDataOutputStream
getPos() -
Method in class org.apache.hadoop.fs.FSInputChecker
getPos() -
Method in class org.apache.hadoop.fs.FSInputStream
Return the current offset from the start of the file
getPos() -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
getPos() -
Method in interface org.apache.hadoop.fs.Seekable
Return the current offset from the start of the file
getPos() -
Method in class org.apache.hadoop.io.compress.CompressionInputStream
This method returns the current position in the stream.
getPos() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Unsupported (returns zero in all cases).
getPos() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request position from proxied RR.
getPos() -
Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
getPos() -
Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
return the amount of data processed
getPos() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
Returns the current position in the input.
getPos() -
Method in class org.apache.hadoop.mapred.LineRecordReader
getPos() -
Method in interface org.apache.hadoop.mapred.RecordReader
Returns the current position in the input.
getPos() -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
getPos() -
Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
getPos() -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
getPos() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Deprecated.
getPos() -
Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
Returns the current position in the input.
getPosition() -
Method in class org.apache.hadoop.io.DataInputBuffer
Returns the current position in the input.
getPosition() -
Method in class org.apache.hadoop.io.InputBuffer
Returns the current position in the input.
getPosition() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Return the current byte position in the input file.
getPreviousIntervalAverageTime() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
Deprecated. The average rate of an operation in the previous interval
getPreviousIntervalNumOps() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
Deprecated. The number of operations in the previous interval
getPreviousIntervalValue() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
Deprecated. The Value at the Previous interval
getPreviousIntervalValue() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
Deprecated. The Value at the Previous interval
getPrincipal() -
Method in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
Returns the Kerberos principal used by the authentication handler.
getPriority() -
Method in class org.apache.hadoop.mapred.JobInProgress
getPrivateDistributedCacheDir(String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getProblems() -
Method in exception org.apache.hadoop.mapred.InvalidInputException
Get the complete list of the problems reported.
getProblems() -
Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
Get the complete list of the problems reported.
getProcess() -
Method in class org.apache.hadoop.util.Shell
get the current sub-process executing the given command
getProcessedByteCount() -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
This method reports the processed bytes so far.
getProcessTree() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get the process-tree with latest state.
getProcessTreeDump() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
Get a dump of the process-tree.
getProcResourceValues() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
getProcResourceValues() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain resource status used by current process tree.
getProfile() -
Method in class org.apache.hadoop.mapred.JobInProgress
getProfileEnabled() -
Method in class org.apache.hadoop.mapred.JobConf
Get whether the task profiling is enabled.
getProfileParams() -
Method in class org.apache.hadoop.mapred.JobConf
Get the profiler configuration arguments.
getProfileTaskRange(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
Get the range of maps or reduces to profile.
getProgress() -
Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
getProgress() -
Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
getProgress() -
Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
Gets the Progress object; this has a float (0.0 - 1.0)
indicating the bytes processed by the iterator so far
getProgress() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Report progress as the minimum of all child RR progress.
getProgress() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request progress from proxied RR.
getProgress() -
Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
getProgress() -
Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
return progress based on the amount of data processed so far.
getProgress() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
How much of the input has the RecordReader
consumed i.e.
getProgress() -
Method in class org.apache.hadoop.mapred.LineRecordReader
Get the progress within the split
getProgress() -
Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
Gets the Progress object; this has a float (0.0 - 1.0)
indicating the bytes processed by the iterator so far
getProgress() -
Method in interface org.apache.hadoop.mapred.RecordReader
How much of the input has the RecordReader
consumed i.e.
getProgress() -
Method in interface org.apache.hadoop.mapred.Reporter
Get the progress of the task.
getProgress() -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Return the progress within the input split
getProgress() -
Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
getProgress() -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
Return the progress within the input split
getProgress() -
Method in class org.apache.hadoop.mapred.Task
getProgress() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
getProgress() -
Method in class org.apache.hadoop.mapred.TaskReport
The amount completed, between zero and one.
getProgress() -
Method in class org.apache.hadoop.mapred.TaskStatus
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
The current progress of the record reader through its data.
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
return progress based on the amount of data processed so far.
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
Get the progress within the split
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Return the progress within the input split
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
getProgress() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
Return the progress within the input split
getProgress() -
Method in class org.apache.hadoop.mapreduce.RecordReader
The current progress of the record reader through its data.
getProgress() -
Method in class org.apache.hadoop.mapreduce.StatusReporter
Get the current progress.
getProgress() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
getProgress() -
Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
getProgress() -
Method in class org.apache.hadoop.util.Progress
Returns progress in this node.
getProgressible() -
Method in class org.apache.hadoop.mapred.JobContext
Get the progress mechanism for reporting progress.
getProgressible() -
Method in class org.apache.hadoop.mapred.TaskAttemptContext
getProperty(String) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
Fetches the value of a property from the configuration file.
getProtocol() -
Method in class org.apache.hadoop.security.authorize.Service
Get the protocol for the service
getProtocolVersion(String, long) -
Method in interface org.apache.hadoop.ipc.VersionedProtocol
Return protocol version corresponding to protocol interface.
getProtocolVersion(String, long) -
Method in class org.apache.hadoop.mapred.JobTracker
getProtocolVersion(String, long) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getProtocolVersion(String, long) -
Method in class org.apache.hadoop.mapred.TaskTracker
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, Configuration, SocketFactory) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, Configuration, SocketFactory, int) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, UserGroupInformation, Configuration, SocketFactory) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, UserGroupInformation, Configuration, SocketFactory, int) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, UserGroupInformation, Configuration, SocketFactory, int, RetryPolicy, boolean) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, Configuration) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a client-side proxy object with the default SocketFactory
getProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, Configuration, int) -
Static method in class org.apache.hadoop.ipc.RPC
getProxySuperuserGroupConfKey(String) -
Static method in class org.apache.hadoop.security.authorize.ProxyUsers
Returns configuration key for effective user groups allowed for a superuser
getProxySuperuserIpConfKey(String) -
Static method in class org.apache.hadoop.security.authorize.ProxyUsers
Return configuration key for superuser ip addresses
getPublicDistributedCacheDir() -
Static method in class org.apache.hadoop.mapred.TaskTracker
getQueueAclsForCurrentUser() -
Method in class org.apache.hadoop.mapred.JobClient
Gets the Queue ACLs for current user
getQueueAclsForCurrentUser() -
Method in class org.apache.hadoop.mapred.JobTracker
getQueueAclsForCurrentUser() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getQueueAdmins(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getQueueAdmins(String) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getQueueInfo(String) -
Method in class org.apache.hadoop.mapred.JobClient
Gets the queue information associated to a particular Job Queue
getQueueInfo(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getQueueInfo(String) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getQueueInfoJson() -
Method in class org.apache.hadoop.mapred.JobTracker
getQueueInfoJson() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getQueueManager() -
Method in class org.apache.hadoop.mapred.JobTracker
Return the QueueManager
associated with the JobTracker.
getQueueMetrics() -
Method in class org.apache.hadoop.mapred.JobInProgress
Get the QueueMetrics object associated with this job
getQueueName() -
Method in class org.apache.hadoop.mapred.JobConf
Return the name of the queue to which this job is submitted.
getQueueName() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the name of the queue to which the job is submitted.
getQueueName() -
Method in class org.apache.hadoop.mapred.JobQueueInfo
Get the queue name from JobQueueInfo
getQueueName() -
Method in class org.apache.hadoop.mapred.QueueAclsInfo
getQueues() -
Method in class org.apache.hadoop.mapred.JobClient
Return an array of queue information objects about all the Job Queues
configured.
getQueues() -
Method in class org.apache.hadoop.mapred.JobTracker
getQueues() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getQueueState() -
Method in class org.apache.hadoop.mapred.JobQueueInfo
Return the queue state
getQuota() -
Method in class org.apache.hadoop.fs.ContentSummary
Return the directory quota
getRack(String) -
Method in class org.apache.hadoop.net.NetworkTopology
Given a string representation of a rack for a specific network
location
To be overridden in subclasses for specific NetworkTopology
implementations, as alternative to overriding the full
NetworkTopology.getRack(String)
method.
getRack(String) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
getRange(String, String) -
Method in class org.apache.hadoop.conf.Configuration
Parse the given attribute as a set of integer ranges
getRaw(String) -
Method in class org.apache.hadoop.conf.Configuration
Get the value of the name
property, without doing
variable expansion.
getRaw() -
Method in class org.apache.hadoop.fs.LocalFileSystem
getRawFileSystem() -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
get the raw file system
getReader() -
Method in class org.apache.hadoop.contrib.failmon.LogParser
Return the BufferedReader, that reads the log file
getReaders(FileSystem, Path, Configuration) -
Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
Open the output generated by this format.
getReaders(Configuration, Path) -
Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
Open the output generated by this format.
getReadOps() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
Get the number of file system read operations such as list files
getReadyJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getReadyJobsList() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getRealm() -
Method in class org.apache.hadoop.security.KerberosName
Get the realm of the name.
getRealUser() -
Method in class org.apache.hadoop.security.UserGroupInformation
get RealUser (vs.
getReasonsForBlacklisting(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getReasonsForGraylisting(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getRecordName() -
Method in interface org.apache.hadoop.metrics.MetricsRecord
Deprecated. Returns the record name.
getRecordName() -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
Deprecated. Returns the record name.
getRecordNum() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
Get the RecordNum corresponding to the entry pointed by the cursor.
getRecordNumNear(long) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
Get the RecordNum for the first key-value pair in a compressed block
whose byte offset in the TFile is greater than or equal to the specified
offset.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.contrib.index.example.LineDocInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.examples.MultiFileWordCount.MyInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.examples.SleepJob.SleepInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.FileInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in interface org.apache.hadoop.mapred.InputFormat
Get the RecordReader
for the given InputSplit
.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in interface org.apache.hadoop.mapred.join.ComposableInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Construct a CompositeRecordReader for the children of this InputFormat
as defined in the init expression.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
This is not implemented yet.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Get the RecordReader
for the given InputSplit
.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.lib.DelegatingInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.MultiFileInputFormat
Deprecated.
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFilter
Create a record reader for the given split
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.mapred.TextInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.streaming.AutoInputFormat
getRecordReader(InputSplit, JobConf, Reporter) -
Method in class org.apache.hadoop.streaming.StreamInputFormat
getRecordReaderQueue() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return sorted list of RecordReaders for this composite.
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.FileOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
Get the RecordWriter
for the given job.
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Create a composite record writer that can write key/value data to different
output files
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.lib.NullOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.MapFileOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in interface org.apache.hadoop.mapred.OutputFormat
Get the RecordWriter
for the given job.
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
getRecordWriter(FileSystem, JobConf, String, Progressable) -
Method in class org.apache.hadoop.mapred.TextOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
Get the RecordWriter
for the given task.
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
getRecordWriter(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.OutputFormat
Get the RecordWriter
for the given task.
getRecoveryDuration() -
Method in class org.apache.hadoop.mapred.JobTracker
How long the jobtracker took to recover from restart.
getReduceCounters(Counters) -
Method in class org.apache.hadoop.mapred.JobInProgress
Returns map phase counters by summing over all map tasks in progress.
getReduceDebugScript() -
Method in class org.apache.hadoop.mapred.JobConf
Get the reduce task's debug Script
getReducerClass() -
Method in class org.apache.hadoop.mapred.JobConf
Get the Reducer
class for the job.
getReducerClass() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the Reducer
class for the job.
getReducerMaxSkipGroups(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of acceptable skip groups surrounding the bad group PER
bad group in reducer.
getReduceSlotCapacity() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of reduce slots in the cluster.
getReduceSpeculativeExecution() -
Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job for reduce tasks?
Defaults to true
.
getReduceTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the reduce tasks of a job.
getReduceTaskReports(String) -
Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getReduceTaskReports(JobID)
getReduceTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getReduceTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getReduceTasks() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of currently running reduce tasks in the cluster.
getRemaining() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
getRemaining() -
Method in interface org.apache.hadoop.io.compress.Decompressor
Returns the number of bytes remaining in the compressed-data buffer;
typically called after the decompressor has finished decompressing
the current gzip stream (a.k.a.
getRemaining() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
Returns 0
.
getRemaining() -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
Returns the number of bytes remaining in the input buffer; normally
called when finished() is true to determine amount of post-gzip-stream
data.
getRemaining() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
Returns the number of bytes remaining in the input buffers; normally
called when finished() is true to determine amount of post-gzip-stream
data.
getRemainingArgs() -
Method in class org.apache.hadoop.util.GenericOptionsParser
Returns an array of Strings containing only application-specific arguments.
getRemoteAddress() -
Static method in class org.apache.hadoop.ipc.Server
Returns remote address as a string when invoked inside an RPC.
getRemoteIp() -
Static method in class org.apache.hadoop.ipc.Server
Returns the remote side ip address when invoked inside an RPC
Returns null incase of an error.
getRenewDate() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager.DelegationTokenInformation
returns renew date
getRenewer() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getReplication() -
Method in class org.apache.hadoop.fs.FileStatus
Get the replication factor of a file.
getReplication(Path) -
Method in class org.apache.hadoop.fs.FileSystem
Deprecated. Use getFileStatus() instead
getReplication(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
Deprecated.
getReport() -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
log the counters
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
getReport() -
Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
getReport() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
getReportDetails() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
getReportItems() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
getRequestURL() -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
Quote the url so that users specifying the HOST HTTP header
can't inject attacks.
getRequestURL(HttpServletRequest) -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the full URL of the request including the query string.
getReservedMapSlots() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get number of reserved map slots in the cluster.
getReservedReduceSlots() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of reserved reduce slots in the cluster.
getResource(String) -
Method in class org.apache.hadoop.conf.Configuration
Get the URL
for the named resource.
getResourceBundle(String) -
Static method in class org.apache.hadoop.mapreduce.CounterGroup
Returns the specified resource bundle, or throws an exception.
getResourceCalculatorPlugin(Class<? extends ResourceCalculatorPlugin>, Configuration) -
Static method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Get the ResourceCalculatorPlugin from the class name and configure it.
getResult() -
Method in class org.apache.hadoop.examples.Sort
Get the last job that was run using this instance.
getRetainHours() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobCompletedEvent
Get the number of hours for which job logs should be retained.
getRevision() -
Static method in class org.apache.hadoop.util.VersionInfo
Get the subversion revision number for the root directory
getRotations() -
Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
getRpcMetrics() -
Method in class org.apache.hadoop.ipc.Server
Returns a handle to the rpcMetrics (required in tests)
getRpcPort() -
Method in class org.apache.hadoop.mapred.TaskTracker
getRpcPort() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getRunAsUser(JobConf) -
Method in class org.apache.hadoop.mapred.TaskController
Returns the local unix user that a given job will run as.
getRunnable() -
Method in class org.apache.hadoop.util.Daemon
getRunningJobList() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getRunningJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getRunningJobs() -
Method in class org.apache.hadoop.mapred.JobTracker
Version that is called from a timer thread, and therefore needs to be
careful to synchronize.
getRunningMaps() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of running map tasks in the cluster.
getRunningReduces() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of running reduce tasks in the cluster.
getRunningTaskAttempts() -
Method in class org.apache.hadoop.mapred.TaskReport
Get the running task attempt IDs for this task
getRunState() -
Method in class org.apache.hadoop.mapred.JobStatus
getRunState() -
Method in class org.apache.hadoop.mapred.TaskStatus
getSample(InputFormat<K, V>, JobConf) -
Method in class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler
For each split sampled, emit when the ratio of the number of records
retained to the total record count is less than the specified
frequency.
getSample(InputFormat<K, V>, JobConf) -
Method in class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
Randomize the split order, then take the specified number of keys from
each split sampled, where each key is selected with the specified
probability and possibly replaced by a subsequently selected key when
the quota of keys from that split is satisfied.
getSample(InputFormat<K, V>, JobConf) -
Method in interface org.apache.hadoop.mapred.lib.InputSampler.Sampler
For a given job, collect and return a subset of the keys from the
input data.
getSample(InputFormat<K, V>, JobConf) -
Method in class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
From each split sampled, take the first numSamples / numSplits records.
getSample(InputFormat<K, V>, Job) -
Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
For each split sampled, emit when the ratio of the number of records
retained to the total record count is less than the specified
frequency.
getSample(InputFormat<K, V>, Job) -
Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
Randomize the split order, then take the specified number of keys from
each split sampled, where each key is selected with the specified
probability and possibly replaced by a subsequently selected key when
the quota of keys from that split is satisfied.
getSample(InputFormat<K, V>, Job) -
Method in interface org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler
For a given job, collect and return a subset of the keys from the
input data.
getSample(InputFormat<K, V>, Job) -
Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
From each split sampled, take the first numSamples / numSplits records.
getSaslQop() -
Method in enum org.apache.hadoop.security.SaslRpcServer.QualityOfProtection
getSchedulingInfo() -
Method in class org.apache.hadoop.mapred.JobInProgress
getSchedulingInfo() -
Method in class org.apache.hadoop.mapred.JobQueueInfo
Gets the scheduling information associated to particular job queue.
getSchedulingInfo() -
Method in class org.apache.hadoop.mapred.JobStatus
Gets the Scheduling information associated to a particular Job.
getScheme() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
Get the uri scheme associated with this statistics object.
getSecond() -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
getSecretAccessKey() -
Method in class org.apache.hadoop.fs.s3.S3Credentials
getSecretKey(Credentials, Text) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
auxiliary method to get user's secret keys..
getSecretKey(Text) -
Method in class org.apache.hadoop.security.Credentials
Returns the key bytes for the alias
getSelectQuery() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
Returns the query for selecting the records,
subclasses can override this for custom behaviour.
getSelectQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader
Returns the query for selecting the records,
subclasses can override this for custom behaviour.
getSelectQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Returns the query for selecting the records,
subclasses can override this for custom behaviour.
getSelectQuery() -
Method in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
Returns the query for selecting the records from an Oracle DB.
getSequenceFileOutputKeyClass(JobConf) -
Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Get the key class for the SequenceFile
getSequenceFileOutputKeyClass(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Get the key class for the SequenceFile
getSequenceFileOutputValueClass(JobConf) -
Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Get the value class for the SequenceFile
getSequenceFileOutputValueClass(JobContext) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Get the value class for the SequenceFile
getSequenceNumber() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
getSequenceWriter(TaskAttemptContext, Class<?>, Class<?>) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
getSerialization(Class<T>) -
Method in class org.apache.hadoop.io.serializer.SerializationFactory
getSerializedLength() -
Method in class org.apache.hadoop.fs.s3.INode
getSerializer(Class<Serializable>) -
Method in class org.apache.hadoop.io.serializer.JavaSerialization
getSerializer(Class<T>) -
Method in interface org.apache.hadoop.io.serializer.Serialization
getSerializer(Class<T>) -
Method in class org.apache.hadoop.io.serializer.SerializationFactory
getSerializer(Class<Writable>) -
Method in class org.apache.hadoop.io.serializer.WritableSerialization
getServer(Object, String, int, Configuration) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a server for a protocol implementation instance listening on a
port and address.
getServer(Object, String, int, int, boolean, Configuration) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a server for a protocol implementation instance listening on a
port and address.
getServer(Object, String, int, int, boolean, Configuration, SecretManager<? extends TokenIdentifier>) -
Static method in class org.apache.hadoop.ipc.RPC
Construct a server for a protocol implementation instance listening on a
port and address, with a secret manager.
getServerAddress(Configuration, String, String, String) -
Static method in class org.apache.hadoop.net.NetUtils
Deprecated.
getServerName() -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
Quote the server name so that users specifying the HOST HTTP header
can't inject attacks.
getServerPrincipal(String, String) -
Static method in class org.apache.hadoop.security.SecurityUtil
Convert Kerberos principal name pattern to valid Kerberos principal
names.
getServerPrincipal(String, InetAddress) -
Static method in class org.apache.hadoop.security.SecurityUtil
Convert Kerberos principal name pattern to valid Kerberos principal names.
getServerVersion() -
Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
Get the server's agreed to version.
getService() -
Method in class org.apache.hadoop.security.token.Token
Get the service on which the token is supposed to be used
getServiceKey() -
Method in class org.apache.hadoop.security.authorize.Service
Get the configuration key for the service.
getServiceName() -
Method in class org.apache.hadoop.security.KerberosName
Get the first component of the name.
getServicePrincipal(String, String) -
Static method in class org.apache.hadoop.security.authentication.util.KerberosUtil
Create Kerberos principal for a given service and hostname.
getServices() -
Method in class org.apache.hadoop.mapred.MapReducePolicyProvider
getServices() -
Method in class org.apache.hadoop.security.authorize.PolicyProvider
Get the Service
definitions from the PolicyProvider
.
getSessionId() -
Method in class org.apache.hadoop.mapred.JobConf
Get the user-specified session identifier.
getSetupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the setup tasks of a job.
getSetupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
getSetupTaskReports(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getShape(boolean, int) -
Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
getShortName() -
Method in class org.apache.hadoop.security.KerberosName
Get the translation of the principal name into an operating system
user name.
getShortUserName() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the user's login name.
getShuffleFinishTime() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get shuffle finish time for the task.
getSize() -
Method in class org.apache.hadoop.io.BytesWritable
Deprecated. Use BytesWritable.getLength()
instead.
getSize() -
Method in interface org.apache.hadoop.io.SequenceFile.ValueBytes
Size of stored data.
getSize() -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
getSize() -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
getSkipOutputPath(Configuration) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the directory to which skipped records are written.
getSkipRanges() -
Method in class org.apache.hadoop.mapred.Task
Get skipRanges.
getSlope(String) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
Deprecated.
getSocketFactory(Configuration, Class<?>) -
Static method in class org.apache.hadoop.net.NetUtils
Get the socket factory for the given class according to its
configuration parameter
hadoop.rpc.socket.factory.class.<ClassName>.
getSocketFactoryFromProperty(Configuration, String) -
Static method in class org.apache.hadoop.net.NetUtils
Get the socket factory corresponding to the given proxy URI.
getSortComparator() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the RawComparator
comparator used to compare keys.
getSortFinishTime() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get sort finish time for the task,.
getSpace(int) -
Static method in class org.apache.hadoop.streaming.StreamUtil
getSpaceConsumed() -
Method in class org.apache.hadoop.fs.ContentSummary
Retuns (disk) space consumed
getSpaceQuota() -
Method in class org.apache.hadoop.fs.ContentSummary
Returns (disk) space quota
getSpeculativeExecution() -
Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job?
Defaults to true
.
getSplit() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getSplitHosts(BlockLocation[], long, long, NetworkTopology) -
Method in class org.apache.hadoop.mapred.FileInputFormat
This function identifies and returns the hosts that contribute
most for a given split.
getSplitIndex() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
getSplitLocation() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
getSplitLocation() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
getSplits(int) -
Method in class org.apache.hadoop.examples.dancing.Pentomino
Generate a list of prefixes to a given depth
getSplits(JobConf, int) -
Method in class org.apache.hadoop.examples.SleepJob.SleepInputFormat
getSplits(JobConf, int) -
Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.FileInputFormat
Splits files returned by FileInputFormat.listStatus(JobConf)
when
they're too big.
getSplits(JobConf, int) -
Method in interface org.apache.hadoop.mapred.InputFormat
Logically split the set of input files for the job.
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Build a CompositeInputSplit from the child InputFormats by assigning the
ith split from each child to the ith composite split.
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Logically split the set of input files for the job.
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.lib.DelegatingInputFormat
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
Logically splits the set of input files for the job, splits N lines
of the input as one split.
getSplits(JobConf, int) -
Method in class org.apache.hadoop.mapred.MultiFileInputFormat
Deprecated.
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.InputFormat
Logically split the set of input files for the job.
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
Logically split the set of input files for the job.
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Logically split the set of input files for the job.
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingInputFormat
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Generate the list of files and make them into FileSplits.
getSplits(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
Logically splits the set of input files for the job, splits N lines
of the input as one split.
getSplitsForFile(FileStatus, Configuration, int) -
Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
getSplitter(int) -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
getSplitter(int) -
Method in class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat
getSrcChecksum() -
Static method in class org.apache.hadoop.util.VersionInfo
Get the checksum of the source files from which Hadoop was
built.
getStackTrace() -
Method in exception org.apache.hadoop.security.authorize.AuthorizationException
getStagingAreaDir() -
Method in class org.apache.hadoop.mapred.JobClient
Grab the jobtracker's view of the staging directory path where
job-specific files will be placed.
getStagingAreaDir() -
Method in class org.apache.hadoop.mapred.JobTracker
getStagingAreaDir() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getStagingDir(JobClient, Configuration) -
Static method in class org.apache.hadoop.mapreduce.JobSubmissionFiles
Initializes the staging directory and returns the path.
getStart() -
Method in class org.apache.hadoop.mapred.FileSplit
The position of the first byte in the file to process.
getStart() -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
getStart() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
getStart() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The position of the first byte in the file to process.
getStartOffset() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
getStartOffset() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
getStartOffset() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
getStartOffsets() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
Returns an array containing the startoffsets of the files in the split
getStartOffsets() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns an array containing the start offsets of the files in the split
getStartTime() -
Method in class org.apache.hadoop.mapred.JobInProgress
getStartTime() -
Method in class org.apache.hadoop.mapred.JobStatus
getStartTime() -
Method in class org.apache.hadoop.mapred.JobTracker
getStartTime() -
Method in class org.apache.hadoop.mapred.TaskReport
Get start time of task.
getStartTime() -
Method in class org.apache.hadoop.mapred.TaskStatus
Get start time of the task.
getState(String) -
Static method in class org.apache.hadoop.contrib.failmon.PersistentState
Read and return the state of parsing for a particular log file.
getState() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
getState() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getState() -
Method in class org.apache.hadoop.mapred.TaskReport
The most recent state, reported by a Reporter
.
getStatement() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
getStatement() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getStateString() -
Method in class org.apache.hadoop.mapred.TaskStatus
getStaticResolution(String) -
Static method in class org.apache.hadoop.net.NetUtils
Retrieves the resolved name for the passed host.
getStatistics() -
Static method in class org.apache.hadoop.fs.FileSystem
Deprecated. use FileSystem.getAllStatistics()
instead
getStatistics(String, Class<? extends FileSystem>) -
Static method in class org.apache.hadoop.fs.FileSystem
Get the statistics for a particular file system
getStatus() -
Method in class org.apache.hadoop.mapred.JobInProgress
getStatus() -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
Get the current TaskTrackerStatus
of the TaskTracker
.
getStatus() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
Get the last set status message.
getStr() -
Method in class org.apache.hadoop.mapred.join.Parser.StrToken
getStr() -
Method in class org.apache.hadoop.mapred.join.Parser.Token
getStringCollection(String) -
Method in class org.apache.hadoop.conf.Configuration
Get the comma delimited values of the name
property as
a collection of String
s.
getStringCollection(String) -
Static method in class org.apache.hadoop.util.StringUtils
Returns a collection of strings.
getStrings(String) -
Method in class org.apache.hadoop.conf.Configuration
Get the comma delimited values of the name
property as
an array of String
s.
getStrings(String, String...) -
Method in class org.apache.hadoop.conf.Configuration
Get the comma delimited values of the name
property as
an array of String
s.
getStrings(String) -
Static method in class org.apache.hadoop.util.StringUtils
Returns an arraylist of strings.
getSubject() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the underlying subject from this ugi.
getSuccessfulJobList() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getSuccessfulJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getSuccessfulTaskAttempt() -
Method in class org.apache.hadoop.mapred.TaskReport
Get the attempt ID that took this task to completion
GetSuffix(int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getSum() -
Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
getSum() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
getSummaryJson() -
Method in class org.apache.hadoop.mapred.JobTracker
getSummaryJson() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getSupportedCompressionAlgorithms() -
Static method in class org.apache.hadoop.io.file.tfile.TFile
Get names of supported compression algorithms.
getSymlink(Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
This method checks to see if symlinks are to be create for the
localized cache files in the current working directory
Used by internal DistributedCache code.
getSystemDir() -
Method in class org.apache.hadoop.mapred.JobClient
Grab the jobtracker system directory path where job-specific files are to be placed.
getSystemDir() -
Method in class org.apache.hadoop.mapred.JobTracker
getSystemDir() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getTableName() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
getTabSize(int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
getTag() -
Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
getTag(String) -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns a tag object which is can be a String, Integer, Short or Byte.
getTag(String) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache.Record
Get the tag value
getTagNames() -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns the set of tag names
getTagsCopy() -
Method in class org.apache.hadoop.metrics.spi.OutputRecord
Deprecated. Returns a copy of this record's tags.
getTask() -
Method in class org.apache.hadoop.mapred.JvmTask
getTask(JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
Called upon startup by the child process, to fetch Task data.
getTask(JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
Called when a child task process starts, to get its task.
getTaskAttemptID() -
Method in class org.apache.hadoop.mapred.TaskAttemptContext
Get the taskAttemptID.
getTaskAttemptId() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns task id.
getTaskAttemptID() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
Get the unique name for this task attempt.
getTaskAttemptIDsPattern(String, Integer, Boolean, Integer, Integer) -
Static method in class org.apache.hadoop.mapred.TaskAttemptID
Deprecated.
getTaskAttemptLogDir(TaskAttemptID, String, String[]) -
Static method in class org.apache.hadoop.mapred.TaskLog
Get attempt log directory path for the given attempt-id under randomly
selected mapred local directory.
getTaskAttempts() -
Method in class org.apache.hadoop.mapred.JobHistory.Task
Returns all task attempts for this task.
getTaskCompletionEvents(int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
getTaskCompletionEvents(JobID, int, int) -
Method in class org.apache.hadoop.mapred.JobTracker
getTaskCompletionEvents(JobID, int, int) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
getTaskCompletionEvents(int) -
Method in interface org.apache.hadoop.mapred.RunningJob
Get events indicating completion (success/failure) of component tasks.
getTaskCompletionEvents(int) -
Method in class org.apache.hadoop.mapreduce.Job
Get events indicating completion (success/failure) of component tasks.
getTaskController() -
Method in class org.apache.hadoop.mapred.TaskTracker
getTaskController() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
Get the taskController for deleting logs.
getTaskDiagnostics(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.JobTracker
Get the diagnostics for a given task
getTaskDiagnostics(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
Returns the diagnostic information for a particular task in the given job.
getTaskDiagnostics(TaskAttemptID) -
Method in interface org.apache.hadoop.mapred.RunningJob
Gets the diagnostic messages for a given task attempt.
getTaskDistributedCacheManager(JobID) -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
getTaskFailures() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the number of tasks that have failed on this tracker.
getTaskID() -
Method in class org.apache.hadoop.mapred.Task
getTaskID() -
Method in class org.apache.hadoop.mapred.TaskAttemptID
getTaskId() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Deprecated. use TaskCompletionEvent.getTaskAttemptId()
instead.
getTaskId() -
Method in class org.apache.hadoop.mapred.TaskLogAppender
Getter/Setter methods for log4j.
getTaskId() -
Method in class org.apache.hadoop.mapred.TaskReport
Deprecated. use TaskReport.getTaskID()
instead
getTaskID() -
Method in class org.apache.hadoop.mapred.TaskReport
The id of the task.
getTaskID() -
Method in class org.apache.hadoop.mapred.TaskStatus
getTaskID() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Returns the TaskID
object that this task attempt belongs to
getTaskIDsPattern(String, Integer, Boolean, Integer) -
Static method in class org.apache.hadoop.mapred.TaskID
Deprecated.
getTaskInfo(JobConf) -
Static method in class org.apache.hadoop.streaming.StreamUtil
getTaskInProgress(TaskID) -
Method in class org.apache.hadoop.mapred.JobInProgress
Return the TaskInProgress that matches the tipid.
getTaskLogFile(TaskAttemptID, boolean, TaskLog.LogName) -
Static method in class org.apache.hadoop.mapred.TaskLog
getTaskLogLength(JobConf) -
Static method in class org.apache.hadoop.mapred.TaskLog
Get the desired maximum length of task's logs.
getTaskLogsUrl(JobHistory.TaskAttempt) -
Static method in class org.apache.hadoop.mapred.JobHistory
Return the TaskLogsUrl of a particular TaskAttempt
getTaskLogUrl(String, String, String) -
Static method in class org.apache.hadoop.mapred.TaskLogServlet
Construct the taskLogUrl
getTaskMemoryManager() -
Method in class org.apache.hadoop.mapred.TaskTracker
getTaskOutputFilter(JobConf) -
Static method in class org.apache.hadoop.mapred.JobClient
Get the task output filter out of the JobConf.
getTaskOutputFilter() -
Method in class org.apache.hadoop.mapred.JobClient
Deprecated.
getTaskOutputPath(JobConf, String) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to create the task's temporary output directory and
return the path to the task's output file.
getTaskReports() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
Get the current tasks at the TaskTracker.
getTaskRunTime() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns time (in millisec) the task took to complete.
getTasksInfoJson() -
Method in class org.apache.hadoop.mapred.TaskTracker
getTasksInfoJson() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getTaskStatus() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns enum Status.SUCESS or Status.FAILURE.
getTaskTracker(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getTaskTracker() -
Method in class org.apache.hadoop.mapred.TaskStatus
getTaskTrackerCount() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of active trackers in the cluster.
getTaskTrackerHttp() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
http location of the tasktracker where this task ran.
getTaskTrackerInstrumentation() -
Method in class org.apache.hadoop.mapred.TaskTracker
getTaskTrackerReportAddress() -
Method in class org.apache.hadoop.mapred.TaskTracker
Return the port at which the tasktracker bound to
getTaskTrackers() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of active task trackers in the cluster.
getTaskTrackerStatus(String) -
Method in class org.apache.hadoop.mapred.JobTracker
getTerm() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the term.
getText() -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Get the text that represents a document.
getText() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
The text of the document id.
getThreadCount() -
Method in class org.apache.hadoop.mapred.JobTracker
getThreadCount() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getThreadState() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getTimeout(Configuration) -
Static method in class org.apache.hadoop.ipc.Client
The time after which a RPC will timeout.
getTimestamp(Configuration, URI) -
Static method in class org.apache.hadoop.filecache.DistributedCache
Returns mtime of a given cache file on hdfs.
getTip(TaskID) -
Method in class org.apache.hadoop.mapred.JobTracker
Returns specified TaskInProgress, or null.
getTmax(String) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
Deprecated.
getToken(int) -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
getToken(HttpServletRequest) -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the AuthenticationToken
for the request.
getToken(Text) -
Method in class org.apache.hadoop.security.Credentials
Returns the Token object for the alias
getTokenIdentifiers() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the set of TokenIdentifiers belonging to this UGI
getTokens() -
Method in class org.apache.hadoop.security.UserGroupInformation
Obtain the collection of tokens associated with this user.
getTokenServiceAddr(Token<?>) -
Static method in class org.apache.hadoop.security.SecurityUtil
Decode the given token's service field into an InetAddress
getTopologyPaths() -
Method in class org.apache.hadoop.fs.BlockLocation
Get the list of network topology paths for each of the hosts.
getTotalJobSubmissions() -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of job submissions in the cluster.
getTotalLogFileSize() -
Method in class org.apache.hadoop.mapred.TaskLogAppender
getTotalSubmissions() -
Method in class org.apache.hadoop.mapred.JobTracker
getTrackerIdentifier() -
Method in class org.apache.hadoop.mapred.JobTracker
Get the unique identifier (ie.
getTrackerName() -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
getTrackerName() -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
Get the unique identifier for the TaskTracker
getTrackerPort() -
Method in class org.apache.hadoop.mapred.JobTracker
getTrackingURL() -
Method in interface org.apache.hadoop.mapred.RunningJob
Get the URL where some job progress information will be displayed.
getTrackingURL() -
Method in class org.apache.hadoop.mapreduce.Job
Get the URL where some job progress information will be displayed.
getTrash() -
Method in class org.apache.hadoop.fs.FsShell
getTTExpiryInterval() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Get the tasktracker expiry interval for the cluster
getType() -
Method in class org.apache.hadoop.mapred.join.Parser.Token
getType() -
Method in interface org.apache.hadoop.security.authentication.server.AuthenticationHandler
Returns the authentication type of the authentication handler.
getType() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
Returns the authentication mechanism of the token.
getType() -
Method in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
Returns the authentication type of the authentication handler, 'kerberos'.
getType() -
Method in class org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler
Returns the authentication type of the authentication handler, 'simple'.
getType() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Get the type code embedded in the first byte.
getTypeID() -
Method in class org.apache.hadoop.record.meta.FieldTypeInfo
get the field's TypeID object
getTypes() -
Method in class org.apache.hadoop.io.GenericWritable
Return all classes that may be wrapped.
getTypeVal() -
Method in class org.apache.hadoop.record.meta.TypeID
Get the type value.
getUlimitMemoryCommand(int) -
Static method in class org.apache.hadoop.util.Shell
Get the Unix command for setting the maximum virtual memory available
to a given child process.
getUlimitMemoryCommand(Configuration) -
Static method in class org.apache.hadoop.util.Shell
Deprecated. Use Shell.getUlimitMemoryCommand(int)
getUMask(Configuration) -
Static method in class org.apache.hadoop.fs.permission.FsPermission
Get the user file creation mask (umask)
UMASK_LABEL
config param has umask value that is either symbolic
or octal.
getUniqueFile(TaskAttemptContext, String, String) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Generate a unique filename, based on the task id, name, and extension
getUniqueItems() -
Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
getUniqueName(JobConf, String) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to generate a name that is unique for the task.
getUnits(String) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
Deprecated.
getUpperClause() -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
getUri() -
Method in class org.apache.hadoop.fs.FileSystem
Returns a URI whose scheme and authority identify this FileSystem.
getUri() -
Method in class org.apache.hadoop.fs.FilterFileSystem
Returns a URI whose scheme and authority identify this FileSystem.
getUri() -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
getUri() -
Method in class org.apache.hadoop.fs.HarFileSystem
Returns the uri of this filesystem.
getUri() -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
getUri() -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
getUri() -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
getUri() -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
getURIs(String, String) -
Method in class org.apache.hadoop.streaming.StreamJob
get the uris of all the files/caches
getURL() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the link to the web-ui for details of the job.
getUrl() -
Static method in class org.apache.hadoop.util.VersionInfo
Get the subversion URL for the root Hadoop directory.
getUsed() -
Method in class org.apache.hadoop.fs.DF
getUsed() -
Method in class org.apache.hadoop.fs.DU
getUsed() -
Method in class org.apache.hadoop.fs.FileSystem
Return the total size of all files in the filesystem.
getUsedMemory() -
Method in class org.apache.hadoop.mapred.ClusterStatus
Deprecated.
getUseNewMapper() -
Method in class org.apache.hadoop.mapred.JobConf
Should the framework use the new context-object code for running
the mapper?
getUseNewReducer() -
Method in class org.apache.hadoop.mapred.JobConf
Should the framework use the new context-object code for running
the reducer?
getUser() -
Method in class org.apache.hadoop.mapred.JobConf
Get the reported username for this job.
getUser() -
Method in class org.apache.hadoop.mapred.JobInProgress
Get the user for the job
getUser() -
Method in class org.apache.hadoop.mapred.JobProfile
Get the user id.
getUser() -
Method in class org.apache.hadoop.mapred.Task
Get the name of the user running the job/task.
getUser() -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
Get the Ugi with the username encoded in the token identifier
getUser() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
Get the username encoded in the token identifier
getUser() -
Method in class org.apache.hadoop.security.token.TokenIdentifier
Get the Ugi with the username encoded in the token identifier
getUser() -
Static method in class org.apache.hadoop.util.VersionInfo
The user that compiled Hadoop.
getUserAction() -
Method in class org.apache.hadoop.fs.permission.FsPermission
Return user FsAction
.
getUserDir(String) -
Static method in class org.apache.hadoop.mapred.TaskTracker
getUserLogCleaner() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
Get UserLogCleaner
.
getUserLogDir() -
Static method in class org.apache.hadoop.mapred.TaskLog
getUserName() -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
Return user name
getUserName(JobConf) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the user name from the job conf
getUsername() -
Method in class org.apache.hadoop.mapred.JobStatus
getUserName() -
Method in class org.apache.hadoop.security.authentication.client.PseudoAuthenticator
Returns the current user name.
getUserName() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
Returns the user name.
getUserName() -
Method in class org.apache.hadoop.security.UserGroupInformation
Get the user's full principal name.
getUsersForNetgroup(String) -
Method in class org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMapping
Calls JNI function to get users for a netgroup, since C functions
are not reentrant we need to make this synchronized (see
documentation for setnetgrent, getnetgrent and endnetgrent)
getUsersForNetgroupCommand(String) -
Static method in class org.apache.hadoop.util.Shell
a Unix command to get a given netgroup's user list
getUserToGroupsMappingService() -
Static method in class org.apache.hadoop.security.Groups
Get the groups being used to map user-to-groups.
getUserToGroupsMappingService(Configuration) -
Static method in class org.apache.hadoop.security.Groups
getVal() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
getVal() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
getVal() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
getVal() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
getValByRegex(String) -
Method in class org.apache.hadoop.conf.Configuration
get keys matching the the regex
getValidity() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
Returns the validity time of the generated tokens.
getValue(BytesWritable) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy the value into BytesWritable.
getValue(byte[]) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy value into user-supplied buffer.
getValue(byte[], int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Copy value into user-supplied buffer.
getValue() -
Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
Gets the current raw value
getValue() -
Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
Gets the current raw value.
getValue() -
Method in class org.apache.hadoop.mapreduce.Counter
What is the current value of this counter?
getValue() -
Method in enum org.apache.hadoop.mapreduce.JobStatus.State
getValue() -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
getValue() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Get the typed bytes as a Java object.
getValue() -
Method in class org.apache.hadoop.util.DataChecksum
getValue() -
Method in enum org.apache.hadoop.util.ProcessTree.Signal
getValue() -
Method in class org.apache.hadoop.util.PureJavaCrc32
getValueClass() -
Method in class org.apache.hadoop.io.ArrayWritable
getValueClass() -
Method in class org.apache.hadoop.io.MapFile.Reader
Returns the class of values in this file.
getValueClass() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the class of values in this file.
getValueClass() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
Returns the class of values in this file.
getValueClass() -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
The class of value that must be passed to SequenceFileRecordReader.next(Object, Object)
..
getValueClassName() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
Returns the name of the value class.
getValueClassName() -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the value class for this SequenceFile.
getValueClassName() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the value class for this SequenceFile.
getValueLength() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Get the length of the value.
getValues() -
Method in class org.apache.hadoop.mapreduce.ReduceContext
Iterate through the values for the current key, reusing the same value
object, which is stored in the context.
getValueStream() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
Stream access to value.
getValueTypeID() -
Method in class org.apache.hadoop.record.meta.MapTypeID
get the TypeID of the map's value element
getVectorSize() -
Method in class org.apache.hadoop.util.bloom.BloomFilter
getVersion() -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the version number of the entire index.
getVersion() -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
getVersion() -
Method in class org.apache.hadoop.io.VersionedWritable
Return the version number of the current implementation.
getVersion() -
Method in class org.apache.hadoop.mapred.JobTracker
getVersion() -
Method in interface org.apache.hadoop.mapred.JobTrackerMXBean
getVersion() -
Method in class org.apache.hadoop.mapred.TaskTracker
getVersion() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
getVersion() -
Static method in class org.apache.hadoop.util.VersionInfo
Get the Hadoop version.
getVIntSize(long) -
Static method in class org.apache.hadoop.io.WritableUtils
Get the encoded length if an integer is stored in a variable-length format
getVIntSize(long) -
Static method in class org.apache.hadoop.record.Utils
Get the encoded length if an integer is stored in a variable-length format
getVirtualMemorySize() -
Method in class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin
Deprecated. Obtain the total size of the virtual memory present in the system.
getVirtualMemorySize() -
Method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
Obtain the total size of the virtual memory present in the system.
getVirtualMemorySize() -
Method in class org.apache.hadoop.util.MemoryCalculatorPlugin
Deprecated. Obtain the total size of the virtual memory present in the system.
getVirtualMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin
Obtain the total size of the virtual memory present in the system.
getVirtualMemorySize() -
Method in class org.apache.hadoop.util.ResourceCalculatorPlugin.ProcResourceValues
Obtain the virtual memory size used by a current process tree.
getVIVersion() -
Method in class org.apache.hadoop.mapred.JobTracker
getWaitingJobList() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
getWaitingJobs() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
getWarn() -
Static method in class org.apache.hadoop.log.metrics.EventCounter
getWebAppsPath() -
Method in class org.apache.hadoop.http.HttpServer
Get the pathname to the webapps files.
getWeight() -
Method in class org.apache.hadoop.util.bloom.Key
getWorkflowAdjacencies(Configuration) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
Get the workflow adjacencies from the job conf
The string returned is of the form "key"="value" "key"="value" ...
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.FileSystem
Get the current working directory for the given file system
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.FilterFileSystem
Get the current working directory for the given file system
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.HarFileSystem
return the top level archive.
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
getWorkingDirectory() -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
getWorkingDirectory() -
Method in class org.apache.hadoop.mapred.JobConf
Get the current working directory for the default file system.
getWorkingDirectory() -
Method in class org.apache.hadoop.mapreduce.JobContext
Get the current working directory for the default file system.
getWorkOutputPath(JobConf) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the Path
to the task's temporary output directory
for the map-reduce job
Tasks' Side-Effect Files
- getWorkOutputPath(TaskInputOutputContext<?, ?, ?, ?>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
- Get the
Path
to the task's temporary output directory
for the map-reduce job
Tasks' Side-Effect Files
- getWorkPath() -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
- Get the directory that the task should write results into
- getWrappedStream() -
Method in class org.apache.hadoop.fs.FSDataOutputStream
-
- getWriteOps() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Get the number of file system write operations such as create, append
rename etc.
- getZlibCompressor(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
- Return the appropriate implementation of the zlib compressor.
- getZlibCompressorType(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
- Return the appropriate type of the zlib compressor.
- getZlibDecompressor(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
- Return the appropriate implementation of the zlib decompressor.
- getZlibDecompressorType(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
- Return the appropriate type of the zlib decompressor.
- GlobFilter - Class in org.apache.hadoop.metrics2.filter
- A glob pattern filter for metrics
- GlobFilter() -
Constructor for class org.apache.hadoop.metrics2.filter.GlobFilter
-
- GlobPattern - Class in org.apache.hadoop.fs
- A class for POSIX glob pattern with brace expansions.
- GlobPattern(String) -
Constructor for class org.apache.hadoop.fs.GlobPattern
- Construct the glob pattern object with a glob pattern string
- globStatus(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Return all the files that match filePattern and are not checksum
files.
- globStatus(Path, PathFilter) -
Method in class org.apache.hadoop.fs.FileSystem
- Return an array of FileStatus objects whose path names match pathPattern
and is accepted by the user-supplied path filter.
- go() -
Method in class org.apache.hadoop.streaming.StreamJob
- Deprecated. use
StreamJob.run(String[])
instead.
- goodClassOrNull(Configuration, String, String) -
Static method in class org.apache.hadoop.streaming.StreamUtil
- It may seem strange to silently switch behaviour when a String
is not a classname; the reason is simplified Usage:
- graylistedTaskTrackers() -
Method in class org.apache.hadoop.mapred.JobTracker
- Get the statuses of the graylisted task trackers in the cluster.
- GREATER_ICOST -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- Grep - Class in org.apache.hadoop.examples
-
- Groups - Class in org.apache.hadoop.security
- A user-to-groups mapping service.
- Groups(Configuration) -
Constructor for class org.apache.hadoop.security.Groups
-
- GT_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- GzipCodec - Class in org.apache.hadoop.io.compress
- This class creates gzip compressors/decompressors.
- GzipCodec() -
Constructor for class org.apache.hadoop.io.compress.GzipCodec
-
- GzipCodec.GzipOutputStream - Class in org.apache.hadoop.io.compress
- A bridge that wraps around a DeflaterOutputStream to make it
a CompressionOutputStream.
- GzipCodec.GzipOutputStream(OutputStream) -
Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- GzipCodec.GzipOutputStream(CompressorStream) -
Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
- Allow children types to put a different type in here.
H
- HADOOP_JETTY_LOGS_SERVE_ALIASES -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- Enable/Disable aliases serving from jetty
- HADOOP_RELAXED_VERSION_CHECK_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- HADOOP_RELAXED_VERSION_CHECK_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_AUTHENTICATION -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_AUTHORIZATION -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_GROUP_MAPPING -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_INSTRUMENTATION_REQUIRES_ADMIN -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_SERVICE_USER_NAME_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_TOKEN_SERVICE_USE_IP -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_SECURITY_TOKEN_SERVICE_USE_IP_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- HADOOP_SECURITY_USE_WEAK_HTTP_CRYPTO_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- HADOOP_SECURITY_USE_WEAK_HTTP_CRYPTO_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- HADOOP_SKIP_VERSION_CHECK_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- HADOOP_SKIP_VERSION_CHECK_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- See src/core/core-default.xml
- HADOOP_TOKEN_FILE_LOCATION -
Static variable in class org.apache.hadoop.security.UserGroupInformation
- Environment variable pointing to the token cache file
- HadoopLogParser - Class in org.apache.hadoop.contrib.failmon
- An object of this class parses a Hadoop log file to create
appropriate EventRecords.
- HadoopLogParser(String) -
Constructor for class org.apache.hadoop.contrib.failmon.HadoopLogParser
- Create a new parser object and try to find the hostname
of the node that generated the log
- HadoopStreaming - Class in org.apache.hadoop.streaming
- The main entrypoint.
- HadoopStreaming() -
Constructor for class org.apache.hadoop.streaming.HadoopStreaming
-
- HadoopVersionAnnotation - Annotation Type in org.apache.hadoop
- A package attribute that captures the version of Hadoop that was compiled.
- halfDigest() -
Method in class org.apache.hadoop.io.MD5Hash
- Construct a half-sized version of this MD5.
- handle(Map<JobHistory.Keys, String>) -
Method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
- handle(JobHistory.RecordTypes, Map<JobHistory.Keys, String>) -
Method in interface org.apache.hadoop.mapred.JobHistory.Listener
- Callback method for history parser.
- handle(Callback[]) -
Method in class org.apache.hadoop.security.SaslRpcServer.SaslDigestCallbackHandler
-
- handle(Callback[]) -
Method in class org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler
-
- handleKind(Text) -
Method in class org.apache.hadoop.mapred.JobClient.Renewer
-
- handleKind(Text) -
Method in class org.apache.hadoop.security.token.Token.TrivialRenewer
-
- handleKind(Text) -
Method in class org.apache.hadoop.security.token.TokenRenewer
- Does this renewer handle this kind of token?
- HardLink - Class in org.apache.hadoop.fs
- Class for creating hardlinks.
- HardLink() -
Constructor for class org.apache.hadoop.fs.HardLink
-
- HardLink.LinkStats - Class in org.apache.hadoop.fs
- HardLink statistics counters and methods.
- HardLink.LinkStats() -
Constructor for class org.apache.hadoop.fs.HardLink.LinkStats
-
- HardLink.OSType - Enum in org.apache.hadoop.fs
-
- HarFileSystem - Class in org.apache.hadoop.fs
- This is an implementation of the Hadoop Archive
Filesystem.
- HarFileSystem() -
Constructor for class org.apache.hadoop.fs.HarFileSystem
- public construction of harfilesystem
- HarFileSystem(FileSystem) -
Constructor for class org.apache.hadoop.fs.HarFileSystem
- Constructor to create a HarFileSystem with an
underlying filesystem.
- has(int) -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- Return true if tuple has an element at the position provided.
- hasAdministratorAccess(ServletContext, HttpServletRequest, HttpServletResponse) -
Static method in class org.apache.hadoop.http.HttpServer
- Does the user sending the HttpServletRequest has the administrator ACLs? If
it isn't the case, response will be modified to send an error to the user.
- HasFileDescriptor - Interface in org.apache.hadoop.fs
- Having a FileDescriptor
- hash -
Variable in class org.apache.hadoop.util.bloom.Filter
- The hash function used to map a key to several positions in the vector.
- hash(Key) -
Method in class org.apache.hadoop.util.bloom.HashFunction
- Hashes a specified key into several integers.
- Hash - Class in org.apache.hadoop.util.hash
- This class represents a common API for hashing functions.
- Hash() -
Constructor for class org.apache.hadoop.util.hash.Hash
-
- hash(byte[]) -
Method in class org.apache.hadoop.util.hash.Hash
- Calculate a hash using all bytes from the input argument, and
a seed of -1.
- hash(byte[], int) -
Method in class org.apache.hadoop.util.hash.Hash
- Calculate a hash using all bytes from the input argument,
and a provided seed value.
- hash(byte[], int, int) -
Method in class org.apache.hadoop.util.hash.Hash
- Calculate a hash using bytes from 0 to
length
, and
the provided seed value
- hash(byte[], int, int) -
Method in class org.apache.hadoop.util.hash.JenkinsHash
- taken from hashlittle() -- hash a variable-length key into a 32-bit value
- hash(byte[], int, int) -
Method in class org.apache.hadoop.util.hash.MurmurHash
-
- HASH_COUNT -
Static variable in class org.apache.hadoop.io.BloomMapFile
-
- hashBytes(byte[], int, int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Compute hash for binary data.
- hashBytes(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Compute hash for binary data.
- hashCode() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- hashCode() -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- hashCode() -
Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- hashCode() -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
- hashCode() -
Method in class org.apache.hadoop.fs.FileChecksum
-
- hashCode() -
Method in class org.apache.hadoop.fs.FileStatus
- Returns a hash code value for the object, which is defined as
the hash code of the path name.
- hashCode() -
Method in class org.apache.hadoop.fs.Path
-
- hashCode() -
Method in class org.apache.hadoop.fs.permission.FsPermission
-
- hashCode() -
Method in class org.apache.hadoop.io.BinaryComparable
- Return a hash of the bytes returned from {#getBytes()}.
- hashCode() -
Method in class org.apache.hadoop.io.BooleanWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.BytesWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.ByteWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.DoubleWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
- hashCode() -
Method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
- hashCode() -
Method in class org.apache.hadoop.io.FloatWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.IntWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.LongWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.MD5Hash
- Returns a hash code value for this object.
- hashCode() -
Method in class org.apache.hadoop.io.NullWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry
-
- hashCode() -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- hashCode() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
- hashCode() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.Text
-
- hashCode() -
Method in class org.apache.hadoop.io.UTF8
- Deprecated.
- hashCode() -
Method in class org.apache.hadoop.io.VIntWritable
-
- hashCode() -
Method in class org.apache.hadoop.io.VLongWritable
-
- hashCode() -
Method in class org.apache.hadoop.mapred.Counters.Group
-
- hashCode() -
Method in class org.apache.hadoop.mapred.Counters
-
- hashCode() -
Method in class org.apache.hadoop.mapred.join.TupleWritable
-
- hashCode() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
- hashCode(byte[], int, int, int) -
Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
-
- hashCode() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- hashCode() -
Method in class org.apache.hadoop.mapred.TaskReport
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.Counter
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.Counters
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.ID
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.JobID
-
- hashCode(byte[], int, int, int) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
- hashCode() -
Method in class org.apache.hadoop.mapreduce.TaskID
-
- hashCode() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- hashCode() -
Method in class org.apache.hadoop.metrics2.Metric
-
- hashCode() -
Method in class org.apache.hadoop.metrics2.MetricsTag
-
- hashCode() -
Method in class org.apache.hadoop.net.SocksSocketFactory
-
- hashCode() -
Method in class org.apache.hadoop.net.StandardSocketFactory
-
- hashCode() -
Method in class org.apache.hadoop.record.Buffer
-
- hashCode() -
Method in class org.apache.hadoop.record.meta.FieldTypeInfo
- We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key
- hashCode() -
Method in class org.apache.hadoop.record.meta.MapTypeID
- We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key
- hashCode() -
Method in class org.apache.hadoop.record.meta.TypeID
- We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key
- hashCode() -
Method in class org.apache.hadoop.record.meta.VectorTypeID
- We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key
- hashCode() -
Method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL.Token
- Return the hashcode for the token.
- hashCode() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- hashCode() -
Method in class org.apache.hadoop.security.token.Token
-
- hashCode() -
Method in class org.apache.hadoop.security.UserGroupInformation
- Return the hash of the subject.
- hashCode() -
Method in class org.apache.hadoop.util.bloom.Key
-
- hashFromString(String, SecretKey) -
Static method in class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
- Aux util to calculate hash of a String
- HashFunction - Class in org.apache.hadoop.util.bloom
- Implements a hash object that returns a certain number of hashed values.
- HashFunction(int, int, int) -
Constructor for class org.apache.hadoop.util.bloom.HashFunction
- Constructor.
- HashingDistributionPolicy - Class in org.apache.hadoop.contrib.index.example
- Choose a shard for each insert or delete based on document id hashing.
- HashingDistributionPolicy() -
Constructor for class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
-
- HashPartitioner<K2,V2> - Class in org.apache.hadoop.mapred.lib
- Partition keys by their
Object.hashCode()
. - HashPartitioner() -
Constructor for class org.apache.hadoop.mapred.lib.HashPartitioner
-
- HashPartitioner<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Partition keys by their
Object.hashCode()
. - HashPartitioner() -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner
-
- hashType -
Variable in class org.apache.hadoop.util.bloom.Filter
- Type of hashing function to use.
- hasKerberosCredentials() -
Method in class org.apache.hadoop.security.UserGroupInformation
- checks if logged in using kerberos
- hasNext() -
Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- hasNext() -
Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- hasNext() -
Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
- Returns true if the stream is not empty, but provides no guarantee that
a call to next(K,V) will succeed.
- hasNext() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Return true if it is possible that this could emit more values.
- hasNext() -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- hasNext() -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- hasNext() -
Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- hasNext() -
Method in interface org.apache.hadoop.mapred.join.ResetableIterator
- True if a call to next may return a value.
- hasNext() -
Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- hasNext() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Return true if the RR- including the k,v pair stored in this object-
is exhausted.
- hasNext() -
Method in class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
-
- hasNext() -
Method in class org.apache.hadoop.metrics2.util.TryIterator
-
- hasRecovered() -
Method in class org.apache.hadoop.mapred.JobTracker
- Whether the JT has recovered upon restart
- hasRestarted() -
Method in class org.apache.hadoop.mapred.JobTracker
- Whether the JT has restarted
- hasSimpleInputSpecs_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- hasWildcard() -
Method in class org.apache.hadoop.fs.GlobPattern
-
- hbMakeCodeLengths(char[], int[], int, int) -
Static method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This method is accessible by subclasses for historical purposes.
- HDFSMerger - Class in org.apache.hadoop.contrib.failmon
-
- HDFSMerger() -
Constructor for class org.apache.hadoop.contrib.failmon.HDFSMerger
-
- HDFSMonitorThread - Class in org.apache.hadoop.mapred
-
- HDFSMonitorThread(Configuration, JobTracker, FileSystem) -
Constructor for class org.apache.hadoop.mapred.HDFSMonitorThread
-
- HEADER -
Static variable in class org.apache.hadoop.ipc.Server
- The first four bytes of Hadoop RPC connections
- HEADER_LEN -
Static variable in class org.apache.hadoop.util.DataChecksum
-
- headMap(WritableComparable) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- HeapSort - Class in org.apache.hadoop.util
- An implementation of the core algorithm of HeapSort.
- HeapSort() -
Constructor for class org.apache.hadoop.util.HeapSort
-
- heartbeat(TaskTrackerStatus, boolean, boolean, boolean, short) -
Method in class org.apache.hadoop.mapred.JobTracker
- The periodic heartbeat mechanism between the
TaskTracker
and
the JobTracker
.
- height -
Variable in class org.apache.hadoop.examples.dancing.Pentomino
-
- height -
Static variable in class org.apache.hadoop.mapred.TaskGraphServlet
- height of the graph w/o margins
- hexchars -
Static variable in class org.apache.hadoop.record.Utils
-
- hexStringToByte(String) -
Static method in class org.apache.hadoop.util.StringUtils
- Given a hexstring this will return the byte array corresponding to the
string
- HOSTNAME_PATTERN -
Static variable in class org.apache.hadoop.security.SecurityUtil
-
- HostsFileReader - Class in org.apache.hadoop.util
-
- HostsFileReader(String, String) -
Constructor for class org.apache.hadoop.util.HostsFileReader
-
- HTML_TAIL -
Static variable in class org.apache.hadoop.util.ServletUtil
-
- htmlFooter() -
Static method in class org.apache.hadoop.util.ServletUtil
- HTML footer to be added in the jsps.
- HtmlQuoting - Class in org.apache.hadoop.http
- This class is responsible for quoting HTML characters.
- HtmlQuoting() -
Constructor for class org.apache.hadoop.http.HtmlQuoting
-
- HTTP_HEADER_REPLY_URL_HASH -
Static variable in class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
-
- HTTP_HEADER_URL_HASH -
Static variable in class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
-
- HttpServer - Class in org.apache.hadoop.http
- Create a Jetty embedded server to answer http requests.
- HttpServer(String, String, int, boolean) -
Constructor for class org.apache.hadoop.http.HttpServer
- Same as this(name, bindAddress, port, findPort, null);
- HttpServer(String, String, int, boolean, Configuration) -
Constructor for class org.apache.hadoop.http.HttpServer
-
- HttpServer(String, String, int, boolean, Configuration, Connector) -
Constructor for class org.apache.hadoop.http.HttpServer
-
- HttpServer(String, String, int, boolean, Configuration, AccessControlList) -
Constructor for class org.apache.hadoop.http.HttpServer
- Create a status server on the given port.
- HttpServer(String, String, int, boolean, Configuration, AccessControlList, Connector) -
Constructor for class org.apache.hadoop.http.HttpServer
-
- HttpServer.QuotingInputFilter - Class in org.apache.hadoop.http
- A Servlet input filter that quotes all HTML active characters in the
parameter names and values.
- HttpServer.QuotingInputFilter() -
Constructor for class org.apache.hadoop.http.HttpServer.QuotingInputFilter
-
- HttpServer.QuotingInputFilter.RequestQuoter - Class in org.apache.hadoop.http
-
- HttpServer.QuotingInputFilter.RequestQuoter(HttpServletRequest) -
Constructor for class org.apache.hadoop.http.HttpServer.QuotingInputFilter.RequestQuoter
-
- HttpServer.StackServlet - Class in org.apache.hadoop.http
- A very simple servlet to serve up a text representation of the current
stack traces.
- HttpServer.StackServlet() -
Constructor for class org.apache.hadoop.http.HttpServer.StackServlet
-
- humanReadableInt(long) -
Static method in class org.apache.hadoop.util.StringUtils
- Given an integer, return a string that is in an approximate, but human
readable format.
I
- ID - Class in org.apache.hadoop.mapred
- A general identifier, which internally stores the id
as an integer.
- ID(int) -
Constructor for class org.apache.hadoop.mapred.ID
- constructs an ID object from the given int
- ID() -
Constructor for class org.apache.hadoop.mapred.ID
-
- id() -
Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
- Return the position in the collector this class occupies.
- id() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Return the position in the collector this class occupies.
- id -
Variable in class org.apache.hadoop.mapred.join.Parser.Node
-
- id() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Return the position in the collector this class occupies.
- ID - Class in org.apache.hadoop.mapreduce
- A general identifier, which internally stores the id
as an integer.
- ID(int) -
Constructor for class org.apache.hadoop.mapreduce.ID
- constructs an ID object from the given int
- ID() -
Constructor for class org.apache.hadoop.mapreduce.ID
-
- id -
Variable in class org.apache.hadoop.mapreduce.ID
-
- ident -
Variable in class org.apache.hadoop.mapred.join.Parser.Node
-
- IDENT_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- IdentifierResolver - Class in org.apache.hadoop.streaming.io
- This class is used to resolve a string identifier into the required IO
classes.
- IdentifierResolver() -
Constructor for class org.apache.hadoop.streaming.io.IdentifierResolver
-
- IdentityLocalAnalysis - Class in org.apache.hadoop.contrib.index.example
- Identity local analysis maps inputs directly into outputs.
- IdentityLocalAnalysis() -
Constructor for class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
-
- IdentityMapper<K,V> - Class in org.apache.hadoop.mapred.lib
- Implements the identity function, mapping inputs directly to outputs.
- IdentityMapper() -
Constructor for class org.apache.hadoop.mapred.lib.IdentityMapper
-
- IdentityReducer<K,V> - Class in org.apache.hadoop.mapred.lib
- Performs no reduction, writing all input values directly to the output.
- IdentityReducer() -
Constructor for class org.apache.hadoop.mapred.lib.IdentityReducer
-
- idFormat -
Static variable in class org.apache.hadoop.mapreduce.JobID
-
- idFormat -
Static variable in class org.apache.hadoop.mapreduce.TaskID
-
- IDistributionPolicy - Interface in org.apache.hadoop.contrib.index.mapred
- A distribution policy decides, given a document with a document id, which
one shard the request should be sent to if the request is an insert, and
which shard(s) the request should be sent to if the request is a delete.
- idWithinJob() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- idx -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- idx -
Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- ifExists(String, Configuration) -
Method in class org.apache.hadoop.fs.LocalDirAllocator
- We search through all the configured dirs for the file's existence
and return true when we find
- ifmt(double) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- IIndexUpdater - Interface in org.apache.hadoop.contrib.index.mapred
- A class implements an index updater interface should create a Map/Reduce job
configuration and run the Map/Reduce job to analyze documents and update
Lucene instances in parallel.
- ILocalAnalysis<K extends WritableComparable,V extends Writable> - Interface in org.apache.hadoop.contrib.index.mapred
- Application specific local analysis.
- image -
Variable in class org.apache.hadoop.record.compiler.generated.Token
- The string image of the token.
- implies(FsAction) -
Method in enum org.apache.hadoop.fs.permission.FsAction
- Return true if this action implies that action.
- in -
Variable in class org.apache.hadoop.io.compress.CompressionInputStream
- The input stream to be compressed.
- inBuf -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- inc(int) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
- Deprecated. Inc metrics for incr vlaue
- inc() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
- Deprecated. Inc metrics by one
- inc(long) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
- Deprecated. Inc metrics for incr vlaue
- inc() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
- Deprecated. Inc metrics by one
- inc(int, long) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Increment the metrics for numOps operations
- inc(long) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Increment the metrics for one operation
- incDfsUsed(long) -
Method in class org.apache.hadoop.fs.DU
- Increase how much disk space we use.
- Include() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- INCLUDE_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounter
- Increment the metric value by 1.
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterInt
-
- incr(int) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterInt
- Increment the value by a delta
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterLong
-
- incr(long) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterLong
- Increment the value by a delta
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGauge
- Increment the value of the metric by 1
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeInt
-
- incr(int) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeInt
- Increment by delta
- incr() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeLong
-
- incr(long) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeLong
- Increment by delta
- incr(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Increment a metric by name.
- incr(String, MetricMutableFactory) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Increment a metric by name.
- incr() -
Method in interface org.apache.hadoop.record.Index
-
- incrAllCounters(Counters) -
Method in class org.apache.hadoop.mapred.Counters
- Increments multiple counters by their amounts in another Counters
instance.
- incrAllCounters(CounterGroup) -
Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- incrAllCounters(Counters) -
Method in class org.apache.hadoop.mapreduce.Counters
- Increments multiple counters by their amounts in another Counters
instance.
- incrAuthenticationFailures() -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- One authentication failure event
- incrAuthenticationSuccesses() -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- One authentication success event
- incrAuthorizationFailures() -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- One authorization failure event
- incrAuthorizationSuccesses() -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- One authorization success event
- incrCounter(Enum, long) -
Method in class org.apache.hadoop.mapred.Counters
- Increments the specified counter by the specified amount, creating it if
it didn't already exist.
- incrCounter(String, String, long) -
Method in class org.apache.hadoop.mapred.Counters
- Increments the specified counter by the specified amount, creating it if
it didn't already exist.
- incrCounter(Enum<?>, long) -
Method in interface org.apache.hadoop.mapred.Reporter
- Increments the counter identified by the key, which can be of
any
Enum
type, by the specified amount.
- incrCounter(String, String, long) -
Method in interface org.apache.hadoop.mapred.Reporter
- Increments the counter identified by the group and counter name
by the specified amount.
- incrCounter(Enum, long) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- incrCounter(String, String, long) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- increment(long) -
Method in class org.apache.hadoop.mapreduce.Counter
- Increment this counter by the given value
- INCREMENT -
Static variable in class org.apache.hadoop.metrics.spi.MetricValue
- Deprecated.
- incrementBytesRead(long) -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Increment the bytes read in the statistics
- incrementBytesWritten(long) -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Increment the bytes written in the statistics
- incrementLargeReadOps(int) -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Increment the number of large read operations
- incrementReadOps(int) -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Increment the number of read operations
- incrementWeight(double) -
Method in class org.apache.hadoop.util.bloom.Key
- Increments the weight of this key with a specified value.
- incrementWeight() -
Method in class org.apache.hadoop.util.bloom.Key
- Increments the weight of this key by one.
- incrementWriteOps(int) -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Increment the number of write operations
- incrMetric(String, int) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, long) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, short) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, byte) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, float) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, int) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, long) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, short) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, byte) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Increments the named metric by the specified value.
- incrMetric(String, float) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Increments the named metric by the specified value.
- incrReceivedBytes(int) -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- Increment received bytes by count
- incrSentBytes(int) -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- Increment sent bytes by count
- Index - Interface in org.apache.hadoop.record
- Interface that acts as an iterator for deserializing maps.
- INDEX_FILE_NAME -
Static variable in class org.apache.hadoop.io.MapFile
- The name of the index file.
- IndexedSortable - Interface in org.apache.hadoop.util
- Interface for collections capable of being sorted by
IndexedSorter
algorithms. - IndexedSorter - Interface in org.apache.hadoop.util
- Interface for sort algorithms accepting
IndexedSortable
items. - IndexUpdateCombiner - Class in org.apache.hadoop.contrib.index.mapred
- This combiner combines multiple intermediate forms into one intermediate
form.
- IndexUpdateCombiner() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
-
- IndexUpdateConfiguration - Class in org.apache.hadoop.contrib.index.mapred
- This class provides the getters and the setters to a number of parameters.
- IndexUpdateConfiguration(Configuration) -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Constructor
- IndexUpdateMapper<K extends WritableComparable,V extends Writable> - Class in org.apache.hadoop.contrib.index.mapred
- This class applies local analysis on a key-value pair and then convert the
result docid-operation pair to a shard-and-intermediate form pair.
- IndexUpdateMapper() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
-
- IndexUpdateOutputFormat - Class in org.apache.hadoop.contrib.index.mapred
- The record writer of this output format simply puts a message in an output
path when a shard update is done.
- IndexUpdateOutputFormat() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateOutputFormat
-
- IndexUpdatePartitioner - Class in org.apache.hadoop.contrib.index.mapred
- This partitioner class puts the values of the same key - in this case the
same shard - in the same partition.
- IndexUpdatePartitioner() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
-
- IndexUpdater - Class in org.apache.hadoop.contrib.index.mapred
- An implementation of an index updater interface which creates a Map/Reduce
job configuration and run the Map/Reduce job to analyze documents and update
Lucene instances in parallel.
- IndexUpdater() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdater
-
- IndexUpdateReducer - Class in org.apache.hadoop.contrib.index.mapred
- This reducer applies to a shard the changes for it.
- IndexUpdateReducer() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
- init(Shard[]) -
Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
-
- init(Shard[]) -
Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
-
- init(Shard[]) -
Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
- Initialization.
- init() -
Method in class org.apache.hadoop.fs.FsShell
-
- init(FilterConfig) -
Method in class org.apache.hadoop.http.HttpServer.QuotingInputFilter
-
- init(FilterConfig) -
Method in class org.apache.hadoop.http.lib.StaticUserWebFilter.StaticUserFilter
-
- init() -
Method in class org.apache.hadoop.jmx.JMXJsonServlet
- Initialize this servlet.
- init(JobConf) -
Method in class org.apache.hadoop.mapred.JobClient
- Connect to the default
JobTracker
.
- init(JobTracker, JobConf, String, long) -
Static method in class org.apache.hadoop.mapred.JobHistory
- Initialize JobHistory files.
- init(ServletConfig) -
Method in class org.apache.hadoop.mapred.RawHistoryFileServlet
-
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.file.FileContext
- Deprecated.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
- Deprecated.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext31
-
- init(String, String) -
Static method in class org.apache.hadoop.metrics.jvm.JvmMetrics
- Deprecated.
- init(String, String, String) -
Static method in class org.apache.hadoop.metrics.jvm.JvmMetrics
- Deprecated.
- init(String, ContextFactory) -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Initialize this context.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Initializes the context.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.spi.NoEmitMetricsContext
- Deprecated.
- init(String, ContextFactory) -
Method in class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
- Deprecated.
- init(String) -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
- Initialized the metrics system with a prefix.
- init(SubsetConfiguration) -
Method in class org.apache.hadoop.metrics2.MetricsFilter
-
- init(SubsetConfiguration) -
Method in interface org.apache.hadoop.metrics2.MetricsPlugin
- Initialize the plugin
- init(SubsetConfiguration) -
Method in class org.apache.hadoop.metrics2.sink.FileSink
-
- init(SubsetConfiguration) -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- init(FilterConfig) -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
- Initializes the authentication filter.
- init(Properties) -
Method in interface org.apache.hadoop.security.authentication.server.AuthenticationHandler
- Initializes the authentication handler instance.
- init(Properties) -
Method in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- Initializes the authentication handler instance.
- init(Properties) -
Method in class org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler
- Initializes the authentication handler instance.
- init(FilterConfig) -
Method in class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector.Krb5SslFilter
-
- init(Configuration) -
Static method in class org.apache.hadoop.security.SaslRpcServer
-
- init() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- init() -
Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- inited() -
Method in class org.apache.hadoop.mapred.JobInProgress
- Check if the job has been initialized.
- initFilter(FilterContainer, Configuration) -
Method in class org.apache.hadoop.http.FilterInitializer
- Initialize a Filter to a FilterContainer.
- initFilter(FilterContainer, Configuration) -
Method in class org.apache.hadoop.http.lib.StaticUserWebFilter
-
- initFilter(FilterContainer, Configuration) -
Method in class org.apache.hadoop.security.AuthenticationFilterInitializer
- Initializes Alfredo AuthenticationFilter.
- initHTML(ServletResponse, String) -
Static method in class org.apache.hadoop.util.ServletUtil
- Initial HTML header
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.FileSystem
- Called after a new FileSystem instance is constructed.
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Called after a new FileSystem instance is constructed.
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.HarFileSystem
- Initialize a Har filesystem per har archive.
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- initialize(URI, Configuration) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- initialize(URI) -
Method in class org.apache.hadoop.fs.s3.MigrationTool
-
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.s3.S3Credentials
-
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- initialize(URI, Configuration) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- initialize(JobConf, JobID, Reporter, boolean) -
Method in class org.apache.hadoop.mapred.Task
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- initialize(InputSplit, TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.RecordReader
- Called once at initialization.
- initialize(String) -
Static method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
- Common static convenience method to initialize the metrics system
- initialize(Subject, CallbackHandler, Map<String, ?>, Map<String, ?>) -
Method in class org.apache.hadoop.security.UserGroupInformation.HadoopLoginModule
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.InputWriter
- Initializes the InputWriter.
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.OutputReader
- Initializes the OutputReader.
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.TextInputWriter
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.TextOutputReader
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
-
- initialize(PipeMapRed) -
Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
-
- initialize(int) -
Method in class org.apache.hadoop.util.PriorityQueue
- Subclass constructors must call this.
- initializeAttemptDirs(String, String, String) -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.Localizer
- Create taskDirs on all the disks.
- initializeJob(String, String, Path, Path, TaskUmbilicalProtocol, InetSocketAddress) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
- This routine initializes the local file system for running a job.
- initializeJob(String, String, Path, Path, TaskUmbilicalProtocol, InetSocketAddress) -
Method in class org.apache.hadoop.mapred.TaskController
- Create all of the directories necessary for the job to start and download
all of the job and private distributed cache files.
- initializeJobLogDir() -
Method in class org.apache.hadoop.mapred.JobLocalizer
- Create job log directory and set appropriate permissions for the directory.
- initializePieces() -
Method in class org.apache.hadoop.examples.dancing.OneSidedPentomino
- Define the one sided pieces.
- initializePieces() -
Method in class org.apache.hadoop.examples.dancing.Pentomino
- Fill in the pieces list.
- initializeUserDirs(String) -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.Localizer
- Initialize the local directories for a particular user on this TT.
- initJob(JobInProgress) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- initNextRecordReader() -
Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
- Get the record reader for the next chunk in this CombineFileSplit.
- initNextRecordReader() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
- Get the record reader for the next chunk in this CombineFileSplit.
- initTasks() -
Method in class org.apache.hadoop.mapred.JobInProgress
- Construct the splits, etc.
- injectToken(HttpURLConnection, AuthenticatedURL.Token) -
Static method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL
- Helper method that injects an authentication token to send with a connection.
- InMemoryFileSystem - Class in org.apache.hadoop.fs
- Deprecated.
- InMemoryFileSystem() -
Constructor for class org.apache.hadoop.fs.InMemoryFileSystem
- Deprecated.
- InMemoryFileSystem(URI, Configuration) -
Constructor for class org.apache.hadoop.fs.InMemoryFileSystem
- Deprecated.
- InnerJoinRecordReader<K extends WritableComparable> - Class in org.apache.hadoop.mapred.join
- Full inner join.
- INode - Class in org.apache.hadoop.fs.s3
- Holds file metadata including type (regular file, or directory),
and the list of blocks that are pointers to the data.
- INode(INode.FileType, Block[]) -
Constructor for class org.apache.hadoop.fs.s3.INode
-
- inodeExists(Path) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- Input() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- INPUT_BOUNDING_QUERY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Input query to get the max and min values of the jdbc.input.query
- INPUT_CLASS_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Class name implementing DBWritable which will hold input tuples
- INPUT_CLASS_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Class name implementing DBWritable which will hold input tuples
- INPUT_CONDITIONS_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- WHERE clause in the input SELECT statement
- INPUT_CONDITIONS_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- WHERE clause in the input SELECT statement
- INPUT_COUNT_QUERY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Input query to get the count of records
- INPUT_COUNT_QUERY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Input query to get the count of records
- INPUT_FIELD_NAMES_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Field names in the Input table
- INPUT_FIELD_NAMES_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Field names in the Input table
- INPUT_FORMAT_CLASS_ATTR -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- INPUT_ORDER_BY_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- ORDER BY clause in the input SELECT statement
- INPUT_ORDER_BY_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- ORDER BY clause in the input SELECT statement
- INPUT_QUERY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Whole input query, exluding LIMIT...OFFSET
- INPUT_QUERY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Whole input query, exluding LIMIT...OFFSET
- input_stream -
Variable in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- INPUT_TABLE_NAME_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Input table name
- INPUT_TABLE_NAME_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Input table name
- InputBuffer - Class in org.apache.hadoop.io
- A reusable
InputStream
implementation that reads from an in-memory
buffer. - InputBuffer() -
Constructor for class org.apache.hadoop.io.InputBuffer
- Constructs a new empty buffer.
- inputCounter -
Variable in class org.apache.hadoop.mapred.Task.CombinerRunner
-
- inputFile -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- inputFile -
Variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- InputFormat<K,V> - Interface in org.apache.hadoop.mapred
InputFormat
describes the input-specification for a
Map-Reduce job.- InputFormat<K,V> - Class in org.apache.hadoop.mapreduce
InputFormat
describes the input-specification for a
Map-Reduce job.- InputFormat() -
Constructor for class org.apache.hadoop.mapreduce.InputFormat
-
- inputFormatSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- InputSampler<K,V> - Class in org.apache.hadoop.mapred.lib
- Utility for collecting samples and writing a partition file for
TotalOrderPartitioner
. - InputSampler(JobConf) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler
-
- InputSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Utility for collecting samples and writing a partition file for
TotalOrderPartitioner
. - InputSampler(Configuration) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler
-
- InputSampler.IntervalSampler<K,V> - Class in org.apache.hadoop.mapred.lib
- Sample from s splits at regular intervals.
- InputSampler.IntervalSampler(double) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler
- Create a new IntervalSampler sampling all splits.
- InputSampler.IntervalSampler(double, int) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler
- Create a new IntervalSampler.
- InputSampler.IntervalSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Sample from s splits at regular intervals.
- InputSampler.IntervalSampler(double) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
- Create a new IntervalSampler sampling all splits.
- InputSampler.IntervalSampler(double, int) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
- Create a new IntervalSampler.
- InputSampler.RandomSampler<K,V> - Class in org.apache.hadoop.mapred.lib
- Sample from random points in the input.
- InputSampler.RandomSampler(double, int) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
- Create a new RandomSampler sampling all splits.
- InputSampler.RandomSampler(double, int, int) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
- Create a new RandomSampler.
- InputSampler.RandomSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Sample from random points in the input.
- InputSampler.RandomSampler(double, int) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
- Create a new RandomSampler sampling all splits.
- InputSampler.RandomSampler(double, int, int) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
- Create a new RandomSampler.
- InputSampler.Sampler<K,V> - Interface in org.apache.hadoop.mapred.lib
- Interface to sample using an
InputFormat
. - InputSampler.Sampler<K,V> - Interface in org.apache.hadoop.mapreduce.lib.partition
- Interface to sample using an
InputFormat
. - InputSampler.SplitSampler<K,V> - Class in org.apache.hadoop.mapred.lib
- Samples the first n records from s splits.
- InputSampler.SplitSampler(int) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
- Create a SplitSampler sampling all splits.
- InputSampler.SplitSampler(int, int) -
Constructor for class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
- Create a new SplitSampler.
- InputSampler.SplitSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Samples the first n records from s splits.
- InputSampler.SplitSampler(int) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
- Create a SplitSampler sampling all splits.
- InputSampler.SplitSampler(int, int) -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
- Create a new SplitSampler.
- inputSpecs_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- InputSplit - Interface in org.apache.hadoop.mapred
InputSplit
represents the data to be processed by an
individual Mapper
.- InputSplit - Class in org.apache.hadoop.mapreduce
InputSplit
represents the data to be processed by an
individual Mapper
.- InputSplit() -
Constructor for class org.apache.hadoop.mapreduce.InputSplit
-
- inputStream -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- inputTag -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- InputWriter<K,V> - Class in org.apache.hadoop.streaming.io
- Abstract base for classes that write the client's input.
- InputWriter() -
Constructor for class org.apache.hadoop.streaming.io.InputWriter
-
- inReaderSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- insert(EventRecord) -
Method in class org.apache.hadoop.contrib.failmon.LocalStore
- Insert an EventRecord to the local storage, after it
gets serialized and anonymized.
- insert(EventRecord[]) -
Method in class org.apache.hadoop.contrib.failmon.LocalStore
- Insert an array of EventRecords to the local storage, after they
get serialized and anonymized.
- INSERT -
Static variable in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
-
- insert(T) -
Method in class org.apache.hadoop.util.PriorityQueue
- Adds element to the PriorityQueue in log(size) time if either
the PriorityQueue is not full, or not lessThan(element, top()).
- instances -
Static variable in class org.apache.hadoop.contrib.failmon.Executor
-
- INT -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- INT_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- IntegerSplitter - Class in org.apache.hadoop.mapreduce.lib.db
- Implement DBSplitter over integer values.
- IntegerSplitter() -
Constructor for class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter
-
- InterfaceAudience - Class in org.apache.hadoop.classification
- Annotation to inform users of a package, class or method's intended audience.
- InterfaceAudience.LimitedPrivate - Annotation Type in org.apache.hadoop.classification
- Intended only for the project(s) specified in the annotation.
- InterfaceAudience.Private - Annotation Type in org.apache.hadoop.classification
- Intended for use only within Hadoop itself.
- InterfaceAudience.Public - Annotation Type in org.apache.hadoop.classification
- Intended for use by any project or application.
- InterfaceStability - Class in org.apache.hadoop.classification
- Annotation to inform users of how much to rely on a particular package,
class or method not changing over time.
- InterfaceStability() -
Constructor for class org.apache.hadoop.classification.InterfaceStability
-
- InterfaceStability.Evolving - Annotation Type in org.apache.hadoop.classification
- Evolving, but can break compatibility at minor release (i.e.
- InterfaceStability.Stable - Annotation Type in org.apache.hadoop.classification
- Can evolve while retaining compatibility for minor release boundaries.;
can break compatibility only at major release (ie.
- InterfaceStability.Unstable - Annotation Type in org.apache.hadoop.classification
- No guarantee is provided as to reliability or stability across any
level of release granularity.
- IntermediateForm - Class in org.apache.hadoop.contrib.index.mapred
- An intermediate form for one or more parsed Lucene documents and/or
delete terms.
- IntermediateForm() -
Constructor for class org.apache.hadoop.contrib.index.mapred.IntermediateForm
- Constructor
- IntSumReducer<Key> - Class in org.apache.hadoop.mapreduce.lib.reduce
-
- IntSumReducer() -
Constructor for class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer
-
- IntTypeID -
Static variable in class org.apache.hadoop.record.meta.TypeID
-
- IntWritable - Class in org.apache.hadoop.io
- A WritableComparable for ints.
- IntWritable() -
Constructor for class org.apache.hadoop.io.IntWritable
-
- IntWritable(int) -
Constructor for class org.apache.hadoop.io.IntWritable
-
- IntWritable.Comparator - Class in org.apache.hadoop.io
- A Comparator optimized for IntWritable.
- IntWritable.Comparator() -
Constructor for class org.apache.hadoop.io.IntWritable.Comparator
-
- INVALID_HASH -
Static variable in class org.apache.hadoop.util.hash.Hash
- Constant to denote invalid hash type.
- InvalidFileTypeException - Exception in org.apache.hadoop.mapred
- Used when file type differs from the desired file type.
- InvalidFileTypeException() -
Constructor for exception org.apache.hadoop.mapred.InvalidFileTypeException
-
- InvalidFileTypeException(String) -
Constructor for exception org.apache.hadoop.mapred.InvalidFileTypeException
-
- InvalidInputException - Exception in org.apache.hadoop.mapred
- This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.
- InvalidInputException(List<IOException>) -
Constructor for exception org.apache.hadoop.mapred.InvalidInputException
- Create the exception with the given list.
- InvalidInputException - Exception in org.apache.hadoop.mapreduce.lib.input
- This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.
- InvalidInputException(List<IOException>) -
Constructor for exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
- Create the exception with the given list.
- InvalidJobConfException - Exception in org.apache.hadoop.mapred
- This exception is thrown when jobconf misses some mendatory attributes
or value of some attributes is invalid.
- InvalidJobConfException() -
Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
-
- InvalidJobConfException(String) -
Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
-
- InvalidJobConfException(String, Throwable) -
Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
-
- InvalidJobConfException(Throwable) -
Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
-
- InverseMapper<K,V> - Class in org.apache.hadoop.mapred.lib
- A
Mapper
that swaps keys and values. - InverseMapper() -
Constructor for class org.apache.hadoop.mapred.lib.InverseMapper
-
- InverseMapper<K,V> - Class in org.apache.hadoop.mapreduce.lib.map
- A
Mapper
that swaps keys and values. - InverseMapper() -
Constructor for class org.apache.hadoop.mapreduce.lib.map.InverseMapper
-
- invoke(String, Object[], String[]) -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
- Deprecated.
- IO_COMPRESSION_CODEC_SNAPPY_BUFFERSIZE_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- Default value for IO_COMPRESSION_CODEC_SNAPPY_BUFFERSIZE_KEY
- IO_COMPRESSION_CODEC_SNAPPY_BUFFERSIZE_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- Internal buffer size for Snappy compressor/decompressors
- IO_NATIVE_LIB_AVAILABLE_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
- Default value for IO_NATIVE_LIB_AVAILABLE_KEY
- IO_NATIVE_LIB_AVAILABLE_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- ioSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- IOUtils - Class in org.apache.hadoop.io
- An utility class for I/O related functionality.
- IOUtils() -
Constructor for class org.apache.hadoop.io.IOUtils
-
- IOUtils.NullOutputStream - Class in org.apache.hadoop.io
- /dev/null of OutputStreams.
- IOUtils.NullOutputStream() -
Constructor for class org.apache.hadoop.io.IOUtils.NullOutputStream
-
- IPC_CLIENT_CONNECT_MAX_RETRIES_DEFAULT -
Static variable in class org.apache.hadoop.ipc.Client
-
- IPC_CLIENT_CONNECT_MAX_RETRIES_KEY -
Static variable in class org.apache.hadoop.ipc.Client
-
- IPC_CLIENT_FALLBACK_TO_SIMPLE_AUTH_ALLOWED_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- IPC_CLIENT_FALLBACK_TO_SIMPLE_AUTH_ALLOWED_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- IPC_SERVER_RPC_READ_THREADS_DEFAULT -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- IPC_SERVER_RPC_READ_THREADS_KEY -
Static variable in class org.apache.hadoop.fs.CommonConfigurationKeys
-
- isAbsolute() -
Method in class org.apache.hadoop.fs.Path
- True if the directory of this path is absolute.
- isAbsolute() -
Method in class org.apache.hadoop.metrics.spi.MetricValue
- Deprecated.
- isAlive(String) -
Static method in class org.apache.hadoop.util.ProcessTree
- Is the process with PID pid still alive?
This method assumes that isAlive is called on a pid that was alive not
too long ago, and hence assumes no chance of pid-wrapping-around.
- isAlive() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
- Is the root-process alive?
- isAllAllowed() -
Method in class org.apache.hadoop.security.authorize.AccessControlList
-
- isAnyProcessInTreeAlive() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
- Is any of the subprocesses in the process-tree alive?
- isAvailable() -
Static method in class org.apache.hadoop.io.compress.snappy.LoadSnappy
- Returns if Snappy native library is loaded.
- isAvailable() -
Static method in class org.apache.hadoop.io.nativeio.NativeIO
- Return true if the JNI-based native IO extensions are available.
- isAvailable() -
Static method in class org.apache.hadoop.util.ProcfsBasedProcessTree
- Checks if the ProcfsBasedProcessTree is available on this system.
- isBlacklisted(String) -
Method in class org.apache.hadoop.mapred.JobTracker
- Whether the tracker is blacklisted or not
- isBlockCompressed() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Returns true if records are block-compressed.
- isCached(String) -
Method in class org.apache.hadoop.security.NetgroupCache
-
- isChecksumFile(Path) -
Static method in class org.apache.hadoop.fs.ChecksumFileSystem
- Return true iff file is a checksum file name.
- isComplete() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Check if the job is finished or not.
- isComplete() -
Method in class org.apache.hadoop.mapreduce.Job
- Check if the job is finished or not.
- isCompleted() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- isCompleted() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
-
- isCompressed() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Returns true if values are compressed.
- isContextValid(String) -
Static method in class org.apache.hadoop.fs.LocalDirAllocator
- Method to check whether a context is valid
- isCygwin() -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- isDir() -
Method in class org.apache.hadoop.fs.FileStatus
- Is this a directory?
- isDirectory(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Deprecated. Use getFileStatus() instead
- isDirectory(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
- Deprecated.
- isDirectory() -
Method in class org.apache.hadoop.fs.s3.INode
-
- isEmpty() -
Method in class org.apache.hadoop.io.MapWritable
-
- isEmpty() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- isExpired() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
- Returns if the token has expired.
- isFile(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- True iff the named path is a regular file.
- isFile(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
- Deprecated.
- isFile() -
Method in class org.apache.hadoop.fs.s3.INode
-
- isFile(Path) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- isFromKeytab() -
Method in class org.apache.hadoop.security.UserGroupInformation
- Is this user logged in from a keytab file?
- isGraylisted(String) -
Method in class org.apache.hadoop.mapred.JobTracker
- Whether the tracker is graylisted or not
- isHealthy() -
Method in class org.apache.hadoop.mapred.TaskTracker
-
- isHealthy() -
Method in interface org.apache.hadoop.mapred.TaskTrackerMXBean
-
- isIdle() -
Method in class org.apache.hadoop.mapred.TaskTracker
- Is this task tracker idle?
- isIdleAndClean() -
Method in class org.apache.hadoop.mapred.TaskTracker
- Is this task tracker idle and clean?
- isIncluded(int) -
Method in class org.apache.hadoop.conf.Configuration.IntegerRanges
- Is the given value in the set of ranges
- isIncrement() -
Method in class org.apache.hadoop.metrics.spi.MetricValue
- Deprecated.
- isInSafeMode() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- isInstrumentationAccessAllowed(ServletContext, HttpServletRequest, HttpServletResponse) -
Static method in class org.apache.hadoop.http.HttpServer
- Checks the user has privileges to access to instrumentation servlets.
- isJobComplete() -
Method in class org.apache.hadoop.mapred.JobStatus
- Returns true if the status is for a completed job.
- isJobDirValid(Path, FileSystem) -
Static method in class org.apache.hadoop.mapred.JobClient
- Checks if the job directory is clean and has all the required components
for (re) starting the job
- isLoaded() -
Static method in class org.apache.hadoop.io.compress.snappy.LoadSnappy
- Returns if Snappy native library is loaded.
- isLocalHadoop() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- isLocalJobTracker(JobConf) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- isLoginKeytabBased() -
Static method in class org.apache.hadoop.security.UserGroupInformation
- Did the login happen via keytab
- isManaged(Token<?>) -
Method in class org.apache.hadoop.mapred.JobClient.Renewer
-
- isManaged() -
Method in class org.apache.hadoop.security.token.Token
- Is this token managed so that it can be renewed or cancelled?
- isManaged(Token<?>) -
Method in class org.apache.hadoop.security.token.Token.TrivialRenewer
-
- isManaged(Token<?>) -
Method in class org.apache.hadoop.security.token.TokenRenewer
- Is the given token managed? Only managed tokens may be renewed or
cancelled.
- isMap() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
- Returns whether this TaskAttemptID is a map ID
- isMap() -
Method in class org.apache.hadoop.mapreduce.TaskID
- Returns whether this TaskID is a map ID
- isMapTask() -
Method in class org.apache.hadoop.mapred.Task
-
- isMapTask() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- isMonitoring() -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Returns true if monitoring is currently in progress.
- isMonitoring() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Returns true if monitoring is currently in progress.
- isMonitoring() -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated. Return true if all subcontexts are monitoring.
- isMultiNamedOutput(JobConf, String) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
- Returns if a named output is multiple.
- isNativeCodeLoaded() -
Static method in class org.apache.hadoop.util.NativeCodeLoader
- Check if native-hadoop code is loaded for this platform.
- isNativeSnappyLoaded(Configuration) -
Static method in class org.apache.hadoop.io.compress.SnappyCodec
- Are the native snappy libraries loaded & initialized?
- isNativeZlibLoaded(Configuration) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
- Check if native-zlib code is loaded & initialized correctly and
can be loaded for this job.
- isNegativeVInt(byte) -
Static method in class org.apache.hadoop.io.WritableUtils
- Given the first byte of a vint/vlong, determine the sign
- isNodeGroupAware() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- isNodeGroupAware() -
Method in class org.apache.hadoop.net.NetworkTopology
- Check if network topology is aware of NodeGroup
- isNodeGroupAware() -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
- Check if network topology is aware of NodeGroup
- IsolationRunner - Class in org.apache.hadoop.mapred
- IsolationRunner is intended to facilitate debugging by re-running a specific
task, given left-over task files for a (typically failed) past job.
- IsolationRunner() -
Constructor for class org.apache.hadoop.mapred.IsolationRunner
-
- isOnSameNodeGroup(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopology
- Return false directly as not aware of NodeGroup, to be override in sub-class
- isOnSameNodeGroup(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
- Check if two nodes are on the same node group (hypervisor) The
assumption here is: each nodes are leaf nodes.
- isOnSameRack(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopology
- Check if two nodes are on the same rack
- isOnSameRack(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
-
- isOpen() -
Method in class org.apache.hadoop.net.SocketInputStream
-
- isOpen() -
Method in class org.apache.hadoop.net.SocketOutputStream
-
- isOriginalTGT(String) -
Static method in class org.apache.hadoop.security.SecurityUtil
-
- isProcessGroupAlive(String) -
Static method in class org.apache.hadoop.util.ProcessTree
- Is the process group with still alive?
This method assumes that isAlive is called on a pid that was alive not
too long ago, and hence assumes no chance of pid-wrapping-around.
- isQueueEmpty() -
Method in class org.apache.hadoop.mapred.CleanupQueue
-
- isRandomSecret() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
- Returns if a random secret is being used.
- isReady() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- isReady() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
-
- isRunning() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
- is secretMgr running
- isSameParents(Node, Node) -
Method in class org.apache.hadoop.net.NetworkTopology
- Compare the parents of each node for equality
- isSecurityEnabled() -
Static method in class org.apache.hadoop.security.UserGroupInformation
- Determine if UserGroupInformation is using Kerberos to determine
user identities or is relying on simple authentication
- isSegmentsFile(String) -
Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
- Check if the file is a segments_N file
- isSegmentsGenFile(String) -
Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
- Check if the file is the segments.gen file
- isSet() -
Method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL.Token
- Returns if a token from the server has been set.
- isSetsidAvailable -
Static variable in class org.apache.hadoop.util.ProcessTree
-
- isSkipping() -
Method in class org.apache.hadoop.mapred.Task
- Is Task in skipping mode.
- isSorted() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
- Is the TFile sorted?
- isSplitable(FileSystem, Path) -
Method in class org.apache.hadoop.mapred.FileInputFormat
- Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
- isSplitable(FileSystem, Path) -
Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
-
- isSplitable(FileSystem, Path) -
Method in class org.apache.hadoop.mapred.TextInputFormat
-
- isSplitable(JobContext, Path) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
-
- isSplitable(JobContext, Path) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
- isSplitable(JobContext, Path) -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat
-
- isSplitable(JobContext, Path) -
Method in class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
-
- isSuccessful() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Check if the job completed successfully.
- isSuccessful() -
Method in class org.apache.hadoop.mapreduce.Job
- Check if the job completed successfully.
- isSupportSparseMetrics() -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- isTaskMemoryManagerEnabled() -
Method in class org.apache.hadoop.mapred.TaskTracker
- Is the TaskMemoryManager Enabled on this system?
- isTimedOut() -
Method in class org.apache.hadoop.util.Shell
- To check if the passed script to shell command executor timed out or
not.
- isUserAllowed(UserGroupInformation) -
Method in class org.apache.hadoop.security.authorize.AccessControlList
-
- isValid() -
Method in class org.apache.hadoop.contrib.failmon.EventRecord
- Check if the EventRecord is a valid one, i.e., whether
it represents meaningful metric values.
- isValid() -
Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
- Check if the SerializedRecord is a valid one, i.e., whether
it represents meaningful metric values.
- isValidSubnet(String) -
Static method in class org.apache.hadoop.net.NetUtils
-
- isValueLengthKnown() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
- Check whether it is safe to call getValueLength().
- iterator() -
Method in class org.apache.hadoop.conf.Configuration
- Get an
Iterator
to go through the list of String
key-value pairs in the configuration.
- iterator() -
Method in class org.apache.hadoop.mapred.Counters.Group
-
- iterator() -
Method in class org.apache.hadoop.mapred.Counters
-
- iterator() -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- Return an iterator over the elements in this tuple.
- iterator() -
Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- iterator() -
Method in class org.apache.hadoop.mapreduce.Counters
-
- iterator() -
Method in class org.apache.hadoop.mapreduce.ReduceContext.ValueIterable
-
- iterator() -
Method in class org.apache.hadoop.util.CyclicIteration
-
J
- jar_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- JarBuilder - Class in org.apache.hadoop.streaming
- This class is the main class for generating job.jar
for Hadoop Streaming jobs.
- JarBuilder() -
Constructor for class org.apache.hadoop.streaming.JarBuilder
-
- JavaSerialization - Class in org.apache.hadoop.io.serializer
-
An experimental
Serialization
for Java Serializable
classes. - JavaSerialization() -
Constructor for class org.apache.hadoop.io.serializer.JavaSerialization
-
- JavaSerializationComparator<T extends Serializable & Comparable<T>> - Class in org.apache.hadoop.io.serializer
-
A
RawComparator
that uses a JavaSerialization
Deserializer
to deserialize objects that are then compared via
their Comparable
interfaces. - JavaSerializationComparator() -
Constructor for class org.apache.hadoop.io.serializer.JavaSerializationComparator
-
- JBoolean - Class in org.apache.hadoop.record.compiler
-
- JBoolean() -
Constructor for class org.apache.hadoop.record.compiler.JBoolean
- Creates a new instance of JBoolean
- JBuffer - Class in org.apache.hadoop.record.compiler
- Code generator for "buffer" type.
- JBuffer() -
Constructor for class org.apache.hadoop.record.compiler.JBuffer
- Creates a new instance of JBuffer
- JByte - Class in org.apache.hadoop.record.compiler
- Code generator for "byte" type.
- JByte() -
Constructor for class org.apache.hadoop.record.compiler.JByte
-
- jc -
Variable in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
- jc -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- jc_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- JDouble - Class in org.apache.hadoop.record.compiler
-
- JDouble() -
Constructor for class org.apache.hadoop.record.compiler.JDouble
- Creates a new instance of JDouble
- JENKINS_HASH -
Static variable in class org.apache.hadoop.util.hash.Hash
- Constant to denote
JenkinsHash
.
- JenkinsHash - Class in org.apache.hadoop.util.hash
- Produces 32-bit hash for hash table lookup.
- JenkinsHash() -
Constructor for class org.apache.hadoop.util.hash.JenkinsHash
-
- JField<T> - Class in org.apache.hadoop.record.compiler
- A thin wrappper around record field.
- JField(String, T) -
Constructor for class org.apache.hadoop.record.compiler.JField
- Creates a new instance of JField
- JFile - Class in org.apache.hadoop.record.compiler
- Container for the Hadoop Record DDL.
- JFile(String, ArrayList<JFile>, ArrayList<JRecord>) -
Constructor for class org.apache.hadoop.record.compiler.JFile
- Creates a new instance of JFile
- JFloat - Class in org.apache.hadoop.record.compiler
-
- JFloat() -
Constructor for class org.apache.hadoop.record.compiler.JFloat
- Creates a new instance of JFloat
- JIFFY_LENGTH_IN_MILLIS -
Static variable in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- JInt - Class in org.apache.hadoop.record.compiler
- Code generator for "int" type
- JInt() -
Constructor for class org.apache.hadoop.record.compiler.JInt
- Creates a new instance of JInt
- jj_nt -
Variable in class org.apache.hadoop.record.compiler.generated.Rcc
-
- jjFillToken() -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- jjnewLexState -
Static variable in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- jjstrLiteralImages -
Static variable in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- JLong - Class in org.apache.hadoop.record.compiler
- Code generator for "long" type
- JLong() -
Constructor for class org.apache.hadoop.record.compiler.JLong
- Creates a new instance of JLong
- JMap - Class in org.apache.hadoop.record.compiler
-
- JMap(JType, JType) -
Constructor for class org.apache.hadoop.record.compiler.JMap
- Creates a new instance of JMap
- JMXJsonServlet - Class in org.apache.hadoop.jmx
- Provides Read only web access to JMX.
- JMXJsonServlet() -
Constructor for class org.apache.hadoop.jmx.JMXJsonServlet
-
- JniBasedUnixGroupsMapping - Class in org.apache.hadoop.security
- A JNI-based implementation of
GroupMappingServiceProvider
that invokes libC calls to get the group
memberships of a given user. - JniBasedUnixGroupsMapping() -
Constructor for class org.apache.hadoop.security.JniBasedUnixGroupsMapping
-
- JniBasedUnixGroupsMappingWithFallback - Class in org.apache.hadoop.security
-
- JniBasedUnixGroupsMappingWithFallback() -
Constructor for class org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback
-
- JniBasedUnixGroupsNetgroupMapping - Class in org.apache.hadoop.security
- A JNI-based implementation of
GroupMappingServiceProvider
that invokes libC calls to get the group
memberships of a given user. - JniBasedUnixGroupsNetgroupMapping() -
Constructor for class org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMapping
-
- JniBasedUnixGroupsNetgroupMappingWithFallback - Class in org.apache.hadoop.security
-
- JniBasedUnixGroupsNetgroupMappingWithFallback() -
Constructor for class org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMappingWithFallback
-
- job -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- job -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- Job - Class in org.apache.hadoop.mapred.jobcontrol
- This class encapsulates a MapReduce job and its dependency.
- Job(JobConf, ArrayList<Job>) -
Constructor for class org.apache.hadoop.mapred.jobcontrol.Job
- Construct a job.
- Job(JobConf) -
Constructor for class org.apache.hadoop.mapred.jobcontrol.Job
- Construct a job.
- job -
Variable in class org.apache.hadoop.mapred.Task.CombinerRunner
-
- Job - Class in org.apache.hadoop.mapreduce
- The job submitter's view of the Job.
- Job() -
Constructor for class org.apache.hadoop.mapreduce.Job
-
- Job(Configuration) -
Constructor for class org.apache.hadoop.mapreduce.Job
-
- Job(Configuration, String) -
Constructor for class org.apache.hadoop.mapreduce.Job
-
- JOB -
Static variable in class org.apache.hadoop.mapreduce.JobID
-
- Job.JobState - Enum in org.apache.hadoop.mapreduce
-
- JOB_ACL_MODIFY_JOB -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- JOB_ACL_VIEW_JOB -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- JOB_CANCEL_DELEGATION_TOKEN -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- JOB_DIR_PERMISSION -
Static variable in class org.apache.hadoop.mapreduce.JobSubmissionFiles
-
- JOB_FILE_PERMISSION -
Static variable in class org.apache.hadoop.mapreduce.JobSubmissionFiles
-
- JOB_LOCAL_CTXT -
Static variable in class org.apache.hadoop.mapred.JobLocalizer
-
- JOB_NAME_TRIM_LENGTH -
Static variable in class org.apache.hadoop.mapred.JobHistory
-
- JOB_NAMENODES -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- JOB_TOKEN_HDFS_FILE -
Static variable in class org.apache.hadoop.mapreduce.security.TokenCache
- file name used on HDFS for generated job token
- JOB_TOKENS_FILENAME -
Static variable in class org.apache.hadoop.mapreduce.security.TokenCache
- conf setting for job tokens cache file name
- JobACL - Enum in org.apache.hadoop.mapreduce
- Job related ACLs
- jobACLNeeded -
Variable in enum org.apache.hadoop.mapred.Operation
-
- JobBase - Class in org.apache.hadoop.contrib.utils.join
- A common base implementing some statics collecting mechanisms that are
commonly used in a typical map/reduce job.
- JobBase() -
Constructor for class org.apache.hadoop.contrib.utils.join.JobBase
-
- jobCleanup -
Variable in class org.apache.hadoop.mapred.Task
-
- JobClient - Class in org.apache.hadoop.mapred
JobClient
is the primary interface for the user-job to interact
with the JobTracker
.- JobClient() -
Constructor for class org.apache.hadoop.mapred.JobClient
- Create a job client.
- JobClient(JobConf) -
Constructor for class org.apache.hadoop.mapred.JobClient
- Build a job client with the given
JobConf
, and connect to the
default JobTracker
.
- JobClient(InetSocketAddress, Configuration) -
Constructor for class org.apache.hadoop.mapred.JobClient
- Build a job client, connect to the indicated job tracker.
- JobClient.Renewer - Class in org.apache.hadoop.mapred
-
- JobClient.Renewer() -
Constructor for class org.apache.hadoop.mapred.JobClient.Renewer
-
- JobClient.TaskStatusFilter - Enum in org.apache.hadoop.mapred
-
- JobCompletedEvent - Class in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
- This is an
UserLogEvent
sent when the job completes - JobCompletedEvent(JobID, long, int) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobCompletedEvent
- Create the event for job completion.
- JobConf - Class in org.apache.hadoop.mapred
- A map/reduce job configuration.
- JobConf() -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce job configuration.
- JobConf(Class) -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce job configuration.
- JobConf(Configuration) -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce job configuration.
- JobConf(Configuration, Class) -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce job configuration.
- JobConf(String) -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce configuration.
- JobConf(Path) -
Constructor for class org.apache.hadoop.mapred.JobConf
- Construct a map/reduce configuration.
- JobConf(boolean) -
Constructor for class org.apache.hadoop.mapred.JobConf
- A new map/reduce configuration where the behavior of reading from the
default resources can be turned off.
- jobConf_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- JobConfigurable - Interface in org.apache.hadoop.mapred
- That what may be configured.
- JobContext - Class in org.apache.hadoop.mapred
-
- jobContext -
Variable in class org.apache.hadoop.mapred.Task
-
- JobContext - Class in org.apache.hadoop.mapreduce
- A read-only view of the job that is provided to the tasks while they
are running.
- JobContext(Configuration, JobID) -
Constructor for class org.apache.hadoop.mapreduce.JobContext
-
- JobControl - Class in org.apache.hadoop.mapred.jobcontrol
- This class encapsulates a set of MapReduce jobs and its dependency.
- JobControl(String) -
Constructor for class org.apache.hadoop.mapred.jobcontrol.JobControl
- Construct a job control for a group of jobs.
- JobControl - Class in org.apache.hadoop.mapreduce.lib.jobcontrol
- This class encapsulates a set of MapReduce jobs and its dependency.
- JobControl(String) -
Constructor for class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
- Construct a job control for a group of jobs.
- JobControl.ThreadState - Enum in org.apache.hadoop.mapreduce.lib.jobcontrol
-
- JobEndNotifier - Class in org.apache.hadoop.mapred
-
- JobEndNotifier() -
Constructor for class org.apache.hadoop.mapred.JobEndNotifier
-
- JobHistory - Class in org.apache.hadoop.mapred
- Provides methods for writing to and reading from job history.
- JobHistory() -
Constructor for class org.apache.hadoop.mapred.JobHistory
-
- JobHistory.HistoryCleaner - Class in org.apache.hadoop.mapred
- Delete history files older than one month.
- JobHistory.HistoryCleaner() -
Constructor for class org.apache.hadoop.mapred.JobHistory.HistoryCleaner
-
- JobHistory.JobInfo - Class in org.apache.hadoop.mapred
- Helper class for logging or reading back events related to job start, finish or failure.
- JobHistory.JobInfo(String) -
Constructor for class org.apache.hadoop.mapred.JobHistory.JobInfo
- Create new JobInfo
- JobHistory.Keys - Enum in org.apache.hadoop.mapred
- Job history files contain key="value" pairs, where keys belong to this enum.
- JobHistory.Listener - Interface in org.apache.hadoop.mapred
- Callback interface for reading back log events from JobHistory.
- JobHistory.MapAttempt - Class in org.apache.hadoop.mapred
- Helper class for logging or reading back events related to start, finish or failure of
a Map Attempt on a node.
- JobHistory.MapAttempt() -
Constructor for class org.apache.hadoop.mapred.JobHistory.MapAttempt
-
- JobHistory.RecordTypes - Enum in org.apache.hadoop.mapred
- Record types are identifiers for each line of log in history files.
- JobHistory.ReduceAttempt - Class in org.apache.hadoop.mapred
- Helper class for logging or reading back events related to start, finish or failure of
a Map Attempt on a node.
- JobHistory.ReduceAttempt() -
Constructor for class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
-
- JobHistory.Task - Class in org.apache.hadoop.mapred
- Helper class for logging or reading back events related to Task's start, finish or failure.
- JobHistory.Task() -
Constructor for class org.apache.hadoop.mapred.JobHistory.Task
-
- JobHistory.TaskAttempt - Class in org.apache.hadoop.mapred
- Base class for Map and Reduce TaskAttempts.
- JobHistory.TaskAttempt() -
Constructor for class org.apache.hadoop.mapred.JobHistory.TaskAttempt
-
- JobHistory.Values - Enum in org.apache.hadoop.mapred
- This enum contains some of the values commonly used by history log events.
- JOBHISTORY_FILENAME_REGEX -
Static variable in class org.apache.hadoop.mapred.JobHistory
-
- JobHistoryServer - Class in org.apache.hadoop.mapred
JobHistoryServer
is responsible for servicing all job history
related requests from client.- JobHistoryServer(JobConf) -
Constructor for class org.apache.hadoop.mapred.JobHistoryServer
- Starts job history server as a independent process
* Initializes ACL Manager
* Starts a webapp to service history requests
- JobHistoryServer(JobConf, ACLsManager, HttpServer) -
Constructor for class org.apache.hadoop.mapred.JobHistoryServer
- Starts job history server as a embedded server within job tracker
* Starts a webapp to service history requests
- JobID - Class in org.apache.hadoop.mapred
- JobID represents the immutable and unique identifier for
the job.
- JobID(String, int) -
Constructor for class org.apache.hadoop.mapred.JobID
- Constructs a JobID object
- JobID() -
Constructor for class org.apache.hadoop.mapred.JobID
-
- JobID - Class in org.apache.hadoop.mapreduce
- JobID represents the immutable and unique identifier for
the job.
- JobID(String, int) -
Constructor for class org.apache.hadoop.mapreduce.JobID
- Constructs a JobID object
- JobID() -
Constructor for class org.apache.hadoop.mapreduce.JobID
-
- jobId_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- JOBID_REGEX -
Static variable in class org.apache.hadoop.mapreduce.JobID
-
- jobInfo() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- JobInProgress - Class in org.apache.hadoop.mapred
- JobInProgress maintains all the info for keeping
a Job on the straight and narrow.
- JobInProgress(JobID, JobConf, JobTracker) -
Constructor for class org.apache.hadoop.mapred.JobInProgress
- Create an almost empty JobInProgress, which can be used only for tests
- JobInProgress.Counter - Enum in org.apache.hadoop.mapred
-
- JobLocalizer - Class in org.apache.hadoop.mapred
- Internal class responsible for initializing the job, not intended for users.
- JobLocalizer(JobConf, String, String) -
Constructor for class org.apache.hadoop.mapred.JobLocalizer
-
- JobLocalizer(JobConf, String, String, String...) -
Constructor for class org.apache.hadoop.mapred.JobLocalizer
-
- JobPriority - Enum in org.apache.hadoop.mapred
- Used to describe the priority of the running job.
- JobProfile - Class in org.apache.hadoop.mapred
- A JobProfile is a MapReduce primitive.
- JobProfile() -
Constructor for class org.apache.hadoop.mapred.JobProfile
- Construct an empty
JobProfile
.
- JobProfile(String, JobID, String, String, String) -
Constructor for class org.apache.hadoop.mapred.JobProfile
- Construct a
JobProfile
the userid, jobid,
job config-file, job-details url and job name.
- JobProfile(String, JobID, String, String, String, String) -
Constructor for class org.apache.hadoop.mapred.JobProfile
- Construct a
JobProfile
the userid, jobid,
job config-file, job-details url and job name.
- JobProfile(String, String, String, String, String) -
Constructor for class org.apache.hadoop.mapred.JobProfile
- Deprecated. use JobProfile(String, JobID, String, String, String) instead
- JobQueueInfo - Class in org.apache.hadoop.mapred
- Class that contains the information regarding the Job Queues which are
maintained by the Hadoop Map/Reduce framework.
- JobQueueInfo() -
Constructor for class org.apache.hadoop.mapred.JobQueueInfo
-
- JobQueueInfo(String, String) -
Constructor for class org.apache.hadoop.mapred.JobQueueInfo
- Construct a new JobQueueInfo object using the queue name and the
scheduling information passed.
- jobRunStateForCleanup -
Variable in class org.apache.hadoop.mapred.Task
-
- jobSetup -
Variable in class org.apache.hadoop.mapred.Task
-
- JobSplit - Class in org.apache.hadoop.mapreduce.split
- This class groups the fundamental classes associated with
reading/writing splits.
- JobSplit() -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit
-
- JobSplit.SplitMetaInfo - Class in org.apache.hadoop.mapreduce.split
- This represents the meta information about the task split.
- JobSplit.SplitMetaInfo() -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- JobSplit.SplitMetaInfo(String[], long, long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- JobSplit.SplitMetaInfo(InputSplit, long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- JobSplit.TaskSplitIndex - Class in org.apache.hadoop.mapreduce.split
- This represents the meta information about the task split that the
task gets
- JobSplit.TaskSplitIndex() -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
-
- JobSplit.TaskSplitIndex(String, long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
-
- JobSplit.TaskSplitMetaInfo - Class in org.apache.hadoop.mapreduce.split
- This represents the meta information about the task split that the
JobTracker creates
- JobSplit.TaskSplitMetaInfo() -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
-
- JobSplit.TaskSplitMetaInfo(JobSplit.TaskSplitIndex, String[], long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
-
- JobSplit.TaskSplitMetaInfo(InputSplit, long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
-
- JobSplit.TaskSplitMetaInfo(String[], long, long) -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
-
- JobSplitWriter - Class in org.apache.hadoop.mapreduce.split
- The class that is used by the Job clients to write splits (both the meta
and the raw bytes parts)
- JobSplitWriter() -
Constructor for class org.apache.hadoop.mapreduce.split.JobSplitWriter
-
- JobStartedEvent - Class in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
- This is an
UserLogEvent
sent when the job starts. - JobStartedEvent(JobID) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JobStartedEvent
- Create the event to inform the job has started.
- JobStatus - Class in org.apache.hadoop.mapred
- Describes the current status of a job.
- JobStatus() -
Constructor for class org.apache.hadoop.mapred.JobStatus
-
- JobStatus(JobID, float, float, float, int) -
Constructor for class org.apache.hadoop.mapred.JobStatus
- Create a job status object for a given jobid.
- JobStatus(JobID, float, float, int) -
Constructor for class org.apache.hadoop.mapred.JobStatus
- Create a job status object for a given jobid.
- JobStatus(JobID, float, float, float, int, JobPriority) -
Constructor for class org.apache.hadoop.mapred.JobStatus
- Create a job status object for a given jobid.
- JobStatus(JobID, float, float, float, float, int, JobPriority) -
Constructor for class org.apache.hadoop.mapred.JobStatus
- Create a job status object for a given jobid.
- JobStatus - Class in org.apache.hadoop.mapreduce
- Describes the current status of a job.
- JobStatus() -
Constructor for class org.apache.hadoop.mapreduce.JobStatus
-
- JobStatus.State - Enum in org.apache.hadoop.mapreduce
- Current state of the job
- jobsToComplete() -
Method in class org.apache.hadoop.mapred.JobClient
- Get the jobs that are not completed and not failed.
- jobsToComplete() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- jobsToComplete() -
Method in class org.apache.hadoop.mapred.LocalJobRunner
-
- JobSubmissionFiles - Class in org.apache.hadoop.mapreduce
- A utility to manage job submission files.
Note that this class is for framework internal usage only and is
not to be used by users directly. - JobSubmissionFiles() -
Constructor for class org.apache.hadoop.mapreduce.JobSubmissionFiles
-
- jobSubmit(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Submit a job to the Map-Reduce framework.
- JobTokenIdentifier - Class in org.apache.hadoop.mapreduce.security.token
- The token identifier for job token
- JobTokenIdentifier() -
Constructor for class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
- Default constructor
- JobTokenIdentifier(Text) -
Constructor for class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
- Create a job token identifier from a jobid
- JobTokenIdentifier.Renewer - Class in org.apache.hadoop.mapreduce.security.token
-
- JobTokenIdentifier.Renewer() -
Constructor for class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier.Renewer
-
- JobTokenSecretManager - Class in org.apache.hadoop.mapreduce.security.token
- SecretManager for job token.
- JobTokenSecretManager() -
Constructor for class org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager
- Default constructor
- JobTokenSelector - Class in org.apache.hadoop.mapreduce.security.token
- Look through tokens to find the first job token that matches the service
and return it.
- JobTokenSelector() -
Constructor for class org.apache.hadoop.mapreduce.security.token.JobTokenSelector
-
- JobTracker - Class in org.apache.hadoop.mapred
- JobTracker is the central location for submitting and
tracking MR jobs in a network environment.
- JobTracker.IllegalStateException - Exception in org.apache.hadoop.mapred
- A client tried to submit a job before the Job Tracker was ready.
- JobTracker.IllegalStateException(String) -
Constructor for exception org.apache.hadoop.mapred.JobTracker.IllegalStateException
-
- JobTracker.SafeModeAction - Enum in org.apache.hadoop.mapred
- JobTracker SafeMode
- JobTracker.State - Enum in org.apache.hadoop.mapred
-
- JobTrackerMXBean - Interface in org.apache.hadoop.mapred
- The MXBean interface for JobTrackerInfo
- JobTrackerNotYetInitializedException - Exception in org.apache.hadoop.mapred
- This exception is thrown when the JobTracker is still initializing and
not yet operational.
- JobTrackerNotYetInitializedException() -
Constructor for exception org.apache.hadoop.mapred.JobTrackerNotYetInitializedException
-
- Join - Class in org.apache.hadoop.examples
- This is the trivial map/reduce program that does absolutely nothing
other than use the framework to fragment and sort the input values.
- Join() -
Constructor for class org.apache.hadoop.examples.Join
-
- join() -
Method in class org.apache.hadoop.http.HttpServer
-
- join() -
Method in class org.apache.hadoop.ipc.Server
- Wait for the server to be stopped.
- join() -
Method in class org.apache.hadoop.mapred.JobHistoryServer
-
- join(CharSequence, Iterable<String>) -
Static method in class org.apache.hadoop.util.StringUtils
- Concatenates strings, using a separator.
- join(CharSequence, String[]) -
Static method in class org.apache.hadoop.util.StringUtils
- Concatenates strings, using a separator.
- join(CharSequence, Object[]) -
Static method in class org.apache.hadoop.util.StringUtils
- Concatenates objects, using a separator.
- JoinRecordReader<K extends WritableComparable> - Class in org.apache.hadoop.mapred.join
- Base class for Composite joins returning Tuples of arbitrary Writables.
- JoinRecordReader(int, JobConf, int, Class<? extends WritableComparator>) -
Constructor for class org.apache.hadoop.mapred.join.JoinRecordReader
-
- JoinRecordReader.JoinDelegationIterator - Class in org.apache.hadoop.mapred.join
- Since the JoinCollector is effecting our operation, we need only
provide an iterator proxy wrapping its operation.
- JoinRecordReader.JoinDelegationIterator() -
Constructor for class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- JRecord - Class in org.apache.hadoop.record.compiler
-
- JRecord(String, ArrayList<JField<JType>>) -
Constructor for class org.apache.hadoop.record.compiler.JRecord
- Creates a new instance of JRecord
- JString - Class in org.apache.hadoop.record.compiler
-
- JString() -
Constructor for class org.apache.hadoop.record.compiler.JString
- Creates a new instance of JString
- JT_KEYTAB_FILE -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- JT_USER_NAME -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- JType - Class in org.apache.hadoop.record.compiler
- Abstract Base class for all types supported by Hadoop Record I/O.
- JType() -
Constructor for class org.apache.hadoop.record.compiler.JType
-
- JVector - Class in org.apache.hadoop.record.compiler
-
- JVector(JType) -
Constructor for class org.apache.hadoop.record.compiler.JVector
- Creates a new instance of JVector
- jvmContext -
Variable in class org.apache.hadoop.mapred.Task
-
- JvmFinishedEvent - Class in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
- This is an
UserLogEvent
sent when the jvm finishes. - JvmFinishedEvent(JVMInfo) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.JvmFinishedEvent
- Create the event to inform that the jvm has finished.
- JVMInfo - Class in org.apache.hadoop.mapreduce.server.tasktracker
-
- JVMInfo(File, List<Task>) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.JVMInfo
-
- JvmMetrics - Class in org.apache.hadoop.metrics.jvm
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - JvmMetricsSource - Class in org.apache.hadoop.metrics2.source
-
- JvmTask - Class in org.apache.hadoop.mapred
-
- JvmTask(Task, boolean) -
Constructor for class org.apache.hadoop.mapred.JvmTask
-
- JvmTask() -
Constructor for class org.apache.hadoop.mapred.JvmTask
-
K
- KerberosAuthenticationHandler - Class in org.apache.hadoop.security.authentication.server
- The
KerberosAuthenticationHandler
implements the Kerberos SPNEGO authentication mechanism for HTTP. - KerberosAuthenticationHandler() -
Constructor for class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
-
- KerberosAuthenticator - Class in org.apache.hadoop.security.authentication.client
- The
KerberosAuthenticator
implements the Kerberos SPNEGO authentication sequence. - KerberosAuthenticator() -
Constructor for class org.apache.hadoop.security.authentication.client.KerberosAuthenticator
-
- KerberosInfo - Annotation Type in org.apache.hadoop.security
- Indicates Kerberos related information to be used
- KerberosName - Class in org.apache.hadoop.security
- This class implements parsing and handling of Kerberos principal names.
- KerberosName(String) -
Constructor for class org.apache.hadoop.security.KerberosName
- Create a name from the full Kerberos principal name.
- KerberosName.BadFormatString - Exception in org.apache.hadoop.security
-
- KerberosName.NoMatchingRule - Exception in org.apache.hadoop.security
-
- KerberosUtil - Class in org.apache.hadoop.security.authentication.util
-
- KerberosUtil() -
Constructor for class org.apache.hadoop.security.authentication.util.KerberosUtil
-
- key() -
Method in class org.apache.hadoop.io.ArrayFile.Reader
- Returns the key associated with the most recent call to
ArrayFile.Reader.seek(long)
, ArrayFile.Reader.next(Writable)
, or ArrayFile.Reader.get(long,Writable)
.
- key() -
Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
- Return the key this RecordReader would supply on a call to next(K,V)
- key(K) -
Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
- Clone the key at the head of this RecordReader into the object provided.
- key() -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Return the key for the current join or the value at the top of the
RecordReader heap.
- key(K) -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Clone the key at the top of this RR into the given object.
- key() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Return the key at the head of this RR.
- key(K) -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Clone the key at the head of this RR into the object supplied.
- Key - Class in org.apache.hadoop.util.bloom
- The general behavior of a key that must be stored in a filter.
- Key() -
Constructor for class org.apache.hadoop.util.bloom.Key
- default constructor - use with readFields
- Key(byte[]) -
Constructor for class org.apache.hadoop.util.bloom.Key
- Constructor.
- Key(byte[], double) -
Constructor for class org.apache.hadoop.util.bloom.Key
- Constructor.
- KEY_CLASS -
Static variable in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
-
- KEY_VALUE_SEPERATOR -
Static variable in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
-
- KeyFieldBasedComparator<K,V> - Class in org.apache.hadoop.mapred.lib
- This comparator implementation provides a subset of the features provided
by the Unix/GNU Sort.
- KeyFieldBasedComparator() -
Constructor for class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator
-
- KeyFieldBasedComparator<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- This comparator implementation provides a subset of the features provided
by the Unix/GNU Sort.
- KeyFieldBasedComparator() -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
-
- KeyFieldBasedPartitioner<K2,V2> - Class in org.apache.hadoop.mapred.lib
- Defines a way to partition keys based on certain key fields (also see
KeyFieldBasedComparator
. - KeyFieldBasedPartitioner() -
Constructor for class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
-
- KeyFieldBasedPartitioner<K2,V2> - Class in org.apache.hadoop.mapreduce.lib.partition
- Defines a way to partition keys based on certain key fields (also see
KeyFieldBasedComparator
. - KeyFieldBasedPartitioner() -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
-
- keySerializer -
Variable in class org.apache.hadoop.io.SequenceFile.Writer
-
- keySet() -
Method in class org.apache.hadoop.io.MapWritable
-
- keySet() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- KEYTAB -
Static variable in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- Constant for the configuration property that indicates the keytab file path.
- KeyValueLineRecordReader - Class in org.apache.hadoop.mapred
- This class treats a line in the input as a key/value pair separated by a
separator character.
- KeyValueLineRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- KeyValueLineRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
- This class treats a line in the input as a key/value pair separated by a
separator character.
- KeyValueLineRecordReader(Configuration) -
Constructor for class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
-
- KeyValueTextInputFormat - Class in org.apache.hadoop.mapred
- An
InputFormat
for plain text files. - KeyValueTextInputFormat() -
Constructor for class org.apache.hadoop.mapred.KeyValueTextInputFormat
-
- KeyValueTextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
- An
InputFormat
for plain text files. - KeyValueTextInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat
-
- kids -
Variable in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
- kill() -
Method in class org.apache.hadoop.mapred.JobInProgress
- Kill the job and all its component tasks.
- KILLED -
Static variable in class org.apache.hadoop.mapred.JobStatus
-
- killJob(JobID) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- killJob(JobID) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
-
- killJob() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Kill the running job.
- killJob() -
Method in class org.apache.hadoop.mapreduce.Job
- Kill the running job.
- killJob() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
-
- killProcess(String, ProcessTree.Signal) -
Static method in class org.apache.hadoop.util.ProcessTree
- Sends signal to process, forcefully terminating the process.
- killProcessGroup(String, ProcessTree.Signal) -
Static method in class org.apache.hadoop.util.ProcessTree
- Sends signal to all process belonging to same process group,
forcefully terminating the process group.
- killTask(TaskAttemptID, boolean) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- killTask(TaskAttemptID, boolean) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
- Throws
UnsupportedOperationException
- killTask(TaskAttemptID, boolean) -
Method in interface org.apache.hadoop.mapred.RunningJob
- Kill indicated task attempt.
- killTask(String, boolean) -
Method in interface org.apache.hadoop.mapred.RunningJob
- Deprecated. Applications should rather use
RunningJob.killTask(TaskAttemptID, boolean)
- killTask(TaskAttemptID) -
Method in class org.apache.hadoop.mapreduce.Job
- Kill indicated task attempt.
- kind -
Variable in class org.apache.hadoop.record.compiler.generated.Token
- An integer that describes the kind of this token.
- KIND_NAME -
Static variable in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
-
- KosmosFileSystem - Class in org.apache.hadoop.fs.kfs
- A FileSystem backed by KFS.
- KosmosFileSystem() -
Constructor for class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- KRB5_CIPHER_SUITES -
Static variable in class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector
-
- KRB5_FILTER -
Static variable in class org.apache.hadoop.http.HttpServer
-
- Krb5AndCertsSslSocketConnector - Class in org.apache.hadoop.security
- Extend Jetty's
SslSocketConnector
to optionally also provide
Kerberos5ized SSL sockets. - Krb5AndCertsSslSocketConnector() -
Constructor for class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector
-
- Krb5AndCertsSslSocketConnector(Krb5AndCertsSslSocketConnector.MODE) -
Constructor for class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector
-
- Krb5AndCertsSslSocketConnector.Krb5SslFilter - Class in org.apache.hadoop.security
- Filter that takes the Kerberos principal identified in the
Krb5AndCertsSslSocketConnector
and provides it the to the servlet
at runtime, setting the principal and short name. - Krb5AndCertsSslSocketConnector.Krb5SslFilter() -
Constructor for class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector.Krb5SslFilter
-
- Krb5AndCertsSslSocketConnector.MODE - Enum in org.apache.hadoop.security
-
L
- largestNumOfValues -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- lastKey() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- launchTask(String, String, String, List<String>, List<String>, File, String, String) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
- Create all of the directories for the task and launches the child jvm.
- launchTask(String, String, String, List<String>, List<String>, File, String, String) -
Method in class org.apache.hadoop.mapred.TaskController
- Create all of the directories for the task and launches the child jvm.
- launchTaskForJob(TaskTracker.TaskInProgress, JobConf, TaskTracker.RunningJob) -
Method in class org.apache.hadoop.mapred.TaskTracker
-
- LazyOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
- A Convenience class that creates output lazily.
- LazyOutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
-
- LBRACE_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- lDirAlloc -
Variable in class org.apache.hadoop.mapred.Task
-
- LENGTH -
Static variable in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
- LESSER_ICOST -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- lessThan(Object, Object) -
Method in class org.apache.hadoop.util.PriorityQueue
- Determines the ordering of objects in this priority queue.
- level -
Variable in class org.apache.hadoop.net.NodeBase
-
- LexicalError(boolean, int, int, int, String, char) -
Static method in error org.apache.hadoop.record.compiler.generated.TokenMgrError
- Returns a detailed message for the Error when it is thrown by the
token manager to indicate a lexical error.
- lexStateNames -
Static variable in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- limitDecimalTo2(double) -
Static method in class org.apache.hadoop.fs.FsShell
- Deprecated. Consider using
StringUtils.limitDecimalTo2(double)
instead.
- limitDecimalTo2(double) -
Static method in class org.apache.hadoop.util.StringUtils
-
- line -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- LineDocInputFormat - Class in org.apache.hadoop.contrib.index.example
- An InputFormat for LineDoc for plain text files where each line is a doc.
- LineDocInputFormat() -
Constructor for class org.apache.hadoop.contrib.index.example.LineDocInputFormat
-
- LineDocLocalAnalysis - Class in org.apache.hadoop.contrib.index.example
- Convert LineDocTextAndOp to DocumentAndOp as required by ILocalAnalysis.
- LineDocLocalAnalysis() -
Constructor for class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
-
- LineDocRecordReader - Class in org.apache.hadoop.contrib.index.example
- A simple RecordReader for LineDoc for plain text files where each line is a
doc.
- LineDocRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.contrib.index.example.LineDocRecordReader
- Constructor
- LineDocTextAndOp - Class in org.apache.hadoop.contrib.index.example
- This class represents an operation.
- LineDocTextAndOp() -
Constructor for class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
- Constructor
- LineReader - Class in org.apache.hadoop.util
- A class that provides a line reader from an input stream.
- LineReader(InputStream) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
default buffer-size (64k).
- LineReader(InputStream, int) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
given buffer-size.
- LineReader(InputStream, Configuration) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
io.file.buffer.size
specified in the given
Configuration
.
- LineReader(InputStream, byte[]) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
default buffer-size, and using a custom delimiter of array of
bytes.
- LineReader(InputStream, int, byte[]) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
given buffer-size, and using a custom delimiter of array of
bytes.
- LineReader(InputStream, Configuration, byte[]) -
Constructor for class org.apache.hadoop.util.LineReader
- Create a line reader that reads from the given stream using the
io.file.buffer.size
specified in the given
Configuration
, and using a custom delimiter of array of
bytes.
- LineRecordReader - Class in org.apache.hadoop.mapred
- Treats keys as offset in file and value as line.
- LineRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.mapred.LineRecordReader
-
- LineRecordReader(InputStream, long, long, int) -
Constructor for class org.apache.hadoop.mapred.LineRecordReader
-
- LineRecordReader(InputStream, long, long, Configuration) -
Constructor for class org.apache.hadoop.mapred.LineRecordReader
-
- LineRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
- Treats keys as offset in file and value as line.
- LineRecordReader() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- LineRecordReader.LineReader - Class in org.apache.hadoop.mapred
- Deprecated. Use
LineReader
instead. - LineRecordReader.LineReader(InputStream, Configuration) -
Constructor for class org.apache.hadoop.mapred.LineRecordReader.LineReader
- Deprecated.
- LINES_PER_MAP -
Static variable in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
-
- LINK_URI -
Static variable in class org.apache.hadoop.streaming.StreamJob
-
- linkStats -
Variable in class org.apache.hadoop.fs.HardLink
-
- LinuxMemoryCalculatorPlugin - Class in org.apache.hadoop.util
- Deprecated. Use
LinuxResourceCalculatorPlugin
instead - LinuxMemoryCalculatorPlugin() -
Constructor for class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin
- Deprecated.
- LinuxResourceCalculatorPlugin - Class in org.apache.hadoop.util
- Plugin to calculate resource information on Linux systems.
- LinuxResourceCalculatorPlugin() -
Constructor for class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
-
- LinuxResourceCalculatorPlugin(String, String, String, long) -
Constructor for class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
- Constructor which allows assigning the /proc/ directories.
- list() -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- list(File) -
Static method in class org.apache.hadoop.fs.FileUtil
- A wrapper for
File.list()
.
- listDeepSubPaths(Path) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- listener -
Variable in class org.apache.hadoop.http.HttpServer
-
- listFiles(File) -
Static method in class org.apache.hadoop.fs.FileUtil
- A wrapper for
File.listFiles()
.
- listJobConfProperties() -
Method in class org.apache.hadoop.streaming.StreamJob
- Prints out the jobconf properties on stdout
when verbose is specified.
- listStatus(Path) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- List the statuses of the files/directories in the given path if the path is
a directory.
- listStatus(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- List the statuses of the files/directories in the given path if the path is
a directory.
- listStatus(Path, PathFilter) -
Method in class org.apache.hadoop.fs.FileSystem
- Filter files/directories in the given path using the user-supplied path
filter.
- listStatus(Path[]) -
Method in class org.apache.hadoop.fs.FileSystem
- Filter files/directories in the given list of paths using default
path filter.
- listStatus(Path[], PathFilter) -
Method in class org.apache.hadoop.fs.FileSystem
- Filter files/directories in the given list of paths using user-supplied
path filter.
- listStatus(Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- List files in a directory.
- listStatus(Path) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- listStatus(Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
- liststatus returns the children of a directory
after looking up the index files.
- listStatus(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- listStatus(Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- listStatus(Path) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- listStatus(Path) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
If
f
is a file, this method will make a single call to S3.
- listStatus(JobConf) -
Method in class org.apache.hadoop.mapred.FileInputFormat
- List input directories.
- listStatus(JobConf) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFormat
-
- listStatus(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- List input directories.
- listStatus(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
-
- listSubPaths(Path) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- ljustify(String, int) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- load(Configuration, String, Class<K>) -
Static method in class org.apache.hadoop.io.DefaultStringifier
- Restores the object from the configuration.
- loadArray(Configuration, String, Class<K>) -
Static method in class org.apache.hadoop.io.DefaultStringifier
- Restores the array of objects from the configuration.
- LoadSnappy - Class in org.apache.hadoop.io.compress.snappy
- Determines if Snappy native library is available and loads it if available.
- LoadSnappy() -
Constructor for class org.apache.hadoop.io.compress.snappy.LoadSnappy
-
- loadTokens(String, Configuration) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
- load job token from a file
- LoadTypedBytes - Class in org.apache.hadoop.streaming
- Utility program that reads typed bytes from standard input and stores them in
a sequence file for which the path is given as an argument.
- LoadTypedBytes(Configuration) -
Constructor for class org.apache.hadoop.streaming.LoadTypedBytes
-
- LoadTypedBytes() -
Constructor for class org.apache.hadoop.streaming.LoadTypedBytes
-
- LOCAL_MAX_MAPS -
Static variable in class org.apache.hadoop.mapred.LocalJobRunner
-
- LocalDirAllocator - Class in org.apache.hadoop.fs
- An implementation of a round-robin scheme for disk allocation for creating
files.
- LocalDirAllocator(String) -
Constructor for class org.apache.hadoop.fs.LocalDirAllocator
- Create an allocator object
- LocalFileSystem - Class in org.apache.hadoop.fs
- Implement the FileSystem API for the checksumed local filesystem.
- LocalFileSystem() -
Constructor for class org.apache.hadoop.fs.LocalFileSystem
-
- LocalFileSystem(FileSystem) -
Constructor for class org.apache.hadoop.fs.LocalFileSystem
-
- localHadoop_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- localizeBin(String) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- localizeConfiguration(JobConf) -
Method in class org.apache.hadoop.mapred.Task
- Localize the given JobConf to be specific for this task.
- localizeJobFiles(JobID, JobConf, Path, TaskUmbilicalProtocol) -
Method in class org.apache.hadoop.mapred.JobLocalizer
-
- localizeJobFiles(JobID, JobConf, Path, Path, TaskUmbilicalProtocol) -
Method in class org.apache.hadoop.mapred.JobLocalizer
-
- Localizer - Class in org.apache.hadoop.mapreduce.server.tasktracker
- NOTE: This class is internal only and not intended for users!!
- Localizer(FileSystem, String[]) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.Localizer
- Create a Localizer instance
- LocalJobRunner - Class in org.apache.hadoop.mapred
- Implements MapReduce locally, in-process, for debugging.
- LocalJobRunner(JobConf) -
Constructor for class org.apache.hadoop.mapred.LocalJobRunner
-
- localRunnerNotification(JobConf, JobStatus) -
Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- localStorage -
Variable in class org.apache.hadoop.mapred.TaskController
-
- LocalStore - Class in org.apache.hadoop.contrib.failmon
- This class takes care of the temporary local storage of
gathered metrics before they get uploaded into HDFS.
- LocalStore() -
Constructor for class org.apache.hadoop.contrib.failmon.LocalStore
- Create an instance of the class and read the configuration
file to determine some output parameters.
- location -
Variable in class org.apache.hadoop.net.NodeBase
-
- lock(Path, boolean) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
- Deprecated.
- LOG -
Static variable in class org.apache.hadoop.contrib.failmon.Environment
-
- LOG -
Static variable in class org.apache.hadoop.contrib.index.main.UpdateIndex
-
- LOG -
Static variable in class org.apache.hadoop.contrib.index.mapred.IndexUpdater
-
- LOG -
Static variable in class org.apache.hadoop.contrib.utils.join.JobBase
-
- LOG -
Static variable in class org.apache.hadoop.fs.FileSystem
-
- LOG -
Static variable in class org.apache.hadoop.fs.FSInputChecker
-
- LOG -
Static variable in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- LOG -
Static variable in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- LOG -
Static variable in class org.apache.hadoop.http.HttpServer
-
- LOG -
Static variable in class org.apache.hadoop.io.compress.CompressionCodecFactory
-
- LOG -
Static variable in class org.apache.hadoop.io.retry.RetryUtils
-
- LOG -
Static variable in class org.apache.hadoop.ipc.Client
-
- LOG -
Static variable in class org.apache.hadoop.ipc.Server
-
- LOG -
Static variable in class org.apache.hadoop.mapred.CleanupQueue
-
- log(Log) -
Method in class org.apache.hadoop.mapred.Counters
- Logs the current counter values.
- LOG -
Static variable in class org.apache.hadoop.mapred.FileInputFormat
-
- LOG -
Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
-
- LOG -
Static variable in class org.apache.hadoop.mapred.HDFSMonitorThread
-
- LOG -
Static variable in class org.apache.hadoop.mapred.JobHistory
-
- LOG -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- LOG -
Static variable in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
-
- LOG -
Static variable in class org.apache.hadoop.mapred.LocalJobRunner
-
- LOG -
Static variable in class org.apache.hadoop.mapred.pipes.Submitter
-
- LOG -
Static variable in class org.apache.hadoop.mapred.ShuffleExceptionTracker
-
- LOG -
Static variable in class org.apache.hadoop.mapred.TaskController
-
- LOG -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- LOG -
Static variable in class org.apache.hadoop.mapred.TaskTrackerStatus
-
- LOG -
Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
-
- LOG -
Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
-
- LOG -
Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
-
- LOG -
Static variable in class org.apache.hadoop.metrics.MetricsUtil
- Deprecated.
- LOG -
Variable in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- LOG -
Variable in class org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30
-
- LOG -
Variable in class org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31
-
- LOG -
Static variable in class org.apache.hadoop.net.DNS
-
- LOG -
Static variable in class org.apache.hadoop.net.NetworkTopology
-
- LOG -
Static variable in class org.apache.hadoop.security.SaslInputStream
-
- LOG -
Static variable in class org.apache.hadoop.security.SaslRpcClient
-
- LOG -
Static variable in class org.apache.hadoop.security.SaslRpcServer
-
- LOG -
Static variable in class org.apache.hadoop.security.SecurityUtil
-
- LOG -
Static variable in class org.apache.hadoop.security.token.Token
-
- LOG -
Static variable in class org.apache.hadoop.streaming.PipeMapRed
-
- LOG -
Static variable in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- LOG -
Static variable in class org.apache.hadoop.streaming.StreamJob
-
- LOG -
Static variable in class org.apache.hadoop.util.Shell
-
- logFailed(JobID, long, int, int, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Logs job failed event.
- logFailed(TaskAttemptID, long, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Deprecated. Use
JobHistory.MapAttempt.logFailed(TaskAttemptID, long, String, String, String)
- logFailed(TaskAttemptID, long, String, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Log task attempt failed event.
- logFailed(TaskAttemptID, long, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Deprecated. Use
JobHistory.ReduceAttempt.logFailed(TaskAttemptID, long, String, String, String)
- logFailed(TaskAttemptID, long, String, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Log failed reduce task attempt.
- logFailed(TaskID, String, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.Task
- Log job failed event.
- logFailed(TaskID, String, long, String, TaskAttemptID) -
Static method in class org.apache.hadoop.mapred.JobHistory.Task
-
- logFinished(JobID, long, int, int, int, int, Counters, Counters, Counters) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Log job finished.
- logFinished(TaskAttemptID, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Deprecated. Use
JobHistory.MapAttempt.logFinished(TaskAttemptID, long, String, String, String, Counters)
- logFinished(TaskAttemptID, long, String, String, String, Counters) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Log finish time of map task attempt.
- logFinished(TaskAttemptID, long, long, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Deprecated. Use
JobHistory.ReduceAttempt.logFinished(TaskAttemptID, long, long, long, String, String, String, Counters)
- logFinished(TaskAttemptID, long, long, long, String, String, String, Counters) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Log finished event of this task.
- logFinished(TaskID, String, long, Counters) -
Static method in class org.apache.hadoop.mapred.JobHistory.Task
- Log finish time of task.
- login(Configuration, String, String) -
Static method in class org.apache.hadoop.security.SecurityUtil
- If a keytab has been provided, login as that user.
- login(Configuration, String, String, String) -
Static method in class org.apache.hadoop.security.SecurityUtil
- If a keytab has been provided, login as that user.
- login() -
Method in class org.apache.hadoop.security.UserGroupInformation.HadoopLoginModule
-
- logInfo(String) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
-
- logInited(JobID, long, int, int) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Logs launch time of job.
- loginUserFromKeytab(String, String) -
Static method in class org.apache.hadoop.security.UserGroupInformation
- Log a user in from a keytab file.
- loginUserFromKeytabAndReturnUGI(String, String) -
Static method in class org.apache.hadoop.security.UserGroupInformation
- Log a user in from a keytab file.
- logJobInfo(JobID, long, long, int) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Deprecated. Use
JobHistory.JobInfo.logJobInfo(JobID, long, long)
instead.
- logJobInfo(JobID, long, long) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
- logJobPriority(JobID, JobPriority) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Log job's priority.
- logKilled(JobID, long, int, int) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Logs job killed event.
- logKilled(TaskAttemptID, long, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Deprecated. Use
JobHistory.MapAttempt.logKilled(TaskAttemptID, long, String, String, String)
- logKilled(TaskAttemptID, long, String, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Log task attempt killed event.
- logKilled(TaskAttemptID, long, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Deprecated. Use
JobHistory.ReduceAttempt.logKilled(TaskAttemptID, long, String, String, String)
- logKilled(TaskAttemptID, long, String, String, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Log killed reduce task attempt.
- LogLevel - Class in org.apache.hadoop.log
- Change log level in runtime.
- LogLevel() -
Constructor for class org.apache.hadoop.log.LogLevel
-
- LogLevel.Servlet - Class in org.apache.hadoop.log
- A servlet implementation
- LogLevel.Servlet() -
Constructor for class org.apache.hadoop.log.LogLevel.Servlet
-
- logout() -
Method in class org.apache.hadoop.security.UserGroupInformation.HadoopLoginModule
-
- logOutput(String) -
Method in class org.apache.hadoop.mapred.TaskController
-
- LogParser - Class in org.apache.hadoop.contrib.failmon
- This class represents objects that provide log parsing
functionality.
- LogParser(String) -
Constructor for class org.apache.hadoop.contrib.failmon.LogParser
- Create a parser that will read from the specified log file.
- logSpec() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
-
- logStarted(JobID, long, int, int) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Deprecated. Use
JobHistory.JobInfo.logInited(JobID, long, int, int)
and
JobHistory.JobInfo.logStarted(JobID)
- logStarted(JobID) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Logs job as running
- logStarted(TaskAttemptID, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Deprecated. Use
JobHistory.MapAttempt.logStarted(TaskAttemptID, long, String, int, String)
- logStarted(TaskAttemptID, long, String, int, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Deprecated.
- logStarted(TaskAttemptID, long, String, int, String, Locality, Avataar) -
Static method in class org.apache.hadoop.mapred.JobHistory.MapAttempt
- Log start time of this map task attempt.
- logStarted(TaskAttemptID, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Deprecated. Use
JobHistory.ReduceAttempt.logStarted(TaskAttemptID, long, String, int, String)
- logStarted(TaskAttemptID, long, String, int, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Deprecated.
- logStarted(TaskAttemptID, long, String, int, String, Locality, Avataar) -
Static method in class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
- Log start time of Reduce task attempt.
- logStarted(TaskID, String, long, String) -
Static method in class org.apache.hadoop.mapred.JobHistory.Task
- Log start time of task (TIP).
- logSubmitted(JobID, JobConf, String, long) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Deprecated. Use
JobHistory.JobInfo.logSubmitted(JobID, JobConf, String, long, boolean)
instead.
- logSubmitted(JobID, JobConf, String, long, boolean) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
- logThreadInfo(Log, String, long) -
Static method in class org.apache.hadoop.util.ReflectionUtils
- Log the current thread stacks at INFO level.
- logUpdateMasterKey(DelegationKey) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
-
- logUpdates(TaskID, long) -
Static method in class org.apache.hadoop.mapred.JobHistory.Task
- Update the finish time of task.
- LONG -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- LONG_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- LONG_VALUE_MAX -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- LONG_VALUE_MIN -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- LONG_VALUE_SUM -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- LongSumReducer<K> - Class in org.apache.hadoop.mapred.lib
- A
Reducer
that sums long values. - LongSumReducer() -
Constructor for class org.apache.hadoop.mapred.lib.LongSumReducer
-
- LongSumReducer<KEY> - Class in org.apache.hadoop.mapreduce.lib.reduce
-
- LongSumReducer() -
Constructor for class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer
-
- LongTypeID -
Static variable in class org.apache.hadoop.record.meta.TypeID
-
- LongValueMax - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that maintain the maximum of
a sequence of long values.
- LongValueMax() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
- the default constructor
- LongValueMin - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that maintain the minimum of
a sequence of long values.
- LongValueMin() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
- the default constructor
- LongValueSum - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that sums up
a sequence of long values.
- LongValueSum() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
- the default constructor
- LongWritable - Class in org.apache.hadoop.io
- A WritableComparable for longs.
- LongWritable() -
Constructor for class org.apache.hadoop.io.LongWritable
-
- LongWritable(long) -
Constructor for class org.apache.hadoop.io.LongWritable
-
- LongWritable.Comparator - Class in org.apache.hadoop.io
- A Comparator optimized for LongWritable.
- LongWritable.Comparator() -
Constructor for class org.apache.hadoop.io.LongWritable.Comparator
-
- LongWritable.DecreasingComparator - Class in org.apache.hadoop.io
- A decreasing Comparator optimized for LongWritable.
- LongWritable.DecreasingComparator() -
Constructor for class org.apache.hadoop.io.LongWritable.DecreasingComparator
-
- lowerBound(byte[]) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is greater than or equal
to the input key.
- lowerBound(byte[], int, int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is greater than or equal
to the input key.
- lowerBound(List<? extends T>, T, Comparator<? super T>) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Lower bound binary search.
- lowerBound(List<? extends Comparable<? super T>>, T) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Lower bound binary search.
- LT_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- LuceneUtil - Class in org.apache.hadoop.contrib.index.lucene
- This class copies some methods from Lucene's SegmentInfos since that class
is not public.
- LuceneUtil() -
Constructor for class org.apache.hadoop.contrib.index.lucene.LuceneUtil
-
M
- main(String[]) -
Static method in class org.apache.hadoop.conf.Configuration
- For debugging.
- main(String[]) -
Static method in class org.apache.hadoop.contrib.failmon.Continuous
-
- main(String[]) -
Static method in class org.apache.hadoop.contrib.failmon.HDFSMerger
-
- main(String[]) -
Static method in class org.apache.hadoop.contrib.failmon.OfflineAnonymizer
-
- main(String[]) -
Static method in class org.apache.hadoop.contrib.failmon.RunOnce
-
- main(String[]) -
Static method in class org.apache.hadoop.contrib.index.main.UpdateIndex
- The main() method
- main(String[]) -
Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.AggregateWordCount
- The main driver for word count map/reduce program.
- main(String[]) -
Static method in class org.apache.hadoop.examples.AggregateWordHistogram
- The main driver for word count map/reduce program.
- main(String[]) -
Static method in class org.apache.hadoop.examples.dancing.DistributedPentomino
- Launch the solver on 9x10 board and the one sided pentominos.
- main(String[]) -
Static method in class org.apache.hadoop.examples.dancing.OneSidedPentomino
- Solve the 3x30 puzzle.
- main(String[]) -
Static method in class org.apache.hadoop.examples.dancing.Pentomino
- Solve the 6x10 pentomino puzzle.
- main(String[]) -
Static method in class org.apache.hadoop.examples.dancing.Sudoku
- Solves a set of sudoku puzzles.
- main(String[]) -
Static method in class org.apache.hadoop.examples.DBCountPageView
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.ExampleDriver
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.Grep
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.Join
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.MultiFileWordCount
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.PiEstimator
- main method for running it as a stand alone command.
- main(String[]) -
Static method in class org.apache.hadoop.examples.RandomTextWriter
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.RandomWriter
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.SecondarySort
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.SleepJob
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.Sort
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.terasort.TeraGen
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.terasort.TeraSort
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.terasort.TeraValidate
-
- main(String[]) -
Static method in class org.apache.hadoop.examples.WordCount
-
- main(String[]) -
Static method in class org.apache.hadoop.fs.DF
-
- main(String[]) -
Static method in class org.apache.hadoop.fs.DU
-
- main(String[]) -
Static method in class org.apache.hadoop.fs.FsShell
- main() has some simple utility methods
- main(String[]) -
Static method in class org.apache.hadoop.fs.s3.MigrationTool
-
- main(String[]) -
Static method in class org.apache.hadoop.fs.Trash
- Run an emptier.
- main(String[]) -
Static method in class org.apache.hadoop.http.HtmlQuoting
-
- main(String[]) -
Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
- A little test program.
- main(String[]) -
Static method in class org.apache.hadoop.io.file.tfile.TFile
- Dumping the TFile information.
- main(String[]) -
Static method in class org.apache.hadoop.io.MapFile
-
- main(String[]) -
Static method in class org.apache.hadoop.log.LogLevel
- A command line implementation
- main(String[]) -
Static method in class org.apache.hadoop.mapred.IsolationRunner
- Run a single task.
- main(String[]) -
Static method in class org.apache.hadoop.mapred.JobClient
-
- main(String[]) -
Static method in class org.apache.hadoop.mapred.JobHistoryServer
- Start job history server as an independent process
- main(String[]) -
Static method in class org.apache.hadoop.mapred.JobLocalizer
-
- main(String[]) -
Static method in class org.apache.hadoop.mapred.JobTracker
- Start the JobTracker process.
- main(String[]) -
Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
- create and run an Aggregate based map/reduce job.
- main(String[]) -
Static method in class org.apache.hadoop.mapred.lib.InputSampler
-
- main(String[]) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Submit a pipes job based on the command line arguments.
- main(String[]) -
Static method in class org.apache.hadoop.mapred.TaskLogsTruncater
-
- main(String[]) -
Static method in class org.apache.hadoop.mapred.TaskTracker
- Start the TaskTracker, point toward the indicated JobTracker
- main(String[]) -
Static method in class org.apache.hadoop.mapred.tools.MRAdmin
-
- main(String[]) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
-
- main(String[]) -
Static method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- main(String[]) -
Static method in class org.apache.hadoop.security.KerberosName
-
- main(String[]) -
Static method in class org.apache.hadoop.security.UserGroupInformation
- A test method to print out the current user's UGI.
- main(String[]) -
Static method in class org.apache.hadoop.streaming.DumpTypedBytes
-
- main(String[]) -
Static method in class org.apache.hadoop.streaming.HadoopStreaming
-
- main(String[]) -
Static method in class org.apache.hadoop.streaming.JarBuilder
- Test program
- main(String[]) -
Static method in class org.apache.hadoop.streaming.LoadTypedBytes
-
- main(String[]) -
Static method in class org.apache.hadoop.streaming.PathFinder
-
- main(String[]) -
Static method in class org.apache.hadoop.util.hash.JenkinsHash
- Compute the hash of the specified file
- main(String[]) -
Static method in class org.apache.hadoop.util.LinuxResourceCalculatorPlugin
- Test the
LinuxResourceCalculatorPlugin
- main(String[]) -
Static method in class org.apache.hadoop.util.PlatformName
-
- main(String[]) -
Static method in class org.apache.hadoop.util.PrintJarMainClass
-
- main(String[]) -
Static method in class org.apache.hadoop.util.RunJar
- Run a Hadoop job jar.
- main(String[]) -
Static method in class org.apache.hadoop.util.VersionInfo
-
- makeClassLoader(ClassLoader) -
Method in class org.apache.hadoop.filecache.TaskDistributedCacheManager
- Creates a class loader that includes the designated
files and archives.
- makeCompactString() -
Method in class org.apache.hadoop.mapred.Counters
- Convert a counters object into a single line that is easy to parse.
- makeComparator(String) -
Static method in class org.apache.hadoop.io.file.tfile.TFile
- Make a raw comparator from a string name.
- makeEscapedCompactString() -
Method in class org.apache.hadoop.mapred.Counters.Counter
- Returns the compact stringified version of the counter in the format
[(actual-name)(display-name)(value)]
- makeEscapedCompactString() -
Method in class org.apache.hadoop.mapred.Counters.Group
- Returns the compact stringified version of the group in the format
{(actual-name)(display-name)(value)[][][]} where [] are compact strings for the
counters within.
- makeEscapedCompactString() -
Method in class org.apache.hadoop.mapred.Counters
- Represent the counter in a textual format that can be converted back to
its object form
- makeJavaCommand(Class, String[]) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- makeLock(String) -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- makeQualified(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Make sure that a path specifies a FileSystem.
- makeQualified(Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Make sure that a path specifies a FileSystem.
- makeQualified(Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
-
- makeQualified(FileSystem) -
Method in class org.apache.hadoop.fs.Path
- Returns a qualified path object.
- makeQualified(URI, Path) -
Method in class org.apache.hadoop.fs.Path
- Returns a qualified path object.
- makeShellPath(String) -
Static method in class org.apache.hadoop.fs.FileUtil
- Convert a os-native filename to a path that works for the shell.
- makeShellPath(File) -
Static method in class org.apache.hadoop.fs.FileUtil
- Convert a os-native filename to a path that works for the shell.
- makeShellPath(File, boolean) -
Static method in class org.apache.hadoop.fs.FileUtil
- Convert a os-native filename to a path that works for the shell.
- makeSocketAddr(String, int) -
Static method in class org.apache.hadoop.net.NetUtils
- Create a socket address with the given host and port.
- map(DocumentID, DocumentAndOp, OutputCollector<DocumentID, DocumentAndOp>, Reporter) -
Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
-
- map(DocumentID, LineDocTextAndOp, OutputCollector<DocumentID, DocumentAndOp>, Reporter) -
Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
-
- map(K, V, OutputCollector<Shard, IntermediateForm>, Reporter) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
- Map a key-value pair to a shard-and-intermediate form pair.
- map(Object, Object, OutputCollector, Reporter) -
Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- map(Object, Object, OutputCollector, Reporter) -
Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- map(WritableComparable, Text, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.examples.dancing.DistributedPentomino.PentMap
- Break the prefix string into moves (a sequence of integer row ids that
will be selected for each column in order).
- map(MultiFileWordCount.WordOffset, Text, OutputCollector<Text, LongWritable>, Reporter) -
Method in class org.apache.hadoop.examples.MultiFileWordCount.MapClass
-
- map(LongWritable, LongWritable, OutputCollector<BooleanWritable, LongWritable>, Reporter) -
Method in class org.apache.hadoop.examples.PiEstimator.PiMapper
- Map method.
- map(LongWritable, Text, Mapper<LongWritable, Text, SecondarySort.IntPair, IntWritable>.Context) -
Method in class org.apache.hadoop.examples.SecondarySort.MapClass
-
- map(IntWritable, IntWritable, OutputCollector<IntWritable, NullWritable>, Reporter) -
Method in class org.apache.hadoop.examples.SleepJob
-
- map(LongWritable, NullWritable, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
-
- map(Object, Text, Mapper<Object, Text, Text, IntWritable>.Context) -
Method in class org.apache.hadoop.examples.WordCount.TokenizerMapper
-
- map(K1, V1, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
- Do nothing.
- map(K1, V1, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
- the map function.
- map(K1, V1, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
- Do nothing.
- map(Object, Object, OutputCollector, Reporter) -
Method in class org.apache.hadoop.mapred.lib.ChainMapper
- Chains the
map(...)
methods of the Mappers in the chain.
- map(K1, V1, OutputCollector<K2, V2>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.DelegatingMapper
-
- map(K, V, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
- The identify function.
- map(K, V, OutputCollector<K, V>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.IdentityMapper
- The identify function.
- map(K, V, OutputCollector<V, K>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.InverseMapper
- The inverse function.
- map(K, Text, OutputCollector<Text, LongWritable>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.RegexMapper
-
- map(K, Text, OutputCollector<Text, LongWritable>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.TokenCountMapper
-
- map(K1, V1, OutputCollector<K2, V2>, Reporter) -
Method in interface org.apache.hadoop.mapred.Mapper
- Maps a single input key/value pair into an intermediate key/value pair.
- map(K, V, Mapper<K, V, Text, Text>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
- The identify function.
- map(K, V, Mapper<K, V, V, K>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.map.InverseMapper
- The inverse function.
- map(Object, Text, Mapper<Object, Text, Text, IntWritable>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper
-
- map(KEYIN, VALUEIN, Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Mapper
- Called once for each key/value pair in the input split.
- Map() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- MAP -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- map(Object, Object, OutputCollector, Reporter) -
Method in class org.apache.hadoop.streaming.PipeMapper
-
- MAP_CLASS_ATTR -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- MAP_OUTPUT_KEY_VALUE_SPEC -
Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
-
- MAP_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- mapCmd_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
- The context that is given to the
Mapper
. - MapContext(Configuration, TaskAttemptID, RecordReader<KEYIN, VALUEIN>, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, InputSplit) -
Constructor for class org.apache.hadoop.mapreduce.MapContext
-
- mapDebugSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- MapFile - Class in org.apache.hadoop.io
- A file-based map from keys to values.
- MapFile() -
Constructor for class org.apache.hadoop.io.MapFile
-
- MapFile.Reader - Class in org.apache.hadoop.io
- Provide access to an existing map.
- MapFile.Reader(FileSystem, String, Configuration) -
Constructor for class org.apache.hadoop.io.MapFile.Reader
- Construct a map reader for the named map.
- MapFile.Reader(FileSystem, String, WritableComparator, Configuration) -
Constructor for class org.apache.hadoop.io.MapFile.Reader
- Construct a map reader for the named map using the named comparator.
- MapFile.Reader(FileSystem, String, WritableComparator, Configuration, boolean) -
Constructor for class org.apache.hadoop.io.MapFile.Reader
- Hook to allow subclasses to defer opening streams until further
initialization is complete.
- MapFile.Writer - Class in org.apache.hadoop.io
- Writes a new map.
- MapFile.Writer(Configuration, FileSystem, String, Class<? extends WritableComparable>, Class) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map for keys of the named class.
- MapFile.Writer(Configuration, FileSystem, String, Class<? extends WritableComparable>, Class, SequenceFile.CompressionType, Progressable) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map for keys of the named class.
- MapFile.Writer(Configuration, FileSystem, String, Class<? extends WritableComparable>, Class, SequenceFile.CompressionType, CompressionCodec, Progressable) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map for keys of the named class.
- MapFile.Writer(Configuration, FileSystem, String, Class<? extends WritableComparable>, Class, SequenceFile.CompressionType) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map for keys of the named class.
- MapFile.Writer(Configuration, FileSystem, String, WritableComparator, Class) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map using the named key comparator.
- MapFile.Writer(Configuration, FileSystem, String, WritableComparator, Class, SequenceFile.CompressionType) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map using the named key comparator.
- MapFile.Writer(Configuration, FileSystem, String, WritableComparator, Class, SequenceFile.CompressionType, Progressable) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map using the named key comparator.
- MapFile.Writer(Configuration, FileSystem, String, WritableComparator, Class, SequenceFile.CompressionType, CompressionCodec, Progressable) -
Constructor for class org.apache.hadoop.io.MapFile.Writer
- Create the named map using the named key comparator.
- MapFileOutputFormat - Class in org.apache.hadoop.mapred
- An
OutputFormat
that writes MapFile
s. - MapFileOutputFormat() -
Constructor for class org.apache.hadoop.mapred.MapFileOutputFormat
-
- mapOutputFile -
Variable in class org.apache.hadoop.mapred.Task
-
- mapOutputLost(TaskAttemptID, String) -
Method in class org.apache.hadoop.mapred.TaskTracker
- A completed map task's output has been lost.
- Mapper<K1,V1,K2,V2> - Interface in org.apache.hadoop.mapred
- Maps input key/value pairs to a set of intermediate key/value pairs.
- Mapper<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
- Maps input key/value pairs to a set of intermediate key/value pairs.
- Mapper() -
Constructor for class org.apache.hadoop.mapreduce.Mapper
-
- Mapper.Context - Class in org.apache.hadoop.mapreduce
-
- Mapper.Context(Configuration, TaskAttemptID, RecordReader<KEYIN, VALUEIN>, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, InputSplit) -
Constructor for class org.apache.hadoop.mapreduce.Mapper.Context
-
- mapProgress() -
Method in class org.apache.hadoop.mapred.JobStatus
-
- mapProgress() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Get the progress of the job's map-tasks, as a float between 0.0
and 1.0.
- mapProgress() -
Method in class org.apache.hadoop.mapreduce.Job
- Get the progress of the job's map-tasks, as a float between 0.0
and 1.0.
- MAPRED_CLUSTER_MAP_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- MAPRED_CLUSTER_MAX_MAP_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- MAPRED_CLUSTER_MAX_REDUCE_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- MAPRED_CLUSTER_REDUCE_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobTracker
-
- MAPRED_HISTORY_SERVER_EMBEDDED -
Static variable in class org.apache.hadoop.mapred.JobHistoryServer
-
- MAPRED_HISTORY_SERVER_HTTP_ADDRESS -
Static variable in class org.apache.hadoop.mapred.JobHistoryServer
-
- MAPRED_JOB_MAP_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- MAPRED_LOCAL_DIR_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
- Property name for the configuration property mapred.local.dir
- MAPRED_MAP_TASK_ENV -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the maximum virutal memory available to the
map tasks.
- MAPRED_MAP_TASK_JAVA_OPTS -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the java command line options for the map tasks.
- MAPRED_MAP_TASK_ULIMIT -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the maximum virutal memory available to the
map tasks (in kilo-bytes).
- MAPRED_REDUCE_TASK_ENV -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the maximum virutal memory available to the
reduce tasks.
- MAPRED_REDUCE_TASK_JAVA_OPTS -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the java command line options for the reduce tasks.
- MAPRED_REDUCE_TASK_ULIMIT -
Static variable in class org.apache.hadoop.mapred.JobConf
- Configuration key to set the maximum virutal memory available to the
reduce tasks (in kilo-bytes).
- MAPRED_TASK_DEFAULT_MAXVMEM_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated.
- MAPRED_TASK_ENV -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated. Use
JobConf.MAPRED_MAP_TASK_ENV
or
JobConf.MAPRED_REDUCE_TASK_ENV
- MAPRED_TASK_JAVA_OPTS -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated. Use
JobConf.MAPRED_MAP_TASK_JAVA_OPTS
or
JobConf.MAPRED_REDUCE_TASK_JAVA_OPTS
- MAPRED_TASK_MAXPMEM_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated.
- MAPRED_TASK_MAXVMEM_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated. Use
JobConf.MAPRED_JOB_MAP_MEMORY_MB_PROPERTY
and
JobConf.MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY
- MAPRED_TASK_ULIMIT -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated. Use
JobConf.MAPRED_MAP_TASK_ULIMIT
or
JobConf.MAPRED_REDUCE_TASK_ULIMIT
- mapRedFinished() -
Method in class org.apache.hadoop.streaming.PipeMapRed
-
- MAPREDUCE_CLIENT_RETRY_POLICY_ENABLED_DEFAULT -
Static variable in class org.apache.hadoop.mapred.JobClient
-
- MAPREDUCE_CLIENT_RETRY_POLICY_ENABLED_KEY -
Static variable in class org.apache.hadoop.mapred.JobClient
-
- MAPREDUCE_CLIENT_RETRY_POLICY_SPEC_DEFAULT -
Static variable in class org.apache.hadoop.mapred.JobClient
-
- MAPREDUCE_CLIENT_RETRY_POLICY_SPEC_KEY -
Static variable in class org.apache.hadoop.mapred.JobClient
-
- MAPREDUCE_DELEGATION_KIND -
Static variable in class org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier
-
- MAPREDUCE_JOBEND_NOTIFICATION_TIMEOUT_DEFAULT -
Static variable in class org.apache.hadoop.mapred.JobEndNotifier
-
- MAPREDUCE_RECOVER_JOB -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- MapReduceBase - Class in org.apache.hadoop.mapred
- Base class for
Mapper
and Reducer
implementations. - MapReduceBase() -
Constructor for class org.apache.hadoop.mapred.MapReduceBase
-
- MapReducePolicyProvider - Class in org.apache.hadoop.mapred
PolicyProvider
for Map-Reduce protocols.- MapReducePolicyProvider() -
Constructor for class org.apache.hadoop.mapred.MapReducePolicyProvider
-
- MapRunnable<K1,V1,K2,V2> - Interface in org.apache.hadoop.mapred
- Expert: Generic interface for
Mapper
s. - MapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.mapred
- Default
MapRunnable
implementation. - MapRunner() -
Constructor for class org.apache.hadoop.mapred.MapRunner
-
- MapTaskCompletionEventsUpdate - Class in org.apache.hadoop.mapred
- A class that represents the communication between the tasktracker and child
tasks w.r.t the map task completion events.
- MapTaskCompletionEventsUpdate() -
Constructor for class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
-
- MapTaskCompletionEventsUpdate(TaskCompletionEvent[], boolean) -
Constructor for class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
-
- MapTypeID - Class in org.apache.hadoop.record.meta
- Represents typeID for a Map
- MapTypeID(TypeID, TypeID) -
Constructor for class org.apache.hadoop.record.meta.MapTypeID
-
- MapWritable - Class in org.apache.hadoop.io
- A Writable Map.
- MapWritable() -
Constructor for class org.apache.hadoop.io.MapWritable
- Default constructor.
- MapWritable(MapWritable) -
Constructor for class org.apache.hadoop.io.MapWritable
- Copy constructor.
- mark(int) -
Method in class org.apache.hadoop.fs.FSInputChecker
-
- mark(int) -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- mark(int) -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- markJobLogsForDeletion(long, int, JobID) -
Method in class org.apache.hadoop.mapred.UserLogCleaner
- Adds job user-log directory to cleanup thread to delete logs after user-log
retain hours.
- markSupported() -
Method in class org.apache.hadoop.fs.FSInputChecker
-
- markSupported() -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- markSupported() -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- markSupported() -
Method in class org.apache.hadoop.security.SaslInputStream
- Tests if this input stream supports the
mark
and
reset
methods, which it does not.
- matches(CharSequence) -
Method in class org.apache.hadoop.fs.GlobPattern
- Match input against the compiled glob pattern
- matches(String) -
Static method in class org.apache.hadoop.fs.shell.Count
- Check if a command is the count command
- max() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- max() -
Method in class org.apache.hadoop.metrics2.util.SampleStat.MinMax
-
- MAX_ALPHA_SIZE -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- MAX_BLOCKSIZE -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- The maximum supported blocksize == 9.
- MAX_CODE_LEN -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- MAX_COUNTER_LIMIT -
Static variable in class org.apache.hadoop.mapred.Counters
- limit on counters
- MAX_GROUP_LIMIT -
Static variable in class org.apache.hadoop.mapred.Counters
- the max groups allowed
- MAX_OUTPUT_LENGTH -
Static variable in class org.apache.hadoop.contrib.failmon.Environment
-
- MAX_SELECTORS -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- MAX_TRIE_DEPTH -
Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
-
- MAX_USER_JOBCONF_SIZE_KEY -
Static variable in class org.apache.hadoop.mapred.JobTracker
- the config key for max user jobconf size
- maxAvailableData -
Variable in class org.apache.hadoop.io.compress.CompressionInputStream
-
- MAXIMUM_FP -
Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
- MaximumFP Selection.
- maxNextCharInd -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- MBeans - Class in org.apache.hadoop.metrics2.util
- This util class provides a method to register an MBean using
our standard naming convention as described in the doc
for {link
MBeans.register(String, String, Object)
- MBeans() -
Constructor for class org.apache.hadoop.metrics2.util.MBeans
-
- mBeanServer -
Variable in class org.apache.hadoop.jmx.JMXJsonServlet
- MBean server.
- MBeanUtil - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MBeans
. - MBeanUtil() -
Constructor for class org.apache.hadoop.metrics.util.MBeanUtil
- Deprecated.
- MD5_LEN -
Static variable in class org.apache.hadoop.io.MD5Hash
-
- MD5_LEN -
Static variable in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
-
- MD5_LEN -
Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
-
- MD5Hash - Class in org.apache.hadoop.io
- A Writable for MD5 hash values.
- MD5Hash() -
Constructor for class org.apache.hadoop.io.MD5Hash
- Constructs an MD5Hash.
- MD5Hash(String) -
Constructor for class org.apache.hadoop.io.MD5Hash
- Constructs an MD5Hash from a hex string.
- MD5Hash(byte[]) -
Constructor for class org.apache.hadoop.io.MD5Hash
- Constructs an MD5Hash with a specified value.
- MD5Hash.Comparator - Class in org.apache.hadoop.io
- A WritableComparator optimized for MD5Hash keys.
- MD5Hash.Comparator() -
Constructor for class org.apache.hadoop.io.MD5Hash.Comparator
-
- MD5MD5CRC32FileChecksum - Class in org.apache.hadoop.fs
- MD5 of MD5 of CRC32.
- MD5MD5CRC32FileChecksum() -
Constructor for class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Same as this(0, 0, null)
- MD5MD5CRC32FileChecksum(int, long, MD5Hash) -
Constructor for class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Create a MD5FileChecksum
- mean() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- mechanismName -
Variable in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
-
- membershipTest(Key) -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- membershipTest(Key) -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- membershipTest(Key) -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- membershipTest(Key) -
Method in class org.apache.hadoop.util.bloom.Filter
- Determines wether a specified key belongs to this filter.
- MemoryCalculatorPlugin - Class in org.apache.hadoop.util
- Deprecated. Use
ResourceCalculatorPlugin
instead - MemoryCalculatorPlugin() -
Constructor for class org.apache.hadoop.util.MemoryCalculatorPlugin
- Deprecated.
- merge(List<SequenceFile.Sorter.SegmentDescriptor>, Path) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Merges the list of segments of type
SegmentDescriptor
- merge(Path[], boolean, Path) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Merges the contents of files passed in Path[] using a max factor value
that is already set
- merge(Path[], boolean, int, Path) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Merges the contents of files passed in Path[]
- merge(Path[], Path, boolean) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Merges the contents of files passed in Path[]
- merge(Path[], Path) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Merge the provided files.
- merge(List, List, String) -
Method in class org.apache.hadoop.streaming.JarBuilder
-
- MergeSort - Class in org.apache.hadoop.util
- An implementation of the core algorithm of MergeSort.
- MergeSort(Comparator<IntWritable>) -
Constructor for class org.apache.hadoop.util.MergeSort
-
- mergeSort(int[], int[], int, int) -
Method in class org.apache.hadoop.util.MergeSort
-
- MetaBlockAlreadyExists - Exception in org.apache.hadoop.io.file.tfile
- Exception - Meta Block with the same name already exists.
- MetaBlockDoesNotExist - Exception in org.apache.hadoop.io.file.tfile
- Exception - No such Meta Block with the given name.
- Metric - Class in org.apache.hadoop.metrics2
- The immutable metric
- Metric(String) -
Constructor for class org.apache.hadoop.metrics2.Metric
- Construct the metric with name only
- Metric(String, String) -
Constructor for class org.apache.hadoop.metrics2.Metric
- Construct the metric with a name and a description
- MetricCounter<T extends Number> - Class in org.apache.hadoop.metrics2
- A generic immutable counter metric type
- MetricCounter(String, String) -
Constructor for class org.apache.hadoop.metrics2.MetricCounter
- Construct a counter metric
- MetricGauge<T extends Number> - Class in org.apache.hadoop.metrics2
- A generic immutable gauge metric
- MetricGauge(String, String) -
Constructor for class org.apache.hadoop.metrics2.MetricGauge
- Construct a gauge metric
- MetricMutable - Class in org.apache.hadoop.metrics2.lib
- The mutable metric interface
- MetricMutable(String, String) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutable
- Construct a mutable metric
- MetricMutableCounter<T extends Number> - Class in org.apache.hadoop.metrics2.lib
- The mutable counter (monotonically increasing) metric interface
- MetricMutableCounter(String, String) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableCounter
- Construct the metric with name and description
- MetricMutableCounterInt - Class in org.apache.hadoop.metrics2.lib
- A mutable int counter for implementing metrics sources
- MetricMutableCounterInt(String, String, int) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableCounterInt
- Construct a mutable int counter
- MetricMutableCounterLong - Class in org.apache.hadoop.metrics2.lib
- A mutable long counter
- MetricMutableCounterLong(String, String, long) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableCounterLong
- Construct a mutable long counter
- MetricMutableFactory - Class in org.apache.hadoop.metrics2.lib
- Factory class for mutable metrics
- MetricMutableFactory() -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableFactory
-
- MetricMutableGauge<T extends Number> - Class in org.apache.hadoop.metrics2.lib
- The mutable gauge metric interface
- MetricMutableGauge(String, String) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableGauge
- Construct the metric with name and description
- MetricMutableGaugeInt - Class in org.apache.hadoop.metrics2.lib
- A mutable int gauge
- MetricMutableGaugeInt(String, String, int) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableGaugeInt
- Construct a mutable int gauge metric
- MetricMutableGaugeLong - Class in org.apache.hadoop.metrics2.lib
- A mutable long gauge
- MetricMutableGaugeLong(String, String, long) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableGaugeLong
- Construct a mutable long gauge metric
- MetricMutableStat - Class in org.apache.hadoop.metrics2.lib
- A mutable metric with stats
Useful for keep throughput/latency stats.
- MetricMutableStat(String, String, String, String, boolean) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableStat
- Construct a sample statistics metric
- MetricMutableStat(String, String, String, String) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricMutableStat
- Construct a snapshot stat metric with extended stat off by default
- metrics() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- metrics() -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Get the metrics
- metrics() -
Method in interface org.apache.hadoop.metrics2.MetricsRecord
- Get the metrics of the record
- metrics() -
Method in class org.apache.hadoop.metrics2.util.MetricsCache.Record
-
- MetricsBase - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsBase(String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsBase
- Deprecated.
- MetricsBase(String, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsBase
- Deprecated.
- MetricsBuilder - Interface in org.apache.hadoop.metrics2
- The metrics builder interface
- metricsCache -
Variable in class org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30
-
- MetricsCache - Class in org.apache.hadoop.metrics2.util
- A metrics cache for sinks that don't support sparse updates.
- MetricsCache() -
Constructor for class org.apache.hadoop.metrics2.util.MetricsCache
-
- MetricsCache.Record - Class in org.apache.hadoop.metrics2.util
- Cached record
- MetricsCache.Record() -
Constructor for class org.apache.hadoop.metrics2.util.MetricsCache.Record
-
- MetricsConfigException - Exception in org.apache.hadoop.metrics2.impl
- The metrics configuration runtime exception
- MetricsContext - Interface in org.apache.hadoop.metrics
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsDynamicMBeanBase - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsDynamicMBeanBase(MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
- Deprecated.
- metricsEntrySet() -
Method in class org.apache.hadoop.metrics2.util.MetricsCache.Record
-
- MetricsException - Exception in org.apache.hadoop.metrics
- Deprecated. in favor of
MetricsException
. - MetricsException() -
Constructor for exception org.apache.hadoop.metrics.MetricsException
- Deprecated. Creates a new instance of MetricsException
- MetricsException(String) -
Constructor for exception org.apache.hadoop.metrics.MetricsException
- Deprecated. Creates a new instance of MetricsException
- MetricsException - Exception in org.apache.hadoop.metrics2
- A general metrics exception wrapper
- MetricsException(String) -
Constructor for exception org.apache.hadoop.metrics2.MetricsException
- Construct the exception with a message
- MetricsException(String, Throwable) -
Constructor for exception org.apache.hadoop.metrics2.MetricsException
- Construct the exception with a message and a cause
- MetricsException(Throwable) -
Constructor for exception org.apache.hadoop.metrics2.MetricsException
- Construct the exception with a cause
- MetricsFilter - Class in org.apache.hadoop.metrics2
- The metrics filter interface
- MetricsFilter() -
Constructor for class org.apache.hadoop.metrics2.MetricsFilter
-
- MetricsIntValue - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MetricMutableGaugeInt
. - MetricsIntValue(String, MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsIntValue
- Deprecated. Constructor - create a new metric
- MetricsIntValue(String, MetricsRegistry) -
Constructor for class org.apache.hadoop.metrics.util.MetricsIntValue
- Deprecated. Constructor - create a new metric
- MetricsLongValue - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsLongValue(String, MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsLongValue
- Deprecated. Constructor - create a new metric
- MetricsLongValue(String, MetricsRegistry) -
Constructor for class org.apache.hadoop.metrics.util.MetricsLongValue
- Deprecated. Constructor - create a new metric
- MetricsPlugin - Interface in org.apache.hadoop.metrics2
- A fairly generic plugin interface
- MetricsRecord - Interface in org.apache.hadoop.metrics
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsRecord - Interface in org.apache.hadoop.metrics2
- An immutable snapshot of metrics with a timestamp
- MetricsRecordBuilder - Class in org.apache.hadoop.metrics2
- The metrics record builder interface
- MetricsRecordBuilder() -
Constructor for class org.apache.hadoop.metrics2.MetricsRecordBuilder
-
- MetricsRecordImpl - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsRecordImpl(String, AbstractMetricsContext) -
Constructor for class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Creates a new instance of FileRecord
- MetricsRecordImpl - Class in org.apache.hadoop.metrics2.impl
-
- MetricsRecordImpl(String, long, Iterable<MetricsTag>, Iterable<Metric>) -
Constructor for class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
- Construct a metrics record
- MetricsRegistry - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MetricsRegistry
. - MetricsRegistry() -
Constructor for class org.apache.hadoop.metrics.util.MetricsRegistry
- Deprecated.
- MetricsRegistry - Class in org.apache.hadoop.metrics2.lib
- An optional metrics registry class for creating and maintaining a
collection of MetricsMutables, making writing metrics source easier.
- MetricsRegistry(String) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Construct the registry with a record name
- MetricsRegistry(String, MetricMutableFactory) -
Constructor for class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Construct the registry with a name and a metric factory
- metricsServers -
Variable in class org.apache.hadoop.metrics.ganglia.GangliaContext
- Deprecated.
- MetricsServlet - Class in org.apache.hadoop.metrics
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsServlet() -
Constructor for class org.apache.hadoop.metrics.MetricsServlet
- Deprecated.
- MetricsSink - Interface in org.apache.hadoop.metrics2
- The metrics sink interface
- MetricsSource - Interface in org.apache.hadoop.metrics2
- The metrics source interface
- MetricsSystem - Interface in org.apache.hadoop.metrics2
- The metrics system interface
- MetricsSystem.AbstractCallback - Class in org.apache.hadoop.metrics2
- Convenient abstract class for implementing callback interface
- MetricsSystem.AbstractCallback() -
Constructor for class org.apache.hadoop.metrics2.MetricsSystem.AbstractCallback
-
- MetricsSystem.Callback - Interface in org.apache.hadoop.metrics2
- The metrics system callback interface
- MetricsSystemImpl - Class in org.apache.hadoop.metrics2.impl
- A base class for metrics system singletons
- MetricsSystemImpl(String) -
Constructor for class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
- Construct the metrics system
- MetricsSystemImpl() -
Constructor for class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
- Construct the system but not initializing (read config etc.) it.
- MetricsSystemMXBean - Interface in org.apache.hadoop.metrics2
- The JMX interface to the metrics system
- MetricsTag - Class in org.apache.hadoop.metrics2
- Immutable tag for metrics (for grouping on host/queue/username etc.)
- MetricsTag(String, String, String) -
Constructor for class org.apache.hadoop.metrics2.MetricsTag
- Construct the tag with name, description and value
- MetricsTimeVaryingInt - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MetricMutableCounterInt
. - MetricsTimeVaryingInt(String, MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
- Deprecated. Constructor - create a new metric
- MetricsTimeVaryingInt(String, MetricsRegistry) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
- Deprecated. Constructor - create a new metric
- MetricsTimeVaryingLong - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MetricMutableCounterLong
. - MetricsTimeVaryingLong(String, MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
- Deprecated. Constructor - create a new metric
- MetricsTimeVaryingLong(String, MetricsRegistry) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
- Deprecated. Constructor - create a new metric
- MetricsTimeVaryingRate - Class in org.apache.hadoop.metrics.util
- Deprecated. in favor of
MetricMutableGauge
. - MetricsTimeVaryingRate(String, MetricsRegistry, String) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Constructor - create a new metric
- MetricsTimeVaryingRate(String, MetricsRegistry) -
Constructor for class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Constructor - create a new metric
- MetricsUtil - Class in org.apache.hadoop.metrics
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricsVisitor - Interface in org.apache.hadoop.metrics2
- A visitor interface for metrics
- MetricValue - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - MetricValue(Number, boolean) -
Constructor for class org.apache.hadoop.metrics.spi.MetricValue
- Deprecated. Creates a new instance of MetricValue
- midKey() -
Method in class org.apache.hadoop.io.MapFile.Reader
- Get the key at approximately the middle of the file.
- MigrationTool - Class in org.apache.hadoop.fs.s3
-
This class is a tool for migrating data from an older to a newer version
of an S3 filesystem.
- MigrationTool() -
Constructor for class org.apache.hadoop.fs.s3.MigrationTool
-
- min() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- min() -
Method in class org.apache.hadoop.metrics2.util.SampleStat.MinMax
-
- MIN_BLOCKSIZE -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- The minimum supported blocksize == 1.
- MIN_INTERVAL -
Static variable in class org.apache.hadoop.contrib.failmon.Environment
-
- MIN_INTERVAL -
Static variable in class org.apache.hadoop.contrib.failmon.Executor
-
- MINIMUM_FN -
Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
- MinimumFN Selection.
- minRecWrittenToEnableSkip_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- mkdirs(Path) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- mkdirs(FileSystem, Path, FsPermission) -
Static method in class org.apache.hadoop.fs.FileSystem
- create a directory with the provided permission
The permission of the directory is set to be the provided permission as in
setPermission, not permission&~umask
- mkdirs(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Call
FileSystem.mkdirs(Path, FsPermission)
with default permission.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.FileSystem
- Make the given file and all non-existent parents into
directories.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Make the given file and all non-existent parents into
directories.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.HarFileSystem
- not implemented.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- mkdirs(Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Creates the specified directory hierarchy.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Make the given file and all non-existent parents into
directories.
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- mkdirs(Path, FsPermission) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- mkdirsWithExistsAndPermissionCheck(LocalFileSystem, Path, FsPermission) -
Static method in class org.apache.hadoop.util.DiskChecker
- Create the directory or check permissions if it already exists.
- mkdirsWithExistsCheck(File) -
Static method in class org.apache.hadoop.util.DiskChecker
- The semantics of mkdirsWithExistsCheck method is different from the mkdirs
method provided in the Sun's java.io.File class in the following way:
While creating the non-existent parent directories, this method checks for
the existence of those directories if the mkdir fails at any point (since
that directory might have just been created by some other process).
- modifFmt -
Static variable in class org.apache.hadoop.fs.FsShell
-
- Module() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- MODULE_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- ModuleName() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.CPUParser
- Invokes query() to do the parsing and handles parsing errors.
- monitor(LocalStore) -
Method in class org.apache.hadoop.contrib.failmon.LogParser
- Insert all EventRecords that can be extracted for
the represented hardware component into a LocalStore.
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.LogParser
- Get an array of all EventRecords that can be extracted for
the represented hardware component.
- monitor() -
Method in interface org.apache.hadoop.contrib.failmon.Monitored
- Get an array of all EventRecords that can be extracted for
the represented hardware component.
- monitor(LocalStore) -
Method in interface org.apache.hadoop.contrib.failmon.Monitored
- Inserts all EventRecords that can be extracted for
the represented hardware component into a LocalStore.
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.NICParser
- Invokes query() to do the parsing and handles parsing errors for
each one of the NICs specified in the configuration.
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.SensorsParser
- Invokes query() to do the parsing and handles parsing errors.
- monitor(LocalStore) -
Method in class org.apache.hadoop.contrib.failmon.ShellParser
- Insert all EventRecords that can be extracted for
the represented hardware component into a LocalStore.
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.ShellParser
-
- monitor() -
Method in class org.apache.hadoop.contrib.failmon.SMARTParser
- Invokes query() to do the parsing and handles parsing errors for
each one of the disks specified in the configuration.
- monitor() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
-
- monitorAndPrintJob(JobConf, RunningJob) -
Method in class org.apache.hadoop.mapred.JobClient
- Monitor a job and print status in real-time as progress is made and tasks
fail.
- Monitored - Interface in org.apache.hadoop.contrib.failmon
- Represents objects that monitor specific hardware resources and
can query them to get EventRecords describing the state of these
resources.
- MonitorJob - Class in org.apache.hadoop.contrib.failmon
- This class is a wrapper for a monitoring job.
- MonitorJob(Monitored, String, int) -
Constructor for class org.apache.hadoop.contrib.failmon.MonitorJob
-
- moveFromLocalFile(Path[], Path) -
Method in class org.apache.hadoop.fs.FileSystem
- The src files is on the local disk.
- moveFromLocalFile(Path, Path) -
Method in class org.apache.hadoop.fs.FileSystem
- The src file is on the local disk.
- moveFromLocalFile(Path, Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- moveToLocalFile(Path, Path) -
Method in class org.apache.hadoop.fs.FileSystem
- The src file is under FS, and the dst is on the local disk.
- moveToTrash(Path) -
Method in class org.apache.hadoop.fs.Trash
- Move a file or directory to the current trash directory.
- MR_CLIENTTRACE_FORMAT -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- MR_COMBINE_RECORDS_BEFORE_PROGRESS -
Static variable in class org.apache.hadoop.mapred.Task
-
- MRAdmin - Class in org.apache.hadoop.mapred.tools
- Administrative access to Hadoop Map-Reduce.
- MRAdmin() -
Constructor for class org.apache.hadoop.mapred.tools.MRAdmin
-
- MRAdmin(Configuration) -
Constructor for class org.apache.hadoop.mapred.tools.MRAdmin
-
- msg(String) -
Method in class org.apache.hadoop.streaming.StreamJob
-
- MultiFileInputFormat<K,V> - Class in org.apache.hadoop.mapred
- Deprecated. Use
CombineFileInputFormat
instead - MultiFileInputFormat() -
Constructor for class org.apache.hadoop.mapred.MultiFileInputFormat
- Deprecated.
- MultiFileSplit - Class in org.apache.hadoop.mapred
- Deprecated. Use
CombineFileSplit
instead - MultiFileSplit(JobConf, Path[], long[]) -
Constructor for class org.apache.hadoop.mapred.MultiFileSplit
- Deprecated.
- MultiFileWordCount - Class in org.apache.hadoop.examples
- MultiFileWordCount is an example to demonstrate the usage of
MultiFileInputFormat.
- MultiFileWordCount() -
Constructor for class org.apache.hadoop.examples.MultiFileWordCount
-
- MultiFileWordCount.MapClass - Class in org.apache.hadoop.examples
- This Mapper is similar to the one in
MultiFileWordCount.MapClass
. - MultiFileWordCount.MapClass() -
Constructor for class org.apache.hadoop.examples.MultiFileWordCount.MapClass
-
- MultiFileWordCount.MultiFileLineRecordReader - Class in org.apache.hadoop.examples
- RecordReader is responsible from extracting records from the InputSplit.
- MultiFileWordCount.MultiFileLineRecordReader(Configuration, MultiFileSplit) -
Constructor for class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- MultiFileWordCount.MyInputFormat - Class in org.apache.hadoop.examples
- To use
MultiFileInputFormat
, one should extend it, to return a
(custom) RecordReader
. - MultiFileWordCount.MyInputFormat() -
Constructor for class org.apache.hadoop.examples.MultiFileWordCount.MyInputFormat
-
- MultiFileWordCount.WordOffset - Class in org.apache.hadoop.examples
- This record keeps <filename,offset> pairs.
- MultiFileWordCount.WordOffset() -
Constructor for class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- MultiFilterRecordReader<K extends WritableComparable,V extends Writable> - Class in org.apache.hadoop.mapred.join
- Base class for Composite join returning values derived from multiple
sources, but generally not tuples.
- MultiFilterRecordReader(int, JobConf, int, Class<? extends WritableComparator>) -
Constructor for class org.apache.hadoop.mapred.join.MultiFilterRecordReader
-
- MultiFilterRecordReader.MultiFilterDelegationIterator - Class in org.apache.hadoop.mapred.join
- Proxy the JoinCollector, but include callback to emit.
- MultiFilterRecordReader.MultiFilterDelegationIterator() -
Constructor for class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- MultipleInputs - Class in org.apache.hadoop.mapred.lib
- This class supports MapReduce jobs that have multiple input paths with
a different
InputFormat
and Mapper
for each path - MultipleInputs() -
Constructor for class org.apache.hadoop.mapred.lib.MultipleInputs
-
- MultipleInputs - Class in org.apache.hadoop.mapreduce.lib.input
- This class supports MapReduce jobs that have multiple input paths with
a different
InputFormat
and Mapper
for each path - MultipleInputs() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
-
- MultipleIOException - Exception in org.apache.hadoop.io
- Encapsulate a list of
IOException
into an IOException
- MultipleOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
- This abstract class extends the FileOutputFormat, allowing to write the
output data to different output files.
- MultipleOutputFormat() -
Constructor for class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
- MultipleOutputs - Class in org.apache.hadoop.mapred.lib
- The MultipleOutputs class simplifies writting to additional outputs other
than the job default output via the
OutputCollector
passed to
the map()
and reduce()
methods of the
Mapper
and Reducer
implementations. - MultipleOutputs(JobConf) -
Constructor for class org.apache.hadoop.mapred.lib.MultipleOutputs
- Creates and initializes multiple named outputs support, it should be
instantiated in the Mapper/Reducer configure method.
- MultipleOutputs<KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.output
- The MultipleOutputs class simplifies writing output data
to multiple outputs
- MultipleOutputs(TaskInputOutputContext<?, ?, KEYOUT, VALUEOUT>) -
Constructor for class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
- Creates and initializes multiple outputs support,
it should be instantiated in the Mapper/Reducer setup method.
- MultipleSequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
- This class extends the MultipleOutputFormat, allowing to write the output data
to different output files in sequence file output format.
- MultipleSequenceFileOutputFormat() -
Constructor for class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat
-
- MultipleTextOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
- This class extends the MultipleOutputFormat, allowing to write the output
data to different output files in Text output format.
- MultipleTextOutputFormat() -
Constructor for class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat
-
- MultithreadedMapper<K1,V1,K2,V2> - Class in org.apache.hadoop.mapreduce.lib.map
- Multithreaded implementation for @link org.apache.hadoop.mapreduce.Mapper.
- MultithreadedMapper() -
Constructor for class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
- MultithreadedMapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.mapred.lib
- Multithreaded implementation for @link org.apache.hadoop.mapred.MapRunnable.
- MultithreadedMapRunner() -
Constructor for class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
-
- MURMUR_HASH -
Static variable in class org.apache.hadoop.util.hash.Hash
- Constant to denote
MurmurHash
.
- MurmurHash - Class in org.apache.hadoop.util.hash
- This is a very fast, non-cryptographic hash suitable for general hash-based
lookup.
- MurmurHash() -
Constructor for class org.apache.hadoop.util.hash.MurmurHash
-
- MySQLDataDrivenDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
- A RecordReader that reads records from a MySQL table via DataDrivenDBRecordReader
- MySQLDataDrivenDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) -
Constructor for class org.apache.hadoop.mapreduce.lib.db.MySQLDataDrivenDBRecordReader
-
- MySQLDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
- A RecordReader that reads records from a MySQL table.
- MySQLDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) -
Constructor for class org.apache.hadoop.mapreduce.lib.db.MySQLDBRecordReader
-
N
- N_GROUPS -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- N_ITERS -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- NAME -
Static variable in class org.apache.hadoop.fs.shell.Count
-
- name() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- name -
Variable in class org.apache.hadoop.metrics2.lib.MetricMutable
- name of the metric
- name() -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
-
- name() -
Method in class org.apache.hadoop.metrics2.Metric
- Get the name of the metric
- name() -
Method in interface org.apache.hadoop.metrics2.MetricsRecord
- Get the record name of the metrics
- name() -
Method in class org.apache.hadoop.metrics2.MetricsTag
- Get the name of the tag
- name -
Variable in class org.apache.hadoop.net.NodeBase
-
- NAME_RULES -
Static variable in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- Constant for the configuration property that indicates the Kerberos name
rules for the Kerberos principals.
- NativeCodeLoader - Class in org.apache.hadoop.util
- A helper to load the native hadoop code i.e.
- NativeCodeLoader() -
Constructor for class org.apache.hadoop.util.NativeCodeLoader
-
- NativeIO - Class in org.apache.hadoop.io.nativeio
- JNI wrappers for various native IO-related calls not available in Java.
- NativeIO() -
Constructor for class org.apache.hadoop.io.nativeio.NativeIO
-
- NativeIO.Stat - Class in org.apache.hadoop.io.nativeio
- Result type of the fstat call
- NativeIOException - Exception in org.apache.hadoop.io.nativeio
- An exception generated by a call to the native IO code.
- NativeIOException(String, Errno) -
Constructor for exception org.apache.hadoop.io.nativeio.NativeIOException
-
- NativeS3FileSystem - Class in org.apache.hadoop.fs.s3native
-
A
FileSystem
for reading and writing files stored on
Amazon S3. - NativeS3FileSystem() -
Constructor for class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- NativeS3FileSystem(NativeFileSystemStore) -
Constructor for class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- NATURAL_ORDER -
Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
-
- nbHash -
Variable in class org.apache.hadoop.util.bloom.Filter
- The number of hash function to consider.
- needChecksum() -
Method in class org.apache.hadoop.fs.FSInputChecker
- Return true if there is a need for checksum verification
- needsDictionary() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- needsDictionary() -
Method in interface org.apache.hadoop.io.compress.Decompressor
- Returns
true
if a preset dictionary is needed for decompression.
- needsDictionary() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Returns
false
.
- needsDictionary() -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
- Returns
true
if a preset dictionary is needed for decompression.
- needsDictionary() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- needsInput() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- needsInput() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- needsInput() -
Method in interface org.apache.hadoop.io.compress.Compressor
- Returns true if the input data buffer is empty and
#setInput() should be called to provide more input.
- needsInput() -
Method in interface org.apache.hadoop.io.compress.Decompressor
- Returns true if the input data buffer is empty and
Decompressor.setInput(byte[], int, int)
should be called to
provide more input.
- needsInput() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Returns true if the input data buffer is empty and
#setInput() should be called to provide more input.
- needsInput() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Returns true if the input data buffer is empty and
SnappyDecompressor.setInput(byte[], int, int)
should be called to
provide more input.
- needsInput() -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
- Returns true if the input data buffer is empty and
Decompressor.setInput(byte[], int, int)
should be called to
provide more input.
- needsInput() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- needsInput() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- needsQuoting(byte[], int, int) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Does the given string need to be quoted?
- needsQuoting(String) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Does the given string need to be quoted?
- needsTaskCommit(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- needsTaskCommit(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- Check whether task needs a commit
- needsTaskCommit(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- This method implements the new interface by calling the old method.
- needsTaskCommit(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
- Did this task write any files in the work directory?
- needsTaskCommit(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.OutputCommitter
- Check whether task needs a commit
- NEGOTIATE -
Static variable in class org.apache.hadoop.security.authentication.client.KerberosAuthenticator
- HTTP header prefix used by the SPNEGO client/server endpoints during an authentication sequence.
- NetgroupCache - Class in org.apache.hadoop.security
- Class that caches the netgroups and inverts group-to-user map
to user-to-group map
- NetgroupCache() -
Constructor for class org.apache.hadoop.security.NetgroupCache
-
- netgroupToUsersMap -
Static variable in class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
-
- netgroupToUsersMapUpdated -
Static variable in class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
-
- netlock -
Variable in class org.apache.hadoop.net.NetworkTopology
-
- NetUtils - Class in org.apache.hadoop.net
-
- NetUtils() -
Constructor for class org.apache.hadoop.net.NetUtils
-
- NetworkTopology - Class in org.apache.hadoop.net
- The class represents a cluster of computer with a tree hierarchical
network topology.
- NetworkTopology() -
Constructor for class org.apache.hadoop.net.NetworkTopology
-
- NetworkTopology.InvalidTopologyException - Exception in org.apache.hadoop.net
-
- NetworkTopology.InvalidTopologyException(String) -
Constructor for exception org.apache.hadoop.net.NetworkTopology.InvalidTopologyException
-
- NetworkTopologyWithNodeGroup - Class in org.apache.hadoop.net
- The class extends NetworkTopology to represents a cluster of computer with
a 4-layers hierarchical network topology.
- NetworkTopologyWithNodeGroup() -
Constructor for class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
-
- newCounter(String, String, int) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable integer counter
- newCounter(String, String, long) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable long integer counter
- newCounter(String, String, int) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable integer counter
- newCounter(String, String, long) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable long integer counter
- newCounterInt(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable integer counter with name only.
- newCounterLong(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable long integer counter with a name
Usually gets overridden.
- newDataChecksum(int, int) -
Static method in class org.apache.hadoop.util.DataChecksum
-
- newDataChecksum(byte[], int) -
Static method in class org.apache.hadoop.util.DataChecksum
- Creates a DataChecksum from HEADER_LEN bytes from arr[offset].
- newDataChecksum(DataInputStream) -
Static method in class org.apache.hadoop.util.DataChecksum
- This constructucts a DataChecksum by reading HEADER_LEN bytes from
input stream in
- newGauge(String, String, int) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable integer gauge
- newGauge(String, String, long) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable long integer gauge
- newGauge(String, String, int) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable integer gauge
- newGauge(String, String, long) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable long integer gauge
- newGaugeInt(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable integer gauge with name only.
- newGaugeLong(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable long integer gauge with name only.
- newInstance(Class<? extends Writable>, Configuration) -
Static method in class org.apache.hadoop.io.WritableFactories
- Create a new instance of a class with a defined factory.
- newInstance(Class<? extends Writable>) -
Static method in class org.apache.hadoop.io.WritableFactories
- Create a new instance of a class with a defined factory.
- newInstance() -
Method in interface org.apache.hadoop.io.WritableFactory
- Return a new instance.
- newInstance(Class<T>, Configuration) -
Static method in class org.apache.hadoop.util.ReflectionUtils
- Create an object for the given class and initialize it from conf
- newKey() -
Method in class org.apache.hadoop.io.WritableComparator
- Construct a new
WritableComparable
instance.
- newMetric(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a new mutable metric by name
Usually overridden by app specific factory
- newRecord(String) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Subclasses should override this if they subclass MetricsRecordImpl.
- newRecord(String) -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- newServerSocket(String, int, int) -
Method in class org.apache.hadoop.security.Krb5AndCertsSslSocketConnector
-
- newStat(String, String, String, String, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable stat metric
- newStat(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableFactory
- Create a mutable stat metric with name only.
- newStat(String, String, String, String, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable metric with stats
- newStat(String, String, String, String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable metric with stats
- newStat(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Create a mutable metric with stats using the name only
- newTaskDistributedCacheManager(JobID, Configuration) -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
-
- newToken(int) -
Static method in class org.apache.hadoop.record.compiler.generated.Token
- Returns a new Token object, by default.
- next(DocumentID, LineDocTextAndOp) -
Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- next() -
Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- next(MultiFileWordCount.WordOffset, Text) -
Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- next(Writable) -
Method in class org.apache.hadoop.io.ArrayFile.Reader
- Read and return the next value in the file.
- next(WritableComparable, Writable) -
Method in class org.apache.hadoop.io.MapFile.Reader
- Read the next key/value pair in the map into
key
and
val
.
- next(Writable) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read the next key in the file into
key
, skipping its
value.
- next(Writable, Writable) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read the next key/value pair in the file into
key
and
val
.
- next(DataOutputBuffer) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Deprecated. Call
SequenceFile.Reader.nextRaw(DataOutputBuffer,SequenceFile.ValueBytes)
.
- next(Object) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read the next key in the file, skipping its
value.
- next() -
Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
- Sets up the current key and value (for getKey and getValue)
- next(WritableComparable) -
Method in class org.apache.hadoop.io.SetFile.Reader
- Read the next key in a set into
key
.
- next(X) -
Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- next(TupleWritable) -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- next(K, TupleWritable) -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader
- Emit the next set of key, value pairs as defined by the child
RecordReaders and operation associated with this composite RR.
- next(V) -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- next(K, V) -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
- Reads the next key/value pair from the input for processing.
- next(U) -
Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- next(T) -
Method in interface org.apache.hadoop.mapred.join.ResetableIterator
- Assign next value to actual.
- next(X) -
Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- next() -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Read the next k,v pair into the head of this object; return true iff
the RR and this are exhausted.
- next(K, U) -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Write key-value pair at the head of this stream to the objects provided;
get next key-value pair from proxied RR.
- next(Text, Text) -
Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
- Read key/value pair in a line.
- next(K, V) -
Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- next(LongWritable, T) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
- Reads the next key/value pair from the input for processing.
- next(LongWritable, Text) -
Method in class org.apache.hadoop.mapred.LineRecordReader
- Read a line.
- next() -
Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
- Sets up the current key and value (for getKey and getValue).
- next(K, V) -
Method in interface org.apache.hadoop.mapred.RecordReader
- Reads the next key/value pair from the input for processing.
- next(BytesWritable, BytesWritable) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
- Read raw bytes from a SequenceFile.
- next(Text, Text) -
Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
- Read key/value pair in a line.
- next(K, V) -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- next(K) -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- next() -
Method in class org.apache.hadoop.mapred.Task.CombineValuesIterator
-
- next(LongWritable, T) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
- Deprecated. Use
DBRecordReader.nextKeyValue()
- next() -
Method in class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
-
- next() -
Method in class org.apache.hadoop.metrics2.util.TryIterator
-
- next -
Variable in class org.apache.hadoop.record.compiler.generated.Token
- A reference to the next regular (non-special) token from the input
stream.
- next(Text, Text) -
Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
- Read a record.
- next(Text, Text) -
Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- nextKey() -
Method in class org.apache.hadoop.mapreduce.ReduceContext
- Start processing next unique key.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
- Read the next key, value pair.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingRecordReader
-
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
- Read key/value pair in a line.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
- Read raw bytes from a SequenceFile.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
- Read key/value pair in a line.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.MapContext
-
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.RecordReader
- Read the next key, value pair.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.ReduceContext
- Advance to the next key/value pair.
- nextKeyValue() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
- Advance to the next key, value pair, returning null if at end.
- nextRaw(DataOutputBuffer, SequenceFile.ValueBytes) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read 'raw' records.
- nextRawKey(DataOutputBuffer) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read 'raw' keys.
- nextRawKey() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
- Fills up the rawKey object with the key returned by the Reader
- nextRawValue(SequenceFile.ValueBytes) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Read 'raw' values.
- nextRawValue(SequenceFile.ValueBytes) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
- Fills up the passed rawValue with the value corresponding to the key
read earlier
- NICParser - Class in org.apache.hadoop.contrib.failmon
- Objects of this class parse the output of ifconfig to
gather information about present Network Interface Cards
in the system.
- NICParser() -
Constructor for class org.apache.hadoop.contrib.failmon.NICParser
- Constructs a NICParser and reads the list of NICs to query
- NLineInputFormat - Class in org.apache.hadoop.mapred.lib
- NLineInputFormat which splits N lines of input as one split.
- NLineInputFormat() -
Constructor for class org.apache.hadoop.mapred.lib.NLineInputFormat
-
- NLineInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
- NLineInputFormat which splits N lines of input as one split.
- NLineInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
-
- NO_DESCRIPTION -
Static variable in class org.apache.hadoop.metrics.util.MetricsBase
- Deprecated.
- NO_DESCRIPTION -
Static variable in class org.apache.hadoop.metrics2.Metric
-
- Node - Interface in org.apache.hadoop.net
- The interface defines a node in a network topology.
- NodeBase - Class in org.apache.hadoop.net
- A base class that implements interface Node
- NodeBase() -
Constructor for class org.apache.hadoop.net.NodeBase
- Default constructor
- NodeBase(String) -
Constructor for class org.apache.hadoop.net.NodeBase
- Construct a node from its path
- NodeBase(String, String) -
Constructor for class org.apache.hadoop.net.NodeBase
- Construct a node from its name and its location
- NodeBase(String, String, Node, int) -
Constructor for class org.apache.hadoop.net.NodeBase
- Construct a node from its name and its location
- NoEmitMetricsContext - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - NoEmitMetricsContext() -
Constructor for class org.apache.hadoop.metrics.spi.NoEmitMetricsContext
- Deprecated. Creates a new instance of NullContextWithUpdateThread
- noInterruptsLock -
Variable in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
- If the delegation token update thread holds this lock, it will
not get interrupted.
- normalize(String) -
Static method in class org.apache.hadoop.net.NodeBase
- Normalize a path
- normalizeHostName(String) -
Static method in class org.apache.hadoop.net.NetUtils
- Given a string representation of a host, return its ip address
in textual presentation.
- normalizeHostNames(Collection<String>) -
Static method in class org.apache.hadoop.net.NetUtils
- Given a collection of string representation of hosts, return a list of
corresponding IP addresses in the textual representation.
- normalizeMemoryConfigValue(long) -
Static method in class org.apache.hadoop.mapred.JobConf
- Normalize the negative values in configuration
- normalizePath(String) -
Static method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- not() -
Method in enum org.apache.hadoop.fs.permission.FsAction
- NOT operation.
- not() -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- not() -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- not() -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- not() -
Method in class org.apache.hadoop.util.bloom.Filter
- Performs a logical NOT on this filter.
- NULL -
Static variable in interface org.apache.hadoop.mapred.Reporter
- A constant of Reporter type that does nothing.
- NullContext - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - NullContext() -
Constructor for class org.apache.hadoop.metrics.spi.NullContext
- Deprecated. Creates a new instance of NullContext
- NullContextWithUpdateThread - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - NullContextWithUpdateThread() -
Constructor for class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
- Deprecated. Creates a new instance of NullContextWithUpdateThread
- NullOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
- Consume all outputs and put them in /dev/null.
- NullOutputFormat() -
Constructor for class org.apache.hadoop.mapred.lib.NullOutputFormat
-
- NullOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
- Consume all outputs and put them in /dev/null.
- NullOutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
-
- NullWritable - Class in org.apache.hadoop.io
- Singleton Writable with no data.
- NullWritable.Comparator - Class in org.apache.hadoop.io
- A Comparator "optimized" for NullWritable.
- NullWritable.Comparator() -
Constructor for class org.apache.hadoop.io.NullWritable.Comparator
-
- NUM_OF_VALUES_FIELD -
Static variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- NUM_OVERSHOOT_BYTES -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- numberOfBytesTillNextMarker(InputStream) -
Static method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
- Returns the number of bytes between the current stream position
and the immediate next BZip2 block marker.
- numberOfSecretKeys() -
Method in class org.apache.hadoop.security.Credentials
-
- numberOfTokens() -
Method in class org.apache.hadoop.security.Credentials
-
- numOfRacks -
Variable in class org.apache.hadoop.net.NetworkTopology
-
- numOfValues -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- numReduceTasksSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- numSamples() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
O
- O_APPEND -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_ASYNC -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_CREAT -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_EXCL -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_FSYNC -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_NDELAY -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_NOCTTY -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_NONBLOCK -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_RDONLY -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_RDWR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_SYNC -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_TRUNC -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- O_WRONLY -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- ObjectWritable - Class in org.apache.hadoop.io
- A polymorphic Writable that writes an instance with it's class name.
- ObjectWritable() -
Constructor for class org.apache.hadoop.io.ObjectWritable
-
- ObjectWritable(Object) -
Constructor for class org.apache.hadoop.io.ObjectWritable
-
- ObjectWritable(Class, Object) -
Constructor for class org.apache.hadoop.io.ObjectWritable
-
- obtainJobCleanupTask(TaskTrackerStatus, int, int, boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a CleanupTask, if appropriate, to run on the given tasktracker
- obtainJobSetupTask(TaskTrackerStatus, int, int, boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a SetupTask, if appropriate, to run on the given tasktracker
- obtainNewMapTask(TaskTrackerStatus, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a MapTask, if appropriate, to run on the given tasktracker
- obtainNewMapTaskCommon(TaskTrackerStatus, int, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a MapTask with locality level that smaller or equal than a given
locality level to tasktracker.
- obtainNewNodeLocalMapTask(TaskTrackerStatus, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- obtainNewNodeOrRackLocalMapTask(TaskTrackerStatus, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- obtainNewNonLocalMapTask(TaskTrackerStatus, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- obtainNewReduceTask(TaskTrackerStatus, int, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a ReduceTask, if appropriate, to run on the given tasktracker.
- obtainTaskCleanupTask(TaskTrackerStatus, boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- obtainTokensForNamenodes(Credentials, Path[], Configuration) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
- Convenience method to obtain delegation tokens from namenodes
corresponding to the paths passed.
- offerService() -
Method in class org.apache.hadoop.mapred.JobTracker
- Run forever
- OfflineAnonymizer - Class in org.apache.hadoop.contrib.failmon
- This class can be used to anonymize logs independently of
Hadoop and the Executor.
- OfflineAnonymizer(OfflineAnonymizer.LogType, String) -
Constructor for class org.apache.hadoop.contrib.failmon.OfflineAnonymizer
- Creates an OfflineAnonymizer for a specific log file.
- OfflineAnonymizer.LogType - Enum in org.apache.hadoop.contrib.failmon
-
- offset() -
Method in class org.apache.hadoop.io.file.tfile.ByteArray
-
- offset() -
Method in interface org.apache.hadoop.io.file.tfile.RawComparable
- Get the offset of the first byte in the byte array.
- offset -
Variable in class org.apache.hadoop.metrics.ganglia.GangliaContext
- Deprecated.
- ONE -
Static variable in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
-
- oneRotation -
Static variable in class org.apache.hadoop.examples.dancing.Pentomino
- Is the piece fixed under rotation?
- OneSidedPentomino - Class in org.apache.hadoop.examples.dancing
- Of the "normal" 12 pentominos, 6 of them have distinct shapes when flipped.
- OneSidedPentomino() -
Constructor for class org.apache.hadoop.examples.dancing.OneSidedPentomino
-
- OneSidedPentomino(int, int) -
Constructor for class org.apache.hadoop.examples.dancing.OneSidedPentomino
-
- open(Path, int) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- Opens an FSDataInputStream at the indicated Path.
- open(Path, int) -
Method in class org.apache.hadoop.fs.FileSystem
- Opens an FSDataInputStream at the indicated Path.
- open(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Opens an FSDataInputStream at the indicated Path.
- open(Path, int) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Opens an FSDataInputStream at the indicated Path.
- open(Path, int) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- open(Path, int) -
Method in class org.apache.hadoop.fs.HarFileSystem
- Returns a har input stream which fakes end of
file.
- open(Path, int) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- open(Path, int) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- open(Path, int) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- open(Path, int) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- open(FileSystem, String, WritableComparator, Configuration) -
Method in class org.apache.hadoop.io.MapFile.Reader
-
- open(String, int, int) -
Static method in class org.apache.hadoop.io.nativeio.NativeIO
- Wrapper around open(2)
- open(InputStream) -
Method in interface org.apache.hadoop.io.serializer.Deserializer
- Prepare the deserializer for reading.
- open(OutputStream) -
Method in interface org.apache.hadoop.io.serializer.Serializer
- Prepare the serializer for writing.
- openConnection(URL, AuthenticatedURL.Token) -
Method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL
- Returns an authenticated
HttpURLConnection
.
- openFile(FileSystem, Path, int, long) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Override this method to specialize the type of
FSDataInputStream
returned.
- openForRead(File, String) -
Static method in class org.apache.hadoop.io.SecureIOUtils
- Open the given File for read access, verifying the expected user
constraints.
- openInput(String) -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- openInput(String, int) -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- openSecureHttpConnection(URL) -
Static method in class org.apache.hadoop.security.SecurityUtil
- Open a (if need be) secure connection to a URL in a secure environment
that is using SPNEGO or KSSL to authenticate its URLs.
- Operation - Enum in org.apache.hadoop.mapred
- Generic operation that maps to the dependent set of ACLs that drive the
authorization of the operation.
- or(FsAction) -
Method in enum org.apache.hadoop.fs.permission.FsAction
- OR operation.
- or(Filter) -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- or(Filter) -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- or(Filter) -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- or(Filter) -
Method in class org.apache.hadoop.util.bloom.Filter
- Peforms a logical OR between this filter and a specified filter.
- OracleDataDrivenDBInputFormat<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
- A InputFormat that reads input data from an SQL table in an Oracle db.
- OracleDataDrivenDBInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat
-
- OracleDataDrivenDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
- A RecordReader that reads records from a Oracle table via DataDrivenDBRecordReader
- OracleDataDrivenDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) -
Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBRecordReader
-
- OracleDateSplitter - Class in org.apache.hadoop.mapreduce.lib.db
- Implement DBSplitter over date/time values returned by an Oracle db.
- OracleDateSplitter() -
Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDateSplitter
-
- OracleDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
- A RecordReader that reads records from an Oracle SQL table.
- OracleDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) -
Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
-
- org.apache.hadoop - package org.apache.hadoop
-
- org.apache.hadoop.classification - package org.apache.hadoop.classification
-
- org.apache.hadoop.conf - package org.apache.hadoop.conf
- Configuration of system parameters.
- org.apache.hadoop.contrib.failmon - package org.apache.hadoop.contrib.failmon
-
- org.apache.hadoop.contrib.index.example - package org.apache.hadoop.contrib.index.example
-
- org.apache.hadoop.contrib.index.lucene - package org.apache.hadoop.contrib.index.lucene
-
- org.apache.hadoop.contrib.index.main - package org.apache.hadoop.contrib.index.main
-
- org.apache.hadoop.contrib.index.mapred - package org.apache.hadoop.contrib.index.mapred
-
- org.apache.hadoop.contrib.utils.join - package org.apache.hadoop.contrib.utils.join
-
- org.apache.hadoop.examples - package org.apache.hadoop.examples
- Hadoop example code.
- org.apache.hadoop.examples.dancing - package org.apache.hadoop.examples.dancing
- This package is a distributed implementation of Knuth's dancing links
algorithm that can run under Hadoop.
- org.apache.hadoop.examples.terasort - package org.apache.hadoop.examples.terasort
- This package consists of 3 map/reduce applications for Hadoop to
compete in the annual terabyte sort
competition.
- org.apache.hadoop.filecache - package org.apache.hadoop.filecache
-
- org.apache.hadoop.fs - package org.apache.hadoop.fs
- An abstract file system API.
- org.apache.hadoop.fs.ftp - package org.apache.hadoop.fs.ftp
-
- org.apache.hadoop.fs.kfs - package org.apache.hadoop.fs.kfs
- A client for the Kosmos filesystem (KFS)
- org.apache.hadoop.fs.permission - package org.apache.hadoop.fs.permission
-
- org.apache.hadoop.fs.s3 - package org.apache.hadoop.fs.s3
- A distributed, block-based implementation of
FileSystem
that uses Amazon S3
as a backing store. - org.apache.hadoop.fs.s3native - package org.apache.hadoop.fs.s3native
-
A distributed implementation of
FileSystem
for reading and writing files on
Amazon S3. - org.apache.hadoop.fs.shell - package org.apache.hadoop.fs.shell
-
- org.apache.hadoop.http - package org.apache.hadoop.http
-
- org.apache.hadoop.http.lib - package org.apache.hadoop.http.lib
-
- org.apache.hadoop.io - package org.apache.hadoop.io
- Generic i/o code for use when reading and writing data to the network,
to databases, and to files.
- org.apache.hadoop.io.compress - package org.apache.hadoop.io.compress
-
- org.apache.hadoop.io.compress.bzip2 - package org.apache.hadoop.io.compress.bzip2
-
- org.apache.hadoop.io.compress.snappy - package org.apache.hadoop.io.compress.snappy
-
- org.apache.hadoop.io.compress.zlib - package org.apache.hadoop.io.compress.zlib
-
- org.apache.hadoop.io.file.tfile - package org.apache.hadoop.io.file.tfile
-
- org.apache.hadoop.io.nativeio - package org.apache.hadoop.io.nativeio
-
- org.apache.hadoop.io.retry - package org.apache.hadoop.io.retry
-
A mechanism for selectively retrying methods that throw exceptions under certain circumstances.
- org.apache.hadoop.io.serializer - package org.apache.hadoop.io.serializer
-
This package provides a mechanism for using different serialization frameworks
in Hadoop.
- org.apache.hadoop.ipc - package org.apache.hadoop.ipc
- Tools to help define network clients and servers.
- org.apache.hadoop.ipc.metrics - package org.apache.hadoop.ipc.metrics
-
- org.apache.hadoop.jmx - package org.apache.hadoop.jmx
- This package provides access to JMX primarily through the
JMXJsonServlet
class. - org.apache.hadoop.log - package org.apache.hadoop.log
-
- org.apache.hadoop.log.metrics - package org.apache.hadoop.log.metrics
-
- org.apache.hadoop.mapred - package org.apache.hadoop.mapred
- A software framework for easily writing applications which process vast
amounts of data (multi-terabyte data-sets) parallelly on large clusters
(thousands of nodes) built of commodity hardware in a reliable, fault-tolerant
manner.
- org.apache.hadoop.mapred.jobcontrol - package org.apache.hadoop.mapred.jobcontrol
- Utilities for managing dependent jobs.
- org.apache.hadoop.mapred.join - package org.apache.hadoop.mapred.join
- Given a set of sorted datasets keyed with the same class and yielding equal
partitions, it is possible to effect a join of those datasets prior to the map.
- org.apache.hadoop.mapred.lib - package org.apache.hadoop.mapred.lib
- Library of generally useful mappers, reducers, and partitioners.
- org.apache.hadoop.mapred.lib.aggregate - package org.apache.hadoop.mapred.lib.aggregate
- Classes for performing various counting and aggregations.
- org.apache.hadoop.mapred.lib.db - package org.apache.hadoop.mapred.lib.db
- org.apache.hadoop.mapred.lib.db Package
- org.apache.hadoop.mapred.pipes - package org.apache.hadoop.mapred.pipes
- Hadoop Pipes allows C++ code to use Hadoop DFS and map/reduce.
- org.apache.hadoop.mapred.tools - package org.apache.hadoop.mapred.tools
-
- org.apache.hadoop.mapreduce - package org.apache.hadoop.mapreduce
-
- org.apache.hadoop.mapreduce.lib.db - package org.apache.hadoop.mapreduce.lib.db
-
- org.apache.hadoop.mapreduce.lib.fieldsel - package org.apache.hadoop.mapreduce.lib.fieldsel
-
- org.apache.hadoop.mapreduce.lib.input - package org.apache.hadoop.mapreduce.lib.input
-
- org.apache.hadoop.mapreduce.lib.jobcontrol - package org.apache.hadoop.mapreduce.lib.jobcontrol
-
- org.apache.hadoop.mapreduce.lib.map - package org.apache.hadoop.mapreduce.lib.map
-
- org.apache.hadoop.mapreduce.lib.output - package org.apache.hadoop.mapreduce.lib.output
-
- org.apache.hadoop.mapreduce.lib.partition - package org.apache.hadoop.mapreduce.lib.partition
-
- org.apache.hadoop.mapreduce.lib.reduce - package org.apache.hadoop.mapreduce.lib.reduce
-
- org.apache.hadoop.mapreduce.security - package org.apache.hadoop.mapreduce.security
-
- org.apache.hadoop.mapreduce.security.token - package org.apache.hadoop.mapreduce.security.token
-
- org.apache.hadoop.mapreduce.security.token.delegation - package org.apache.hadoop.mapreduce.security.token.delegation
-
- org.apache.hadoop.mapreduce.server.jobtracker - package org.apache.hadoop.mapreduce.server.jobtracker
-
- org.apache.hadoop.mapreduce.server.tasktracker - package org.apache.hadoop.mapreduce.server.tasktracker
-
- org.apache.hadoop.mapreduce.server.tasktracker.userlogs - package org.apache.hadoop.mapreduce.server.tasktracker.userlogs
-
- org.apache.hadoop.mapreduce.split - package org.apache.hadoop.mapreduce.split
-
- org.apache.hadoop.metrics - package org.apache.hadoop.metrics
- Note, this package is deprecated in favor of
org.apache.hadoop.metrics2
usage. - org.apache.hadoop.metrics.file - package org.apache.hadoop.metrics.file
- Implementation of the metrics package that writes the metrics to a file.
- org.apache.hadoop.metrics.ganglia - package org.apache.hadoop.metrics.ganglia
- Implementation of the metrics package that sends metric data to
Ganglia.
- org.apache.hadoop.metrics.jvm - package org.apache.hadoop.metrics.jvm
-
- org.apache.hadoop.metrics.spi - package org.apache.hadoop.metrics.spi
- Note, this package is deprecated in favor of
org.apache.hadoop.metrics2
usage. - org.apache.hadoop.metrics.util - package org.apache.hadoop.metrics.util
-
- org.apache.hadoop.metrics2 - package org.apache.hadoop.metrics2
- Metrics 2.0
- org.apache.hadoop.metrics2.filter - package org.apache.hadoop.metrics2.filter
-
- org.apache.hadoop.metrics2.impl - package org.apache.hadoop.metrics2.impl
-
- org.apache.hadoop.metrics2.lib - package org.apache.hadoop.metrics2.lib
-
- org.apache.hadoop.metrics2.sink - package org.apache.hadoop.metrics2.sink
-
- org.apache.hadoop.metrics2.sink.ganglia - package org.apache.hadoop.metrics2.sink.ganglia
-
- org.apache.hadoop.metrics2.source - package org.apache.hadoop.metrics2.source
-
- org.apache.hadoop.metrics2.util - package org.apache.hadoop.metrics2.util
-
- org.apache.hadoop.net - package org.apache.hadoop.net
- Network-related classes.
- org.apache.hadoop.record - package org.apache.hadoop.record
- Hadoop record I/O contains classes and a record description language
translator for simplifying serialization and deserialization of records in a
language-neutral manner.
- org.apache.hadoop.record.compiler - package org.apache.hadoop.record.compiler
- This package contains classes needed for code generation
from the hadoop record compiler.
- org.apache.hadoop.record.compiler.ant - package org.apache.hadoop.record.compiler.ant
-
- org.apache.hadoop.record.compiler.generated - package org.apache.hadoop.record.compiler.generated
- This package contains code generated by JavaCC from the
Hadoop record syntax file rcc.jj.
- org.apache.hadoop.record.meta - package org.apache.hadoop.record.meta
-
- org.apache.hadoop.security - package org.apache.hadoop.security
-
- org.apache.hadoop.security.authentication.client - package org.apache.hadoop.security.authentication.client
-
- org.apache.hadoop.security.authentication.server - package org.apache.hadoop.security.authentication.server
-
- org.apache.hadoop.security.authentication.util - package org.apache.hadoop.security.authentication.util
-
- org.apache.hadoop.security.authorize - package org.apache.hadoop.security.authorize
-
- org.apache.hadoop.security.token - package org.apache.hadoop.security.token
-
- org.apache.hadoop.security.token.delegation - package org.apache.hadoop.security.token.delegation
-
- org.apache.hadoop.streaming - package org.apache.hadoop.streaming
- Hadoop Streaming is a utility which allows users to create and run
Map-Reduce jobs with any executables (e.g.
- org.apache.hadoop.streaming.io - package org.apache.hadoop.streaming.io
-
- org.apache.hadoop.typedbytes - package org.apache.hadoop.typedbytes
- Typed bytes are sequences of bytes in which the first byte is a type code.
- org.apache.hadoop.util - package org.apache.hadoop.util
- Common utilities.
- org.apache.hadoop.util.bloom - package org.apache.hadoop.util.bloom
-
- org.apache.hadoop.util.hash - package org.apache.hadoop.util.hash
-
- osType -
Static variable in class org.apache.hadoop.fs.HardLink
-
- out -
Variable in class org.apache.hadoop.io.compress.CompressionOutputStream
- The output stream to be compressed.
- out -
Variable in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
-
- out -
Variable in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
-
- OuterJoinRecordReader<K extends WritableComparable> - Class in org.apache.hadoop.mapred.join
- Full outer join.
- outerrThreadsThrowable -
Variable in class org.apache.hadoop.streaming.PipeMapRed
-
- output_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- OUTPUT_FIELD_COUNT_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Number of fields in the Output table
- OUTPUT_FIELD_NAMES_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Field names in the Output table
- OUTPUT_FIELD_NAMES_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Field names in the Output table
- OUTPUT_FORMAT -
Static variable in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
-
- OUTPUT_FORMAT_CLASS_ATTR -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- OUTPUT_TABLE_NAME_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Output table name
- OUTPUT_TABLE_NAME_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Output table name
- OutputBuffer - Class in org.apache.hadoop.io
- A reusable
OutputStream
implementation that writes to an in-memory
buffer. - OutputBuffer() -
Constructor for class org.apache.hadoop.io.OutputBuffer
- Constructs a new empty buffer.
- OutputCollector<K,V> - Interface in org.apache.hadoop.mapred
- Collects the
<key, value>
pairs output by Mapper
s
and Reducer
s. - OutputCommitter - Class in org.apache.hadoop.mapred
OutputCommitter
describes the commit of task output for a
Map-Reduce job.- OutputCommitter() -
Constructor for class org.apache.hadoop.mapred.OutputCommitter
-
- OutputCommitter - Class in org.apache.hadoop.mapreduce
OutputCommitter
describes the commit of task output for a
Map-Reduce job.- OutputCommitter() -
Constructor for class org.apache.hadoop.mapreduce.OutputCommitter
-
- OutputFormat<K,V> - Interface in org.apache.hadoop.mapred
OutputFormat
describes the output-specification for a
Map-Reduce job.- outputFormat -
Variable in class org.apache.hadoop.mapred.Task
-
- OutputFormat<K,V> - Class in org.apache.hadoop.mapreduce
OutputFormat
describes the output-specification for a
Map-Reduce job.- OutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.OutputFormat
-
- outputFormatSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- OutputLogFilter - Class in org.apache.hadoop.mapred
- Deprecated. Use
Utils.OutputFileUtils.OutputLogFilter
instead. - OutputLogFilter() -
Constructor for class org.apache.hadoop.mapred.OutputLogFilter
- Deprecated.
- OutputReader<K,V> - Class in org.apache.hadoop.streaming.io
- Abstract base for classes that read the client's output.
- OutputReader() -
Constructor for class org.apache.hadoop.streaming.io.OutputReader
-
- OutputRecord - Class in org.apache.hadoop.metrics.spi
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - outputSingleNode_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- OverrideRecordReader<K extends WritableComparable,V extends Writable> - Class in org.apache.hadoop.mapred.join
- Prefer the "rightmost" data source for this key.
- overrideSchedulingOpportunities() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
P
- pack(SerializedRecord) -
Static method in class org.apache.hadoop.contrib.failmon.LocalStore
- Pack a SerializedRecord into an array of bytes
- packageFiles_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- packageJobJar() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- PAGE_SIZE -
Static variable in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- parent -
Variable in class org.apache.hadoop.net.NodeBase
-
- parse(String[], int) -
Method in class org.apache.hadoop.fs.shell.CommandFormat
- Parse parameters starting from the given position
- parse(String, int) -
Static method in class org.apache.hadoop.metrics.spi.Util
- Parses a space and/or comma separated sequence of server specifications
of the form hostname or hostname:port.
- parse(String, int) -
Static method in class org.apache.hadoop.metrics2.util.Servers
- Parses a space and/or comma separated sequence of server specifications
of the form hostname or hostname:port.
- parse(String) -
Static method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
- Parses a string into an authentication token.
- parseAndSetPeriod(String) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. If a period is set in the attribute passed in, override
the default with it.
- parseCommaSeparatedString(String) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry
- Parse the given string as a MultipleLinearRandomRetry object.
- parseDate(String, String) -
Method in class org.apache.hadoop.contrib.failmon.HadoopLogParser
- Parse a date found in the Hadoop log.
- parseDate(String, String) -
Method in class org.apache.hadoop.contrib.failmon.LogParser
- Parse a date found in Hadoop log file.
- parseDate(String, String) -
Method in class org.apache.hadoop.contrib.failmon.SystemLogParser
- Parse a date found in the system log.
- ParseException - Exception in org.apache.hadoop.record.compiler.generated
- This exception is thrown when parse errors are encountered.
- ParseException(Token, int[][], String[]) -
Constructor for exception org.apache.hadoop.record.compiler.generated.ParseException
- This constructor is used by the method "generateParseException"
in the generated parser.
- ParseException() -
Constructor for exception org.apache.hadoop.record.compiler.generated.ParseException
- The following constructors are for use by you for whatever
purpose you can think of.
- ParseException(String) -
Constructor for exception org.apache.hadoop.record.compiler.generated.ParseException
-
- parseExecResult(BufferedReader) -
Method in class org.apache.hadoop.fs.DF
-
- parseExecResult(BufferedReader) -
Method in class org.apache.hadoop.fs.DU
-
- parseExecResult(BufferedReader) -
Method in class org.apache.hadoop.util.Shell
- Parse the execution result
- parseExecResult(BufferedReader) -
Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- parseHashType(String) -
Static method in class org.apache.hadoop.util.hash.Hash
- This utility method converts String representation of hash function name
to a symbolic constant.
- parseHistoryFromFS(String, JobHistory.Listener, FileSystem) -
Static method in class org.apache.hadoop.mapred.JobHistory
- Parses history file and invokes Listener.handle() for
each line of history.
- parseJobTasks(String, JobHistory.JobInfo, FileSystem) -
Static method in class org.apache.hadoop.mapred.DefaultJobHistoryParser
- Populates a JobInfo object from the job's history log file.
- parseLine(String) -
Method in class org.apache.hadoop.contrib.failmon.HadoopLogParser
- Parses one line of the log.
- parseLine(String) -
Method in class org.apache.hadoop.contrib.failmon.LogParser
- Parses one line of the log.
- parseLine(String) -
Method in class org.apache.hadoop.contrib.failmon.SystemLogParser
- Parses one line of the log.
- parseOutputKeyValueSpec(String, List<Integer>, List<Integer>) -
Static method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
-
- Parser - Class in org.apache.hadoop.mapred.join
- Very simple shift-reduce parser for join expressions.
- Parser() -
Constructor for class org.apache.hadoop.mapred.join.Parser
-
- Parser.Node - Class in org.apache.hadoop.mapred.join
-
- Parser.Node(String) -
Constructor for class org.apache.hadoop.mapred.join.Parser.Node
-
- Parser.NodeToken - Class in org.apache.hadoop.mapred.join
-
- Parser.NumToken - Class in org.apache.hadoop.mapred.join
-
- Parser.NumToken(double) -
Constructor for class org.apache.hadoop.mapred.join.Parser.NumToken
-
- Parser.StrToken - Class in org.apache.hadoop.mapred.join
-
- Parser.StrToken(Parser.TType, String) -
Constructor for class org.apache.hadoop.mapred.join.Parser.StrToken
-
- Parser.Token - Class in org.apache.hadoop.mapred.join
- Tagged-union type for tokens from the join expression.
- Parser.TType - Enum in org.apache.hadoop.mapred.join
-
- PART -
Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- Partitioner<K2,V2> - Interface in org.apache.hadoop.mapred
- Partitions the key space.
- Partitioner<KEY,VALUE> - Class in org.apache.hadoop.mapreduce
- Partitions the key space.
- Partitioner() -
Constructor for class org.apache.hadoop.mapreduce.Partitioner
-
- PARTITIONER_CLASS_ATTR -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- PARTITIONER_OPTIONS -
Static variable in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
-
- PARTITIONER_PATH -
Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
-
- partitionerSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- PASSWORD_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- Password to access the database
- PASSWORD_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- Password to access the database
- Path - Class in org.apache.hadoop.fs
- Names a file or directory in a
FileSystem
. - Path(String, String) -
Constructor for class org.apache.hadoop.fs.Path
- Resolve a child path against a parent path.
- Path(Path, String) -
Constructor for class org.apache.hadoop.fs.Path
- Resolve a child path against a parent path.
- Path(String, Path) -
Constructor for class org.apache.hadoop.fs.Path
- Resolve a child path against a parent path.
- Path(Path, Path) -
Constructor for class org.apache.hadoop.fs.Path
- Resolve a child path against a parent path.
- Path(String) -
Constructor for class org.apache.hadoop.fs.Path
- Construct a path from a String.
- Path(String, String, String) -
Constructor for class org.apache.hadoop.fs.Path
- Construct a Path from components.
- Path(URI) -
Constructor for class org.apache.hadoop.fs.Path
- Construct a path from a URI
- PATH_SEPARATOR -
Static variable in class org.apache.hadoop.net.NodeBase
-
- PATH_SEPARATOR_STR -
Static variable in class org.apache.hadoop.net.NodeBase
-
- PathFilter - Interface in org.apache.hadoop.fs
-
- PathFinder - Class in org.apache.hadoop.streaming
- Maps a relative pathname to an absolute pathname using the
PATH enviroment.
- PathFinder() -
Constructor for class org.apache.hadoop.streaming.PathFinder
- Construct a PathFinder object using the path from
java.class.path
- PathFinder(String) -
Constructor for class org.apache.hadoop.streaming.PathFinder
- Construct a PathFinder object using the path from
the specified system environment variable.
- pathToFile(Path) -
Method in class org.apache.hadoop.fs.LocalFileSystem
- Convert a path to a File.
- pathToFile(Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Convert a path to a File.
- pendingMaps() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- pendingReduces() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- Pentomino - Class in org.apache.hadoop.examples.dancing
-
- Pentomino(int, int) -
Constructor for class org.apache.hadoop.examples.dancing.Pentomino
- Create the model for a given pentomino set of pieces and board size.
- Pentomino() -
Constructor for class org.apache.hadoop.examples.dancing.Pentomino
- Create the object without initialization.
- Pentomino.ColumnName - Interface in org.apache.hadoop.examples.dancing
- This interface just is a marker for what types I expect to get back
as column names.
- Pentomino.Piece - Class in org.apache.hadoop.examples.dancing
- Maintain information about a puzzle piece.
- Pentomino.Piece(String, String, boolean, int[]) -
Constructor for class org.apache.hadoop.examples.dancing.Pentomino.Piece
-
- Pentomino.SolutionCategory - Enum in org.apache.hadoop.examples.dancing
-
- percentageGraph(int, int) -
Static method in class org.apache.hadoop.util.ServletUtil
- Generate the percentage graph and returns HTML representation string
of the same.
- percentageGraph(float, int) -
Static method in class org.apache.hadoop.util.ServletUtil
- Generate the percentage graph and returns HTML representation string
of the same.
- PERIOD_PROPERTY -
Static variable in class org.apache.hadoop.metrics.file.FileContext
- Deprecated.
- PermissionStatus - Class in org.apache.hadoop.fs.permission
- Store permission related information.
- PermissionStatus(String, String, FsPermission) -
Constructor for class org.apache.hadoop.fs.permission.PermissionStatus
- Constructor
- PersistentState - Class in org.apache.hadoop.contrib.failmon
- This class takes care of the information that needs to be
persistently stored locally on nodes.
- PersistentState() -
Constructor for class org.apache.hadoop.contrib.failmon.PersistentState
-
- phase() -
Method in class org.apache.hadoop.util.Progress
- Returns the current sub-node executing.
- pieces -
Variable in class org.apache.hadoop.examples.dancing.Pentomino
-
- PiEstimator - Class in org.apache.hadoop.examples
- A Map-reduce program to estimate the value of Pi
using quasi-Monte Carlo method.
- PiEstimator() -
Constructor for class org.apache.hadoop.examples.PiEstimator
-
- PiEstimator.PiMapper - Class in org.apache.hadoop.examples
- Mapper class for Pi estimation.
- PiEstimator.PiMapper() -
Constructor for class org.apache.hadoop.examples.PiEstimator.PiMapper
-
- PiEstimator.PiReducer - Class in org.apache.hadoop.examples
- Reducer class for Pi estimation.
- PiEstimator.PiReducer() -
Constructor for class org.apache.hadoop.examples.PiEstimator.PiReducer
-
- ping(TaskAttemptID, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
- Child checking to see if we're alive.
- ping(TaskAttemptID, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Periodically called by child to check if parent is still alive.
- PipeCombiner - Class in org.apache.hadoop.streaming
-
- PipeCombiner() -
Constructor for class org.apache.hadoop.streaming.PipeCombiner
-
- PipeMapper - Class in org.apache.hadoop.streaming
- A generic Mapper bridge.
- PipeMapper() -
Constructor for class org.apache.hadoop.streaming.PipeMapper
-
- PipeMapRed - Class in org.apache.hadoop.streaming
- Shared functionality for PipeMapper, PipeReducer.
- PipeMapRed() -
Constructor for class org.apache.hadoop.streaming.PipeMapRed
-
- PipeMapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.streaming
-
- PipeMapRunner() -
Constructor for class org.apache.hadoop.streaming.PipeMapRunner
-
- PipeReducer - Class in org.apache.hadoop.streaming
- A generic Reducer bridge.
- PipeReducer() -
Constructor for class org.apache.hadoop.streaming.PipeReducer
-
- PlatformName - Class in org.apache.hadoop.util
- A helper class for getting build-info of the java-vm.
- PlatformName() -
Constructor for class org.apache.hadoop.util.PlatformName
-
- POLICY_PROVIDER_CONFIG -
Static variable in class org.apache.hadoop.security.authorize.PolicyProvider
- Configuration key for the
PolicyProvider
implementation.
- PolicyProvider - Class in org.apache.hadoop.security.authorize
PolicyProvider
provides the Service
definitions to the
security Policy
in effect for Hadoop.- PolicyProvider() -
Constructor for class org.apache.hadoop.security.authorize.PolicyProvider
-
- pop() -
Method in class org.apache.hadoop.util.PriorityQueue
- Removes and returns the least element of the PriorityQueue in log(size)
time.
- PositionedReadable - Interface in org.apache.hadoop.fs
- Stream that permits positional reading.
- POSIX_FADV_DONTNEED -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- POSIX_FADV_NOREUSE -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- POSIX_FADV_NORMAL -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- POSIX_FADV_RANDOM -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- POSIX_FADV_SEQUENTIAL -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- POSIX_FADV_WILLNEED -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- posixFadviseIfPossible(FileDescriptor, long, long, int) -
Static method in class org.apache.hadoop.io.nativeio.NativeIO
- Call posix_fadvise on the given file descriptor.
- postCheckIsLastCycle(int) -
Method in class org.apache.hadoop.util.QueueProcessingStatistics
- See
QueueProcessingStatistics.preCheckIsLastCycle(int)
.
- postStart() -
Method in class org.apache.hadoop.metrics2.MetricsSystem.AbstractCallback
-
- postStart() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem.Callback
- Called after start()
- postStop() -
Method in class org.apache.hadoop.metrics2.MetricsSystem.AbstractCallback
-
- postStop() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem.Callback
- Called after stop()
- preCheckIsLastCycle(int) -
Method in class org.apache.hadoop.util.QueueProcessingStatistics
- The termination condition is to identify the last cycle that will
empty the queue.
- PREP -
Static variable in class org.apache.hadoop.mapred.JobStatus
-
- prepare(String) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
- Initializes structures needed by other methods.
- prepareAppendKey(int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Writer
- Obtain an output stream for writing a key into TFile.
- prepareAppendValue(int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Writer
- Obtain an output stream for writing a value into TFile.
- prepareMetaBlock(String, String) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Writer
- Obtain an output stream for creating a meta block.
- prepareMetaBlock(String) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Writer
- Obtain an output stream for creating a meta block.
- prependPathComponent(String) -
Method in class org.apache.hadoop.streaming.PathFinder
- Appends the specified component to the path list
- preserveInput(boolean) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
- Whether to delete the files when no longer needed
- preStart() -
Method in class org.apache.hadoop.metrics2.MetricsSystem.AbstractCallback
-
- preStart() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem.Callback
- Called before start()
- preStop() -
Method in class org.apache.hadoop.metrics2.MetricsSystem.AbstractCallback
-
- preStop() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem.Callback
- Called before stop()
- prevCharIsCR -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- prevCharIsLF -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- PRINCIPAL -
Static variable in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- Constant for the configuration property that indicates the kerberos principal.
- printGenericCommandUsage(PrintStream) -
Static method in class org.apache.hadoop.util.GenericOptionsParser
- Print the usage message for generic command-line options supported.
- printGenericCommandUsage(PrintStream) -
Static method in class org.apache.hadoop.util.ToolRunner
- Prints generic command-line argurments and usage information.
- PrintJarMainClass - Class in org.apache.hadoop.util
- A micro-application that prints the main class name out of a jar file.
- PrintJarMainClass() -
Constructor for class org.apache.hadoop.util.PrintJarMainClass
-
- printRules() -
Static method in class org.apache.hadoop.security.KerberosName
-
- printStackTrace() -
Method in exception org.apache.hadoop.security.authorize.AuthorizationException
-
- printStackTrace(PrintStream) -
Method in exception org.apache.hadoop.security.authorize.AuthorizationException
-
- printStackTrace(PrintWriter) -
Method in exception org.apache.hadoop.security.authorize.AuthorizationException
-
- printStatistics() -
Static method in class org.apache.hadoop.fs.FileSystem
-
- printThreadInfo(PrintWriter, String) -
Static method in class org.apache.hadoop.util.ReflectionUtils
- Print all of the thread's information and stack traces.
- PriorityQueue<T> - Class in org.apache.hadoop.util
- A PriorityQueue maintains a partial ordering of its elements such that the
least element can always be found in constant time.
- PriorityQueue() -
Constructor for class org.apache.hadoop.util.PriorityQueue
-
- probablyHasKey(WritableComparable) -
Method in class org.apache.hadoop.io.BloomMapFile.Reader
- Checks if this MapFile has the indicated key.
- process(IntermediateForm) -
Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
- Process an intermediate form by carrying out, on the Lucene instance of
the shard, the deletes and the inserts (a ram index) in the form.
- process(DocumentAndOp, Analyzer) -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
- This method is used by the index update mapper and process a document
operation into the current intermediate form.
- process(IntermediateForm) -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
- This method is used by the index update combiner and process an
intermediate form into the current intermediate form.
- processDeleteOnExit() -
Method in class org.apache.hadoop.fs.FileSystem
- Delete all files that were marked as delete-on-exit.
- processEvent(UserLogEvent) -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
-
- ProcessTree - Class in org.apache.hadoop.util
- Process tree related operations
- ProcessTree() -
Constructor for class org.apache.hadoop.util.ProcessTree
-
- ProcessTree.Signal - Enum in org.apache.hadoop.util
- The constants for the signals.
- ProcfsBasedProcessTree - Class in org.apache.hadoop.util
- A Proc file-system based ProcessTree.
- ProcfsBasedProcessTree(String) -
Constructor for class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- ProcfsBasedProcessTree(String, boolean) -
Constructor for class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- ProcfsBasedProcessTree(String, String) -
Constructor for class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- ProgramDriver - Class in org.apache.hadoop.util
- A driver that is used to run programs added to it
- ProgramDriver() -
Constructor for class org.apache.hadoop.util.ProgramDriver
-
- progress -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- progress() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- progress() -
Method in class org.apache.hadoop.mapred.TaskAttemptContext
-
- progress -
Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- progress() -
Method in class org.apache.hadoop.mapreduce.StatusReporter
-
- progress() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
- Report progress.
- progress() -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- Progress - Class in org.apache.hadoop.util
- Utility to assist with generation of progress reports.
- Progress() -
Constructor for class org.apache.hadoop.util.Progress
- Creates a new root node.
- progress() -
Method in interface org.apache.hadoop.util.Progressable
- Report progress to the Hadoop framework.
- PROGRESS_INTERVAL -
Static variable in class org.apache.hadoop.mapred.Task
- The number of milliseconds between progress reports.
- Progressable - Interface in org.apache.hadoop.util
- A facility for reporting progress.
- ProxyUsers - Class in org.apache.hadoop.security.authorize
-
- ProxyUsers() -
Constructor for class org.apache.hadoop.security.authorize.ProxyUsers
-
- PseudoAuthenticationHandler - Class in org.apache.hadoop.security.authentication.server
- The
PseudoAuthenticationHandler
provides a pseudo authentication mechanism that accepts
the user name specified as a query string parameter. - PseudoAuthenticationHandler() -
Constructor for class org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler
-
- PseudoAuthenticator - Class in org.apache.hadoop.security.authentication.client
- The
PseudoAuthenticator
implementation provides an authentication equivalent to Hadoop's
Simple authentication, it trusts the value of the 'user.name' Java System property. - PseudoAuthenticator() -
Constructor for class org.apache.hadoop.security.authentication.client.PseudoAuthenticator
-
- pseudoSortByDistance(Node, Node[]) -
Method in class org.apache.hadoop.net.NetworkTopology
- Sort nodes array by their distances to reader
It linearly scans the array, if a local node is found, swap it with
the first element of the array.
- pseudoSortByDistance(Node, Node[]) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
- Sort nodes array by their distances to reader
It linearly scans the array, if a local node is found, swap it with
the first element of the array.
- publishMetricsNow() -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
- Requests an immediate publish of all metrics from sources to sinks.
- publishMetricsNow() -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- publishMetricsNow() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem
- Requests an immediate publish of all metrics from sources to sinks.
- PureJavaCrc32 - Class in org.apache.hadoop.util
- A pure-java implementation of the CRC32 checksum that uses
the same polynomial as the built-in native CRC32.
- PureJavaCrc32() -
Constructor for class org.apache.hadoop.util.PureJavaCrc32
-
- purge() -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
- Delete everything.
- purgeCache() -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
- Clear the entire contents of the cache and delete the backing files.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsBase
- Deprecated.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsIntValue
- Deprecated. Push the metric to the mr.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsLongValue
- Deprecated. Push the metric to the mr.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
- Deprecated. Push the delta metrics to the mr.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
- Deprecated. Push the delta metrics to the mr.
- pushMetric(MetricsRecord) -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Push the delta metrics to the mr.
- put(Writable, Writable) -
Method in class org.apache.hadoop.io.MapWritable
-
- put(WritableComparable, Writable) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- put(T) -
Method in class org.apache.hadoop.util.PriorityQueue
- Adds an Object to a PriorityQueue in log(size) time.
- putAll(Map<? extends Writable, ? extends Writable>) -
Method in class org.apache.hadoop.io.MapWritable
-
- putAll(Map<? extends WritableComparable, ? extends Writable>) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- putMetrics(MetricsRecord) -
Method in interface org.apache.hadoop.metrics2.MetricsSink
- Put a metrics record in the sink
- putMetrics(MetricsRecord) -
Method in class org.apache.hadoop.metrics2.sink.FileSink
-
- putMetrics(MetricsRecord) -
Method in class org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30
-
Q
- qACLNeeded -
Variable in enum org.apache.hadoop.mapred.Operation
-
- QSORT_STACK_SIZE -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- quarterDigest() -
Method in class org.apache.hadoop.io.MD5Hash
- Return a 32-bit digest of the MD5.
- query(String) -
Method in class org.apache.hadoop.contrib.failmon.CPUParser
- Reads and parses /proc/cpuinfo and creates an appropriate
EventRecord that holds the desirable information.
- query(String) -
Method in class org.apache.hadoop.contrib.failmon.NICParser
- Reads and parses the output of ifconfig for a specified NIC and
creates an appropriate EventRecord that holds the desirable
information for it.
- query(String) -
Method in class org.apache.hadoop.contrib.failmon.SensorsParser
- Reads and parses the output of the 'sensors' command
and creates an appropriate EventRecord that holds
the desirable information.
- query(String) -
Method in class org.apache.hadoop.contrib.failmon.ShellParser
-
- query(String) -
Method in class org.apache.hadoop.contrib.failmon.SMARTParser
- Reads and parses the output of smartctl for a specified disk and
creates an appropriate EventRecord that holds the desirable
information for it.
- QueueAclsInfo - Class in org.apache.hadoop.mapred
- Class to encapsulate Queue ACLs for a particular
user.
- QueueAclsInfo() -
Constructor for class org.apache.hadoop.mapred.QueueAclsInfo
- Default constructor for QueueAclsInfo.
- QueueAclsInfo(String, String[]) -
Constructor for class org.apache.hadoop.mapred.QueueAclsInfo
- Construct a new QueueAclsInfo object using the queue name and the
queue operations array
- QueueProcessingStatistics - Class in org.apache.hadoop.util
- Hadoop has several work queues, such as
FSNamesystem.neededReplications
With a properly throttled queue, a worker thread cycles repeatedly,
doing a chunk of work each cycle then resting a bit, until the queue is
empty. - QueueProcessingStatistics(String, String, Log) -
Constructor for class org.apache.hadoop.util.QueueProcessingStatistics
-
- QueueProcessingStatistics.State - Enum in org.apache.hadoop.util
- This enum provides the "states" of a state machine for
QueueProcessingStatistics
. - QuickSort - Class in org.apache.hadoop.util
- An implementation of the core algorithm of QuickSort.
- QuickSort() -
Constructor for class org.apache.hadoop.util.QuickSort
-
- quoteHtmlChars(OutputStream, byte[], int, int) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Quote all of the active HTML characters in the given string as they
are added to the buffer.
- quoteHtmlChars(String) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Quote the given item to make it html-safe.
- quoteOutputStream(OutputStream) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Return an output stream that quotes all of the output.
R
- r -
Static variable in class org.apache.hadoop.net.NetworkTopology
-
- RAMDirectoryUtil - Class in org.apache.hadoop.contrib.index.lucene
- A utility class which writes an index in a ram dir into a DataOutput and
read from a DataInput an index into a ram dir.
- RAMDirectoryUtil() -
Constructor for class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
-
- RANDOM -
Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
- Random selection.
- RandomTextWriter - Class in org.apache.hadoop.examples
- This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task writes a large unsorted
random sequence of words.
- RandomTextWriter() -
Constructor for class org.apache.hadoop.examples.RandomTextWriter
-
- RandomWriter - Class in org.apache.hadoop.examples
- This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task write a large unsorted
random binary sequence file of BytesWritable.
- RandomWriter() -
Constructor for class org.apache.hadoop.examples.RandomWriter
-
- RATIO -
Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
- Ratio Selection.
- RAW_BYTES_ID -
Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
-
- RawBytesInputWriter - Class in org.apache.hadoop.streaming.io
- InputWriter that writes the client's input as raw bytes.
- RawBytesInputWriter() -
Constructor for class org.apache.hadoop.streaming.io.RawBytesInputWriter
-
- RawBytesOutputReader - Class in org.apache.hadoop.streaming.io
- OutputReader that reads the client's output as raw bytes.
- RawBytesOutputReader() -
Constructor for class org.apache.hadoop.streaming.io.RawBytesOutputReader
-
- RawComparable - Interface in org.apache.hadoop.io.file.tfile
- Interface for objects that can be compared through
RawComparator
. - RawComparator<T> - Interface in org.apache.hadoop.io
-
A
Comparator
that operates directly on byte representations of
objects. - RawHistoryFileServlet - Class in org.apache.hadoop.mapred
-
- RawHistoryFileServlet() -
Constructor for class org.apache.hadoop.mapred.RawHistoryFileServlet
-
- RawKeyValueIterator - Interface in org.apache.hadoop.mapred
RawKeyValueIterator
is an iterator used to iterate over
the raw keys and values during sort/merge of intermediate data.- RawLocalFileSystem - Class in org.apache.hadoop.fs
- Implement the FileSystem API for the raw local filesystem.
- RawLocalFileSystem() -
Constructor for class org.apache.hadoop.fs.RawLocalFileSystem
-
- rawMapping -
Variable in class org.apache.hadoop.net.CachedDNSToSwitchMapping
-
- rawWriter -
Variable in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
-
- RBRACE_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- Rcc - Class in org.apache.hadoop.record.compiler.generated
-
- Rcc(InputStream) -
Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
- Rcc(InputStream, String) -
Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
- Rcc(Reader) -
Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
- Rcc(RccTokenManager) -
Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
- RccConstants - Interface in org.apache.hadoop.record.compiler.generated
-
- RccTask - Class in org.apache.hadoop.record.compiler.ant
- Hadoop record compiler ant Task
- RccTask() -
Constructor for class org.apache.hadoop.record.compiler.ant.RccTask
- Creates a new instance of RccTask
- RccTokenManager - Class in org.apache.hadoop.record.compiler.generated
-
- RccTokenManager(SimpleCharStream) -
Constructor for class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- RccTokenManager(SimpleCharStream, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- read(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- read(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.FSDataInputStream
-
- read() -
Method in class org.apache.hadoop.fs.FSInputChecker
- Read one checksum-verified byte
- read(byte[], int, int) -
Method in class org.apache.hadoop.fs.FSInputChecker
- Read checksum verified bytes from this byte-input stream into
the specified byte array, starting at the given offset.
- read(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.FSInputStream
-
- read() -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- read(byte[], int, int) -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- read(DataInput) -
Static method in class org.apache.hadoop.fs.permission.FsPermission
- Create and initialize a
FsPermission
from DataInput
.
- read(DataInput) -
Static method in class org.apache.hadoop.fs.permission.PermissionStatus
- Create and initialize a
PermissionStatus
from DataInput
.
- read(long, byte[], int, int) -
Method in interface org.apache.hadoop.fs.PositionedReadable
- Read upto the specified number of bytes, from a given
position within a file, and return the number of bytes read.
- read() -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
- read(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
- In CONTINOUS reading mode, this read method starts from the
start of the compressed stream and end at the end of file by
emitting un-compressed data.
- read(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.CompressionInputStream
- Read bytes from the stream.
- read() -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- read(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- read(DataInput) -
Static method in class org.apache.hadoop.io.MD5Hash
- Constructs, reads and returns an instance.
- read(DataInput) -
Static method in class org.apache.hadoop.mapred.JobID
- Deprecated.
- read(DataInput) -
Static method in class org.apache.hadoop.mapred.TaskAttemptID
- Deprecated.
- read(DataInput) -
Static method in class org.apache.hadoop.mapred.TaskID
- Deprecated.
- read() -
Method in class org.apache.hadoop.net.SocketInputStream
-
- read(byte[], int, int) -
Method in class org.apache.hadoop.net.SocketInputStream
-
- read(ByteBuffer) -
Method in class org.apache.hadoop.net.SocketInputStream
-
- read() -
Method in class org.apache.hadoop.security.SaslInputStream
- Reads the next byte of data from this input stream.
- read(byte[]) -
Method in class org.apache.hadoop.security.SaslInputStream
- Reads up to
b.length
bytes of data from this input stream into
an array of bytes.
- read(byte[], int, int) -
Method in class org.apache.hadoop.security.SaslInputStream
- Reads up to
len
bytes of data from this input stream into an
array of bytes.
- read(DataInput) -
Static method in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
- Read from in
- read() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads a typed bytes sequence and converts it to a Java object.
- read() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- ReadaheadPool - Class in org.apache.hadoop.io
- Manages a pool of threads which can issue readahead requests on file descriptors.
- ReadaheadPool.ReadaheadRequest - Interface in org.apache.hadoop.io
- An outstanding readahead request that has been submitted to
the pool.
- readaheadStream(String, FileDescriptor, long, long, long, ReadaheadPool.ReadaheadRequest) -
Method in class org.apache.hadoop.io.ReadaheadPool
- Issue a request to readahead on the given file descriptor.
- readAndProcess() -
Method in class org.apache.hadoop.ipc.Server.Connection
-
- readArray(ArrayWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readArray() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readBool(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readBool(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readBool(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a boolean from serialized record.
- readBool(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readBool() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the boolean following a
Type.BOOL
code.
- readBool(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readBoolean(BooleanWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readBoolean() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readBuffer(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readBuffer(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readBuffer(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read byte array from serialized record.
- readBuffer(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readBuffer(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readByte(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readByte(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readByte(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a byte from serialized record.
- readByte(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readByte() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the byte following a
Type.BYTE
code.
- readByte(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readByte(ByteWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readByte() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readBytes() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the bytes following a
Type.BYTES
code.
- readBytes(BytesWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readBytes() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readChar() -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- readChunk(long, byte[], int, int, byte[]) -
Method in class org.apache.hadoop.fs.FSInputChecker
- Reads in next checksum chunk data into
buf
at offset
and checksum into checksum
.
- readCompressedByteArray(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- readCompressedString(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- readCompressedStringArray(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- readDouble(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Parse a double from a byte array.
- readDouble(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readDouble(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readDouble(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a double-precision number from serialized record.
- readDouble(byte[], int) -
Static method in class org.apache.hadoop.record.Utils
- Parse a double from a byte array.
- readDouble(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readDouble() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the double following a
Type.DOUBLE
code.
- readDouble(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readDouble(DoubleWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readDouble() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readEnum(DataInput, Class<T>) -
Static method in class org.apache.hadoop.io.WritableUtils
- Read an Enum value from DataInput, Enums are read and written
using String values.
- readFields(DataInput) -
Method in class org.apache.hadoop.conf.Configuration
-
- readFields(DataInput) -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
- readFields(DataInput) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
- readFields(DataInput) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- readFields(DataInput) -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
- readFields(DataInput) -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- readFields(DataInput) -
Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- readFields(DataInput) -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
- Read the two integers.
- readFields(DataInput) -
Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.BlockLocation
- Implement readFields of Writable
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.ContentSummary
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.FileStatus
-
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.permission.FsPermission
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.io.AbstractMapWritable
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.io.ArrayWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.BooleanWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.BytesWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.ByteWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.CompressedWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.DoubleWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.FloatWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.GenericWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.IntWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.LongWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.MapWritable
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.io.MD5Hash
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.NullWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.ObjectWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.SortedMapWritable
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.io.Text
- deserialize
- readFields(DataInput) -
Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.UTF8
- Deprecated.
- readFields(DataInput) -
Method in class org.apache.hadoop.io.VersionedWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.VIntWritable
-
- readFields(DataInput) -
Method in class org.apache.hadoop.io.VLongWritable
-
- readFields(DataInput) -
Method in interface org.apache.hadoop.io.Writable
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.ClusterStatus
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.Counters.Group
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.Counters
- Read a set of groups.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.FileSplit
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.JobProfile
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.JobQueueInfo
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.JobStatus
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.JvmTask
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) -
Method in interface org.apache.hadoop.mapred.lib.db.DBWritable
- Reads the fields of the object from the
ResultSet
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.QueueAclsInfo
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.Task
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.TaskReport
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.Counter
- Read the binary representation of the counter
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.Counters
- Read a set of groups.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.ID
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.JobID
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) -
Method in interface org.apache.hadoop.mapreduce.lib.db.DBWritable
- Reads the fields of the object from the
ResultSet
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
- readFields(DataInput) -
Method in class org.apache.hadoop.mapreduce.TaskID
-
- readFields(DataInput) -
Method in class org.apache.hadoop.record.Record
-
- readFields(DataInput) -
Method in class org.apache.hadoop.security.authorize.AccessControlList
-
- readFields(DataInput) -
Method in class org.apache.hadoop.security.Credentials
- Loads all the keys
- readFields(DataInput) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- readFields(DataInput) -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
-
- readFields(DataInput) -
Method in class org.apache.hadoop.security.token.Token
- Deserialize the fields of this object from
in
.
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.Filter
-
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.Key
-
- readFields(DataInput) -
Method in class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
- readFieldsCompressed(DataInput) -
Method in class org.apache.hadoop.io.CompressedWritable
- Subclasses implement this instead of
CompressedWritable.readFields(DataInput)
.
- readFileChannelFully(FileChannel, byte[], int, int) -
Static method in class org.apache.hadoop.io.IOUtils
- Reads len bytes in a loop using the channel of the stream
- readFloat(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Parse a float from a byte array.
- readFloat(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readFloat(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readFloat(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a single-precision float from serialized record.
- readFloat(byte[], int) -
Static method in class org.apache.hadoop.record.Utils
- Parse a float from a byte array.
- readFloat(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readFloat() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the float following a
Type.FLOAT
code.
- readFloat(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readFloat(FloatWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readFloat() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readFully(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- readFully(long, byte[]) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- readFully(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.FSDataInputStream
-
- readFully(long, byte[]) -
Method in class org.apache.hadoop.fs.FSDataInputStream
-
- readFully(InputStream, byte[], int, int) -
Static method in class org.apache.hadoop.fs.FSInputChecker
- A utility function that tries to read up to
len
bytes from
stm
- readFully(long, byte[], int, int) -
Method in class org.apache.hadoop.fs.FSInputStream
-
- readFully(long, byte[]) -
Method in class org.apache.hadoop.fs.FSInputStream
-
- readFully(long, byte[], int, int) -
Method in interface org.apache.hadoop.fs.PositionedReadable
- Read the specified number of bytes, from a given
position within a file.
- readFully(long, byte[]) -
Method in interface org.apache.hadoop.fs.PositionedReadable
- Read number of bytes equalt to the length of the buffer, from a given
position within a file.
- readFully(InputStream, byte[], int, int) -
Static method in class org.apache.hadoop.io.IOUtils
- Reads len bytes in a loop.
- readInt(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Parse an integer from a byte array.
- readInt(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readInt(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readInt(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read an integer from serialized record.
- readInt(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readInt() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the integer following a
Type.INT
code.
- readInt(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readInt(IntWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readInt() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readKeyValue() -
Method in class org.apache.hadoop.streaming.io.OutputReader
- Read the next key/value pair outputted by the client.
- readKeyValue() -
Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
-
- readKeyValue() -
Method in class org.apache.hadoop.streaming.io.TextOutputReader
-
- readKeyValue() -
Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
-
- readLine(LineReader, Text) -
Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
- Read a utf8 encoded line from a data input stream.
- readLine(LineReader, Text) -
Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated. use
StreamKeyValUtil.readLine(LineReader, Text)
- readLine(Text, int, int) -
Method in class org.apache.hadoop.util.LineReader
- Read one line from the InputStream into the given Text.
- readLine(Text, int) -
Method in class org.apache.hadoop.util.LineReader
- Read from the InputStream into the given Text.
- readLine(Text) -
Method in class org.apache.hadoop.util.LineReader
- Read from the InputStream into the given Text.
- readList() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the list following a
Type.LIST
code.
- readLong(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Parse a long from a byte array.
- readLong(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readLong(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readLong(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a long integer from serialized record.
- readLong(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readLong() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the long following a
Type.LONG
code.
- readLong(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readLong(LongWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readLong() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readMap() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the map following a
Type.MAP
code.
- readMap(MapWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readMap() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readMapHeader() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the header following a
Type.MAP
code.
- readObject(DataInput, Configuration) -
Static method in class org.apache.hadoop.io.ObjectWritable
- Read a
Writable
, String
, primitive type, or an array of
the preceding.
- readObject(DataInput, ObjectWritable, Configuration) -
Static method in class org.apache.hadoop.io.ObjectWritable
- Read a
Writable
, String
, primitive type, or an array of
the preceding.
- readRAMFiles(DataInput, RAMDirectory) -
Static method in class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
- Read a number of files from a data input to a ram directory.
- readRaw() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads a typed bytes sequence.
- readRawBool() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.BOOL
code.
- readRawByte() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw byte following a
Type.BYTE
code.
- readRawBytes(int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a custom code.
- readRawBytes() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.BYTES
code.
- readRawDouble() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.DOUBLE
code.
- readRawFloat() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.FLOAT
code.
- readRawInt() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.INT
code.
- readRawList() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.LIST
code.
- readRawLong() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.LONG
code.
- readRawMap() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.MAP
code.
- readRawString() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.STRING
code.
- readRawVector() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the raw bytes following a
Type.VECTOR
code.
- readSortedMap(SortedMapWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readSortedMap() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readSplitMetaInfo(JobID, FileSystem, Configuration, Path) -
Static method in class org.apache.hadoop.mapreduce.split.SplitMetaInfoReader
-
- readState(String) -
Static method in class org.apache.hadoop.contrib.failmon.PersistentState
- Read the state of parsing for all open log files from a property
file.
- readString(DataInput) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Read a String as a VInt n, followed by n Bytes in Text format.
- readString(DataInput) -
Static method in class org.apache.hadoop.io.Text
- Read a UTF8 encoded string from in
- readString(DataInput) -
Static method in class org.apache.hadoop.io.UTF8
- Deprecated. Read a UTF-8 encoded string.
- readString(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- readString(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readString(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- readString(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Read a UTF-8 encoded string from serialized record.
- readString(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- readString() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the string following a
Type.STRING
code.
- readString(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- readStringArray(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- readText(Text) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readText() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readTokenStorageFile(Path, Configuration) -
Static method in class org.apache.hadoop.security.Credentials
- Convenience method for reading a token storage file, and loading the Tokens
therein in the passed UGI
- readTokenStorageStream(DataInputStream) -
Method in class org.apache.hadoop.security.Credentials
- Convenience method for reading a token storage file directly from a
datainputstream
- readType() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads a type byte and returns the corresponding
Type
.
- readType() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readUnsignedShort(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Parse an unsigned short from a byte array.
- readVector() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the vector following a
Type.VECTOR
code.
- readVectorHeader() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Reads the header following a
Type.VECTOR
code.
- readVInt(DataInput) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Decoding the variable-length integer.
- readVInt(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Reads a zero-compressed encoded integer from a byte array and returns it.
- readVInt(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
- Reads a zero-compressed encoded integer from input stream and returns it.
- readVInt(byte[], int) -
Static method in class org.apache.hadoop.record.Utils
- Reads a zero-compressed encoded integer from a byte array and returns it.
- readVInt(DataInput) -
Static method in class org.apache.hadoop.record.Utils
- Reads a zero-compressed encoded integer from a stream and returns it.
- readVInt(VIntWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readVInt() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readVLong(DataInput) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Decoding the variable-length integer.
- readVLong(byte[], int) -
Static method in class org.apache.hadoop.io.WritableComparator
- Reads a zero-compressed encoded long from a byte array and returns it.
- readVLong(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
- Reads a zero-compressed encoded long from input stream and returns it.
- readVLong(byte[], int) -
Static method in class org.apache.hadoop.record.Utils
- Reads a zero-compressed encoded long from a byte array and returns it.
- readVLong(DataInput) -
Static method in class org.apache.hadoop.record.Utils
- Reads a zero-compressed encoded long from a stream and return it.
- readVLong(VLongWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readVLong() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readWritable(Writable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- readWritable() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- READY -
Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- Record() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- Record - Class in org.apache.hadoop.record
- Abstract class that is extended by generated classes.
- Record() -
Constructor for class org.apache.hadoop.record.Record
-
- RECORD_INPUT -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RECORD_OUTPUT -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RECORD_SEPARATOR -
Static variable in class org.apache.hadoop.contrib.failmon.LocalStore
-
- RECORD_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- RecordComparator - Class in org.apache.hadoop.record
- A raw record comparator base class
- RecordComparator(Class<? extends WritableComparable>) -
Constructor for class org.apache.hadoop.record.RecordComparator
- Construct a raw
Record
comparison implementation.
- RecordInput - Interface in org.apache.hadoop.record
- Interface that all the Deserializers have to implement.
- RecordList() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- RecordOutput - Interface in org.apache.hadoop.record
- Interface that alll the serializers have to implement.
- RecordReader<K,V> - Interface in org.apache.hadoop.mapred
RecordReader
reads <key, value> pairs from an
InputSplit
.- RecordReader<KEYIN,VALUEIN> - Class in org.apache.hadoop.mapreduce
- The record reader breaks the data into key/value pairs for input to the
Mapper
. - RecordReader() -
Constructor for class org.apache.hadoop.mapreduce.RecordReader
-
- RecordTypeInfo - Class in org.apache.hadoop.record.meta
- A record's Type Information object which can read/write itself.
- RecordTypeInfo() -
Constructor for class org.apache.hadoop.record.meta.RecordTypeInfo
- Create an empty RecordTypeInfo object.
- RecordTypeInfo(String) -
Constructor for class org.apache.hadoop.record.meta.RecordTypeInfo
- Create a RecordTypeInfo object representing a record with the given name
- RecordWriter<K,V> - Interface in org.apache.hadoop.mapred
RecordWriter
writes the output <key, value> pairs
to an output file.- RecordWriter<K,V> - Class in org.apache.hadoop.mapreduce
RecordWriter
writes the output <key, value> pairs
to an output file.- RecordWriter() -
Constructor for class org.apache.hadoop.mapreduce.RecordWriter
-
- recoverJobHistoryFile(JobConf, Path) -
Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
- Selects one of the two files generated as a part of recovery.
- redCmd_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, IntermediateForm>, Reporter) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
-
- reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, Text>, Reporter) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
- reduce(Object, Iterator, OutputCollector, Reporter) -
Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- reduce(Object, Iterator, OutputCollector, Reporter) -
Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- reduce(BooleanWritable, Iterator<LongWritable>, OutputCollector<WritableComparable<?>, Writable>, Reporter) -
Method in class org.apache.hadoop.examples.PiEstimator.PiReducer
- Accumulate number of points inside/outside results from the mappers.
- reduce(SecondarySort.IntPair, Iterable<IntWritable>, Reducer<SecondarySort.IntPair, IntWritable, Text, IntWritable>.Context) -
Method in class org.apache.hadoop.examples.SecondarySort.Reduce
-
- reduce(IntWritable, Iterator<NullWritable>, OutputCollector<NullWritable, NullWritable>, Reporter) -
Method in class org.apache.hadoop.examples.SleepJob
-
- reduce(Text, Iterable<IntWritable>, Reducer<Text, IntWritable, Text, IntWritable>.Context) -
Method in class org.apache.hadoop.examples.WordCount.IntSumReducer
-
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
- Combines values for a given key.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
- Do nothing.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
-
- reduce(Object, Iterator, OutputCollector, Reporter) -
Method in class org.apache.hadoop.mapred.lib.ChainReducer
- Chains the
reduce(...)
method of the Reducer with the
map(...)
methods of the Mappers in the chain.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
-
- reduce(K, Iterator<V>, OutputCollector<K, V>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.IdentityReducer
- Writes all keys and values directly to output.
- reduce(K, Iterator<LongWritable>, OutputCollector<K, LongWritable>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.LongSumReducer
-
- reduce(K2, Iterator<V2>, OutputCollector<K3, V3>, Reporter) -
Method in interface org.apache.hadoop.mapred.Reducer
- Reduces values for a given key.
- reduce(Text, Iterable<Text>, Reducer<Text, Text, Text, Text>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
-
- reduce(Key, Iterable<IntWritable>, Reducer<Key, IntWritable, Key, IntWritable>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer
-
- reduce(KEY, Iterable<LongWritable>, Reducer<KEY, LongWritable, KEY, LongWritable>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer
-
- reduce(KEYIN, Iterable<VALUEIN>, Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Reducer
- This method is called once for each key.
- reduce(Object, Iterator, OutputCollector, Reporter) -
Method in class org.apache.hadoop.streaming.PipeReducer
-
- REDUCE_CLASS_ATTR -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- REDUCE_OUTPUT_KEY_VALUE_SPEC -
Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
-
- ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
- The context passed to the
Reducer
. - ReduceContext(Configuration, TaskAttemptID, RawKeyValueIterator, Counter, Counter, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, RawComparator<KEYIN>, Class<KEYIN>, Class<VALUEIN>) -
Constructor for class org.apache.hadoop.mapreduce.ReduceContext
-
- ReduceContext.ValueIterable - Class in org.apache.hadoop.mapreduce
-
- ReduceContext.ValueIterable() -
Constructor for class org.apache.hadoop.mapreduce.ReduceContext.ValueIterable
-
- ReduceContext.ValueIterator - Class in org.apache.hadoop.mapreduce
-
- ReduceContext.ValueIterator() -
Constructor for class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
-
- reduceDebugSpec_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- reduceProgress() -
Method in class org.apache.hadoop.mapred.JobStatus
-
- reduceProgress() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Get the progress of the job's reduce-tasks, as a float between 0.0
and 1.0.
- reduceProgress() -
Method in class org.apache.hadoop.mapreduce.Job
- Get the progress of the job's reduce-tasks, as a float between 0.0
and 1.0.
- Reducer<K2,V2,K3,V3> - Interface in org.apache.hadoop.mapred
- Reduces a set of intermediate values which share a key to a smaller set of
values.
- Reducer<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
- Reduces a set of intermediate values which share a key to a smaller set of
values.
- Reducer() -
Constructor for class org.apache.hadoop.mapreduce.Reducer
-
- Reducer.Context - Class in org.apache.hadoop.mapreduce
-
- Reducer.Context(Configuration, TaskAttemptID, RawKeyValueIterator, Counter, Counter, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, RawComparator<KEYIN>, Class<KEYIN>, Class<VALUEIN>) -
Constructor for class org.apache.hadoop.mapreduce.Reducer.Context
-
- ReflectionUtils - Class in org.apache.hadoop.util
- General reflection utils
- ReflectionUtils() -
Constructor for class org.apache.hadoop.util.ReflectionUtils
-
- refresh(Configuration, PolicyProvider) -
Static method in class org.apache.hadoop.security.authorize.ServiceAuthorizationManager
-
- refresh() -
Method in class org.apache.hadoop.security.Groups
- Refresh all user-to-groups mappings.
- refresh() -
Method in class org.apache.hadoop.util.HostsFileReader
-
- RefreshAuthorizationPolicyProtocol - Interface in org.apache.hadoop.security.authorize
- Protocol which is used to refresh the authorization policy in use currently.
- refreshMBeans() -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- refreshMBeans() -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- refreshMBeans() -
Method in interface org.apache.hadoop.metrics2.MetricsSystemMXBean
- Force a refresh of MBeans
- refreshNodes() -
Method in interface org.apache.hadoop.mapred.AdminOperationsProtocol
- Refresh the node list at the
JobTracker
- refreshNodes() -
Method in class org.apache.hadoop.mapred.JobTracker
- Rereads the config to get hosts and exclude list file names.
- refreshQueues() -
Method in interface org.apache.hadoop.mapred.AdminOperationsProtocol
- Refresh the queue acls in use currently.
- refreshQueues() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- refreshServiceAcl() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- refreshServiceAcl() -
Method in interface org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol
- Refresh the service-level authorization policy in-effect.
- refreshSuperUserGroupsConfiguration() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- refreshSuperUserGroupsConfiguration() -
Method in class org.apache.hadoop.mapred.tools.MRAdmin
- refreshSuperUserGroupsConfiguration
JobTracker
.
- refreshSuperUserGroupsConfiguration() -
Static method in class org.apache.hadoop.security.authorize.ProxyUsers
- reread the conf and get new values for "hadoop.proxyuser.*.groups/hosts"
- refreshSuperUserGroupsConfiguration(Configuration) -
Static method in class org.apache.hadoop.security.authorize.ProxyUsers
- refresh configuration
- refreshSuperUserGroupsConfiguration() -
Method in interface org.apache.hadoop.security.RefreshUserMappingsProtocol
- Refresh superuser proxy group list
- RefreshUserMappingsProtocol - Interface in org.apache.hadoop.security
- Protocol use
- refreshUserToGroupsMappings() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- refreshUserToGroupsMappings() -
Method in interface org.apache.hadoop.security.RefreshUserMappingsProtocol
- Refresh user to group mappings.
- RegexFilter - Class in org.apache.hadoop.metrics2.filter
- A regex pattern filter for metrics
- RegexFilter() -
Constructor for class org.apache.hadoop.metrics2.filter.RegexFilter
-
- RegexMapper<K> - Class in org.apache.hadoop.mapred.lib
- A
Mapper
that extracts text matching a regular expression. - RegexMapper() -
Constructor for class org.apache.hadoop.mapred.lib.RegexMapper
-
- regexpEscape(String) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- register(String, String, T) -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- register(String, String, T) -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- register(MetricsSystem.Callback) -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- register(String, String, T) -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- register(String, String, T) -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- register(MetricsSystem.Callback) -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- register(String, String, T) -
Method in interface org.apache.hadoop.metrics2.MetricsSystem
- Register a metrics source
- register(String, String, T) -
Method in interface org.apache.hadoop.metrics2.MetricsSystem
- Register a metrics sink
- register(MetricsSystem.Callback) -
Method in interface org.apache.hadoop.metrics2.MetricsSystem
- Register a callback interface for JMX events
- register(String, String, Object) -
Static method in class org.apache.hadoop.metrics2.util.MBeans
- Register the MBean using our standard MBeanName format
"hadoop:service=
,name="
Where the and are the supplied parameters
- registerDelegationTokensForRenewal(JobID, Credentials, Configuration) -
Static method in class org.apache.hadoop.mapreduce.security.token.DelegationTokenRenewal
-
- registerMBean(String, String, Object) -
Static method in class org.apache.hadoop.metrics.util.MBeanUtil
- Deprecated. Register the MBean using our standard MBeanName format
"hadoop:service=
,name="
Where the and are the supplied parameters
- registerNotification(JobConf, JobStatus) -
Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- registerSource(String, String, T) -
Static method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
- Common static method to register a source
- registerUpdater(Updater) -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Registers a callback to be called at regular time intervals, as
determined by the implementation-class specific configuration.
- registerUpdater(Updater) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Registers a callback to be called at time intervals determined by
the configuration.
- registerUpdater(Updater) -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- registry -
Variable in class org.apache.hadoop.metrics2.lib.AbstractMetricsSource
-
- registry() -
Method in class org.apache.hadoop.metrics2.lib.AbstractMetricsSource
-
- reinit(Configuration) -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- reinit(Configuration) -
Method in interface org.apache.hadoop.io.compress.Compressor
- Prepare the compressor to be used in a new stream with settings defined in
the given Configuration
- reinit(Configuration) -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Prepare the compressor to be used in a new stream with settings defined in
the given Configuration
- reinit(Configuration) -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater
- reinit the compressor with the given configuration.
- reinit(Configuration) -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
- Prepare the compressor to be used in a new stream with settings defined in
the given Configuration.
- ReInit(InputStream) -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- ReInit(InputStream, String) -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- ReInit(Reader) -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- ReInit(RccTokenManager) -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- ReInit(SimpleCharStream) -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- ReInit(SimpleCharStream, int) -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- ReInit(Reader, int, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(Reader, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(Reader) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream, String, int, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream, int, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream, String) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream, String, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- ReInit(InputStream, int, int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- release() -
Method in class org.apache.hadoop.filecache.TaskDistributedCacheManager
- Releases the cached files/archives, so that space
can be reclaimed by the
TrackerDistributedCacheManager
.
- release(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
- Deprecated.
- reloadConfiguration() -
Method in class org.apache.hadoop.conf.Configuration
- Reload configuration from previously added resources.
- reloginFromKeytab() -
Method in class org.apache.hadoop.security.UserGroupInformation
- Re-Login a user in from a keytab file.
- reloginFromTicketCache() -
Method in class org.apache.hadoop.security.UserGroupInformation
- Re-Login a user in from the ticket cache.
- RemoteException - Exception in org.apache.hadoop.ipc
-
- RemoteException(String, String) -
Constructor for exception org.apache.hadoop.ipc.RemoteException
-
- remove() -
Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- remove(Object) -
Method in class org.apache.hadoop.io.MapWritable
-
- remove(Object) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- remove() -
Method in class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
-
- remove() -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Removes, from the buffered data table, all rows having tags
that equal the tags that have been set on this record.
- remove(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Called by MetricsRecordImpl.remove().
- remove() -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Removes the row, if it exists, in the buffered data table having tags
that equal the tags that have been set on this record.
- remove(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.NullContext
- Deprecated. Do-nothing version of remove
- remove(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
- Deprecated. Do-nothing version of remove
- remove() -
Method in class org.apache.hadoop.metrics2.util.TryIterator
- Guaranteed to throw UnsupportedOperationException
- remove(Node) -
Method in class org.apache.hadoop.net.NetworkTopology
- Remove a node
Update node counter & rack counter if neccessary
- remove(Node) -
Method in class org.apache.hadoop.net.NetworkTopologyWithNodeGroup
- Remove a node
Update node counter and rack counter if necessary
- removeAttribute(String) -
Method in class org.apache.hadoop.metrics.ContextFactory
- Deprecated. Removes the named attribute if it exists.
- removeDelegationTokenRenewalForJob(JobID) -
Static method in class org.apache.hadoop.mapreduce.security.token.DelegationTokenRenewal
- removing DT for completed jobs
- removeJobInProgressListener(JobInProgressListener) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- RemoveScheme - Interface in org.apache.hadoop.util.bloom
- Defines the different remove scheme for retouched Bloom filters.
- removeSuffix(String, String) -
Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
- Removes a suffix from a filename, if it has it.
- removeTag(String) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Removes any tag of the specified name.
- removeTag(String) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Removes any tag of the specified name.
- removeTaskDistributedCacheManager(JobID) -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
-
- removeTokenForJob(String) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager
- Remove the cached job token of a job from cache
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- Rename files/dirs
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Renames Path src to Path dst.
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Renames Path src to Path dst.
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- rename(Path, Path) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- rename(FileSystem, String, String) -
Static method in class org.apache.hadoop.io.MapFile
- Renames an existing map directory.
- renameFile(String, String) -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- renew(Token<?>, Configuration) -
Method in class org.apache.hadoop.mapred.JobClient.Renewer
-
- renew(Configuration) -
Method in class org.apache.hadoop.security.token.Token
- Renew this delegation token
- renew(Token<?>, Configuration) -
Method in class org.apache.hadoop.security.token.Token.TrivialRenewer
-
- renew(Token<?>, Configuration) -
Method in class org.apache.hadoop.security.token.TokenRenewer
- Renew the given token.
- renewDelegationToken(Token<DelegationTokenIdentifier>) -
Method in class org.apache.hadoop.mapred.JobClient
- Renew a delegation token
- renewDelegationToken(Token<DelegationTokenIdentifier>) -
Method in class org.apache.hadoop.mapred.JobTracker
- Renew a delegation token to extend its lifetime.
- renewDelegationToken(Token<DelegationTokenIdentifier>) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
-
- renewToken(Token<TokenIdent>, String) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
- Renew a delegation token.
- replaceFile(File, File) -
Static method in class org.apache.hadoop.fs.FileUtil
- Move the src file to the name specified by target.
- replay(X) -
Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- replay(TupleWritable) -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- replay(V) -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- replay(U) -
Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- replay(T) -
Method in interface org.apache.hadoop.mapred.join.ResetableIterator
- Assign last value returned to actual.
- replay(X) -
Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- report() -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
- log the counters
- report() -
Method in class org.apache.hadoop.fs.HardLink.LinkStats
-
- reportChecksumFailure(Path, FSDataInputStream, long, FSDataInputStream, long) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- Report a checksum error to the file system.
- reportChecksumFailure(Path, FSDataInputStream, long, FSDataInputStream, long) -
Method in class org.apache.hadoop.fs.LocalFileSystem
- Moves files to a bad file directory on the same device, so that their
storage will not be reused.
- reportCleanupTIPs(boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a vector of cleanup TaskInProgress objects
- reportCRCError() -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
- reportDiagnosticInfo(TaskAttemptID, String, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
- Called when the task dies before completion, and we want to report back
diagnostic info
- reportDiagnosticInfo(TaskAttemptID, String, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Report error messages back to parent.
- reporter -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- reporter -
Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- reporter -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- Reporter - Interface in org.apache.hadoop.mapred
- A facility for Map-Reduce applications to report progress and update
counters, status information etc.
- reporter -
Variable in class org.apache.hadoop.mapred.Task.CombinerRunner
-
- reporter -
Variable in class org.apache.hadoop.mapreduce.ReduceContext
-
- reportFatalError(TaskAttemptID, Throwable, String) -
Method in class org.apache.hadoop.mapred.Task
- Report a fatal error to the parent (task) tracker.
- reportNextRecordRange(TaskUmbilicalProtocol, long) -
Method in class org.apache.hadoop.mapred.Task
- Reports the next executing record range to TaskTracker.
- reportNextRecordRange(TaskAttemptID, SortedRanges.Range, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
-
- reportNextRecordRange(TaskAttemptID, SortedRanges.Range, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Report the record range which is going to process next by the Task.
- reportSetupTIPs(boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a vector of setup TaskInProgress objects
- reportTasksInProgress(boolean, boolean) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Return a vector of completed TaskInProgress objects
- reportTaskTrackerError(String, String, String) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- requiresLayout() -
Method in class org.apache.hadoop.log.metrics.EventCounter
-
- reserveSlots(TaskType, JobInProgress, int) -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
- Reserve specified number of slots for a given
job
.
- reserveSpaceWithCheckSum(Path, long) -
Method in class org.apache.hadoop.fs.InMemoryFileSystem
- Deprecated. Register a file with its size.
- reserveTaskTracker(TaskTracker, TaskType, int) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- reset() -
Method in class org.apache.hadoop.contrib.failmon.MonitorJob
-
- reset() -
Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- reset() -
Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
-
- reset() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
- Reset the counts of bytes to 0.
- reset() -
Method in class org.apache.hadoop.fs.FSInputChecker
-
- reset() -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- reset() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- reset() -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- reset() -
Method in interface org.apache.hadoop.io.compress.Compressor
- Resets compressor so that a new set of input data can be processed.
- reset() -
Method in interface org.apache.hadoop.io.compress.Decompressor
- Resets decompressor and input and output buffers so that a new set of
input data can be processed.
- reset() -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- reset() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Resets compressor so that a new set of input data can be processed.
- reset() -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
-
- reset() -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
- Resets everything, including the input buffer, regardless of whether the
current gzip substream is finished.
- reset() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- reset() -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
- Resets everything including the input buffers (user and direct).
- reset(byte[], int) -
Method in class org.apache.hadoop.io.DataInputBuffer
- Resets the data that the buffer reads.
- reset(byte[], int, int) -
Method in class org.apache.hadoop.io.DataInputBuffer
- Resets the data that the buffer reads.
- reset() -
Method in class org.apache.hadoop.io.DataOutputBuffer
- Resets the buffer to empty.
- reset(byte[], int) -
Method in class org.apache.hadoop.io.InputBuffer
- Resets the data that the buffer reads.
- reset(byte[], int, int) -
Method in class org.apache.hadoop.io.InputBuffer
- Resets the data that the buffer reads.
- reset() -
Method in class org.apache.hadoop.io.MapFile.Reader
- Re-positions the reader before its first key.
- reset() -
Method in class org.apache.hadoop.io.OutputBuffer
- Resets the buffer to empty.
- reset() -
Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- reset() -
Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- reset() -
Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- reset() -
Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- reset() -
Method in interface org.apache.hadoop.mapred.join.ResetableIterator
- Set iterator to return to the start of its range.
- reset() -
Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
- reset the aggregator
- reset() -
Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
- reset the aggregator
- reset() -
Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
- reset the aggregator
- reset(BytesWritable) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- reset(BytesWritable) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- reset() -
Method in class org.apache.hadoop.metrics2.util.SampleStat.MinMax
-
- reset(SampleStat.MinMax) -
Method in class org.apache.hadoop.metrics2.util.SampleStat.MinMax
-
- reset() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- reset() -
Method in class org.apache.hadoop.record.Buffer
- Reset the buffer to 0 size
- reset() -
Method in class org.apache.hadoop.util.DataChecksum
-
- reset() -
Method in class org.apache.hadoop.util.PureJavaCrc32
-
- ResetableIterator - Interface in org.apache.hadoop.contrib.utils.join
- This defines an iterator interface that will help the reducer class
re-group its input by source tags.
- ResetableIterator<T extends Writable> - Interface in org.apache.hadoop.mapred.join
- This defines an interface to a stateful Iterator that can replay elements
added to it directly.
- ResetableIterator.EMPTY<U extends Writable> - Class in org.apache.hadoop.mapred.join
-
- ResetableIterator.EMPTY() -
Constructor for class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- resetChecksumChunk(int) -
Method in class org.apache.hadoop.fs.FSOutputSummer
- Resets existing buffer with a new one of the specified size.
- resetMinMax() -
Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
- Deprecated. Reset the min max values
- resetMinMax() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableStat
- Reset the all time min max of the metric
- resetSchedulingOpportunities() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- resetState() -
Method in class org.apache.hadoop.io.compress.BlockDecompressorStream
-
- resetState() -
Method in class org.apache.hadoop.io.compress.CompressionInputStream
- Reset the decompressor to its initial state and discard any buffered data,
as the underlying stream may have been repositioned.
- resetState() -
Method in class org.apache.hadoop.io.compress.CompressionOutputStream
- Reset the compression to the initial state.
- resetState() -
Method in class org.apache.hadoop.io.compress.CompressorStream
-
- resetState() -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- resetState() -
Method in class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- resolve(List<String>) -
Method in class org.apache.hadoop.net.CachedDNSToSwitchMapping
-
- resolve(List<String>) -
Method in interface org.apache.hadoop.net.DNSToSwitchMapping
- Resolves a list of DNS-names/IP-addresses and returns back a list of
switch information (network paths).
- resolve(String) -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
- Resolves a given identifier.
- resolveAndAddToTopology(String) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- ResourceCalculatorPlugin - Class in org.apache.hadoop.util
- Plugin to calculate resource information on the system.
- ResourceCalculatorPlugin() -
Constructor for class org.apache.hadoop.util.ResourceCalculatorPlugin
-
- ResourceCalculatorPlugin.ProcResourceValues - Class in org.apache.hadoop.util
-
- ResourceCalculatorPlugin.ProcResourceValues(long, long, long) -
Constructor for class org.apache.hadoop.util.ResourceCalculatorPlugin.ProcResourceValues
-
- resume() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
- resume the suspended thread
- resume() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
- resume the suspended thread
- RetouchedBloomFilter - Class in org.apache.hadoop.util.bloom
- Implements a retouched Bloom filter, as defined in the CoNEXT 2006 paper.
- RetouchedBloomFilter() -
Constructor for class org.apache.hadoop.util.bloom.RetouchedBloomFilter
- Default constructor - use with readFields
- RetouchedBloomFilter(int, int, int) -
Constructor for class org.apache.hadoop.util.bloom.RetouchedBloomFilter
- Constructor
- retrieveBlock(Block, long) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- retrieveINode(Path) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- retrievePassword(JobTokenIdentifier) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager
- Look up the token password/secret for the given job token identifier.
- retrievePassword(TokenIdent) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
-
- retrievePassword(T) -
Method in class org.apache.hadoop.security.token.SecretManager
- Retrieve the password for the given token identifier.
- retrieveTokenSecret(String) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenSecretManager
- Look up the token password/secret for the given jobId.
- RETRY_FOREVER -
Static variable in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying forever.
- retryByException(RetryPolicy, Map<Class<? extends Exception>, RetryPolicy>) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Set a default policy with some explicit handlers for specific exceptions.
- retryByRemoteException(RetryPolicy, Map<Class<? extends Exception>, RetryPolicy>) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
A retry policy for RemoteException
Set a default policy with some explicit handlers for specific exceptions.
- RetryPolicies - Class in org.apache.hadoop.io.retry
-
A collection of useful implementations of
RetryPolicy
. - RetryPolicies() -
Constructor for class org.apache.hadoop.io.retry.RetryPolicies
-
- RetryPolicies.MultipleLinearRandomRetry - Class in org.apache.hadoop.io.retry
- Given pairs of number of retries and sleep time (n0, t0), (n1, t1), ...,
the first n0 retries sleep t0 milliseconds on average,
the following n1 retries sleep t1 milliseconds on average, and so on.
- RetryPolicies.MultipleLinearRandomRetry(List<RetryPolicies.MultipleLinearRandomRetry.Pair>) -
Constructor for class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry
-
- RetryPolicies.MultipleLinearRandomRetry.Pair - Class in org.apache.hadoop.io.retry
- Pairs of numRetries and sleepSeconds
- RetryPolicies.MultipleLinearRandomRetry.Pair(int, int) -
Constructor for class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry.Pair
-
- RetryPolicy - Interface in org.apache.hadoop.io.retry
-
Specifies a policy for retrying method failures.
- RetryProxy - Class in org.apache.hadoop.io.retry
-
A factory for creating retry proxies.
- RetryProxy() -
Constructor for class org.apache.hadoop.io.retry.RetryProxy
-
- retryUpToMaximumCountWithFixedSleep(int, long, TimeUnit) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying a limited number of times, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
- retryUpToMaximumCountWithProportionalSleep(int, long, TimeUnit) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying a limited number of times, waiting a growing amount of time between attempts,
and then fail by re-throwing the exception.
- retryUpToMaximumTimeWithFixedSleep(long, long, TimeUnit) -
Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying for a maximum time, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
- RetryUtils - Class in org.apache.hadoop.io.retry
-
- RetryUtils() -
Constructor for class org.apache.hadoop.io.retry.RetryUtils
-
- returnCompressor(Compressor) -
Static method in class org.apache.hadoop.io.compress.CodecPool
- Return the
Compressor
to the pool.
- returnDecompressor(Decompressor) -
Static method in class org.apache.hadoop.io.compress.CodecPool
- Return the
Decompressor
to the pool.
- reverseDns(InetAddress, String) -
Static method in class org.apache.hadoop.net.DNS
- Returns the hostname associated with the specified IP address by the
provided nameserver.
- rewind() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Rewind to the first entry in the scanner.
- RIO_PREFIX -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- rjustify(String, int) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- rNums -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
- This array really shouldn't be here.
- ROOT -
Static variable in class org.apache.hadoop.net.NodeBase
-
- RoundRobinDistributionPolicy - Class in org.apache.hadoop.contrib.index.example
- Choose a shard for each insert in a round-robin fashion.
- RoundRobinDistributionPolicy() -
Constructor for class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
-
- RPC - Class in org.apache.hadoop.ipc
- A simple RPC mechanism.
- RPC.Server - Class in org.apache.hadoop.ipc
- An RPC Server.
- RPC.Server(Object, Configuration, String, int) -
Constructor for class org.apache.hadoop.ipc.RPC.Server
- Construct an RPC server.
- RPC.Server(Object, Configuration, String, int, int, boolean, SecretManager<? extends TokenIdentifier>) -
Constructor for class org.apache.hadoop.ipc.RPC.Server
- Construct an RPC server.
- RPC.VersionMismatch - Exception in org.apache.hadoop.ipc
- A version mismatch for the RPC protocol.
- RPC.VersionMismatch(String, long, long) -
Constructor for exception org.apache.hadoop.ipc.RPC.VersionMismatch
- Create a version mismatch exception
- RpcInstrumentation - Class in org.apache.hadoop.ipc.metrics
- The RPC metrics instrumentation
- RpcInstrumentation.Detailed - Class in org.apache.hadoop.ipc.metrics
- Use a separate source for detailed (per call) RPC metrics for
easy and efficient filtering
- rpcMetrics -
Variable in class org.apache.hadoop.ipc.Server
-
- rrClass -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- rrClass -
Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- rrConstructor -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- rrConstructor -
Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- rrCstrMap -
Static variable in class org.apache.hadoop.mapred.join.Parser.Node
-
- RTI_FILTER -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RTI_FILTER_FIELDS -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RTI_VAR -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- run() -
Method in class org.apache.hadoop.contrib.failmon.Executor
-
- run(Configuration, Path[], Path, int, Shard[]) -
Method in interface org.apache.hadoop.contrib.index.mapred.IIndexUpdater
- Create a Map/Reduce job configuration and run the Map/Reduce job to
analyze documents and update Lucene instances in parallel.
- run(Configuration, Path[], Path, int, Shard[]) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdater
-
- run(String[]) -
Method in class org.apache.hadoop.examples.dancing.DistributedPentomino
-
- run(String[]) -
Method in class org.apache.hadoop.examples.DBCountPageView
-
- run(String[]) -
Method in class org.apache.hadoop.examples.Grep
-
- run(String[]) -
Method in class org.apache.hadoop.examples.Join
- The main driver for sort program.
- run(String[]) -
Method in class org.apache.hadoop.examples.MultiFileWordCount
-
- run(String[]) -
Method in class org.apache.hadoop.examples.PiEstimator
- Parse arguments and then runs a map/reduce job.
- run(String[]) -
Method in class org.apache.hadoop.examples.RandomTextWriter
- This is the main routine for launching a distributed random write job.
- run(String[]) -
Method in class org.apache.hadoop.examples.RandomWriter
- This is the main routine for launching a distributed random write job.
- run(int, int, long, int, long, int) -
Method in class org.apache.hadoop.examples.SleepJob
-
- run(String[]) -
Method in class org.apache.hadoop.examples.SleepJob
-
- run(String[]) -
Method in class org.apache.hadoop.examples.Sort
- The main driver for sort program.
- run(String[]) -
Method in class org.apache.hadoop.examples.terasort.TeraGen
-
- run(String[]) -
Method in class org.apache.hadoop.examples.terasort.TeraSort
-
- run(String[]) -
Method in class org.apache.hadoop.examples.terasort.TeraValidate
-
- run() -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager.CleanupThread
-
- run(String[]) -
Method in class org.apache.hadoop.fs.FsShell
- run
- run(String[]) -
Method in class org.apache.hadoop.fs.s3.MigrationTool
-
- run(Path) -
Method in class org.apache.hadoop.fs.shell.Command
- Execute the command on the input path
- run(Path) -
Method in class org.apache.hadoop.fs.shell.Count
-
- run() -
Method in class org.apache.hadoop.mapred.HDFSMonitorThread
-
- run(String[]) -
Method in class org.apache.hadoop.mapred.JobClient
-
- run() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
- The main loop for the thread.
- run() -
Method in class org.apache.hadoop.mapred.JobHistory.HistoryCleaner
- Cleans up history data.
- run(String[]) -
Method in class org.apache.hadoop.mapred.lib.InputSampler
- Driver for InputSampler from the command line.
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) -
Method in class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
-
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) -
Method in interface org.apache.hadoop.mapred.MapRunnable
- Start mapping input <key, value> pairs.
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) -
Method in class org.apache.hadoop.mapred.MapRunner
-
- run(String[]) -
Method in class org.apache.hadoop.mapred.pipes.Submitter
-
- run(JobConf, TaskUmbilicalProtocol) -
Method in class org.apache.hadoop.mapred.Task
- Run this task as a part of the named job.
- run() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
- The communication thread handles communication with the parent (Task Tracker).
- run() -
Method in class org.apache.hadoop.mapred.TaskTracker
- The server retry loop.
- run(String[]) -
Method in class org.apache.hadoop.mapred.tools.MRAdmin
-
- run() -
Method in class org.apache.hadoop.mapred.UserLogCleaner
-
- run(Mapper<K1, V1, K2, V2>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingMapper
-
- run() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
- The main loop for the thread.
- run(Mapper<K1, V1, K2, V2>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
- Run the application's maps using a thread pool.
- run(String[]) -
Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
- Driver for InputSampler from the command line.
- run(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Mapper
- Expert users can override this method for more complete control over the
execution of the Mapper.
- run(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Reducer
- Advanced application writers can use the
Reducer.run(org.apache.hadoop.mapreduce.Reducer.Context)
method to
control how the reduce task works.
- run(String[]) -
Method in class org.apache.hadoop.streaming.DumpTypedBytes
- The main driver for
DumpTypedBytes
.
- run(String[]) -
Method in class org.apache.hadoop.streaming.LoadTypedBytes
- The main driver for
LoadTypedBytes
.
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) -
Method in class org.apache.hadoop.streaming.PipeMapRunner
-
- run(String[]) -
Method in class org.apache.hadoop.streaming.StreamJob
-
- run() -
Method in class org.apache.hadoop.util.Shell
- check to see if a command needs to be executed and execute if needed
- run(String[]) -
Method in interface org.apache.hadoop.util.Tool
- Execute the command with the given arguments.
- run(Configuration, Tool, String[]) -
Static method in class org.apache.hadoop.util.ToolRunner
- Runs the given
Tool
by Tool.run(String[])
, after
parsing with the given generic arguments.
- run(Tool, String[]) -
Static method in class org.apache.hadoop.util.ToolRunner
- Runs the
Tool
with its Configuration
.
- RUNA -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- runAll() -
Method in class org.apache.hadoop.fs.shell.Command
- For each source path, execute the command
- RUNB -
Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- runCommand(String[]) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
- Runs a shell command in the system and provides a StringBuffer
with the output of the command.
- runCommand(String) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
- Runs a shell command in the system and provides a StringBuffer
with the output of the command.
- RunJar - Class in org.apache.hadoop.util
- Run a Hadoop job jar.
- RunJar() -
Constructor for class org.apache.hadoop.util.RunJar
-
- runJob(JobConf) -
Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
- Submit/run a map/reduce job.
- runJob(JobConf) -
Static method in class org.apache.hadoop.mapred.JobClient
- Utility that submits a job, then polls for progress until the job is
complete.
- runJob(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Submit a job to the map/reduce cluster.
- runJobCleanupTask(TaskUmbilicalProtocol, Task.TaskReporter) -
Method in class org.apache.hadoop.mapred.Task
-
- runJobSetupTask(TaskUmbilicalProtocol, Task.TaskReporter) -
Method in class org.apache.hadoop.mapred.Task
-
- RUNNING -
Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- RUNNING -
Static variable in class org.apache.hadoop.mapred.JobStatus
-
- running -
Variable in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
-
- running_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- RunningJob - Interface in org.apache.hadoop.mapred
RunningJob
is the user-interface to query for details on a
running Map-Reduce job.- runningJobs() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- runningMaps() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- runningReduces() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- RunOnce - Class in org.apache.hadoop.contrib.failmon
- Runs a set of monitoring jobs once for the local node.
- RunOnce(String) -
Constructor for class org.apache.hadoop.contrib.failmon.RunOnce
-
- runSetup(String, String, Path, TaskUmbilicalProtocol) -
Method in class org.apache.hadoop.mapred.JobLocalizer
-
- runTaskCleanupTask(TaskUmbilicalProtocol, Task.TaskReporter) -
Method in class org.apache.hadoop.mapred.Task
-
S
- S3Credentials - Class in org.apache.hadoop.fs.s3
-
Extracts AWS credentials from the filesystem URI or configuration.
- S3Credentials() -
Constructor for class org.apache.hadoop.fs.s3.S3Credentials
-
- S3Exception - Exception in org.apache.hadoop.fs.s3
- Thrown if there is a problem communicating with Amazon S3.
- S3Exception(Throwable) -
Constructor for exception org.apache.hadoop.fs.s3.S3Exception
-
- S3FileSystem - Class in org.apache.hadoop.fs.s3
-
A block-based
FileSystem
backed by
Amazon S3. - S3FileSystem() -
Constructor for class org.apache.hadoop.fs.s3.S3FileSystem
-
- S3FileSystem(FileSystemStore) -
Constructor for class org.apache.hadoop.fs.s3.S3FileSystem
-
- S3FileSystemException - Exception in org.apache.hadoop.fs.s3
- Thrown when there is a fatal exception while using
S3FileSystem
. - S3FileSystemException(String) -
Constructor for exception org.apache.hadoop.fs.s3.S3FileSystemException
-
- S_IFBLK -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFCHR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFDIR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFIFO -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFLNK -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFMT -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFREG -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFSOCK -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IFWHT -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IRUSR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_ISGID -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_ISUID -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_ISVTX -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IWUSR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- S_IXUSR -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- safeGetCanonicalPath(File) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- SafeModeException - Exception in org.apache.hadoop.mapred
- This exception is thrown when the JobTracker is in safe mode.
- SafeModeException(String) -
Constructor for exception org.apache.hadoop.mapred.SafeModeException
- SafeModeException
- SampleStat - Class in org.apache.hadoop.metrics2.util
- Helper to compute running sample stats
- SampleStat() -
Constructor for class org.apache.hadoop.metrics2.util.SampleStat
- Construct a new running sample stat
- SampleStat.MinMax - Class in org.apache.hadoop.metrics2.util
- Helper to keep running min/max
- SampleStat.MinMax() -
Constructor for class org.apache.hadoop.metrics2.util.SampleStat.MinMax
-
- SASL_DEFAULT_REALM -
Static variable in class org.apache.hadoop.security.SaslRpcServer
-
- SASL_PROPS -
Static variable in class org.apache.hadoop.security.SaslRpcServer
-
- saslConnect(InputStream, OutputStream) -
Method in class org.apache.hadoop.security.SaslRpcClient
- Do client side SASL authentication with server via the given InputStream
and OutputStream
- SaslInputStream - Class in org.apache.hadoop.security
- A SaslInputStream is composed of an InputStream and a SaslServer (or
SaslClient) so that read() methods return data that are read in from the
underlying InputStream but have been additionally processed by the SaslServer
(or SaslClient) object.
- SaslInputStream(InputStream, SaslServer) -
Constructor for class org.apache.hadoop.security.SaslInputStream
- Constructs a SASLInputStream from an InputStream and a SaslServer
Note: if the specified InputStream or SaslServer is null, a
NullPointerException may be thrown later when they are used.
- SaslInputStream(InputStream, SaslClient) -
Constructor for class org.apache.hadoop.security.SaslInputStream
- Constructs a SASLInputStream from an InputStream and a SaslClient
Note: if the specified InputStream or SaslClient is null, a
NullPointerException may be thrown later when they are used.
- SaslOutputStream - Class in org.apache.hadoop.security
- A SaslOutputStream is composed of an OutputStream and a SaslServer (or
SaslClient) so that write() methods first process the data before writing
them out to the underlying OutputStream.
- SaslOutputStream(OutputStream, SaslServer) -
Constructor for class org.apache.hadoop.security.SaslOutputStream
- Constructs a SASLOutputStream from an OutputStream and a SaslServer
Note: if the specified OutputStream or SaslServer is null, a
NullPointerException may be thrown later when they are used.
- SaslOutputStream(OutputStream, SaslClient) -
Constructor for class org.apache.hadoop.security.SaslOutputStream
- Constructs a SASLOutputStream from an OutputStream and a SaslClient
Note: if the specified OutputStream or SaslClient is null, a
NullPointerException may be thrown later when they are used.
- saslQop -
Variable in enum org.apache.hadoop.security.SaslRpcServer.QualityOfProtection
-
- SaslRpcClient - Class in org.apache.hadoop.security
- A utility class that encapsulates SASL logic for RPC client
- SaslRpcClient(SaslRpcServer.AuthMethod, Token<? extends TokenIdentifier>, String, boolean) -
Constructor for class org.apache.hadoop.security.SaslRpcClient
- Create a SaslRpcClient for an authentication method
- SaslRpcServer - Class in org.apache.hadoop.security
- A utility class for dealing with SASL on RPC server
- SaslRpcServer() -
Constructor for class org.apache.hadoop.security.SaslRpcServer
-
- SaslRpcServer.AuthMethod - Enum in org.apache.hadoop.security
- Authentication method
- SaslRpcServer.QualityOfProtection - Enum in org.apache.hadoop.security
-
- SaslRpcServer.SaslDigestCallbackHandler - Class in org.apache.hadoop.security
- CallbackHandler for SASL DIGEST-MD5 mechanism
- SaslRpcServer.SaslDigestCallbackHandler(SecretManager<TokenIdentifier>, Server.Connection) -
Constructor for class org.apache.hadoop.security.SaslRpcServer.SaslDigestCallbackHandler
-
- SaslRpcServer.SaslGssCallbackHandler - Class in org.apache.hadoop.security
- CallbackHandler for SASL GSSAPI Kerberos mechanism
- SaslRpcServer.SaslGssCallbackHandler() -
Constructor for class org.apache.hadoop.security.SaslRpcServer.SaslGssCallbackHandler
-
- SaslRpcServer.SaslStatus - Enum in org.apache.hadoop.security
-
- scheduleMap(TaskInProgress) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Adds a map tip to the list of running maps.
- scheduleOffSwitch(int) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Check if we can schedule an off-switch task for this job.
- scheduleReduce(TaskInProgress) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Adds a reduce tip to the list of running reduces
- scheduleReduces() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- schedulingOpportunity() -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- SCHEME -
Static variable in class org.apache.hadoop.mapreduce.security.token.DelegationTokenRenewal
-
- ScriptBasedMapping - Class in org.apache.hadoop.net
- This class implements the
DNSToSwitchMapping
interface using a
script configured via topology.script.file.name . - ScriptBasedMapping() -
Constructor for class org.apache.hadoop.net.ScriptBasedMapping
-
- ScriptBasedMapping(Configuration) -
Constructor for class org.apache.hadoop.net.ScriptBasedMapping
-
- SecondarySort - Class in org.apache.hadoop.examples
- This is an example Hadoop Map/Reduce application.
- SecondarySort() -
Constructor for class org.apache.hadoop.examples.SecondarySort
-
- SecondarySort.FirstGroupingComparator - Class in org.apache.hadoop.examples
- Compare only the first part of the pair, so that reduce is called once
for each value of the first part.
- SecondarySort.FirstGroupingComparator() -
Constructor for class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
-
- SecondarySort.FirstPartitioner - Class in org.apache.hadoop.examples
- Partition based on the first part of the pair.
- SecondarySort.FirstPartitioner() -
Constructor for class org.apache.hadoop.examples.SecondarySort.FirstPartitioner
-
- SecondarySort.IntPair - Class in org.apache.hadoop.examples
- Define a pair of integers that are writable.
- SecondarySort.IntPair() -
Constructor for class org.apache.hadoop.examples.SecondarySort.IntPair
-
- SecondarySort.IntPair.Comparator - Class in org.apache.hadoop.examples
- A Comparator that compares serialized IntPair.
- SecondarySort.IntPair.Comparator() -
Constructor for class org.apache.hadoop.examples.SecondarySort.IntPair.Comparator
-
- SecondarySort.MapClass - Class in org.apache.hadoop.examples
- Read two integers from each line and generate a key, value pair
as ((left, right), right).
- SecondarySort.MapClass() -
Constructor for class org.apache.hadoop.examples.SecondarySort.MapClass
-
- SecondarySort.Reduce - Class in org.apache.hadoop.examples
- A reducer class that just emits the sum of the input values.
- SecondarySort.Reduce() -
Constructor for class org.apache.hadoop.examples.SecondarySort.Reduce
-
- SecretManager<T extends TokenIdentifier> - Class in org.apache.hadoop.security.token
- The server-side secret manager for each token type.
- SecretManager() -
Constructor for class org.apache.hadoop.security.token.SecretManager
-
- SecretManager.InvalidToken - Exception in org.apache.hadoop.security.token
- The token was invalid and the message explains why.
- SecretManager.InvalidToken(String) -
Constructor for exception org.apache.hadoop.security.token.SecretManager.InvalidToken
-
- SecureIOUtils - Class in org.apache.hadoop.io
- This class provides secure APIs for opening and creating files on the local
disk.
- SecureIOUtils() -
Constructor for class org.apache.hadoop.io.SecureIOUtils
-
- SecureIOUtils.AlreadyExistsException - Exception in org.apache.hadoop.io
- Signals that an attempt to create a file at a given pathname has failed
because another file already existed at that path.
- SecureIOUtils.AlreadyExistsException(String) -
Constructor for exception org.apache.hadoop.io.SecureIOUtils.AlreadyExistsException
-
- SecureIOUtils.AlreadyExistsException(Throwable) -
Constructor for exception org.apache.hadoop.io.SecureIOUtils.AlreadyExistsException
-
- SecureShuffleUtils - Class in org.apache.hadoop.mapreduce.security
- utilities for generating kyes, hashes and verifying them for shuffle
- SecureShuffleUtils() -
Constructor for class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
-
- SecurityUtil - Class in org.apache.hadoop.security
-
- SecurityUtil() -
Constructor for class org.apache.hadoop.security.SecurityUtil
-
- SecurityUtil.QualifiedHostResolver - Class in org.apache.hadoop.security
- This an alternate resolver with important properties that the standard
java resolver lacks:
1) The hostname is fully qualified.
- SecurityUtil.QualifiedHostResolver() -
Constructor for class org.apache.hadoop.security.SecurityUtil.QualifiedHostResolver
-
- seek(long) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- seek(long) -
Method in class org.apache.hadoop.fs.FSDataInputStream
-
- seek(long) -
Method in class org.apache.hadoop.fs.FSInputChecker
- Seek to the given position in the stream.
- seek(long) -
Method in class org.apache.hadoop.fs.FSInputStream
- Seek to the given offset from the start of the file.
- seek(long) -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- seek(long) -
Method in interface org.apache.hadoop.fs.Seekable
- Seek to the given offset from the start of the file.
- seek(long) -
Method in class org.apache.hadoop.io.ArrayFile.Reader
- Positions the reader before its
n
th value.
- seek(long) -
Method in class org.apache.hadoop.io.compress.CompressionInputStream
- This method is current not supported.
- seek(WritableComparable) -
Method in class org.apache.hadoop.io.MapFile.Reader
- Positions the reader at the named key, or if none such exists, at the
first entry after the named key.
- seek(long) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Set the current byte position in the input file.
- seek(WritableComparable) -
Method in class org.apache.hadoop.io.SetFile.Reader
-
- seek(long) -
Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- Seekable - Interface in org.apache.hadoop.fs
- Stream that permits seeking.
- seekNextRecordBoundary() -
Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
- Implementation should seek forward in_ to the first byte of the next record.
- seekNextRecordBoundary() -
Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- seekTo(byte[]) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is greater than or equal
to the input key.
- seekTo(byte[], int, int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is greater than or equal
to the input key.
- seekToEnd() -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Seek to the end of the scanner.
- seekToNewSource(long) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- seekToNewSource(long) -
Method in class org.apache.hadoop.fs.FSDataInputStream
-
- seekToNewSource(long) -
Method in class org.apache.hadoop.fs.FSInputStream
- Seeks a different copy of the data.
- seekToNewSource(long) -
Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- seekToNewSource(long) -
Method in interface org.apache.hadoop.fs.Seekable
- Seeks a different copy of the data.
- seekToNewSource(long) -
Method in class org.apache.hadoop.io.compress.CompressionInputStream
- This method is current not supported.
- seenPrimary_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- selectiveClearing(Key, short) -
Method in class org.apache.hadoop.util.bloom.RetouchedBloomFilter
- Performs the selective clearing for a given key.
- selectToken(Text, Collection<Token<? extends TokenIdentifier>>) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenSelector
-
- selectToken(Text, Collection<Token<? extends TokenIdentifier>>) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSelector
-
- selectToken(Text, Collection<Token<? extends TokenIdentifier>>) -
Method in interface org.apache.hadoop.security.token.TokenSelector
-
- SEMICOLON_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- SensorsParser - Class in org.apache.hadoop.contrib.failmon
- Objects of this class parse the output of the lm-sensors utility
to gather information about fan speed, temperatures for cpus
and motherboard etc.
- SensorsParser() -
Constructor for class org.apache.hadoop.contrib.failmon.SensorsParser
-
- SEPARATOR -
Static variable in class org.apache.hadoop.fs.Path
- The directory separator, a slash.
- SEPARATOR -
Static variable in class org.apache.hadoop.mapreduce.ID
-
- SEPARATOR_CHAR -
Static variable in class org.apache.hadoop.fs.Path
-
- SequenceFile - Class in org.apache.hadoop.io
SequenceFile
s are flat files consisting of binary key/value
pairs.- SequenceFile.CompressionType - Enum in org.apache.hadoop.io
- The compression type used to compress key/value pairs in the
SequenceFile
. - SequenceFile.Metadata - Class in org.apache.hadoop.io
- The class encapsulating with the metadata of a file.
- SequenceFile.Metadata() -
Constructor for class org.apache.hadoop.io.SequenceFile.Metadata
-
- SequenceFile.Metadata(TreeMap<Text, Text>) -
Constructor for class org.apache.hadoop.io.SequenceFile.Metadata
-
- SequenceFile.Reader - Class in org.apache.hadoop.io
- Reads key/value pairs from a sequence-format file.
- SequenceFile.Reader(FileSystem, Path, Configuration) -
Constructor for class org.apache.hadoop.io.SequenceFile.Reader
- Open the named file.
- SequenceFile.Sorter - Class in org.apache.hadoop.io
- Sorts key/value pairs in a sequence-format file.
- SequenceFile.Sorter(FileSystem, Class<? extends WritableComparable>, Class, Configuration) -
Constructor for class org.apache.hadoop.io.SequenceFile.Sorter
- Sort and merge files containing the named classes.
- SequenceFile.Sorter(FileSystem, RawComparator, Class, Class, Configuration) -
Constructor for class org.apache.hadoop.io.SequenceFile.Sorter
- Sort and merge using an arbitrary
RawComparator
.
- SequenceFile.Sorter.RawKeyValueIterator - Interface in org.apache.hadoop.io
- The interface to iterate over raw keys/values of SequenceFiles.
- SequenceFile.Sorter.SegmentDescriptor - Class in org.apache.hadoop.io
- This class defines a merge segment.
- SequenceFile.Sorter.SegmentDescriptor(long, long, Path) -
Constructor for class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
- Constructs a segment
- SequenceFile.ValueBytes - Interface in org.apache.hadoop.io
- The interface to 'raw' values of SequenceFiles.
- SequenceFile.Writer - Class in org.apache.hadoop.io
- Write key/value pairs to a sequence-format file.
- SequenceFile.Writer(FileSystem, Configuration, Path, Class, Class) -
Constructor for class org.apache.hadoop.io.SequenceFile.Writer
- Create the named file.
- SequenceFile.Writer(FileSystem, Configuration, Path, Class, Class, Progressable, SequenceFile.Metadata) -
Constructor for class org.apache.hadoop.io.SequenceFile.Writer
- Create the named file with write-progress reporter.
- SequenceFile.Writer(FileSystem, Configuration, Path, Class, Class, int, short, long, Progressable, SequenceFile.Metadata) -
Constructor for class org.apache.hadoop.io.SequenceFile.Writer
- Create the named file with write-progress reporter.
- SequenceFileAsBinaryInputFormat - Class in org.apache.hadoop.mapred
- InputFormat reading keys, values from SequenceFiles in binary (raw)
format.
- SequenceFileAsBinaryInputFormat() -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
-
- SequenceFileAsBinaryInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
- InputFormat reading keys, values from SequenceFiles in binary (raw)
format.
- SequenceFileAsBinaryInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat
-
- SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader - Class in org.apache.hadoop.mapred
- Read records from a SequenceFile as binary (raw) bytes.
- SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
- Read records from a SequenceFile as binary (raw) bytes.
- SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- SequenceFileAsBinaryOutputFormat - Class in org.apache.hadoop.mapred
- An
OutputFormat
that writes keys, values to
SequenceFile
s in binary(raw) format - SequenceFileAsBinaryOutputFormat() -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- SequenceFileAsBinaryOutputFormat - Class in org.apache.hadoop.mapreduce.lib.output
- An
OutputFormat
that writes keys,
values to SequenceFile
s in binary(raw) format - SequenceFileAsBinaryOutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
-
- SequenceFileAsBinaryOutputFormat.WritableValueBytes - Class in org.apache.hadoop.mapred
- Inner class used for appendRaw
- SequenceFileAsBinaryOutputFormat.WritableValueBytes() -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- SequenceFileAsBinaryOutputFormat.WritableValueBytes(BytesWritable) -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- SequenceFileAsBinaryOutputFormat.WritableValueBytes - Class in org.apache.hadoop.mapreduce.lib.output
- Inner class used for appendRaw
- SequenceFileAsBinaryOutputFormat.WritableValueBytes() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- SequenceFileAsBinaryOutputFormat.WritableValueBytes(BytesWritable) -
Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- SequenceFileAsTextInputFormat - Class in org.apache.hadoop.mapred
- This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader
which converts the input keys and values to their String forms by calling toString() method.
- SequenceFileAsTextInputFormat() -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
-
- SequenceFileAsTextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
- This class is similar to SequenceFileInputFormat, except it generates
SequenceFileAsTextRecordReader which converts the input keys and values
to their String forms by calling toString() method.
- SequenceFileAsTextInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat
-
- SequenceFileAsTextRecordReader - Class in org.apache.hadoop.mapred
- This class converts the input keys and values to their String forms by calling toString()
method.
- SequenceFileAsTextRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- SequenceFileAsTextRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
- This class converts the input keys and values to their String forms by
calling toString() method.
- SequenceFileAsTextRecordReader() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
-
- SequenceFileInputFilter<K,V> - Class in org.apache.hadoop.mapred
- A class that allows a map/red job to work on a sample of sequence files.
- SequenceFileInputFilter() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter
-
- SequenceFileInputFilter<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
- A class that allows a map/red job to work on a sample of sequence files.
- SequenceFileInputFilter() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
-
- SequenceFileInputFilter.Filter - Interface in org.apache.hadoop.mapred
- filter interface
- SequenceFileInputFilter.Filter - Interface in org.apache.hadoop.mapreduce.lib.input
- filter interface
- SequenceFileInputFilter.FilterBase - Class in org.apache.hadoop.mapred
- base class for Filters
- SequenceFileInputFilter.FilterBase() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase
-
- SequenceFileInputFilter.FilterBase - Class in org.apache.hadoop.mapreduce.lib.input
- base class for Filters
- SequenceFileInputFilter.FilterBase() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase
-
- SequenceFileInputFilter.MD5Filter - Class in org.apache.hadoop.mapred
- This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency f.
- SequenceFileInputFilter.MD5Filter() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
-
- SequenceFileInputFilter.MD5Filter - Class in org.apache.hadoop.mapreduce.lib.input
- This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency f.
- SequenceFileInputFilter.MD5Filter() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
-
- SequenceFileInputFilter.PercentFilter - Class in org.apache.hadoop.mapred
- This class returns a percentage of records
The percentage is determined by a filtering frequency f using
the criteria record# % f == 0.
- SequenceFileInputFilter.PercentFilter() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
-
- SequenceFileInputFilter.PercentFilter - Class in org.apache.hadoop.mapreduce.lib.input
- This class returns a percentage of records
The percentage is determined by a filtering frequency f using
the criteria record# % f == 0.
- SequenceFileInputFilter.PercentFilter() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
-
- SequenceFileInputFilter.RegexFilter - Class in org.apache.hadoop.mapred
- Records filter by matching key to regex
- SequenceFileInputFilter.RegexFilter() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
-
- SequenceFileInputFilter.RegexFilter - Class in org.apache.hadoop.mapreduce.lib.input
- Records filter by matching key to regex
- SequenceFileInputFilter.RegexFilter() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
-
- SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapred
- An
InputFormat
for SequenceFile
s. - SequenceFileInputFormat() -
Constructor for class org.apache.hadoop.mapred.SequenceFileInputFormat
-
- SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
- An
InputFormat
for SequenceFile
s. - SequenceFileInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
-
- SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapred
- An
OutputFormat
that writes SequenceFile
s. - SequenceFileOutputFormat() -
Constructor for class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
- SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
- An
OutputFormat
that writes SequenceFile
s. - SequenceFileOutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
-
- SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapred
- An
RecordReader
for SequenceFile
s. - SequenceFileRecordReader(Configuration, FileSplit) -
Constructor for class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
- An
RecordReader
for SequenceFile
s. - SequenceFileRecordReader() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- Serialization<T> - Interface in org.apache.hadoop.io.serializer
-
Encapsulates a
Serializer
/Deserializer
pair. - SerializationFactory - Class in org.apache.hadoop.io.serializer
-
A factory for
Serialization
s. - SerializationFactory(Configuration) -
Constructor for class org.apache.hadoop.io.serializer.SerializationFactory
-
Serializations are found by reading the
io.serializations
property from conf
, which is a comma-delimited list of
classnames.
- serialize() -
Method in class org.apache.hadoop.fs.s3.INode
-
- serialize(T) -
Method in interface org.apache.hadoop.io.serializer.Serializer
- Serialize
t
to the underlying output stream.
- serialize(RecordOutput, String) -
Method in class org.apache.hadoop.record.meta.RecordTypeInfo
- Serialize the type information for a record
- serialize(RecordOutput, String) -
Method in class org.apache.hadoop.record.Record
- Serialize a record with tag (ususally field name)
- serialize(RecordOutput) -
Method in class org.apache.hadoop.record.Record
- Serialize a record without a tag
- SerializedRecord - Class in org.apache.hadoop.contrib.failmon
- Objects of this class hold the serialized representations
of EventRecords.
- SerializedRecord(EventRecord) -
Constructor for class org.apache.hadoop.contrib.failmon.SerializedRecord
- Create the SerializedRecord given an EventRecord.
- Serializer<T> - Interface in org.apache.hadoop.io.serializer
-
Provides a facility for serializing objects of type
to an
OutputStream
. - Server - Class in org.apache.hadoop.ipc
- An abstract IPC service.
- Server(String, int, Class<? extends Writable>, int, Configuration) -
Constructor for class org.apache.hadoop.ipc.Server
-
- Server(String, int, Class<? extends Writable>, int, Configuration, String) -
Constructor for class org.apache.hadoop.ipc.Server
-
- Server(String, int, Class<? extends Writable>, int, Configuration, String, SecretManager<? extends TokenIdentifier>) -
Constructor for class org.apache.hadoop.ipc.Server
- Constructs a server listening on the named port and address.
- Server.Connection - Class in org.apache.hadoop.ipc
- Reads calls from a connection and queues them for handling.
- Server.Connection(SelectionKey, SocketChannel, long) -
Constructor for class org.apache.hadoop.ipc.Server.Connection
-
- Servers - Class in org.apache.hadoop.metrics2.util
- Helpers to handle server addresses
- SERVERS_PROPERTY -
Static variable in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- Service - Class in org.apache.hadoop.security.authorize
- An abstract definition of service as related to
Service Level Authorization for Hadoop.
- Service(String, Class<?>) -
Constructor for class org.apache.hadoop.security.authorize.Service
-
- SERVICE_AUTHORIZATION_CONFIG -
Static variable in class org.apache.hadoop.security.authorize.ServiceAuthorizationManager
- Deprecated. Use
CommonConfigurationKeys.HADOOP_SECURITY_AUTHORIZATION
instead.
- ServiceAuthorizationManager - Class in org.apache.hadoop.security.authorize
- An authorization manager which handles service-level authorization
for incoming service requests.
- ServiceAuthorizationManager() -
Constructor for class org.apache.hadoop.security.authorize.ServiceAuthorizationManager
-
- ServicePlugin - Interface in org.apache.hadoop.util
- Service plug-in interface.
- ServletUtil - Class in org.apache.hadoop.util
-
- ServletUtil() -
Constructor for class org.apache.hadoop.util.ServletUtil
-
- SESSION_TIMEZONE_KEY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
- Configuration key to set to a timezone string.
- set(String, String) -
Method in class org.apache.hadoop.conf.Configuration
- Set the
value
of the name
property.
- set(String, Object) -
Method in class org.apache.hadoop.contrib.failmon.EventRecord
- Set the value of a property of the EventRecord.
- set(String, String) -
Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
- Set the value of a property of the EventRecord.
- set(int, int) -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
- Set the left and right values.
- set(boolean, Checksum, int, int) -
Method in class org.apache.hadoop.fs.FSInputChecker
- Set the checksum related parameters
- set(String) -
Method in class org.apache.hadoop.fs.GlobPattern
- Set and compile a glob pattern
- set(Writable[]) -
Method in class org.apache.hadoop.io.ArrayWritable
-
- set(boolean) -
Method in class org.apache.hadoop.io.BooleanWritable
- Set the value of the BooleanWritable
- set(BytesWritable) -
Method in class org.apache.hadoop.io.BytesWritable
- Set the BytesWritable to the contents of the given newData.
- set(byte[], int, int) -
Method in class org.apache.hadoop.io.BytesWritable
- Set the value to a copy of the given byte range
- set(byte) -
Method in class org.apache.hadoop.io.ByteWritable
- Set the value of this ByteWritable.
- set(double) -
Method in class org.apache.hadoop.io.DoubleWritable
-
- set(float) -
Method in class org.apache.hadoop.io.FloatWritable
- Set the value of this FloatWritable.
- set(Writable) -
Method in class org.apache.hadoop.io.GenericWritable
- Set the instance that is wrapped.
- set(int) -
Method in class org.apache.hadoop.io.IntWritable
- Set the value of this IntWritable.
- set(long) -
Method in class org.apache.hadoop.io.LongWritable
- Set the value of this LongWritable.
- set(MD5Hash) -
Method in class org.apache.hadoop.io.MD5Hash
- Copy the contents of another instance into this instance.
- set(Object) -
Method in class org.apache.hadoop.io.ObjectWritable
- Reset the instance.
- set(Text, Text) -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- set(String) -
Method in class org.apache.hadoop.io.Text
- Set to contain the contents of a string.
- set(byte[]) -
Method in class org.apache.hadoop.io.Text
- Set to a utf8 byte array
- set(Text) -
Method in class org.apache.hadoop.io.Text
- copy a text.
- set(byte[], int, int) -
Method in class org.apache.hadoop.io.Text
- Set the Text to range of bytes
- set(Writable[][]) -
Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- set(String) -
Method in class org.apache.hadoop.io.UTF8
- Deprecated. Set to contain the contents of a string.
- set(UTF8) -
Method in class org.apache.hadoop.io.UTF8
- Deprecated. Set to contain the contents of a string.
- set(int) -
Method in class org.apache.hadoop.io.VIntWritable
- Set the value of this VIntWritable.
- set(long) -
Method in class org.apache.hadoop.io.VLongWritable
- Set the value of this LongWritable.
- set(int) -
Method in class org.apache.hadoop.metrics.util.MetricsIntValue
- Deprecated. Set the value
- set(long) -
Method in class org.apache.hadoop.metrics.util.MetricsLongValue
- Deprecated. Set the value
- set(int) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeInt
- Set the value of the metric
- set(long) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeLong
- Set the value of the metric
- set(byte[]) -
Method in class org.apache.hadoop.record.Buffer
- Use the specified bytes array as underlying sequence.
- set(byte[], double) -
Method in class org.apache.hadoop.util.bloom.Key
-
- set(float) -
Method in class org.apache.hadoop.util.Progress
- Called during execution on a leaf node to set its progress.
- SET_GROUP_COMMAND -
Static variable in class org.apache.hadoop.util.Shell
-
- SET_OWNER_COMMAND -
Static variable in class org.apache.hadoop.util.Shell
- a Unix command to set owner
- SET_PERMISSION_COMMAND -
Static variable in class org.apache.hadoop.util.Shell
- a Unix command to set permission
- setAggregatorDescriptors(JobConf, Class<? extends ValueAggregatorDescriptor>[]) -
Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- setArchiveSizes(JobID, long[]) -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
- Set the sizes for any archives, files, or directories in the private
distributed cache.
- setArchiveTimestamps(Configuration, String) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- This is to check the timestamp of the archives to be localized.
- setAssignedJobID(JobID) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the mapred ID for this job as assigned by the
mapred framework.
- setAttemptsToStartSkipping(Configuration, int) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the number of Task attempts AFTER which skip mode
will be kicked off.
- setAttribute(String, Object) -
Method in class org.apache.hadoop.http.HttpServer
- Set a value in the webapp context.
- setAttribute(Context, String, Object) -
Method in class org.apache.hadoop.http.HttpServer
- Set a value in the webapp context.
- setAttribute(String, Object) -
Method in class org.apache.hadoop.metrics.ContextFactory
- Deprecated. Sets the named factory attribute to the specified value, creating it
if it did not already exist.
- setAttribute(Attribute) -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
- Deprecated.
- setAttributes(AttributeList) -
Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
- Deprecated.
- setAuthenticationMethod(UserGroupInformation.AuthenticationMethod) -
Method in class org.apache.hadoop.security.UserGroupInformation
- Sets the authentication method in the subject
- setAutoIncrMapperProcCount(Configuration, boolean) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the flag which if set to true,
SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS
is incremented
by MapRunner after invoking the map function.
- setAutoIncrReducerProcCount(Configuration, boolean) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the flag which if set to true,
SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS
is incremented
by framework after invoking the reduce function.
- setBoolean(String, boolean) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to a boolean
.
- setBooleanIfUnset(String, boolean) -
Method in class org.apache.hadoop.conf.Configuration
- Set the given property, if it is currently unset.
- setBoundingQuery(Configuration, String) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
- Set the user-defined bounding query to use with a user-defined query.
- setCacheArchives(URI[], Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- Set the configuration with the given set of archives.
- setCacheFiles(URI[], Configuration) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- Set the configuration with the given set of files.
- setCancelDelegationTokenUponJobCompletion(boolean) -
Method in class org.apache.hadoop.mapreduce.Job
- Sets the flag that will allow the JobTracker to cancel the HDFS delegation
tokens upon job completion.
- setCapacity(int) -
Method in class org.apache.hadoop.io.BytesWritable
- Change the capacity of the backing storage.
- setCapacity(int) -
Method in class org.apache.hadoop.record.Buffer
- Change the capacity of the backing storage.
- setChanged() -
Method in class org.apache.hadoop.metrics2.lib.MetricMutable
- Set the changed flag in mutable operations
- setClass(String, Class<?>, Class<?>) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to the name of a
theClass
implementing the given interface xface
.
- setClassLoader(ClassLoader) -
Method in class org.apache.hadoop.conf.Configuration
- Set the class loader that will be used to load the various objects.
- setCodecClasses(Configuration, List<Class>) -
Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
- Sets a list of codec classes in the configuration.
- setCombinerClass(Class<? extends Reducer>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the user-defined combiner class used to combine map-outputs
before being sent to the reducers.
- setCombinerClass(Class<? extends Reducer>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the combiner class for the job.
- setCompressionLevel(Configuration, ZlibCompressor.CompressionLevel) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
- setCompressionStrategy(Configuration, ZlibCompressor.CompressionStrategy) -
Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
- setCompressionType(Configuration, SequenceFile.CompressionType) -
Static method in class org.apache.hadoop.io.SequenceFile
- Deprecated. Use the one of the many SequenceFile.createWriter methods to specify
the
SequenceFile.CompressionType
while creating the SequenceFile
or
SequenceFileOutputFormat.setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)
to specify the SequenceFile.CompressionType
for job-outputs.
or
- setCompressMapOutput(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Should the map outputs be compressed before transfer?
Uses the SequenceFile compression.
- setCompressOutput(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
- Set whether the output of the job is compressed.
- setCompressOutput(Job, boolean) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
- Set whether the output of the job is compressed.
- setConf(Configuration) -
Method in interface org.apache.hadoop.conf.Configurable
- Set the configuration to be used by this object.
- setConf(Configuration) -
Method in class org.apache.hadoop.conf.Configured
-
- setConf(Configuration) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- setConf(Configuration) -
Method in class org.apache.hadoop.io.AbstractMapWritable
-
- setConf(Configuration) -
Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- setConf(Configuration) -
Method in class org.apache.hadoop.io.compress.SnappyCodec
- Set the configuration to be used by this object.
- setConf(Configuration) -
Method in class org.apache.hadoop.io.GenericWritable
-
- setConf(Configuration) -
Method in class org.apache.hadoop.io.ObjectWritable
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Set the configuration to be used by this object.
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.lib.InputSampler
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
- configure the filter according to configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
- configure the filter by checking the configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
- configure the Filter by checking the configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.Task
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapred.TaskController
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
- Set the configuration to be used by this object.
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
- configure the filter according to configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
- configure the filter by checking the configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
- configure the Filter by checking the configuration
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
-
- setConf(Configuration) -
Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
- Read in the partition file and build indexing data structures.
- setConf(Configuration) -
Method in class org.apache.hadoop.net.ScriptBasedMapping
-
- setConf(Configuration) -
Method in class org.apache.hadoop.net.SocksSocketFactory
-
- setConf(Configuration) -
Method in class org.apache.hadoop.streaming.DumpTypedBytes
-
- setConf(Configuration) -
Method in class org.apache.hadoop.streaming.LoadTypedBytes
-
- setConf(Configuration) -
Method in class org.apache.hadoop.streaming.StreamJob
-
- setConf(Configuration) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
-
- setConf(Object, Configuration) -
Static method in class org.apache.hadoop.util.ReflectionUtils
- Check and set 'configuration' if necessary.
- setConfiguration(Configuration) -
Static method in class org.apache.hadoop.security.KerberosName
- Set the static configuration to get the rules.
- setConfiguration(Configuration) -
Static method in class org.apache.hadoop.security.UserGroupInformation
- Set the static configuration for UGI.
- setContentionTracing(boolean) -
Static method in class org.apache.hadoop.util.ReflectionUtils
-
- setContext(String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Set the metrics context tag
- setContext(String) -
Method in class org.apache.hadoop.metrics2.MetricsRecordBuilder
- Set the context tag
- setCounters(Counters) -
Method in class org.apache.hadoop.mapred.TaskStatus
- Set the task's counters.
- setCountersEnabled(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
- Enables or disables counters for the named outputs.
- setCountersEnabled(Job, boolean) -
Static method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
- Enables or disables counters for the named outputs.
- setDebugStream(PrintStream) -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- setDefaultAuthenticator(Class<? extends Authenticator>) -
Static method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL
- Sets the default
Authenticator
class to use when an AuthenticatedURL
instance
is created without specifying an authenticator.
- setDefaultUri(Configuration, URI) -
Static method in class org.apache.hadoop.fs.FileSystem
- Set the default filesystem URI in a configuration.
- setDefaultUri(Configuration, String) -
Static method in class org.apache.hadoop.fs.FileSystem
- Set the default filesystem URI in a configuration.
- setDelete(Term) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
- Set the instance to be a delete operation.
- setDestdir(File) -
Method in class org.apache.hadoop.record.compiler.ant.RccTask
- Sets directory where output files will be generated
- setDiagnosticInfo(String) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- setDictionary(byte[], int, int) -
Method in interface org.apache.hadoop.io.compress.Compressor
- Sets preset dictionary for compression.
- setDictionary(byte[], int, int) -
Method in interface org.apache.hadoop.io.compress.Decompressor
- Sets preset dictionary for compression.
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Does nothing.
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Does nothing.
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
- Sets preset dictionary for compression.
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- setDictionary(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- setDigest(String) -
Method in class org.apache.hadoop.io.MD5Hash
- Sets the digest value from a hex string.
- setDisplayName(String) -
Method in class org.apache.hadoop.mapred.Counters.Counter
-
- setDisplayName(String) -
Method in class org.apache.hadoop.mapred.Counters.Group
- Set the display name
- setDisplayName(String) -
Method in class org.apache.hadoop.mapreduce.Counter
- Deprecated.
- setDistributionPolicyClass(Class<? extends IDistributionPolicy>) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the distribution policy class.
- setDocumentAnalyzerClass(Class<? extends Analyzer>) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the analyzer class.
- setDoubleValue(Object, double) -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
- Set the given counter to the given value
- setEnd(long) -
Method in class org.apache.hadoop.io.compress.SplitCompressionInputStream
-
- setEnum(String, T) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to the given type.
- setEnvironment(Map<String, String>) -
Method in class org.apache.hadoop.util.Shell
- set the environment for the command
- setEventId(int) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- set event Id.
- setExcludesFile(String) -
Method in class org.apache.hadoop.util.HostsFileReader
-
- setExecutable(JobConf, String) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set the URI for the application's executable.
- setExpires(long) -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
- Sets the expiration of the token.
- setExpiryDate(long) -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
-
- setFactor(int) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Set the number of streams to merge at once.
- setFactory(Class, WritableFactory) -
Static method in class org.apache.hadoop.io.WritableFactories
- Define a factory for a class.
- setFailonerror(boolean) -
Method in class org.apache.hadoop.record.compiler.ant.RccTask
- Given multiple files (via fileset), set the error handling behavior
- setFailureInfo(String) -
Method in class org.apache.hadoop.mapred.JobStatus
- set the reason for failuire of this job
- SetFile - Class in org.apache.hadoop.io
- A file-based set of keys.
- SetFile() -
Constructor for class org.apache.hadoop.io.SetFile
-
- setFile(File) -
Method in class org.apache.hadoop.record.compiler.ant.RccTask
- Sets the record definition file attribute
- SetFile.Reader - Class in org.apache.hadoop.io
- Provide access to an existing set file.
- SetFile.Reader(FileSystem, String, Configuration) -
Constructor for class org.apache.hadoop.io.SetFile.Reader
- Construct a set reader for the named set.
- SetFile.Reader(FileSystem, String, WritableComparator, Configuration) -
Constructor for class org.apache.hadoop.io.SetFile.Reader
- Construct a set reader for the named set using the named comparator.
- SetFile.Writer - Class in org.apache.hadoop.io
- Write a new set file.
- SetFile.Writer(FileSystem, String, Class<? extends WritableComparable>) -
Constructor for class org.apache.hadoop.io.SetFile.Writer
- Deprecated. pass a Configuration too
- SetFile.Writer(Configuration, FileSystem, String, Class<? extends WritableComparable>, SequenceFile.CompressionType) -
Constructor for class org.apache.hadoop.io.SetFile.Writer
- Create a set naming the element class and compression type.
- SetFile.Writer(Configuration, FileSystem, String, WritableComparator, SequenceFile.CompressionType) -
Constructor for class org.apache.hadoop.io.SetFile.Writer
- Create a set naming the element comparator and compression type.
- setFileTimestamps(Configuration, String) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- This is to check the timestamp of the files to be localized.
- setFilterClass(Configuration, Class) -
Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter
- set the filter class
- setFilterClass(Job, Class<?>) -
Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
- set the filter class
- setFinalSync(JobConf, boolean) -
Static method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
- Set the requirement for a final sync before the stream is closed.
- setFloat(String, float) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to a float
.
- setFormat(JobConf) -
Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
- Interpret a given string as a composite expression.
- setFrequency(Configuration, int) -
Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
- set the filtering frequency in configuration
- setFrequency(Configuration, int) -
Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
- set the frequency and stores it in conf
- setFrequency(Configuration, int) -
Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
- set the filtering frequency in configuration
- setFrequency(Configuration, int) -
Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
- set the frequency and stores it in conf
- setGroup(String) -
Method in class org.apache.hadoop.fs.FileStatus
- Sets group.
- setGroupingComparatorClass(Class<? extends RawComparator>) -
Method in class org.apache.hadoop.mapreduce.Job
- Define the comparator that controls which keys are grouped together
for a single call to
Reducer.reduce(Object, Iterable,
org.apache.hadoop.mapreduce.Reducer.Context)
- setHosts(String[]) -
Method in class org.apache.hadoop.fs.BlockLocation
- Set the hosts hosting this block
- setID(int) -
Method in class org.apache.hadoop.mapred.join.Parser.Node
-
- setIfUnset(String, String) -
Method in class org.apache.hadoop.conf.Configuration
- Sets a property if it is currently unset.
- setIncludeCounters(boolean) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setIncludesFile(String) -
Method in class org.apache.hadoop.util.HostsFileReader
-
- setIndexInputFormatClass(Class<? extends InputFormat>) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the index input format class.
- setIndexInterval(int) -
Method in class org.apache.hadoop.io.MapFile.Writer
- Sets the index interval.
- setIndexInterval(Configuration, int) -
Static method in class org.apache.hadoop.io.MapFile.Writer
- Sets the index interval and stores it in conf
- setIndexMaxFieldLength(int) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the max field length for a Lucene instance.
- setIndexMaxNumSegments(int) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the max number of segments for a Lucene instance.
- setIndexShards(String) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the string representation of a number of shards.
- setIndexShards(IndexUpdateConfiguration, Shard[]) -
Static method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- setIndexUpdaterClass(Class<? extends IIndexUpdater>) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the index updater class.
- setIndexUseCompoundFile(boolean) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set whether use the compound file format for a Lucene instance.
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- setInput(byte[], int, int) -
Method in interface org.apache.hadoop.io.compress.Compressor
- Sets input data for compression.
- setInput(byte[], int, int) -
Method in interface org.apache.hadoop.io.compress.Decompressor
- Sets input data for decompression.
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Sets input data for compression.
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Sets input data for decompression.
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor
- Sets input data for decompression.
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- setInput(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- setInput(JobConf, Class<? extends DBWritable>, String, String, String, String...) -
Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
- Initializes the map-part of the job with the appropriate input settings.
- setInput(JobConf, Class<? extends DBWritable>, String, String) -
Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
- Initializes the map-part of the job with the appropriate input settings.
- setInput(Job, Class<? extends DBWritable>, String, String, String, String...) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
- Note that the "orderBy" column is called the "splitBy" in this version.
- setInput(Job, Class<? extends DBWritable>, String, String) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
- setInput() takes a custom query and a separate "bounding query" to use
instead of the custom "count query" used by DBInputFormat.
- setInput(Job, Class<? extends DBWritable>, String, String, String, String...) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
- Initializes the map-part of the job with the appropriate input settings.
- setInput(Job, Class<? extends DBWritable>, String, String) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
- Initializes the map-part of the job with the appropriate input settings.
- setInputBoundingQuery(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputClass(Class<? extends DBWritable>) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputConditions(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputCountQuery(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputDataLength(long) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- setInputDataLocations(String[]) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- setInputFieldNames(String...) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputFormat(Class<? extends InputFormat>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
InputFormat
implementation for the map-reduce job.
- setInputFormatClass(Class<? extends InputFormat>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the
InputFormat
for the job.
- setInputOrderBy(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputPathFilter(JobConf, Class<? extends PathFilter>) -
Static method in class org.apache.hadoop.mapred.FileInputFormat
- Set a PathFilter to be applied to the input paths for the map-reduce job.
- setInputPathFilter(Job, Class<? extends PathFilter>) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Set a PathFilter to be applied to the input paths for the map-reduce job.
- setInputPaths(JobConf, String) -
Static method in class org.apache.hadoop.mapred.FileInputFormat
- Sets the given comma separated paths as the list of inputs
for the map-reduce job.
- setInputPaths(JobConf, Path...) -
Static method in class org.apache.hadoop.mapred.FileInputFormat
- Set the array of
Path
s as the list of inputs
for the map-reduce job.
- setInputPaths(Job, String) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Sets the given comma separated paths as the list of inputs
for the map-reduce job.
- setInputPaths(Job, Path...) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Set the array of
Path
s as the list of inputs
for the map-reduce job.
- setInputQuery(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputSplit(InputSplit) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- setInputTableName(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setInputWriterClass(Class<? extends InputWriter>) -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
- Sets the
InputWriter
class.
- setInsert(Document) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
- Set the instance to be an insert operation.
- setInt(String, int) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to an int
.
- setIOSortMB(int) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the IO sort space in MB.
- setIsCleanup(boolean) -
Method in class org.apache.hadoop.mapred.TaskLogAppender
- Set whether the task is a cleanup attempt or not.
- setIsJavaMapper(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set whether the Mapper is written in Java.
- setIsJavaRecordReader(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set whether the job is using a Java RecordReader.
- setIsJavaRecordWriter(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set whether the job will use a Java RecordWriter.
- setIsJavaReducer(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set whether the Reducer is written in Java.
- setIssueDate(long) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- setJar(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the user jar for the map-reduce job.
- setJarByClass(Class) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the job's jar file by finding an example class location.
- setJarByClass(Class<?>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the Jar by finding where a given class came from.
- setJob(Job) -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Set the mapreduce job
- setJobACLs(Map<JobACL, AccessControlList>) -
Method in class org.apache.hadoop.mapred.JobStatus
- Set the job acls
- setJobConf(JobConf) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the mapred job conf for this job.
- setJobConf() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- setJobEndNotificationURI(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
- setJobFile(String) -
Method in class org.apache.hadoop.mapred.Task
-
- setJobID(String) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the job ID for this job.
- setJobID(String) -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Set the job ID for this job.
- setJobName(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the user-specified job name.
- setJobName(String) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the job name for this job.
- setJobName(String) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the user-specified job name.
- setJobName(String) -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Set the job name for this job.
- setJobPriority(JobPriority) -
Method in class org.apache.hadoop.mapred.JobConf
- Set
JobPriority
for this job.
- setJobPriority(JobPriority) -
Method in class org.apache.hadoop.mapred.JobStatus
- Set the priority of the job, defaulting to NORMAL.
- setJobPriority(JobID, String) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- setJobPriority(JobID, String) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
-
- setJobPriority(String) -
Method in interface org.apache.hadoop.mapred.RunningJob
- Set the priority of a running job.
- setJobState(ControlledJob.State) -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Set the state for this job.
- setJobToken(Token<? extends TokenIdentifier>, Credentials) -
Static method in class org.apache.hadoop.mapreduce.security.TokenCache
- store job token
- setJobTokenSecret(SecretKey) -
Method in class org.apache.hadoop.mapred.Task
- Set the job token secret
- setJvmContext(JvmContext) -
Method in class org.apache.hadoop.mapred.Task
- Set the task JvmContext
- setKeepCommandFile(JobConf, boolean) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Set whether to keep the command file for debugging
- setKeepFailedTaskFiles(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Set whether the framework should keep the intermediate files for
failed tasks.
- setKeepTaskFilesPattern(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set a regular expression for task names that should be kept.
- setKeyComparator(Class<? extends WritableComparator>) -
Method in class org.apache.hadoop.mapred.join.Parser.Node
-
- setKeyFieldComparatorOptions(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
KeyFieldBasedComparator
options used to compare keys.
- setKeyFieldComparatorOptions(Job, String) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
- Set the
KeyFieldBasedComparator
options used to compare keys.
- setKeyFieldPartitionerOptions(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
KeyFieldBasedPartitioner
options used for
Partitioner
- setKeyFieldPartitionerOptions(Job, String) -
Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
- Set the
KeyFieldBasedPartitioner
options used for
Partitioner
- setKeyValue(Text, Text, byte[], int, int) -
Static method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
-
- setKind(Text) -
Method in class org.apache.hadoop.security.token.Token
- Set the token kind.
- setLanguage(String) -
Method in class org.apache.hadoop.record.compiler.ant.RccTask
- Sets the output language option
- setLastContact(long) -
Method in class org.apache.hadoop.ipc.Server.Connection
-
- setLastSeen(long) -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
-
- setLeftOffset(Configuration, int) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
- Set the subarray to be used for partitioning to
bytes[offset:]
in Python syntax.
- setLength(long) -
Method in class org.apache.hadoop.fs.BlockLocation
- Set the length of block
- setLevel(int) -
Method in interface org.apache.hadoop.net.Node
- Set this node's level in the tree.
- setLevel(int) -
Method in class org.apache.hadoop.net.NodeBase
- Set this node's level in the tree
- setLoadNativeLibraries(Configuration, boolean) -
Method in class org.apache.hadoop.util.NativeCodeLoader
- Set if native hadoop libraries, if present, can be used for this job.
- setLocalAnalysisClass(Class<? extends ILocalAnalysis>) -
Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
- Set the local analysis class.
- setLocalArchives(Configuration, String) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- Set the conf to contain the location for localized archives.
- setLocalFiles(Configuration, String) -
Static method in class org.apache.hadoop.filecache.DistributedCache
- Set the conf to contain the location for localized files.
- setLocalMaxRunningMaps(JobContext, int) -
Static method in class org.apache.hadoop.mapred.LocalJobRunner
- Set the max number of map tasks to run concurrently in the LocalJobRunner.
- setLong(String, long) -
Method in class org.apache.hadoop.conf.Configuration
- Set the value of the
name
property to a long
.
- setLongValue(Object, long) -
Method in class org.apache.hadoop.contrib.utils.join.JobBase
- Set the given counter to the given value
- setMapDebugScript(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the debug script to run when the map tasks fail.
- setMapOutputCompressorClass(Class<? extends CompressionCodec>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the given class as the
CompressionCodec
for the map outputs.
- setMapOutputKeyClass(Class<?>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the key class for the map output data.
- setMapOutputKeyClass(Class<?>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the key class for the map output data.
- setMapOutputValueClass(Class<?>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the value class for the map output data.
- setMapOutputValueClass(Class<?>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the value class for the map output data.
- setMapperClass(Class<? extends Mapper>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
Mapper
class for the job.
- setMapperClass(Class<? extends Mapper>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the
Mapper
for the job.
- setMapperClass(Job, Class<? extends Mapper<K1, V1, K2, V2>>) -
Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
- Set the application's mapper class.
- setMapperMaxSkipRecords(Configuration, long) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the number of acceptable skip records surrounding the bad record PER
bad record in mapper.
- setMapredJobID(String) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Deprecated. use
Job.setAssignedJobID(JobID)
instead
- setMapRunnerClass(Class<? extends MapRunnable>) -
Method in class org.apache.hadoop.mapred.JobConf
- Expert: Set the
MapRunnable
class for the job.
- setMapSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Turn speculative execution on or off for this job for map tasks.
- setMapSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapreduce.Job
- Turn speculative execution on or off for this job for map tasks.
- SETMASK -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- setMasterKeyId(int) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- setMaxDate(long) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- setMaxInputSplitSize(Job, long) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Set the maximum split size
- setMaxItems(long) -
Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
- Set the limit on the number of unique values
- setMaxMapAttempts(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Expert: Set the number of maximum attempts that will be made to run a
map task.
- setMaxMapTaskFailuresPercent(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Expert: Set the maximum percentage of map tasks that can fail without the
job being aborted.
- setMaxPhysicalMemoryForTask(long) -
Method in class org.apache.hadoop.mapred.JobConf
- Deprecated.
- setMaxReduceAttempts(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Expert: Set the number of maximum attempts that will be made to run a
reduce task.
- setMaxReduceTaskFailuresPercent(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the maximum percentage of reduce tasks that can fail without the job
being aborted.
- setMaxSplitSize(long) -
Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
- Specify the maximum size (in bytes) of each split.
- setMaxSplitSize(long) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
- Specify the maximum size (in bytes) of each split.
- setMaxTaskFailuresPerTracker(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the maximum no.
- setMaxVirtualMemoryForTask(long) -
Method in class org.apache.hadoop.mapred.JobConf
- Deprecated. Use
JobConf.setMemoryForMapTask(long mem)
and
Use JobConf.setMemoryForReduceTask(long mem)
- setMemory(int) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Set the total amount of buffer memory, in bytes.
- setMemoryForMapTask(long) -
Method in class org.apache.hadoop.mapred.JobConf
-
- setMemoryForReduceTask(long) -
Method in class org.apache.hadoop.mapred.JobConf
-
- setMessage(String) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the message for this job.
- setMessage(String) -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Set the message for this job.
- setMetric(String, int) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, long) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, short) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, byte) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, float) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, int) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, long) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, short) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, byte) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named metric to the specified value.
- setMetric(String, float) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named metric to the specified value.
- setMinInputSplitSize(Job, long) -
Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
- Set the minimum input split size
- setMinSplitSize(long) -
Method in class org.apache.hadoop.mapred.FileInputFormat
-
- setMinSplitSizeNode(long) -
Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
- Specify the minimum size (in bytes) of each split per node.
- setMinSplitSizeNode(long) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
- Specify the minimum size (in bytes) of each split per node.
- setMinSplitSizeRack(long) -
Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
- Specify the minimum size (in bytes) of each split per rack.
- setMinSplitSizeRack(long) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
- Specify the minimum size (in bytes) of each split per rack.
- setName(Class, String) -
Static method in class org.apache.hadoop.io.WritableName
- Set the name that a class should be known as to something other than the
class name.
- setName(String) -
Method in class org.apache.hadoop.record.meta.RecordTypeInfo
- set the name of the record
- setNames(String[]) -
Method in class org.apache.hadoop.fs.BlockLocation
- Set the names (host:port) hosting this block
- setNetworkLocation(String) -
Method in interface org.apache.hadoop.net.Node
- Set the node's network location
- setNetworkLocation(String) -
Method in class org.apache.hadoop.net.NodeBase
- Set this node's network location
- setNetworkProperties() -
Method in class org.apache.hadoop.contrib.failmon.LogParser
-
- setNextRecordRange(SortedRanges.Range) -
Method in class org.apache.hadoop.mapred.TaskStatus
- Set the next record range which is going to be processed by Task.
- setNumberOfThreads(Job, int) -
Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
- Set the number of threads in the pool for running maps.
- setNumLinesPerSplit(Job, int) -
Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
- Set the number of lines per split
- setNumMapTasks(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the number of map tasks for this job.
- setNumReduceTasks(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the requisite number of reduce tasks for this job.
- setNumReduceTasks(int) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the number of reduce tasks for the job.
- setNumTasksToExecutePerJvm(int) -
Method in class org.apache.hadoop.mapred.JobConf
- Sets the number of tasks that a spawned task JVM should run
before it exits
- setOffset(long) -
Method in class org.apache.hadoop.fs.BlockLocation
- Set the start offset of file associated with this block
- setOffsets(Configuration, int, int) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
- Set the subarray to be used for partitioning to
bytes[left:(right+1)]
in Python syntax.
- setOp(DocumentAndOp.Op) -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
- Set the type of the operation.
- setOutput(JobConf, String, String...) -
Static method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
- Initializes the reduce-part of the job with the appropriate output settings
- setOutput(Job, String, String...) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
- Initializes the reduce-part of the job with
the appropriate output settings
- setOutput(Job, String, int) -
Static method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
- Initializes the reduce-part of the job
with the appropriate output settings
- setOutputCommitter(Class<? extends OutputCommitter>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
OutputCommitter
implementation for the map-reduce job.
- setOutputCompressionType(JobConf, SequenceFile.CompressionType) -
Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
- Set the
SequenceFile.CompressionType
for the output SequenceFile
.
- setOutputCompressionType(Job, SequenceFile.CompressionType) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
- Set the
SequenceFile.CompressionType
for the output SequenceFile
.
- setOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
- Set the
CompressionCodec
to be used to compress job outputs.
- setOutputCompressorClass(Job, Class<? extends CompressionCodec>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
- Set the
CompressionCodec
to be used to compress job outputs.
- setOutputFieldCount(int) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setOutputFieldNames(String...) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setOutputFormat(Class<? extends OutputFormat>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
OutputFormat
implementation for the map-reduce job.
- setOutputFormatClass(Class<? extends OutputFormat>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the
OutputFormat
for the job.
- setOutputFormatClass(Job, Class<? extends OutputFormat>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
- Set the underlying output format for LazyOutputFormat.
- setOutputKeyClass(Class<?>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the key class for the job output data.
- setOutputKeyClass(Class<?>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the key class for the job output data.
- setOutputKeyClass(Class) -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
- Sets the output key class class.
- setOutputKeyComparatorClass(Class<? extends RawComparator>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
RawComparator
comparator used to compare keys.
- setOutputName(JobContext, String) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
- Set the base output name for output file to be created.
- setOutputPath(JobConf, Path) -
Static method in class org.apache.hadoop.mapred.FileOutputFormat
- Set the
Path
of the output directory for the map-reduce job.
- setOutputPath(Job, Path) -
Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
- Set the
Path
of the output directory for the map-reduce job.
- setOutputReaderClass(Class<? extends OutputReader>) -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
- Sets the
OutputReader
class.
- setOutputTableName(String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
-
- setOutputValueClass(Class<?>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the value class for job outputs.
- setOutputValueClass(Class<?>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the value class for job outputs.
- setOutputValueClass(Class) -
Method in class org.apache.hadoop.streaming.io.IdentifierResolver
- Sets the output value class.
- setOutputValueGroupingComparator(Class<? extends RawComparator>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the user defined
RawComparator
comparator for
grouping keys in the input to the reduce.
- setOwner(String) -
Method in class org.apache.hadoop.fs.FileStatus
- Sets owner.
- setOwner(Path, String, String) -
Method in class org.apache.hadoop.fs.FileSystem
- Set owner of a path (i.e.
- setOwner(Path, String, String) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Set owner of a path (i.e.
- setOwner(Path, String, String) -
Method in class org.apache.hadoop.fs.HarFileSystem
- not implemented.
- setOwner(Path, String, String) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Use the command chown to set owner.
- setParent(Node) -
Method in interface org.apache.hadoop.net.Node
- Set this node's parent
- setParent(Node) -
Method in class org.apache.hadoop.net.NodeBase
- Set this node's parent
- setPartitionerClass(Class<? extends Partitioner>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
Partitioner
class used to partition
Mapper
-outputs to be sent to the Reducer
s.
- setPartitionerClass(Class<? extends Partitioner>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the
Partitioner
for the job.
- setPartitionFile(JobConf, Path) -
Static method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
- Set the path to the SequenceFile storing the sorted partition keyset.
- setPartitionFile(Configuration, Path) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
- Set the path to the SequenceFile storing the sorted partition keyset.
- setPattern(Configuration, String) -
Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
- Define the filtering regex and stores it in conf
- setPattern(Configuration, String) -
Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
- Define the filtering regex and stores it in conf
- setPeriod(int) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Sets the timer period
- setPermission(FsPermission) -
Method in class org.apache.hadoop.fs.FileStatus
- Sets permission.
- setPermission(Path, FsPermission) -
Method in class org.apache.hadoop.fs.FileSystem
- Set permission of a path.
- setPermission(File, FsPermission) -
Static method in class org.apache.hadoop.fs.FileUtil
- Set permissions to the required value.
- setPermission(Path, FsPermission) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Set permission of a path.
- setPermission(Path, FsPermission) -
Method in class org.apache.hadoop.fs.HarFileSystem
- Not implemented.
- setPermission(Path, FsPermission) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Use the command chmod to set permission.
- setPhase(TaskStatus.Phase) -
Method in class org.apache.hadoop.mapred.Task
- Set current phase of the task.
- setPingInterval(Configuration, int) -
Static method in class org.apache.hadoop.ipc.Client
- set the ping interval value in configuration
- setPrinter(DancingLinks.SolutionAcceptor<Pentomino.ColumnName>) -
Method in class org.apache.hadoop.examples.dancing.Pentomino
- Set the printer for the puzzle.
- setPriority(JobPriority) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- setProfileEnabled(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Set whether the system should collect profiler information for some of
the tasks in this job? The information is stored in the user log
directory.
- setProfileParams(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the profiler configuration arguments.
- setProfileTaskRange(boolean, String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the ranges of maps or reduces to profile.
- setProgress(float) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- setProgress(float) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setProgressable(Progressable) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Set the progressable object in order to report progress.
- setProperty(String, String) -
Static method in class org.apache.hadoop.contrib.failmon.Environment
- Sets the value of a property inthe configuration file.
- setQueueName(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the name of the queue to which this job should be submitted.
- setQueueName(String) -
Method in class org.apache.hadoop.mapred.JobQueueInfo
- Set the queue name of the JobQueueInfo
- setQueueName(String) -
Method in class org.apache.hadoop.mapred.QueueAclsInfo
-
- setQueueState(String) -
Method in class org.apache.hadoop.mapred.JobQueueInfo
- Set the state of the queue
- setQuietMode(boolean) -
Method in class org.apache.hadoop.conf.Configuration
- Set the quietness-mode.
- setReduceDebugScript(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the debug script to run when the reduce tasks fail.
- setReducer(JobConf, Class<? extends Reducer<K1, V1, K2, V2>>, Class<? extends K1>, Class<? extends V1>, Class<? extends K2>, Class<? extends V2>, boolean, JobConf) -
Static method in class org.apache.hadoop.mapred.lib.ChainReducer
- Sets the Reducer class to the chain job's JobConf.
- setReducerClass(Class<? extends Reducer>) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the
Reducer
class for the job.
- setReducerClass(Class<? extends Reducer>) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the
Reducer
for the job.
- setReducerMaxSkipGroups(Configuration, long) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the number of acceptable skip groups surrounding the bad group PER
bad group in reducer.
- setReduceSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Turn speculative execution on or off for this job for reduce tasks.
- setReduceSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapreduce.Job
- Turn speculative execution on or off for this job for reduce tasks.
- setReplication(Path, short) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- Set replication for an existing file.
- setReplication(Path, short) -
Method in class org.apache.hadoop.fs.FileSystem
- Set replication for an existing file.
- setReplication(Path, short) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Set replication for an existing file.
- setReplication(Path, short) -
Method in class org.apache.hadoop.fs.HarFileSystem
- Not implemented.
- setReplication(Path, short) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- setRightOffset(Configuration, int) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
- Set the subarray to be used for partitioning to
bytes[:(offset+1)]
in Python syntax.
- setRules(String) -
Static method in class org.apache.hadoop.security.KerberosName
- Set the rules.
- setRunningTaskAttempts(Collection<TaskAttemptID>) -
Method in class org.apache.hadoop.mapred.TaskReport
- set running attempt(s) of the task.
- setRunState(int) -
Method in class org.apache.hadoop.mapred.JobStatus
- Change the current run state of the job.
- setRunState(TaskStatus.State) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setSafeMode(JobTracker.SafeModeAction) -
Method in interface org.apache.hadoop.mapred.AdminOperationsProtocol
- Set safe mode for the JobTracker.
- setSafeMode(JobTracker.SafeModeAction) -
Method in class org.apache.hadoop.mapred.JobTracker
-
- setSchedulingInfo(Object) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- setSchedulingInfo(String) -
Method in class org.apache.hadoop.mapred.JobQueueInfo
- Set the scheduling information associated to particular job queue
- setSchedulingInfo(String) -
Method in class org.apache.hadoop.mapred.JobStatus
- Used to set the scheduling information associated to a particular Job.
- setSequenceFileOutputKeyClass(JobConf, Class<?>) -
Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
- Set the key class for the
SequenceFile
- setSequenceFileOutputKeyClass(Job, Class<?>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
- Set the key class for the
SequenceFile
- setSequenceFileOutputValueClass(JobConf, Class<?>) -
Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
- Set the value class for the
SequenceFile
- setSequenceFileOutputValueClass(Job, Class<?>) -
Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
- Set the value class for the
SequenceFile
- setSequenceNumber(int) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- setService(Text) -
Method in class org.apache.hadoop.security.token.Token
- Set the service on which the token is supposed to be used
- setSessionId(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the user-specified session identifier.
- setSessionTimeZone(Configuration, Connection) -
Static method in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
- Set session time zone
- setSize(int) -
Method in class org.apache.hadoop.io.BytesWritable
- Change the size of the buffer.
- setSizes(long[]) -
Method in class org.apache.hadoop.filecache.TaskDistributedCacheManager
-
- setSkipOutputPath(JobConf, Path) -
Static method in class org.apache.hadoop.mapred.SkipBadRecords
- Set the directory to which skipped records are written.
- setSkipping(boolean) -
Method in class org.apache.hadoop.mapred.Task
- Sets whether to run Task in skipping mode.
- setSkipRanges(SortedRanges) -
Method in class org.apache.hadoop.mapred.Task
- Set skipRanges.
- setSocketSendBufSize(int) -
Method in class org.apache.hadoop.ipc.Server
- Sets the socket buffer size used for responding to RPCs
- setSortComparatorClass(Class<? extends RawComparator>) -
Method in class org.apache.hadoop.mapreduce.Job
- Define the comparator that controls how the keys are sorted before they
are passed to the
Reducer
.
- setSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Turn speculative execution on or off for this job.
- setSpeculativeExecution(boolean) -
Method in class org.apache.hadoop.mapreduce.Job
- Turn speculative execution on or off for this job.
- setStart(long) -
Method in class org.apache.hadoop.io.compress.SplitCompressionInputStream
-
- setState(ParseState) -
Static method in class org.apache.hadoop.contrib.failmon.PersistentState
- Set the state of parsing for a particular log file.
- setState(int) -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Set the state for this job.
- setStatement(PreparedStatement) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
-
- setStateString(String) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setStatus(String) -
Method in interface org.apache.hadoop.mapred.Reporter
- Set the status description for the task.
- setStatus(String) -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- setStatus(TaskTrackerStatus) -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
- Set the current
TaskTrackerStatus
of the TaskTracker
.
- setStatus(String) -
Method in class org.apache.hadoop.mapreduce.StatusReporter
-
- setStatus(String) -
Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
- Set the current status of the task to the given string.
- setStatus(String) -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- setStatus(String) -
Method in class org.apache.hadoop.util.Progress
-
- setStrings(String, String...) -
Method in class org.apache.hadoop.conf.Configuration
- Set the array of string values for the
name
property as
as comma delimited values.
- setSuccessfulAttempt(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.TaskReport
- set successful attempt ID of the task.
- setTabSize(int) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- setTag(Text) -
Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- setTag(String, String) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named tag to the specified value.
- setTag(String, int) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named tag to the specified value.
- setTag(String, long) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named tag to the specified value.
- setTag(String, short) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named tag to the specified value.
- setTag(String, byte) -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Sets the named tag to the specified value.
- setTag(String, String) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named tag to the specified value.
- setTag(String, int) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named tag to the specified value.
- setTag(String, long) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named tag to the specified value.
- setTag(String, short) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named tag to the specified value.
- setTag(String, byte) -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Sets the named tag to the specified value.
- setTaskId(String) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- Deprecated. use
TaskCompletionEvent.setTaskID(TaskAttemptID)
instead.
- setTaskID(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- Sets task id.
- setTaskId(String) -
Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- setTaskOutputFilter(JobClient.TaskStatusFilter) -
Method in class org.apache.hadoop.mapred.JobClient
- Deprecated.
- setTaskOutputFilter(JobConf, JobClient.TaskStatusFilter) -
Static method in class org.apache.hadoop.mapred.JobClient
- Modify the JobConf to set the task output filter.
- setTaskRunTime(int) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- Set the task completion time
- setTaskStatus(TaskCompletionEvent.Status) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- Set task status.
- setTaskTracker(String) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- setTaskTrackerHttp(String) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
- Set task tracker http location.
- setThreads(int, int) -
Method in class org.apache.hadoop.http.HttpServer
- Set the min, max number of worker threads (simultaneous connections).
- setTimes(Path, long, long) -
Method in class org.apache.hadoop.fs.FileSystem
- Set access time of a file
- setTokenService(Token<?>, InetSocketAddress) -
Static method in class org.apache.hadoop.security.SecurityUtil
- Set the given token's service to the format expected by the RPC client
- setTopologyPaths(String[]) -
Method in class org.apache.hadoop.fs.BlockLocation
- Set the network topology paths of the hosts
- setTotalLogFileSize(long) -
Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- setUMask(Configuration, FsPermission) -
Static method in class org.apache.hadoop.fs.permission.FsPermission
- Set the user file creation mask (umask)
- setup(LocalDirAllocator, TaskTracker.LocalStorage) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
-
- setup(LocalDirAllocator, TaskTracker.LocalStorage) -
Method in class org.apache.hadoop.mapred.TaskController
- Does initialization and setup.
- setup(Mapper<K, V, Text, Text>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
-
- setup(Reducer<Text, Text, Text, Text>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
-
- setup(Mapper<K1, V1, K2, V2>.Context) -
Method in class org.apache.hadoop.mapreduce.lib.input.DelegatingMapper
-
- setup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Mapper
- Called once at the beginning of the task.
- setup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) -
Method in class org.apache.hadoop.mapreduce.Reducer
- Called once at the start of the task.
- setupCache(Configuration, String, String) -
Method in class org.apache.hadoop.filecache.TaskDistributedCacheManager
- Retrieve public distributed cache files into the local cache and updates
the task configuration (which has been passed in via the constructor).
- setUpdate(Document, Term) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
- Set the instance to be an update operation.
- setupJob(JobContext) -
Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- setupJob(JobContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- For the framework to setup the job output during initialization
- setupJob(JobContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- This method implements the new interface by calling the old method.
- setupJob(JobContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
- Create the temporary directory that is the root of all of the task
work directories.
- setupJob(JobContext) -
Method in class org.apache.hadoop.mapreduce.OutputCommitter
- For the framework to setup the job output during initialization
- setupJobConf(int, int, long, int, long, int) -
Method in class org.apache.hadoop.examples.SleepJob
-
- setupProgress() -
Method in class org.apache.hadoop.mapred.JobStatus
-
- setupProgress() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Get the progress of the job's setup-tasks, as a float between 0.0
and 1.0.
- setupProgress() -
Method in class org.apache.hadoop.mapreduce.Job
- Get the progress of the job's setup, as a float between 0.0
and 1.0.
- setupTask(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- setupTask(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- Sets up output for the task.
- setupTask(TaskAttemptContext) -
Method in class org.apache.hadoop.mapred.OutputCommitter
- This method implements the new interface by calling the old method.
- setupTask(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
- No task setup required.
- setupTask(TaskAttemptContext) -
Method in class org.apache.hadoop.mapreduce.OutputCommitter
- Sets up output for the task.
- setUseNewMapper(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Set whether the framework should use the new api for the mapper.
- setUseNewReducer(boolean) -
Method in class org.apache.hadoop.mapred.JobConf
- Set whether the framework should use the new api for the reducer.
- setUser(String) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the reported username for this job.
- setValue(long) -
Method in class org.apache.hadoop.mapreduce.Counter
- Set this counter by the given value
- setValue(Object) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
- Set the typed bytes from a given Java object.
- setVerbose(boolean) -
Method in class org.apache.hadoop.streaming.JarBuilder
-
- setVerifyChecksum(boolean) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
- Set whether to verify checksum.
- setVerifyChecksum(boolean) -
Method in class org.apache.hadoop.fs.FileSystem
- Set the verify checksum flag.
- setVerifyChecksum(boolean) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Set the verify checksum flag.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Set the current working directory for the given file system.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Set the current working directory for the given file system.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
-
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
- Set the working directory to the given directory.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
- Set the working directory to the given directory.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.mapred.JobConf
- Set the current working directory for the default file system.
- setWorkingDirectory(Path) -
Method in class org.apache.hadoop.mapreduce.Job
- Set the current working directory for the default file system.
- setWorkingDirectory(File) -
Method in class org.apache.hadoop.util.Shell
- set the working directory
- setWriter(IFile.Writer<K, V>) -
Method in class org.apache.hadoop.mapred.Task.CombineOutputCollector
-
- setWriteSkipRecs(boolean) -
Method in class org.apache.hadoop.mapred.Task
- Set whether to write skip records.
- Shard - Class in org.apache.hadoop.contrib.index.mapred
- This class represents the metadata of a shard.
- Shard() -
Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
- Constructor.
- Shard(long, String, long) -
Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
- Construct a shard from a versio number, a directory and a generation
number.
- Shard(Shard) -
Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
- Construct using a shard object.
- ShardWriter - Class in org.apache.hadoop.contrib.index.lucene
- The initial version of an index is stored in the perm dir.
- ShardWriter(FileSystem, Shard, String, IndexUpdateConfiguration) -
Constructor for class org.apache.hadoop.contrib.index.lucene.ShardWriter
- Constructor
- Shell - Class in org.apache.hadoop.util
- A base class for running a Unix command.
- Shell() -
Constructor for class org.apache.hadoop.util.Shell
-
- Shell(long) -
Constructor for class org.apache.hadoop.util.Shell
-
- Shell.ExitCodeException - Exception in org.apache.hadoop.util
- This is an IOException with exit code added.
- Shell.ExitCodeException(int, String) -
Constructor for exception org.apache.hadoop.util.Shell.ExitCodeException
-
- Shell.ShellCommandExecutor - Class in org.apache.hadoop.util
- A simple shell command executor.
- Shell.ShellCommandExecutor(String[]) -
Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- Shell.ShellCommandExecutor(String[], File) -
Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- Shell.ShellCommandExecutor(String[], File, Map<String, String>) -
Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- Shell.ShellCommandExecutor(String[], File, Map<String, String>, long) -
Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
- Create a new instance of the ShellCommandExecutor to execute a command.
- ShellBasedUnixGroupsMapping - Class in org.apache.hadoop.security
- A simple shell-based implementation of
GroupMappingServiceProvider
that exec's the groups
shell command to fetch the group
memberships of a given user. - ShellBasedUnixGroupsMapping() -
Constructor for class org.apache.hadoop.security.ShellBasedUnixGroupsMapping
-
- ShellBasedUnixGroupsNetgroupMapping - Class in org.apache.hadoop.security
- A simple shell-based implementation of
GroupMappingServiceProvider
that exec's the groups
shell command to fetch the group
memberships of a given user. - ShellBasedUnixGroupsNetgroupMapping() -
Constructor for class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
-
- ShellParser - Class in org.apache.hadoop.contrib.failmon
- Objects of this class parse the output of system command-line
utilities that can give information about the state of
various hardware components in the system.
- ShellParser() -
Constructor for class org.apache.hadoop.contrib.failmon.ShellParser
-
- shippedCanonFiles_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- shouldDie() -
Method in class org.apache.hadoop.mapred.JvmTask
-
- shouldPreserveInput() -
Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
- shouldReset() -
Method in class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
-
- shouldRetry(Exception, int) -
Method in class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry
-
- shouldRetry(Exception, int) -
Method in interface org.apache.hadoop.io.retry.RetryPolicy
-
Determines whether the framework should retry a
method for the given exception, and the number
of retries that have been made for that operation
so far.
- shouldTruncateLogs(JVMInfo) -
Method in class org.apache.hadoop.mapred.TaskLogsTruncater
- Check the log file sizes generated by the attempts that ran in a
particular JVM
- shuffleError(TaskAttemptID, String, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
- A reduce-task failed to shuffle the map-outputs.
- shuffleError(TaskAttemptID, String, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Report that a reduce-task couldn't shuffle map-outputs.
- ShuffleExceptionTracker - Class in org.apache.hadoop.mapred
- This class is used to track shuffle exceptions.
- shutdown() -
Method in class org.apache.hadoop.fs.DU
- Shut down the refreshing thread.
- shutdown() -
Method in class org.apache.hadoop.ipc.metrics.RpcInstrumentation
- Shutdown the instrumentation for the process
- shutdown() -
Method in class org.apache.hadoop.mapred.JobHistoryServer
- Shutsdown the history server if already initialized
- shutdown() -
Method in class org.apache.hadoop.mapred.TaskTracker
-
- shutdown() -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- shutdown() -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- shutdown() -
Method in interface org.apache.hadoop.metrics2.MetricsSystem
- Shutdown the metrics system completely (usually during server shutdown.)
The MetricsSystemMXBean will be unregistered.
- sign(String) -
Method in class org.apache.hadoop.security.authentication.util.Signer
- Returns a signed string.
- signalTask(String, int, ProcessTree.Signal) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
-
- signalTask(String, int, ProcessTree.Signal) -
Method in class org.apache.hadoop.mapred.TaskController
- Send a signal to a task pid as the user.
- SIGNATURE_SECRET -
Static variable in class org.apache.hadoop.security.authentication.server.AuthenticationFilter
- Constant for the property that specifies the secret to use for signing the HTTP Cookies.
- Signer - Class in org.apache.hadoop.security.authentication.util
- Signs strings and verifies signed strings using a SHA digest.
- Signer(byte[]) -
Constructor for class org.apache.hadoop.security.authentication.util.Signer
- Creates a Signer instance using the specified secret.
- SignerException - Exception in org.apache.hadoop.security.authentication.util
- Exception thrown by
Signer
when a string signature is invalid. - SignerException(String) -
Constructor for exception org.apache.hadoop.security.authentication.util.SignerException
- Creates an exception instance.
- SimpleCharStream - Class in org.apache.hadoop.record.compiler.generated
- An implementation of interface CharStream, where the stream is assumed to
contain only ASCII characters (without unicode processing).
- SimpleCharStream(Reader, int, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(Reader, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(Reader) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream, String, int, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream, int, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream, String, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream, int, int) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream, String) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- SimpleCharStream(InputStream) -
Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- simpleHostname(String) -
Static method in class org.apache.hadoop.util.StringUtils
- Given a full hostname, return the word upto the first dot.
- size() -
Method in class org.apache.hadoop.conf.Configuration
- Return the number of keys in the configuration.
- size() -
Method in class org.apache.hadoop.io.file.tfile.ByteArray
-
- size() -
Method in interface org.apache.hadoop.io.file.tfile.RawComparable
- Get the size of the byte range in the byte array.
- size() -
Static method in class org.apache.hadoop.io.file.tfile.Utils.Version
- Get the size of the serialized Version object.
- size() -
Method in class org.apache.hadoop.io.MapWritable
-
- size() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- size() -
Method in class org.apache.hadoop.mapred.Counters.Group
- Returns the number of counters in this group.
- size() -
Method in class org.apache.hadoop.mapred.Counters
- Returns the total number of counters, by summing the number of counters
in each group.
- size() -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- The number of children in this Tuple.
- size() -
Method in class org.apache.hadoop.mapreduce.CounterGroup
- Returns the number of counters in this group.
- size() -
Method in class org.apache.hadoop.metrics.util.MetricsRegistry
- Deprecated.
- size() -
Method in class org.apache.hadoop.util.PriorityQueue
- Returns the number of elements currently stored in the PriorityQueue.
- SIZE_OF_INTEGER -
Static variable in class org.apache.hadoop.util.DataChecksum
-
- SIZE_UNKNOWN -
Static variable in class org.apache.hadoop.fs.LocalDirAllocator
- Used when size of file to be allocated is unknown.
- skip(long) -
Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- skip(long) -
Method in class org.apache.hadoop.fs.FSInputChecker
- Skips over and discards
n
bytes of data from the
input stream.
- skip(long) -
Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- skip(DataInput) -
Static method in class org.apache.hadoop.io.Text
- Skips over one Text in the input.
- skip(DataInput) -
Static method in class org.apache.hadoop.io.UTF8
- Deprecated. Skips over one UTF8 in the input.
- skip(K) -
Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
- Skip key-value pairs with keys less than or equal to the key provided.
- skip(K) -
Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
- Pass skip key to child RRs.
- skip(K) -
Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
- Skip key-value pairs with keys less than or equal to the key provided.
- skip(RecordInput, String, TypeID) -
Static method in class org.apache.hadoop.record.meta.Utils
- read/skip bytes from stream based on a type
- skip(long) -
Method in class org.apache.hadoop.security.SaslInputStream
- Skips
n
bytes of input from the bytes that can be read from
this input stream without blocking.
- SkipBadRecords - Class in org.apache.hadoop.mapred
- Utility class for skip bad records functionality.
- SkipBadRecords() -
Constructor for class org.apache.hadoop.mapred.SkipBadRecords
-
- skipCompressedByteArray(DataInput) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- skipFully(InputStream, long) -
Static method in class org.apache.hadoop.io.IOUtils
- Similar to readFully().
- skipFully(DataInput, int) -
Static method in class org.apache.hadoop.io.WritableUtils
- Skip len number of bytes in input streamin
- skipToNextMarker(long, int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
- This method tries to find the marker (passed to it as the first parameter)
in the stream.
- skipType() -
Method in class org.apache.hadoop.typedbytes.TypedBytesInput
- Skips a type byte.
- SleepJob - Class in org.apache.hadoop.examples
- Dummy class for testing MR framefork.
- SleepJob() -
Constructor for class org.apache.hadoop.examples.SleepJob
-
- SleepJob.EmptySplit - Class in org.apache.hadoop.examples
-
- SleepJob.EmptySplit() -
Constructor for class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- SleepJob.SleepInputFormat - Class in org.apache.hadoop.examples
-
- SleepJob.SleepInputFormat() -
Constructor for class org.apache.hadoop.examples.SleepJob.SleepInputFormat
-
- SMALL_THRESH -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- SMARTParser - Class in org.apache.hadoop.contrib.failmon
- Objects of this class parse the output of smartmontools to
gather information about the state of disks in the system.
- SMARTParser() -
Constructor for class org.apache.hadoop.contrib.failmon.SMARTParser
- Constructs a SMARTParser and reads the list of disk
devices to query
- SnappyCodec - Class in org.apache.hadoop.io.compress
- This class creates snappy compressors/decompressors.
- SnappyCodec() -
Constructor for class org.apache.hadoop.io.compress.SnappyCodec
-
- SnappyCompressor - Class in org.apache.hadoop.io.compress.snappy
- A
Compressor
based on the snappy compression algorithm. - SnappyCompressor(int) -
Constructor for class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Creates a new compressor.
- SnappyCompressor() -
Constructor for class org.apache.hadoop.io.compress.snappy.SnappyCompressor
- Creates a new compressor with the default buffer size.
- SnappyDecompressor - Class in org.apache.hadoop.io.compress.snappy
- A
Decompressor
based on the snappy compression algorithm. - SnappyDecompressor(int) -
Constructor for class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Creates a new compressor.
- SnappyDecompressor() -
Constructor for class org.apache.hadoop.io.compress.snappy.SnappyDecompressor
- Creates a new decompressor with the default buffer size.
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutable
- Get a snapshot/snapshot of the metric
- snapshot(MetricsRecordBuilder) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutable
- Get a sampe/snapshot of metric if changed
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterInt
-
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableCounterLong
-
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeInt
-
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableGaugeLong
-
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricMutableStat
-
- snapshot(MetricsRecordBuilder, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Sample all the mutable metrics and put the snapshot in the builder
- SocketInputStream - Class in org.apache.hadoop.net
- This implements an input stream that can have a timeout while reading.
- SocketInputStream(ReadableByteChannel, long) -
Constructor for class org.apache.hadoop.net.SocketInputStream
- Create a new input stream with the given timeout.
- SocketInputStream(Socket, long) -
Constructor for class org.apache.hadoop.net.SocketInputStream
- Same as SocketInputStream(socket.getChannel(), timeout):
Create a new input stream with the given timeout.
- SocketInputStream(Socket) -
Constructor for class org.apache.hadoop.net.SocketInputStream
- Same as SocketInputStream(socket.getChannel(), socket.getSoTimeout())
:
Create a new input stream with the given timeout.
- SocketOutputStream - Class in org.apache.hadoop.net
- This implements an output stream that can have a timeout while writing.
- SocketOutputStream(WritableByteChannel, long) -
Constructor for class org.apache.hadoop.net.SocketOutputStream
- Create a new ouput stream with the given timeout.
- SocketOutputStream(Socket, long) -
Constructor for class org.apache.hadoop.net.SocketOutputStream
- Same as SocketOutputStream(socket.getChannel(), timeout):
Create a new ouput stream with the given timeout.
- SocksSocketFactory - Class in org.apache.hadoop.net
- Specialized SocketFactory to create sockets with a SOCKS proxy
- SocksSocketFactory() -
Constructor for class org.apache.hadoop.net.SocksSocketFactory
- Default empty constructor (for use with the reflection API).
- SocksSocketFactory(Proxy) -
Constructor for class org.apache.hadoop.net.SocksSocketFactory
- Constructor with a supplied Proxy
- solution(List<List<ColumnName>>) -
Method in interface org.apache.hadoop.examples.dancing.DancingLinks.SolutionAcceptor
- A callback to return a solution to the application.
- solve(int[], DancingLinks.SolutionAcceptor<ColumnName>) -
Method in class org.apache.hadoop.examples.dancing.DancingLinks
- Given a prefix, find solutions under it.
- solve(DancingLinks.SolutionAcceptor<ColumnName>) -
Method in class org.apache.hadoop.examples.dancing.DancingLinks
- Solve a complete problem
- solve(int[]) -
Method in class org.apache.hadoop.examples.dancing.Pentomino
- Find all of the solutions that start with the given prefix.
- solve() -
Method in class org.apache.hadoop.examples.dancing.Pentomino
- Find all of the solutions to the puzzle.
- solve() -
Method in class org.apache.hadoop.examples.dancing.Sudoku
-
- Sort<K,V> - Class in org.apache.hadoop.examples
- This is the trivial map/reduce program that does absolutely nothing
other than use the framework to fragment and sort the input values.
- Sort() -
Constructor for class org.apache.hadoop.examples.Sort
-
- sort(Path[], Path, boolean) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Perform a file sort from a set of input files into an output file.
- sort(Path, Path) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- The backwards compatible interface to sort.
- sort(IndexedSortable, int, int) -
Method in class org.apache.hadoop.util.HeapSort
- Sort the given range of items using heap sort.
- sort(IndexedSortable, int, int, Progressable) -
Method in class org.apache.hadoop.util.HeapSort
- Same as
IndexedSorter.sort(IndexedSortable,int,int)
, but indicate progress
periodically.
- sort(IndexedSortable, int, int) -
Method in interface org.apache.hadoop.util.IndexedSorter
- Sort the items accessed through the given IndexedSortable over the given
range of logical indices.
- sort(IndexedSortable, int, int, Progressable) -
Method in interface org.apache.hadoop.util.IndexedSorter
- Same as
IndexedSorter.sort(IndexedSortable,int,int)
, but indicate progress
periodically.
- sort(IndexedSortable, int, int) -
Method in class org.apache.hadoop.util.QuickSort
- Sort the given range of items using quick sort.
- sort(IndexedSortable, int, int, Progressable) -
Method in class org.apache.hadoop.util.QuickSort
- Same as
IndexedSorter.sort(IndexedSortable,int,int)
, but indicate progress
periodically.
- sortAndIterate(Path[], Path, boolean) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Perform a file sort from a set of input files and return an iterator.
- SortedMapWritable - Class in org.apache.hadoop.io
- A Writable SortedMap.
- SortedMapWritable() -
Constructor for class org.apache.hadoop.io.SortedMapWritable
- default constructor.
- SortedMapWritable(SortedMapWritable) -
Constructor for class org.apache.hadoop.io.SortedMapWritable
- Copy constructor.
- SOURCE_TAGS_FIELD -
Static variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- specialConstructor -
Variable in exception org.apache.hadoop.record.compiler.generated.ParseException
- This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
- specialToken -
Variable in class org.apache.hadoop.record.compiler.generated.Token
- This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
- specToString(String, String, int, List<Integer>, List<Integer>) -
Static method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
-
- spilledRecordsCounter -
Variable in class org.apache.hadoop.mapred.Task
-
- split(int) -
Method in class org.apache.hadoop.examples.dancing.DancingLinks
- Generate a list of row choices to cover the first moves.
- split -
Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter
-
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.BooleanSplitter
-
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.DateSplitter
-
- split(Configuration, ResultSet, String) -
Method in interface org.apache.hadoop.mapreduce.lib.db.DBSplitter
- Given a ResultSet containing one record (and already advanced to that record)
with two columns (a low value, and a high value, both of the same type), determine
a set of splits that span the given values.
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.FloatSplitter
-
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter
-
- split(Configuration, ResultSet, String) -
Method in class org.apache.hadoop.mapreduce.lib.db.TextSplitter
- This method needs to determine the splits between two user-provided strings.
- split -
Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
-
- split(String) -
Static method in class org.apache.hadoop.util.StringUtils
- Split a string using the default separator
- split(String, char, char) -
Static method in class org.apache.hadoop.util.StringUtils
- Split a string using the given separator
- SPLIT_MINSIZE_PERNODE -
Static variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
-
- SPLIT_MINSIZE_PERRACK -
Static variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
-
- SplitCompressionInputStream - Class in org.apache.hadoop.io.compress
- An InputStream covering a range of compressed data.
- SplitCompressionInputStream(InputStream, long, long) -
Constructor for class org.apache.hadoop.io.compress.SplitCompressionInputStream
-
- splitKerberosName(String) -
Static method in class org.apache.hadoop.security.SaslRpcServer
- Splitting fully qualified Kerberos name into parts
- splitKeyVal(byte[], int, int, Text, Text, int, int) -
Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
- split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], int, int, Text, Text, int) -
Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
- split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], Text, Text, int, int) -
Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
- split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], Text, Text, int) -
Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
- split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], int, int, Text, Text, int, int) -
Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated. use
StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text,
int, int)
- splitKeyVal(byte[], int, int, Text, Text, int) -
Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated. use
StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int)
- splitKeyVal(byte[], Text, Text, int, int) -
Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated. use
StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int, int)
- splitKeyVal(byte[], Text, Text, int) -
Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated. use
StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int)
- SplitMetaInfoReader - Class in org.apache.hadoop.mapreduce.split
- A (internal) utility that reads the split meta info and creates
split meta info objects
- SplitMetaInfoReader() -
Constructor for class org.apache.hadoop.mapreduce.split.SplitMetaInfoReader
-
- SplittableCompressionCodec - Interface in org.apache.hadoop.io.compress
- This interface is meant to be implemented by those compression codecs
which are capable to compress / decompress a stream starting at any
arbitrary position.
- SplittableCompressionCodec.READ_MODE - Enum in org.apache.hadoop.io.compress
- During decompression, data can be read off from the decompressor in two
modes, namely continuous and blocked.
- SPNEGO_FILTER -
Static variable in class org.apache.hadoop.http.HttpServer
-
- StandardSocketFactory - Class in org.apache.hadoop.net
- Specialized SocketFactory to create sockets with a SOCKS proxy
- StandardSocketFactory() -
Constructor for class org.apache.hadoop.net.StandardSocketFactory
- Default empty constructor (for use with the reflection API).
- start() -
Method in class org.apache.hadoop.fs.DU
- Start the disk usage checking thread.
- start() -
Method in class org.apache.hadoop.http.HttpServer
- Start the server.
- start() -
Method in class org.apache.hadoop.ipc.Server
- Starts the service.
- start() -
Method in class org.apache.hadoop.mapred.JobHistoryServer
-
- start() -
Method in class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
- Starts managing the logs
- start() -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- start() -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- start() -
Method in interface org.apache.hadoop.metrics2.MetricsSystemMXBean
- Start the metrics system
- start(Object) -
Method in interface org.apache.hadoop.util.ServicePlugin
- This method is invoked when the service instance has been started.
- startCleanupThread() -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
- Start the background thread
- startCommunicationThread() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- startCycle(int) -
Method in class org.apache.hadoop.util.QueueProcessingStatistics
-
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.FileSystem
- Returns a local File that the user can write output to.
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.FilterFileSystem
- Returns a local File that the user can write output to.
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.HarFileSystem
- not implemented.
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- startLocalOutput(Path, Path) -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- startMap(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startMap(TreeMap, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startMap(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- startMap(TreeMap, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startMap(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Check the mark for start of the serialized map.
- startMap(TreeMap, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Mark the start of a map to be serialized.
- startMap(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- startMap(TreeMap, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- startMap(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- startMap(TreeMap, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- startMonitoring() -
Method in class org.apache.hadoop.metrics.file.FileContext
- Deprecated. Starts or restarts monitoring, by opening in append-mode, the
file specified by the
fileName
attribute,
if specified.
- startMonitoring() -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Starts or restarts monitoring, the emitting of metrics records as they are
updated.
- startMonitoring() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Starts or restarts monitoring, the emitting of metrics records.
- startMonitoring() -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- startMonitoring() -
Method in class org.apache.hadoop.metrics.spi.NullContext
- Deprecated. Do-nothing version of startMonitoring
- startNextPhase() -
Method in class org.apache.hadoop.util.Progress
- Called during execution to move to the next phase at this level in the
tree.
- startNotifier() -
Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- startRecord(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startRecord(Record, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startRecord(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- startRecord(Record, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startRecord(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Check the mark for start of the serialized record.
- startRecord(Record, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Mark the start of a record to be serialized.
- startRecord(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- startRecord(Record, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- startRecord(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- startRecord(Record, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- startThreads() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
- should be called before this object is used
- startTracker(JobConf) -
Static method in class org.apache.hadoop.mapred.JobTracker
- Start the JobTracker with given configuration.
- startTracker(JobConf, String) -
Static method in class org.apache.hadoop.mapred.JobTracker
-
- startTracker(JobConf, String, boolean) -
Static method in class org.apache.hadoop.mapred.JobTracker
-
- startupShutdownMessage(Class<?>, String[], Log) -
Static method in class org.apache.hadoop.util.StringUtils
- Print a log message for starting up and shutting down
- startVector(String) -
Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startVector(ArrayList, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startVector(String) -
Method in class org.apache.hadoop.record.CsvRecordInput
-
- startVector(ArrayList, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startVector(String) -
Method in interface org.apache.hadoop.record.RecordInput
- Check the mark for start of the serialized vector.
- startVector(ArrayList, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Mark the start of a vector to be serialized.
- startVector(String) -
Method in class org.apache.hadoop.record.XmlRecordInput
-
- startVector(ArrayList, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- startVector(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
-
- startVector(ArrayList, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- stat2Paths(FileStatus[]) -
Static method in class org.apache.hadoop.fs.FileUtil
- convert an array of FileStatus to an array of Path
- stat2Paths(FileStatus[], Path) -
Static method in class org.apache.hadoop.fs.FileUtil
- convert an array of FileStatus to an array of Path.
- state -
Variable in enum org.apache.hadoop.security.SaslRpcServer.SaslStatus
-
- statement -
Variable in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
-
- staticFlag -
Static variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- StaticUserWebFilter - Class in org.apache.hadoop.http.lib
- Provides a servlet filter that pretends to authenticate a fake user (Dr.Who)
so that the web UI is usable for a secure cluster without authentication.
- StaticUserWebFilter() -
Constructor for class org.apache.hadoop.http.lib.StaticUserWebFilter
-
- StaticUserWebFilter.StaticUserFilter - Class in org.apache.hadoop.http.lib
-
- StaticUserWebFilter.StaticUserFilter() -
Constructor for class org.apache.hadoop.http.lib.StaticUserWebFilter.StaticUserFilter
-
- statistics -
Variable in class org.apache.hadoop.fs.FileSystem
- The statistics for this file system.
- StatusReporter - Class in org.apache.hadoop.mapreduce
-
- StatusReporter() -
Constructor for class org.apache.hadoop.mapreduce.StatusReporter
-
- statusUpdate(TaskUmbilicalProtocol) -
Method in class org.apache.hadoop.mapred.Task
-
- statusUpdate(TaskAttemptID, TaskStatus, JvmContext) -
Method in class org.apache.hadoop.mapred.TaskTracker
- Called periodically to report Task progress, from 0.0 to 1.0.
- statusUpdate(TaskAttemptID, TaskStatus, JvmContext) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Report child's progress to parent.
- stddev() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- stop() -
Method in class org.apache.hadoop.http.HttpServer
- stop the server
- stop() -
Method in class org.apache.hadoop.ipc.Client
- Stop all threads related to this client.
- stop() -
Method in class org.apache.hadoop.ipc.Server
- Stops the service.
- stop() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
- set the thread state to STOPPING so that the
thread will stop when it wakes up.
- stop() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
- set the thread state to STOPPING so that the
thread will stop when it wakes up.
- stop() -
Method in class org.apache.hadoop.metrics2.impl.MetricsSystemImpl
-
- stop() -
Method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
-
- stop() -
Method in interface org.apache.hadoop.metrics2.MetricsSystemMXBean
- Stop the metrics system
- stop() -
Method in interface org.apache.hadoop.util.ServicePlugin
- This method is invoked when the service instance is about to be shut down.
- stopCleanupThread() -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
- Stop the background thread
- stopCommunicationThread() -
Method in class org.apache.hadoop.mapred.Task.TaskReporter
-
- stopMonitoring() -
Method in class org.apache.hadoop.metrics.file.FileContext
- Deprecated. Stops monitoring, closing the file.
- stopMonitoring() -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Stops monitoring.
- stopMonitoring() -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Stops monitoring.
- stopMonitoring() -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- stopNotifier() -
Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- stopProxy(VersionedProtocol) -
Static method in class org.apache.hadoop.ipc.RPC
- Stop this proxy and release its invoker's resource
- stopRunning() -
Method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager.CleanupThread
-
- stopThreads() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
-
- stopTracker() -
Method in class org.apache.hadoop.mapred.JobTracker
-
- store(Configuration, K, String) -
Static method in class org.apache.hadoop.io.DefaultStringifier
- Stores the item in the configuration with the given keyName.
- storeArray(Configuration, K[], String) -
Static method in class org.apache.hadoop.io.DefaultStringifier
- Stores the array of items in the configuration with the given keyName.
- storeBlock(Block, File) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- storeINode(Path, INode) -
Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- StreamBackedIterator<X extends Writable> - Class in org.apache.hadoop.mapred.join
- This class provides an implementation of ResetableIterator.
- StreamBackedIterator() -
Constructor for class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- StreamBaseRecordReader - Class in org.apache.hadoop.streaming
- Shared functionality for hadoopStreaming formats.
- StreamBaseRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) -
Constructor for class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- StreamInputFormat - Class in org.apache.hadoop.streaming
- An input format that selects a RecordReader based on a JobConf property.
- StreamInputFormat() -
Constructor for class org.apache.hadoop.streaming.StreamInputFormat
-
- StreamJob - Class in org.apache.hadoop.streaming
- All the client-side work happens here.
- StreamJob(String[], boolean) -
Constructor for class org.apache.hadoop.streaming.StreamJob
- Deprecated. use StreamJob() with ToolRunner or set the
Configuration using
StreamJob.setConf(Configuration)
and
run with StreamJob.run(String[])
.
- StreamJob() -
Constructor for class org.apache.hadoop.streaming.StreamJob
-
- StreamKeyValUtil - Class in org.apache.hadoop.streaming
-
- StreamKeyValUtil() -
Constructor for class org.apache.hadoop.streaming.StreamKeyValUtil
-
- StreamUtil - Class in org.apache.hadoop.streaming
- Utilities not available elsewhere in Hadoop.
- StreamUtil() -
Constructor for class org.apache.hadoop.streaming.StreamUtil
-
- StreamXmlRecordReader - Class in org.apache.hadoop.streaming
- A way to interpret XML fragments as Mapper input records.
- StreamXmlRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) -
Constructor for class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- STRING -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- string2long(String) -
Static method in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
- Convert a string to long.
- STRING_VALUE_MAX -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- STRING_VALUE_MIN -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- Stringifier<T> - Interface in org.apache.hadoop.io
- Stringifier interface offers two methods to convert an object
to a string representation and restore the object given its
string representation.
- stringifyException(Throwable) -
Static method in class org.apache.hadoop.util.StringUtils
- Make a string representation of the exception.
- stringifySolution(int, int, List<List<Pentomino.ColumnName>>) -
Static method in class org.apache.hadoop.examples.dancing.Pentomino
- Convert a solution to the puzzle returned by the model into a string
that represents the placement of the pieces onto the board.
- stringToPath(String[]) -
Static method in class org.apache.hadoop.util.StringUtils
-
- stringToURI(String[]) -
Static method in class org.apache.hadoop.util.StringUtils
-
- StringTypeID -
Static variable in class org.apache.hadoop.record.meta.TypeID
-
- StringUtils - Class in org.apache.hadoop.util
- General string utils
- StringUtils() -
Constructor for class org.apache.hadoop.util.StringUtils
-
- StringUtils.TraditionalBinaryPrefix - Enum in org.apache.hadoop.util
- The traditional binary prefixes, kilo, mega, ..., exa,
which can be represented by a 64-bit integer.
- StringValueMax - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that maintain the biggest of
a sequence of strings.
- StringValueMax() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
- the default constructor
- StringValueMin - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that maintain the smallest of
a sequence of strings.
- StringValueMin() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
- the default constructor
- STRUCT -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- StructTypeID - Class in org.apache.hadoop.record.meta
- Represents typeID for a struct
- StructTypeID(RecordTypeInfo) -
Constructor for class org.apache.hadoop.record.meta.StructTypeID
- Create a StructTypeID based on the RecordTypeInfo of some record
- SUBDIR -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- subMap(WritableComparable, WritableComparable) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- submit() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
- Submit this job to mapred.
- submit() -
Method in class org.apache.hadoop.mapreduce.Job
- Submit the job to the cluster and return immediately.
- submit() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
- Submit this job to mapred.
- submitAndMonitorJob() -
Method in class org.apache.hadoop.streaming.StreamJob
-
- submitJob(String) -
Method in class org.apache.hadoop.mapred.JobClient
- Submit a job to the MR system.
- submitJob(JobConf) -
Method in class org.apache.hadoop.mapred.JobClient
- Submit a job to the MR system.
- submitJob(JobID, String, Credentials) -
Method in class org.apache.hadoop.mapred.JobTracker
- JobTracker.submitJob() kicks off a new job.
- submitJob(JobID, String, Credentials) -
Method in class org.apache.hadoop.mapred.LocalJobRunner
-
- submitJob(JobConf) -
Static method in class org.apache.hadoop.mapred.pipes.Submitter
- Deprecated. Use
Submitter.runJob(JobConf)
- submitJobInternal(JobConf) -
Method in class org.apache.hadoop.mapred.JobClient
- Internal method for submitting jobs to the system.
- submitReadahead(String, FileDescriptor, long, long) -
Method in class org.apache.hadoop.io.ReadaheadPool
- Submit a request to readahead on the given file descriptor.
- Submitter - Class in org.apache.hadoop.mapred.pipes
- The main entry point and job submitter.
- Submitter() -
Constructor for class org.apache.hadoop.mapred.pipes.Submitter
-
- Submitter(Configuration) -
Constructor for class org.apache.hadoop.mapred.pipes.Submitter
-
- SUBSTITUTE_TOKEN -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
- If users are providing their own query, the following string is expected to
appear in the WHERE clause, which will be substituted with a pair of conditions
on the input to allow input splits to parallelise the import.
- SUCCEEDED -
Static variable in class org.apache.hadoop.mapred.JobStatus
-
- SUCCEEDED_FILE_NAME -
Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
-
- SUCCEEDED_FILE_NAME -
Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
- SUCCESS -
Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- success() -
Method in class org.apache.hadoop.mapred.ShuffleExceptionTracker
- Mark the request as success.
- Sudoku - Class in org.apache.hadoop.examples.dancing
- This class uses the dancing links algorithm from Knuth to solve sudoku
puzzles.
- Sudoku(InputStream) -
Constructor for class org.apache.hadoop.examples.dancing.Sudoku
- Set up a puzzle board to the given size.
- Sudoku.ColumnName - Interface in org.apache.hadoop.examples.dancing
- This interface is a marker class for the columns created for the
Sudoku solver.
- suffix(String) -
Method in class org.apache.hadoop.fs.Path
- Adds a suffix to the final name in the path.
- sum(Counters, Counters) -
Static method in class org.apache.hadoop.mapred.Counters
- Convenience method for computing the sum of two sets of counters.
- SUPPORT_SPARSE_METRICS_DEFAULT -
Static variable in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- SUPPORT_SPARSE_METRICS_PROPERTY -
Static variable in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
-
- supportIsolationRunner(JobConf) -
Method in class org.apache.hadoop.mapred.Task
-
- suspend() -
Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
- suspend the running thread
- suspend() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
- suspend the running thread
- swap(Node[], int, int) -
Static method in class org.apache.hadoop.net.NetworkTopology
-
- swap(int, int) -
Method in interface org.apache.hadoop.util.IndexedSortable
- Swap items at the given addresses.
- SWITCH_TO_SIMPLE_AUTH -
Static variable in class org.apache.hadoop.security.SaslRpcServer
-
- SwitchTo(int) -
Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- SYMBOL -
Variable in enum org.apache.hadoop.fs.permission.FsAction
- Symbolic representation
- symbol -
Variable in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
-
- symLink(String, String) -
Static method in class org.apache.hadoop.fs.FileUtil
- Create a soft link between a src and destination
only on a local disk.
- sync() -
Method in class org.apache.hadoop.fs.FSDataOutputStream
- Synchronize all buffer with the underlying devices.
- sync() -
Method in interface org.apache.hadoop.fs.Syncable
- Synchronize all buffer with the underlying devices.
- sync(long) -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Seek to the next sync mark past a given position.
- sync() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
- create a sync point
- SYNC_FILE_RANGE_WAIT_AFTER -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- SYNC_FILE_RANGE_WAIT_BEFORE -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- SYNC_FILE_RANGE_WRITE -
Static variable in class org.apache.hadoop.io.nativeio.NativeIO
-
- SYNC_INTERVAL -
Static variable in class org.apache.hadoop.io.SequenceFile
- The number of bytes between sync points.
- Syncable - Interface in org.apache.hadoop.fs
- This interface declare the sync() operation.
- syncFileRangeIfPossible(FileDescriptor, long, long, int) -
Static method in class org.apache.hadoop.io.nativeio.NativeIO
- Call sync_file_range on the given file descriptor.
- syncFs() -
Method in class org.apache.hadoop.io.SequenceFile.Writer
- flush all currently written data to the file system
- syncLogs(String, TaskAttemptID, boolean, boolean) -
Static method in class org.apache.hadoop.mapred.TaskLog
-
- syncSeen() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Returns true iff the previous call to next passed a sync mark.
- SystemLogParser - Class in org.apache.hadoop.contrib.failmon
- An object of this class parses a Unix system log file to create
appropriate EventRecords.
- SystemLogParser(String) -
Constructor for class org.apache.hadoop.contrib.failmon.SystemLogParser
- Create a new parser object .
T
- tabSize -
Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- tag -
Variable in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- tag(String, String, String) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Add a tag to the metrics
- tag(String, String, String, boolean) -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Add a tag to the metrics
- tag(String, String, String) -
Method in class org.apache.hadoop.metrics2.MetricsRecordBuilder
- Add a metrics tag
- TAG -
Static variable in class org.apache.hadoop.record.compiler.Consts
-
- TaggedMapOutput - Class in org.apache.hadoop.contrib.utils.join
- This abstract class serves as the base class for the values that
flow from the mappers to the reducers in a data join job.
- TaggedMapOutput() -
Constructor for class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- tags() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- tags() -
Method in class org.apache.hadoop.metrics2.lib.MetricsRegistry
- Get the tags
- tags() -
Method in interface org.apache.hadoop.metrics2.MetricsRecord
- Get the tags of the record
- tailMap(WritableComparable) -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- Task - Class in org.apache.hadoop.mapred
- Base class for tasks.
- Task() -
Constructor for class org.apache.hadoop.mapred.Task
-
- Task(String, TaskAttemptID, int, int) -
Constructor for class org.apache.hadoop.mapred.Task
-
- TASK -
Static variable in class org.apache.hadoop.mapreduce.TaskID
-
- Task.CombineOutputCollector<K,V> - Class in org.apache.hadoop.mapred
- OutputCollector for the combiner.
- Task.CombineOutputCollector(Counters.Counter, Progressable, Configuration) -
Constructor for class org.apache.hadoop.mapred.Task.CombineOutputCollector
-
- Task.CombinerRunner<K,V> - Class in org.apache.hadoop.mapred
-
- Task.CombineValuesIterator<KEY,VALUE> - Class in org.apache.hadoop.mapred
-
- Task.CombineValuesIterator(RawKeyValueIterator, RawComparator<KEY>, Class<KEY>, Class<VALUE>, Configuration, Reporter, Counters.Counter) -
Constructor for class org.apache.hadoop.mapred.Task.CombineValuesIterator
-
- Task.Counter - Enum in org.apache.hadoop.mapred
-
- Task.NewCombinerRunner<K,V> - Class in org.apache.hadoop.mapred
-
- Task.OldCombinerRunner<K,V> - Class in org.apache.hadoop.mapred
-
- Task.OldCombinerRunner(Class<? extends Reducer<K, V, K, V>>, JobConf, Counters.Counter, Task.TaskReporter) -
Constructor for class org.apache.hadoop.mapred.Task.OldCombinerRunner
-
- Task.TaskReporter - Class in org.apache.hadoop.mapred
-
- TASK_LAUNCH_SCRIPT_PERMISSION -
Static variable in class org.apache.hadoop.mapred.TaskController
-
- TaskAttemptContext - Class in org.apache.hadoop.mapred
-
- TaskAttemptContext - Class in org.apache.hadoop.mapreduce
- The context for task attempts.
- TaskAttemptContext(Configuration, TaskAttemptID) -
Constructor for class org.apache.hadoop.mapreduce.TaskAttemptContext
-
- TaskAttemptID - Class in org.apache.hadoop.mapred
- TaskAttemptID represents the immutable and unique identifier for
a task attempt.
- TaskAttemptID(TaskID, int) -
Constructor for class org.apache.hadoop.mapred.TaskAttemptID
- Constructs a TaskAttemptID object from given
TaskID
.
- TaskAttemptID(String, int, boolean, int, int) -
Constructor for class org.apache.hadoop.mapred.TaskAttemptID
- Constructs a TaskId object from given parts.
- TaskAttemptID() -
Constructor for class org.apache.hadoop.mapred.TaskAttemptID
-
- TaskAttemptID - Class in org.apache.hadoop.mapreduce
- TaskAttemptID represents the immutable and unique identifier for
a task attempt.
- TaskAttemptID(TaskID, int) -
Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
- Constructs a TaskAttemptID object from given
TaskID
.
- TaskAttemptID(String, int, boolean, int, int) -
Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
- Constructs a TaskId object from given parts.
- TaskAttemptID() -
Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
-
- taskCleanup -
Variable in class org.apache.hadoop.mapred.Task
-
- TaskCompletionEvent - Class in org.apache.hadoop.mapred
- This is used to track task completion events on
job tracker.
- TaskCompletionEvent() -
Constructor for class org.apache.hadoop.mapred.TaskCompletionEvent
- Default constructor for Writable.
- TaskCompletionEvent(int, TaskAttemptID, int, boolean, TaskCompletionEvent.Status, String) -
Constructor for class org.apache.hadoop.mapred.TaskCompletionEvent
- Constructor.
- TaskCompletionEvent.Status - Enum in org.apache.hadoop.mapred
-
- taskContext -
Variable in class org.apache.hadoop.mapred.Task
-
- TaskController - Class in org.apache.hadoop.mapred
- Controls initialization, finalization and clean up of tasks, and
also the launching and killing of task JVMs.
- TaskController() -
Constructor for class org.apache.hadoop.mapred.TaskController
-
- TaskDistributedCacheManager - Class in org.apache.hadoop.filecache
- Helper class of
TrackerDistributedCacheManager
that represents
the cached files of a single job. - taskFailedPing(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.TaskTrackerMetricsSource
-
- TaskGraphServlet - Class in org.apache.hadoop.mapred
- The servlet that outputs svg graphics for map / reduce task
statuses
- TaskGraphServlet() -
Constructor for class org.apache.hadoop.mapred.TaskGraphServlet
-
- TaskID - Class in org.apache.hadoop.mapred
- TaskID represents the immutable and unique identifier for
a Map or Reduce Task.
- TaskID(JobID, boolean, int) -
Constructor for class org.apache.hadoop.mapred.TaskID
- Constructs a TaskID object from given
JobID
.
- TaskID(String, int, boolean, int) -
Constructor for class org.apache.hadoop.mapred.TaskID
- Constructs a TaskInProgressId object from given parts.
- TaskID() -
Constructor for class org.apache.hadoop.mapred.TaskID
-
- TaskID - Class in org.apache.hadoop.mapreduce
- TaskID represents the immutable and unique identifier for
a Map or Reduce Task.
- TaskID(JobID, boolean, int) -
Constructor for class org.apache.hadoop.mapreduce.TaskID
- Constructs a TaskID object from given
JobID
.
- TaskID(String, int, boolean, int) -
Constructor for class org.apache.hadoop.mapreduce.TaskID
- Constructs a TaskInProgressId object from given parts.
- TaskID() -
Constructor for class org.apache.hadoop.mapreduce.TaskID
-
- TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
- A context object that allows input and output from the task.
- TaskInputOutputContext(Configuration, TaskAttemptID, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter) -
Constructor for class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- TaskLog - Class in org.apache.hadoop.mapred
- A simple logger to handle the task-specific user logs.
- TaskLog() -
Constructor for class org.apache.hadoop.mapred.TaskLog
-
- TaskLog.LogName - Enum in org.apache.hadoop.mapred
- The filter for userlogs.
- TaskLogAppender - Class in org.apache.hadoop.mapred
- A simple log4j-appender for the task child's
map-reduce system logs.
- TaskLogAppender() -
Constructor for class org.apache.hadoop.mapred.TaskLogAppender
-
- TaskLogServlet - Class in org.apache.hadoop.mapred
- A servlet that is run by the TaskTrackers to provide the task logs via http.
- TaskLogServlet() -
Constructor for class org.apache.hadoop.mapred.TaskLogServlet
-
- TaskLogsTruncater - Class in org.apache.hadoop.mapred
- The class for truncating the user logs.
- TaskLogsTruncater(Configuration) -
Constructor for class org.apache.hadoop.mapred.TaskLogsTruncater
-
- TaskReport - Class in org.apache.hadoop.mapred
- A report on the state of a task.
- TaskReport() -
Constructor for class org.apache.hadoop.mapred.TaskReport
-
- TaskStatus - Class in org.apache.hadoop.mapred
- Describes the current status of a task.
- TaskStatus() -
Constructor for class org.apache.hadoop.mapred.TaskStatus
-
- TaskStatus(TaskAttemptID, float, int, TaskStatus.State, String, String, String, TaskStatus.Phase, Counters) -
Constructor for class org.apache.hadoop.mapred.TaskStatus
-
- TaskStatus.Phase - Enum in org.apache.hadoop.mapred
-
- TaskStatus.State - Enum in org.apache.hadoop.mapred
-
- TaskTracker - Class in org.apache.hadoop.mapred
- TaskTracker is a process that starts and tracks MR Tasks
in a networked environment.
- TaskTracker(JobConf) -
Constructor for class org.apache.hadoop.mapred.TaskTracker
- Start with the local machine name, and the default JobTracker
- TaskTracker - Class in org.apache.hadoop.mapreduce.server.jobtracker
- The representation of a single
TaskTracker
as seen by
the JobTracker
. - TaskTracker(String) -
Constructor for class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
- Create a new
TaskTracker
.
- TaskTracker.MapOutputServlet - Class in org.apache.hadoop.mapred
- This class is used in TaskTracker's Jetty to serve the map outputs
to other nodes.
- TaskTracker.MapOutputServlet() -
Constructor for class org.apache.hadoop.mapred.TaskTracker.MapOutputServlet
-
- TaskTrackerMetricsSource - Class in org.apache.hadoop.mapred
- Instrumentation for metrics v2
- TaskTrackerMetricsSource(TaskTracker) -
Constructor for class org.apache.hadoop.mapred.TaskTrackerMetricsSource
-
- TaskTrackerMXBean - Interface in org.apache.hadoop.mapred
- MXBean interface for TaskTracker
- taskTrackerNames() -
Method in class org.apache.hadoop.mapred.JobTracker
- Get the active, blacklisted, and graylisted task tracker names in the
cluster.
- taskTrackers() -
Method in class org.apache.hadoop.mapred.JobTracker
- Get all the task trackers in the cluster
- TaskTrackerStatus - Class in org.apache.hadoop.mapred
- A TaskTrackerStatus is a MapReduce primitive.
- TaskTrackerStatus() -
Constructor for class org.apache.hadoop.mapred.TaskTrackerStatus
-
- TaskTrackerStatus(String, String, int, List<TaskStatus>, int, int, int, int) -
Constructor for class org.apache.hadoop.mapred.TaskTrackerStatus
-
- TaskType - Enum in org.apache.hadoop.mapreduce
- Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.
- TaskUmbilicalProtocol - Interface in org.apache.hadoop.mapred
- Protocol that task child process uses to contact its parent process.
- TEMP_DIR_NAME -
Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
- Temporary directory name
- TEMP_DIR_NAME -
Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
- Temporary directory name
- TeraGen - Class in org.apache.hadoop.examples.terasort
- Generate the official terasort input data set.
- TeraGen() -
Constructor for class org.apache.hadoop.examples.terasort.TeraGen
-
- TeraGen.SortGenMapper - Class in org.apache.hadoop.examples.terasort
- The Mapper class that given a row number, will generate the appropriate
output line.
- TeraGen.SortGenMapper() -
Constructor for class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
-
- TeraInputFormat - Class in org.apache.hadoop.examples.terasort
- An input format that reads the first 10 characters of each line as the key
and the rest of the line as the value.
- TeraInputFormat() -
Constructor for class org.apache.hadoop.examples.terasort.TeraInputFormat
-
- TeraOutputFormat - Class in org.apache.hadoop.examples.terasort
- A streamlined text output format that writes key, value, and "\r\n".
- TeraOutputFormat() -
Constructor for class org.apache.hadoop.examples.terasort.TeraOutputFormat
-
- TeraSort - Class in org.apache.hadoop.examples.terasort
- Generates the sampled split points, launches the job, and waits for it to
finish.
- TeraSort() -
Constructor for class org.apache.hadoop.examples.terasort.TeraSort
-
- TeraValidate - Class in org.apache.hadoop.examples.terasort
- Generate 1 mapper per a file that checks to make sure the keys
are sorted within each file.
- TeraValidate() -
Constructor for class org.apache.hadoop.examples.terasort.TeraValidate
-
- Text - Class in org.apache.hadoop.io
- This class stores text using standard UTF8 encoding.
- Text() -
Constructor for class org.apache.hadoop.io.Text
-
- Text(String) -
Constructor for class org.apache.hadoop.io.Text
- Construct from a string.
- Text(Text) -
Constructor for class org.apache.hadoop.io.Text
- Construct from another text.
- Text(byte[]) -
Constructor for class org.apache.hadoop.io.Text
- Construct from a byte array.
- Text.Comparator - Class in org.apache.hadoop.io
- A WritableComparator optimized for Text keys.
- Text.Comparator() -
Constructor for class org.apache.hadoop.io.Text.Comparator
-
- TEXT_ID -
Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
-
- TextInputFormat - Class in org.apache.hadoop.mapred
- An
InputFormat
for plain text files. - TextInputFormat() -
Constructor for class org.apache.hadoop.mapred.TextInputFormat
-
- TextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
- An
InputFormat
for plain text files. - TextInputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
-
- TextInputWriter - Class in org.apache.hadoop.streaming.io
- InputWriter that writes the client's input as text.
- TextInputWriter() -
Constructor for class org.apache.hadoop.streaming.io.TextInputWriter
-
- TextOutputFormat<K,V> - Class in org.apache.hadoop.mapred
- An
OutputFormat
that writes plain text files. - TextOutputFormat() -
Constructor for class org.apache.hadoop.mapred.TextOutputFormat
-
- TextOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
- An
OutputFormat
that writes plain text files. - TextOutputFormat() -
Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
-
- TextOutputFormat.LineRecordWriter<K,V> - Class in org.apache.hadoop.mapred
-
- TextOutputFormat.LineRecordWriter(DataOutputStream, String) -
Constructor for class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
-
- TextOutputFormat.LineRecordWriter(DataOutputStream) -
Constructor for class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
-
- TextOutputFormat.LineRecordWriter<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
-
- TextOutputFormat.LineRecordWriter(DataOutputStream, String) -
Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
-
- TextOutputFormat.LineRecordWriter(DataOutputStream) -
Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
-
- TextOutputReader - Class in org.apache.hadoop.streaming.io
- OutputReader that reads the client's output as text.
- TextOutputReader() -
Constructor for class org.apache.hadoop.streaming.io.TextOutputReader
-
- TextSplitter - Class in org.apache.hadoop.mapreduce.lib.db
- Implement DBSplitter over text strings.
- TextSplitter() -
Constructor for class org.apache.hadoop.mapreduce.lib.db.TextSplitter
-
- TFile - Class in org.apache.hadoop.io.file.tfile
- A TFile is a container of key-value pairs.
- TFile.Reader - Class in org.apache.hadoop.io.file.tfile
- TFile Reader.
- TFile.Reader(FSDataInputStream, long, Configuration) -
Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader
- Constructor
- TFile.Reader.Scanner - Class in org.apache.hadoop.io.file.tfile
- The TFile Scanner.
- TFile.Reader.Scanner(TFile.Reader, long, long) -
Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Constructor
- TFile.Reader.Scanner(TFile.Reader, RawComparable, RawComparable) -
Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Constructor
- TFile.Reader.Scanner.Entry - Class in org.apache.hadoop.io.file.tfile
- Entry to a <Key, Value> pair.
- TFile.Reader.Scanner.Entry() -
Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
- TFile.Writer - Class in org.apache.hadoop.io.file.tfile
- TFile Writer.
- TFile.Writer(FSDataOutputStream, int, String, String, Configuration) -
Constructor for class org.apache.hadoop.io.file.tfile.TFile.Writer
- Constructor
- timedoutTask(TaskAttemptID) -
Method in class org.apache.hadoop.mapred.TaskTrackerMetricsSource
-
- timeOutInterval -
Variable in class org.apache.hadoop.util.Shell
- Time after which the executing script would be timedout
- timestamp() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- timestamp() -
Method in interface org.apache.hadoop.metrics2.MetricsRecord
- Get the timestamp of the metrics
- TIPStatus - Enum in org.apache.hadoop.mapred
- The states of a
TaskInProgress
as seen by the JobTracker. - toArray() -
Method in class org.apache.hadoop.io.ArrayWritable
-
- toArray() -
Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- toArray(Class<T>, List<T>) -
Static method in class org.apache.hadoop.util.GenericsUtil
- Converts the given
List<T>
to a an array of
T[]
.
- toArray(List<T>) -
Static method in class org.apache.hadoop.util.GenericsUtil
- Converts the given
List<T>
to a an array of
T[]
.
- toByteArray(Writable...) -
Static method in class org.apache.hadoop.io.WritableUtils
- Convert writables to a byte array
- toHex(byte[]) -
Static method in class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
- byte array to Hex String
- token -
Variable in class org.apache.hadoop.record.compiler.generated.Rcc
-
- Token - Class in org.apache.hadoop.record.compiler.generated
- Describes the input token stream.
- Token() -
Constructor for class org.apache.hadoop.record.compiler.generated.Token
-
- Token<T extends TokenIdentifier> - Class in org.apache.hadoop.security.token
- The client-side form of the token.
- Token(T, SecretManager<T>) -
Constructor for class org.apache.hadoop.security.token.Token
- Construct a token given a token identifier and a secret manager for the
type of the token identifier.
- Token(byte[], byte[], Text, Text) -
Constructor for class org.apache.hadoop.security.token.Token
- Construct a token from the components.
- Token() -
Constructor for class org.apache.hadoop.security.token.Token
- Default constructor
- Token(Token<T>) -
Constructor for class org.apache.hadoop.security.token.Token
- Clone a token.
- Token.TrivialRenewer - Class in org.apache.hadoop.security.token
- A trivial renewer for token kinds that aren't managed.
- Token.TrivialRenewer() -
Constructor for class org.apache.hadoop.security.token.Token.TrivialRenewer
-
- token_source -
Variable in class org.apache.hadoop.record.compiler.generated.Rcc
-
- TokenCache - Class in org.apache.hadoop.mapreduce.security
- This class provides user facing APIs for transferring secrets from
the job client to the tasks.
- TokenCache() -
Constructor for class org.apache.hadoop.mapreduce.security.TokenCache
-
- TokenCounterMapper - Class in org.apache.hadoop.mapreduce.lib.map
- Tokenize the input values and emit each word with a count of 1.
- TokenCounterMapper() -
Constructor for class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper
-
- TokenCountMapper<K> - Class in org.apache.hadoop.mapred.lib
- A
Mapper
that maps text values into pairs. - TokenCountMapper() -
Constructor for class org.apache.hadoop.mapred.lib.TokenCountMapper
-
- TokenIdentifier - Class in org.apache.hadoop.security.token
- An identifier that identifies a token, may contain public information
about a token, including its kind (or type).
- TokenIdentifier() -
Constructor for class org.apache.hadoop.security.token.TokenIdentifier
-
- tokenImage -
Variable in exception org.apache.hadoop.record.compiler.generated.ParseException
- This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
- tokenImage -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- TokenInfo - Annotation Type in org.apache.hadoop.security.token
- Indicates Token related information to be used
- TokenMgrError - Error in org.apache.hadoop.record.compiler.generated
-
- TokenMgrError() -
Constructor for error org.apache.hadoop.record.compiler.generated.TokenMgrError
-
- TokenMgrError(String, int) -
Constructor for error org.apache.hadoop.record.compiler.generated.TokenMgrError
-
- TokenMgrError(boolean, int, int, int, String, char, int) -
Constructor for error org.apache.hadoop.record.compiler.generated.TokenMgrError
-
- TokenRenewer - Class in org.apache.hadoop.security.token
- This is the interface for plugins that handle tokens.
- TokenRenewer() -
Constructor for class org.apache.hadoop.security.token.TokenRenewer
-
- tokenSecret -
Variable in class org.apache.hadoop.mapred.Task
-
- TokenSelector<T extends TokenIdentifier> - Interface in org.apache.hadoop.security.token
- Select token of type T from tokens for use with named service
- tokenStorage -
Variable in class org.apache.hadoop.mapred.JobInProgress
-
- toMap() -
Method in class org.apache.hadoop.streaming.Environment
-
- Tool - Interface in org.apache.hadoop.util
- A tool interface that supports handling of generic command-line options.
- ToolRunner - Class in org.apache.hadoop.util
- A utility to help run
Tool
s. - ToolRunner() -
Constructor for class org.apache.hadoop.util.ToolRunner
-
- top() -
Method in class org.apache.hadoop.util.PriorityQueue
- Returns the least element of the PriorityQueue in constant time.
- toShort() -
Method in class org.apache.hadoop.fs.permission.FsPermission
- Encode the object to a short.
- toString() -
Method in class org.apache.hadoop.conf.Configuration.IntegerRanges
-
- toString() -
Method in class org.apache.hadoop.conf.Configuration
-
- toString() -
Method in class org.apache.hadoop.contrib.failmon.EventRecord
- Creates and returns a string representation of the object.
- toString() -
Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
- Creates and returns a string reperssentation of the object
- toString() -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
- toString() -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- toString() -
Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
-
- toString() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
-
- toString() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
- toString() -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- toString() -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
- toString() -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- toString() -
Method in class org.apache.hadoop.fs.BlockLocation
-
- toString() -
Method in class org.apache.hadoop.fs.ContentSummary
-
- toString(boolean) -
Method in class org.apache.hadoop.fs.ContentSummary
- Return the string representation of the object in the output format.
- toString() -
Method in class org.apache.hadoop.fs.DF
-
- toString() -
Method in class org.apache.hadoop.fs.DU
-
- toString() -
Method in class org.apache.hadoop.fs.FileSystem.Statistics
-
- toString() -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
- toString() -
Method in class org.apache.hadoop.fs.Path
-
- toString() -
Method in class org.apache.hadoop.fs.permission.FsPermission
-
- toString() -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
-
- toString() -
Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- toString() -
Method in class org.apache.hadoop.fs.s3.Block
-
- toString() -
Method in class org.apache.hadoop.io.BooleanWritable
-
- toString() -
Method in class org.apache.hadoop.io.BytesWritable
- Generate the stream of bytes as hex pairs separated by ' '.
- toString() -
Method in class org.apache.hadoop.io.ByteWritable
-
- toString() -
Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
- Print the extension map out as a string.
- toString(T) -
Method in class org.apache.hadoop.io.DefaultStringifier
-
- toString() -
Method in class org.apache.hadoop.io.DoubleWritable
-
- toString() -
Method in class org.apache.hadoop.io.file.tfile.Utils.Version
- Return a string representation of the version.
- toString() -
Method in class org.apache.hadoop.io.FloatWritable
-
- toString() -
Method in class org.apache.hadoop.io.GenericWritable
-
- toString() -
Method in class org.apache.hadoop.io.IntWritable
-
- toString() -
Method in class org.apache.hadoop.io.LongWritable
-
- toString() -
Method in class org.apache.hadoop.io.MD5Hash
- Returns a string representation of this object.
- toString() -
Method in class org.apache.hadoop.io.nativeio.NativeIO.Stat
-
- toString() -
Method in exception org.apache.hadoop.io.nativeio.NativeIOException
-
- toString() -
Method in class org.apache.hadoop.io.NullWritable
-
- toString() -
Method in class org.apache.hadoop.io.ObjectWritable
-
- toString() -
Method in class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry.Pair
-
- toString() -
Method in class org.apache.hadoop.io.retry.RetryPolicies.MultipleLinearRandomRetry
-
- toString() -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- toString() -
Method in class org.apache.hadoop.io.SequenceFile.Reader
- Returns the name of the file.
- toString(T) -
Method in interface org.apache.hadoop.io.Stringifier
- Converts the object to a string representation
- toString() -
Method in class org.apache.hadoop.io.Text
- Convert text back to string
- toString() -
Method in class org.apache.hadoop.io.UTF8
- Deprecated. Convert to a String.
- toString() -
Method in exception org.apache.hadoop.io.VersionMismatchException
- Returns a string representation of this object.
- toString() -
Method in class org.apache.hadoop.io.VIntWritable
-
- toString() -
Method in class org.apache.hadoop.io.VLongWritable
-
- toString() -
Method in class org.apache.hadoop.ipc.Server.Connection
-
- toString() -
Method in class org.apache.hadoop.mapred.Counters
- Return textual representation of the counter values.
- toString() -
Method in class org.apache.hadoop.mapred.FileSplit
-
- toString() -
Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- toString() -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- Convert Tuple to String as in the following.
- toString() -
Method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
-
- toString() -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- toString() -
Method in class org.apache.hadoop.mapred.MultiFileSplit
- Deprecated.
- toString() -
Method in class org.apache.hadoop.mapred.Task
-
- toString() -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- toString() -
Method in enum org.apache.hadoop.mapred.TaskLog.LogName
-
- toString() -
Method in class org.apache.hadoop.mapreduce.Counters
- Return textual representation of the counter values.
- toString() -
Method in class org.apache.hadoop.mapreduce.ID
-
- toString() -
Method in class org.apache.hadoop.mapreduce.JobID
-
- toString() -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
-
- toString() -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
- toString() -
Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
-
- toString() -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- toString() -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
- toString() -
Method in class org.apache.hadoop.mapreduce.TaskID
-
- toString() -
Method in class org.apache.hadoop.metrics2.impl.MetricsRecordImpl
-
- toString() -
Method in class org.apache.hadoop.metrics2.Metric
-
- toString() -
Method in class org.apache.hadoop.metrics2.MetricsTag
-
- toString() -
Method in class org.apache.hadoop.net.NetworkTopology
- convert a network tree to a string
- toString() -
Method in class org.apache.hadoop.net.NodeBase
- Return this node's string representation
- toString() -
Method in class org.apache.hadoop.record.Buffer
-
- toString(String) -
Method in class org.apache.hadoop.record.Buffer
- Convert the byte buffer to a string an specific character encoding
- toString() -
Method in class org.apache.hadoop.record.compiler.CodeBuffer
-
- toString() -
Method in class org.apache.hadoop.record.compiler.generated.Token
- Returns the image.
- toString() -
Method in class org.apache.hadoop.record.Record
-
- toString() -
Method in class org.apache.hadoop.security.authentication.client.AuthenticatedURL.Token
- Returns the string representation of the token.
- toString() -
Method in class org.apache.hadoop.security.authentication.server.AuthenticationToken
- Returns the string representation of the token.
- toString() -
Method in class org.apache.hadoop.security.authorize.AccessControlList
-
- toString() -
Method in class org.apache.hadoop.security.KerberosName
- Put the name back together from the parts.
- toString() -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- toString() -
Method in class org.apache.hadoop.security.token.Token
-
- toString() -
Method in class org.apache.hadoop.security.UserGroupInformation
- Return the username.
- toString() -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
- Generate a suitable string representation.
- toString() -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- toString() -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- toString() -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- toString() -
Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
- Returns a string printing PIDs of process present in the
ProcfsBasedProcessTree.
- toString() -
Method in class org.apache.hadoop.util.Progress
-
- toString() -
Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
- Returns the commands of this instance.
- toStringChecked() -
Method in class org.apache.hadoop.io.UTF8
- Deprecated. Convert to a string, checking for valid UTF8.
- toStrings() -
Method in class org.apache.hadoop.io.ArrayWritable
-
- TotalOrderPartitioner<K extends WritableComparable,V> - Class in org.apache.hadoop.mapred.lib
- Partitioner effecting a total order by reading split points from
an externally generated source.
- TotalOrderPartitioner() -
Constructor for class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
-
- TotalOrderPartitioner<K extends WritableComparable<?>,V> - Class in org.apache.hadoop.mapreduce.lib.partition
- Partitioner effecting a total order by reading split points from
an externally generated source.
- TotalOrderPartitioner() -
Constructor for class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
-
- touch(File) -
Static method in class org.apache.hadoop.streaming.StreamUtil
-
- touchFile(String) -
Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- toUri() -
Method in class org.apache.hadoop.fs.Path
- Convert this to a URI.
- toWriteSkipRecs() -
Method in class org.apache.hadoop.mapred.Task
- Get whether to write skip records.
- TrackerDistributedCacheManager - Class in org.apache.hadoop.filecache
- Manages a single machine's instance of a cross-job
cache.
- TrackerDistributedCacheManager(Configuration, TaskController) -
Constructor for class org.apache.hadoop.filecache.TrackerDistributedCacheManager
-
- TrackerDistributedCacheManager.BaseDirManager - Class in org.apache.hadoop.filecache
- This class holds properties of each base directories and is responsible
for clean up unused cache files in base directories.
- TrackerDistributedCacheManager.BaseDirManager() -
Constructor for class org.apache.hadoop.filecache.TrackerDistributedCacheManager.BaseDirManager
-
- TrackerDistributedCacheManager.CleanupThread - Class in org.apache.hadoop.filecache
- A thread to check and cleanup the unused files periodically
- TrackerDistributedCacheManager.CleanupThread(Configuration) -
Constructor for class org.apache.hadoop.filecache.TrackerDistributedCacheManager.CleanupThread
-
- transferToFully(FileChannel, long, int) -
Method in class org.apache.hadoop.net.SocketOutputStream
- Transfers data from FileChannel using
FileChannel.transferTo(long, long, WritableByteChannel)
.
- transform(InputStream, InputStream, Writer) -
Static method in class org.apache.hadoop.util.XMLUtils
- Transform input xml given a stylesheet.
- Trash - Class in org.apache.hadoop.fs
- Provides a trash feature.
- Trash(Configuration) -
Constructor for class org.apache.hadoop.fs.Trash
- Construct a trash can accessor.
- Trash(FileSystem, Configuration) -
Constructor for class org.apache.hadoop.fs.Trash
- Construct a trash can accessor for the FileSystem provided.
- truncate() -
Method in class org.apache.hadoop.record.Buffer
- Change the capacity of the backing store to be the same as the current
count of buffer.
- truncateLogs(JVMInfo) -
Method in class org.apache.hadoop.mapred.TaskLogsTruncater
- Process the removed task's logs.
- truncateLogsAsUser(String, List<Task>) -
Method in class org.apache.hadoop.mapred.DefaultTaskController
-
- truncateLogsAsUser(String, List<Task>) -
Method in class org.apache.hadoop.mapred.TaskController
- Run the passed command as the user
- TRY_ONCE_DONT_FAIL -
Static variable in class org.apache.hadoop.io.retry.RetryPolicies
-
Try once, and fail silently for
void
methods, or by
re-throwing the exception for non-void
methods.
- TRY_ONCE_THEN_FAIL -
Static variable in class org.apache.hadoop.io.retry.RetryPolicies
-
Try once, and fail by re-throwing the exception.
- tryDivide(BigDecimal, BigDecimal) -
Method in class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter
- Divide numerator by denominator.
- TryIterator<T> - Class in org.apache.hadoop.metrics2.util
- A base class for unmodifiable iterators (throws on remove)
This class also makes writing filtering iterators easier, where the only
way to discover the end of data is by trying to read it.
- TryIterator() -
Constructor for class org.apache.hadoop.metrics2.util.TryIterator
-
- tryNext() -
Method in class org.apache.hadoop.metrics2.util.TryIterator
- Return the next element.
- TT_KEYTAB_FILE -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- TT_LOG_TMP_DIR -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- TT_RESOURCE_CALCULATOR_PLUGIN -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- TT_USER_NAME -
Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- ttConf -
Variable in class org.apache.hadoop.mapred.JobLocalizer
-
- TupleWritable - Class in org.apache.hadoop.mapred.join
- Writable type storing multiple
Writable
s. - TupleWritable() -
Constructor for class org.apache.hadoop.mapred.join.TupleWritable
- Create an empty tuple with no allocated storage for writables.
- TupleWritable(Writable[]) -
Constructor for class org.apache.hadoop.mapred.join.TupleWritable
- Initialize tuple with storage; unknown whether any of them contain
"written" values.
- TwoDArrayWritable - Class in org.apache.hadoop.io
- A Writable for 2D arrays containing a matrix of instances of a class.
- TwoDArrayWritable(Class) -
Constructor for class org.apache.hadoop.io.TwoDArrayWritable
-
- TwoDArrayWritable(Class, Writable[][]) -
Constructor for class org.apache.hadoop.io.TwoDArrayWritable
-
- twoRotations -
Static variable in class org.apache.hadoop.examples.dancing.Pentomino
- Is the piece identical if rotated 180 degrees?
- Type() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- TYPE -
Static variable in class org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler
- Constant that identifies the authentication mechanism.
- TYPE -
Static variable in class org.apache.hadoop.security.authentication.server.PseudoAuthenticationHandler
- Constant that identifies the authentication mechanism.
- Type - Enum in org.apache.hadoop.typedbytes
- The possible type codes.
- TYPE_SEPARATOR -
Static variable in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
-
- TYPED_BYTES_ID -
Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
-
- TypedBytesInput - Class in org.apache.hadoop.typedbytes
- Provides functionality for reading typed bytes.
- TypedBytesInput(DataInput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesInput
- Creates a new instance of TypedBytesInput.
- TypedBytesInputWriter - Class in org.apache.hadoop.streaming.io
- InputWriter that writes the client's input as typed bytes.
- TypedBytesInputWriter() -
Constructor for class org.apache.hadoop.streaming.io.TypedBytesInputWriter
-
- TypedBytesOutput - Class in org.apache.hadoop.typedbytes
- Provides functionality for writing typed bytes.
- TypedBytesOutput(DataOutput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesOutput
- Creates a new instance of TypedBytesOutput.
- TypedBytesOutputReader - Class in org.apache.hadoop.streaming.io
- OutputReader that reads the client's output as typed bytes.
- TypedBytesOutputReader() -
Constructor for class org.apache.hadoop.streaming.io.TypedBytesOutputReader
-
- TypedBytesRecordInput - Class in org.apache.hadoop.typedbytes
- Serializer for records that writes typed bytes.
- TypedBytesRecordInput(TypedBytesInput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordInput
- Creates a new instance of TypedBytesRecordInput.
- TypedBytesRecordInput(DataInput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordInput
- Creates a new instance of TypedBytesRecordInput.
- TypedBytesRecordOutput - Class in org.apache.hadoop.typedbytes
- Deserialized for records that reads typed bytes.
- TypedBytesRecordOutput(TypedBytesOutput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
- Creates a new instance of TypedBytesRecordOutput.
- TypedBytesRecordOutput(DataOutput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
- Creates a new instance of TypedBytesRecordOutput.
- TypedBytesWritable - Class in org.apache.hadoop.typedbytes
- Writable for typed bytes.
- TypedBytesWritable() -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritable
- Create a TypedBytesWritable.
- TypedBytesWritable(byte[]) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritable
- Create a TypedBytesWritable with a given byte array as initial value.
- TypedBytesWritableInput - Class in org.apache.hadoop.typedbytes
- Provides functionality for reading typed bytes as Writable objects.
- TypedBytesWritableInput(TypedBytesInput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableInput
- Creates a new instance of TypedBytesWritableInput.
- TypedBytesWritableInput(DataInput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableInput
- Creates a new instance of TypedBytesWritableInput.
- TypedBytesWritableOutput - Class in org.apache.hadoop.typedbytes
- Provides functionality for writing Writable objects as typed bytes.
- TypedBytesWritableOutput(TypedBytesOutput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
- Creates a new instance of TypedBytesWritableOutput.
- TypedBytesWritableOutput(DataOutput) -
Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
- Creates a new instance of TypedBytesWritableOutput.
- TypeID - Class in org.apache.hadoop.record.meta
- Represents typeID for basic types.
- TypeID.RIOType - Class in org.apache.hadoop.record.meta
- constants representing the IDL types we support
- TypeID.RIOType() -
Constructor for class org.apache.hadoop.record.meta.TypeID.RIOType
-
- typeVal -
Variable in class org.apache.hadoop.record.meta.TypeID
-
U
- ugi -
Variable in class org.apache.hadoop.mapreduce.JobContext
- The UserGroupInformation object that has a reference to the current user
- ULIMIT_COMMAND -
Static variable in class org.apache.hadoop.util.Shell
- a Unix command to get ulimit of a process.
- UMASK_LABEL -
Static variable in class org.apache.hadoop.fs.permission.FsPermission
-
- umbilical -
Variable in class org.apache.hadoop.mapred.Task
-
- UNAVAILABLE -
Static variable in class org.apache.hadoop.mapred.TaskTrackerStatus
-
- uncompressedValSerializer -
Variable in class org.apache.hadoop.io.SequenceFile.Writer
-
- unEscapeString(String) -
Static method in class org.apache.hadoop.util.StringUtils
- Unescape commas in the string using the default escape char
- unEscapeString(String, char, char) -
Static method in class org.apache.hadoop.util.StringUtils
- Unescape
charToEscape
in the string
with the escape char escapeChar
- unEscapeString(String, char, char[]) -
Static method in class org.apache.hadoop.util.StringUtils
-
- UNINITIALIZED_MEMORY_VALUE -
Static variable in class org.apache.hadoop.mapred.ClusterStatus
-
- UNIQ_VALUE_COUNT -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- UniqValueCount - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that dedupes a sequence of objects.
- UniqValueCount() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
- the default constructor
- UniqValueCount(long) -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
- constructor
- unJar(File, File) -
Static method in class org.apache.hadoop.util.RunJar
- Unpack a jar file into a directory.
- unmarkJobFromLogDeletion(JobID) -
Method in class org.apache.hadoop.mapred.UserLogCleaner
- Remove job from user log deletion.
- unquoteHtmlChars(String) -
Static method in class org.apache.hadoop.http.HtmlQuoting
- Remove HTML quoting from a string.
- unregister(ObjectName) -
Static method in class org.apache.hadoop.metrics2.util.MBeans
-
- unregisterMBean(ObjectName) -
Static method in class org.apache.hadoop.metrics.util.MBeanUtil
- Deprecated.
- unregisterUpdater(Updater) -
Method in interface org.apache.hadoop.metrics.MetricsContext
- Deprecated. Removes a callback, if it exists.
- unregisterUpdater(Updater) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Removes a callback, if it exists.
- unregisterUpdater(Updater) -
Method in class org.apache.hadoop.metrics.spi.CompositeContext
- Deprecated.
- unreserveSlots(TaskType, JobInProgress) -
Method in class org.apache.hadoop.mapreduce.server.jobtracker.TaskTracker
- Free map slots on this
TaskTracker
which were reserved for
taskType
.
- unreserveTaskTracker(TaskTracker, TaskType) -
Method in class org.apache.hadoop.mapred.JobInProgress
-
- unset(String) -
Method in class org.apache.hadoop.conf.Configuration
- Unset a previously set property.
- unTar(File, File) -
Static method in class org.apache.hadoop.fs.FileUtil
- Given a Tar File as input it will untar the file in a the untar directory
passed as the second parameter
This utility will untar ".tar" files and ".tar.gz","tgz" files.
- unwrapRemoteException(Class<?>...) -
Method in exception org.apache.hadoop.ipc.RemoteException
- If this remote exception wraps up one of the lookupTypes
then return this exception.
- unwrapRemoteException() -
Method in exception org.apache.hadoop.ipc.RemoteException
- Instantiate and return the exception wrapped up by this remote exception.
- unZip(File, File) -
Static method in class org.apache.hadoop.fs.FileUtil
- Given a File input it will unzip the file in a the unzip directory
passed as the second parameter
- UPDATE -
Static variable in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
-
- update() -
Method in interface org.apache.hadoop.metrics.MetricsRecord
- Deprecated. Updates the table of buffered data which is to be sent periodically.
- update(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
- Deprecated. Called by MetricsRecordImpl.update().
- update() -
Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
- Deprecated. Updates the table of buffered data which is to be sent periodically.
- update(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.NullContext
- Deprecated. Do-nothing version of update
- update(MetricsRecordImpl) -
Method in class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
- Deprecated. Do-nothing version of update
- update(MetricsRecord, boolean) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache
- Update the cache and return the cached record
- update(MetricsRecord) -
Method in class org.apache.hadoop.metrics2.util.MetricsCache
-
- update(byte[], int, int) -
Method in class org.apache.hadoop.util.DataChecksum
-
- update(int) -
Method in class org.apache.hadoop.util.DataChecksum
-
- update(byte[], int, int) -
Method in class org.apache.hadoop.util.PureJavaCrc32
-
- update(int) -
Method in class org.apache.hadoop.util.PureJavaCrc32
-
- updateChunkChecksum(byte[], int, int, int, DataChecksum) -
Static method in class org.apache.hadoop.util.ChecksumUtil
- updates the checksum for a buffer
- updateFileNames(String, String) -
Method in class org.apache.hadoop.util.HostsFileReader
-
- UpdateIndex - Class in org.apache.hadoop.contrib.index.main
- A distributed "index" is partitioned into "shards".
- UpdateIndex() -
Constructor for class org.apache.hadoop.contrib.index.main.UpdateIndex
-
- UpdateLineColumn(char) -
Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- updatePrivateDistributedCacheSizes(JobID, long[]) -
Method in class org.apache.hadoop.mapred.TaskTracker
-
- updatePrivateDistributedCacheSizes(JobID, long[]) -
Method in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- The job initializer needs to report the sizes of the archive
objects and directories in the private distributed cache.
- updateProcessedByteCount(int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
- This method keeps track of raw processed compressed
bytes.
- Updater - Interface in org.apache.hadoop.metrics
- Deprecated. in favor of
org.apache.hadoop.metrics2
usage. - updateReportedByteCount(int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
- This method is called by the client of this
class in case there are any corrections in
the stream position.
- updateState(String, String, long) -
Static method in class org.apache.hadoop.contrib.failmon.PersistentState
- Upadate the state of parsing for a particular log file.
- updateTaskStatus(TaskInProgress, TaskStatus) -
Method in class org.apache.hadoop.mapred.JobInProgress
- Assuming
JobTracker
is locked on entry.
- upload() -
Method in class org.apache.hadoop.contrib.failmon.LocalStore
- Upload the local file store into HDFS, after it
compressing it.
- UPLOAD_INTERVAL -
Static variable in class org.apache.hadoop.contrib.failmon.LocalStore
-
- UPPER_LIMIT_ON_TASK_VMEM_PROPERTY -
Static variable in class org.apache.hadoop.mapred.JobConf
- Deprecated.
- upperBound(byte[]) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is strictly greater than
the input key.
- upperBound(byte[], int, int) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
- Move the cursor to the first entry whose key is strictly greater than
the input key.
- upperBound(List<? extends T>, T, Comparator<? super T>) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Upper bound binary search.
- upperBound(List<? extends Comparable<? super T>>, T) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Upper bound binary search.
- uriToString(URI[]) -
Static method in class org.apache.hadoop.util.StringUtils
-
- URL_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- JDBC Database access URL
- URL_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- JDBC Database access URL
- USAGE -
Static variable in class org.apache.hadoop.fs.shell.Count
-
- usage() -
Static method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- USAGES -
Static variable in class org.apache.hadoop.log.LogLevel
-
- useKsslAuth() -
Static method in class org.apache.hadoop.security.SecurityUtil
-
- USER_LOG_RETAIN_HOURS -
Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- USER_NAME -
Static variable in class org.apache.hadoop.security.authentication.client.PseudoAuthenticator
- Name of the additional parameter that carries the 'user.name' value.
- USER_NAME_COMMAND -
Static variable in class org.apache.hadoop.util.Shell
- a Unix command to get the current user's name
- UserDefinedValueAggregatorDescriptor - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a wrapper for a user defined value aggregator descriptor.
- UserDefinedValueAggregatorDescriptor(String, JobConf) -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
-
- UserGroupInformation - Class in org.apache.hadoop.security
- User and group information for Hadoop.
- UserGroupInformation.AuthenticationMethod - Enum in org.apache.hadoop.security
- existing types of authentications' methods
- UserGroupInformation.HadoopLoginModule - Class in org.apache.hadoop.security
- A login module that looks at the Kerberos, Unix, or Windows principal and
adds the corresponding UserName.
- UserGroupInformation.HadoopLoginModule() -
Constructor for class org.apache.hadoop.security.UserGroupInformation.HadoopLoginModule
-
- UserLogCleaner - Class in org.apache.hadoop.mapred
- This is used only in UserLogManager, to manage cleanup of user logs.
- UserLogCleaner(UserLogManager, Configuration) -
Constructor for class org.apache.hadoop.mapred.UserLogCleaner
-
- UserLogEvent - Class in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
- A directive from the various components of
TaskTracker
to the
UserLogManager
to inform about an event. - UserLogEvent(UserLogEvent.EventType) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogEvent
-
- UserLogEvent.EventType - Enum in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
-
- UserLogManager - Class in org.apache.hadoop.mapreduce.server.tasktracker.userlogs
- This manages user logs on the
TaskTracker
. - UserLogManager(Configuration, TaskController) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
- Create the user log manager to manage user logs on
TaskTracker
.
- UserLogManager(Configuration) -
Constructor for class org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogManager
- Create the user log manager to manage user logs on
TaskTracker
.
- USERNAME_PROPERTY -
Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
- User name to access the database
- USERNAME_PROPERTY -
Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
- User name to access the database
- userToNetgroupsMap -
Static variable in class org.apache.hadoop.security.ShellBasedUnixGroupsNetgroupMapping
-
- USTRING_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- UTF8 - Class in org.apache.hadoop.io
- Deprecated. replaced by Text
- UTF8() -
Constructor for class org.apache.hadoop.io.UTF8
- Deprecated.
- UTF8(String) -
Constructor for class org.apache.hadoop.io.UTF8
- Deprecated. Construct from a given string.
- UTF8(UTF8) -
Constructor for class org.apache.hadoop.io.UTF8
- Deprecated. Construct from a given string.
- UTF8.Comparator - Class in org.apache.hadoop.io
- Deprecated. A WritableComparator optimized for UTF8 keys.
- UTF8.Comparator() -
Constructor for class org.apache.hadoop.io.UTF8.Comparator
- Deprecated.
- UTF8ByteArrayUtils - Class in org.apache.hadoop.streaming
- Deprecated. use
UTF8ByteArrayUtils
and
StreamKeyValUtil
instead - UTF8ByteArrayUtils() -
Constructor for class org.apache.hadoop.streaming.UTF8ByteArrayUtils
- Deprecated.
- UTF8ByteArrayUtils - Class in org.apache.hadoop.util
-
- UTF8ByteArrayUtils() -
Constructor for class org.apache.hadoop.util.UTF8ByteArrayUtils
-
- utf8Length(String) -
Static method in class org.apache.hadoop.io.Text
- For the given string, returns the number of UTF-8 bytes
required to encode the string.
- Util - Class in org.apache.hadoop.metrics.spi
- Static utility methods
- Utils - Class in org.apache.hadoop.io.file.tfile
- Supporting Utility classes used by TFile, and shared by users of TFile.
- Utils - Class in org.apache.hadoop.mapred
- A utility class.
- Utils() -
Constructor for class org.apache.hadoop.mapred.Utils
-
- Utils - Class in org.apache.hadoop.record.meta
- Various utility functions for Hadooop record I/O platform.
- Utils - Class in org.apache.hadoop.record
- Various utility functions for Hadooop record I/O runtime.
- Utils.OutputFileUtils - Class in org.apache.hadoop.mapred
-
- Utils.OutputFileUtils() -
Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils
-
- Utils.OutputFileUtils.OutputFilesFilter - Class in org.apache.hadoop.mapred
- This class filters output(part) files from the given directory
It does not accept files with filenames _logs and _SUCCESS.
- Utils.OutputFileUtils.OutputFilesFilter() -
Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter
-
- Utils.OutputFileUtils.OutputLogFilter - Class in org.apache.hadoop.mapred
- This class filters log files from directory given
It doesnt accept paths having _logs.
- Utils.OutputFileUtils.OutputLogFilter() -
Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter
-
- Utils.Version - Class in org.apache.hadoop.io.file.tfile
- A generic Version class.
- Utils.Version(DataInput) -
Constructor for class org.apache.hadoop.io.file.tfile.Utils.Version
- Construct the Version object by reading from the input stream.
- Utils.Version(short, short) -
Constructor for class org.apache.hadoop.io.file.tfile.Utils.Version
- Constructor.
V
- validate(Configuration) -
Static method in class org.apache.hadoop.filecache.TrackerDistributedCacheManager
- This is part of the framework API.
- validateUTF8(byte[]) -
Static method in class org.apache.hadoop.io.Text
- Check if a byte array contains valid utf-8
- validateUTF8(byte[], int, int) -
Static method in class org.apache.hadoop.io.Text
- Check to see if a byte array is valid utf-8
- value() -
Method in class org.apache.hadoop.metrics2.Metric
- Get the value of the metric
- value() -
Method in class org.apache.hadoop.metrics2.MetricCounter
-
- value() -
Method in class org.apache.hadoop.metrics2.MetricGauge
-
- value() -
Method in class org.apache.hadoop.metrics2.MetricsTag
- Get the value of the tag
- value -
Variable in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
-
- VALUE_CLASS -
Static variable in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
-
- VALUE_HISTOGRAM -
Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- ValueAggregator - Interface in org.apache.hadoop.mapred.lib.aggregate
- This interface defines the minimal protocol for value aggregators.
- ValueAggregatorBaseDescriptor - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements the common functionalities of
the subclasses of ValueAggregatorDescriptor class.
- ValueAggregatorBaseDescriptor() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- ValueAggregatorCombiner<K1 extends WritableComparable,V1 extends Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements the generic combiner of Aggregate.
- ValueAggregatorCombiner() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
-
- ValueAggregatorDescriptor - Interface in org.apache.hadoop.mapred.lib.aggregate
- This interface defines the contract a value aggregator descriptor must
support.
- ValueAggregatorJob - Class in org.apache.hadoop.mapred.lib.aggregate
- This is the main class for creating a map/reduce job using Aggregate
framework.
- ValueAggregatorJob() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- ValueAggregatorJobBase<K1 extends WritableComparable,V1 extends Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
- This abstract class implements some common functionalities of the
the generic mapper, reducer and combiner classes of Aggregate.
- ValueAggregatorJobBase() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
-
- ValueAggregatorMapper<K1 extends WritableComparable,V1 extends Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements the generic mapper of Aggregate.
- ValueAggregatorMapper() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
-
- ValueAggregatorReducer<K1 extends WritableComparable,V1 extends Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements the generic reducer of Aggregate.
- ValueAggregatorReducer() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
-
- ValueHistogram - Class in org.apache.hadoop.mapred.lib.aggregate
- This class implements a value aggregator that computes the
histogram of a sequence of strings.
- ValueHistogram() -
Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
- valueOf(String) -
Static method in enum org.apache.hadoop.contrib.failmon.OfflineAnonymizer.LogType
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.examples.dancing.Pentomino.SolutionCategory
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.fs.HardLink.OSType
- Returns the enum constant of this type with the specified name.
- valueOf(Attributes) -
Static method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Return the object represented in the attributes.
- valueOf(String) -
Static method in enum org.apache.hadoop.fs.permission.FsAction
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in class org.apache.hadoop.fs.permission.FsPermission
- Create a FsPermission from a Unix symbolic permission string
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.bzip2.CBZip2InputStream.STATE
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.SplittableCompressionCodec.READ_MODE
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.nativeio.Errno
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.io.SequenceFile.CompressionType
- Returns the enum constant of this type with the specified name.
- valueOf(Attributes) -
Static method in exception org.apache.hadoop.ipc.RemoteException
- Create RemoteException from attributes
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.FileInputFormat.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.FileOutputFormat.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobClient.TaskStatusFilter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobHistory.Keys
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobHistory.RecordTypes
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobHistory.Values
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobInProgress.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobPriority
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobTracker.SafeModeAction
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.JobTracker.State
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.join.Parser.TType
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.Operation
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.Task.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.TaskCompletionEvent.Status
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.TaskLog.LogName
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.TaskStatus.Phase
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.TaskStatus.State
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapred.TIPStatus
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.Job.JobState
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.JobACL
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.JobStatus.State
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.lib.input.FileInputFormat.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.Counter
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogEvent.EventType
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.mapreduce.TaskType
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink.GangliaConfType
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink.GangliaSlope
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.security.Krb5AndCertsSslSocketConnector.MODE
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.security.SaslRpcServer.QualityOfProtection
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.security.SaslRpcServer.SaslStatus
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.typedbytes.Type
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.util.ProcessTree.Signal
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.util.QueueProcessingStatistics.State
- Returns the enum constant of this type with the specified name.
- valueOf(String) -
Static method in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
- Returns the enum constant of this type with the specified name.
- valueOf(char) -
Static method in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
-
- values() -
Static method in enum org.apache.hadoop.contrib.failmon.OfflineAnonymizer.LogType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.examples.dancing.Pentomino.SolutionCategory
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.fs.HardLink.OSType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.fs.permission.FsAction
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.bzip2.CBZip2InputStream.STATE
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.SplittableCompressionCodec.READ_MODE
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Method in class org.apache.hadoop.io.MapWritable
-
- values() -
Static method in enum org.apache.hadoop.io.nativeio.Errno
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.io.SequenceFile.CompressionType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Method in class org.apache.hadoop.io.SortedMapWritable
-
- values() -
Static method in enum org.apache.hadoop.mapred.FileInputFormat.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.FileOutputFormat.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobClient.TaskStatusFilter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobHistory.Keys
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobHistory.RecordTypes
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobHistory.Values
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobInProgress.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobPriority
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobTracker.SafeModeAction
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.JobTracker.State
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.join.Parser.TType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.Operation
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.Task.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.TaskCompletionEvent.Status
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.TaskLog.LogName
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.TaskStatus.Phase
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.TaskStatus.State
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapred.TIPStatus
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.Job.JobState
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.JobACL
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.JobStatus.State
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.lib.input.FileInputFormat.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.Counter
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.server.tasktracker.userlogs.UserLogEvent.EventType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.mapreduce.TaskType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.metrics2.lib.DefaultMetricsSystem
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink.GangliaConfType
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink.GangliaSlope
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.security.Krb5AndCertsSslSocketConnector.MODE
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.security.SaslRpcServer.QualityOfProtection
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.security.SaslRpcServer.SaslStatus
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.security.UserGroupInformation.AuthenticationMethod
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.typedbytes.Type
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.util.ProcessTree.Signal
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.util.QueueProcessingStatistics.State
- Returns an array containing the constants of this enum type, in
the order they are declared.
- values() -
Static method in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
- Returns an array containing the constants of this enum type, in
the order they are declared.
- variance() -
Method in class org.apache.hadoop.metrics2.util.SampleStat
-
- Vector() -
Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- VECTOR -
Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- VECTOR_TKN -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- vectorSize -
Variable in class org.apache.hadoop.util.bloom.Filter
- The vector size of this filter.
- VectorTypeID - Class in org.apache.hadoop.record.meta
- Represents typeID for vector.
- VectorTypeID(TypeID) -
Constructor for class org.apache.hadoop.record.meta.VectorTypeID
-
- verbose -
Variable in class org.apache.hadoop.streaming.JarBuilder
-
- verbose_ -
Variable in class org.apache.hadoop.streaming.StreamJob
-
- verifyAndExtract(String) -
Method in class org.apache.hadoop.security.authentication.util.Signer
- Verifies a signed string and extracts the original string.
- verifyHostnames(String[]) -
Static method in class org.apache.hadoop.net.NetUtils
- Performs a sanity check on the list of hostnames/IPs to verify they at least
appear to be valid.
- verifyReply(String, String, SecretKey) -
Static method in class org.apache.hadoop.mapreduce.security.SecureShuffleUtils
- verify that base64Hash is same as HMacHash(msg)
- verifyToken(TokenIdent, byte[]) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenSecretManager
- Verifies that the given identifier and password are valid and match.
- VERSION -
Static variable in class org.apache.hadoop.fs.HarFileSystem
-
- VersionedProtocol - Interface in org.apache.hadoop.ipc
- Superclass of all protocols that use Hadoop RPC.
- VersionedWritable - Class in org.apache.hadoop.io
- A base class for Writables that provides version checking.
- VersionedWritable() -
Constructor for class org.apache.hadoop.io.VersionedWritable
-
- versionID -
Static variable in interface org.apache.hadoop.mapred.AdminOperationsProtocol
- Version 1: Initial version.
- versionID -
Static variable in interface org.apache.hadoop.mapred.TaskUmbilicalProtocol
- Changed the version to 2, since we have a new method getMapOutputs
Changed version to 3 to have progress() return a boolean
Changed the version to 4, since we have replaced
TaskUmbilicalProtocol.progress(String, float, String,
org.apache.hadoop.mapred.TaskStatus.Phase, Counters)
with statusUpdate(String, TaskStatus)
Version 5 changed counters representation for HADOOP-2248
Version 6 changes the TaskStatus representation for HADOOP-2208
Version 7 changes the done api (via HADOOP-3140).
- versionID -
Static variable in interface org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol
- Version 1: Initial version
- versionID -
Static variable in interface org.apache.hadoop.security.RefreshUserMappingsProtocol
- Version 1: Initial version.
- VersionInfo - Class in org.apache.hadoop.util
- This class finds the package info for Hadoop and the HadoopVersionAnnotation
information.
- VersionInfo() -
Constructor for class org.apache.hadoop.util.VersionInfo
-
- VersionMismatchException - Exception in org.apache.hadoop.fs.s3
- Thrown when Hadoop cannot read the version of the data stored
in
S3FileSystem
. - VersionMismatchException(String, String) -
Constructor for exception org.apache.hadoop.fs.s3.VersionMismatchException
-
- VersionMismatchException - Exception in org.apache.hadoop.io
- Thrown by
VersionedWritable.readFields(DataInput)
when the
version of an object being read does not match the current implementation
version as returned by VersionedWritable.getVersion()
. - VersionMismatchException(byte, byte) -
Constructor for exception org.apache.hadoop.io.VersionMismatchException
-
- VIntWritable - Class in org.apache.hadoop.io
- A WritableComparable for integer values stored in variable-length format.
- VIntWritable() -
Constructor for class org.apache.hadoop.io.VIntWritable
-
- VIntWritable(int) -
Constructor for class org.apache.hadoop.io.VIntWritable
-
- visit(MetricsVisitor) -
Method in class org.apache.hadoop.metrics2.Metric
- Accept a visitor interface
- VLongWritable - Class in org.apache.hadoop.io
- A WritableComparable for longs in a variable-length format.
- VLongWritable() -
Constructor for class org.apache.hadoop.io.VLongWritable
-
- VLongWritable(long) -
Constructor for class org.apache.hadoop.io.VLongWritable
-
W
- waitForCompletion() -
Method in interface org.apache.hadoop.mapred.RunningJob
- Blocks until the job is complete.
- waitForCompletion(boolean) -
Method in class org.apache.hadoop.mapreduce.Job
- Submit the job to the cluster and wait for it to finish.
- waitForProxy(Class<? extends VersionedProtocol>, long, InetSocketAddress, Configuration) -
Static method in class org.apache.hadoop.ipc.RPC
-
- waitForReadable() -
Method in class org.apache.hadoop.net.SocketInputStream
- waits for the underlying channel to be ready for reading.
- waitForWritable() -
Method in class org.apache.hadoop.net.SocketOutputStream
- waits for the underlying channel to be ready for writing.
- WAITING -
Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- webAppContext -
Variable in class org.apache.hadoop.http.HttpServer
-
- webServer -
Variable in class org.apache.hadoop.http.HttpServer
-
- width -
Variable in class org.apache.hadoop.examples.dancing.Pentomino
-
- width -
Static variable in class org.apache.hadoop.mapred.TaskGraphServlet
- height of the graph w/o margins
- WILDCARD_ACL_VALUE -
Static variable in class org.apache.hadoop.security.authorize.AccessControlList
-
- windowBits() -
Method in enum org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader
-
- windowBits() -
Method in enum org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader
-
- WINDOWS -
Static variable in class org.apache.hadoop.util.Shell
- Set to true on Windows platforms
- WithinMultiLineComment -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- WithinOneLineComment -
Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
- WordCount - Class in org.apache.hadoop.examples
-
- WordCount() -
Constructor for class org.apache.hadoop.examples.WordCount
-
- WordCount.IntSumReducer - Class in org.apache.hadoop.examples
-
- WordCount.IntSumReducer() -
Constructor for class org.apache.hadoop.examples.WordCount.IntSumReducer
-
- WordCount.TokenizerMapper - Class in org.apache.hadoop.examples
-
- WordCount.TokenizerMapper() -
Constructor for class org.apache.hadoop.examples.WordCount.TokenizerMapper
-
- WORK_FACTOR -
Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
- This constant is accessible by subclasses for historical purposes.
- WORKFLOW_ADJACENCY_PREFIX_PATTERN -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- WORKFLOW_ADJACENCY_PREFIX_STRING -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- WORKFLOW_ID -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- WORKFLOW_NAME -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- WORKFLOW_NODE_NAME -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- WORKFLOW_TAGS -
Static variable in class org.apache.hadoop.mapred.JobConf
-
- wrappedReadForCompressedData(InputStream, byte[], int, int) -
Static method in class org.apache.hadoop.io.IOUtils
- Utility wrapper for reading from
InputStream
.
- WrappedRecordReader<K extends WritableComparable,U extends Writable> - Class in org.apache.hadoop.mapred.join
- Proxy class for a RecordReader participating in the join framework.
- Writable - Interface in org.apache.hadoop.io
- A serializable object which implements a simple, efficient, serialization
protocol, based on
DataInput
and DataOutput
. - WritableComparable<T> - Interface in org.apache.hadoop.io
- A
Writable
which is also Comparable
. - WritableComparator - Class in org.apache.hadoop.io
- A Comparator for
WritableComparable
s. - WritableComparator(Class<? extends WritableComparable>) -
Constructor for class org.apache.hadoop.io.WritableComparator
- Construct for a
WritableComparable
implementation.
- WritableComparator(Class<? extends WritableComparable>, boolean) -
Constructor for class org.apache.hadoop.io.WritableComparator
-
- WritableFactories - Class in org.apache.hadoop.io
- Factories for non-public writables.
- WritableFactory - Interface in org.apache.hadoop.io
- A factory for a class of Writable.
- WritableName - Class in org.apache.hadoop.io
- Utility to permit renaming of Writable implementation classes without
invalidiating files that contain their class name.
- WritableSerialization - Class in org.apache.hadoop.io.serializer
- A
Serialization
for Writable
s that delegates to
Writable.write(java.io.DataOutput)
and
Writable.readFields(java.io.DataInput)
. - WritableSerialization() -
Constructor for class org.apache.hadoop.io.serializer.WritableSerialization
-
- WritableUtils - Class in org.apache.hadoop.io
-
- WritableUtils() -
Constructor for class org.apache.hadoop.io.WritableUtils
-
- write(DataOutput) -
Method in class org.apache.hadoop.conf.Configuration
-
- write(DataOutput) -
Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
- write(DataOutput) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
- write(DataOutput) -
Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- write(DataOutput) -
Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
- write(DataOutput) -
Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- write(DataOutput) -
Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- write(DataOutput) -
Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
- write(DataOutput) -
Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- write(DataOutput) -
Method in class org.apache.hadoop.fs.BlockLocation
- Implement write of Writable
- write(DataOutput) -
Method in class org.apache.hadoop.fs.ContentSummary
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.fs.FileStatus
-
- write(int) -
Method in class org.apache.hadoop.fs.FSOutputSummer
- Write one byte
- write(byte[], int, int) -
Method in class org.apache.hadoop.fs.FSOutputSummer
- Writes
len
bytes from the specified byte array
starting at offset off
and generate a checksum for
each data chunk.
- write(DataOutput) -
Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Serialize the fields of this object to
out
.
- write(XMLOutputter, MD5MD5CRC32FileChecksum) -
Static method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
- Write that object to xml output.
- write(DataOutput) -
Method in class org.apache.hadoop.fs.permission.FsPermission
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.fs.permission.PermissionStatus
- Serialize the fields of this object to
out
.
- write(DataOutput, String, String, FsPermission) -
Static method in class org.apache.hadoop.fs.permission.PermissionStatus
- Serialize a
PermissionStatus
from its base components.
- write(DataOutput) -
Method in class org.apache.hadoop.io.AbstractMapWritable
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.io.ArrayWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.BooleanWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.BytesWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.ByteWritable
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.BlockCompressorStream
- Write the data provided to the compression codec, compressing no more
than the buffer size less the compression overhead as specified during
construction for each block.
- write(int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.CompressionOutputStream
- Write compressed bytes to the stream.
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.CompressorStream
-
- write(int) -
Method in class org.apache.hadoop.io.compress.CompressorStream
-
- write(int) -
Method in class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.CompressedWritable
-
- write(DataInput, int) -
Method in class org.apache.hadoop.io.DataOutputBuffer
- Writes bytes from a DataInput directly into the buffer.
- write(DataOutput) -
Method in class org.apache.hadoop.io.DoubleWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.file.tfile.Utils.Version
- Write the objec to a DataOutput.
- write(DataOutput) -
Method in class org.apache.hadoop.io.FloatWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.GenericWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.IntWritable
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.io.IOUtils.NullOutputStream
-
- write(int) -
Method in class org.apache.hadoop.io.IOUtils.NullOutputStream
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.LongWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.MapWritable
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.io.MD5Hash
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.NullWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.ObjectWritable
-
- write(InputStream, int) -
Method in class org.apache.hadoop.io.OutputBuffer
- Writes bytes from a InputStream directly into the buffer.
- write(DataOutput) -
Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.SortedMapWritable
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.io.Text
- serialize
write this object to out
length uses zero-compressed encoding
- write(DataOutput) -
Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.UTF8
- Deprecated.
- write(DataOutput) -
Method in class org.apache.hadoop.io.VersionedWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.VIntWritable
-
- write(DataOutput) -
Method in class org.apache.hadoop.io.VLongWritable
-
- write(DataOutput) -
Method in interface org.apache.hadoop.io.Writable
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.ClusterStatus
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.Counters.Group
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.Counters
- Write the set of groups.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.FileSplit
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.JobProfile
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.JobQueueInfo
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.JobStatus
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
- Write splits in the following format.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.join.TupleWritable
- Writes each Writable to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.JvmTask
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- write(PreparedStatement) -
Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- write(K, V) -
Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter
- Writes a key/value pair.
- write(PreparedStatement) -
Method in interface org.apache.hadoop.mapred.lib.db.DBWritable
- Sets the fields of the object in the
PreparedStatement
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.QueueAclsInfo
-
- write(K, V) -
Method in interface org.apache.hadoop.mapred.RecordWriter
- Writes a key/value pair.
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.Task
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.TaskReport
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.TaskStatus
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapred.TaskTrackerStatus
-
- write(K, V) -
Method in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.ClusterMetrics
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.Counter
- Write the binary representation of the counter
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.Counters
- Write the set of groups.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.ID
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.JobID
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
-
- write(PreparedStatement) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
-
- write(K, V) -
Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
- Writes a key/value pair.
- write(PreparedStatement) -
Method in interface org.apache.hadoop.mapreduce.lib.db.DBWritable
- Sets the fields of the object in the
PreparedStatement
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
- write(K, V) -
Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
-
- write(String, K, V) -
Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
- Write key and value to the namedOutput.
- write(String, K, V, String) -
Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
- Write key and value to baseOutputPath using the namedOutput.
- write(KEYOUT, VALUEOUT, String) -
Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
- Write key value to an output file name.
- write(K, V) -
Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
-
- write(K, V) -
Method in class org.apache.hadoop.mapreduce.RecordWriter
- Writes a key/value pair.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.security.token.JobTokenIdentifier
- Serialize the fields of this object to
out
.
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
- write(DataOutput) -
Method in class org.apache.hadoop.mapreduce.TaskID
-
- write(KEYOUT, VALUEOUT) -
Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
- Generate an output key/value pair.
- write(int) -
Method in class org.apache.hadoop.net.SocketOutputStream
-
- write(byte[], int, int) -
Method in class org.apache.hadoop.net.SocketOutputStream
-
- write(ByteBuffer) -
Method in class org.apache.hadoop.net.SocketOutputStream
-
- write(DataOutput) -
Method in class org.apache.hadoop.record.Record
-
- write(DataOutput) -
Method in class org.apache.hadoop.security.authorize.AccessControlList
-
- write(DataOutput) -
Method in class org.apache.hadoop.security.Credentials
- Stores all the keys to DataOutput
- write(int) -
Method in class org.apache.hadoop.security.SaslOutputStream
- Writes the specified byte to this output stream.
- write(byte[]) -
Method in class org.apache.hadoop.security.SaslOutputStream
- Writes
b.length
bytes from the specified byte array to this
output stream.
- write(byte[], int, int) -
Method in class org.apache.hadoop.security.SaslOutputStream
- Writes
len
bytes from the specified byte array starting at
offset off
to this output stream.
- write(DataOutput) -
Method in enum org.apache.hadoop.security.SaslRpcServer.AuthMethod
- Write to out
- write(DataOutput) -
Method in class org.apache.hadoop.security.token.delegation.AbstractDelegationTokenIdentifier
-
- write(DataOutput) -
Method in class org.apache.hadoop.security.token.delegation.DelegationKey
-
- write(DataOutput) -
Method in class org.apache.hadoop.security.token.Token
- Serialize the fields of this object to
out
.
- write(Object) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a Java object as a typed bytes sequence.
- write(Writable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.Filter
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.Key
-
- write(DataOutput) -
Method in class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
- writeArray(ArrayWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeBool(boolean, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeBool(boolean, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeBool(boolean, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a boolean to serialized record.
- writeBool(boolean, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeBool(boolean) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a boolean as a typed bytes sequence.
- writeBool(boolean, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeBoolean(BooleanWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeBuffer(Buffer, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeBuffer(Buffer, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeBuffer(Buffer, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a buffer to serialized record.
- writeBuffer(Buffer, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeBuffer(Buffer, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeByte(byte, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeByte(byte, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeByte(byte, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a byte to serialized record.
- writeByte(byte, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeByte(byte) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a byte as a typed bytes sequence.
- writeByte(byte, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeByte(ByteWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeBytes(byte[], int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a bytes array as a typed bytes sequence, using a given typecode.
- writeBytes(byte[]) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a bytes array as a typed bytes sequence.
- writeBytes(BytesWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeChunk(byte[], int, int, byte[]) -
Method in class org.apache.hadoop.fs.FSOutputSummer
-
- writeCommand(String, FileSystem, Path) -
Static method in class org.apache.hadoop.mapred.TaskController
-
- writeCompressed(DataOutput) -
Method in class org.apache.hadoop.io.CompressedWritable
- Subclasses implement this instead of
CompressedWritable.write(DataOutput)
.
- writeCompressedByteArray(DataOutput, byte[]) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- writeCompressedBytes(DataOutputStream) -
Method in interface org.apache.hadoop.io.SequenceFile.ValueBytes
- Write compressed bytes to outStream.
- writeCompressedBytes(DataOutputStream) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- writeCompressedBytes(DataOutputStream) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- writeCompressedString(DataOutput, String) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- writeCompressedStringArray(DataOutput, String[]) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- writeDouble(double, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeDouble(double, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeDouble(double, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a double precision floating point number to serialized record.
- writeDouble(double, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeDouble(double) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a double as a typed bytes sequence.
- writeDouble(double, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeDouble(DoubleWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeEnum(DataOutput, Enum<?>) -
Static method in class org.apache.hadoop.io.WritableUtils
- writes String value of enum to DataOutput.
- writeFile(SequenceFile.Sorter.RawKeyValueIterator, SequenceFile.Writer) -
Method in class org.apache.hadoop.io.SequenceFile.Sorter
- Writes records from RawKeyValueIterator into a file represented by the
passed writer
- writeFloat(float, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeFloat(float, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeFloat(float, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a single-precision float to serialized record.
- writeFloat(float, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeFloat(float) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a float as a typed bytes sequence.
- writeFloat(float, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeFloat(FloatWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeHeader(DataOutputStream) -
Method in class org.apache.hadoop.util.DataChecksum
- Writes the checksum header to the output stream out.
- writeInt(int, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeInt(int, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeInt(int, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write an integer to serialized record.
- writeInt(int, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeInt(int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes an integer as a typed bytes sequence.
- writeInt(int, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeInt(IntWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeKey(OutputStream) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
- Writing the key to the output stream.
- writeKey(K) -
Method in class org.apache.hadoop.streaming.io.InputWriter
- Writes an input key.
- writeKey(Writable) -
Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
-
- writeKey(Object) -
Method in class org.apache.hadoop.streaming.io.TextInputWriter
-
- writeKey(Object) -
Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
-
- writeList(List) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a list as a typed bytes sequence.
- writeListFooter() -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a list footer.
- writeListHeader() -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a list header.
- writeLocalJobFile(Path, JobConf) -
Static method in class org.apache.hadoop.mapred.JobLocalizer
- Write the task specific job-configuration file.
- writeLong(long, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeLong(long, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeLong(long, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a long integer to serialized record.
- writeLong(long, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeLong(long) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a long as a typed bytes sequence.
- writeLong(long, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeLong(LongWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeMap(Map) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a map as a typed bytes sequence.
- writeMap(MapWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeMapHeader(int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a map header.
- writeObject(DataOutput, Object, Class, Configuration) -
Static method in class org.apache.hadoop.io.ObjectWritable
- Write a
Writable
, String
, primitive type, or an array of
the preceding.
- writePartitionFile(JobConf, Path) -
Static method in class org.apache.hadoop.examples.terasort.TeraInputFormat
- Use the input splits to take samples of the input and generate sample
keys.
- writePartitionFile(JobConf, InputSampler.Sampler<K, V>) -
Static method in class org.apache.hadoop.mapred.lib.InputSampler
- Write a partition file for the given job, using the Sampler provided.
- writePartitionFile(Job, InputSampler.Sampler<K, V>) -
Static method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
- Write a partition file for the given job, using the Sampler provided.
- writeRAMFiles(DataOutput, RAMDirectory, String[]) -
Static method in class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
- Write a number of files from a ram directory to a data output.
- writeRaw(byte[]) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a raw sequence of typed bytes.
- writeRaw(byte[], int, int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a raw sequence of typed bytes.
- writeRecord(Record) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeSortedMap(SortedMapWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeState(String) -
Static method in class org.apache.hadoop.contrib.failmon.PersistentState
- Write the state of parsing for all open log files to a property
file on disk.
- writeString(DataOutput, String) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Write a String as a VInt n, followed by n Bytes as in Text format.
- writeString(DataOutput, String) -
Static method in class org.apache.hadoop.io.Text
- Write a UTF8 encoded string to out
- writeString(DataOutput, String) -
Static method in class org.apache.hadoop.io.UTF8
- Deprecated. Write a UTF-8 encoded string.
- writeString(DataOutput, String) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- writeString(String, String) -
Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- writeString(String, String) -
Method in class org.apache.hadoop.record.CsvRecordOutput
-
- writeString(String, String) -
Method in interface org.apache.hadoop.record.RecordOutput
- Write a unicode string to serialized record.
- writeString(String, String) -
Method in class org.apache.hadoop.record.XmlRecordOutput
-
- writeString(String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a string as a typed bytes sequence.
- writeString(String, String) -
Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
-
- writeStringArray(DataOutput, String[]) -
Static method in class org.apache.hadoop.io.WritableUtils
-
- writeText(Text) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeTo(OutputStream) -
Method in class org.apache.hadoop.io.DataOutputBuffer
- Write to a file stream
- writeTokenStorageFile(Path, Configuration) -
Method in class org.apache.hadoop.security.Credentials
-
- writeTokenStorageToStream(DataOutputStream) -
Method in class org.apache.hadoop.security.Credentials
-
- writeTypedBytes(TypedBytesWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeUncompressedBytes(DataOutputStream) -
Method in interface org.apache.hadoop.io.SequenceFile.ValueBytes
- Writes the uncompressed bytes to the outStream.
- writeUncompressedBytes(DataOutputStream) -
Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- writeUncompressedBytes(DataOutputStream) -
Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- writeValue(OutputStream) -
Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
- Writing the value to the output stream.
- writeValue(V) -
Method in class org.apache.hadoop.streaming.io.InputWriter
- Writes an input value.
- writeValue(Writable) -
Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
-
- writeValue(Object) -
Method in class org.apache.hadoop.streaming.io.TextInputWriter
-
- writeValue(Object) -
Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
-
- writeValue(DataOutputStream, boolean) -
Method in class org.apache.hadoop.util.DataChecksum
- Writes the current checksum to the stream.
- writeValue(byte[], int, boolean) -
Method in class org.apache.hadoop.util.DataChecksum
- Writes the current checksum to a buffer.
- writeVector(ArrayList) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a vector as a typed bytes sequence.
- writeVectorHeader(int) -
Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
- Writes a vector header.
- writeVInt(DataOutput, int) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Encoding an integer into a variable-length encoding format.
- writeVInt(DataOutput, int) -
Static method in class org.apache.hadoop.io.WritableUtils
- Serializes an integer to a binary stream with zero-compressed encoding.
- writeVInt(DataOutput, int) -
Static method in class org.apache.hadoop.record.Utils
- Serializes an int to a binary stream with zero-compressed encoding.
- writeVInt(VIntWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeVLong(DataOutput, long) -
Static method in class org.apache.hadoop.io.file.tfile.Utils
- Encoding a Long integer into a variable-length encoding format.
- writeVLong(DataOutput, long) -
Static method in class org.apache.hadoop.io.WritableUtils
- Serializes a long to a binary stream with zero-compressed encoding.
- writeVLong(DataOutput, long) -
Static method in class org.apache.hadoop.record.Utils
- Serializes a long to a binary stream with zero-compressed encoding.
- writeVLong(VLongWritable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeWritable(Writable) -
Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
-
- writeXml(OutputStream) -
Method in class org.apache.hadoop.conf.Configuration
- Write out the non-default properties in this configuration to the given
OutputStream
.
- writeXml(Writer) -
Method in class org.apache.hadoop.conf.Configuration
- Write out the non-default properties in this configuration to the given
Writer
.
- WWW_AUTHENTICATE -
Static variable in class org.apache.hadoop.security.authentication.client.KerberosAuthenticator
- HTTP header used by the SPNEGO server endpoint during an authentication sequence.
X
- xdr_int(int) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
- Deprecated. Puts an integer into the buffer as 4 bytes, big-endian.
- xdr_int(int) -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
- Puts an integer into the buffer as 4 bytes, big-endian.
- xdr_string(String) -
Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
- Deprecated. Puts a string into the buffer by first writing the size of the string
as an int, followed by the bytes of the string, padded if necessary to
a multiple of 4.
- xdr_string(String) -
Method in class org.apache.hadoop.metrics2.sink.ganglia.AbstractGangliaSink
- Puts a string into the buffer by first writing the size of the string as an
int, followed by the bytes of the string, padded if necessary to a multiple
of 4.
- xmargin -
Static variable in class org.apache.hadoop.mapred.TaskGraphServlet
- margin space on x axis
- XmlRecordInput - Class in org.apache.hadoop.record
- XML Deserializer.
- XmlRecordInput(InputStream) -
Constructor for class org.apache.hadoop.record.XmlRecordInput
- Creates a new instance of XmlRecordInput
- XmlRecordOutput - Class in org.apache.hadoop.record
- XML Serializer.
- XmlRecordOutput(OutputStream) -
Constructor for class org.apache.hadoop.record.XmlRecordOutput
- Creates a new instance of XmlRecordOutput
- XMLUtils - Class in org.apache.hadoop.util
- General xml utilities.
- XMLUtils() -
Constructor for class org.apache.hadoop.util.XMLUtils
-
- xor(Filter) -
Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- xor(Filter) -
Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- xor(Filter) -
Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- xor(Filter) -
Method in class org.apache.hadoop.util.bloom.Filter
- Peforms a logical XOR between this filter and a specified filter.
Y
- ymargin -
Static variable in class org.apache.hadoop.mapred.TaskGraphServlet
- margin space on y axis
Z
- zipCompress(String) -
Static method in class org.apache.hadoop.contrib.failmon.LocalStore
- Compress a text file using the ZIP compressing algorithm.
- ZlibCompressor - Class in org.apache.hadoop.io.compress.zlib
- A
Compressor
based on the popular
zlib compression algorithm. - ZlibCompressor() -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibCompressor
- Creates a new compressor with the default compression level.
- ZlibCompressor(Configuration) -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibCompressor
- Creates a new compressor, taking settings from the configuration.
- ZlibCompressor(ZlibCompressor.CompressionLevel, ZlibCompressor.CompressionStrategy, ZlibCompressor.CompressionHeader, int) -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibCompressor
- Creates a new compressor using the specified compression level.
- ZlibCompressor.CompressionHeader - Enum in org.apache.hadoop.io.compress.zlib
- The type of header for compressed data.
- ZlibCompressor.CompressionLevel - Enum in org.apache.hadoop.io.compress.zlib
- The compression level for zlib library.
- ZlibCompressor.CompressionStrategy - Enum in org.apache.hadoop.io.compress.zlib
- The compression level for zlib library.
- ZlibDecompressor - Class in org.apache.hadoop.io.compress.zlib
- A
Decompressor
based on the popular
zlib compression algorithm. - ZlibDecompressor(ZlibDecompressor.CompressionHeader, int) -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
- Creates a new decompressor.
- ZlibDecompressor() -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- ZlibDecompressor.CompressionHeader - Enum in org.apache.hadoop.io.compress.zlib
- The headers to detect from compressed data.
- ZlibFactory - Class in org.apache.hadoop.io.compress.zlib
- A collection of factories to create the right
zlib/gzip compressor/decompressor instances.
- ZlibFactory() -
Constructor for class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z
Overview
Package
Class
Use
Tree
Deprecated
Index
Help
PREV
NEXT
FRAMES
NO FRAMES
All Classes
Copyright © 2009 The Apache Software Foundation