|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Uses of Configuration in org.apache.hadoop.conf |
---|
Methods in org.apache.hadoop.conf that return Configuration | |
---|---|
Configuration |
Configured.getConf()
|
Configuration |
Configurable.getConf()
Return the configuration used by this object. |
Methods in org.apache.hadoop.conf with parameters of type Configuration | |
---|---|
static void |
Configuration.dumpConfiguration(Configuration conf,
Writer out)
Writes out all the parameters and their properties (final and resource) to the given Writer
The format of the output would be
{ "properties" : [ {key1,value1,key1.isFinal,key1.resource}, {key2,value2,
key2.isFinal,key2.resource}... |
void |
Configured.setConf(Configuration conf)
|
void |
Configurable.setConf(Configuration conf)
Set the configuration to be used by this object. |
Constructors in org.apache.hadoop.conf with parameters of type Configuration | |
---|---|
Configuration(Configuration other)
A new configuration with the same settings cloned from another. |
|
Configured(Configuration conf)
Construct a Configured. |
Uses of Configuration in org.apache.hadoop.contrib.failmon |
---|
Constructors in org.apache.hadoop.contrib.failmon with parameters of type Configuration | |
---|---|
Executor(Configuration conf)
Create an instance of the class and read the configuration file to determine the set of jobs that will be run and the maximum interval for which the thread can sleep before it wakes up to execute a monitoring job on the node. |
Uses of Configuration in org.apache.hadoop.contrib.index.example |
---|
Constructors in org.apache.hadoop.contrib.index.example with parameters of type Configuration | |
---|---|
LineDocRecordReader(Configuration job,
FileSplit split)
Constructor |
Uses of Configuration in org.apache.hadoop.contrib.index.lucene |
---|
Constructors in org.apache.hadoop.contrib.index.lucene with parameters of type Configuration | |
---|---|
FileSystemDirectory(FileSystem fs,
Path directory,
boolean create,
Configuration conf)
Constructor |
Uses of Configuration in org.apache.hadoop.contrib.index.mapred |
---|
Methods in org.apache.hadoop.contrib.index.mapred that return Configuration | |
---|---|
Configuration |
IndexUpdateConfiguration.getConfiguration()
Get the underlying configuration object. |
Methods in org.apache.hadoop.contrib.index.mapred with parameters of type Configuration | |
---|---|
void |
IIndexUpdater.run(Configuration conf,
Path[] inputPaths,
Path outputPath,
int numMapTasks,
Shard[] shards)
Create a Map/Reduce job configuration and run the Map/Reduce job to analyze documents and update Lucene instances in parallel. |
void |
IndexUpdater.run(Configuration conf,
Path[] inputPaths,
Path outputPath,
int numMapTasks,
Shard[] shards)
|
Constructors in org.apache.hadoop.contrib.index.mapred with parameters of type Configuration | |
---|---|
IndexUpdateConfiguration(Configuration conf)
Constructor |
Uses of Configuration in org.apache.hadoop.examples |
---|
Constructors in org.apache.hadoop.examples with parameters of type Configuration | |
---|---|
MultiFileWordCount.MultiFileLineRecordReader(Configuration conf,
MultiFileSplit split)
|
Uses of Configuration in org.apache.hadoop.filecache |
---|
Methods in org.apache.hadoop.filecache with parameters of type Configuration | |
---|---|
static void |
DistributedCache.addArchiveToClassPath(Path archive,
Configuration conf)
Deprecated. Please use DistributedCache.addArchiveToClassPath(Path, Configuration, FileSystem)
instead. The FileSystem should be obtained within an
appropriate doAs . |
static void |
DistributedCache.addArchiveToClassPath(Path archive,
Configuration conf,
FileSystem fs)
Add an archive path to the current set of classpath entries. |
static void |
DistributedCache.addCacheArchive(URI uri,
Configuration conf)
Add a archives to be localized to the conf. |
static void |
DistributedCache.addCacheFile(URI uri,
Configuration conf)
Add a file to be localized to the conf. |
static void |
DistributedCache.addFileToClassPath(Path file,
Configuration conf)
Deprecated. Please use DistributedCache.addFileToClassPath(Path, Configuration, FileSystem)
instead. The FileSystem should be obtained within an
appropriate doAs . |
static void |
DistributedCache.addFileToClassPath(Path file,
Configuration conf,
FileSystem fs)
Add a file path to the current set of classpath entries. |
static void |
DistributedCache.addLocalArchives(Configuration conf,
String str)
Add a archive that has been localized to the conf. |
static void |
DistributedCache.addLocalFiles(Configuration conf,
String str)
Add a file that has been localized to the conf.. |
static void |
TrackerDistributedCacheManager.createAllSymlink(Configuration conf,
File jobCacheDir,
File workDir)
This method create symlinks for all files in a given dir in another directory. |
static void |
DistributedCache.createAllSymlink(Configuration conf,
File jobCacheDir,
File workDir)
Deprecated. Internal to MapReduce framework. Use DistributedCacheManager instead. |
static void |
DistributedCache.createSymlink(Configuration conf)
This method allows you to create symlinks in the current working directory of the task to all the cache files/archives. |
static void |
TrackerDistributedCacheManager.determineCacheVisibilities(Configuration job)
Determines the visibilities of the distributed cache files and archives. |
static void |
TrackerDistributedCacheManager.determineTimestamps(Configuration job)
Determines timestamps of files to be cached, and stores those in the configuration. |
static long |
TrackerDistributedCacheManager.downloadCacheObject(Configuration conf,
URI source,
Path destination,
long desiredTimestamp,
boolean isArchive,
FsPermission permission)
Download a given path to the local file system. |
static Path[] |
DistributedCache.getArchiveClassPaths(Configuration conf)
Get the archive entries in classpath as an array of Path. |
static long[] |
DistributedCache.getArchiveTimestamps(Configuration conf)
Get the timestamps of the archives. |
static boolean[] |
TrackerDistributedCacheManager.getArchiveVisibilities(Configuration conf)
Get the booleans on whether the archives are public or not. |
static URI[] |
DistributedCache.getCacheArchives(Configuration conf)
Get cache archives set in the Configuration. |
static URI[] |
DistributedCache.getCacheFiles(Configuration conf)
Get cache files set in the Configuration. |
static void |
TrackerDistributedCacheManager.getDelegationTokens(Configuration job,
Credentials credentials)
For each archive or cache file - get the corresponding delegation token |
static Path[] |
DistributedCache.getFileClassPaths(Configuration conf)
Get the file entries in classpath as an array of Path. |
static FileStatus |
DistributedCache.getFileStatus(Configuration conf,
URI cache)
Returns FileStatus of a given cache file on hdfs. |
static long[] |
DistributedCache.getFileTimestamps(Configuration conf)
Get the timestamps of the files. |
static boolean[] |
TrackerDistributedCacheManager.getFileVisibilities(Configuration conf)
Get the booleans on whether the files are public or not. |
static Path[] |
DistributedCache.getLocalCacheArchives(Configuration conf)
Return the path array of the localized caches. |
static Path[] |
DistributedCache.getLocalCacheFiles(Configuration conf)
Return the path array of the localized files. |
static boolean |
DistributedCache.getSymlink(Configuration conf)
This method checks to see if symlinks are to be create for the localized cache files in the current working directory Used by internal DistributedCache code. |
static long |
DistributedCache.getTimestamp(Configuration conf,
URI cache)
Returns mtime of a given cache file on hdfs. |
TaskDistributedCacheManager |
TrackerDistributedCacheManager.newTaskDistributedCacheManager(JobID jobId,
Configuration taskConf)
|
static void |
DistributedCache.setArchiveTimestamps(Configuration conf,
String timestamps)
This is to check the timestamp of the archives to be localized. |
static void |
DistributedCache.setCacheArchives(URI[] archives,
Configuration conf)
Set the configuration with the given set of archives. |
static void |
DistributedCache.setCacheFiles(URI[] files,
Configuration conf)
Set the configuration with the given set of files. |
static void |
DistributedCache.setFileTimestamps(Configuration conf,
String timestamps)
This is to check the timestamp of the files to be localized. |
static void |
DistributedCache.setLocalArchives(Configuration conf,
String str)
Set the conf to contain the location for localized archives. |
static void |
DistributedCache.setLocalFiles(Configuration conf,
String str)
Set the conf to contain the location for localized files. |
void |
TaskDistributedCacheManager.setupCache(Configuration taskConf,
String publicCacheSubdir,
String privateCacheSubdir)
Retrieve public distributed cache files into the local cache and updates the task configuration (which has been passed in via the constructor). |
static void |
TrackerDistributedCacheManager.validate(Configuration conf)
This is part of the framework API. |
Constructors in org.apache.hadoop.filecache with parameters of type Configuration | |
---|---|
TrackerDistributedCacheManager.CleanupThread(Configuration conf)
|
|
TrackerDistributedCacheManager(Configuration conf,
TaskController controller)
|
Uses of Configuration in org.apache.hadoop.fs |
---|
Methods in org.apache.hadoop.fs that return Configuration | |
---|---|
Configuration |
FilterFileSystem.getConf()
|
Methods in org.apache.hadoop.fs with parameters of type Configuration | |
---|---|
static void |
FileSystem.addFileSystemForTesting(URI uri,
Configuration conf,
FileSystem fs)
This method adds a file system for testing so that we can find it later. |
static boolean |
FileUtil.copy(File src,
FileSystem dstFS,
Path dst,
boolean deleteSource,
Configuration conf)
Copy local files to a FileSystem. |
static boolean |
FileUtil.copy(FileSystem srcFS,
Path[] srcs,
FileSystem dstFS,
Path dst,
boolean deleteSource,
boolean overwrite,
Configuration conf)
|
static boolean |
FileUtil.copy(FileSystem srcFS,
Path src,
File dst,
boolean deleteSource,
Configuration conf)
Copy FileSystem files to local files. |
static boolean |
FileUtil.copy(FileSystem srcFS,
Path src,
FileSystem dstFS,
Path dst,
boolean deleteSource,
boolean overwrite,
Configuration conf)
Copy files between FileSystems. |
static boolean |
FileUtil.copy(FileSystem srcFS,
Path src,
FileSystem dstFS,
Path dst,
boolean deleteSource,
Configuration conf)
Copy files between FileSystems. |
static boolean |
FileUtil.copyMerge(FileSystem srcFS,
Path srcDir,
FileSystem dstFS,
Path dstFile,
boolean deleteSource,
Configuration conf,
String addString)
Copy all files in a directory to one output file (merge). |
File |
LocalDirAllocator.createTmpFileForWrite(String pathStr,
long size,
Configuration conf)
Creates a temporary file in the local FS. |
static FileSystem |
FileSystem.get(Configuration conf)
Returns the configured filesystem implementation. |
static FileSystem |
FileSystem.get(URI uri,
Configuration conf)
Returns the FileSystem for this URI's scheme and authority. |
static FileSystem |
FileSystem.get(URI uri,
Configuration conf,
String user)
|
Iterable<Path> |
LocalDirAllocator.getAllLocalPathsToRead(String pathStr,
Configuration conf)
Get all of the paths that currently exist in the working directories. |
static URI |
FileSystem.getDefaultUri(Configuration conf)
Get the default filesystem URI from a configuration. |
FileSystem |
Path.getFileSystem(Configuration conf)
Return the FileSystem that owns this Path. |
static LocalFileSystem |
FileSystem.getLocal(Configuration conf)
Get the local file syste |
Path |
LocalDirAllocator.getLocalPathForWrite(String pathStr,
Configuration conf)
Get a path from the local FS. |
Path |
LocalDirAllocator.getLocalPathForWrite(String pathStr,
long size,
Configuration conf)
Get a path from the local FS. |
Path |
LocalDirAllocator.getLocalPathForWrite(String pathStr,
long size,
Configuration conf,
boolean checkWrite)
Get a path from the local FS. |
Path |
LocalDirAllocator.getLocalPathToRead(String pathStr,
Configuration conf)
Get a path from the local FS for reading. |
static FileSystem |
FileSystem.getNamed(String name,
Configuration conf)
Deprecated. call #get(URI,Configuration) instead. |
boolean |
LocalDirAllocator.ifExists(String pathStr,
Configuration conf)
We search through all the configured dirs for the file's existence and return true when we find |
void |
HarFileSystem.initialize(URI name,
Configuration conf)
Initialize a Har filesystem per har archive. |
void |
FilterFileSystem.initialize(URI name,
Configuration conf)
Called after a new FileSystem instance is constructed. |
void |
RawLocalFileSystem.initialize(URI uri,
Configuration conf)
|
void |
FileSystem.initialize(URI name,
Configuration conf)
Called after a new FileSystem instance is constructed. |
void |
ChecksumFileSystem.setConf(Configuration conf)
|
static void |
FileSystem.setDefaultUri(Configuration conf,
String uri)
Set the default filesystem URI in a configuration. |
static void |
FileSystem.setDefaultUri(Configuration conf,
URI uri)
Set the default filesystem URI in a configuration. |
Constructors in org.apache.hadoop.fs with parameters of type Configuration | |
---|---|
DF(File path,
Configuration conf)
|
|
DU(File path,
Configuration conf)
Keeps track of disk usage. |
|
FsShell(Configuration conf)
|
|
FsUrlStreamHandlerFactory(Configuration conf)
|
|
InMemoryFileSystem(URI uri,
Configuration conf)
Deprecated. |
|
Trash(Configuration conf)
Construct a trash can accessor. |
|
Trash(FileSystem fs,
Configuration conf)
Construct a trash can accessor for the FileSystem provided. |
Uses of Configuration in org.apache.hadoop.fs.ftp |
---|
Methods in org.apache.hadoop.fs.ftp with parameters of type Configuration | |
---|---|
void |
FTPFileSystem.initialize(URI uri,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.fs.kfs |
---|
Methods in org.apache.hadoop.fs.kfs with parameters of type Configuration | |
---|---|
void |
KosmosFileSystem.initialize(URI uri,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.fs.permission |
---|
Methods in org.apache.hadoop.fs.permission with parameters of type Configuration | |
---|---|
static FsPermission |
FsPermission.getUMask(Configuration conf)
Get the user file creation mask (umask) UMASK_LABEL config param has umask value that is either symbolic
or octal. |
static void |
FsPermission.setUMask(Configuration conf,
FsPermission umask)
Set the user file creation mask (umask) |
Uses of Configuration in org.apache.hadoop.fs.s3 |
---|
Methods in org.apache.hadoop.fs.s3 with parameters of type Configuration | |
---|---|
void |
S3FileSystem.initialize(URI uri,
Configuration conf)
|
void |
S3Credentials.initialize(URI uri,
Configuration conf)
|
void |
FileSystemStore.initialize(URI uri,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.fs.s3native |
---|
Methods in org.apache.hadoop.fs.s3native with parameters of type Configuration | |
---|---|
void |
NativeS3FileSystem.initialize(URI uri,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.fs.shell |
---|
Constructors in org.apache.hadoop.fs.shell with parameters of type Configuration | |
---|---|
Command(Configuration conf)
Constructor |
|
Count(String[] cmd,
int pos,
Configuration conf)
Constructor |
Uses of Configuration in org.apache.hadoop.http |
---|
Methods in org.apache.hadoop.http with parameters of type Configuration | |
---|---|
void |
HttpServer.addSslListener(InetSocketAddress addr,
Configuration sslConf,
boolean needClientAuth)
Configure an ssl listener on the server. |
void |
HttpServer.addSslListener(InetSocketAddress addr,
Configuration sslConf,
boolean needCertsAuth,
boolean needKrbAuth)
Configure an ssl listener on the server. |
org.mortbay.jetty.Connector |
HttpServer.createBaseListener(Configuration conf)
Create a required listener for the Jetty instance listening on the port provided. |
abstract void |
FilterInitializer.initFilter(FilterContainer container,
Configuration conf)
Initialize a Filter to a FilterContainer. |
Constructors in org.apache.hadoop.http with parameters of type Configuration | |
---|---|
HttpServer(String name,
String bindAddress,
int port,
boolean findPort,
Configuration conf)
|
|
HttpServer(String name,
String bindAddress,
int port,
boolean findPort,
Configuration conf,
AccessControlList adminsAcl)
Create a status server on the given port. |
|
HttpServer(String name,
String bindAddress,
int port,
boolean findPort,
Configuration conf,
AccessControlList adminsAcl,
org.mortbay.jetty.Connector connector)
|
|
HttpServer(String name,
String bindAddress,
int port,
boolean findPort,
Configuration conf,
org.mortbay.jetty.Connector connector)
|
Uses of Configuration in org.apache.hadoop.http.lib |
---|
Methods in org.apache.hadoop.http.lib with parameters of type Configuration | |
---|---|
void |
StaticUserWebFilter.initFilter(FilterContainer container,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.io |
---|
Methods in org.apache.hadoop.io that return Configuration | |
---|---|
Configuration |
AbstractMapWritable.getConf()
|
Configuration |
ObjectWritable.getConf()
|
Configuration |
GenericWritable.getConf()
|
Methods in org.apache.hadoop.io with parameters of type Configuration | ||
---|---|---|
static
|
WritableUtils.clone(T orig,
Configuration conf)
Make a copy of a writable object using serialization to a buffer. |
|
static void |
IOUtils.copyBytes(InputStream in,
OutputStream out,
Configuration conf)
Copies from one stream to another. |
|
static void |
IOUtils.copyBytes(InputStream in,
OutputStream out,
Configuration conf,
boolean close)
Copies from one stream to another. |
|
protected SequenceFile.Reader |
MapFile.Reader.createDataFileReader(FileSystem fs,
Path dataFile,
Configuration conf)
Override this method to specialize the type of SequenceFile.Reader returned. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(Configuration conf,
FSDataOutputStream out,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
CompressionCodec codec)
Construct the preferred type of 'raw' SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(Configuration conf,
FSDataOutputStream out,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
CompressionCodec codec,
SequenceFile.Metadata metadata)
Construct the preferred type of 'raw' SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
int bufferSize,
short replication,
long blockSize,
boolean createParent,
SequenceFile.CompressionType compressionType,
CompressionCodec codec,
SequenceFile.Metadata metadata)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
int bufferSize,
short replication,
long blockSize,
SequenceFile.CompressionType compressionType,
CompressionCodec codec,
Progressable progress,
SequenceFile.Metadata metadata)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
CompressionCodec codec)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
CompressionCodec codec,
Progressable progress)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
CompressionCodec codec,
Progressable progress,
SequenceFile.Metadata metadata)
Construct the preferred type of SequenceFile Writer. |
|
static SequenceFile.Writer |
SequenceFile.createWriter(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
SequenceFile.CompressionType compressionType,
Progressable progress)
Construct the preferred type of SequenceFile Writer. |
|
static long |
MapFile.fix(FileSystem fs,
Path dir,
Class<? extends Writable> keyClass,
Class<? extends Writable> valueClass,
boolean dryrun,
Configuration conf)
This method attempts to fix a corrupt MapFile by re-creating its index. |
|
static Class<?> |
WritableName.getClass(String name,
Configuration conf)
Return the class for a name. |
|
static SequenceFile.CompressionType |
SequenceFile.getCompressionType(Configuration job)
Deprecated. Use SequenceFileOutputFormat.getOutputCompressionType(org.apache.hadoop.mapred.JobConf)
to get SequenceFile.CompressionType for job-outputs. |
|
static
|
DefaultStringifier.load(Configuration conf,
String keyName,
Class<K> itemClass)
Restores the object from the configuration. |
|
static
|
DefaultStringifier.loadArray(Configuration conf,
String keyName,
Class<K> itemClass)
Restores the array of objects from the configuration. |
|
static Writable |
WritableFactories.newInstance(Class<? extends Writable> c,
Configuration conf)
Create a new instance of a class with a defined factory. |
|
protected void |
MapFile.Reader.open(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf)
|
|
static Object |
ObjectWritable.readObject(DataInput in,
Configuration conf)
Read a Writable , String , primitive type, or an array of
the preceding. |
|
static Object |
ObjectWritable.readObject(DataInput in,
ObjectWritable objectWritable,
Configuration conf)
Read a Writable , String , primitive type, or an array of
the preceding. |
|
static void |
SequenceFile.setCompressionType(Configuration job,
SequenceFile.CompressionType val)
Deprecated. Use the one of the many SequenceFile.createWriter methods to specify the SequenceFile.CompressionType while creating the SequenceFile or
SequenceFileOutputFormat.setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)
to specify the SequenceFile.CompressionType for job-outputs.
or |
|
void |
AbstractMapWritable.setConf(Configuration conf)
|
|
void |
ObjectWritable.setConf(Configuration conf)
|
|
void |
GenericWritable.setConf(Configuration conf)
|
|
static void |
MapFile.Writer.setIndexInterval(Configuration conf,
int interval)
Sets the index interval and stores it in conf |
|
static
|
DefaultStringifier.store(Configuration conf,
K item,
String keyName)
Stores the item in the configuration with the given keyName. |
|
static
|
DefaultStringifier.storeArray(Configuration conf,
K[] items,
String keyName)
Stores the array of items in the configuration with the given keyName. |
|
static void |
ObjectWritable.writeObject(DataOutput out,
Object instance,
Class declaredClass,
Configuration conf)
Write a Writable , String , primitive type, or an array of
the preceding. |
Constructors in org.apache.hadoop.io with parameters of type Configuration | |
---|---|
ArrayFile.Reader(FileSystem fs,
String file,
Configuration conf)
Construct an array reader for the named file. |
|
ArrayFile.Writer(Configuration conf,
FileSystem fs,
String file,
Class<? extends Writable> valClass)
Create the named file for values of the named class. |
|
ArrayFile.Writer(Configuration conf,
FileSystem fs,
String file,
Class<? extends Writable> valClass,
SequenceFile.CompressionType compress,
Progressable progress)
Create the named file for values of the named class. |
|
BloomMapFile.Reader(FileSystem fs,
String dirName,
Configuration conf)
|
|
BloomMapFile.Reader(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf)
|
|
BloomMapFile.Reader(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf,
boolean open)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class<? extends Writable> valClass,
SequenceFile.CompressionType compress,
CompressionCodec codec,
Progressable progress)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass,
SequenceFile.CompressionType compress)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass,
SequenceFile.CompressionType compress,
Progressable progress)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress,
CompressionCodec codec,
Progressable progress)
|
|
BloomMapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress,
Progressable progress)
|
|
DefaultStringifier(Configuration conf,
Class<T> c)
|
|
MapFile.Reader(FileSystem fs,
String dirName,
Configuration conf)
Construct a map reader for the named map. |
|
MapFile.Reader(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf)
Construct a map reader for the named map using the named comparator. |
|
MapFile.Reader(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf,
boolean open)
Hook to allow subclasses to defer opening streams until further initialization is complete. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass)
Create the named map for keys of the named class. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass,
SequenceFile.CompressionType compress)
Create the named map for keys of the named class. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass,
SequenceFile.CompressionType compress,
CompressionCodec codec,
Progressable progress)
Create the named map for keys of the named class. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
Class valClass,
SequenceFile.CompressionType compress,
Progressable progress)
Create the named map for keys of the named class. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass)
Create the named map using the named key comparator. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress)
Create the named map using the named key comparator. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress,
CompressionCodec codec,
Progressable progress)
Create the named map using the named key comparator. |
|
MapFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
Class valClass,
SequenceFile.CompressionType compress,
Progressable progress)
Create the named map using the named key comparator. |
|
SequenceFile.Reader(FileSystem fs,
Path file,
Configuration conf)
Open the named file. |
|
SequenceFile.Sorter(FileSystem fs,
Class<? extends WritableComparable> keyClass,
Class valClass,
Configuration conf)
Sort and merge files containing the named classes. |
|
SequenceFile.Sorter(FileSystem fs,
RawComparator comparator,
Class keyClass,
Class valClass,
Configuration conf)
Sort and merge using an arbitrary RawComparator . |
|
SequenceFile.Writer(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass)
Create the named file. |
|
SequenceFile.Writer(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
int bufferSize,
short replication,
long blockSize,
Progressable progress,
SequenceFile.Metadata metadata)
Create the named file with write-progress reporter. |
|
SequenceFile.Writer(FileSystem fs,
Configuration conf,
Path name,
Class keyClass,
Class valClass,
Progressable progress,
SequenceFile.Metadata metadata)
Create the named file with write-progress reporter. |
|
SetFile.Reader(FileSystem fs,
String dirName,
Configuration conf)
Construct a set reader for the named set. |
|
SetFile.Reader(FileSystem fs,
String dirName,
WritableComparator comparator,
Configuration conf)
Construct a set reader for the named set using the named comparator. |
|
SetFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
Class<? extends WritableComparable> keyClass,
SequenceFile.CompressionType compress)
Create a set naming the element class and compression type. |
|
SetFile.Writer(Configuration conf,
FileSystem fs,
String dirName,
WritableComparator comparator,
SequenceFile.CompressionType compress)
Create a set naming the element comparator and compression type. |
Uses of Configuration in org.apache.hadoop.io.compress |
---|
Methods in org.apache.hadoop.io.compress that return Configuration | |
---|---|
Configuration |
DefaultCodec.getConf()
|
Configuration |
SnappyCodec.getConf()
Return the configuration used by this object. |
Methods in org.apache.hadoop.io.compress with parameters of type Configuration | |
---|---|
static List<Class<? extends CompressionCodec>> |
CompressionCodecFactory.getCodecClasses(Configuration conf)
Get the list of codecs listed in the configuration |
static Compressor |
CodecPool.getCompressor(CompressionCodec codec,
Configuration conf)
Get a Compressor for the given CompressionCodec from the
pool or a new one. |
static boolean |
SnappyCodec.isNativeSnappyLoaded(Configuration conf)
Are the native snappy libraries loaded & initialized? |
void |
Compressor.reinit(Configuration conf)
Prepare the compressor to be used in a new stream with settings defined in the given Configuration |
static void |
CompressionCodecFactory.setCodecClasses(Configuration conf,
List<Class> classes)
Sets a list of codec classes in the configuration. |
void |
DefaultCodec.setConf(Configuration conf)
|
void |
SnappyCodec.setConf(Configuration conf)
Set the configuration to be used by this object. |
Constructors in org.apache.hadoop.io.compress with parameters of type Configuration | |
---|---|
CompressionCodecFactory(Configuration conf)
Find the codecs specified in the config value io.compression.codecs and register them. |
Uses of Configuration in org.apache.hadoop.io.compress.bzip2 |
---|
Methods in org.apache.hadoop.io.compress.bzip2 with parameters of type Configuration | |
---|---|
void |
BZip2DummyCompressor.reinit(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.io.compress.snappy |
---|
Methods in org.apache.hadoop.io.compress.snappy with parameters of type Configuration | |
---|---|
void |
SnappyCompressor.reinit(Configuration conf)
Prepare the compressor to be used in a new stream with settings defined in the given Configuration |
Uses of Configuration in org.apache.hadoop.io.compress.zlib |
---|
Methods in org.apache.hadoop.io.compress.zlib with parameters of type Configuration | |
---|---|
static ZlibCompressor.CompressionLevel |
ZlibFactory.getCompressionLevel(Configuration conf)
|
static ZlibCompressor.CompressionStrategy |
ZlibFactory.getCompressionStrategy(Configuration conf)
|
static Compressor |
ZlibFactory.getZlibCompressor(Configuration conf)
Return the appropriate implementation of the zlib compressor. |
static Class<? extends Compressor> |
ZlibFactory.getZlibCompressorType(Configuration conf)
Return the appropriate type of the zlib compressor. |
static Decompressor |
ZlibFactory.getZlibDecompressor(Configuration conf)
Return the appropriate implementation of the zlib decompressor. |
static Class<? extends Decompressor> |
ZlibFactory.getZlibDecompressorType(Configuration conf)
Return the appropriate type of the zlib decompressor. |
static boolean |
ZlibFactory.isNativeZlibLoaded(Configuration conf)
Check if native-zlib code is loaded & initialized correctly and can be loaded for this job. |
void |
BuiltInZlibDeflater.reinit(Configuration conf)
reinit the compressor with the given configuration. |
void |
ZlibCompressor.reinit(Configuration conf)
Prepare the compressor to be used in a new stream with settings defined in the given Configuration. |
static void |
ZlibFactory.setCompressionLevel(Configuration conf,
ZlibCompressor.CompressionLevel level)
|
static void |
ZlibFactory.setCompressionStrategy(Configuration conf,
ZlibCompressor.CompressionStrategy strategy)
|
Constructors in org.apache.hadoop.io.compress.zlib with parameters of type Configuration | |
---|---|
ZlibCompressor(Configuration conf)
Creates a new compressor, taking settings from the configuration. |
Uses of Configuration in org.apache.hadoop.io.file.tfile |
---|
Constructors in org.apache.hadoop.io.file.tfile with parameters of type Configuration | |
---|---|
TFile.Reader(FSDataInputStream fsdis,
long fileLength,
Configuration conf)
Constructor |
|
TFile.Writer(FSDataOutputStream fsdos,
int minBlockSize,
String compressName,
String comparator,
Configuration conf)
Constructor |
Uses of Configuration in org.apache.hadoop.io.serializer |
---|
Constructors in org.apache.hadoop.io.serializer with parameters of type Configuration | |
---|---|
SerializationFactory(Configuration conf)
Serializations are found by reading the io.serializations
property from conf , which is a comma-delimited list of
classnames. |
Uses of Configuration in org.apache.hadoop.ipc |
---|
Methods in org.apache.hadoop.ipc with parameters of type Configuration | |
---|---|
static Object[] |
RPC.call(Method method,
Object[][] params,
InetSocketAddress[] addrs,
Configuration conf)
Deprecated. Use RPC.call(Method, Object[][], InetSocketAddress[], UserGroupInformation, Configuration) instead |
static Object[] |
RPC.call(Method method,
Object[][] params,
InetSocketAddress[] addrs,
UserGroupInformation ticket,
Configuration conf)
Expert: Make multiple, parallel calls to a set of servers. |
Writable[] |
Client.call(Writable[] params,
InetSocketAddress[] addresses,
Class<?> protocol,
UserGroupInformation ticket,
Configuration conf)
Makes a set of calls in parallel. |
Writable |
Client.call(Writable param,
InetSocketAddress addr,
Class<?> protocol,
UserGroupInformation ticket,
int rpcTimeout,
Configuration conf)
Make a call, passing param , to the IPC server running at
address which is servicing the protocol protocol,
with the ticket credentials, rpcTimeout as timeout
and conf as configuration for this connection, returning the
value. |
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
Configuration conf)
Construct a client-side proxy object with the default SocketFactory |
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
Configuration conf,
int rpcTimeout)
|
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
Configuration conf,
SocketFactory factory)
Construct a client-side proxy object that implements the named protocol, talking to a server at the named address. |
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
Configuration conf,
SocketFactory factory,
int rpcTimeout)
Construct a client-side proxy object that implements the named protocol, talking to a server at the named address. |
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
UserGroupInformation ticket,
Configuration conf,
SocketFactory factory)
Construct a client-side proxy object that implements the named protocol, talking to a server at the named address. |
static VersionedProtocol |
RPC.getProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
UserGroupInformation ticket,
Configuration conf,
SocketFactory factory,
int rpcTimeout)
Construct a client-side proxy object that implements the named protocol, talking to a server at the named address. |
static RPC.Server |
RPC.getServer(Object instance,
String bindAddress,
int port,
Configuration conf)
Construct a server for a protocol implementation instance listening on a port and address. |
static RPC.Server |
RPC.getServer(Object instance,
String bindAddress,
int port,
int numHandlers,
boolean verbose,
Configuration conf)
Construct a server for a protocol implementation instance listening on a port and address. |
static RPC.Server |
RPC.getServer(Object instance,
String bindAddress,
int port,
int numHandlers,
boolean verbose,
Configuration conf,
SecretManager<? extends TokenIdentifier> secretManager)
Construct a server for a protocol implementation instance listening on a port and address, with a secret manager. |
static void |
Client.setPingInterval(Configuration conf,
int pingInterval)
set the ping interval value in configuration |
static VersionedProtocol |
RPC.waitForProxy(Class<? extends VersionedProtocol> protocol,
long clientVersion,
InetSocketAddress addr,
Configuration conf)
|
Constructors in org.apache.hadoop.ipc with parameters of type Configuration | |
---|---|
Client(Class<? extends Writable> valueClass,
Configuration conf)
Construct an IPC client with the default SocketFactory |
|
Client(Class<? extends Writable> valueClass,
Configuration conf,
SocketFactory factory)
Construct an IPC client whose values are of the given Writable
class. |
|
RPC.Server(Object instance,
Configuration conf,
String bindAddress,
int port)
Construct an RPC server. |
|
RPC.Server(Object instance,
Configuration conf,
String bindAddress,
int port,
int numHandlers,
boolean verbose,
SecretManager<? extends TokenIdentifier> secretManager)
Construct an RPC server. |
|
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf)
|
|
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf,
String serverName)
|
|
Server(String bindAddress,
int port,
Class<? extends Writable> paramClass,
int handlerCount,
Configuration conf,
String serverName,
SecretManager<? extends TokenIdentifier> secretManager)
Constructs a server listening on the named port and address. |
Uses of Configuration in org.apache.hadoop.mapred |
---|
Subclasses of Configuration in org.apache.hadoop.mapred | |
---|---|
class |
JobConf
A map/reduce job configuration. |
Fields in org.apache.hadoop.mapred declared as Configuration | |
---|---|
protected Configuration |
SequenceFileRecordReader.conf
|
Methods in org.apache.hadoop.mapred that return Configuration | |
---|---|
Configuration |
TaskController.getConf()
|
Configuration |
SequenceFileInputFilter.FilterBase.getConf()
|
Configuration |
Task.getConf()
|
Methods in org.apache.hadoop.mapred with parameters of type Configuration | ||
---|---|---|
void |
UserLogCleaner.addOldUserLogsForDeletion(File loc,
Configuration conf)
Adds the job log directories for deletion with default retain hours. |
|
void |
JobClient.Renewer.cancel(Token<?> token,
Configuration conf)
|
|
void |
UserLogCleaner.clearOldUserLogs(Configuration conf)
Clears all the logs in userlogs directory. |
|
protected static
|
Task.createReduceContext(Reducer<INKEY,INVALUE,OUTKEY,OUTVALUE> reducer,
Configuration job,
TaskAttemptID taskId,
RawKeyValueIterator rIter,
Counter inputKeyCounter,
Counter inputValueCounter,
RecordWriter<OUTKEY,OUTVALUE> output,
OutputCommitter committer,
StatusReporter reporter,
RawComparator<INKEY> comparator,
Class<INKEY> keyClass,
Class<INVALUE> valueClass)
|
|
static long[] |
JobLocalizer.downloadPrivateCache(Configuration conf)
Download the parts of the distributed cache that are private. |
|
static InetSocketAddress |
JobTracker.getAddress(Configuration conf)
|
|
static int |
SkipBadRecords.getAttemptsToStartSkipping(Configuration conf)
Get the number of Task attempts AFTER which skip mode will be kicked off. |
|
static boolean |
SkipBadRecords.getAutoIncrMapperProcCount(Configuration conf)
Get the flag which if set to true, SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS is incremented
by MapRunner after invoking the map function. |
|
static boolean |
SkipBadRecords.getAutoIncrReducerProcCount(Configuration conf)
Get the flag which if set to true, SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS is incremented
by framework after invoking the reduce function. |
|
protected static FileSystem.Statistics |
Task.getFsStatistics(Path path,
Configuration conf)
Gets a handle to the Statistics instance based on the scheme associated with path. |
|
static long |
SkipBadRecords.getMapperMaxSkipRecords(Configuration conf)
Get the number of acceptable skip records surrounding the bad record PER bad record in mapper. |
|
static SequenceFile.Reader[] |
SequenceFileOutputFormat.getReaders(Configuration conf,
Path dir)
Open the output generated by this format. |
|
static MapFile.Reader[] |
MapFileOutputFormat.getReaders(FileSystem ignored,
Path dir,
Configuration conf)
Open the output generated by this format. |
|
static long |
SkipBadRecords.getReducerMaxSkipGroups(Configuration conf)
Get the number of acceptable skip groups surrounding the bad group PER bad group in reducer. |
|
static Path |
SkipBadRecords.getSkipOutputPath(Configuration conf)
Get the directory to which skipped records are written. |
|
long |
JobClient.Renewer.renew(Token<?> token,
Configuration conf)
|
|
static void |
SkipBadRecords.setAttemptsToStartSkipping(Configuration conf,
int attemptsToStartSkipping)
Set the number of Task attempts AFTER which skip mode will be kicked off. |
|
static void |
SkipBadRecords.setAutoIncrMapperProcCount(Configuration conf,
boolean autoIncr)
Set the flag which if set to true, SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS is incremented
by MapRunner after invoking the map function. |
|
static void |
SkipBadRecords.setAutoIncrReducerProcCount(Configuration conf,
boolean autoIncr)
Set the flag which if set to true, SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS is incremented
by framework after invoking the reduce function. |
|
void |
TaskController.setConf(Configuration conf)
|
|
void |
SequenceFileInputFilter.RegexFilter.setConf(Configuration conf)
configure the Filter by checking the configuration |
|
void |
SequenceFileInputFilter.PercentFilter.setConf(Configuration conf)
configure the filter by checking the configuration |
|
void |
SequenceFileInputFilter.MD5Filter.setConf(Configuration conf)
configure the filter according to configuration |
|
void |
DefaultTaskController.setConf(Configuration conf)
|
|
void |
Task.setConf(Configuration conf)
|
|
static void |
SequenceFileInputFilter.setFilterClass(Configuration conf,
Class filterClass)
set the filter class |
|
static void |
SequenceFileInputFilter.PercentFilter.setFrequency(Configuration conf,
int frequency)
set the frequency and stores it in conf |
|
static void |
SequenceFileInputFilter.MD5Filter.setFrequency(Configuration conf,
int frequency)
set the filtering frequency in configuration |
|
static void |
SkipBadRecords.setMapperMaxSkipRecords(Configuration conf,
long maxSkipRecs)
Set the number of acceptable skip records surrounding the bad record PER bad record in mapper. |
|
static void |
SequenceFileInputFilter.RegexFilter.setPattern(Configuration conf,
String regex)
Define the filtering regex and stores it in conf |
|
static void |
SkipBadRecords.setReducerMaxSkipGroups(Configuration conf,
long maxSkipGrps)
Set the number of acceptable skip groups surrounding the bad group PER bad group in reducer. |
Uses of Configuration in org.apache.hadoop.mapred.join |
---|
Methods in org.apache.hadoop.mapred.join that return Configuration | |
---|---|
Configuration |
CompositeRecordReader.getConf()
Return the configuration used by this object. |
Methods in org.apache.hadoop.mapred.join with parameters of type Configuration | |
---|---|
void |
CompositeRecordReader.setConf(Configuration conf)
Set the configuration to be used by this object. |
Uses of Configuration in org.apache.hadoop.mapred.lib |
---|
Methods in org.apache.hadoop.mapred.lib that return Configuration | |
---|---|
Configuration |
InputSampler.getConf()
|
Methods in org.apache.hadoop.mapred.lib with parameters of type Configuration | |
---|---|
void |
InputSampler.setConf(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapred.pipes |
---|
Constructors in org.apache.hadoop.mapred.pipes with parameters of type Configuration | |
---|---|
Submitter(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapred.tools |
---|
Constructors in org.apache.hadoop.mapred.tools with parameters of type Configuration | |
---|---|
MRAdmin(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapreduce |
---|
Methods in org.apache.hadoop.mapreduce that return Configuration | |
---|---|
Configuration |
JobContext.getConfiguration()
Return the configuration for the job. |
Methods in org.apache.hadoop.mapreduce with parameters of type Configuration | |
---|---|
static Path |
JobSubmissionFiles.getStagingDir(JobClient client,
Configuration conf)
Initializes the staging directory and returns the path. |
Uses of Configuration in org.apache.hadoop.mapreduce.lib.db |
---|
Methods in org.apache.hadoop.mapreduce.lib.db that return Configuration | |
---|---|
Configuration |
DBInputFormat.getConf()
|
Configuration |
DBConfiguration.getConf()
|
Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type Configuration | |
---|---|
static void |
DBConfiguration.configureDB(Configuration job,
String driverClass,
String dbUrl)
Sets the DB access related fields in the JobConf. |
static void |
DBConfiguration.configureDB(Configuration conf,
String driverClass,
String dbUrl,
String userName,
String passwd)
Sets the DB access related fields in the Configuration . |
protected RecordReader<LongWritable,T> |
DBInputFormat.createDBRecordReader(DBInputFormat.DBInputSplit split,
Configuration conf)
|
protected RecordReader<LongWritable,T> |
DataDrivenDBInputFormat.createDBRecordReader(DBInputFormat.DBInputSplit split,
Configuration conf)
|
protected RecordReader<LongWritable,T> |
OracleDataDrivenDBInputFormat.createDBRecordReader(DBInputFormat.DBInputSplit split,
Configuration conf)
|
static void |
DataDrivenDBInputFormat.setBoundingQuery(Configuration conf,
String query)
Set the user-defined bounding query to use with a user-defined query. |
void |
DBInputFormat.setConf(Configuration conf)
Set the configuration to be used by this object. |
static void |
OracleDBRecordReader.setSessionTimeZone(Configuration conf,
Connection conn)
Set session time zone |
List<InputSplit> |
DBSplitter.split(Configuration conf,
ResultSet results,
String colName)
Given a ResultSet containing one record (and already advanced to that record) with two columns (a low value, and a high value, both of the same type), determine a set of splits that span the given values. |
List<InputSplit> |
BooleanSplitter.split(Configuration conf,
ResultSet results,
String colName)
|
List<InputSplit> |
BigDecimalSplitter.split(Configuration conf,
ResultSet results,
String colName)
|
List<InputSplit> |
DateSplitter.split(Configuration conf,
ResultSet results,
String colName)
|
List<InputSplit> |
IntegerSplitter.split(Configuration conf,
ResultSet results,
String colName)
|
List<InputSplit> |
TextSplitter.split(Configuration conf,
ResultSet results,
String colName)
This method needs to determine the splits between two user-provided strings. |
List<InputSplit> |
FloatSplitter.split(Configuration conf,
ResultSet results,
String colName)
|
Uses of Configuration in org.apache.hadoop.mapreduce.lib.input |
---|
Fields in org.apache.hadoop.mapreduce.lib.input declared as Configuration | |
---|---|
protected Configuration |
SequenceFileRecordReader.conf
|
Methods in org.apache.hadoop.mapreduce.lib.input that return Configuration | |
---|---|
Configuration |
SequenceFileInputFilter.FilterBase.getConf()
|
Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type Configuration | |
---|---|
static List<FileSplit> |
NLineInputFormat.getSplitsForFile(FileStatus status,
Configuration conf,
int numLinesPerSplit)
|
void |
SequenceFileInputFilter.RegexFilter.setConf(Configuration conf)
configure the Filter by checking the configuration |
void |
SequenceFileInputFilter.PercentFilter.setConf(Configuration conf)
configure the filter by checking the configuration |
void |
SequenceFileInputFilter.MD5Filter.setConf(Configuration conf)
configure the filter according to configuration |
static void |
SequenceFileInputFilter.PercentFilter.setFrequency(Configuration conf,
int frequency)
set the frequency and stores it in conf |
static void |
SequenceFileInputFilter.MD5Filter.setFrequency(Configuration conf,
int frequency)
set the filtering frequency in configuration |
static void |
SequenceFileInputFilter.RegexFilter.setPattern(Configuration conf,
String regex)
Define the filtering regex and stores it in conf |
Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type Configuration | |
---|---|
KeyValueLineRecordReader(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapreduce.lib.jobcontrol |
---|
Constructors in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Configuration | |
---|---|
ControlledJob(Configuration conf)
Construct a job. |
Uses of Configuration in org.apache.hadoop.mapreduce.lib.partition |
---|
Methods in org.apache.hadoop.mapreduce.lib.partition that return Configuration | |
---|---|
Configuration |
BinaryPartitioner.getConf()
|
Configuration |
KeyFieldBasedComparator.getConf()
|
Configuration |
TotalOrderPartitioner.getConf()
|
Configuration |
KeyFieldBasedPartitioner.getConf()
|
Methods in org.apache.hadoop.mapreduce.lib.partition with parameters of type Configuration | |
---|---|
static String |
TotalOrderPartitioner.getPartitionFile(Configuration conf)
Get the path to the SequenceFile storing the sorted partition keyset. |
void |
BinaryPartitioner.setConf(Configuration conf)
|
void |
KeyFieldBasedComparator.setConf(Configuration conf)
|
void |
TotalOrderPartitioner.setConf(Configuration conf)
Read in the partition file and build indexing data structures. |
void |
KeyFieldBasedPartitioner.setConf(Configuration conf)
|
static void |
BinaryPartitioner.setLeftOffset(Configuration conf,
int offset)
Set the subarray to be used for partitioning to bytes[offset:] in Python syntax. |
static void |
BinaryPartitioner.setOffsets(Configuration conf,
int left,
int right)
Set the subarray to be used for partitioning to bytes[left:(right+1)] in Python syntax. |
static void |
TotalOrderPartitioner.setPartitionFile(Configuration conf,
Path p)
Set the path to the SequenceFile storing the sorted partition keyset. |
static void |
BinaryPartitioner.setRightOffset(Configuration conf,
int offset)
Set the subarray to be used for partitioning to bytes[:(offset+1)] in Python syntax. |
Constructors in org.apache.hadoop.mapreduce.lib.partition with parameters of type Configuration | |
---|---|
InputSampler(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapreduce.security |
---|
Methods in org.apache.hadoop.mapreduce.security with parameters of type Configuration | |
---|---|
static Credentials |
TokenCache.loadTokens(String jobTokenFile,
Configuration conf)
load job token from a file |
static void |
TokenCache.obtainTokensForNamenodes(Credentials credentials,
Path[] ps,
Configuration conf)
Convenience method to obtain delegation tokens from namenodes corresponding to the paths passed. |
Uses of Configuration in org.apache.hadoop.mapreduce.security.token |
---|
Methods in org.apache.hadoop.mapreduce.security.token with parameters of type Configuration | |
---|---|
static void |
DelegationTokenRenewal.registerDelegationTokensForRenewal(JobID jobId,
Credentials ts,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.mapreduce.server.tasktracker.userlogs |
---|
Methods in org.apache.hadoop.mapreduce.server.tasktracker.userlogs with parameters of type Configuration | |
---|---|
void |
UserLogManager.clearOldUserLogs(Configuration conf)
Called during TaskTracker restart/re-init. |
Constructors in org.apache.hadoop.mapreduce.server.tasktracker.userlogs with parameters of type Configuration | |
---|---|
UserLogManager(Configuration conf)
Create the user log manager to manage user logs on TaskTracker . |
|
UserLogManager(Configuration conf,
TaskController taskController)
Create the user log manager to manage user logs on TaskTracker . |
Uses of Configuration in org.apache.hadoop.mapreduce.split |
---|
Methods in org.apache.hadoop.mapreduce.split with parameters of type Configuration | ||
---|---|---|
static void |
JobSplitWriter.createSplitFiles(Path jobSubmitDir,
Configuration conf,
FileSystem fs,
InputSplit[] splits)
|
|
static
|
JobSplitWriter.createSplitFiles(Path jobSubmitDir,
Configuration conf,
FileSystem fs,
List<InputSplit> splits)
|
|
static
|
JobSplitWriter.createSplitFiles(Path jobSubmitDir,
Configuration conf,
FileSystem fs,
T[] splits)
|
|
static JobSplit.TaskSplitMetaInfo[] |
SplitMetaInfoReader.readSplitMetaInfo(JobID jobId,
FileSystem fs,
Configuration conf,
Path jobSubmitDir)
|
Uses of Configuration in org.apache.hadoop.net |
---|
Methods in org.apache.hadoop.net that return Configuration | |
---|---|
Configuration |
SocksSocketFactory.getConf()
|
Configuration |
ScriptBasedMapping.getConf()
|
Methods in org.apache.hadoop.net with parameters of type Configuration | |
---|---|
static SocketFactory |
NetUtils.getDefaultSocketFactory(Configuration conf)
Get the default socket factory as specified by the configuration parameter hadoop.rpc.socket.factory.default |
static String |
NetUtils.getServerAddress(Configuration conf,
String oldBindAddressName,
String oldPortName,
String newBindAddressName)
Deprecated. |
static SocketFactory |
NetUtils.getSocketFactory(Configuration conf,
Class<?> clazz)
Get the socket factory for the given class according to its configuration parameter hadoop.rpc.socket.factory.class.<ClassName>. |
static SocketFactory |
NetUtils.getSocketFactoryFromProperty(Configuration conf,
String propValue)
Get the socket factory corresponding to the given proxy URI. |
void |
SocksSocketFactory.setConf(Configuration conf)
|
void |
ScriptBasedMapping.setConf(Configuration conf)
|
Constructors in org.apache.hadoop.net with parameters of type Configuration | |
---|---|
ScriptBasedMapping(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.security |
---|
Methods in org.apache.hadoop.security with parameters of type Configuration | |
---|---|
static AccessControlList |
SecurityUtil.getAdminAcls(Configuration conf,
String configKey)
Get the ACL object representing the cluster administrators The user who starts the daemon is automatically added as an admin |
static Groups |
Groups.getUserToGroupsMappingService(Configuration conf)
|
static void |
SaslRpcServer.init(Configuration conf)
|
void |
AuthenticationFilterInitializer.initFilter(FilterContainer container,
Configuration conf)
Initializes Alfredo AuthenticationFilter. |
static void |
SecurityUtil.login(Configuration conf,
String keytabFileKey,
String userNameKey)
If a keytab has been provided, login as that user. |
static void |
SecurityUtil.login(Configuration conf,
String keytabFileKey,
String userNameKey,
String hostname)
If a keytab has been provided, login as that user. |
static Credentials |
Credentials.readTokenStorageFile(Path filename,
Configuration conf)
Convenience method for reading a token storage file, and loading the Tokens therein in the passed UGI |
static void |
KerberosName.setConfiguration(Configuration conf)
Set the static configuration to get the rules. |
static void |
UserGroupInformation.setConfiguration(Configuration conf)
Set the static configuration for UGI. |
void |
Credentials.writeTokenStorageFile(Path filename,
Configuration conf)
|
Constructors in org.apache.hadoop.security with parameters of type Configuration | |
---|---|
Groups(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.security.authorize |
---|
Methods in org.apache.hadoop.security.authorize with parameters of type Configuration | |
---|---|
static void |
ServiceAuthorizationManager.authorize(UserGroupInformation user,
Class<?> protocol,
Configuration conf,
InetAddress addr)
Authorize the user to access the protocol being used. |
static void |
ProxyUsers.authorize(UserGroupInformation user,
String remoteAddress,
Configuration newConf)
Authorize the superuser which is doing doAs |
static void |
ServiceAuthorizationManager.refresh(Configuration conf,
PolicyProvider provider)
|
static void |
ProxyUsers.refreshSuperUserGroupsConfiguration(Configuration conf)
refresh configuration |
Uses of Configuration in org.apache.hadoop.security.token |
---|
Methods in org.apache.hadoop.security.token with parameters of type Configuration | |
---|---|
void |
Token.cancel(Configuration conf)
Cancel this delegation token |
abstract void |
TokenRenewer.cancel(Token<?> token,
Configuration conf)
Cancel the given token |
void |
Token.TrivialRenewer.cancel(Token<?> token,
Configuration conf)
|
long |
Token.renew(Configuration conf)
Renew this delegation token |
abstract long |
TokenRenewer.renew(Token<?> token,
Configuration conf)
Renew the given token. |
long |
Token.TrivialRenewer.renew(Token<?> token,
Configuration conf)
|
Uses of Configuration in org.apache.hadoop.streaming |
---|
Fields in org.apache.hadoop.streaming declared as Configuration | |
---|---|
protected Configuration |
StreamJob.config_
|
Methods in org.apache.hadoop.streaming that return Configuration | |
---|---|
Configuration |
DumpTypedBytes.getConf()
|
Configuration |
LoadTypedBytes.getConf()
|
Configuration |
StreamJob.getConf()
|
Configuration |
PipeMapRed.getConfiguration()
Returns the Configuration. |
Methods in org.apache.hadoop.streaming with parameters of type Configuration | |
---|---|
static Class |
StreamUtil.goodClassOrNull(Configuration conf,
String className,
String defaultPackage)
It may seem strange to silently switch behaviour when a String is not a classname; the reason is simplified Usage: |
void |
DumpTypedBytes.setConf(Configuration conf)
|
void |
LoadTypedBytes.setConf(Configuration conf)
|
void |
StreamJob.setConf(Configuration conf)
|
Constructors in org.apache.hadoop.streaming with parameters of type Configuration | |
---|---|
DumpTypedBytes(Configuration conf)
|
|
LoadTypedBytes(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.typedbytes |
---|
Methods in org.apache.hadoop.typedbytes that return Configuration | |
---|---|
Configuration |
TypedBytesWritableInput.getConf()
|
Methods in org.apache.hadoop.typedbytes with parameters of type Configuration | |
---|---|
void |
TypedBytesWritableInput.setConf(Configuration conf)
|
Uses of Configuration in org.apache.hadoop.util |
---|
Methods in org.apache.hadoop.util that return Configuration | |
---|---|
Configuration |
GenericOptionsParser.getConfiguration()
Get the modified configuration |
Methods in org.apache.hadoop.util with parameters of type Configuration | ||
---|---|---|
static
|
ReflectionUtils.copy(Configuration conf,
T src,
T dst)
Make a copy of the writable object using serialization to a buffer |
|
static URL[] |
GenericOptionsParser.getLibJars(Configuration conf)
If libjars are set in the conf, parse the libjars. |
|
boolean |
NativeCodeLoader.getLoadNativeLibraries(Configuration conf)
Return if native hadoop libraries, if present, can be used for this job. |
|
static MemoryCalculatorPlugin |
MemoryCalculatorPlugin.getMemoryCalculatorPlugin(Class<? extends MemoryCalculatorPlugin> clazz,
Configuration conf)
Deprecated. Get the MemoryCalculatorPlugin from the class name and configure it. |
|
static ResourceCalculatorPlugin |
ResourceCalculatorPlugin.getResourceCalculatorPlugin(Class<? extends ResourceCalculatorPlugin> clazz,
Configuration conf)
Get the ResourceCalculatorPlugin from the class name and configure it. |
|
static String[] |
Shell.getUlimitMemoryCommand(Configuration conf)
Deprecated. Use Shell.getUlimitMemoryCommand(int) |
|
static
|
ReflectionUtils.newInstance(Class<T> theClass,
Configuration conf)
Create an object for the given class and initialize it from conf |
|
static int |
ToolRunner.run(Configuration conf,
Tool tool,
String[] args)
Runs the given Tool by Tool.run(String[]) , after
parsing with the given generic arguments. |
|
static void |
ReflectionUtils.setConf(Object theObject,
Configuration conf)
Check and set 'configuration' if necessary. |
|
void |
NativeCodeLoader.setLoadNativeLibraries(Configuration conf,
boolean loadNativeLibraries)
Set if native hadoop libraries, if present, can be used for this job. |
Constructors in org.apache.hadoop.util with parameters of type Configuration | |
---|---|
GenericOptionsParser(Configuration conf,
org.apache.commons.cli.Options options,
String[] args)
Create a GenericOptionsParser to parse given options as well
as generic Hadoop options. |
|
GenericOptionsParser(Configuration conf,
String[] args)
Create a GenericOptionsParser |
|
LineReader(InputStream in,
Configuration conf)
Create a line reader that reads from the given stream using the io.file.buffer.size specified in the given
Configuration . |
Uses of Configuration in org.apache.hadoop.util.hash |
---|
Methods in org.apache.hadoop.util.hash with parameters of type Configuration | |
---|---|
static int |
Hash.getHashType(Configuration conf)
This utility method converts the name of the configured hash type to a symbolic constant. |
static Hash |
Hash.getInstance(Configuration conf)
Get a singleton instance of hash function of a type defined in the configuration. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |