- cacheArchives - Variable in class org.apache.hadoop.streaming.StreamJob
-
- CachedDNSToSwitchMapping - Class in org.apache.hadoop.net
-
A cached implementation of DNSToSwitchMapping that takes an
raw DNSToSwitchMapping and stores the resolved network location in
a cache.
- CachedDNSToSwitchMapping(DNSToSwitchMapping) - Constructor for class org.apache.hadoop.net.CachedDNSToSwitchMapping
-
- cacheFiles - Variable in class org.apache.hadoop.streaming.StreamJob
-
- call(Writable, InetSocketAddress) - Method in class org.apache.hadoop.ipc.Client
-
- call(Writable, InetSocketAddress, UserGroupInformation) - Method in class org.apache.hadoop.ipc.Client
-
- call(Writable, InetSocketAddress, Class<?>, UserGroupInformation) - Method in class org.apache.hadoop.ipc.Client
-
Make a call, passing param
, to the IPC server running at
address
which is servicing the protocol
protocol,
with the ticket
credentials, returning the value.
- call(Writable[], InetSocketAddress[]) - Method in class org.apache.hadoop.ipc.Client
-
- call(Writable[], InetSocketAddress[], Class<?>, UserGroupInformation) - Method in class org.apache.hadoop.ipc.Client
-
Makes a set of calls in parallel.
- call(Method, Object[][], InetSocketAddress[], Configuration) - Static method in class org.apache.hadoop.ipc.RPC
-
- call(Method, Object[][], InetSocketAddress[], UserGroupInformation, Configuration) - Static method in class org.apache.hadoop.ipc.RPC
-
Expert: Make multiple, parallel calls to a set of servers.
- call(Class<?>, Writable, long) - Method in class org.apache.hadoop.ipc.RPC.Server
-
- call(Writable, long) - Method in class org.apache.hadoop.ipc.Server
-
- call(Class<?>, Writable, long) - Method in class org.apache.hadoop.ipc.Server
-
Called for each call.
- callQueueLen - Variable in class org.apache.hadoop.ipc.metrics.RpcMetrics
-
- canCommit(TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskTracker
-
Child checking whether it can commit
- captureDebugOut(List<String>, File) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Wrap a command in a shell to capture debug script's
stdout and stderr to debugout.
- captureOutAndError(List<String>, File, File, long) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Wrap a command in a shell to capture stdout and stderr to files.
- captureOutAndError(List<String>, List<String>, File, File, long) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Wrap a command in a shell to capture stdout and stderr to files.
- captureOutAndError(List<String>, List<String>, File, File, long, String) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Wrap a command in a shell to capture stdout and stderr to files.
- CBZip2InputStream - Class in org.apache.hadoop.io.compress.bzip2
-
An input stream that decompresses from the BZip2 format (without the file
header chars) to be read as any other stream.
- CBZip2InputStream(InputStream) - Constructor for class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
Constructs a new CBZip2InputStream which decompresses bytes read from the
specified stream.
- CBZip2OutputStream - Class in org.apache.hadoop.io.compress.bzip2
-
An output stream that compresses into the BZip2 format (without the file
header chars) into another stream.
- CBZip2OutputStream(OutputStream) - Constructor for class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
Constructs a new CBZip2OutputStream with a blocksize of 900k.
- CBZip2OutputStream(OutputStream, int) - Constructor for class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
Constructs a new CBZip2OutputStream with specified blocksize.
- ChainMapper - Class in org.apache.hadoop.mapred.lib
-
The ChainMapper class allows to use multiple Mapper classes within a single
Map task.
- ChainMapper() - Constructor for class org.apache.hadoop.mapred.lib.ChainMapper
-
Constructor.
- ChainReducer - Class in org.apache.hadoop.mapred.lib
-
The ChainReducer class allows to chain multiple Mapper classes after a
Reducer within the Reducer task.
- ChainReducer() - Constructor for class org.apache.hadoop.mapred.lib.ChainReducer
-
Constructor.
- charAt(int) - Method in class org.apache.hadoop.io.Text
-
Returns the Unicode Scalar Value (32-bit integer value)
for the character at position
.
- checkDir(File) - Static method in class org.apache.hadoop.util.DiskChecker
-
- checkExistence(String) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Checks whether a specific shell command is available
in the system.
- checkForRotation() - Method in class org.apache.hadoop.contrib.failmon.LogParser
-
Check whether the log file has been rotated.
- checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.FileOutputFormat
-
- checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
-
Check for validity of the output-specification for the job.
- checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.NullOutputFormat
-
Deprecated.
- checkOutputSpecs(FileSystem, JobConf) - Method in interface org.apache.hadoop.mapred.OutputFormat
-
Deprecated.
Check for validity of the output-specification for the job.
- checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
-
- checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
-
Check for validity of the output-specification for the job.
- checkPath(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Check that a Path belongs to this FileSystem.
- checkPath(Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Check that a Path belongs to this FileSystem.
- checkpoint() - Method in class org.apache.hadoop.fs.Trash
-
Create a trash checkpoint.
- checkStream() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- checksum2long(byte[]) - Static method in class org.apache.hadoop.fs.FSInputChecker
-
Convert a checksum byte array to a long
- CHECKSUM_CRC32 - Static variable in class org.apache.hadoop.util.DataChecksum
-
- CHECKSUM_NULL - Static variable in class org.apache.hadoop.util.DataChecksum
-
- ChecksumException - Exception in org.apache.hadoop.fs
-
Thrown for checksum errors.
- ChecksumException(String, long) - Constructor for exception org.apache.hadoop.fs.ChecksumException
-
- ChecksumFileSystem - Class in org.apache.hadoop.fs
-
Abstract Checksumed FileSystem.
- ChecksumFileSystem(FileSystem) - Constructor for class org.apache.hadoop.fs.ChecksumFileSystem
-
- checkURIs(URI[], URI[]) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This method checks if there is a conflict in the fragment names
of the uris.
- chmod(String, String) - Static method in class org.apache.hadoop.fs.FileUtil
-
Change the permissions on a filename.
- chooseBlockSize(long) - Static method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
Chooses a blocksize based on the given length of the data to compress.
- chooseRandom(String) - Method in class org.apache.hadoop.net.NetworkTopology
-
randomly choose one node from scope
if scope starts with ~, choose one from the all nodes except for the
ones in scope; otherwise, choose one from scope
- chooseShardForDelete(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
-
- chooseShardForDelete(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
-
- chooseShardForDelete(DocumentID) - Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
-
Choose a shard or all shards to send a delete request.
- chooseShardForInsert(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
-
- chooseShardForInsert(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
-
- chooseShardForInsert(DocumentID) - Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
-
Choose a shard to send an insert request.
- cleanup() - Method in class org.apache.hadoop.contrib.failmon.Executor
-
- cleanup() - Method in class org.apache.hadoop.contrib.failmon.RunOnce
-
- cleanup(Log, Closeable...) - Static method in class org.apache.hadoop.io.IOUtils
-
Close the Closeable objects and
ignore any
IOException
or
null pointers.
- cleanup() - Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
The default cleanup.
- cleanup(int) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Purge old user logs.
- cleanup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
-
Called once at the end of the task.
- cleanup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
-
Called once at the end of the task.
- cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
For cleaning up the job's output after job completion
- cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
This method implements the new interface by calling the old method.
- cleanupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
Delete the temporary directory, including all of the work directories.
- cleanupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
-
For cleaning up the job's output after job completion
- cleanupProgress() - Method in class org.apache.hadoop.mapred.JobStatus
-
- cleanupProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the progress of the job's cleanup-tasks, as a float between 0.0
and 1.0.
- cleanupStorage() - Method in class org.apache.hadoop.mapred.TaskTracker
-
Removes all contents of temporary storage.
- clear() - Method in class org.apache.hadoop.conf.Configuration
-
Clears all keys from the configuration.
- clear() - Method in class org.apache.hadoop.io.MapWritable
- clear() - Method in class org.apache.hadoop.io.SortedMapWritable
- clear() - Method in class org.apache.hadoop.io.Text
-
Clear the string to empty.
- clear() - Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- clear() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- clear() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- clear() - Method in interface org.apache.hadoop.mapred.join.ResetableIterator
-
Close datasources, but do not release internal resources.
- clear() - Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- clear() - Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- clear() - Method in class org.apache.hadoop.util.bloom.HashFunction
-
Clears this hash function.
- clear() - Method in class org.apache.hadoop.util.PriorityQueue
-
Removes all entries from the PriorityQueue.
- CLEARMASK - Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
This constant is accessible by subclasses for historical purposes.
- clearStatistics() - Static method in class org.apache.hadoop.fs.FileSystem
-
- Client - Class in org.apache.hadoop.ipc
-
A client for an IPC service.
- Client(Class<? extends Writable>, Configuration, SocketFactory) - Constructor for class org.apache.hadoop.ipc.Client
-
Construct an IPC client whose values are of the given
Writable
class.
- Client(Class<? extends Writable>, Configuration) - Constructor for class org.apache.hadoop.ipc.Client
-
Construct an IPC client with the default SocketFactory
- ClientTraceLog - Static variable in class org.apache.hadoop.mapred.TaskTracker
-
- clone(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- clone(T, Configuration) - Static method in class org.apache.hadoop.io.WritableUtils
-
Make a copy of a writable object using serialization to a buffer.
- clone() - Method in class org.apache.hadoop.mapred.JobStatus
-
- clone() - Method in class org.apache.hadoop.record.Buffer
-
- cloneFileAttributes(Path, Path, Progressable) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Clones the attributes (like compression of the input file and creates a
corresponding Writer
- cloneInto(Writable, Writable) - Static method in class org.apache.hadoop.io.WritableUtils
-
- cloneWritableInto(Writable, Writable) - Static method in class org.apache.hadoop.util.ReflectionUtils
-
Deprecated.
- close() - Method in class org.apache.hadoop.contrib.failmon.LocalStore
-
Close the temporary local file
- close() - Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
-
- close() - Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
-
- close() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- close() - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- close() - Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
-
Close the shard writer.
- close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
-
- close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
-
- close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
- close() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- close() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- close() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- close() - Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
-
- close() - Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- close() - Method in class org.apache.hadoop.examples.PiEstimator.PiReducer
-
Reduce task done, write output to a file.
- close() - Method in class org.apache.hadoop.examples.SleepJob
-
- close() - Method in class org.apache.hadoop.fs.FileSystem
-
No more filesystem operations are needed.
- close() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
- close() - Method in class org.apache.hadoop.fs.FSDataOutputStream
-
- close() - Method in class org.apache.hadoop.fs.FsShell
-
- close() - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- close() - Method in class org.apache.hadoop.fs.HarFileSystem
-
- close() - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- close() - Method in class org.apache.hadoop.io.BloomMapFile.Writer
-
- close() - Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
- close() - Method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
- close() - Method in class org.apache.hadoop.io.compress.CompressionInputStream
-
- close() - Method in class org.apache.hadoop.io.compress.CompressionOutputStream
-
- close() - Method in class org.apache.hadoop.io.compress.CompressorStream
-
- close() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- close() - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- close() - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- close() - Method in class org.apache.hadoop.io.DefaultStringifier
-
- close() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Close the reader.
- close() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Close the scanner.
- close() - Method in class org.apache.hadoop.io.file.tfile.TFile.Writer
-
Close the Writer.
- close() - Method in class org.apache.hadoop.io.MapFile.Reader
-
Close the map.
- close() - Method in class org.apache.hadoop.io.MapFile.Writer
-
Close the map.
- close() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Close the file.
- close() - Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
-
closes the iterator so that the underlying streams can be closed
- close() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
Close the file.
- close() - Method in interface org.apache.hadoop.io.serializer.Deserializer
-
Close the underlying input stream and clear up any resources.
- close() - Method in interface org.apache.hadoop.io.serializer.Serializer
-
Close the underlying output stream and clear up any resources.
- close() - Method in interface org.apache.hadoop.io.Stringifier
-
Closes this object.
- close() - Method in class org.apache.hadoop.mapred.JobClient
-
Close the JobClient
.
- close() - Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- close() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Close all child RRs.
- close() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- close() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- close() - Method in interface org.apache.hadoop.mapred.join.ResetableIterator
-
Close datasources and release resources.
- close() - Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- close() - Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- close() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Forward close request to proxied RR.
- close() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- close() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
-
Do nothing.
- close() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
-
- close() - Method in class org.apache.hadoop.mapred.lib.ChainMapper
-
Closes the ChainMapper and all the Mappers in the chain.
- close() - Method in class org.apache.hadoop.mapred.lib.ChainReducer
-
Closes the ChainReducer, the Reducer and all the Mappers in the chain.
- close() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- close() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
- close(Reporter) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter
-
Close this RecordWriter
to future operations.
- close() - Method in class org.apache.hadoop.mapred.lib.DelegatingMapper
-
- close() - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
-
- close() - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Closes all the opened named outputs.
- close() - Method in class org.apache.hadoop.mapred.LineRecordReader
-
Deprecated.
- close() - Method in class org.apache.hadoop.mapred.MapReduceBase
-
Deprecated.
Default implementation that does nothing.
- close() - Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
-
Closes the iterator so that the underlying streams can be closed.
- close() - Method in interface org.apache.hadoop.mapred.RecordReader
-
- close(Reporter) - Method in interface org.apache.hadoop.mapred.RecordWriter
-
Close this RecordWriter
to future operations.
- close() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- close() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- close() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- close() - Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- close() - Method in class org.apache.hadoop.mapred.TaskTracker
-
Close down the TaskTracker and all its components.
- close(Reporter) - Method in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
-
Deprecated.
- close() - Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- close() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
-
- close() - Method in class org.apache.hadoop.mapreduce.RecordReader
-
Close the record reader.
- close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.RecordWriter
-
Close this RecordWriter
to future operations.
- close() - Method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- close() - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Stops monitoring and also frees any buffered data, returning this
object to its initial state.
- close() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Stops monitoring and frees buffered data, returning this
object to its initial state.
- close() - Method in class org.apache.hadoop.metrics.spi.CompositeContext
-
- close() - Method in class org.apache.hadoop.net.SocketInputStream
-
- close() - Method in class org.apache.hadoop.net.SocketOutputStream
-
- close() - Method in class org.apache.hadoop.streaming.PipeMapper
-
- close() - Method in class org.apache.hadoop.streaming.PipeReducer
-
- close() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
Close this to future operations.
- close() - Method in class org.apache.hadoop.util.LineReader
-
Close the underlying stream.
- Closeable - Interface in org.apache.hadoop.io
-
- closeAll() - Static method in class org.apache.hadoop.fs.FileSystem
-
Close all cached filesystems.
- closed - Variable in class org.apache.hadoop.io.compress.CompressorStream
-
- closed - Variable in class org.apache.hadoop.io.compress.DecompressorStream
-
- closeSocket(Socket) - Static method in class org.apache.hadoop.io.IOUtils
-
- closeStream(Closeable) - Static method in class org.apache.hadoop.io.IOUtils
-
- closeWriter() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
Close the Lucene index writer associated with the intermediate form,
if created.
- ClusterStatus - Class in org.apache.hadoop.mapred
-
Status information on the current state of the Map-Reduce cluster.
- cmpcl - Variable in class org.apache.hadoop.mapred.join.Parser.Node
-
- CodeBuffer - Class in org.apache.hadoop.record.compiler
-
A wrapper around StringBuffer that automatically does indentation
- CodecPool - Class in org.apache.hadoop.io.compress
-
A global compressor/decompressor pool used to save and reuse
(possibly native) compression/decompression codecs.
- CodecPool() - Constructor for class org.apache.hadoop.io.compress.CodecPool
-
- collate(Object[], String) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- collate(List, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- collect(Object, TaggedMapOutput, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
The subclass can overwrite this method to perform additional filtering
and/or other processing logic before a value is collected.
- collect(K, V) - Method in interface org.apache.hadoop.mapred.OutputCollector
-
Adds a key/value pair to the output.
- collected - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- column - Variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- combine(Object[], Object[]) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
- combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.InnerJoinRecordReader
-
Return true iff the tuple is full (all data sources contain this key).
- combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
-
- combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.OuterJoinRecordReader
-
Emit everything from the collector.
- COMBINE_CLASS_ATTR - Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- CombineFileInputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
-
- CombineFileInputFormat() - Constructor for class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
default constructor
- CombineFileRecordReader<K,V> - Class in org.apache.hadoop.mapred.lib
-
A generic RecordReader that can hand out different recordReaders
for each chunk in a
CombineFileSplit
.
- CombineFileRecordReader(JobConf, CombineFileSplit, Reporter, Class<RecordReader<K, V>>) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
A generic RecordReader that can hand out different recordReaders
for each chunk in the CombineFileSplit.
- CombineFileSplit - Class in org.apache.hadoop.mapred.lib
-
A sub-collection of input files.
- CombineFileSplit() - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
-
default constructor
- CombineFileSplit(JobConf, Path[], long[], long[], String[]) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- CombineFileSplit(JobConf, Path[], long[]) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- CombineFileSplit(CombineFileSplit) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Copy constructor
- comCmd_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- COMMA - Static variable in class org.apache.hadoop.util.StringUtils
-
- COMMA_STR - Static variable in class org.apache.hadoop.util.StringUtils
-
- COMMA_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- Command - Class in org.apache.hadoop.fs.shell
-
An abstract class for the execution of a file system command
- Command(Configuration) - Constructor for class org.apache.hadoop.fs.shell.Command
-
Constructor
- CommandFormat - Class in org.apache.hadoop.fs.shell
-
Parse the args of a command and check the format of args.
- CommandFormat(String, int, int, String...) - Constructor for class org.apache.hadoop.fs.shell.CommandFormat
-
constructor
- commitPending(TaskAttemptID, TaskStatus) - Method in class org.apache.hadoop.mapred.TaskTracker
-
Task is reporting that it is in commit_pending
and it is waiting for the commit Response
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
To promote the task's temporary output to final output location
The task's output is moved to the job's output directory.
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
This method implements the new interface by calling the old method.
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
Move the files from the work directory to the job output directory
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
-
To promote the task's temporary output to final output location
The task's output is moved to the job's output directory.
- Comparator() - Constructor for class org.apache.hadoop.examples.SecondarySort.IntPair.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.BooleanWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.BytesWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.ByteWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.DoubleWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.FloatWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.IntWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.LongWritable.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.MD5Hash.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.NullWritable.Comparator
-
- comparator() - Method in class org.apache.hadoop.io.SortedMapWritable
- Comparator() - Constructor for class org.apache.hadoop.io.Text.Comparator
-
- Comparator() - Constructor for class org.apache.hadoop.io.UTF8.Comparator
-
Deprecated.
- COMPARATOR_JCLASS - Static variable in class org.apache.hadoop.io.file.tfile.TFile
-
comparator prefix: java class
- COMPARATOR_MEMCMP - Static variable in class org.apache.hadoop.io.file.tfile.TFile
-
comparator: memcmp
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
-
- compare(SecondarySort.IntPair, SecondarySort.IntPair) - Method in class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.BooleanWritable.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.BytesWritable.Comparator
-
Compare the buffers in serialized form.
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.ByteWritable.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.DoubleWritable.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.FloatWritable.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.IntWritable.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.LongWritable.Comparator
-
- compare(WritableComparable, WritableComparable) - Method in class org.apache.hadoop.io.LongWritable.DecreasingComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.LongWritable.DecreasingComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.MD5Hash.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.NullWritable.Comparator
-
Compare the buffers in serialized form.
- compare(byte[], int, int, byte[], int, int) - Method in interface org.apache.hadoop.io.RawComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.serializer.DeserializerComparator
-
- compare(T, T) - Method in class org.apache.hadoop.io.serializer.JavaSerializationComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.Text.Comparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.UTF8.Comparator
-
Deprecated.
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.io.WritableComparator
-
Optimization hook.
- compare(WritableComparable, WritableComparable) - Method in class org.apache.hadoop.io.WritableComparator
-
Compare two WritableComparables.
- compare(Object, Object) - Method in class org.apache.hadoop.io.WritableComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.record.RecordComparator
-
- compare(byte[], int) - Method in class org.apache.hadoop.util.DataChecksum
-
Compares the checksum located at buf[offset] with the current checksum.
- compare(int, int) - Method in interface org.apache.hadoop.util.IndexedSortable
-
- compareBytes(byte[], int, int, byte[], int, int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Lexicographic order of binary data.
- compareBytes(byte[], int, int, byte[], int, int) - Static method in class org.apache.hadoop.record.Utils
-
Lexicographic order of binary data.
- compareTo(Object) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- compareTo(Object) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- compareTo(Shard) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
Compare to another shard.
- compareTo(Object) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- compareTo(SecondarySort.IntPair) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
- compareTo(Object) - Method in class org.apache.hadoop.fs.FileStatus
-
Compare this object to another object
- compareTo(Object) - Method in class org.apache.hadoop.fs.Path
-
- compareTo(BinaryComparable) - Method in class org.apache.hadoop.io.BinaryComparable
-
Compare bytes from {#getBytes()}.
- compareTo(byte[], int, int) - Method in class org.apache.hadoop.io.BinaryComparable
-
Compare bytes from {#getBytes()} to those provided.
- compareTo(Object) - Method in class org.apache.hadoop.io.BooleanWritable
-
- compareTo(Object) - Method in class org.apache.hadoop.io.ByteWritable
-
Compares two ByteWritables.
- compareTo(Object) - Method in class org.apache.hadoop.io.DoubleWritable
-
- compareTo(byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Compare the entry key to another key.
- compareTo(byte[], int, int) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Compare the entry key to another key.
- compareTo(RawComparable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Compare an entry with a RawComparable object.
- compareTo(Utils.Version) - Method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
Compare this version with another version.
- compareTo(Object) - Method in class org.apache.hadoop.io.FloatWritable
-
Compares two FloatWritables.
- compareTo(Object) - Method in class org.apache.hadoop.io.IntWritable
-
Compares two IntWritables.
- compareTo(Object) - Method in class org.apache.hadoop.io.LongWritable
-
Compares two LongWritables.
- compareTo(MD5Hash) - Method in class org.apache.hadoop.io.MD5Hash
-
Compares this object with the specified object for order.
- compareTo(Object) - Method in class org.apache.hadoop.io.NullWritable
-
- compareTo(Object) - Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
- compareTo(Object) - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Compare two UTF8s.
- compareTo(Object) - Method in class org.apache.hadoop.io.VIntWritable
-
Compares two VIntWritables.
- compareTo(Object) - Method in class org.apache.hadoop.io.VLongWritable
-
Compares two VLongWritables.
- compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Implement Comparable contract (compare key of join or head of heap
with that of another).
- compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Implement Comparable contract (compare key at head of proxied RR
with that of another).
- compareTo(ID) - Method in class org.apache.hadoop.mapreduce.ID
-
Compare IDs by associated numbers
- compareTo(ID) - Method in class org.apache.hadoop.mapreduce.JobID
-
Compare JobIds by first jtIdentifiers, then by job numbers
- compareTo(ID) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
Compare TaskIds by first tipIds, then by task numbers.
- compareTo(ID) - Method in class org.apache.hadoop.mapreduce.TaskID
-
Compare TaskInProgressIds by first jobIds, then by tip numbers.
- compareTo(Object) - Method in class org.apache.hadoop.record.Buffer
-
Define the sort order of the Buffer.
- compareTo(Object) - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
This class doesn't implement Comparable as it's not meant to be used
for anything besides de/serializing.
- compareTo(Object) - Method in class org.apache.hadoop.record.Record
-
- compareTo(Key) - Method in class org.apache.hadoop.util.bloom.Key
-
- compatibleWith(Utils.Version) - Method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
Test compatibility.
- complete() - Method in class org.apache.hadoop.util.Progress
-
Completes this node, moving the parent node to its next child.
- completedJobs() - Method in class org.apache.hadoop.mapred.JobTracker
-
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Called when we're all done writing to the target.
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Called when we're all done writing to the target.
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
not implemented.
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- completeLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- ComposableInputFormat<K extends WritableComparable,V extends Writable> - Interface in org.apache.hadoop.mapred.join
-
Refinement of InputFormat requiring implementors to provide
ComposableRecordReader instead of RecordReader.
- ComposableRecordReader<K extends WritableComparable,V extends Writable> - Interface in org.apache.hadoop.mapred.join
-
Additional operations required of a RecordReader to participate in a join.
- compose(Class<? extends InputFormat>, String) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Convenience method for constructing composite formats.
- compose(String, Class<? extends InputFormat>, String...) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Convenience method for constructing composite formats.
- compose(String, Class<? extends InputFormat>, Path...) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Convenience method for constructing composite formats.
- CompositeContext - Class in org.apache.hadoop.metrics.spi
-
- CompositeContext() - Constructor for class org.apache.hadoop.metrics.spi.CompositeContext
-
- CompositeInputFormat<K extends WritableComparable> - Class in org.apache.hadoop.mapred.join
-
An InputFormat capable of performing joins over a set of data sources sorted
and partitioned the same way.
- CompositeInputFormat() - Constructor for class org.apache.hadoop.mapred.join.CompositeInputFormat
-
- CompositeInputSplit - Class in org.apache.hadoop.mapred.join
-
This InputSplit contains a set of child InputSplits.
- CompositeInputSplit() - Constructor for class org.apache.hadoop.mapred.join.CompositeInputSplit
-
- CompositeInputSplit(int) - Constructor for class org.apache.hadoop.mapred.join.CompositeInputSplit
-
- CompositeRecordReader<K extends WritableComparable,V extends Writable,X extends Writable> - Class in org.apache.hadoop.mapred.join
-
A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.
- CompositeRecordReader(int, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Create a RecordReader with capacity children to position
id in the parent reader.
- compress() - Method in class org.apache.hadoop.io.compress.BlockCompressorStream
-
- compress(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- compress(byte[], int, int) - Method in interface org.apache.hadoop.io.compress.Compressor
-
Fills specified buffer with compressed data.
- compress() - Method in class org.apache.hadoop.io.compress.CompressorStream
-
- compress(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater
-
- compress(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- compressedValSerializer - Variable in class org.apache.hadoop.io.SequenceFile.Writer
-
- CompressedWritable - Class in org.apache.hadoop.io
-
A base-class for Writables which store themselves compressed and lazily
inflate on field access.
- CompressedWritable() - Constructor for class org.apache.hadoop.io.CompressedWritable
-
- COMPRESSION_GZ - Static variable in class org.apache.hadoop.io.file.tfile.TFile
-
compression: gzip
- COMPRESSION_LZO - Static variable in class org.apache.hadoop.io.file.tfile.TFile
-
compression: lzo
- COMPRESSION_NONE - Static variable in class org.apache.hadoop.io.file.tfile.TFile
-
compression: none
- COMPRESSION_SUFFIX - Static variable in class org.apache.hadoop.contrib.failmon.LocalStore
-
- CompressionCodec - Interface in org.apache.hadoop.io.compress
-
This class encapsulates a streaming compression/decompression pair.
- CompressionCodecFactory - Class in org.apache.hadoop.io.compress
-
A factory that will find the correct codec for a given filename.
- CompressionCodecFactory(Configuration) - Constructor for class org.apache.hadoop.io.compress.CompressionCodecFactory
-
Find the codecs specified in the config value io.compression.codecs
and register them.
- CompressionInputStream - Class in org.apache.hadoop.io.compress
-
A compression input stream.
- CompressionInputStream(InputStream) - Constructor for class org.apache.hadoop.io.compress.CompressionInputStream
-
Create a compression input stream that reads
the decompressed bytes from the given stream.
- CompressionOutputStream - Class in org.apache.hadoop.io.compress
-
A compression output stream.
- CompressionOutputStream(OutputStream) - Constructor for class org.apache.hadoop.io.compress.CompressionOutputStream
-
Create a compression output stream that writes
the compressed bytes to the given stream.
- Compressor - Interface in org.apache.hadoop.io.compress
-
- compressor - Variable in class org.apache.hadoop.io.compress.CompressorStream
-
- CompressorStream - Class in org.apache.hadoop.io.compress
-
- CompressorStream(OutputStream, Compressor, int) - Constructor for class org.apache.hadoop.io.compress.CompressorStream
-
- CompressorStream(OutputStream, Compressor) - Constructor for class org.apache.hadoop.io.compress.CompressorStream
-
- CompressorStream(OutputStream) - Constructor for class org.apache.hadoop.io.compress.CompressorStream
-
Allow derived classes to directly set the underlying stream.
- computeSplitSize(long, long, long) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
- computeSplitSize(long, long, long) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
- conf - Variable in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- conf - Variable in class org.apache.hadoop.mapreduce.JobContext
-
- conf - Variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- config_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- Configurable - Interface in org.apache.hadoop.conf
-
- Configuration - Class in org.apache.hadoop.conf
-
Provides access to configuration parameters.
- Configuration() - Constructor for class org.apache.hadoop.conf.Configuration
-
A new configuration.
- Configuration(boolean) - Constructor for class org.apache.hadoop.conf.Configuration
-
A new configuration where the behavior of reading from the default
resources can be turned off.
- Configuration(Configuration) - Constructor for class org.apache.hadoop.conf.Configuration
-
A new configuration with the same settings cloned from another.
- Configuration.IntegerRanges - Class in org.apache.hadoop.conf
-
A class that represents a set of positive integer ranges.
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
- configure(IndexUpdateConfiguration) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
Configure using an index update configuration.
- configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
Initializes a new instance from a
JobConf
.
- configure(JobConf) - Method in class org.apache.hadoop.examples.dancing.DistributedPentomino.PentMap
-
- configure(JobConf) - Method in class org.apache.hadoop.examples.PiEstimator.PiReducer
-
Store job configuration.
- configure(JobConf) - Method in class org.apache.hadoop.examples.SleepJob
-
- configure(JobConf) - Method in interface org.apache.hadoop.mapred.JobConfigurable
-
Deprecated.
Initializes a new instance from a
JobConf
.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
-
Do nothing.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
get the input file name.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
-
Combiner does not need to configure.
- configure(JobConf) - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
-
Configure the object
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.ChainMapper
-
Configures the ChainMapper and all the Mappers in the chain.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.ChainReducer
-
Configures the ChainReducer, the Reducer and all the Mappers in the chain.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
Initializes a new instance from a
JobConf
.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.DelegatingMapper
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.HashPartitioner
-
Deprecated.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.RegexMapper
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
-
Read in the partition file and build indexing data structures.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.MapReduceBase
-
Deprecated.
Default implementation that does nothing.
- configure(JobConf) - Method in class org.apache.hadoop.mapred.MapRunner
-
- configure(JobConf) - Method in class org.apache.hadoop.mapred.TextInputFormat
-
Deprecated.
- configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeMapper
-
- configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeMapRed
-
- configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeReducer
-
- Configured - Class in org.apache.hadoop.conf
-
Base class for things that may be configured with a
Configuration
.
- Configured() - Constructor for class org.apache.hadoop.conf.Configured
-
Construct a Configured.
- Configured(Configuration) - Constructor for class org.apache.hadoop.conf.Configured
-
Construct a Configured.
- configureDB(JobConf, String, String, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBConfiguration
-
Sets the DB access related fields in the JobConf.
- configureDB(JobConf, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBConfiguration
-
Sets the DB access related fields in the JobConf.
- ConfiguredPolicy - Class in org.apache.hadoop.security.authorize
-
- ConfiguredPolicy(Configuration, PolicyProvider) - Constructor for class org.apache.hadoop.security.authorize.ConfiguredPolicy
-
- connect(Socket, SocketAddress, int) - Static method in class org.apache.hadoop.net.NetUtils
-
- ConnectionPermission - Class in org.apache.hadoop.security.authorize
-
Permission
to initiate a connection to a given service.
- ConnectionPermission(Class<?>) - Constructor for class org.apache.hadoop.security.authorize.ConnectionPermission
-
- constructQuery(String, String[]) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
-
Constructs the query used as the prepared statement to insert data.
- Consts - Class in org.apache.hadoop.record.compiler
-
const definitions for Record I/O compiler
- contains(Node) - Method in class org.apache.hadoop.net.NetworkTopology
-
Check if the tree contains node node
- containsKey(Object) - Method in class org.apache.hadoop.io.MapWritable
- containsKey(Object) - Method in class org.apache.hadoop.io.SortedMapWritable
- containsValue(Object) - Method in class org.apache.hadoop.io.MapWritable
- containsValue(Object) - Method in class org.apache.hadoop.io.SortedMapWritable
- ContentSummary - Class in org.apache.hadoop.fs
-
Store the summary of a content (a directory or a file).
- ContentSummary() - Constructor for class org.apache.hadoop.fs.ContentSummary
-
Constructor
- ContentSummary(long, long, long) - Constructor for class org.apache.hadoop.fs.ContentSummary
-
Constructor
- ContentSummary(long, long, long, long, long, long) - Constructor for class org.apache.hadoop.fs.ContentSummary
-
Constructor
- Context(Configuration, TaskAttemptID, RecordReader<KEYIN, VALUEIN>, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, InputSplit) - Constructor for class org.apache.hadoop.mapreduce.Mapper.Context
-
- Context(Configuration, TaskAttemptID, RawKeyValueIterator, Counter, Counter, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, RawComparator<KEYIN>, Class<KEYIN>, Class<VALUEIN>) - Constructor for class org.apache.hadoop.mapreduce.Reducer.Context
-
- ContextFactory - Class in org.apache.hadoop.metrics
-
Factory class for creating MetricsContext objects.
- ContextFactory() - Constructor for class org.apache.hadoop.metrics.ContextFactory
-
Creates a new instance of ContextFactory
- Continuous - Class in org.apache.hadoop.contrib.failmon
-
This class runs FailMon in a continuous mode on the local
node.
- Continuous() - Constructor for class org.apache.hadoop.contrib.failmon.Continuous
-
- convertToByteStream(Checksum, int) - Static method in class org.apache.hadoop.fs.FSOutputSummer
-
Converts a checksum integer value to a byte stream
- copy(FileSystem, Path, FileSystem, Path, boolean, Configuration) - Static method in class org.apache.hadoop.fs.FileUtil
-
Copy files between FileSystems.
- copy(FileSystem, Path[], FileSystem, Path, boolean, boolean, Configuration) - Static method in class org.apache.hadoop.fs.FileUtil
-
- copy(FileSystem, Path, FileSystem, Path, boolean, boolean, Configuration) - Static method in class org.apache.hadoop.fs.FileUtil
-
Copy files between FileSystems.
- copy(File, FileSystem, Path, boolean, Configuration) - Static method in class org.apache.hadoop.fs.FileUtil
-
Copy local files to a FileSystem.
- copy(FileSystem, Path, File, boolean, Configuration) - Static method in class org.apache.hadoop.fs.FileUtil
-
Copy FileSystem files to local files.
- copy(Writable) - Method in class org.apache.hadoop.io.AbstractMapWritable
-
Used by child copy constructors.
- copy(byte[], int, int) - Method in class org.apache.hadoop.record.Buffer
-
Copy the specified byte array to the Buffer.
- copy(Configuration, T, T) - Static method in class org.apache.hadoop.util.ReflectionUtils
-
Make a copy of the writable object using serialization to a buffer
- copyBytes(InputStream, OutputStream, int, boolean) - Static method in class org.apache.hadoop.io.IOUtils
-
Copies from one stream to another.
- copyBytes(InputStream, OutputStream, Configuration) - Static method in class org.apache.hadoop.io.IOUtils
-
Copies from one stream to another.
- copyBytes(InputStream, OutputStream, Configuration, boolean) - Static method in class org.apache.hadoop.io.IOUtils
-
Copies from one stream to another.
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- copyFromLocalFile(Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src file is on the local disk.
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src file is on the local disk.
- copyFromLocalFile(boolean, boolean, Path[], Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src files are on the local disk.
- copyFromLocalFile(boolean, boolean, Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src file is on the local disk.
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
The src file is on the local disk.
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
not implemented.
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- copyFromLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.LocalFileSystem
-
- copyMerge(FileSystem, Path, FileSystem, Path, boolean, Configuration, String) - Static method in class org.apache.hadoop.fs.FileUtil
-
Copy all files in a directory to one output file (merge).
- copyToHDFS(String, String) - Static method in class org.apache.hadoop.contrib.failmon.LocalStore
-
Copy a local file to HDFS
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
The src file is under FS, and the dst is on the local disk.
- copyToLocalFile(Path, Path, boolean) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
The src file is under FS, and the dst is on the local disk.
- copyToLocalFile(Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src file is under FS, and the dst is on the local disk.
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
The src file is under FS, and the dst is on the local disk.
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
The src file is under FS, and the dst is on the local disk.
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
copies the file in the har filesystem to a local file.
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- copyToLocalFile(boolean, Path, Path) - Method in class org.apache.hadoop.fs.LocalFileSystem
-
- Count - Class in org.apache.hadoop.fs.shell
-
Count the number of directories, files, bytes, quota, and remaining quota.
- Count(String[], int, Configuration) - Constructor for class org.apache.hadoop.fs.shell.Count
-
Constructor
- countCounters() - Method in class org.apache.hadoop.mapreduce.Counters
-
Returns the total number of counters, by summing the number of counters
in each group.
- Counter - Class in org.apache.hadoop.mapreduce
-
A named counter that tracks the progress of a map/reduce job.
- Counter() - Constructor for class org.apache.hadoop.mapreduce.Counter
-
- Counter(String, String) - Constructor for class org.apache.hadoop.mapreduce.Counter
-
- COUNTER_GROUP - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
-
Special counters which are written by the application and are
used by the framework for detecting bad records.
- COUNTER_MAP_PROCESSED_RECORDS - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
-
Number of processed map records.
- COUNTER_REDUCE_PROCESSED_GROUPS - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
-
Number of processed reduce groups.
- CounterGroup - Class in org.apache.hadoop.mapreduce
-
A group of
Counter
s that logically belong together.
- CounterGroup(String) - Constructor for class org.apache.hadoop.mapreduce.CounterGroup
-
- CounterGroup(String, String) - Constructor for class org.apache.hadoop.mapreduce.CounterGroup
-
- Counters - Class in org.apache.hadoop.mapred
-
- Counters() - Constructor for class org.apache.hadoop.mapred.Counters
-
Deprecated.
- Counters - Class in org.apache.hadoop.mapreduce
-
- Counters() - Constructor for class org.apache.hadoop.mapreduce.Counters
-
- Counters.Counter - Class in org.apache.hadoop.mapred
-
Deprecated.
A counter record, comprising its name and value.
- Counters.Group - Class in org.apache.hadoop.mapred
-
Deprecated.
Group
of counters, comprising of counters from a particular
counter
Enum
class.
- CountingBloomFilter - Class in org.apache.hadoop.util.bloom
-
Implements a counting Bloom filter, as defined by Fan et al.
- CountingBloomFilter() - Constructor for class org.apache.hadoop.util.bloom.CountingBloomFilter
-
Default constructor - use with readFields
- CountingBloomFilter(int, int, int) - Constructor for class org.apache.hadoop.util.bloom.CountingBloomFilter
-
Constructor
- countNumOfAvailableNodes(String, List<Node>) - Method in class org.apache.hadoop.net.NetworkTopology
-
return the number of leaves in scope but not in excludedNodes
if scope starts with ~, return the number of nodes that are not
in scope and excludedNodes;
- CPUParser - Class in org.apache.hadoop.contrib.failmon
-
Objects of this class parse the /proc/cpuinfo file to
gather information about present processors in the system.
- CPUParser() - Constructor for class org.apache.hadoop.contrib.failmon.CPUParser
-
Constructs a CPUParser
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(FileSystem, Path, FsPermission) - Static method in class org.apache.hadoop.fs.FileSystem
-
create a file with the provided permission
The permission of the file is set to be the provided permission as in
setPermission, not permission&~umask
It is implemented using two RPCs.
- create(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path.
- create(Path, boolean) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path.
- create(Path, Progressable) - Method in class org.apache.hadoop.fs.FileSystem
-
Create an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, short) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path.
- create(Path, short, Progressable) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, boolean, int) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path.
- create(Path, boolean, int, Progressable) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, boolean, int, short, long) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path.
- create(Path, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.FileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
A stream obtained via this call must be closed before using other APIs of
this class or else the invocation will block.
- create(Path, int) - Method in class org.apache.hadoop.fs.HarFileSystem
-
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.HarFileSystem
-
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- create(Path, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- create(Path, FsPermission, boolean, int, short, long, Progressable) - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- create(Class<?>, Object, RetryPolicy) - Static method in class org.apache.hadoop.io.retry.RetryProxy
-
Create a proxy for an interface of an implementation class
using the same retry policy for each method in the interface.
- create(Class<?>, Object, Map<String, RetryPolicy>) - Static method in class org.apache.hadoop.io.retry.RetryProxy
-
Create a proxy for an interface of an implementation class
using the a set of retry policies specified by method name.
- createAllSymlink(Configuration, File, File) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This method create symlinks for all files in a given dir in another directory
- createBaseListener(Configuration) - Method in class org.apache.hadoop.http.HttpServer
-
Create a required listener for the Jetty instance listening on the port
provided.
- createCompressor() - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- createCompressor() - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- createCompressor() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createCompressor() - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createDataFileReader(FileSystem, Path, Configuration) - Method in class org.apache.hadoop.io.MapFile.Reader
-
- createDataJoinJob(String[]) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
-
- createDecompressor() - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- createDecompressor() - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- createDecompressor() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createDecompressor() - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createHardLink(File, File) - Static method in class org.apache.hadoop.fs.FileUtil.HardLink
-
Creates a hardlink
- createImmutable(short) - Static method in class org.apache.hadoop.fs.permission.FsPermission
-
- createImmutable(String, String, FsPermission) - Static method in class org.apache.hadoop.fs.permission.PermissionStatus
-
- createImmutable(String[]) - Static method in class org.apache.hadoop.security.UnixUserGroupInformation
-
- createInputStream(InputStream) - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
Creates CompressionInputStream to be used to read off uncompressed data.
- createInputStream(InputStream, Decompressor) - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- createInputStream(InputStream) - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
Create a stream decompressor that will read from the given input stream.
- createInputStream(InputStream, Decompressor) - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- createInputStream(InputStream) - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createInputStream(InputStream, Decompressor) - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createInputStream(InputStream) - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createInputStream(InputStream, Decompressor) - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createInstance(String) - Static method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
-
Create an instance of the given class
- createInternalValue() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Create a value to be used internally for joins.
- createIOException(List<IOException>) - Static method in exception org.apache.hadoop.io.MultipleIOException
-
- createJob(String[]) - Static method in class org.apache.hadoop.streaming.StreamJob
-
This method creates a streaming job from the given argument list.
- createKey() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- createKey() - Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- createKey() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Create a new key value common to all child RRs.
- createKey() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Request new key from proxied RR.
- createKey() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- createKey() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- createKey() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
Create an object of the appropriate type to be used as a key.
- createKey() - Method in class org.apache.hadoop.mapred.LineRecordReader
-
Deprecated.
- createKey() - Method in interface org.apache.hadoop.mapred.RecordReader
-
Create an object of the appropriate type to be used as a key.
- createKey() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- createKey() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- createKey() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- createKey() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- createLocalTempFile(File, String, boolean) - Static method in class org.apache.hadoop.fs.FileUtil
-
Create a tmp file for a base file.
- createNewFile(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Creates the given Path as a brand-new zero-length file.
- createOutput(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- createOutputStream(OutputStream) - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
Creates CompressionOutputStream for BZip2
- createOutputStream(OutputStream, Compressor) - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- createOutputStream(OutputStream) - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- createOutputStream(OutputStream, Compressor) - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- createOutputStream(OutputStream) - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createOutputStream(OutputStream, Compressor) - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- createOutputStream(OutputStream) - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createOutputStream(OutputStream, Compressor) - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- createPool(JobConf, List<PathFilter>) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
Create a new pool and add the filters to it.
- createPool(JobConf, PathFilter...) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
Create a new pool and add the filters to it.
- createRecord(String) - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Creates a new MetricsRecord instance with the given recordName
.
- createRecord(MetricsContext, String) - Static method in class org.apache.hadoop.metrics.MetricsUtil
-
Utility method to create and return new metrics record instance within the
given context.
- createRecord(String) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Creates a new AbstractMetricsRecord instance with the given recordName
.
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.InputFormat
-
Create a record reader for a given split.
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
-
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
-
- createResetableIterator() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
The subclass can provide a different implementation on ResetableIterator.
- createScanner() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a scanner than can scan the whole TFile.
- createScanner(byte[], byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
- createScanner(RawComparable, RawComparable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
- createScannerByByteRange(long, long) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a scanner that covers a portion of TFile based on byte offsets.
- createScannerByKey(byte[], byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a scanner that covers a portion of TFile based on keys.
- createScannerByKey(RawComparable, RawComparable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a scanner that covers a specific key range.
- createScannerByRecordNum(long, long) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Create a scanner that covers a range of records.
- createSocket() - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- createSocket(InetAddress, int) - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- createSocket(InetAddress, int, InetAddress, int) - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- createSocket(String, int) - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- createSocket(String, int, InetAddress, int) - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- createSocket() - Method in class org.apache.hadoop.net.StandardSocketFactory
-
- createSocket(InetAddress, int) - Method in class org.apache.hadoop.net.StandardSocketFactory
-
- createSocket(InetAddress, int, InetAddress, int) - Method in class org.apache.hadoop.net.StandardSocketFactory
-
- createSocket(String, int) - Method in class org.apache.hadoop.net.StandardSocketFactory
-
- createSocket(String, int, InetAddress, int) - Method in class org.apache.hadoop.net.StandardSocketFactory
-
- createSocketAddr(String) - Static method in class org.apache.hadoop.net.NetUtils
-
Util method to build socket addr from either:
:
://:/
- createSocketAddr(String, int) - Static method in class org.apache.hadoop.net.NetUtils
-
Util method to build socket addr from either:
:
://:/
- createSymlink(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This method allows you to create symlinks in the current working directory
of the task to all the cache files/archives
- createTmpFileForWrite(String, long, Configuration) - Method in class org.apache.hadoop.fs.LocalDirAllocator
-
Creates a temporary file in the local FS.
- createURLStreamHandler(String) - Method in class org.apache.hadoop.fs.FsUrlStreamHandlerFactory
-
- createValue() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- createValue() - Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- createValue() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader
-
Create an object of the appropriate type to be used as a value.
- createValue() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
-
Create an object of the appropriate type to be used as a value.
- createValue() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Request new value from proxied RR.
- createValue() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- createValue() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- createValue() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
Create an object of the appropriate type to be used as a value.
- createValue() - Method in class org.apache.hadoop.mapred.LineRecordReader
-
Deprecated.
- createValue() - Method in interface org.apache.hadoop.mapred.RecordReader
-
Create an object of the appropriate type to be used as a value.
- createValue() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- createValue() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- createValue() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- createValue() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- createValueAggregatorJob(String[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
Create an Aggregate based map/reduce job.
- createValueAggregatorJob(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- createValueAggregatorJobs(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- createValueAggregatorJobs(String[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- createValueBytes() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
- createWriter(FileSystem, Configuration, Path, Class, Class) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, SequenceFile.CompressionType, Progressable) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, SequenceFile.CompressionType, CompressionCodec) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, SequenceFile.CompressionType, CompressionCodec, Progressable, SequenceFile.Metadata) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, int, short, long, SequenceFile.CompressionType, CompressionCodec, Progressable, SequenceFile.Metadata) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(FileSystem, Configuration, Path, Class, Class, SequenceFile.CompressionType, CompressionCodec, Progressable) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of SequenceFile Writer.
- createWriter(Configuration, FSDataOutputStream, Class, Class, SequenceFile.CompressionType, CompressionCodec, SequenceFile.Metadata) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of 'raw' SequenceFile Writer.
- createWriter(Configuration, FSDataOutputStream, Class, Class, SequenceFile.CompressionType, CompressionCodec) - Static method in class org.apache.hadoop.io.SequenceFile
-
Construct the preferred type of 'raw' SequenceFile Writer.
- CSTRING_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- CsvRecordInput - Class in org.apache.hadoop.record
-
- CsvRecordInput(InputStream) - Constructor for class org.apache.hadoop.record.CsvRecordInput
-
Creates a new instance of CsvRecordInput
- CsvRecordOutput - Class in org.apache.hadoop.record
-
- CsvRecordOutput(OutputStream) - Constructor for class org.apache.hadoop.record.CsvRecordOutput
-
Creates a new instance of CsvRecordOutput
- CUR_DIR - Static variable in class org.apache.hadoop.fs.Path
-
- curChar - Variable in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
- curReader - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- CURRENT_VERSION - Static variable in class org.apache.hadoop.ipc.Server
-
- currentToken - Variable in exception org.apache.hadoop.record.compiler.generated.ParseException
-
This is the last token that has been consumed successfully.
- CyclicIteration<K,V> - Class in org.apache.hadoop.util
-
- CyclicIteration(NavigableMap<K, V>, K) - Constructor for class org.apache.hadoop.util.CyclicIteration
-
- G_SIZE - Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- GangliaContext - Class in org.apache.hadoop.metrics.ganglia
-
Context for sending metrics to Ganglia.
- GangliaContext() - Constructor for class org.apache.hadoop.metrics.ganglia.GangliaContext
-
Creates a new instance of GangliaContext
- GangliaContext31 - Class in org.apache.hadoop.metrics.ganglia
-
Context for sending metrics to Ganglia version 3.1.x.
- GangliaContext31() - Constructor for class org.apache.hadoop.metrics.ganglia.GangliaContext31
-
- gcd(int, int) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Determines the greatest common divisor (GCD) of two integers.
- gcd(int[]) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Determines the greatest common divisor (GCD) of a list
of integers.
- genCode(String, String, ArrayList<String>) - Method in class org.apache.hadoop.record.compiler.JFile
-
Generate record code in given language.
- generateActualKey(K, V) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Generate the actual key from the given key/value.
- generateActualValue(K, V) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Generate the actual value from the given key and value.
- generateEntry(String, String, Text) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- generateFileNameForKeyValue(K, V, String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Generate the file output file name based on the given key and the leaf file
name.
- generateGroupKey(TaggedMapOutput) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
Generate a map output key.
- generateInputTag(String) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
Determine the source tag based on the input file name.
- generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.examples.AggregateWordCount.WordCountPlugInClass
-
- generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.examples.AggregateWordHistogram.AggregateWordHistogramPlugin
-
Parse the given value, generate an aggregation-id/value pair per word.
- generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
-
Generate a list of aggregation-id/value pairs for the given key/value pairs
by delegating the invocation to the real object.
- generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair.
- generateKeyValPairs(Object, Object) - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
-
Generate a list of aggregation-id/value pairs for the given key/value pair.
- generateLeafFileName(String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Generate the leaf name for the output file name.
- generateParseException() - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Generate ParseException.
- generateTaggedMapOutput(Object) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
Generate a tagged map output value.
- generateValueAggregator(String) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- generationFromSegmentsFileName(String) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
-
Parse the generation off the segments file name and return it.
- GenericOptionsParser - Class in org.apache.hadoop.util
-
GenericOptionsParser
is a utility to parse command line
arguments generic to the Hadoop framework.
- GenericOptionsParser(Options, String[]) - Constructor for class org.apache.hadoop.util.GenericOptionsParser
-
Create an options parser with the given options to parse the args.
- GenericOptionsParser(String[]) - Constructor for class org.apache.hadoop.util.GenericOptionsParser
-
Create an options parser to parse the args.
- GenericOptionsParser(Configuration, String[]) - Constructor for class org.apache.hadoop.util.GenericOptionsParser
-
Create a GenericOptionsParser to parse only the generic Hadoop
arguments.
- GenericOptionsParser(Configuration, Options, String[]) - Constructor for class org.apache.hadoop.util.GenericOptionsParser
-
Create a GenericOptionsParser
to parse given options as well
as generic Hadoop options.
- GenericsUtil - Class in org.apache.hadoop.util
-
Contains utility methods for dealing with Java Generics.
- GenericsUtil() - Constructor for class org.apache.hadoop.util.GenericsUtil
-
- GenericWritable - Class in org.apache.hadoop.io
-
A wrapper for Writable instances.
- GenericWritable() - Constructor for class org.apache.hadoop.io.GenericWritable
-
- get(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property, null
if
no such property exists.
- get(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property.
- get(String) - Method in class org.apache.hadoop.contrib.failmon.EventRecord
-
Get the value of a property of the EventRecord.
- get(String) - Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
-
Get the value of a property of the EventRecord.
- get(Configuration) - Static method in class org.apache.hadoop.fs.FileSystem
-
Returns the configured filesystem implementation.
- get(URI, Configuration) - Static method in class org.apache.hadoop.fs.FileSystem
-
Returns the FileSystem for this URI's scheme and authority.
- get(long, Writable) - Method in class org.apache.hadoop.io.ArrayFile.Reader
-
Return the n
th value in the file.
- get() - Method in class org.apache.hadoop.io.ArrayWritable
-
- get(WritableComparable, Writable) - Method in class org.apache.hadoop.io.BloomMapFile.Reader
-
- get() - Method in class org.apache.hadoop.io.BooleanWritable
-
Returns the value of the BooleanWritable
- get() - Method in class org.apache.hadoop.io.BytesWritable
-
- get() - Method in class org.apache.hadoop.io.ByteWritable
-
Return the value of this ByteWritable.
- get() - Method in class org.apache.hadoop.io.DoubleWritable
-
- get(BytesWritable, BytesWritable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy the key and value in one shot into BytesWritables.
- get() - Method in class org.apache.hadoop.io.FloatWritable
-
Return the value of this FloatWritable.
- get() - Method in class org.apache.hadoop.io.GenericWritable
-
Return the wrapped instance.
- get() - Method in class org.apache.hadoop.io.IntWritable
-
Return the value of this IntWritable.
- get() - Method in class org.apache.hadoop.io.LongWritable
-
Return the value of this LongWritable.
- get(WritableComparable, Writable) - Method in class org.apache.hadoop.io.MapFile.Reader
-
Return the value for the named key, or null if none exists.
- get(Object) - Method in class org.apache.hadoop.io.MapWritable
- get() - Static method in class org.apache.hadoop.io.NullWritable
-
Returns the single instance of this class.
- get() - Method in class org.apache.hadoop.io.ObjectWritable
-
Return the instance, or null if none.
- get(Text) - Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- get(WritableComparable) - Method in class org.apache.hadoop.io.SetFile.Reader
-
Read the matching key from a set into key
.
- get(Object) - Method in class org.apache.hadoop.io.SortedMapWritable
- get() - Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- get() - Method in class org.apache.hadoop.io.VIntWritable
-
Return the value of this VIntWritable.
- get() - Method in class org.apache.hadoop.io.VLongWritable
-
Return the value of this LongWritable.
- get(Class<? extends WritableComparable>) - Static method in class org.apache.hadoop.io.WritableComparator
-
- get() - Static method in class org.apache.hadoop.ipc.Server
-
Returns the server instance called under or null.
- get(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
Get ith child InputSplit.
- get(int) - Method in class org.apache.hadoop.mapred.join.TupleWritable
-
Get ith Writable from Tuple.
- get() - Method in class org.apache.hadoop.metrics.util.MetricsIntValue
-
Get value
- get() - Method in class org.apache.hadoop.metrics.util.MetricsLongValue
-
Get value
- get(String) - Method in class org.apache.hadoop.metrics.util.MetricsRegistry
-
- get(DataInput) - Static method in class org.apache.hadoop.record.BinaryRecordInput
-
Get a thread-local record input for the supplied DataInput.
- get(DataOutput) - Static method in class org.apache.hadoop.record.BinaryRecordOutput
-
Get a thread-local record output for the supplied DataOutput.
- get() - Method in class org.apache.hadoop.record.Buffer
-
Get the data from the Buffer.
- get() - Method in class org.apache.hadoop.util.Progress
-
Returns the overall progress of the root.
- getAbsolutePath(String) - Method in class org.apache.hadoop.streaming.PathFinder
-
Returns the full path name of this file if it is listed in the
path
- getAccessKey() - Method in class org.apache.hadoop.fs.s3.S3Credentials
-
- getAccessTime() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the access time of the file.
- getActions() - Method in class org.apache.hadoop.security.authorize.ConnectionPermission
-
- getActiveTrackerNames() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the names of task trackers in the cluster.
- getAddress(Configuration) - Static method in class org.apache.hadoop.mapred.JobTracker
-
- getAlgorithmName() - Method in class org.apache.hadoop.fs.FileChecksum
-
The checksum algorithm name
- getAlgorithmName() - Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
The checksum algorithm name
- getAllJobs() - Method in class org.apache.hadoop.mapred.JobClient
-
Get the jobs that are submitted.
- getAllJobs() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getAllStaticResolutions() - Static method in class org.apache.hadoop.net.NetUtils
-
- getAllStatistics() - Static method in class org.apache.hadoop.fs.FileSystem
-
Return the FileSystem classes that have Statistics
- getAllTasks() - Method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Returns all map and reduce tasks .
- getApproxChkSumLength(long) - Static method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- getArchiveClassPaths(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the archive entries in classpath as an array of Path
- getArchiveTimestamps(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the timestamps of the archives
- getAssignedJobID() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getAssignedTracker(TaskAttemptID) - Method in class org.apache.hadoop.mapred.JobTracker
-
Get tracker name for a given task id.
- getAttemptsToStartSkipping(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Get the number of Task attempts AFTER which skip mode
will be kicked off.
- getAttribute(String) - Method in class org.apache.hadoop.http.HttpServer
-
Get the value in the webapp context.
- getAttribute(String) - Method in class org.apache.hadoop.metrics.ContextFactory
-
Returns the value of the named attribute, or null if there is no
attribute of that name.
- getAttribute(String) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Convenience method for subclasses to access factory attributes.
- getAttribute(String) - Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
-
- getAttributeNames() - Method in class org.apache.hadoop.metrics.ContextFactory
-
Returns the names of all the factory's attributes.
- getAttributes(String[]) - Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
-
- getAttributeTable(String) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Returns an attribute-value map derived from the factory attributes
by finding all factory attributes that begin with
contextName.tableName.
- getAutoIncrMapperProcCount(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
- getAutoIncrReducerProcCount(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
- getAvailable() - Method in class org.apache.hadoop.fs.DF
-
- getBasePathInJarOut(String) - Method in class org.apache.hadoop.streaming.JarBuilder
-
- getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
- getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat
-
- getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat
-
- getBeginColumn() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get token beginning column number.
- getBeginLine() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get token beginning line number.
- getBlacklistedTrackerNames() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the names of task trackers in the cluster.
- getBlacklistedTrackers() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the number of blacklisted task trackers in the cluster.
- getBlockIndex(BlockLocation[], long) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
- getBlockIndex(BlockLocation[], long) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
- getBlocks() - Method in class org.apache.hadoop.fs.s3.INode
-
- getBlockSize() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the block size of the file.
- getBlockSize(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
- getBlockSize() - Method in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
Returns the blocksize parameter specified at construction time.
- getBloomFilter() - Method in class org.apache.hadoop.io.BloomMapFile.Reader
-
Retrieve the Bloom filter used by this instance of the Reader.
- getBoolean(String, boolean) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as a boolean
.
- getBoundAntProperty(String, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- getBuildVersion() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getBuildVersion() - Static method in class org.apache.hadoop.util.VersionInfo
-
Returns the buildVersion which includes version,
revision, user and date.
- getBytes() - Method in class org.apache.hadoop.fs.FileChecksum
-
The value of the checksum in bytes
- getBytes() - Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
The value of the checksum in bytes
- getBytes() - Method in class org.apache.hadoop.io.BinaryComparable
-
Return representative byte array for this instance.
- getBytes() - Method in class org.apache.hadoop.io.BytesWritable
-
Get the data from the BytesWritable.
- getBytes() - Method in class org.apache.hadoop.io.Text
-
- getBytes() - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
The raw bytes.
- getBytes(String) - Static method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Convert a string to a UTF-8 encoded byte array.
- getBytes() - Method in class org.apache.hadoop.util.bloom.Key
-
- getBytesPerChecksum() - Method in class org.apache.hadoop.util.DataChecksum
-
- getBytesPerSum() - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Return the bytes Per Checksum
- getBytesRead() - Method in class org.apache.hadoop.fs.FileSystem.Statistics
-
Get the total number of bytes read
- getBytesRead() - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- getBytesRead() - Method in interface org.apache.hadoop.io.compress.Compressor
-
Return number of uncompressed bytes input so far.
- getBytesRead() - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
Returns the total number of uncompressed bytes input so far.
- getBytesRead() - Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
Returns the total number of uncompressed bytes input so far.
- getBytesWritten() - Method in class org.apache.hadoop.fs.FileSystem.Statistics
-
Get the total number of bytes written
- getBytesWritten() - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- getBytesWritten() - Method in interface org.apache.hadoop.io.compress.Compressor
-
Return number of compressed bytes output so far.
- getBytesWritten() - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
Returns the total number of compressed bytes output so far.
- getBytesWritten() - Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
Returns the total number of compressed bytes output so far.
- getCacheArchives(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get cache archives set in the Configuration
- getCacheFiles(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get cache files set in the Configuration
- getCallQueueLen() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The number of rpc calls in the queue.
- getCallQueueLen() - Method in class org.apache.hadoop.ipc.Server
-
The number of rpc calls in the queue.
- getCapacity() - Method in class org.apache.hadoop.fs.DF
-
- getCapacity() - Method in class org.apache.hadoop.io.BytesWritable
-
Get the capacity, which is the maximum size that could handled without
resizing the backing storage.
- getCapacity() - Method in class org.apache.hadoop.record.Buffer
-
Get the capacity, which is the maximum count that could handled without
resizing the backing storage.
- getCategory(List<List<Pentomino.ColumnName>>) - Method in class org.apache.hadoop.examples.dancing.Pentomino
-
Find whether the solution has the x in the upper left quadrant, the
x-midline, the y-midline or in the center.
- getChannel() - Method in class org.apache.hadoop.net.SocketInputStream
-
Returns underlying channel used by inputstream.
- getChannel() - Method in class org.apache.hadoop.net.SocketOutputStream
-
Returns underlying channel used by this stream.
- getChecksumFile(Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Return the name of the checksum file associated with a file.
- getChecksumFileLength(Path, long) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Return the length of the checksum file given the size of the
actual file.
- getChecksumHeaderSize() - Static method in class org.apache.hadoop.util.DataChecksum
-
- getChecksumLength(long, int) - Static method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Calculated the length of the checksum file in bytes.
- getChecksumSize() - Method in class org.apache.hadoop.util.DataChecksum
-
- getChecksumType() - Method in class org.apache.hadoop.util.DataChecksum
-
- getChunkPosition(long) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Return position of beginning of chunk containing pos.
- getClass(String, Class<?>) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as a Class
.
- getClass(String, Class<? extends U>, Class<U>) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as a Class
implementing the interface specified by xface
.
- getClass(byte) - Method in class org.apache.hadoop.io.AbstractMapWritable
-
- getClass(String, Configuration) - Static method in class org.apache.hadoop.io.WritableName
-
Return the class for a name.
- getClass(T) - Static method in class org.apache.hadoop.util.GenericsUtil
-
Returns the Class object (of type Class<T>
) of the
argument of type T
.
- getClass(T) - Static method in class org.apache.hadoop.util.ReflectionUtils
-
Return the correctly-typed
Class
of the given object.
- getClassByName(String) - Method in class org.apache.hadoop.conf.Configuration
-
Load a class by name.
- getClassByName(String) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
-
- getClasses(String, Class<?>...) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property
as an array of Class
.
- getClassLoader() - Method in class org.apache.hadoop.conf.Configuration
-
- getClassName() - Method in exception org.apache.hadoop.ipc.RemoteException
-
- getCleanupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
-
Get the information of the current state of the cleanup tasks of a job.
- getCleanupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getClientVersion() - Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
-
Get the client's preferred version
- getClosest(WritableComparable, Writable) - Method in class org.apache.hadoop.io.MapFile.Reader
-
Finds the record that is the closest match to the specified key.
- getClosest(WritableComparable, Writable, boolean) - Method in class org.apache.hadoop.io.MapFile.Reader
-
Finds the record that is the closest match to the specified key.
- getClusterNick() - Method in class org.apache.hadoop.streaming.StreamJob
-
Deprecated.
- getClusterStatus() - Method in class org.apache.hadoop.mapred.JobClient
-
Get status information about the Map-Reduce cluster.
- getClusterStatus(boolean) - Method in class org.apache.hadoop.mapred.JobClient
-
Get status information about the Map-Reduce cluster.
- getClusterStatus() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getClusterStatus(boolean) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getCodec(Path) - Method in class org.apache.hadoop.io.compress.CompressionCodecFactory
-
Find the relevant compression codec for the given file based on its
filename suffix.
- getCodecClasses(Configuration) - Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
-
Get the list of codecs listed in the configuration
- getCollector(String, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Gets the output collector for a named output.
- getCollector(String, String, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Gets the output collector for a multi named output.
- getColumn() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Deprecated.
- getColumnName(int) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
-
Get the name of a given column as a string
- getCombinerClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the user-defined combiner class used to combine map-outputs
before being sent to the reducers.
- getCombinerClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the combiner class for the job.
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
-
- getCombinerOutput() - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
-
- getCombinerOutput() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
- getCommandLine() - Method in class org.apache.hadoop.util.GenericOptionsParser
-
Returns the commons-cli CommandLine
object
to process the parsed arguments.
- getCommandName() - Method in class org.apache.hadoop.fs.shell.Command
-
Return the command's name excluding the leading character -
- getCommandName() - Method in class org.apache.hadoop.fs.shell.Count
-
- getComparator() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get an instance of the RawComparator that is constructed based on the
string comparator representation.
- getComparator() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Return comparator defining the ordering for RecordReaders in this
composite.
- getComparatorName() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get the string representation of the comparator.
- getCompressedData() - Method in class org.apache.hadoop.io.compress.BlockDecompressorStream
-
- getCompressedData() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- getCompressionCodec() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the compression codec of data in this file.
- getCompressionCodec() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
Returns the compression codec of data in this file.
- getCompressionType(Configuration) - Static method in class org.apache.hadoop.io.SequenceFile
-
- getCompressMapOutput() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Are the outputs of the maps be compressed?
- getCompressor(CompressionCodec) - Static method in class org.apache.hadoop.io.compress.CodecPool
-
- getCompressorType() - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- getCompressorType() - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- getCompressorType() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- getCompressorType() - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- getCompressOutput(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Is the job output compressed?
- getCompressOutput(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Is the job output compressed?
- getConf() - Method in interface org.apache.hadoop.conf.Configurable
-
Return the configuration used by this object.
- getConf() - Method in class org.apache.hadoop.conf.Configured
-
- getConf() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
- getConf() - Method in class org.apache.hadoop.io.AbstractMapWritable
-
- getConf() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- getConf() - Method in class org.apache.hadoop.io.GenericWritable
-
- getConf() - Method in class org.apache.hadoop.io.ObjectWritable
-
- getConf() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Return the configuration used by this object.
- getConf() - Method in class org.apache.hadoop.mapred.lib.InputSampler
-
- getConf() - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase
-
- getConf() - Method in class org.apache.hadoop.net.ScriptBasedMapping
-
- getConf() - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- getConf() - Method in class org.apache.hadoop.security.authorize.ConfiguredPolicy
-
- getConf() - Method in class org.apache.hadoop.streaming.StreamJob
-
- getConfiguration() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the underlying configuration object.
- getConfiguration() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Return the configuration for the job.
- getConfiguration() - Method in class org.apache.hadoop.util.GenericOptionsParser
-
Get the modified configuration
- getConfResourceAsInputStream(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get an input stream attached to the configuration resource with the
given name
.
- getConfResourceAsReader(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get a
Reader
attached to the configuration resource with the
given
name
.
- getConnectAddress(Server) - Static method in class org.apache.hadoop.net.NetUtils
-
Returns InetSocketAddress that a client can use to
connect to the server.
- getContentSummary(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
- getContext(String, String) - Method in class org.apache.hadoop.metrics.ContextFactory
-
Returns the named MetricsContext instance, constructing it if necessary
using the factory's current configuration attributes.
- getContext(String) - Method in class org.apache.hadoop.metrics.ContextFactory
-
- getContext(String) - Static method in class org.apache.hadoop.metrics.MetricsUtil
-
- getContext(String, String) - Static method in class org.apache.hadoop.metrics.MetricsUtil
-
Utility method to return the named context.
- getContext() - Method in class org.apache.hadoop.streaming.PipeMapRed
-
- getContextFactory() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Returns the factory by which this context was created.
- getContextName() - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Returns the context name.
- getContextName() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Returns the context name.
- getCount() - Method in class org.apache.hadoop.record.Buffer
-
Get the current count of the buffer.
- getCounter() - Method in class org.apache.hadoop.mapred.Counters.Counter
-
Deprecated.
What is the current value of this counter?
- getCounter(Enum) - Method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Returns current value of the specified counter, or 0 if the counter
does not exist.
- getCounter(String) - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Returns the value of the specified counter, or 0 if the counter does
not exist.
- getCounter(int, String) - Method in class org.apache.hadoop.mapred.Counters.Group
-
- getCounter(Enum<?>) - Method in interface org.apache.hadoop.mapred.Reporter
-
- getCounter(String, String) - Method in interface org.apache.hadoop.mapred.Reporter
-
- getCounter(Enum<?>) - Method in class org.apache.hadoop.mapreduce.StatusReporter
-
- getCounter(String, String) - Method in class org.apache.hadoop.mapreduce.StatusReporter
-
- getCounter(Enum<?>) - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- getCounter(String, String) - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- getCounterForName(String) - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Get the counter for the given name and create it if it doesn't exist.
- getCounters() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Gets the counters for this job.
- getCounters() - Method in class org.apache.hadoop.mapred.TaskReport
-
A table of counters.
- getCounters() - Method in class org.apache.hadoop.mapreduce.Job
-
Gets the counters for this job.
- getCountersEnabled(JobConf) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns if the counters for the named outputs are enabled or not.
- getCountQuery() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
Returns the query for getting the total number of rows,
subclasses can override this for custom behaviour.
- getCumulativeVmem() - Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
Get the cumulative virtual memory used by all the processes in the
process-tree.
- getCumulativeVmem(int) - Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
Get the cumulative virtual memory used by all the processes in the
process-tree that are older than the passed in age.
- getCurrentIntervalValue() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
-
The Value at the current interval
- getCurrentIntervalValue() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
-
The Value at the current interval
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.MapContext
-
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.RecordReader
-
Get the current key
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.ReduceContext
-
- getCurrentKey() - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
Get the current key.
- getCurrentSegmentGeneration(Directory) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
-
Get the generation (N) of the current segments_N file in the directory.
- getCurrentSegmentGeneration(String[]) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
-
Get the generation (N) of the current segments_N file from a list of
files.
- getCurrentSplit(JobConf) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- getCurrentStatus() - Method in class org.apache.hadoop.mapred.TaskReport
-
The current status
- getCurrentTrashDir() - Method in class org.apache.hadoop.fs.FsShell
-
Returns the Trash object associated with this shell.
- getCurrentUGI() - Static method in class org.apache.hadoop.security.UserGroupInformation
-
- getCurrentValue(Writable) - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Get the 'value' corresponding to the last read 'key'.
- getCurrentValue(Object) - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Get the 'value' corresponding to the last read 'key'.
- getCurrentValue(V) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.MapContext
-
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.RecordReader
-
Get the current value.
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.ReduceContext
-
- getCurrentValue() - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
Get the current value.
- getData() - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- getData() - Method in class org.apache.hadoop.io.DataInputBuffer
-
- getData() - Method in class org.apache.hadoop.io.DataOutputBuffer
-
Returns the current contents of the buffer.
- getData() - Method in class org.apache.hadoop.io.OutputBuffer
-
Returns the current contents of the buffer.
- getDate() - Static method in class org.apache.hadoop.util.VersionInfo
-
The date that Hadoop was compiled.
- getDeclaredClass() - Method in class org.apache.hadoop.io.ObjectWritable
-
Return the class this is meant to be.
- getDecompressor(CompressionCodec) - Static method in class org.apache.hadoop.io.compress.CodecPool
-
- getDecompressorType() - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
This functionality is currently not supported.
- getDecompressorType() - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
- getDecompressorType() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- getDecompressorType() - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- getDefault() - Static method in class org.apache.hadoop.fs.permission.FsPermission
-
Get the default permission.
- getDefaultBlockSize() - Method in class org.apache.hadoop.fs.FileSystem
-
Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.
- getDefaultBlockSize() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.
- getDefaultBlockSize() - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- getDefaultExtension() - Method in class org.apache.hadoop.io.compress.BZip2Codec
-
.bz2 is recognized as the default extension for compressed BZip2 files
- getDefaultExtension() - Method in interface org.apache.hadoop.io.compress.CompressionCodec
-
Get the default filename extension for this kind of compression.
- getDefaultExtension() - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- getDefaultExtension() - Method in class org.apache.hadoop.io.compress.GzipCodec
-
- getDefaultHost(String, String) - Static method in class org.apache.hadoop.net.DNS
-
Returns the default (first) host name associated by the provided
nameserver with the address bound to the specified network interface
- getDefaultHost(String) - Static method in class org.apache.hadoop.net.DNS
-
Returns the default (first) host name associated by the default
nameserver with the address bound to the specified network interface
- getDefaultIP(String) - Static method in class org.apache.hadoop.net.DNS
-
Returns the first available IP address associated with the provided
network interface
- getDefaultMaps() - Method in class org.apache.hadoop.mapred.JobClient
-
Get status information about the max available Maps in the cluster.
- getDefaultReduces() - Method in class org.apache.hadoop.mapred.JobClient
-
Get status information about the max available Reduces in the cluster.
- getDefaultReplication() - Method in class org.apache.hadoop.fs.FileSystem
-
Get the default replication.
- getDefaultReplication() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Get the default replication.
- getDefaultReplication() - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- getDefaultSocketFactory(Configuration) - Static method in class org.apache.hadoop.net.NetUtils
-
Get the default socket factory as specified by the configuration
parameter hadoop.rpc.socket.factory.default
- getDefaultUri(Configuration) - Static method in class org.apache.hadoop.fs.FileSystem
-
Get the default filesystem URI from a configuration.
- getDefaultWorkFile(TaskAttemptContext, String) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Get the default path and filename for the output format.
- getDelegate() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Obtain an iterator over the child RRs apropos of the value type
ultimately emitted from this join.
- getDelegate() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader
-
Return an iterator wrapping the JoinCollector.
- getDelegate() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
-
Return an iterator returning a single value from the tuple.
- getDependingJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getDescription() - Method in class org.apache.hadoop.metrics.util.MetricsBase
-
- getDeserializer(Class<Serializable>) - Method in class org.apache.hadoop.io.serializer.JavaSerialization
-
- getDeserializer(Class<T>) - Method in interface org.apache.hadoop.io.serializer.Serialization
-
- getDeserializer(Class<T>) - Method in class org.apache.hadoop.io.serializer.SerializationFactory
-
- getDeserializer(Class<Writable>) - Method in class org.apache.hadoop.io.serializer.WritableSerialization
-
- getDiagnostics() - Method in class org.apache.hadoop.mapred.TaskReport
-
A list of error messages.
- getDigest() - Method in class org.apache.hadoop.io.MD5Hash
-
Returns the digest bytes.
- getDirectory() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
Get the ram directory of the intermediate form.
- getDirectory() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
Get the directory where this shard resides.
- getDirectoryCount() - Method in class org.apache.hadoop.fs.ContentSummary
-
- getDirPath() - Method in class org.apache.hadoop.fs.DF
-
- getDirPath() - Method in class org.apache.hadoop.fs.DU
-
- getDisplayName() - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Returns localized name of the group.
- getDisplayName() - Method in class org.apache.hadoop.mapreduce.Counter
-
Get the name of the counter.
- getDisplayName() - Method in class org.apache.hadoop.mapreduce.CounterGroup
-
Get the display name of the group.
- getDistance(Node, Node) - Method in class org.apache.hadoop.net.NetworkTopology
-
Return the distance between two nodes
It is assumed that the distance from one node to its parent is 1
The distance between two nodes is calculated by summing up their distances
to their closest common ancestor.
- getDistributionPolicyClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the distribution policy class.
- getDmax(String) - Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
-
- getDocument() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Get the document.
- getDocumentAnalyzerClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the analyzer class.
- getDoubleValue(Object) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
- getDU(File) - Static method in class org.apache.hadoop.fs.FileUtil
-
Takes an input dir and returns the du on that local directory.
- getElementTypeID() - Method in class org.apache.hadoop.record.meta.VectorTypeID
-
- getEmptier() - Method in class org.apache.hadoop.fs.Trash
-
Return a
Runnable
that periodically empties the trash of all
users, intended to be run by the superuser.
- getEnd() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
-
- getEndColumn() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get token end column number.
- getEndLine() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get token end line number.
- getEntry(MapFile.Reader[], Partitioner<K, V>, K, V) - Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
-
Get an entry from output generated by this class.
- getEntryComparator() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a Comparator object to compare Entries.
- getEntryCount() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get the number of key-value pair entries in TFile.
- getError() - Static method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- getEventId() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Returns event Id.
- getExceptions() - Method in exception org.apache.hadoop.io.MultipleIOException
-
- getExcludedHosts() - Method in class org.apache.hadoop.util.HostsFileReader
-
- getExecString() - Method in class org.apache.hadoop.fs.DF
-
- getExecString() - Method in class org.apache.hadoop.fs.DU
-
- getExecString() - Method in class org.apache.hadoop.util.Shell
-
return an array containing the command name & its parameters
- getExecString() - Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- getExecutable(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Get the URI of the application's executable.
- getExitCode() - Method in exception org.apache.hadoop.util.Shell.ExitCodeException
-
- getExitCode() - Method in class org.apache.hadoop.util.Shell
-
get the exit code
- getFactor() - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Get the number of streams to merge at once.
- getFactory(Class) - Static method in class org.apache.hadoop.io.WritableFactories
-
Define a factory for a class.
- getFactory() - Static method in class org.apache.hadoop.metrics.ContextFactory
-
Returns the singleton ContextFactory instance, constructing it if
necessary.
- getFailedJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getFatal() - Static method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- getFieldID() - Method in class org.apache.hadoop.record.meta.FieldTypeInfo
-
get the field's id (name)
- getFieldTypeInfos() - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
Return a collection of field type infos
- getFieldTypeInfos() - Method in class org.apache.hadoop.record.meta.StructTypeID
-
- getFile(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Get a local file name under a directory named in dirsProp with
the given path.
- getFileBlockLocations(FileStatus, long, long) - Method in class org.apache.hadoop.fs.FileSystem
-
Return an array containing hostnames, offset and size of
portions of the given file.
- getFileBlockLocations(FileStatus, long, long) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
- getFileBlockLocations(FileStatus, long, long) - Method in class org.apache.hadoop.fs.HarFileSystem
-
get block locations from the underlying fs
- getFileBlockLocations(FileStatus, long, long) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
Return null if the file doesn't exist; otherwise, get the
locations of the various chunks of the file file from KFS.
- getFileChecksum(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Get the checksum of a file.
- getFileChecksum(Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Get the checksum of a file.
- getFileClassPaths(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the file entries in classpath as an array of Path
- getFileCount() - Method in class org.apache.hadoop.fs.ContentSummary
-
- getFileName() - Method in class org.apache.hadoop.metrics.file.FileContext
-
Returns the configured file name, or null.
- getFiles(PathFilter) - Method in class org.apache.hadoop.fs.InMemoryFileSystem
-
Deprecated.
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Return a file status object that represents the path.
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Get file status.
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
return the filestatus of files in har archive.
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
FileStatus for S3 file systems.
- getFileStatus(Path) - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- getFilesystem() - Method in class org.apache.hadoop.fs.DF
-
- getFileSystem(Configuration) - Method in class org.apache.hadoop.fs.Path
-
Return the FileSystem that owns this Path.
- getFilesystemName() - Method in class org.apache.hadoop.mapred.JobTracker
-
Grab the local fs name
- getFileTimestamps(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the timestamps of the files
- getFileType() - Method in class org.apache.hadoop.fs.s3.INode
-
- getFinalSync(JobConf) - Static method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
-
Does the user want a final sync at close?
- getFinishTime() - Method in class org.apache.hadoop.mapred.TaskReport
-
Get finish time of task.
- getFirst() - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
- getFirstKey() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get the first key in the TFile.
- getFlippable() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
-
- getFloat(String, float) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as a float
.
- getFormatMinSplitSize() - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Get the lower bound on split size imposed by the format.
- getFormatMinSplitSize() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
-
- getFormattedTimeWithDiff(DateFormat, long, long) - Static method in class org.apache.hadoop.util.StringUtils
-
Formats time in ms and appends difference (finishTime - startTime)
as returned by formatTimeDiff().
- getFs() - Method in class org.apache.hadoop.mapred.JobClient
-
Get a filesystem handle.
- getFSSize() - Method in class org.apache.hadoop.fs.InMemoryFileSystem
-
Deprecated.
- getGeneration() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
Get the generation of the Lucene instance.
- getGET_PERMISSION_COMMAND() - Static method in class org.apache.hadoop.util.Shell
-
Return a Unix command to get permission information.
- getGroup() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the group associated with the file.
- getGroup(String) - Method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Returns the named counter group, or an empty group if there is none
with the specified name.
- getGroup(String) - Method in class org.apache.hadoop.mapreduce.Counters
-
Returns the named counter group, or an empty group if there is none
with the specified name.
- getGroupAction() - Method in class org.apache.hadoop.fs.permission.FsPermission
-
- getGroupingComparator() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the user defined
RawComparator
comparator for
grouping keys of inputs to the reduce.
- getGroupName() - Method in class org.apache.hadoop.fs.permission.PermissionStatus
-
Return group name
- getGroupNames() - Method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Returns the names of all counter classes.
- getGroupNames() - Method in class org.apache.hadoop.mapreduce.Counters
-
Returns the names of all counter classes.
- getGroupNames() - Method in class org.apache.hadoop.security.UnixUserGroupInformation
-
Return an array of group names
- getGroupNames() - Method in class org.apache.hadoop.security.UserGroupInformation
-
Get the name of the groups that the user belong to
- getGroups() - Method in class org.apache.hadoop.security.SecurityUtil.AccessControlList
-
- getGROUPS_COMMAND() - Static method in class org.apache.hadoop.util.Shell
-
a Unix command to get the current user's groups list
- getHadoopClientHome() - Method in class org.apache.hadoop.streaming.StreamJob
-
- getHarHash(Path) - Static method in class org.apache.hadoop.fs.HarFileSystem
-
the hash of the path p inside iniside
the filesystem
- getHarVersion() - Method in class org.apache.hadoop.fs.HarFileSystem
-
- getHashType(Configuration) - Static method in class org.apache.hadoop.util.hash.Hash
-
This utility method converts the name of the configured
hash type to a symbolic constant.
- getHeader(boolean) - Static method in class org.apache.hadoop.fs.ContentSummary
-
Return the header of the output.
- getHeader() - Method in class org.apache.hadoop.util.DataChecksum
-
- getHomeDirectory() - Method in class org.apache.hadoop.fs.FileSystem
-
Return the current user's home directory in this filesystem.
- getHomeDirectory() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
- getHomeDirectory() - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- getHomeDirectory() - Method in class org.apache.hadoop.fs.HarFileSystem
-
return the top level archive path.
- getHomeDirectory() - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- getHost() - Method in class org.apache.hadoop.streaming.Environment
-
- getHostname() - Static method in class org.apache.hadoop.util.StringUtils
-
Return hostname without throwing exception.
- getHosts() - Method in class org.apache.hadoop.fs.BlockLocation
-
Get the list of hosts (hostname) hosting this block
- getHosts(String, String) - Static method in class org.apache.hadoop.net.DNS
-
Returns all the host names associated by the provided nameserver with the
address bound to the specified network interface
- getHosts(String) - Static method in class org.apache.hadoop.net.DNS
-
Returns all the host names associated by the default nameserver with the
address bound to the specified network interface
- getHosts() - Method in class org.apache.hadoop.util.HostsFileReader
-
- getId() - Method in class org.apache.hadoop.fs.s3.Block
-
- getId(Class) - Method in class org.apache.hadoop.io.AbstractMapWritable
-
- getID() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the job identifier.
- getId() - Method in class org.apache.hadoop.mapreduce.ID
-
returns the int which represents the identifier
- GetImage() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get token literal value.
- getIndexFile(String) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- getIndexFile(String, boolean) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- getIndexInputFormatClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the index input format class.
- getIndexInterval() - Method in class org.apache.hadoop.io.MapFile.Writer
-
The number of entries that are added before an index entry is added.
- getIndexMaxFieldLength() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the max field length for a Lucene instance.
- getIndexMaxNumSegments() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the max number of segments for a Lucene instance.
- getIndexShards() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the string representation of a number of shards.
- getIndexShards(IndexUpdateConfiguration) - Static method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- getIndexUpdaterClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the index updater class.
- getIndexUseCompoundFile() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Check whether to use the compound file format for a Lucene instance.
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.CPUParser
-
Return a String with information about this class
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.HadoopLogParser
-
Return a String with information about this class
- getInfo() - Method in interface org.apache.hadoop.contrib.failmon.Monitored
-
Return a String with information about the implementing
class
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.NICParser
-
Return a String with information about this class
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.SensorsParser
-
Return a String with information about this class
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.SMARTParser
-
Return a String with information about this class
- getInfo() - Method in class org.apache.hadoop.contrib.failmon.SystemLogParser
-
Return a String with information about this class
- getInfo() - Static method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- getInfoPort() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getInputFileBasedOutputFileName(JobConf, String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Generate the outfile name based on a given anme and the input file name.
- getInputFormat() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getInputFormatClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
- getInputPathFilter(JobConf) - Static method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
Get a PathFilter instance of the filter set for the input paths.
- getInputPathFilter(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Get a PathFilter instance of the filter set for the input paths.
- getInputPaths(JobConf) - Static method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
Get the list of input
Path
s for the map-reduce job.
- getInputPaths(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Get the list of input
Path
s for the map-reduce job.
- getInputSplit() - Method in interface org.apache.hadoop.mapred.Reporter
-
- getInputSplit() - Method in class org.apache.hadoop.mapreduce.MapContext
-
Get the input split for this map.
- getInputStream(Socket) - Static method in class org.apache.hadoop.net.NetUtils
-
- getInputStream(Socket, long) - Static method in class org.apache.hadoop.net.NetUtils
-
Returns InputStream for the socket.
- getInstance(int) - Static method in class org.apache.hadoop.util.hash.Hash
-
Get a singleton instance of hash function of a given type.
- getInstance(Configuration) - Static method in class org.apache.hadoop.util.hash.Hash
-
Get a singleton instance of hash function of a type
defined in the configuration.
- getInstance() - Static method in class org.apache.hadoop.util.hash.JenkinsHash
-
- getInstance() - Static method in class org.apache.hadoop.util.hash.MurmurHash
-
- getInstrumentationClass(Configuration) - Static method in class org.apache.hadoop.mapred.JobTracker
-
- getInstrumentationClass(Configuration) - Static method in class org.apache.hadoop.mapred.TaskTracker
-
- getInt(String, int) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as an int
.
- getInterfaceName() - Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
-
Get the interface name
- getInterval(ArrayList<MonitorJob>) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Determines the minimum interval at which the executor thread
needs to wake upto execute jobs.
- getIOSortMB() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the IO sort space in MB.
- getIPs(String) - Static method in class org.apache.hadoop.net.DNS
-
Returns all the IPs associated with the provided interface, if any, in
textual form.
- getIsJavaMapper(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Check whether the job is using a Java Mapper.
- getIsJavaRecordReader(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Check whether the job is using a Java RecordReader
- getIsJavaRecordWriter(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Will the reduce use a Java RecordWriter?
- getIsJavaReducer(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Check whether the job is using a Java Reducer.
- getJar() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the user jar for the map-reduce job.
- getJar() - Method in class org.apache.hadoop.mapreduce.Job
-
Get the pathname of the job's jar.
- getJar() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the pathname of the job's jar.
- getJob(JobID) - Method in class org.apache.hadoop.mapred.JobClient
-
- getJob(String) - Method in class org.apache.hadoop.mapred.JobClient
-
- getJob(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJob() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- getJobClient() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getJobClient() - Method in class org.apache.hadoop.mapred.TaskTracker
-
The connection to the JobTracker, used by the TaskRunner
for locating remote files.
- getJobConf() - Method in class org.apache.hadoop.mapred.JobContext
-
Deprecated.
Get the job Configuration
- getJobConf() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getJobConf() - Method in class org.apache.hadoop.mapred.TaskAttemptContext
-
Deprecated.
- getJobCounters(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJobEndNotificationURI() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
- getJobFile() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the configuration file for the job.
- getJobFile() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the path of the submitted job configuration.
- getJobHistoryFileName(JobConf, JobID) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Recover the job history filename from the history folder.
- getJobHistoryLogLocation(String) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Get the job history file path given the history filename
- getJobHistoryLogLocationForUser(String, JobConf) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Get the user job history file path
- getJobID() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getJobID() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the job id.
- getJobId() - Method in class org.apache.hadoop.mapred.JobProfile
-
- getJobId() - Method in class org.apache.hadoop.mapred.JobStatus
-
- getJobID() - Method in class org.apache.hadoop.mapred.JobStatus
-
- getJobID() - Method in interface org.apache.hadoop.mapred.RunningJob
-
- getJobID() - Method in class org.apache.hadoop.mapred.TaskAttemptID
-
Deprecated.
- getJobID() - Method in class org.apache.hadoop.mapred.TaskID
-
Deprecated.
- getJobID() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the unique ID for the job.
- getJobID() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
Returns the
JobID
object that this task attempt belongs to
- getJobID() - Method in class org.apache.hadoop.mapreduce.TaskID
-
Returns the
JobID
object that this tip belongs to
- getJobIDsPattern(String, Integer) - Static method in class org.apache.hadoop.mapred.JobID
-
Deprecated.
- getJobLocalDir() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get job-specific shared directory for use as scratch space
- getJobName() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the user-specified job name.
- getJobName() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getJobName() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the user-specified job name.
- getJobName() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the name of the job.
- getJobName() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the user-specified job name.
- getJobPriority() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getJobPriority() - Method in class org.apache.hadoop.mapred.JobStatus
-
Return the priority of the job
- getJobProfile(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJobs() - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Scans the configuration file to determine which monitoring
utilities are available in the system.
- getJobsFromQueue(String) - Method in class org.apache.hadoop.mapred.JobClient
-
Gets all the jobs which were added to particular Job Queue
- getJobsFromQueue(String) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJobState() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Returns the current state of the Job.
- getJobStatus(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJobTrackerHostPort() - Method in class org.apache.hadoop.streaming.StreamJob
-
- getJobTrackerMachine() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getJobTrackerState() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
- getJtIdentifier() - Method in class org.apache.hadoop.mapreduce.JobID
-
- getJvmManagerInstance() - Method in class org.apache.hadoop.mapred.TaskTracker
-
- getKeepCommandFile(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Does the user want to keep the command file for debugging? If this is
true, pipes will write a copy of the command data to a file in the
task directory named "downlink.data", which may be used to run the C++
program under the debugger.
- getKeepFailedTaskFiles() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should the temporary files for failed tasks be kept?
- getKeepTaskFilesPattern() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the regular expression that is matched against the task names
to see if we need to keep the files.
- getKey(BytesWritable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy the key into BytesWritable.
- getKey(byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy the key into user supplied buffer.
- getKey(byte[], int) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy the key into user supplied buffer.
- getKey() - Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
-
Gets the current raw key
- getKey() - Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
Returns the stored rawKey
- getKey() - Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
-
Gets the current raw key.
- getKeyClass() - Method in class org.apache.hadoop.io.MapFile.Reader
-
Returns the class of keys in this file.
- getKeyClass() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the class of keys in this file.
- getKeyClass() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
Returns the class of keys in this file.
- getKeyClass() - Method in class org.apache.hadoop.io.WritableComparator
-
Returns the WritableComparable implementation class.
- getKeyClass() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- getKeyClass() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- getKeyClassName() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the name of the key class.
- getKeyClassName() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
Retrieve the name of the key class for this SequenceFile.
- getKeyFieldComparatorOption() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getKeyFieldPartitionerOption() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getKeyLength() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Get the length of the key.
- getKeyList() - Method in class org.apache.hadoop.metrics.util.MetricsRegistry
-
- getKeyNear(long) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get a sample key that is within a block whose starting offset is greater
than or equal to the specified offset.
- getKeyStream() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Streaming access to the key.
- getKeyTypeID() - Method in class org.apache.hadoop.record.meta.MapTypeID
-
get the TypeID of the map's key element
- getLastKey() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get the last key in the TFile.
- getLen() - Method in class org.apache.hadoop.fs.FileStatus
-
- getLength() - Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- getLength() - Method in class org.apache.hadoop.fs.BlockLocation
-
Get the length of the block
- getLength() - Method in class org.apache.hadoop.fs.ContentSummary
-
- getLength() - Method in class org.apache.hadoop.fs.FileChecksum
-
The length of the checksum in bytes
- getLength(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
- getLength(Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
Deprecated.
- getLength() - Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
The length of the checksum in bytes
- getLength() - Method in class org.apache.hadoop.fs.s3.Block
-
- getLength() - Method in class org.apache.hadoop.io.BinaryComparable
-
Return n st bytes 0..n-1 from {#getBytes()} are valid.
- getLength() - Method in class org.apache.hadoop.io.BytesWritable
-
Get the current size of the buffer.
- getLength() - Method in class org.apache.hadoop.io.DataInputBuffer
-
Returns the length of the input.
- getLength() - Method in class org.apache.hadoop.io.DataOutputBuffer
-
Returns the length of the valid data currently in the buffer.
- getLength() - Method in class org.apache.hadoop.io.InputBuffer
-
Returns the length of the input.
- getLength() - Method in class org.apache.hadoop.io.OutputBuffer
-
Returns the length of the valid data currently in the buffer.
- getLength() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
Returns the current length of the output file.
- getLength() - Method in class org.apache.hadoop.io.Text
-
Returns the number of bytes in the byte array
- getLength() - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
The number of bytes in the encoded string.
- getLength() - Method in class org.apache.hadoop.mapred.FileSplit
-
Deprecated.
The number of bytes in the file to process.
- getLength() - Method in interface org.apache.hadoop.mapred.InputSplit
-
Deprecated.
Get the total number of bytes in the data of the InputSplit
.
- getLength() - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
Return the aggregate length of all child InputSplits currently added.
- getLength(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
Get the length of ith child InputSplit.
- getLength() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- getLength(int) - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns the length of the ith Path
- getLength() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
-
- getLength() - Method in class org.apache.hadoop.mapreduce.InputSplit
-
Get the size of the split, so that the input splits can be sorted by size.
- getLength() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
The number of bytes in the file to process.
- getLengths() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns an array containing the lengths of the files in the split
- getLevel() - Method in interface org.apache.hadoop.net.Node
-
Return this node's level in the tree.
- getLevel() - Method in class org.apache.hadoop.net.NodeBase
-
Return this node's level in the tree.
- getLibJars(Configuration) - Static method in class org.apache.hadoop.util.GenericOptionsParser
-
If libjars are set in the conf, parse the libjars.
- getLine() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Deprecated.
- getLinkCount(File) - Static method in class org.apache.hadoop.fs.FileUtil.HardLink
-
Retrieves the number of links to the specified file.
- getListenerAddress() - Method in class org.apache.hadoop.ipc.Server
-
Return the socket (ip+port) on which the RPC server is listening to.
- getLoadNativeLibraries(Configuration) - Method in class org.apache.hadoop.util.NativeCodeLoader
-
Return if native hadoop libraries, if present, can be used for this job.
- getLocal(Configuration) - Static method in class org.apache.hadoop.fs.FileSystem
-
Get the local file syste
- getLocalAnalysisClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the local analysis class.
- getLocalCache(URI, Configuration, Path, FileStatus, boolean, long, Path) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the locally cached file or archive; it could either be
previously cached (and valid) or copy it from the
FileSystem
now.
- getLocalCache(URI, Configuration, Path, FileStatus, boolean, long, Path, boolean) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the locally cached file or archive; it could either be
previously cached (and valid) or copy it from the
FileSystem
now.
- getLocalCache(URI, Configuration, Path, boolean, long, Path) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Get the locally cached file or archive; it could either be
previously cached (and valid) or copy it from the
FileSystem
now.
- getLocalCacheArchives(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Return the path array of the localized caches
- getLocalCacheFiles(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Return the path array of the localized files
- getLocalDirs() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getLocalJobFilePath(JobID) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Get the path of the locally stored job file
- getLocalJobFilePath(JobID) - Static method in class org.apache.hadoop.mapred.JobTracker
-
Get the localized job file path on the job trackers local file system
- getLocalPath(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Get a local file under a directory named by dirsProp with
the given path.
- getLocalPath(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Constructs a local file name.
- getLocalPathForWrite(String, Configuration) - Method in class org.apache.hadoop.fs.LocalDirAllocator
-
Get a path from the local FS.
- getLocalPathForWrite(String, long, Configuration) - Method in class org.apache.hadoop.fs.LocalDirAllocator
-
Get a path from the local FS.
- getLocalPathToRead(String, Configuration) - Method in class org.apache.hadoop.fs.LocalDirAllocator
-
Get a path from the local FS for reading.
- getLocation(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
getLocations from ith InputSplit.
- getLocations() - Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- getLocations() - Method in class org.apache.hadoop.mapred.FileSplit
-
Deprecated.
- getLocations() - Method in interface org.apache.hadoop.mapred.InputSplit
-
Deprecated.
Get the list of hostnames where the input split is located.
- getLocations() - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
Collect a set of hosts from all child InputSplits.
- getLocations() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns all the Paths where this input-split resides
- getLocations() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
-
Get the list of hostnames where the input split is located.
- getLocations() - Method in class org.apache.hadoop.mapred.MultiFileSplit
-
Deprecated.
- getLocations() - Method in class org.apache.hadoop.mapreduce.InputSplit
-
Get the list of nodes by name where the data for the split would be local.
- getLocations() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
- getLong(String, long) - Method in class org.apache.hadoop.conf.Configuration
-
Get the value of the name
property as a long
.
- getLongValue(Object) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
- getMajor() - Method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
Get the major version.
- getMap() - Method in class org.apache.hadoop.contrib.failmon.EventRecord
-
Return the HashMap of properties of the EventRecord.
- getMapCompletionEvents(JobID, int, int, TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskTracker
-
- getMapDebugScript() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the map task's debug script.
- getMapOutputCompressorClass(Class<? extends CompressionCodec>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getMapOutputKeyClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
-
Get the map output key class.
- getMapOutputKeyClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the key class for the map output data.
- getMapOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the key class for the map output data.
- getMapOutputValueClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
-
Get the map output value class.
- getMapOutputValueClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the value class for the map output data.
- getMapOutputValueClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the value class for the map output data.
- getMapper() - Method in class org.apache.hadoop.mapred.MapRunner
-
- getMapperClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the
Mapper
class for the job.
- getMapperClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the
Mapper
class for the job.
- getMapperClass(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
Get the application's mapper class.
- getMapperMaxSkipRecords(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Get the number of acceptable skip records surrounding the bad record PER
bad record in mapper.
- getMapredJobID() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getMapredTempDir() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Get the Map/Reduce temp directory.
- getMapRunnerClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getMapSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should speculative execution be used for this job for map tasks?
Defaults to true
.
- getMapTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
-
Get the information of the current state of the map tasks of a job.
- getMapTaskReports(String) - Method in class org.apache.hadoop.mapred.JobClient
-
- getMapTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getMapTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the number of currently running map tasks in the cluster.
- getMaxDepth(int) - Static method in class org.apache.hadoop.util.QuickSort
-
Deepest recursion before giving up and doing a heapsort.
- getMaxMapAttempts() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the configured number of maximum attempts that will be made to run a
map task, as specified by the mapred.map.max.attempts
property.
- getMaxMapTaskFailuresPercent() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the maximum percentage of map tasks that can fail without
the job being aborted.
- getMaxMapTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the maximum capacity for running map tasks in the cluster.
- getMaxMemory() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the maximum configured heap memory that can be used by the JobTracker
- getMaxPhysicalMemoryForTask() - Method in class org.apache.hadoop.mapred.JobConf
-
- getMaxReduceAttempts() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the configured number of maximum attempts that will be made to run a
reduce task, as specified by the mapred.reduce.max.attempts
property.
- getMaxReduceTaskFailuresPercent() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the maximum percentage of reduce tasks that can fail without
the job being aborted.
- getMaxReduceTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the maximum capacity for running reduce tasks in the cluster.
- getMaxSplitSize(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Get the maximum split size.
- getMaxTaskFailuresPerTracker() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Expert: Get the maximum no.
- getMaxTime() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
-
- getMaxVirtualMemoryForTask() - Method in class org.apache.hadoop.mapred.JobConf
-
- getMBeanInfo() - Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
-
- getMD5Hash(String) - Static method in class org.apache.hadoop.contrib.failmon.Anonymizer
-
Create the MD5 digest of an input text.
- getMemory() - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Get the total amount of buffer memory, in bytes.
- getMemoryCalculatorPlugin(Class<? extends MemoryCalculatorPlugin>, Configuration) - Static method in class org.apache.hadoop.util.MemoryCalculatorPlugin
-
Get the MemoryCalculatorPlugin from the class name and configure it.
- getMemoryForMapTask() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get memory required to run a map task of the job, in MB.
- getMemoryForReduceTask() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get memory required to run a reduce task of the job, in MB.
- getMessage() - Method in exception org.apache.hadoop.mapred.InvalidInputException
-
Get a summary message of the problems found.
- getMessage() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getMessage() - Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
-
Get a summary message of the problems found.
- getMessage() - Method in error org.apache.hadoop.record.compiler.generated.TokenMgrError
-
You can also modify the body of this method to customize your error messages.
- getMetaBlock(String) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Stream access to a meta block.``
- getMetadata() - Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- getMetadata() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the metadata object of the file
- getMetric(String) - Method in class org.apache.hadoop.metrics.spi.OutputRecord
-
Returns the metric object which can be a Float, Integer, Short or Byte.
- getMetricNames() - Method in class org.apache.hadoop.metrics.spi.OutputRecord
-
Returns the set of metric names.
- getMetricsList() - Method in class org.apache.hadoop.metrics.util.MetricsRegistry
-
- getMinor() - Method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
Get the minor version.
- getMinSplitSize(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Get the minimum split size
- getMinTime() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
-
- getModificationTime() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the modification time of the file.
- getMount() - Method in class org.apache.hadoop.fs.DF
-
- getName() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
-
- getName() - Method in class org.apache.hadoop.fs.FileSystem
-
- getName() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
- getName() - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
Deprecated.
- getName() - Method in class org.apache.hadoop.fs.Path
-
Returns the final component of this path.
- getName() - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- getName(Class) - Static method in class org.apache.hadoop.io.WritableName
-
Return the name for a class.
- getName() - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Returns raw name of the group.
- getName() - Method in class org.apache.hadoop.mapreduce.Counter
-
- getName() - Method in class org.apache.hadoop.mapreduce.CounterGroup
-
Get the internal name of the group
- getName() - Method in class org.apache.hadoop.metrics.util.MetricsBase
-
- getName() - Method in interface org.apache.hadoop.net.Node
-
Return this node's name
- getName() - Method in class org.apache.hadoop.net.NodeBase
-
Return this node's name
- getName() - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
return the name of the record
- getName() - Method in class org.apache.hadoop.security.Group
-
- getName() - Method in class org.apache.hadoop.security.UnixUserGroupInformation
-
- getName() - Method in class org.apache.hadoop.security.User
-
- getNamed(String, Configuration) - Static method in class org.apache.hadoop.fs.FileSystem
-
- getNamedOutputFormatClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns the named output OutputFormat.
- getNamedOutputKeyClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns the key class for a named output.
- getNamedOutputs() - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns iterator with the defined name outputs.
- getNamedOutputsList(JobConf) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns list of channel names.
- getNamedOutputValueClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Returns the value class for a named output.
- getNames() - Method in class org.apache.hadoop.fs.BlockLocation
-
Get the list of names (hostname:port) hosting this block
- getNestedStructTypeInfo(String) - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
Return the type info of a nested record.
- getNetworkLocation() - Method in interface org.apache.hadoop.net.Node
-
Return the string representation of this node's network location
- getNetworkLocation() - Method in class org.apache.hadoop.net.NodeBase
-
Return this node's network location
- getNewJobId() - Method in class org.apache.hadoop.mapred.JobTracker
-
Allocates a new JobId string.
- getNext() - Method in class org.apache.hadoop.contrib.failmon.LogParser
-
Continue parsing the log file until a valid log entry is identified.
- getNextHeartbeatInterval() - Method in class org.apache.hadoop.mapred.JobTracker
-
Calculates next heartbeat interval using cluster size.
- getNextToken() - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Get the next Token.
- getNextToken() - Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Get the next Token.
- getNode(String) - Method in class org.apache.hadoop.mapred.JobTracker
-
Return the Node in the network topology that corresponds to the hostname
- getNode() - Method in class org.apache.hadoop.mapred.join.Parser.NodeToken
-
- getNode() - Method in class org.apache.hadoop.mapred.join.Parser.Token
-
- getNode(String) - Method in class org.apache.hadoop.net.NetworkTopology
-
Given a string representation of a node, return its reference
- getNodesAtMaxLevel() - Method in class org.apache.hadoop.mapred.JobTracker
-
Returns a collection of nodes at the max level
- getNullContext(String) - Static method in class org.apache.hadoop.metrics.ContextFactory
-
Returns a "null" context - one which does nothing.
- getNum() - Method in class org.apache.hadoop.mapred.join.Parser.NumToken
-
- getNum() - Method in class org.apache.hadoop.mapred.join.Parser.Token
-
- getNumber() - Method in class org.apache.hadoop.metrics.spi.MetricValue
-
- getNumberColumns() - Method in class org.apache.hadoop.examples.dancing.DancingLinks
-
Get the number of columns.
- getNumberOfThreads(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
The number of threads in the thread pool that will run the map function.
- getNumberOfUniqueHosts() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getNumBytesInSum() - Method in class org.apache.hadoop.util.DataChecksum
-
- getNumFiles(PathFilter) - Method in class org.apache.hadoop.fs.InMemoryFileSystem
-
Deprecated.
- getNumMapTasks() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get configured the number of reduce tasks for this job.
- getNumOfLeaves() - Method in class org.apache.hadoop.net.NetworkTopology
-
Return the total number of nodes
- getNumOfRacks() - Method in class org.apache.hadoop.net.NetworkTopology
-
Return the total number of racks
- getNumOpenConnections() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The number of open RPC conections
- getNumOpenConnections() - Method in class org.apache.hadoop.ipc.Server
-
The number of open RPC conections
- getNumPaths() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns the number of Paths in the split
- getNumReduceTasks() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get configured the number of reduce tasks for this job.
- getNumReduceTasks() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get configured the number of reduce tasks for this job.
- getNumResolvedTaskTrackers() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getNumTaskCacheLevels() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getNumTasksToExecutePerJvm() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the number of tasks that a spawned JVM should execute
- getOffset() - Method in class org.apache.hadoop.fs.BlockLocation
-
Get the start offset of file associated with this block
- getOffset(int) - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns the start offset of the ith Path
- getOp() - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
Get the type of the operation.
- getOp() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Get the type of operation.
- getOpt(String) - Method in class org.apache.hadoop.fs.shell.CommandFormat
-
Return if the option is set or not
- getOtherAction() - Method in class org.apache.hadoop.fs.permission.FsPermission
-
- getOutput() - Method in class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
Get the output of the shell command.
- getOutputCommitter() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
-
- getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
-
Get the output committer for this output format.
- getOutputCommitter() - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- getOutputCompressionType(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
Deprecated.
- getOutputCompressionType(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
-
- getOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
- getOutputCompressorClass(JobContext, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- getOutputFormat() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getOutputFormatClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
- getOutputKeyClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
Get the reduce output key class.
- getOutputKeyClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the key class for the job output data.
- getOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the key class for the job output data.
- getOutputKeyComparator() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getOutputPath(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Get the
Path
to the output directory for the map-reduce job.
- getOutputPath(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Get the
Path
to the output directory for the map-reduce job.
- getOutputStream(Socket) - Static method in class org.apache.hadoop.net.NetUtils
-
Same as getOutputStream(socket, 0).
- getOutputStream(Socket, long) - Static method in class org.apache.hadoop.net.NetUtils
-
Returns OutputStream for the socket.
- getOutputValueClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
Get the reduce output value class.
- getOutputValueClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the value class for job outputs.
- getOutputValueClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the value class for job outputs.
- getOutputValueGroupingComparator() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the user defined
WritableComparable
comparator for
grouping keys of inputs to the reduce.
- getOwner() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the owner of the file.
- getParameter(ServletRequest, String) - Static method in class org.apache.hadoop.util.ServletUtil
-
Get a parameter from a ServletRequest.
- getParent() - Method in class org.apache.hadoop.fs.Path
-
Returns the parent of a path or null if at root.
- getParent() - Method in interface org.apache.hadoop.net.Node
-
Return this node's parent
- getParent() - Method in class org.apache.hadoop.net.NodeBase
-
Return this node's parent
- getParentNode(Node, int) - Static method in class org.apache.hadoop.mapred.JobTracker
-
- getPartition(Shard, IntermediateForm, int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
-
- getPartition(SecondarySort.IntPair, IntWritable, int) - Method in class org.apache.hadoop.examples.SecondarySort.FirstPartitioner
-
- getPartition(IntWritable, NullWritable, int) - Method in class org.apache.hadoop.examples.SleepJob
-
- getPartition(K2, V2, int) - Method in class org.apache.hadoop.mapred.lib.HashPartitioner
-
Deprecated.
- getPartition(K2, V2, int) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
-
- getPartition(int, int) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
-
- getPartition(K, V, int) - Method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
-
- getPartition(K2, V2, int) - Method in interface org.apache.hadoop.mapred.Partitioner
-
Deprecated.
Get the paritition number for a given key (hence record) given the total
number of partitions i.e.
- getPartition(K, V, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner
-
- getPartition(KEY, VALUE, int) - Method in class org.apache.hadoop.mapreduce.Partitioner
-
Get the partition number for a given key (hence record) given the total
number of partitions i.e.
- getPartitionerClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getPartitionerClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
- getPartitionFile(JobConf) - Static method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
-
Get the path to the SequenceFile storing the sorted partition keyset.
- getPath() - Method in class org.apache.hadoop.fs.FileStatus
-
- getPath() - Method in class org.apache.hadoop.mapred.FileSplit
-
Deprecated.
The file containing this split's data.
- getPath(int) - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns the ith Path
- getPath() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
The file containing this split's data.
- getPath(Node) - Static method in class org.apache.hadoop.net.NodeBase
-
Return this node's path
- getPathForCustomFile(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Helper function to generate a
Path
for a file that is unique for
the task within the job output directory.
- getPathForWorkFile(TaskInputOutputContext<?, ?, ?, ?>, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Helper function to generate a
Path
for a file that is unique for
the task within the job output directory.
- getPaths() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns all the Paths in the split
- getPercentUsed() - Method in class org.apache.hadoop.fs.DF
-
- getPercentUsed() - Method in class org.apache.hadoop.fs.InMemoryFileSystem
-
Deprecated.
- getPeriod() - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Returns the timer period.
- getPeriod() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Returns the timer period.
- getPermission() - Method in class org.apache.hadoop.fs.FileStatus
-
Get FsPermission associated with the file.
- getPermission() - Method in class org.apache.hadoop.fs.permission.PermissionStatus
-
Return permission
- getPermission() - Method in class org.apache.hadoop.security.authorize.Service
-
Get the
Permission
required to access the service.
- getPermissions(ProtectionDomain) - Method in class org.apache.hadoop.security.authorize.ConfiguredPolicy
-
- getPhysicalMemorySize() - Method in class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin
-
Obtain the total size of the physical memory present in the system.
- getPhysicalMemorySize() - Method in class org.apache.hadoop.util.MemoryCalculatorPlugin
-
Obtain the total size of the physical memory present in the system.
- getPidFromPidFile(String) - Static method in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
Get PID from a pid-file.
- getPlatformName() - Static method in class org.apache.hadoop.util.PlatformName
-
Get the complete platform as per the java-vm.
- getPolicy() - Static method in class org.apache.hadoop.security.SecurityUtil
-
Get the current global security policy for Hadoop.
- getPort() - Method in class org.apache.hadoop.http.HttpServer
-
Get the port that the server is on
- getPos() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- getPos() - Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- getPos() - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- getPos() - Method in exception org.apache.hadoop.fs.ChecksumException
-
- getPos() - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- getPos() - Method in class org.apache.hadoop.fs.FSDataOutputStream
-
- getPos() - Method in class org.apache.hadoop.fs.FSInputChecker
-
- getPos() - Method in class org.apache.hadoop.fs.FSInputStream
-
Return the current offset from the start of the file
- getPos() - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- getPos() - Method in interface org.apache.hadoop.fs.Seekable
-
Return the current offset from the start of the file
- getPos() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Unsupported (returns zero in all cases).
- getPos() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Request position from proxied RR.
- getPos() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- getPos() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
return the amount of data processed
- getPos() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
Returns the current position in the input.
- getPos() - Method in class org.apache.hadoop.mapred.LineRecordReader
-
Deprecated.
- getPos() - Method in interface org.apache.hadoop.mapred.RecordReader
-
Returns the current position in the input.
- getPos() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- getPos() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- getPos() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- getPos() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
Returns the current position in the input.
- getPosition() - Method in class org.apache.hadoop.io.DataInputBuffer
-
Returns the current position in the input.
- getPosition() - Method in class org.apache.hadoop.io.InputBuffer
-
Returns the current position in the input.
- getPosition() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Return the current byte position in the input file.
- getPreviousIntervalAverageTime() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
-
The average rate of an operation in the previous interval
- getPreviousIntervalNumOps() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
-
The number of operations in the previous interval
- getPreviousIntervalValue() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt
-
The Value at the Previous interval
- getPreviousIntervalValue() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong
-
The Value at the Previous interval
- getProblems() - Method in exception org.apache.hadoop.mapred.InvalidInputException
-
Get the complete list of the problems reported.
- getProblems() - Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
-
Get the complete list of the problems reported.
- getProcess() - Method in class org.apache.hadoop.util.Shell
-
get the current sub-process executing the given command
- getProcessTree() - Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
Get the process-tree with latest state.
- getProfileEnabled() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get whether the task profiling is enabled.
- getProfileParams() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the profiler configuration arguments.
- getProfileTaskRange(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the range of maps or reduces to profile.
- getProgress() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
-
- getProgress() - Method in class org.apache.hadoop.examples.MultiFileWordCount.MultiFileLineRecordReader
-
- getProgress() - Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
-
Gets the Progress object; this has a float (0.0 - 1.0)
indicating the bytes processed by the iterator so far
- getProgress() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Report progress as the minimum of all child RR progress.
- getProgress() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Request progress from proxied RR.
- getProgress() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
-
- getProgress() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
return progress based on the amount of data processed so far.
- getProgress() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
- getProgress() - Method in class org.apache.hadoop.mapred.LineRecordReader
-
Deprecated.
Get the progress within the split
- getProgress() - Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
-
Gets the Progress object; this has a float (0.0 - 1.0)
indicating the bytes processed by the iterator so far
- getProgress() - Method in interface org.apache.hadoop.mapred.RecordReader
-
- getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
Return the progress within the input split
- getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
Return the progress within the input split
- getProgress() - Method in class org.apache.hadoop.mapred.TaskReport
-
The amount completed, between zero and one.
- getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.LineRecordReader
-
Get the progress within the split
- getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
Return the progress within the input split
- getProgress() - Method in class org.apache.hadoop.mapreduce.RecordReader
-
The current progress of the record reader through its data.
- getProgress() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- getProgressible() - Method in class org.apache.hadoop.mapred.JobContext
-
Deprecated.
Get the progress mechanism for reporting progress.
- getProgressible() - Method in class org.apache.hadoop.mapred.TaskAttemptContext
-
Deprecated.
- getProperty(String) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Fetches the value of a property from the configuration file.
- getProtocolVersion(String, long) - Method in interface org.apache.hadoop.ipc.VersionedProtocol
-
Return protocol version corresponding to protocol interface.
- getProtocolVersion(String, long) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getProtocolVersion(String, long) - Method in class org.apache.hadoop.mapred.TaskTracker
-
- getProxy(Class<?>, long, InetSocketAddress, Configuration, SocketFactory) - Static method in class org.apache.hadoop.ipc.RPC
-
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
- getProxy(Class<?>, long, InetSocketAddress, UserGroupInformation, Configuration, SocketFactory) - Static method in class org.apache.hadoop.ipc.RPC
-
Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.
- getProxy(Class<?>, long, InetSocketAddress, Configuration) - Static method in class org.apache.hadoop.ipc.RPC
-
Construct a client-side proxy object with the default SocketFactory
- getQueueInfo(String) - Method in class org.apache.hadoop.mapred.JobClient
-
Gets the queue information associated to a particular Job Queue
- getQueueInfo(String) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getQueueManager() - Method in class org.apache.hadoop.mapred.JobTracker
-
Return the QueueManager
associated with the JobTracker.
- getQueueName() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Return the name of the queue to which this job is submitted.
- getQueueName() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the name of the queue to which the job is submitted.
- getQueueName() - Method in class org.apache.hadoop.mapred.JobQueueInfo
-
Get the queue name from JobQueueInfo
- getQueues() - Method in class org.apache.hadoop.mapred.JobClient
-
Return an array of queue information objects about all the Job Queues
configured.
- getQueues() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getQuota() - Method in class org.apache.hadoop.fs.ContentSummary
-
Return the directory quota
- getRange(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Parse the given attribute as a set of integer ranges
- getRaw(String) - Method in class org.apache.hadoop.conf.Configuration
-
- getRaw() - Method in class org.apache.hadoop.fs.LocalFileSystem
-
- getRawFileSystem() - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
get the raw file system
- getReader() - Method in class org.apache.hadoop.contrib.failmon.LogParser
-
Return the BufferedReader, that reads the log file
- getReaders(FileSystem, Path, Configuration) - Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
-
Open the output generated by this format.
- getReaders(Configuration, Path) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
Deprecated.
Open the output generated by this format.
- getReadyJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getRealTaskLogFileLocation(TaskAttemptID, TaskLog.LogName) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- getRecordName() - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Returns the record name.
- getRecordName() - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Returns the record name.
- getRecordNum() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Get the RecordNum corresponding to the entry pointed by the cursor.
- getRecordNumNear(long) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Get the RecordNum for the first key-value pair in a compressed block
whose byte offset in the TFile is greater than or equal to the specified
offset.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.contrib.index.example.LineDocInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.examples.MultiFileWordCount.MyInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.examples.SleepJob.SleepInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in interface org.apache.hadoop.mapred.InputFormat
-
Deprecated.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in interface org.apache.hadoop.mapred.join.ComposableInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Construct a CompositeRecordReader for the children of this InputFormat
as defined in the init expression.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
This is not implemented yet.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.DelegatingInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.MultiFileInputFormat
-
Deprecated.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter
-
Create a record reader for the given split
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileInputFormat
-
Deprecated.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.TextInputFormat
-
Deprecated.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.streaming.StreamInputFormat
-
- getRecordReaderQueue() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Return sorted list of RecordReaders for this composite.
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.FileOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
-
Create a composite record writer that can write key/value data to different
output files
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.NullOutputFormat
-
Deprecated.
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.MapFileOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in interface org.apache.hadoop.mapred.OutputFormat
-
Deprecated.
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
Deprecated.
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.TextOutputFormat
-
Deprecated.
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
-
- getRecoveryDuration() - Method in class org.apache.hadoop.mapred.JobTracker
-
How long the jobtracker took to recover from restart.
- getReduceDebugScript() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the reduce task's debug Script
- getReducerClass() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- getReducerClass() - Method in class org.apache.hadoop.mapreduce.JobContext
-
- getReducerMaxSkipGroups(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Get the number of acceptable skip groups surrounding the bad group PER
bad group in reducer.
- getReduceSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should speculative execution be used for this job for reduce tasks?
Defaults to true
.
- getReduceTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
-
Get the information of the current state of the reduce tasks of a job.
- getReduceTaskReports(String) - Method in class org.apache.hadoop.mapred.JobClient
-
- getReduceTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getReduceTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the number of currently running reduce tasks in the cluster.
- getRemainingArgs() - Method in class org.apache.hadoop.util.GenericOptionsParser
-
Returns an array of Strings containing only application-specific arguments.
- getRemoteAddress() - Static method in class org.apache.hadoop.ipc.Server
-
Returns remote address as a string when invoked inside an RPC.
- getRemoteIp() - Static method in class org.apache.hadoop.ipc.Server
-
Returns the remote side ip address when invoked inside an RPC
Returns null incase of an error.
- getReplication() - Method in class org.apache.hadoop.fs.FileStatus
-
Get the replication factor of a file.
- getReplication(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
- getReplication(Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
Deprecated.
- getReport() - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
log the counters
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
-
- getReport() - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
-
- getReport() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
- getReportDetails() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
- getReportItems() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
- getResource(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get the
URL
for the named resource.
- getResult() - Method in class org.apache.hadoop.examples.Sort
-
Get the last job that was run using this instance.
- getRevision() - Static method in class org.apache.hadoop.util.VersionInfo
-
Get the subversion revision number for the root directory
- getRotations() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
-
- getRpcOpsAvgProcessingTime() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
Average time for RPC Operations in last interval
- getRpcOpsAvgProcessingTimeMax() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The Maximum RPC Operation Processing Time since reset was called
- getRpcOpsAvgProcessingTimeMin() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The Minimum RPC Operation Processing Time since reset was called
- getRpcOpsAvgQueueTime() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The Average RPC Operation Queued Time in the last interval
- getRpcOpsAvgQueueTimeMax() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The Maximum RPC Operation Queued Time since reset was called
- getRpcOpsAvgQueueTimeMin() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
The Minimum RPC Operation Queued Time since reset was called
- getRpcOpsNumber() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
Number of RPC Operations in the last interval
- getRunnable() - Method in class org.apache.hadoop.util.Daemon
-
- getRunningJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getRunningJobs() - Method in class org.apache.hadoop.mapred.JobTracker
-
Version that is called from a timer thread, and therefore needs to be
careful to synchronize.
- getRunningTaskAttempts() - Method in class org.apache.hadoop.mapred.TaskReport
-
Get the running task attempt IDs for this task
- getRunState() - Method in class org.apache.hadoop.mapred.JobStatus
-
- getSample(InputFormat<K, V>, JobConf) - Method in class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler
-
For each split sampled, emit when the ratio of the number of records
retained to the total record count is less than the specified
frequency.
- getSample(InputFormat<K, V>, JobConf) - Method in class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
-
Randomize the split order, then take the specified number of keys from
each split sampled, where each key is selected with the specified
probability and possibly replaced by a subsequently selected key when
the quota of keys from that split is satisfied.
- getSample(InputFormat<K, V>, JobConf) - Method in interface org.apache.hadoop.mapred.lib.InputSampler.Sampler
-
For a given job, collect and return a subset of the keys from the
input data.
- getSample(InputFormat<K, V>, JobConf) - Method in class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
-
From each split sampled, take the first numSamples / numSplits records.
- getSchedulingInfo() - Method in class org.apache.hadoop.mapred.JobQueueInfo
-
Gets the scheduling information associated to particular job queue.
- getSchedulingInfo() - Method in class org.apache.hadoop.mapred.JobStatus
-
Gets the Scheduling information associated to a particular Job.
- getScheme() - Method in class org.apache.hadoop.fs.FileSystem.Statistics
-
Get the uri scheme associated with this statistics object.
- getSecond() - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
- getSecretAccessKey() - Method in class org.apache.hadoop.fs.s3.S3Credentials
-
- getSelectQuery() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
-
Returns the query for selecting the records,
subclasses can override this for custom behaviour.
- getSequenceFileOutputKeyClass(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- getSequenceFileOutputValueClass(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- getSerialization(Class<T>) - Method in class org.apache.hadoop.io.serializer.SerializationFactory
-
- getSerializedLength() - Method in class org.apache.hadoop.fs.s3.INode
-
- getSerializer(Class<Serializable>) - Method in class org.apache.hadoop.io.serializer.JavaSerialization
-
- getSerializer(Class<T>) - Method in interface org.apache.hadoop.io.serializer.Serialization
-
- getSerializer(Class<T>) - Method in class org.apache.hadoop.io.serializer.SerializationFactory
-
- getSerializer(Class<Writable>) - Method in class org.apache.hadoop.io.serializer.WritableSerialization
-
- getServer(Object, String, int, Configuration) - Static method in class org.apache.hadoop.ipc.RPC
-
Construct a server for a protocol implementation instance listening on a
port and address.
- getServer(Object, String, int, int, boolean, Configuration) - Static method in class org.apache.hadoop.ipc.RPC
-
Construct a server for a protocol implementation instance listening on a
port and address.
- getServerAddress(Configuration, String, String, String) - Static method in class org.apache.hadoop.net.NetUtils
-
Deprecated.
- getServerVersion() - Method in exception org.apache.hadoop.ipc.RPC.VersionMismatch
-
Get the server's agreed to version.
- getServiceKey() - Method in class org.apache.hadoop.security.authorize.Service
-
Get the configuration key for the service.
- getServices() - Method in class org.apache.hadoop.mapred.MapReducePolicyProvider
-
- getServices() - Method in class org.apache.hadoop.security.authorize.PolicyProvider
-
- getSessionId() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the user-specified session identifier.
- getSetupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
-
Get the information of the current state of the setup tasks of a job.
- getSetupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getShape(boolean, int) - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
-
- getSize() - Method in class org.apache.hadoop.io.BytesWritable
-
- getSize() - Method in interface org.apache.hadoop.io.SequenceFile.ValueBytes
-
Size of stored data.
- getSize() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- getSkipOutputPath(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Get the directory to which skipped records are written.
- getSlope(String) - Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
-
- getSocketFactory(Configuration, Class<?>) - Static method in class org.apache.hadoop.net.NetUtils
-
Get the socket factory for the given class according to its
configuration parameter
hadoop.rpc.socket.factory.class.<ClassName>.
- getSocketFactoryFromProperty(Configuration, String) - Static method in class org.apache.hadoop.net.NetUtils
-
Get the socket factory corresponding to the given proxy URI.
- getSortComparator() - Method in class org.apache.hadoop.mapreduce.JobContext
-
- getSpace(int) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- getSpaceConsumed() - Method in class org.apache.hadoop.fs.ContentSummary
-
Retuns (disk) space consumed
- getSpaceQuota() - Method in class org.apache.hadoop.fs.ContentSummary
-
Returns (disk) space quota
- getSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should speculative execution be used for this job?
Defaults to true
.
- getSplitHosts(BlockLocation[], long, long, NetworkTopology) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
This function identifies and returns the hosts that contribute
most for a given split.
- getSplits(int) - Method in class org.apache.hadoop.examples.dancing.Pentomino
-
Generate a list of prefixes to a given depth
- getSplits(JobConf, int) - Method in class org.apache.hadoop.examples.SleepJob.SleepInputFormat
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
- getSplits(JobConf, int) - Method in interface org.apache.hadoop.mapred.InputFormat
-
Deprecated.
Logically split the set of input files for the job.
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Build a CompositeInputSplit from the child InputFormats by assigning the
ith split from each child to the ith composite split.
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
Logically split the set of input files for the job.
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.DelegatingInputFormat
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
-
Logically splits the set of input files for the job, splits N lines
of the input as one split.
- getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.MultiFileInputFormat
-
Deprecated.
- getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.InputFormat
-
Logically split the set of input files for the job.
- getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Generate the list of files and make them into FileSplits.
- getStackTrace() - Method in exception org.apache.hadoop.security.authorize.AuthorizationException
-
- getStart() - Method in class org.apache.hadoop.mapred.FileSplit
-
Deprecated.
The position of the first byte in the file to process.
- getStart() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
-
- getStart() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
The position of the first byte in the file to process.
- getStartOffsets() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
Returns an array containing the startoffsets of the files in the split
- getStartTime() - Method in class org.apache.hadoop.mapred.JobStatus
-
- getStartTime() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getStartTime() - Method in class org.apache.hadoop.mapred.TaskReport
-
Get start time of task.
- getState(String) - Static method in class org.apache.hadoop.contrib.failmon.PersistentState
-
Read and return the state of parsing for a particular log file.
- getState() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- getState() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getState() - Method in class org.apache.hadoop.mapred.TaskReport
-
The most recent state, reported by a
Reporter
.
- getStaticResolution(String) - Static method in class org.apache.hadoop.net.NetUtils
-
Retrieves the resolved name for the passed host.
- getStatistics() - Static method in class org.apache.hadoop.fs.FileSystem
-
- getStatistics(String, Class<? extends FileSystem>) - Static method in class org.apache.hadoop.fs.FileSystem
-
Get the statistics for a particular file system
- getStatus() - Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
-
Get the last set status message.
- getStr() - Method in class org.apache.hadoop.mapred.join.Parser.StrToken
-
- getStr() - Method in class org.apache.hadoop.mapred.join.Parser.Token
-
- getStringCollection(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get the comma delimited values of the name
property as
a collection of String
s.
- getStringCollection(String) - Static method in class org.apache.hadoop.util.StringUtils
-
Returns a collection of strings.
- getStrings(String) - Method in class org.apache.hadoop.conf.Configuration
-
Get the comma delimited values of the name
property as
an array of String
s.
- getStrings(String, String...) - Method in class org.apache.hadoop.conf.Configuration
-
Get the comma delimited values of the name
property as
an array of String
s.
- getStrings(String) - Static method in class org.apache.hadoop.util.StringUtils
-
Returns an arraylist of strings.
- getSubject(UserGroupInformation) - Static method in class org.apache.hadoop.security.SecurityUtil
-
Get the
Subject
for the user identified by
ugi
.
- getSuccessfulJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getSuccessfulTaskAttempt() - Method in class org.apache.hadoop.mapred.TaskReport
-
Get the attempt ID that took this task to completion
- GetSuffix(int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Get the suffix.
- getSum() - Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
-
- getSum() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
-
- getSupportedCompressionAlgorithms() - Static method in class org.apache.hadoop.io.file.tfile.TFile
-
Get names of supported compression algorithms.
- getSymlink(Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This method checks to see if symlinks are to be create for the
localized cache files in the current working directory
- getSystemDir() - Method in class org.apache.hadoop.mapred.JobClient
-
Grab the jobtracker system directory path where job-specific files are to be placed.
- getSystemDir() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getTabSize(int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- getTag() - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- getTag(String) - Method in class org.apache.hadoop.metrics.spi.OutputRecord
-
Returns a tag object which is can be a String, Integer, Short or Byte.
- getTagNames() - Method in class org.apache.hadoop.metrics.spi.OutputRecord
-
Returns the set of tag names
- getTask(JVMId) - Method in class org.apache.hadoop.mapred.TaskTracker
-
Called upon startup by the child process, to fetch Task data.
- getTaskAttemptID() - Method in class org.apache.hadoop.mapred.TaskAttemptContext
-
Deprecated.
Get the taskAttemptID.
- getTaskAttemptId() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Returns task id.
- getTaskAttemptID() - Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
-
Get the unique name for this task attempt.
- getTaskAttemptIDsPattern(String, Integer, Boolean, Integer, Integer) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
-
Deprecated.
- getTaskAttempts() - Method in class org.apache.hadoop.mapred.JobHistory.Task
-
Returns all task attempts for this task.
- getTaskCompletionEvents(JobID, int, int) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getTaskCompletionEvents(int) - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get events indicating completion (success/failure) of component tasks.
- getTaskCompletionEvents(int) - Method in class org.apache.hadoop.mapreduce.Job
-
Get events indicating completion (success/failure) of component tasks.
- getTaskDiagnostics(TaskAttemptID) - Method in class org.apache.hadoop.mapred.JobTracker
-
Get the diagnostics for a given task
- getTaskDiagnostics(TaskAttemptID) - Method in interface org.apache.hadoop.mapred.RunningJob
-
Gets the diagnostic messages for a given task attempt.
- getTaskID() - Method in class org.apache.hadoop.mapred.TaskAttemptID
-
Deprecated.
- getTaskId() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- getTaskId() - Method in class org.apache.hadoop.mapred.TaskLogAppender
-
Getter/Setter methods for log4j.
- getTaskId() - Method in class org.apache.hadoop.mapred.TaskReport
-
- getTaskID() - Method in class org.apache.hadoop.mapred.TaskReport
-
The id of the task.
- getTaskID() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
Returns the
TaskID
object that this task attempt belongs to
- getTaskIDsPattern(String, Integer, Boolean, Integer) - Static method in class org.apache.hadoop.mapred.TaskID
-
Deprecated.
- getTaskInfo(JobConf) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- getTaskLogFile(TaskAttemptID, TaskLog.LogName) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- getTaskLogLength(JobConf) - Static method in class org.apache.hadoop.mapred.TaskLog
-
Get the desired maximum length of task's logs.
- getTaskLogsUrl(JobHistory.TaskAttempt) - Static method in class org.apache.hadoop.mapred.JobHistory
-
Return the TaskLogsUrl of a particular TaskAttempt
- getTaskLogUrl(String, String, String) - Static method in class org.apache.hadoop.mapred.TaskLogServlet
-
Construct the taskLogUrl
- getTaskMemoryManager() - Method in class org.apache.hadoop.mapred.TaskTracker
-
- getTaskOutputFilter(JobConf) - Static method in class org.apache.hadoop.mapred.JobClient
-
Get the task output filter out of the JobConf.
- getTaskOutputFilter() - Method in class org.apache.hadoop.mapred.JobClient
-
Deprecated.
- getTaskOutputPath(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Helper function to create the task's temporary output directory and
return the path to the task's output file.
- getTaskRunTime() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Returns time (in millisec) the task took to complete.
- getTaskStatus() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Returns enum Status.SUCESS or Status.FAILURE.
- getTaskTracker(String) - Method in class org.apache.hadoop.mapred.JobTracker
-
- getTaskTrackerHttp() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
http location of the tasktracker where this task ran.
- getTaskTrackerInstrumentation() - Method in class org.apache.hadoop.mapred.TaskTracker
-
- getTaskTrackerReportAddress() - Method in class org.apache.hadoop.mapred.TaskTracker
-
Return the port at which the tasktracker bound to
- getTaskTrackers() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the number of task trackers in the cluster.
- getTerm() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Get the term.
- getText() - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
Get the text that represents a document.
- getText() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
The text of the document id.
- getTimestamp(Configuration, URI) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Returns mtime of a given cache file on hdfs.
- getTip(TaskID) - Method in class org.apache.hadoop.mapred.JobTracker
-
Returns specified TaskInProgress, or null.
- getTmax(String) - Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
-
- getToken(int) - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Get the specific Token.
- getTopologyPaths() - Method in class org.apache.hadoop.fs.BlockLocation
-
Get the list of network topology paths for each of the hosts.
- getTotalLogFileSize() - Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- getTotalSubmissions() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getTrackerIdentifier() - Method in class org.apache.hadoop.mapred.JobTracker
-
Get the unique identifier (ie.
- getTrackerPort() - Method in class org.apache.hadoop.mapred.JobTracker
-
- getTrackingURL() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the URL where some job progress information will be displayed.
- getTrackingURL() - Method in class org.apache.hadoop.mapreduce.Job
-
Get the URL where some job progress information will be displayed.
- getTTExpiryInterval() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the tasktracker expiry interval for the cluster
- getType() - Method in class org.apache.hadoop.mapred.join.Parser.Token
-
- getTypeID() - Method in class org.apache.hadoop.record.meta.FieldTypeInfo
-
get the field's TypeID object
- getTypes() - Method in class org.apache.hadoop.io.GenericWritable
-
Return all classes that may be wrapped.
- getTypeVal() - Method in class org.apache.hadoop.record.meta.TypeID
-
Get the type value.
- getUlimitMemoryCommand(Configuration) - Static method in class org.apache.hadoop.util.Shell
-
Get the Unix command for setting the maximum virtual memory available
to a given child process.
- getUMask(Configuration) - Static method in class org.apache.hadoop.fs.permission.FsPermission
-
Get the user file creation mask (umask)
- getUniqueFile(TaskAttemptContext, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Generate a unique filename, based on the task id, name, and extension
- getUniqueItems() - Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
-
- getUniqueName(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Helper function to generate a name that is unique for the task.
- getUnits(String) - Method in class org.apache.hadoop.metrics.ganglia.GangliaContext
-
- getUri() - Method in class org.apache.hadoop.fs.FileSystem
-
Returns a URI whose scheme and authority identify this FileSystem.
- getUri() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Returns a URI whose scheme and authority identify this FileSystem.
- getUri() - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- getUri() - Method in class org.apache.hadoop.fs.HarFileSystem
-
Returns the uri of this filesystem.
- getUri() - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- getUri() - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- getUri() - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- getUri() - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- getURIs(String, String) - Method in class org.apache.hadoop.streaming.StreamJob
-
get the uris of all the files/caches
- getURL() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the link to the web-ui for details of the job.
- getUrl() - Static method in class org.apache.hadoop.util.VersionInfo
-
Get the subversion URL for the root Hadoop directory.
- getUsed() - Method in class org.apache.hadoop.fs.DF
-
- getUsed() - Method in class org.apache.hadoop.fs.DU
-
- getUsed() - Method in class org.apache.hadoop.fs.FileSystem
-
Return the total size of all files in the filesystem.
- getUsedMemory() - Method in class org.apache.hadoop.mapred.ClusterStatus
-
Get the total heap memory used by the JobTracker
- getUseNewMapper() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should the framework use the new context-object code for running
the mapper?
- getUseNewReducer() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should the framework use the new context-object code for running
the reducer?
- getUser() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the reported username for this job.
- getUser() - Method in class org.apache.hadoop.mapred.JobProfile
-
Get the user id.
- getUser() - Static method in class org.apache.hadoop.util.VersionInfo
-
The user that compiled Hadoop.
- getUserAction() - Method in class org.apache.hadoop.fs.permission.FsPermission
-
- getUserName() - Method in class org.apache.hadoop.fs.permission.PermissionStatus
-
Return user name
- getUserName(JobConf) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Get the user name from the job conf
- getUsername() - Method in class org.apache.hadoop.mapred.JobStatus
-
- getUserName() - Method in class org.apache.hadoop.security.UnixUserGroupInformation
-
Return the user's name
- getUserName() - Method in class org.apache.hadoop.security.UserGroupInformation
-
Get username
- getUsers() - Method in class org.apache.hadoop.security.SecurityUtil.AccessControlList
-
- getVal() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
-
- getVal() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
-
- getVal() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
-
- getVal() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
-
- getValue(BytesWritable) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy the value into BytesWritable.
- getValue(byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy value into user-supplied buffer.
- getValue(byte[], int) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Copy value into user-supplied buffer.
- getValue() - Method in interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator
-
Gets the current raw value
- getValue() - Method in interface org.apache.hadoop.mapred.RawKeyValueIterator
-
Gets the current raw value.
- getValue() - Method in class org.apache.hadoop.mapreduce.Counter
-
What is the current value of this counter?
- getValue() - Method in class org.apache.hadoop.record.compiler.generated.Token
-
An optional attribute value of the Token.
- getValue() - Method in class org.apache.hadoop.util.DataChecksum
-
- getValueClass() - Method in class org.apache.hadoop.io.ArrayWritable
-
- getValueClass() - Method in class org.apache.hadoop.io.MapFile.Reader
-
Returns the class of values in this file.
- getValueClass() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the class of values in this file.
- getValueClass() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
Returns the class of values in this file.
- getValueClass() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- getValueClassName() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns the name of the value class.
- getValueClassName() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
Retrieve the name of the value class for this SequenceFile.
- getValueLength() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Get the length of the value.
- getValues() - Method in class org.apache.hadoop.mapreduce.ReduceContext
-
Iterate through the values for the current key, reusing the same value
object, which is stored in the context.
- getValueStream() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry
-
Stream access to value.
- getValueTypeID() - Method in class org.apache.hadoop.record.meta.MapTypeID
-
get the TypeID of the map's value element
- getVectorSize() - Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- getVersion() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
Get the version number of the entire index.
- getVersion() - Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- getVersion() - Method in class org.apache.hadoop.io.VersionedWritable
-
Return the version number of the current implementation.
- getVersion() - Static method in class org.apache.hadoop.util.VersionInfo
-
Get the Hadoop version.
- getVIntSize(long) - Static method in class org.apache.hadoop.io.WritableUtils
-
Get the encoded length if an integer is stored in a variable-length format
- getVIntSize(long) - Static method in class org.apache.hadoop.record.Utils
-
Get the encoded length if an integer is stored in a variable-length format
- getVirtualMemorySize() - Method in class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin
-
Obtain the total size of the virtual memory present in the system.
- getVirtualMemorySize() - Method in class org.apache.hadoop.util.MemoryCalculatorPlugin
-
Obtain the total size of the virtual memory present in the system.
- getWaitingJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
- getWarn() - Static method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- getWebAppsPath() - Method in class org.apache.hadoop.http.HttpServer
-
Get the pathname to the webapps files.
- getWeight() - Method in class org.apache.hadoop.util.bloom.Key
-
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.FileSystem
-
Get the current working directory for the given file system
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Get the current working directory for the given file system
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.HarFileSystem
-
return the top level archive.
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- getWorkingDirectory() - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- getWorkingDirectory() - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Get the current working directory for the default file system.
- getWorkingDirectory() - Method in class org.apache.hadoop.mapreduce.JobContext
-
Get the current working directory for the default file system.
- getWorkOutputPath(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Get the
Path
to the task's temporary output directory
for the map-reduce job
Tasks' Side-Effect Files
- getWorkOutputPath(TaskInputOutputContext<?, ?, ?, ?>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Get the
Path
to the task's temporary output directory
for the map-reduce job
Tasks' Side-Effect Files
- getWorkPath() - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
Get the directory that the task should write results into
- getWrappedStream() - Method in class org.apache.hadoop.fs.FSDataOutputStream
-
- getZlibCompressor(Configuration) - Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
Return the appropriate implementation of the zlib compressor.
- getZlibCompressorType(Configuration) - Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
Return the appropriate type of the zlib compressor.
- getZlibDecompressor(Configuration) - Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
Return the appropriate implementation of the zlib decompressor.
- getZlibDecompressorType(Configuration) - Static method in class org.apache.hadoop.io.compress.zlib.ZlibFactory
-
Return the appropriate type of the zlib decompressor.
- globStatus(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Return all the files that match filePattern and are not checksum
files.
- globStatus(Path, PathFilter) - Method in class org.apache.hadoop.fs.FileSystem
-
Return an array of FileStatus objects whose path names match pathPattern
and is accepted by the user-supplied path filter.
- go() - Method in class org.apache.hadoop.streaming.StreamJob
-
- goodClassOrNull(Configuration, String, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
It may seem strange to silently switch behaviour when a String
is not a classname; the reason is simplified Usage:
- GREATER_ICOST - Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
This constant is accessible by subclasses for historical purposes.
- Grep - Class in org.apache.hadoop.examples
-
- Group - Class in org.apache.hadoop.security
-
A group to which a user belongs to.
- Group(String) - Constructor for class org.apache.hadoop.security.Group
-
Create a new Group
with the given groupname.
- GT_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- GzipCodec - Class in org.apache.hadoop.io.compress
-
This class creates gzip compressors/decompressors.
- GzipCodec() - Constructor for class org.apache.hadoop.io.compress.GzipCodec
-
- GzipCodec.GzipInputStream - Class in org.apache.hadoop.io.compress
-
- GzipCodec.GzipOutputStream - Class in org.apache.hadoop.io.compress
-
A bridge that wraps around a DeflaterOutputStream to make it
a CompressionOutputStream.
- GzipInputStream(InputStream) - Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- GzipInputStream(DecompressorStream) - Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
Allow subclasses to directly set the inflater stream.
- GzipOutputStream(OutputStream) - Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- GzipOutputStream(CompressorStream) - Constructor for class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
Allow children types to put a different type in here.
- RAMDirectoryUtil - Class in org.apache.hadoop.contrib.index.lucene
-
A utility class which writes an index in a ram dir into a DataOutput and
read from a DataInput an index into a ram dir.
- RAMDirectoryUtil() - Constructor for class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
-
- RANDOM - Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
-
Random selection.
- RandomSampler(double, int) - Constructor for class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
-
Create a new RandomSampler sampling all splits.
- RandomSampler(double, int, int) - Constructor for class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler
-
Create a new RandomSampler.
- RandomTextWriter - Class in org.apache.hadoop.examples
-
This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task writes a large unsorted
random sequence of words.
- RandomTextWriter() - Constructor for class org.apache.hadoop.examples.RandomTextWriter
-
- RandomWriter - Class in org.apache.hadoop.examples
-
This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task write a large unsorted
random binary sequence file of BytesWritable.
- RandomWriter() - Constructor for class org.apache.hadoop.examples.RandomWriter
-
- RATIO - Static variable in interface org.apache.hadoop.util.bloom.RemoveScheme
-
Ratio Selection.
- RawComparable - Interface in org.apache.hadoop.io.file.tfile
-
Interface for objects that can be compared through
RawComparator
.
- RawComparator<T> - Interface in org.apache.hadoop.io
-
A
Comparator
that operates directly on byte representations of
objects.
- RawKeyValueIterator - Interface in org.apache.hadoop.mapred
-
RawKeyValueIterator
is an iterator used to iterate over
the raw keys and values during sort/merge of intermediate data.
- RawLocalFileSystem - Class in org.apache.hadoop.fs
-
Implement the FileSystem API for the raw local filesystem.
- RawLocalFileSystem() - Constructor for class org.apache.hadoop.fs.RawLocalFileSystem
-
- rawMapping - Variable in class org.apache.hadoop.net.CachedDNSToSwitchMapping
-
- RBRACE_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- Rcc - Class in org.apache.hadoop.record.compiler.generated
-
- Rcc(InputStream) - Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
Constructor with InputStream.
- Rcc(InputStream, String) - Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
Constructor with InputStream and supplied encoding
- Rcc(Reader) - Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
Constructor.
- Rcc(RccTokenManager) - Constructor for class org.apache.hadoop.record.compiler.generated.Rcc
-
Constructor with generated Token Manager.
- RccConstants - Interface in org.apache.hadoop.record.compiler.generated
-
Token literal values and constants.
- RccTask - Class in org.apache.hadoop.record.compiler.ant
-
Hadoop record compiler ant Task
- RccTask() - Constructor for class org.apache.hadoop.record.compiler.ant.RccTask
-
Creates a new instance of RccTask
- RccTokenManager - Class in org.apache.hadoop.record.compiler.generated
-
Token Manager.
- RccTokenManager(SimpleCharStream) - Constructor for class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Constructor.
- RccTokenManager(SimpleCharStream, int) - Constructor for class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Constructor.
- read(long, byte[], int, int) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- read(long, byte[], int, int) - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- read() - Method in class org.apache.hadoop.fs.FSInputChecker
-
Read one checksum-verified byte
- read(byte[], int, int) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Read checksum verified bytes from this byte-input stream into
the specified byte array, starting at the given offset.
- read(long, byte[], int, int) - Method in class org.apache.hadoop.fs.FSInputStream
-
- read() - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- read(DataInput) - Static method in class org.apache.hadoop.fs.permission.FsPermission
-
- read(DataInput) - Static method in class org.apache.hadoop.fs.permission.PermissionStatus
-
- read(long, byte[], int, int) - Method in interface org.apache.hadoop.fs.PositionedReadable
-
Read upto the specified number of bytes, from a given
position within a file, and return the number of bytes read.
- read() - Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.io.compress.CompressionInputStream
-
Read bytes from the stream.
- read() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- read() - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- read(DataInput) - Static method in class org.apache.hadoop.io.MD5Hash
-
Constructs, reads and returns an instance.
- read(DataInput) - Static method in class org.apache.hadoop.mapred.JobID
-
Deprecated.
- read(DataInput) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
-
Deprecated.
- read(DataInput) - Static method in class org.apache.hadoop.mapred.TaskID
-
Deprecated.
- read() - Method in class org.apache.hadoop.net.SocketInputStream
-
- read(byte[], int, int) - Method in class org.apache.hadoop.net.SocketInputStream
-
- read(ByteBuffer) - Method in class org.apache.hadoop.net.SocketInputStream
-
- readBool(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readBool(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readBool(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a boolean from serialized record.
- readBool(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readBuffer(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readBuffer(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readBuffer(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read byte array from serialized record.
- readBuffer(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readByte(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readByte(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readByte(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a byte from serialized record.
- readByte(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readChar() - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Read a character.
- readChunk(long, byte[], int, int, byte[]) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Reads in next checksum chunk data into buf
at offset
and checksum into checksum
.
- readCompressedByteArray(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- readCompressedString(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- readCompressedStringArray(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- readDouble(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Parse a double from a byte array.
- readDouble(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readDouble(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readDouble(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a double-precision number from serialized record.
- readDouble(byte[], int) - Static method in class org.apache.hadoop.record.Utils
-
Parse a double from a byte array.
- readDouble(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readEnum(DataInput, Class<T>) - Static method in class org.apache.hadoop.io.WritableUtils
-
Read an Enum value from DataInput, Enums are read and written
using String values.
- Reader(FileSystem, String, Configuration) - Constructor for class org.apache.hadoop.io.ArrayFile.Reader
-
Construct an array reader for the named file.
- Reader(FileSystem, String, Configuration) - Constructor for class org.apache.hadoop.io.BloomMapFile.Reader
-
- Reader(FileSystem, String, WritableComparator, Configuration, boolean) - Constructor for class org.apache.hadoop.io.BloomMapFile.Reader
-
- Reader(FileSystem, String, WritableComparator, Configuration) - Constructor for class org.apache.hadoop.io.BloomMapFile.Reader
-
- Reader(FSDataInputStream, long, Configuration) - Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader
-
Constructor
- Reader(FileSystem, String, Configuration) - Constructor for class org.apache.hadoop.io.MapFile.Reader
-
Construct a map reader for the named map.
- Reader(FileSystem, String, WritableComparator, Configuration) - Constructor for class org.apache.hadoop.io.MapFile.Reader
-
Construct a map reader for the named map using the named comparator.
- Reader(FileSystem, String, WritableComparator, Configuration, boolean) - Constructor for class org.apache.hadoop.io.MapFile.Reader
-
Hook to allow subclasses to defer opening streams until further
initialization is complete.
- Reader(FileSystem, Path, Configuration) - Constructor for class org.apache.hadoop.io.SequenceFile.Reader
-
Open the named file.
- Reader(FileSystem, String, Configuration) - Constructor for class org.apache.hadoop.io.SetFile.Reader
-
Construct a set reader for the named set.
- Reader(FileSystem, String, WritableComparator, Configuration) - Constructor for class org.apache.hadoop.io.SetFile.Reader
-
Construct a set reader for the named set using the named comparator.
- readFields(DataInput) - Method in class org.apache.hadoop.conf.Configuration
-
- readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
- readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
- readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
-
- readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
-
- readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- readFields(DataInput) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
-
- readFields(DataInput) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
Read the two integers.
- readFields(DataInput) - Method in class org.apache.hadoop.examples.SleepJob.EmptySplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.fs.BlockLocation
-
Implement readFields of Writable
- readFields(DataInput) - Method in class org.apache.hadoop.fs.ContentSummary
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.fs.FileStatus
-
- readFields(DataInput) - Method in class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.fs.permission.FsPermission
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.fs.permission.PermissionStatus
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.io.AbstractMapWritable
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.io.ArrayWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.BooleanWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.BytesWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.ByteWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.CompressedWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.DoubleWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.FloatWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.GenericWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.IntWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.LongWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.MapWritable
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.io.MD5Hash
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.NullWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.ObjectWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.SortedMapWritable
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.io.Text
-
deserialize
- readFields(DataInput) - Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
- readFields(DataInput) - Method in class org.apache.hadoop.io.VersionedWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.VIntWritable
-
- readFields(DataInput) - Method in class org.apache.hadoop.io.VLongWritable
-
- readFields(DataInput) - Method in interface org.apache.hadoop.io.Writable
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.ClusterStatus
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Read a set of groups.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.FileSplit
-
Deprecated.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.JobProfile
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.JobQueueInfo
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.JobStatus
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.join.TupleWritable
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
-
Deserialize the fields of this object from in
.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
-
- readFields(ResultSet) - Method in interface org.apache.hadoop.mapred.lib.db.DBWritable
-
Reads the fields of the object from the
ResultSet
.
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapred.TaskReport
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.Counter
-
Read the binary representation of the counter
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.CounterGroup
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.Counters
-
Read a set of groups.
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.ID
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.JobID
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
-
- readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskID
-
- readFields(DataInput) - Method in class org.apache.hadoop.record.Record
-
- readFields(DataInput) - Method in class org.apache.hadoop.security.UnixUserGroupInformation
-
Deserialize this object
First check if this is a UGI in the string format.
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.BloomFilter
-
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.CountingBloomFilter
-
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.DynamicBloomFilter
-
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.Filter
-
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.Key
-
- readFields(DataInput) - Method in class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
- readFieldsCompressed(DataInput) - Method in class org.apache.hadoop.io.CompressedWritable
-
- readFloat(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Parse a float from a byte array.
- readFloat(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readFloat(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readFloat(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a single-precision float from serialized record.
- readFloat(byte[], int) - Static method in class org.apache.hadoop.record.Utils
-
Parse a float from a byte array.
- readFloat(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readFrom(Configuration) - Static method in class org.apache.hadoop.security.UserGroupInformation
-
- readFromConf(Configuration, String) - Static method in class org.apache.hadoop.security.UnixUserGroupInformation
-
Read a UGI from the given conf
The object is expected to store with the property name attr
as a comma separated string that starts
with the user name followed by group names.
- readFully(long, byte[], int, int) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- readFully(long, byte[]) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- readFully(long, byte[], int, int) - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- readFully(long, byte[]) - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- readFully(InputStream, byte[], int, int) - Static method in class org.apache.hadoop.fs.FSInputChecker
-
A utility function that tries to read up to len
bytes from
stm
- readFully(long, byte[], int, int) - Method in class org.apache.hadoop.fs.FSInputStream
-
- readFully(long, byte[]) - Method in class org.apache.hadoop.fs.FSInputStream
-
- readFully(long, byte[], int, int) - Method in interface org.apache.hadoop.fs.PositionedReadable
-
Read the specified number of bytes, from a given
position within a file.
- readFully(long, byte[]) - Method in interface org.apache.hadoop.fs.PositionedReadable
-
Read number of bytes equalt to the length of the buffer, from a given
position within a file.
- readFully(InputStream, byte[], int, int) - Static method in class org.apache.hadoop.io.IOUtils
-
Reads len bytes in a loop.
- readInt(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Parse an integer from a byte array.
- readInt(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readInt(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readInt(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read an integer from serialized record.
- readInt(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readLine(LineReader, Text) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
-
Read a utf8 encoded line from a data input stream.
- readLine(LineReader, Text) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
-
- readLine(Text, int, int) - Method in class org.apache.hadoop.util.LineReader
-
Read one line from the InputStream into the given Text.
- readLine(Text, int) - Method in class org.apache.hadoop.util.LineReader
-
Read from the InputStream into the given Text.
- readLine(Text) - Method in class org.apache.hadoop.util.LineReader
-
Read from the InputStream into the given Text.
- readLong(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Parse a long from a byte array.
- readLong(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readLong(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readLong(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a long integer from serialized record.
- readLong(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readObject(DataInput, Configuration) - Static method in class org.apache.hadoop.io.ObjectWritable
-
Read a
Writable
,
String
, primitive type, or an array of
the preceding.
- readObject(DataInput, ObjectWritable, Configuration) - Static method in class org.apache.hadoop.io.ObjectWritable
-
Read a
Writable
,
String
, primitive type, or an array of
the preceding.
- readRAMFiles(DataInput, RAMDirectory) - Static method in class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
-
Read a number of files from a data input to a ram directory.
- readState(String) - Static method in class org.apache.hadoop.contrib.failmon.PersistentState
-
Read the state of parsing for all open log files from a property
file.
- readString(DataInput) - Static method in class org.apache.hadoop.io.file.tfile.Utils
-
Read a String as a VInt n, followed by n Bytes in Text format.
- readString(DataInput) - Static method in class org.apache.hadoop.io.Text
-
Read a UTF8 encoded string from in
- readString(DataInput) - Static method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Read a UTF-8 encoded string.
- readString(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- readString(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- readString(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- readString(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Read a UTF-8 encoded string from serialized record.
- readString(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- readStringArray(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- readUnsignedShort(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Parse an unsigned short from a byte array.
- readVInt(DataInput) - Static method in class org.apache.hadoop.io.file.tfile.Utils
-
Decoding the variable-length integer.
- readVInt(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Reads a zero-compressed encoded integer from a byte array and returns it.
- readVInt(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
Reads a zero-compressed encoded integer from input stream and returns it.
- readVInt(byte[], int) - Static method in class org.apache.hadoop.record.Utils
-
Reads a zero-compressed encoded integer from a byte array and returns it.
- readVInt(DataInput) - Static method in class org.apache.hadoop.record.Utils
-
Reads a zero-compressed encoded integer from a stream and returns it.
- readVLong(DataInput) - Static method in class org.apache.hadoop.io.file.tfile.Utils
-
Decoding the variable-length integer.
- readVLong(byte[], int) - Static method in class org.apache.hadoop.io.WritableComparator
-
Reads a zero-compressed encoded long from a byte array and returns it.
- readVLong(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
Reads a zero-compressed encoded long from input stream and returns it.
- readVLong(byte[], int) - Static method in class org.apache.hadoop.record.Utils
-
Reads a zero-compressed encoded long from a byte array and returns it.
- readVLong(DataInput) - Static method in class org.apache.hadoop.record.Utils
-
Reads a zero-compressed encoded long from a stream and return it.
- READY - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- Record() - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- Record - Class in org.apache.hadoop.record
-
Abstract class that is extended by generated classes.
- Record() - Constructor for class org.apache.hadoop.record.Record
-
- RECORD_INPUT - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RECORD_OUTPUT - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RECORD_SEPARATOR - Static variable in class org.apache.hadoop.contrib.failmon.LocalStore
-
- RECORD_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- RecordComparator - Class in org.apache.hadoop.record
-
A raw record comparator base class
- RecordComparator(Class<? extends WritableComparable>) - Constructor for class org.apache.hadoop.record.RecordComparator
-
Construct a raw
Record
comparison implementation.
- RecordInput - Interface in org.apache.hadoop.record
-
Interface that all the Deserializers have to implement.
- RecordList() - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
- RecordOutput - Interface in org.apache.hadoop.record
-
Interface that alll the serializers have to implement.
- RecordReader<K,V> - Interface in org.apache.hadoop.mapred
-
RecordReader
reads <key, value> pairs from an
InputSplit
.
- RecordReader<KEYIN,VALUEIN> - Class in org.apache.hadoop.mapreduce
-
The record reader breaks the data into key/value pairs for input to the
Mapper
.
- RecordReader() - Constructor for class org.apache.hadoop.mapreduce.RecordReader
-
- RecordTypeInfo - Class in org.apache.hadoop.record.meta
-
A record's Type Information object which can read/write itself.
- RecordTypeInfo() - Constructor for class org.apache.hadoop.record.meta.RecordTypeInfo
-
Create an empty RecordTypeInfo object.
- RecordTypeInfo(String) - Constructor for class org.apache.hadoop.record.meta.RecordTypeInfo
-
Create a RecordTypeInfo object representing a record with the given name
- RecordWriter<K,V> - Interface in org.apache.hadoop.mapred
-
RecordWriter
writes the output <key, value> pairs
to an output file.
- RecordWriter<K,V> - Class in org.apache.hadoop.mapreduce
-
RecordWriter
writes the output <key, value> pairs
to an output file.
- RecordWriter() - Constructor for class org.apache.hadoop.mapreduce.RecordWriter
-
- recoverJobHistoryFile(JobConf, Path) - Static method in class org.apache.hadoop.mapred.JobHistory.JobInfo
-
Selects one of the two files generated as a part of recovery.
- redCmd_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, IntermediateForm>, Reporter) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
-
- reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, Text>, Reporter) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
-
- reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- reduce(BooleanWritable, Iterator<LongWritable>, OutputCollector<WritableComparable<?>, Writable>, Reporter) - Method in class org.apache.hadoop.examples.PiEstimator.PiReducer
-
Accumulate number of points inside/outside results from the mappers.
- Reduce() - Constructor for class org.apache.hadoop.examples.SecondarySort.Reduce
-
- reduce(SecondarySort.IntPair, Iterable<IntWritable>, Reducer<SecondarySort.IntPair, IntWritable, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.SecondarySort.Reduce
-
- reduce(IntWritable, Iterator<NullWritable>, OutputCollector<NullWritable, NullWritable>, Reporter) - Method in class org.apache.hadoop.examples.SleepJob
-
- reduce(Text, Iterable<IntWritable>, Reducer<Text, IntWritable, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.WordCount.IntSumReducer
-
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
-
Combines values for a given key.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
-
Do nothing.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
-
- reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.mapred.lib.ChainReducer
-
Chains the reduce(...)
method of the Reducer with the
map(...)
methods of the Mappers in the chain.
- reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
-
- reduce(K, Iterator<V>, OutputCollector<K, V>, Reporter) - Method in class org.apache.hadoop.mapred.lib.IdentityReducer
-
Deprecated.
Writes all keys and values directly to output.
- reduce(K, Iterator<LongWritable>, OutputCollector<K, LongWritable>, Reporter) - Method in class org.apache.hadoop.mapred.lib.LongSumReducer
-
Deprecated.
- reduce(K2, Iterator<V2>, OutputCollector<K3, V3>, Reporter) - Method in interface org.apache.hadoop.mapred.Reducer
-
Deprecated.
Reduces values for a given key.
- reduce(Key, Iterable<IntWritable>, Reducer<Key, IntWritable, Key, IntWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer
-
- reduce(KEY, Iterable<LongWritable>, Reducer<KEY, LongWritable, KEY, LongWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer
-
- reduce(KEYIN, Iterable<VALUEIN>, Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
-
This method is called once for each key.
- reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.streaming.PipeReducer
-
- REDUCE_CLASS_ATTR - Static variable in class org.apache.hadoop.mapreduce.JobContext
-
- ReduceAttempt() - Constructor for class org.apache.hadoop.mapred.JobHistory.ReduceAttempt
-
- ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
-
- ReduceContext(Configuration, TaskAttemptID, RawKeyValueIterator, Counter, Counter, RecordWriter<KEYOUT, VALUEOUT>, OutputCommitter, StatusReporter, RawComparator<KEYIN>, Class<KEYIN>, Class<VALUEIN>) - Constructor for class org.apache.hadoop.mapreduce.ReduceContext
-
- ReduceContext.ValueIterable - Class in org.apache.hadoop.mapreduce
-
- ReduceContext.ValueIterator - Class in org.apache.hadoop.mapreduce
-
- reduceDebugSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- reduceProgress() - Method in class org.apache.hadoop.mapred.JobStatus
-
- reduceProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the progress of the job's reduce-tasks, as a float between 0.0
and 1.0.
- reduceProgress() - Method in class org.apache.hadoop.mapreduce.Job
-
Get the progress of the job's reduce-tasks, as a float between 0.0
and 1.0.
- Reducer<K2,V2,K3,V3> - Interface in org.apache.hadoop.mapred
-
- Reducer<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
-
Reduces a set of intermediate values which share a key to a smaller set of
values.
- Reducer() - Constructor for class org.apache.hadoop.mapreduce.Reducer
-
- Reducer.Context - Class in org.apache.hadoop.mapreduce
-
- ReflectionUtils - Class in org.apache.hadoop.util
-
General reflection utils
- ReflectionUtils() - Constructor for class org.apache.hadoop.util.ReflectionUtils
-
- refresh() - Method in class org.apache.hadoop.security.authorize.ConfiguredPolicy
-
- refresh() - Method in class org.apache.hadoop.util.HostsFileReader
-
- RefreshAuthorizationPolicyProtocol - Interface in org.apache.hadoop.security.authorize
-
Protocol which is used to refresh the authorization policy in use currently.
- refreshServiceAcl() - Method in class org.apache.hadoop.mapred.JobTracker
-
- refreshServiceAcl() - Method in interface org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol
-
Refresh the service-level authorization policy in-effect.
- RegexFilter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
-
- RegexMapper<K> - Class in org.apache.hadoop.mapred.lib
-
A
Mapper
that extracts text matching a regular expression.
- RegexMapper() - Constructor for class org.apache.hadoop.mapred.lib.RegexMapper
-
- regexpEscape(String) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- registerMBean(String, String, Object) - Static method in class org.apache.hadoop.metrics.util.MBeanUtil
-
Register the MBean using our standard MBeanName format
"hadoop:service=,name="
Where the and are the supplied parameters
- registerNotification(JobConf, JobStatus) - Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- registerUpdater(Updater) - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Registers a callback to be called at regular time intervals, as
determined by the implementation-class specific configuration.
- registerUpdater(Updater) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Registers a callback to be called at time intervals determined by
the configuration.
- registerUpdater(Updater) - Method in class org.apache.hadoop.metrics.spi.CompositeContext
-
- registry - Variable in class org.apache.hadoop.ipc.metrics.RpcMetrics
-
- ReInit(InputStream) - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Reinitialise.
- ReInit(InputStream, String) - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Reinitialise.
- ReInit(Reader) - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Reinitialise.
- ReInit(RccTokenManager) - Method in class org.apache.hadoop.record.compiler.generated.Rcc
-
Reinitialise.
- ReInit(SimpleCharStream) - Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Reinitialise parser.
- ReInit(SimpleCharStream, int) - Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Reinitialise parser.
- ReInit(Reader, int, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(Reader, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(Reader) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream, String, int, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream, int, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream, String) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream, String, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- ReInit(InputStream, int, int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Reinitialise.
- release(Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
Deprecated.
- releaseCache(URI, Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This is the opposite of getlocalcache.
- reloadConfiguration() - Method in class org.apache.hadoop.conf.Configuration
-
Reload configuration from previously added resources.
- RemoteException - Exception in org.apache.hadoop.ipc
-
- RemoteException(String, String) - Constructor for exception org.apache.hadoop.ipc.RemoteException
-
- remove() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- remove(Object) - Method in class org.apache.hadoop.io.MapWritable
- remove(Object) - Method in class org.apache.hadoop.io.SortedMapWritable
- remove() - Method in class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
-
- remove() - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Removes, from the buffered data table, all rows having tags
that equal the tags that have been set on this record.
- remove(MetricsRecordImpl) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Called by MetricsRecordImpl.remove().
- remove() - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Removes the row, if it exists, in the buffered data table having tags
that equal the tags that have been set on this record.
- remove(MetricsRecordImpl) - Method in class org.apache.hadoop.metrics.spi.NullContext
-
Do-nothing version of remove
- remove(MetricsRecordImpl) - Method in class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
-
Do-nothing version of remove
- remove(Node) - Method in class org.apache.hadoop.net.NetworkTopology
-
Remove a node
Update node counter & rack counter if neccessary
- removeAttribute(String) - Method in class org.apache.hadoop.metrics.ContextFactory
-
Removes the named attribute if it exists.
- removeJobInProgressListener(JobInProgressListener) - Method in class org.apache.hadoop.mapred.JobTracker
-
- RemoveScheme - Interface in org.apache.hadoop.util.bloom
-
Defines the different remove scheme for retouched Bloom filters.
- removeSuffix(String, String) - Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
-
Removes a suffix from a filename, if it has it.
- removeTag(String) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Removes any tag of the specified name.
- removeTag(String) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Removes any tag of the specified name.
- rename(Path, Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Rename files/dirs
- rename(Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Renames Path src to Path dst.
- rename(Path, Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Renames Path src to Path dst.
- rename(Path, Path) - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- rename(Path, Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- rename(Path, Path) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- rename(Path, Path) - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- rename(Path, Path) - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
- rename(FileSystem, String, String) - Static method in class org.apache.hadoop.io.MapFile
-
Renames an existing map directory.
- renameFile(String, String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
-
- replaceFile(File, File) - Static method in class org.apache.hadoop.fs.FileUtil
-
Move the src file to the name specified by target.
- replay(X) - Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- replay(TupleWritable) - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- replay(V) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- replay(U) - Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- replay(T) - Method in interface org.apache.hadoop.mapred.join.ResetableIterator
-
Assign last value returned to actual.
- replay(X) - Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- report() - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
log the counters
- reportChecksumFailure(Path, FSDataInputStream, long, FSDataInputStream, long) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Report a checksum error to the file system.
- reportChecksumFailure(Path, FSDataInputStream, long, FSDataInputStream, long) - Method in class org.apache.hadoop.fs.LocalFileSystem
-
Moves files to a bad file directory on the same device, so that their
storage will not be reused.
- reportDiagnosticInfo(TaskAttemptID, String) - Method in class org.apache.hadoop.mapred.TaskTracker
-
Called when the task dies before completion, and we want to report back
diagnostic info
- reporter - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
-
- reporter - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- reporter - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- Reporter - Interface in org.apache.hadoop.mapred
-
A facility for Map-Reduce applications to report progress and update
counters, status information etc.
- reporter - Variable in class org.apache.hadoop.mapreduce.ReduceContext
-
- reportNextRecordRange(TaskAttemptID, SortedRanges.Range) - Method in class org.apache.hadoop.mapred.TaskTracker
-
- reportTaskTrackerError(String, String, String) - Method in class org.apache.hadoop.mapred.JobTracker
-
- requiresLayout() - Method in class org.apache.hadoop.metrics.jvm.EventCounter
-
- reserveSpaceWithCheckSum(Path, long) - Method in class org.apache.hadoop.fs.InMemoryFileSystem
-
Deprecated.
Register a file with its size.
- reset() - Method in class org.apache.hadoop.contrib.failmon.MonitorJob
-
- reset() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
-
- reset() - Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
-
- reset() - Method in class org.apache.hadoop.fs.FileSystem.Statistics
-
Reset the counts of bytes to 0.
- reset() - Method in class org.apache.hadoop.fs.FSInputChecker
-
- reset() - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- reset() - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- reset() - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- reset() - Method in interface org.apache.hadoop.io.compress.Compressor
-
Resets compressor so that a new set of input data can be processed.
- reset() - Method in interface org.apache.hadoop.io.compress.Decompressor
-
Resets decompressor so that a new set of input data can be processed.
- reset() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- reset() - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- reset() - Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- reset(byte[], int) - Method in class org.apache.hadoop.io.DataInputBuffer
-
Resets the data that the buffer reads.
- reset(byte[], int, int) - Method in class org.apache.hadoop.io.DataInputBuffer
-
Resets the data that the buffer reads.
- reset() - Method in class org.apache.hadoop.io.DataOutputBuffer
-
Resets the buffer to empty.
- reset(byte[], int) - Method in class org.apache.hadoop.io.InputBuffer
-
Resets the data that the buffer reads.
- reset(byte[], int, int) - Method in class org.apache.hadoop.io.InputBuffer
-
Resets the data that the buffer reads.
- reset() - Method in class org.apache.hadoop.io.MapFile.Reader
-
Re-positions the reader before its first key.
- reset() - Method in class org.apache.hadoop.io.OutputBuffer
-
Resets the buffer to empty.
- reset() - Method in class org.apache.hadoop.mapred.join.ArrayListBackedIterator
-
- reset() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
-
- reset() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
-
- reset() - Method in class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
-
- reset() - Method in interface org.apache.hadoop.mapred.join.ResetableIterator
-
Set iterator to return to the start of its range.
- reset() - Method in class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
-
reset the aggregator
- reset() - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator
-
reset the aggregator
- reset() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
-
reset the aggregator
- reset(BytesWritable) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
-
- reset() - Method in class org.apache.hadoop.record.Buffer
-
Reset the buffer to 0 size
- reset() - Method in class org.apache.hadoop.util.DataChecksum
-
- ResetableIterator - Interface in org.apache.hadoop.contrib.utils.join
-
This defines an iterator interface that will help the reducer class
re-group its input by source tags.
- ResetableIterator<T extends Writable> - Interface in org.apache.hadoop.mapred.join
-
This defines an interface to a stateful Iterator that can replay elements
added to it directly.
- ResetableIterator.EMPTY<U extends Writable> - Class in org.apache.hadoop.mapred.join
-
- resetAllMinMax() - Method in interface org.apache.hadoop.ipc.metrics.RpcMgtMBean
-
Reset all min max times
- resetChecksumChunk(int) - Method in class org.apache.hadoop.fs.FSOutputSummer
-
Resets existing buffer with a new one of the specified size.
- resetMinMax() - Method in class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate
-
Reset the min max values
- resetState() - Method in class org.apache.hadoop.io.compress.BlockDecompressorStream
-
- resetState() - Method in class org.apache.hadoop.io.compress.CompressionInputStream
-
Reset the decompressor to its initial state and discard any buffered data,
as the underlying stream may have been repositioned.
- resetState() - Method in class org.apache.hadoop.io.compress.CompressionOutputStream
-
Reset the compression to the initial state.
- resetState() - Method in class org.apache.hadoop.io.compress.CompressorStream
-
- resetState() - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- resetState() - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- resetState() - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream
-
- resolve(List<String>) - Method in class org.apache.hadoop.net.CachedDNSToSwitchMapping
-
- resolve(List<String>) - Method in interface org.apache.hadoop.net.DNSToSwitchMapping
-
Resolves a list of DNS-names/IP-addresses and returns back a list of
switch information (network paths).
- resolveAndAddToTopology(String) - Method in class org.apache.hadoop.mapred.JobTracker
-
- resume() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
resume the suspended thread
- RetouchedBloomFilter - Class in org.apache.hadoop.util.bloom
-
Implements a retouched Bloom filter, as defined in the CoNEXT 2006 paper.
- RetouchedBloomFilter() - Constructor for class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
Default constructor - use with readFields
- RetouchedBloomFilter(int, int, int) - Constructor for class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
Constructor
- retrieveBlock(Block, long) - Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- retrieveINode(Path) - Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- RETRY_FOREVER - Static variable in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying forever.
- retryByException(RetryPolicy, Map<Class<? extends Exception>, RetryPolicy>) - Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Set a default policy with some explicit handlers for specific exceptions.
- retryByRemoteException(RetryPolicy, Map<Class<? extends Exception>, RetryPolicy>) - Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
A retry policy for RemoteException
Set a default policy with some explicit handlers for specific exceptions.
- RetryPolicies - Class in org.apache.hadoop.io.retry
-
- RetryPolicies() - Constructor for class org.apache.hadoop.io.retry.RetryPolicies
-
- RetryPolicy - Interface in org.apache.hadoop.io.retry
-
Specifies a policy for retrying method failures.
- RetryProxy - Class in org.apache.hadoop.io.retry
-
A factory for creating retry proxies.
- RetryProxy() - Constructor for class org.apache.hadoop.io.retry.RetryProxy
-
- retryUpToMaximumCountWithFixedSleep(int, long, TimeUnit) - Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying a limited number of times, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
- retryUpToMaximumCountWithProportionalSleep(int, long, TimeUnit) - Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying a limited number of times, waiting a growing amount of time between attempts,
and then fail by re-throwing the exception.
- retryUpToMaximumTimeWithFixedSleep(long, long, TimeUnit) - Static method in class org.apache.hadoop.io.retry.RetryPolicies
-
Keep trying for a maximum time, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
- returnCompressor(Compressor) - Static method in class org.apache.hadoop.io.compress.CodecPool
-
- returnDecompressor(Decompressor) - Static method in class org.apache.hadoop.io.compress.CodecPool
-
- reverseDns(InetAddress, String) - Static method in class org.apache.hadoop.net.DNS
-
Returns the hostname associated with the specified IP address by the
provided nameserver.
- rewind() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Rewind to the first entry in the scanner.
- RIO_PREFIX - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RIOType() - Constructor for class org.apache.hadoop.record.meta.TypeID.RIOType
-
- rjustify(String, int) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- rNums - Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
This array really shouldn't be here.
- ROOT - Static variable in class org.apache.hadoop.net.NodeBase
-
- RoundRobinDistributionPolicy - Class in org.apache.hadoop.contrib.index.example
-
Choose a shard for each insert in a round-robin fashion.
- RoundRobinDistributionPolicy() - Constructor for class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
-
- RPC - Class in org.apache.hadoop.ipc
-
A simple RPC mechanism.
- RPC.Server - Class in org.apache.hadoop.ipc
-
An RPC Server.
- RPC.VersionMismatch - Exception in org.apache.hadoop.ipc
-
A version mismatch for the RPC protocol.
- RpcActivityMBean - Class in org.apache.hadoop.ipc.metrics
-
This is the JMX MBean for reporting the RPC layer Activity.
- RpcActivityMBean(MetricsRegistry, String, String) - Constructor for class org.apache.hadoop.ipc.metrics.RpcActivityMBean
-
- RpcMetrics - Class in org.apache.hadoop.ipc.metrics
-
This class is for maintaining the various RPC statistics
and publishing them through the metrics interfaces.
- RpcMetrics(String, String, Server) - Constructor for class org.apache.hadoop.ipc.metrics.RpcMetrics
-
- rpcMetrics - Variable in class org.apache.hadoop.ipc.Server
-
- RpcMgtMBean - Interface in org.apache.hadoop.ipc.metrics
-
This is the JMX management interface for the RPC layer.
- rpcProcessingTime - Variable in class org.apache.hadoop.ipc.metrics.RpcMetrics
-
- rpcQueueTime - Variable in class org.apache.hadoop.ipc.metrics.RpcMetrics
-
The metrics variables are public:
- they can be set directly by calling their set/inc methods
-they can also be read directly - e.g.
- rrClass - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- rrConstructor - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- rrCstrMap - Static variable in class org.apache.hadoop.mapred.join.Parser.Node
-
- RTI_FILTER - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RTI_FILTER_FIELDS - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- RTI_VAR - Static variable in class org.apache.hadoop.record.compiler.Consts
-
- run() - Method in class org.apache.hadoop.contrib.failmon.Executor
-
- run(Configuration, Path[], Path, int, Shard[]) - Method in interface org.apache.hadoop.contrib.index.mapred.IIndexUpdater
-
Create a Map/Reduce job configuration and run the Map/Reduce job to
analyze documents and update Lucene instances in parallel.
- run(Configuration, Path[], Path, int, Shard[]) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdater
-
- run(String[]) - Method in class org.apache.hadoop.examples.dancing.DistributedPentomino
-
- run(String[]) - Method in class org.apache.hadoop.examples.DBCountPageView
-
- run(String[]) - Method in class org.apache.hadoop.examples.Grep
-
- run(String[]) - Method in class org.apache.hadoop.examples.Join
-
The main driver for sort program.
- run(String[]) - Method in class org.apache.hadoop.examples.MultiFileWordCount
-
- run(String[]) - Method in class org.apache.hadoop.examples.PiEstimator
-
Parse arguments and then runs a map/reduce job.
- run(String[]) - Method in class org.apache.hadoop.examples.RandomTextWriter
-
This is the main routine for launching a distributed random write job.
- run(String[]) - Method in class org.apache.hadoop.examples.RandomWriter
-
This is the main routine for launching a distributed random write job.
- run(int, int, long, int, long, int) - Method in class org.apache.hadoop.examples.SleepJob
-
- run(String[]) - Method in class org.apache.hadoop.examples.SleepJob
-
- run(String[]) - Method in class org.apache.hadoop.examples.Sort
-
The main driver for sort program.
- run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraGen
-
- run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraSort
-
- run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraValidate
-
- run(String[]) - Method in class org.apache.hadoop.fs.FsShell
-
run
- run(String[]) - Method in class org.apache.hadoop.fs.s3.MigrationTool
-
- run(Path) - Method in class org.apache.hadoop.fs.shell.Command
-
Execute the command on the input path
- run(Path) - Method in class org.apache.hadoop.fs.shell.Count
-
- run(String[]) - Method in class org.apache.hadoop.mapred.JobClient
-
- run() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
The main loop for the thread.
- run() - Method in class org.apache.hadoop.mapred.JobHistory.HistoryCleaner
-
Cleans up history data.
- run(String[]) - Method in class org.apache.hadoop.mapred.lib.InputSampler
-
Driver for InputSampler from the command line.
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
-
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in interface org.apache.hadoop.mapred.MapRunnable
-
Deprecated.
Start mapping input <key, value> pairs.
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.mapred.MapRunner
-
- run(String[]) - Method in class org.apache.hadoop.mapred.pipes.Submitter
-
- run() - Method in class org.apache.hadoop.mapred.TaskTracker
-
The server retry loop.
- run(String[]) - Method in class org.apache.hadoop.mapred.tools.MRAdmin
-
- run(Mapper<K1, V1, K2, V2>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
Run the application's maps using a thread pool.
- run(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
-
Expert users can override this method for more complete control over the
execution of the Mapper.
- run(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
-
- run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.streaming.PipeMapRunner
-
- run(String[]) - Method in class org.apache.hadoop.streaming.StreamJob
-
- run() - Method in class org.apache.hadoop.util.Shell
-
check to see if a command needs to be executed and execute if needed
- run(String[]) - Method in interface org.apache.hadoop.util.Tool
-
Execute the command with the given arguments.
- run(Configuration, Tool, String[]) - Static method in class org.apache.hadoop.util.ToolRunner
-
Runs the given
Tool
by
Tool.run(String[])
, after
parsing with the given generic arguments.
- run(Tool, String[]) - Static method in class org.apache.hadoop.util.ToolRunner
-
Runs the Tool
with its Configuration
.
- RUNA - Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- runAll() - Method in class org.apache.hadoop.fs.shell.Command
-
For each source path, execute the command
- RUNB - Static variable in interface org.apache.hadoop.io.compress.bzip2.BZip2Constants
-
- runCommand(String[]) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Runs a shell command in the system and provides a StringBuffer
with the output of the command.
- runCommand(String) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Runs a shell command in the system and provides a StringBuffer
with the output of the command.
- RunJar - Class in org.apache.hadoop.util
-
Run a Hadoop job jar.
- RunJar() - Constructor for class org.apache.hadoop.util.RunJar
-
- runJob(JobConf) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
-
Submit/run a map/reduce job.
- runJob(JobConf) - Static method in class org.apache.hadoop.mapred.JobClient
-
Utility that submits a job, then polls for progress until the job is
complete.
- runJob(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Submit a job to the map/reduce cluster.
- RUNNING - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- RUNNING - Static variable in class org.apache.hadoop.mapred.JobStatus
-
- running_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- RunningJob - Interface in org.apache.hadoop.mapred
-
RunningJob
is the user-interface to query for details on a
running Map-Reduce job.
- runningJobs() - Method in class org.apache.hadoop.mapred.JobTracker
-
- RunOnce - Class in org.apache.hadoop.contrib.failmon
-
Runs a set of monitoring jobs once for the local node.
- RunOnce(String) - Constructor for class org.apache.hadoop.contrib.failmon.RunOnce
-
- S3Credentials - Class in org.apache.hadoop.fs.s3
-
Extracts AWS credentials from the filesystem URI or configuration.
- S3Credentials() - Constructor for class org.apache.hadoop.fs.s3.S3Credentials
-
- S3Exception - Exception in org.apache.hadoop.fs.s3
-
Thrown if there is a problem communicating with Amazon S3.
- S3Exception(Throwable) - Constructor for exception org.apache.hadoop.fs.s3.S3Exception
-
- S3FileSystem - Class in org.apache.hadoop.fs.s3
-
- S3FileSystem() - Constructor for class org.apache.hadoop.fs.s3.S3FileSystem
-
- S3FileSystem(FileSystemStore) - Constructor for class org.apache.hadoop.fs.s3.S3FileSystem
-
- S3FileSystemException - Exception in org.apache.hadoop.fs.s3
-
Thrown when there is a fatal exception while using
S3FileSystem
.
- S3FileSystemException(String) - Constructor for exception org.apache.hadoop.fs.s3.S3FileSystemException
-
- safeGetCanonicalPath(File) - Static method in class org.apache.hadoop.streaming.StreamUtil
-
- saveToConf(Configuration, String, UnixUserGroupInformation) - Static method in class org.apache.hadoop.security.UnixUserGroupInformation
-
Store the given ugi
as a comma separated string in
conf
as a property attr
The String starts with the user name followed by the default group names,
and other group names.
- Scanner(TFile.Reader, long, long) - Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Constructor
- Scanner(TFile.Reader, RawComparable, RawComparable) - Constructor for class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Constructor
- ScriptBasedMapping - Class in org.apache.hadoop.net
-
This class implements the
DNSToSwitchMapping
interface using a
script configured via topology.script.file.name .
- ScriptBasedMapping() - Constructor for class org.apache.hadoop.net.ScriptBasedMapping
-
- ScriptBasedMapping(Configuration) - Constructor for class org.apache.hadoop.net.ScriptBasedMapping
-
- SecondarySort - Class in org.apache.hadoop.examples
-
This is an example Hadoop Map/Reduce application.
- SecondarySort() - Constructor for class org.apache.hadoop.examples.SecondarySort
-
- SecondarySort.FirstGroupingComparator - Class in org.apache.hadoop.examples
-
Compare only the first part of the pair, so that reduce is called once
for each value of the first part.
- SecondarySort.FirstPartitioner - Class in org.apache.hadoop.examples
-
Partition based on the first part of the pair.
- SecondarySort.IntPair - Class in org.apache.hadoop.examples
-
Define a pair of integers that are writable.
- SecondarySort.IntPair.Comparator - Class in org.apache.hadoop.examples
-
A Comparator that compares serialized IntPair.
- SecondarySort.MapClass - Class in org.apache.hadoop.examples
-
Read two integers from each line and generate a key, value pair
as ((left, right), right).
- SecondarySort.Reduce - Class in org.apache.hadoop.examples
-
A reducer class that just emits the sum of the input values.
- SecurityUtil - Class in org.apache.hadoop.security
-
- SecurityUtil() - Constructor for class org.apache.hadoop.security.SecurityUtil
-
- SecurityUtil.AccessControlList - Class in org.apache.hadoop.security
-
Class representing a configured access control list.
- seek(long) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- seek(long) - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- seek(long) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Seek to the given position in the stream.
- seek(long) - Method in class org.apache.hadoop.fs.FSInputStream
-
Seek to the given offset from the start of the file.
- seek(long) - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- seek(long) - Method in interface org.apache.hadoop.fs.Seekable
-
Seek to the given offset from the start of the file.
- seek(long) - Method in class org.apache.hadoop.io.ArrayFile.Reader
-
Positions the reader before its n
th value.
- seek(WritableComparable) - Method in class org.apache.hadoop.io.MapFile.Reader
-
Positions the reader at the named key, or if none such exists, at the
first entry after the named key.
- seek(long) - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Set the current byte position in the input file.
- seek(WritableComparable) - Method in class org.apache.hadoop.io.SetFile.Reader
-
- seek(long) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- Seekable - Interface in org.apache.hadoop.fs
-
Stream that permits seeking.
- seekNextRecordBoundary() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
-
Implementation should seek forward in_ to the first byte of the next record.
- seekNextRecordBoundary() - Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- seekTo(byte[]) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Move the cursor to the first entry whose key is greater than or equal
to the input key.
- seekTo(byte[], int, int) - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Move the cursor to the first entry whose key is greater than or equal
to the input key.
- seekToEnd() - Method in class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner
-
Seek to the end of the scanner.
- seekToNewSource(long) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- seekToNewSource(long) - Method in class org.apache.hadoop.fs.FSDataInputStream
-
- seekToNewSource(long) - Method in class org.apache.hadoop.fs.FSInputStream
-
Seeks a different copy of the data.
- seekToNewSource(long) - Method in class org.apache.hadoop.fs.ftp.FTPInputStream
-
- seekToNewSource(long) - Method in interface org.apache.hadoop.fs.Seekable
-
Seeks a different copy of the data.
- seenPrimary_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- SegmentDescriptor(long, long, Path) - Constructor for class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
Constructs a segment
- selectiveClearing(Key, short) - Method in class org.apache.hadoop.util.bloom.RetouchedBloomFilter
-
Performs the selective clearing for a given key.
- SEMICOLON_TKN - Static variable in interface org.apache.hadoop.record.compiler.generated.RccConstants
-
RegularExpression Id.
- SensorsParser - Class in org.apache.hadoop.contrib.failmon
-
Objects of this class parse the output of the lm-sensors utility
to gather information about fan speed, temperatures for cpus
and motherboard etc.
- SensorsParser() - Constructor for class org.apache.hadoop.contrib.failmon.SensorsParser
-
- SEPARATOR - Static variable in class org.apache.hadoop.fs.Path
-
The directory separator, a slash.
- SEPARATOR - Static variable in class org.apache.hadoop.mapreduce.ID
-
- SEPARATOR_CHAR - Static variable in class org.apache.hadoop.fs.Path
-
- SequenceFile - Class in org.apache.hadoop.io
-
SequenceFile
s are flat files consisting of binary key/value
pairs.
- SequenceFile.CompressionType - Enum in org.apache.hadoop.io
-
The compression type used to compress key/value pairs in the
SequenceFile
.
- SequenceFile.Metadata - Class in org.apache.hadoop.io
-
The class encapsulating with the metadata of a file.
- SequenceFile.Reader - Class in org.apache.hadoop.io
-
Reads key/value pairs from a sequence-format file.
- SequenceFile.Sorter - Class in org.apache.hadoop.io
-
Sorts key/value pairs in a sequence-format file.
- SequenceFile.Sorter.RawKeyValueIterator - Interface in org.apache.hadoop.io
-
The interface to iterate over raw keys/values of SequenceFiles.
- SequenceFile.Sorter.SegmentDescriptor - Class in org.apache.hadoop.io
-
This class defines a merge segment.
- SequenceFile.ValueBytes - Interface in org.apache.hadoop.io
-
The interface to 'raw' values of SequenceFiles.
- SequenceFile.Writer - Class in org.apache.hadoop.io
-
Write key/value pairs to a sequence-format file.
- SequenceFileAsBinaryInputFormat - Class in org.apache.hadoop.mapred
-
InputFormat reading keys, values from SequenceFiles in binary (raw)
format.
- SequenceFileAsBinaryInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
-
- SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader - Class in org.apache.hadoop.mapred
-
Read records from a SequenceFile as binary (raw) bytes.
- SequenceFileAsBinaryOutputFormat - Class in org.apache.hadoop.mapred
-
- SequenceFileAsBinaryOutputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- SequenceFileAsBinaryOutputFormat.WritableValueBytes - Class in org.apache.hadoop.mapred
-
Inner class used for appendRaw
- SequenceFileAsBinaryRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
-
- SequenceFileAsTextInputFormat - Class in org.apache.hadoop.mapred
-
This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader
which converts the input keys and values to their String forms by calling toString() method.
- SequenceFileAsTextInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
-
- SequenceFileAsTextRecordReader - Class in org.apache.hadoop.mapred
-
This class converts the input keys and values to their String forms by calling toString()
method.
- SequenceFileAsTextRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
-
- SequenceFileInputFilter<K,V> - Class in org.apache.hadoop.mapred
-
A class that allows a map/red job to work on a sample of sequence files.
- SequenceFileInputFilter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter
-
- SequenceFileInputFilter.Filter - Interface in org.apache.hadoop.mapred
-
filter interface
- SequenceFileInputFilter.FilterBase - Class in org.apache.hadoop.mapred
-
base class for Filters
- SequenceFileInputFilter.MD5Filter - Class in org.apache.hadoop.mapred
-
This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency f.
- SequenceFileInputFilter.PercentFilter - Class in org.apache.hadoop.mapred
-
This class returns a percentage of records
The percentage is determined by a filtering frequency f using
the criteria record# % f == 0.
- SequenceFileInputFilter.RegexFilter - Class in org.apache.hadoop.mapred
-
Records filter by matching key to regex
- SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapred
-
- SequenceFileInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFormat
-
Deprecated.
- SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
-
- SequenceFileInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
-
- SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapred
-
- SequenceFileOutputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
Deprecated.
- SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
-
- SequenceFileOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
-
- SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapred
-
- SequenceFileRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileRecordReader
-
- SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
-
- SequenceFileRecordReader() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
-
- Serialization<T> - Interface in org.apache.hadoop.io.serializer
-
- SerializationFactory - Class in org.apache.hadoop.io.serializer
-
- SerializationFactory(Configuration) - Constructor for class org.apache.hadoop.io.serializer.SerializationFactory
-
Serializations are found by reading the io.serializations
property from conf
, which is a comma-delimited list of
classnames.
- serialize() - Method in class org.apache.hadoop.fs.s3.INode
-
- serialize(T) - Method in interface org.apache.hadoop.io.serializer.Serializer
-
Serialize t
to the underlying output stream.
- serialize(RecordOutput, String) - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
Serialize the type information for a record
- serialize(RecordOutput, String) - Method in class org.apache.hadoop.record.Record
-
Serialize a record with tag (ususally field name)
- serialize(RecordOutput) - Method in class org.apache.hadoop.record.Record
-
Serialize a record without a tag
- SerializedRecord - Class in org.apache.hadoop.contrib.failmon
-
Objects of this class hold the serialized representations
of EventRecords.
- SerializedRecord(EventRecord) - Constructor for class org.apache.hadoop.contrib.failmon.SerializedRecord
-
Create the SerializedRecord given an EventRecord.
- Serializer<T> - Interface in org.apache.hadoop.io.serializer
-
Provides a facility for serializing objects of type
to an
OutputStream
.
- Server(Object, Configuration, String, int) - Constructor for class org.apache.hadoop.ipc.RPC.Server
-
Construct an RPC server.
- Server(Object, Configuration, String, int, int, boolean) - Constructor for class org.apache.hadoop.ipc.RPC.Server
-
Construct an RPC server.
- Server - Class in org.apache.hadoop.ipc
-
An abstract IPC service.
- Server(String, int, Class<? extends Writable>, int, Configuration) - Constructor for class org.apache.hadoop.ipc.Server
-
- Server(String, int, Class<? extends Writable>, int, Configuration, String) - Constructor for class org.apache.hadoop.ipc.Server
-
Constructs a server listening on the named port and address.
- Service - Class in org.apache.hadoop.security.authorize
-
An abstract definition of service as related to
Service Level Authorization for Hadoop.
- Service(String, Class<?>) - Constructor for class org.apache.hadoop.security.authorize.Service
-
- SERVICE_AUTHORIZATION_CONFIG - Static variable in class org.apache.hadoop.security.authorize.ServiceAuthorizationManager
-
Configuration key for controlling service-level authorization for Hadoop.
- ServiceAuthorizationManager - Class in org.apache.hadoop.security.authorize
-
An authorization manager which handles service-level authorization
for incoming service requests.
- ServiceAuthorizationManager() - Constructor for class org.apache.hadoop.security.authorize.ServiceAuthorizationManager
-
- Servlet() - Constructor for class org.apache.hadoop.log.LogLevel.Servlet
-
- ServletUtil - Class in org.apache.hadoop.util
-
- ServletUtil() - Constructor for class org.apache.hadoop.util.ServletUtil
-
- set(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value
of the name
property.
- set(String, Object) - Method in class org.apache.hadoop.contrib.failmon.EventRecord
-
Set the value of a property of the EventRecord.
- set(String, String) - Method in class org.apache.hadoop.contrib.failmon.SerializedRecord
-
Set the value of a property of the EventRecord.
- set(int, int) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
-
Set the left and right values.
- set(boolean, Checksum, int, int) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Set the checksum related parameters
- set(Writable[]) - Method in class org.apache.hadoop.io.ArrayWritable
-
- set(boolean) - Method in class org.apache.hadoop.io.BooleanWritable
-
Set the value of the BooleanWritable
- set(BytesWritable) - Method in class org.apache.hadoop.io.BytesWritable
-
Set the BytesWritable to the contents of the given newData.
- set(byte[], int, int) - Method in class org.apache.hadoop.io.BytesWritable
-
Set the value to a copy of the given byte range
- set(byte) - Method in class org.apache.hadoop.io.ByteWritable
-
Set the value of this ByteWritable.
- set(double) - Method in class org.apache.hadoop.io.DoubleWritable
-
- set(float) - Method in class org.apache.hadoop.io.FloatWritable
-
Set the value of this FloatWritable.
- set(Writable) - Method in class org.apache.hadoop.io.GenericWritable
-
Set the instance that is wrapped.
- set(int) - Method in class org.apache.hadoop.io.IntWritable
-
Set the value of this IntWritable.
- set(long) - Method in class org.apache.hadoop.io.LongWritable
-
Set the value of this LongWritable.
- set(MD5Hash) - Method in class org.apache.hadoop.io.MD5Hash
-
Copy the contents of another instance into this instance.
- set(Object) - Method in class org.apache.hadoop.io.ObjectWritable
-
Reset the instance.
- set(Text, Text) - Method in class org.apache.hadoop.io.SequenceFile.Metadata
-
- set(String) - Method in class org.apache.hadoop.io.Text
-
Set to contain the contents of a string.
- set(byte[]) - Method in class org.apache.hadoop.io.Text
-
Set to a utf8 byte array
- set(Text) - Method in class org.apache.hadoop.io.Text
-
copy a text.
- set(byte[], int, int) - Method in class org.apache.hadoop.io.Text
-
Set the Text to range of bytes
- set(Writable[][]) - Method in class org.apache.hadoop.io.TwoDArrayWritable
-
- set(String) - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Set to contain the contents of a string.
- set(UTF8) - Method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Set to contain the contents of a string.
- set(int) - Method in class org.apache.hadoop.io.VIntWritable
-
Set the value of this VIntWritable.
- set(long) - Method in class org.apache.hadoop.io.VLongWritable
-
Set the value of this LongWritable.
- set(int) - Method in class org.apache.hadoop.metrics.util.MetricsIntValue
-
Set the value
- set(long) - Method in class org.apache.hadoop.metrics.util.MetricsLongValue
-
Set the value
- set(byte[]) - Method in class org.apache.hadoop.record.Buffer
-
Use the specified bytes array as underlying sequence.
- set(byte[], double) - Method in class org.apache.hadoop.util.bloom.Key
-
- set(float) - Method in class org.apache.hadoop.util.Progress
-
Called during execution on a leaf node to set its progress.
- SET_GROUP_COMMAND - Static variable in class org.apache.hadoop.util.Shell
-
- SET_OWNER_COMMAND - Static variable in class org.apache.hadoop.util.Shell
-
a Unix command to set owner
- SET_PERMISSION_COMMAND - Static variable in class org.apache.hadoop.util.Shell
-
a Unix command to set permission
- setAggregatorDescriptors(JobConf, Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
-
- setArchiveTimestamps(Configuration, String) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This is to check the timestamp of the archives to be localized
- setAssignedJobID(JobID) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the mapred ID for this job as assigned by the
mapred framework.
- setAttemptsToStartSkipping(Configuration, int) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Set the number of Task attempts AFTER which skip mode
will be kicked off.
- setAttribute(String, Object) - Method in class org.apache.hadoop.http.HttpServer
-
Set a value in the webapp context.
- setAttribute(String, Object) - Method in class org.apache.hadoop.metrics.ContextFactory
-
Sets the named factory attribute to the specified value, creating it
if it did not already exist.
- setAttribute(Attribute) - Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
-
- setAttributes(AttributeList) - Method in class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase
-
- setAutoIncrMapperProcCount(Configuration, boolean) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
- setAutoIncrReducerProcCount(Configuration, boolean) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
- setBoolean(String, boolean) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value of the name
property to a boolean
.
- setBooleanIfUnset(String, boolean) - Method in class org.apache.hadoop.conf.Configuration
-
Set the given property, if it is currently unset.
- setCacheArchives(URI[], Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Set the configuration with the given set of archives
- setCacheFiles(URI[], Configuration) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Set the configuration with the given set of files
- setCapacity(int) - Method in class org.apache.hadoop.io.BytesWritable
-
Change the capacity of the backing storage.
- setCapacity(int) - Method in class org.apache.hadoop.record.Buffer
-
Change the capacity of the backing storage.
- setClass(String, Class<?>, Class<?>) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value of the name
property to the name of a
theClass
implementing the given interface xface
.
- setClassLoader(ClassLoader) - Method in class org.apache.hadoop.conf.Configuration
-
Set the class loader that will be used to load the various objects.
- setCodecClasses(Configuration, List<Class>) - Static method in class org.apache.hadoop.io.compress.CompressionCodecFactory
-
Sets a list of codec classes in the configuration.
- setCombinerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the user-defined combiner class used to combine map-outputs
before being sent to the reducers.
- setCombinerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the combiner class for the job.
- setCompressionType(Configuration, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.io.SequenceFile
-
- setCompressMapOutput(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Should the map outputs be compressed before transfer?
Uses the SequenceFile compression.
- setCompressOutput(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Set whether the output of the job is compressed.
- setCompressOutput(Job, boolean) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Set whether the output of the job is compressed.
- setConf(Configuration) - Method in interface org.apache.hadoop.conf.Configurable
-
Set the configuration to be used by this object.
- setConf(Configuration) - Method in class org.apache.hadoop.conf.Configured
-
- setConf(Configuration) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- setConf(Configuration) - Method in class org.apache.hadoop.io.AbstractMapWritable
-
- setConf(Configuration) - Method in class org.apache.hadoop.io.compress.DefaultCodec
-
- setConf(Configuration) - Method in class org.apache.hadoop.io.GenericWritable
-
- setConf(Configuration) - Method in class org.apache.hadoop.io.ObjectWritable
-
- setConf(Configuration) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Set the configuration to be used by this object.
- setConf(Configuration) - Method in class org.apache.hadoop.mapred.lib.InputSampler
-
- setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
-
configure the filter according to configuration
- setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
-
configure the filter by checking the configuration
- setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
-
configure the Filter by checking the configuration
- setConf(Configuration) - Method in class org.apache.hadoop.net.ScriptBasedMapping
-
- setConf(Configuration) - Method in class org.apache.hadoop.net.SocksSocketFactory
-
- setConf(Configuration) - Method in class org.apache.hadoop.security.authorize.ConfiguredPolicy
-
- setConf(Configuration) - Method in class org.apache.hadoop.streaming.StreamJob
-
- setConf(Object, Configuration) - Static method in class org.apache.hadoop.util.ReflectionUtils
-
Check and set 'configuration' if necessary.
- setContentionTracing(boolean) - Static method in class org.apache.hadoop.util.ReflectionUtils
-
- setCountersEnabled(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
-
Enables or disables counters for the named outputs.
- setCurrentUGI(UserGroupInformation) - Static method in class org.apache.hadoop.security.UserGroupInformation
-
- setCurrentUser(UserGroupInformation) - Static method in class org.apache.hadoop.security.UserGroupInformation
-
Set the
UserGroupInformation
for the current thread
WARNING - This method should be used only in test cases and other exceptional
cases!
- setDebugStream(PrintStream) - Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Set debug output.
- setDefaultUri(Configuration, URI) - Static method in class org.apache.hadoop.fs.FileSystem
-
Set the default filesystem URI in a configuration.
- setDefaultUri(Configuration, String) - Static method in class org.apache.hadoop.fs.FileSystem
-
Set the default filesystem URI in a configuration.
- setDelete(Term) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Set the instance to be a delete operation.
- setDestdir(File) - Method in class org.apache.hadoop.record.compiler.ant.RccTask
-
Sets directory where output files will be generated
- setDictionary(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- setDictionary(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- setDictionary(byte[], int, int) - Method in interface org.apache.hadoop.io.compress.Compressor
-
Sets preset dictionary for compression.
- setDictionary(byte[], int, int) - Method in interface org.apache.hadoop.io.compress.Decompressor
-
Sets preset dictionary for compression.
- setDictionary(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- setDictionary(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- setDigest(String) - Method in class org.apache.hadoop.io.MD5Hash
-
Sets the digest value from a hex string.
- setDisableHistory(boolean) - Static method in class org.apache.hadoop.mapred.JobHistory
-
Enable/disable history logging.
- setDisplayName(String) - Method in class org.apache.hadoop.mapred.Counters.Counter
-
Deprecated.
- setDisplayName(String) - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Set the display name
- setDisplayName(String) - Method in class org.apache.hadoop.mapreduce.Counter
-
Deprecated.
- setDistributionPolicyClass(Class<? extends IDistributionPolicy>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the distribution policy class.
- setDocumentAnalyzerClass(Class<? extends Analyzer>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the analyzer class.
- setDoubleValue(Object, double) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
Set the given counter to the given value
- setEnvironment(Map<String, String>) - Method in class org.apache.hadoop.util.Shell
-
set the environment for the command
- setEventId(int) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
set event Id.
- setExcludesFile(String) - Method in class org.apache.hadoop.util.HostsFileReader
-
- setExecutable(JobConf, String) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set the URI for the application's executable.
- setFactor(int) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Set the number of streams to merge at once.
- setFactory(Class, WritableFactory) - Static method in class org.apache.hadoop.io.WritableFactories
-
Define a factory for a class.
- setFailonerror(boolean) - Method in class org.apache.hadoop.record.compiler.ant.RccTask
-
Given multiple files (via fileset), set the error handling behavior
- SetFile - Class in org.apache.hadoop.io
-
A file-based set of keys.
- SetFile() - Constructor for class org.apache.hadoop.io.SetFile
-
- setFile(File) - Method in class org.apache.hadoop.record.compiler.ant.RccTask
-
Sets the record definition file attribute
- SetFile.Reader - Class in org.apache.hadoop.io
-
Provide access to an existing set file.
- SetFile.Writer - Class in org.apache.hadoop.io
-
Write a new set file.
- setFileTimestamps(Configuration, String) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
This is to check the timestamp of the files to be localized
- setFilterClass(Configuration, Class) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter
-
set the filter class
- setFinalSync(JobConf, boolean) - Static method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
-
Set the requirement for a final sync before the stream is closed.
- setFloat(String, float) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value of the name
property to a float
.
- setFormat(JobConf) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
-
Interpret a given string as a composite expression.
- setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
-
set the filtering frequency in configuration
- setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
-
set the frequency and stores it in conf
- setGroup(String) - Method in class org.apache.hadoop.fs.FileStatus
-
Sets group.
- setGroupingComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setHosts(String[]) - Method in class org.apache.hadoop.fs.BlockLocation
-
Set the hosts hosting this block
- setID(int) - Method in class org.apache.hadoop.mapred.join.Parser.Node
-
- setIfUnset(String, String) - Method in class org.apache.hadoop.conf.Configuration
-
Sets a property if it is currently unset.
- setIncludesFile(String) - Method in class org.apache.hadoop.util.HostsFileReader
-
- setIndexInputFormatClass(Class<? extends InputFormat>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the index input format class.
- setIndexInterval(int) - Method in class org.apache.hadoop.io.MapFile.Writer
-
Sets the index interval.
- setIndexInterval(Configuration, int) - Static method in class org.apache.hadoop.io.MapFile.Writer
-
Sets the index interval and stores it in conf
- setIndexMaxFieldLength(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the max field length for a Lucene instance.
- setIndexMaxNumSegments(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the max number of segments for a Lucene instance.
- setIndexShards(String) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the string representation of a number of shards.
- setIndexShards(IndexUpdateConfiguration, Shard[]) - Static method in class org.apache.hadoop.contrib.index.mapred.Shard
-
- setIndexUpdaterClass(Class<? extends IIndexUpdater>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the index updater class.
- setIndexUseCompoundFile(boolean) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set whether use the compound file format for a Lucene instance.
- setInput(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor
-
- setInput(byte[], int, int) - Method in class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor
-
- setInput(byte[], int, int) - Method in interface org.apache.hadoop.io.compress.Compressor
-
Sets input data for compression.
- setInput(byte[], int, int) - Method in interface org.apache.hadoop.io.compress.Decompressor
-
Sets input data for decompression.
- setInput(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.ZlibCompressor
-
- setInput(byte[], int, int) - Method in class org.apache.hadoop.io.compress.zlib.ZlibDecompressor
-
- setInput(JobConf, Class<? extends DBWritable>, String, String, String, String...) - Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
Initializes the map-part of the job with the appropriate input settings.
- setInput(JobConf, Class<? extends DBWritable>, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
-
Initializes the map-part of the job with the appropriate input settings.
- setInputFormat(Class<? extends InputFormat>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the
InputFormat
implementation for the map-reduce job.
- setInputFormatClass(Class<? extends InputFormat>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setInputPathFilter(JobConf, Class<? extends PathFilter>) - Static method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
Set a PathFilter to be applied to the input paths for the map-reduce job.
- setInputPathFilter(Job, Class<? extends PathFilter>) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Set a PathFilter to be applied to the input paths for the map-reduce job.
- setInputPaths(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
Sets the given comma separated paths as the list of inputs
for the map-reduce job.
- setInputPaths(JobConf, Path...) - Static method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
Set the array of
Path
s as the list of inputs
for the map-reduce job.
- setInputPaths(Job, String) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Sets the given comma separated paths as the list of inputs
for the map-reduce job.
- setInputPaths(Job, Path...) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Set the array of
Path
s as the list of inputs
for the map-reduce job.
- setInsert(Document) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Set the instance to be an insert operation.
- setInstrumentationClass(Configuration, Class<? extends JobTrackerInstrumentation>) - Static method in class org.apache.hadoop.mapred.JobTracker
-
- setInstrumentationClass(Configuration, Class<? extends TaskTrackerInstrumentation>) - Static method in class org.apache.hadoop.mapred.TaskTracker
-
- setInt(String, int) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value of the name
property to an int
.
- setIOSortMB(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the IO sort space in MB.
- setIsJavaMapper(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set whether the Mapper is written in Java.
- setIsJavaRecordReader(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set whether the job is using a Java RecordReader.
- setIsJavaRecordWriter(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set whether the job will use a Java RecordWriter.
- setIsJavaReducer(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set whether the Reducer is written in Java.
- setJar(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the user jar for the map-reduce job.
- setJarByClass(Class) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the job's jar file by finding an example class location.
- setJarByClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the Jar by finding where a given class came from.
- setJobConf(JobConf) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the mapred job conf for this job.
- setJobConf() - Method in class org.apache.hadoop.streaming.StreamJob
-
- setJobEndNotificationURI(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
- setJobID(String) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the job ID for this job.
- setJobName(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the user-specified job name.
- setJobName(String) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the job name for this job.
- setJobName(String) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the user-specified job name.
- setJobPriority(JobPriority) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setJobPriority(JobPriority) - Method in class org.apache.hadoop.mapred.JobStatus
-
Set the priority of the job, defaulting to NORMAL.
- setJobPriority(JobID, String) - Method in class org.apache.hadoop.mapred.JobTracker
-
Set the priority of a job
- setJobPriority(String) - Method in interface org.apache.hadoop.mapred.RunningJob
-
Set the priority of a running job.
- setKeepCommandFile(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
Set whether to keep the command file for debugging
- setKeepFailedTaskFiles(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set whether the framework should keep the intermediate files for
failed tasks.
- setKeepTaskFilesPattern(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set a regular expression for task names that should be kept.
- setKeyComparator(Class<? extends WritableComparator>) - Method in class org.apache.hadoop.mapred.join.Parser.Node
-
- setKeyFieldComparatorOptions(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setKeyFieldPartitionerOptions(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setLanguage(String) - Method in class org.apache.hadoop.record.compiler.ant.RccTask
-
Sets the output language option
- setLength(long) - Method in class org.apache.hadoop.fs.BlockLocation
-
Set the length of block
- setLevel(int) - Method in interface org.apache.hadoop.net.Node
-
Set this node's level in the tree.
- setLevel(int) - Method in class org.apache.hadoop.net.NodeBase
-
Set this node's level in the tree
- setLoadNativeLibraries(Configuration, boolean) - Method in class org.apache.hadoop.util.NativeCodeLoader
-
Set if native hadoop libraries, if present, can be used for this job.
- setLocalAnalysisClass(Class<? extends ILocalAnalysis>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
-
Set the local analysis class.
- setLocalArchives(Configuration, String) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Set the conf to contain the location for localized archives
- setLocalFiles(Configuration, String) - Static method in class org.apache.hadoop.filecache.DistributedCache
-
Set the conf to contain the location for localized files
- setLong(String, long) - Method in class org.apache.hadoop.conf.Configuration
-
Set the value of the name
property to a long
.
- setLongValue(Object, long) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
-
Set the given counter to the given value
- setMapDebugScript(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the debug script to run when the map tasks fail.
- setMapOutputCompressorClass(Class<? extends CompressionCodec>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setMapOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the key class for the map output data.
- setMapOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the key class for the map output data.
- setMapOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the value class for the map output data.
- setMapOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the value class for the map output data.
- setMapperClass(Class<? extends Mapper>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the
Mapper
class for the job.
- setMapperClass(Class<? extends Mapper>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setMapperClass(Job, Class<? extends Mapper<K1, V1, K2, V2>>) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
Set the application's mapper class.
- setMapperMaxSkipRecords(Configuration, long) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Set the number of acceptable skip records surrounding the bad record PER
bad record in mapper.
- setMapredJobID(String) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
- setMapRunnerClass(Class<? extends MapRunnable>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setMapSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Turn speculative execution on or off for this job for map tasks.
- SETMASK - Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
This constant is accessible by subclasses for historical purposes.
- setMaxInputSplitSize(Job, long) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Set the maximum split size
- setMaxItems(long) - Method in class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
-
Set the limit on the number of unique values
- setMaxMapAttempts(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Expert: Set the number of maximum attempts that will be made to run a
map task.
- setMaxMapTaskFailuresPercent(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Expert: Set the maximum percentage of map tasks that can fail without the
job being aborted.
- setMaxPhysicalMemoryForTask(long) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setMaxReduceAttempts(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Expert: Set the number of maximum attempts that will be made to run a
reduce task.
- setMaxReduceTaskFailuresPercent(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the maximum percentage of reduce tasks that can fail without the job
being aborted.
- setMaxSplitSize(long) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
Specify the maximum size (in bytes) of each split.
- setMaxTaskFailuresPerTracker(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the maximum no.
- setMaxVirtualMemoryForTask(long) - Method in class org.apache.hadoop.mapred.JobConf
-
- setMemory(int) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Set the total amount of buffer memory, in bytes.
- setMemoryForMapTask(long) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setMemoryForReduceTask(long) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setMessage(String) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the message for this job.
- setMetric(String, int) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named metric to the specified value.
- setMetric(String, long) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named metric to the specified value.
- setMetric(String, short) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named metric to the specified value.
- setMetric(String, byte) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named metric to the specified value.
- setMetric(String, float) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named metric to the specified value.
- setMetric(String, int) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named metric to the specified value.
- setMetric(String, long) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named metric to the specified value.
- setMetric(String, short) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named metric to the specified value.
- setMetric(String, byte) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named metric to the specified value.
- setMetric(String, float) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named metric to the specified value.
- setMinInputSplitSize(Job, long) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
-
Set the minimum input split size
- setMinSplitSize(long) - Method in class org.apache.hadoop.mapred.FileInputFormat
-
Deprecated.
- setMinSplitSizeNode(long) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
Specify the minimum size (in bytes) of each split per node.
- setMinSplitSizeRack(long) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
-
Specify the minimum size (in bytes) of each split per rack.
- setName(Class, String) - Static method in class org.apache.hadoop.io.WritableName
-
Set the name that a class should be known as to something other than the
class name.
- setName(String) - Method in class org.apache.hadoop.record.meta.RecordTypeInfo
-
set the name of the record
- setNames(String[]) - Method in class org.apache.hadoop.fs.BlockLocation
-
Set the names (host:port) hosting this block
- setNetworkLocation(String) - Method in interface org.apache.hadoop.net.Node
-
Set the node's network location
- setNetworkLocation(String) - Method in class org.apache.hadoop.net.NodeBase
-
Set this node's network location
- setNetworkProperties() - Method in class org.apache.hadoop.contrib.failmon.LogParser
-
- setNumberOfThreads(Job, int) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
-
Set the number of threads in the pool for running maps.
- setNumMapTasks(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the number of map tasks for this job.
- setNumReduceTasks(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the requisite number of reduce tasks for this job.
- setNumReduceTasks(int) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the number of reduce tasks for the job.
- setNumTasksToExecutePerJvm(int) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Sets the number of tasks that a spawned task JVM should run
before it exits
- setOffset(long) - Method in class org.apache.hadoop.fs.BlockLocation
-
Set the start offset of file associated with this block
- setOp(DocumentAndOp.Op) - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
-
Set the type of the operation.
- setOutput(JobConf, String, String...) - Static method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
-
Initializes the reduce-part of the job with the appropriate output settings
- setOutputCommitter(Class<? extends OutputCommitter>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setOutputCompressionType(JobConf, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
-
Deprecated.
- setOutputCompressionType(Job, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
-
- setOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
- setOutputCompressorClass(Job, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
- setOutputFormat(Class<? extends OutputFormat>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setOutputFormatClass(Class<? extends OutputFormat>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the key class for the job output data.
- setOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the key class for the job output data.
- setOutputKeyComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setOutputPath(JobConf, Path) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
-
Set the
Path
of the output directory for the map-reduce job.
- setOutputPath(Job, Path) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
-
Set the
Path
of the output directory for the map-reduce job.
- setOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the value class for job outputs.
- setOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the value class for job outputs.
- setOutputValueGroupingComparator(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the user defined
RawComparator
comparator for
grouping keys in the input to the reduce.
- setOwner(String) - Method in class org.apache.hadoop.fs.FileStatus
-
Sets owner.
- setOwner(Path, String, String) - Method in class org.apache.hadoop.fs.FileSystem
-
Set owner of a path (i.e.
- setOwner(Path, String, String) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Set owner of a path (i.e.
- setOwner(Path, String, String) - Method in class org.apache.hadoop.fs.HarFileSystem
-
not implemented.
- setOwner(Path, String, String) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
Use the command chown to set owner.
- setParent(Node) - Method in interface org.apache.hadoop.net.Node
-
Set this node's parent
- setParent(Node) - Method in class org.apache.hadoop.net.NodeBase
-
Set this node's parent
- setPartitionerClass(Class<? extends Partitioner>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setPartitionerClass(Class<? extends Partitioner>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setPartitionFile(JobConf, Path) - Static method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
-
Set the path to the SequenceFile storing the sorted partition keyset.
- setPattern(Configuration, String) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
-
Define the filtering regex and stores it in conf
- setPeriod(int) - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Sets the timer period
- setPermission(FsPermission) - Method in class org.apache.hadoop.fs.FileStatus
-
Sets permission.
- setPermission(Path, FsPermission) - Method in class org.apache.hadoop.fs.FileSystem
-
Set permission of a path.
- setPermission(Path, FsPermission) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Set permission of a path.
- setPermission(Path, FsPermission) - Method in class org.apache.hadoop.fs.HarFileSystem
-
Not implemented.
- setPermission(Path, FsPermission) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
Use the command chmod to set permission.
- setPingInterval(Configuration, int) - Static method in class org.apache.hadoop.ipc.Client
-
set the ping interval value in configuration
- setPolicy(Policy) - Static method in class org.apache.hadoop.security.SecurityUtil
-
Set the global security policy for Hadoop.
- setPrinter(DancingLinks.SolutionAcceptor<Pentomino.ColumnName>) - Method in class org.apache.hadoop.examples.dancing.Pentomino
-
Set the printer for the puzzle.
- setProfileEnabled(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set whether the system should collect profiler information for some of
the tasks in this job? The information is stored in the user log
directory.
- setProfileParams(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the profiler configuration arguments.
- setProfileTaskRange(boolean, String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the ranges of maps or reduces to profile.
- setProgressable(Progressable) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Set the progressable object in order to report progress.
- setProperty(String, String) - Static method in class org.apache.hadoop.contrib.failmon.Environment
-
Sets the value of a property inthe configuration file.
- setQueueName(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the name of the queue to which this job should be submitted.
- setQueueName(String) - Method in class org.apache.hadoop.mapred.JobQueueInfo
-
Set the queue name of the JobQueueInfo
- setQuietMode(boolean) - Method in class org.apache.hadoop.conf.Configuration
-
Set the quietness-mode.
- setReduceDebugScript(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the debug script to run when the reduce tasks fail.
- setReducer(JobConf, Class<? extends Reducer<K1, V1, K2, V2>>, Class<? extends K1>, Class<? extends V1>, Class<? extends K2>, Class<? extends V2>, boolean, JobConf) - Static method in class org.apache.hadoop.mapred.lib.ChainReducer
-
Sets the Reducer class to the chain job's JobConf.
- setReducerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
- setReducerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapreduce.Job
-
- setReducerMaxSkipGroups(Configuration, long) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Set the number of acceptable skip groups surrounding the bad group PER
bad group in reducer.
- setReduceSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Turn speculative execution on or off for this job for reduce tasks.
- setReplication(Path, short) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Set replication for an existing file.
- setReplication(Path, short) - Method in class org.apache.hadoop.fs.FileSystem
-
Set replication for an existing file.
- setReplication(Path, short) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Set replication for an existing file.
- setReplication(Path, short) - Method in class org.apache.hadoop.fs.HarFileSystem
-
Not implemented.
- setReplication(Path, short) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- setRunningTaskAttempts(Collection<TaskAttemptID>) - Method in class org.apache.hadoop.mapred.TaskReport
-
set running attempt(s) of the task.
- setRunState(int) - Method in class org.apache.hadoop.mapred.JobStatus
-
Change the current run state of the job.
- setSchedulingInfo(String) - Method in class org.apache.hadoop.mapred.JobQueueInfo
-
Set the scheduling information associated to particular job queue
- setSchedulingInfo(String) - Method in class org.apache.hadoop.mapred.JobStatus
-
Used to set the scheduling information associated to a particular Job.
- setSequenceFileOutputKeyClass(JobConf, Class<?>) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- setSequenceFileOutputValueClass(JobConf, Class<?>) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
-
- setSessionId(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the user-specified session identifier.
- setSigKillInterval(long) - Method in class org.apache.hadoop.util.ProcfsBasedProcessTree
-
- setSize(int) - Method in class org.apache.hadoop.io.BytesWritable
-
Change the size of the buffer.
- setSkipOutputPath(JobConf, Path) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
-
Set the directory to which skipped records are written.
- setSocketSendBufSize(int) - Method in class org.apache.hadoop.ipc.Server
-
Sets the socket buffer size used for responding to RPCs
- setSortComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapreduce.Job
-
Define the comparator that controls how the keys are sorted before they
are passed to the
Reducer
.
- setSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Turn speculative execution on or off for this job.
- setState(ParseState) - Static method in class org.apache.hadoop.contrib.failmon.PersistentState
-
Set the state of parsing for a particular log file.
- setState(int) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Set the state for this job.
- setStatus(String) - Method in interface org.apache.hadoop.mapred.Reporter
-
Set the status description for the task.
- setStatus(String) - Method in class org.apache.hadoop.mapreduce.StatusReporter
-
- setStatus(String) - Method in class org.apache.hadoop.mapreduce.TaskAttemptContext
-
Set the current status of the task to the given string.
- setStatus(String) - Method in class org.apache.hadoop.mapreduce.TaskInputOutputContext
-
- setStatus(String) - Method in class org.apache.hadoop.util.Progress
-
- setStrings(String, String...) - Method in class org.apache.hadoop.conf.Configuration
-
Set the array of string values for the name
property as
as comma delimited values.
- setSuccessfulAttempt(TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskReport
-
set successful attempt ID of the task.
- setTabSize(int) - Method in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
- setTag(Text) - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
-
- setTag(String, String) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named tag to the specified value.
- setTag(String, int) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named tag to the specified value.
- setTag(String, long) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named tag to the specified value.
- setTag(String, short) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named tag to the specified value.
- setTag(String, byte) - Method in interface org.apache.hadoop.metrics.MetricsRecord
-
Sets the named tag to the specified value.
- setTag(String, String) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named tag to the specified value.
- setTag(String, int) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named tag to the specified value.
- setTag(String, long) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named tag to the specified value.
- setTag(String, short) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named tag to the specified value.
- setTag(String, byte) - Method in class org.apache.hadoop.metrics.spi.MetricsRecordImpl
-
Sets the named tag to the specified value.
- setTaskId(String) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
- setTaskID(TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Sets task id.
- setTaskId(String) - Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- setTaskOutputFilter(JobClient.TaskStatusFilter) - Method in class org.apache.hadoop.mapred.JobClient
-
Deprecated.
- setTaskOutputFilter(JobConf, JobClient.TaskStatusFilter) - Static method in class org.apache.hadoop.mapred.JobClient
-
Modify the JobConf to set the task output filter.
- setTaskRunTime(int) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Set the task completion time
- setTaskStatus(TaskCompletionEvent.Status) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Set task status.
- setTaskTrackerHttp(String) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
-
Set task tracker http location.
- setThreads(int, int) - Method in class org.apache.hadoop.http.HttpServer
-
Set the min, max number of worker threads (simultaneous connections).
- setTimes(Path, long, long) - Method in class org.apache.hadoop.fs.FileSystem
-
Set access time of a file
- setTopologyPaths(String[]) - Method in class org.apache.hadoop.fs.BlockLocation
-
Set the network topology paths of the hosts
- setTotalLogFileSize(long) - Method in class org.apache.hadoop.mapred.TaskLogAppender
-
- setUMask(Configuration, FsPermission) - Static method in class org.apache.hadoop.fs.permission.FsPermission
-
Set the user file creation mask (umask)
- setup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
-
Called once at the beginning of the task.
- setup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
-
Called once at the start of the task.
- setUpdate(Document, Term) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
-
Set the instance to be an update operation.
- setupJob(JobContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- setupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
For the framework to setup the job output during initialization
- setupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
This method implements the new interface by calling the old method.
- setupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
Create the temporary directory that is the root of all of the task
work directories.
- setupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
-
For the framework to setup the job output during initialization
- setupJobConf(int, int, long, int, long, int) - Method in class org.apache.hadoop.examples.SleepJob
-
- setupProgress() - Method in class org.apache.hadoop.mapred.JobStatus
-
- setupProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
-
Get the progress of the job's setup-tasks, as a float between 0.0
and 1.0.
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
-
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
Sets up output for the task.
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
-
Deprecated.
This method implements the new interface by calling the old method.
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
-
No task setup required.
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
-
Sets up output for the task.
- setUseNewMapper(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set whether the framework should use the new api for the mapper.
- setUseNewReducer(boolean) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set whether the framework should use the new api for the reducer.
- setUser(String) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the reported username for this job.
- setVerbose(boolean) - Method in class org.apache.hadoop.streaming.JarBuilder
-
- setVerifyChecksum(boolean) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
Set whether to verify checksum.
- setVerifyChecksum(boolean) - Method in class org.apache.hadoop.fs.FileSystem
-
Set the verify checksum flag.
- setVerifyChecksum(boolean) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Set the verify checksum flag.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Set the current working directory for the given file system.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Set the current working directory for the given file system.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.ftp.FTPFileSystem
-
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
Set the working directory to the given directory.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.s3.S3FileSystem
-
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.fs.s3native.NativeS3FileSystem
-
Set the working directory to the given directory.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.mapred.JobConf
-
Deprecated.
Set the current working directory for the default file system.
- setWorkingDirectory(Path) - Method in class org.apache.hadoop.mapreduce.Job
-
Set the current working directory for the default file system.
- setWorkingDirectory(File) - Method in class org.apache.hadoop.util.Shell
-
set the working directory
- Shard - Class in org.apache.hadoop.contrib.index.mapred
-
This class represents the metadata of a shard.
- Shard() - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
-
Constructor.
- Shard(long, String, long) - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
-
Construct a shard from a versio number, a directory and a generation
number.
- Shard(Shard) - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
-
Construct using a shard object.
- ShardWriter - Class in org.apache.hadoop.contrib.index.lucene
-
The initial version of an index is stored in the perm dir.
- ShardWriter(FileSystem, Shard, String, IndexUpdateConfiguration) - Constructor for class org.apache.hadoop.contrib.index.lucene.ShardWriter
-
Constructor
- Shell - Class in org.apache.hadoop.util
-
A base class for running a Unix command.
- Shell() - Constructor for class org.apache.hadoop.util.Shell
-
- Shell(long) - Constructor for class org.apache.hadoop.util.Shell
-
- Shell.ExitCodeException - Exception in org.apache.hadoop.util
-
This is an IOException with exit code added.
- Shell.ShellCommandExecutor - Class in org.apache.hadoop.util
-
A simple shell command executor.
- ShellCommandExecutor(String[]) - Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- ShellCommandExecutor(String[], File) - Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- ShellCommandExecutor(String[], File, Map<String, String>) - Constructor for class org.apache.hadoop.util.Shell.ShellCommandExecutor
-
- ShellParser - Class in org.apache.hadoop.contrib.failmon
-
Objects of this class parse the output of system command-line
utilities that can give information about the state of
various hardware components in the system.
- ShellParser() - Constructor for class org.apache.hadoop.contrib.failmon.ShellParser
-
- shippedCanonFiles_ - Variable in class org.apache.hadoop.streaming.StreamJob
-
- shouldPreserveInput() - Method in class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor
-
- shouldRetry(Exception, int) - Method in interface org.apache.hadoop.io.retry.RetryPolicy
-
Determines whether the framework should retry a
method for the given exception, and the number
of retries that have been made for that operation
so far.
- shuffleError(TaskAttemptID, String) - Method in class org.apache.hadoop.mapred.TaskTracker
-
A reduce-task failed to shuffle the map-outputs.
- shutdown() - Method in class org.apache.hadoop.fs.DU
-
Shut down the refreshing thread.
- shutdown() - Method in class org.apache.hadoop.ipc.metrics.RpcActivityMBean
-
- shutdown() - Method in class org.apache.hadoop.ipc.metrics.RpcMetrics
-
- shutdown() - Method in class org.apache.hadoop.mapred.TaskTracker
-
- SimpleCharStream - Class in org.apache.hadoop.record.compiler.generated
-
An implementation of interface CharStream, where the stream is assumed to
contain only ASCII characters (without unicode processing).
- SimpleCharStream(Reader, int, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(Reader, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(Reader) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream, String, int, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream, int, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream, String, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream, int, int) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream, String) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- SimpleCharStream(InputStream) - Constructor for class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Constructor.
- simpleHostname(String) - Static method in class org.apache.hadoop.util.StringUtils
-
Given a full hostname, return the word upto the first dot.
- size() - Method in class org.apache.hadoop.conf.Configuration
-
Return the number of keys in the configuration.
- size() - Method in class org.apache.hadoop.io.file.tfile.ByteArray
-
- size() - Method in interface org.apache.hadoop.io.file.tfile.RawComparable
-
Get the size of the byte range in the byte array.
- size() - Static method in class org.apache.hadoop.io.file.tfile.Utils.Version
-
Get the size of the serialized Version object.
- size() - Method in class org.apache.hadoop.io.MapWritable
- size() - Method in class org.apache.hadoop.io.SortedMapWritable
- size() - Method in class org.apache.hadoop.mapred.Counters.Group
-
Deprecated.
Returns the number of counters in this group.
- size() - Method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Returns the total number of counters, by summing the number of counters
in each group.
- size() - Method in class org.apache.hadoop.mapred.join.TupleWritable
-
The number of children in this Tuple.
- size() - Method in class org.apache.hadoop.mapreduce.CounterGroup
-
Returns the number of counters in this group.
- size() - Method in class org.apache.hadoop.metrics.util.MetricsRegistry
-
- size() - Method in class org.apache.hadoop.util.PriorityQueue
-
Returns the number of elements currently stored in the PriorityQueue.
- SIZE_OF_INTEGER - Static variable in class org.apache.hadoop.util.DataChecksum
-
- skip(long) - Method in class org.apache.hadoop.fs.BufferedFSInputStream
-
- skip(long) - Method in class org.apache.hadoop.fs.FSInputChecker
-
Skips over and discards n
bytes of data from the
input stream.
- skip(long) - Method in class org.apache.hadoop.io.compress.DecompressorStream
-
- skip(long) - Method in class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream
-
- skip(DataInput) - Static method in class org.apache.hadoop.io.Text
-
Skips over one Text in the input.
- skip(DataInput) - Static method in class org.apache.hadoop.io.UTF8
-
Deprecated.
Skips over one UTF8 in the input.
- skip(K) - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
-
Skip key-value pairs with keys less than or equal to the key provided.
- skip(K) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
-
Pass skip key to child RRs.
- skip(K) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
-
Skip key-value pairs with keys less than or equal to the key provided.
- skip(RecordInput, String, TypeID) - Static method in class org.apache.hadoop.record.meta.Utils
-
read/skip bytes from stream based on a type
- SkipBadRecords - Class in org.apache.hadoop.mapred
-
Utility class for skip bad records functionality.
- SkipBadRecords() - Constructor for class org.apache.hadoop.mapred.SkipBadRecords
-
- skipCompressedByteArray(DataInput) - Static method in class org.apache.hadoop.io.WritableUtils
-
- skipFully(InputStream, long) - Static method in class org.apache.hadoop.io.IOUtils
-
Similar to readFully().
- skipFully(DataInput, int) - Static method in class org.apache.hadoop.io.WritableUtils
-
Skip len number of bytes in input streamin
- SleepInputFormat() - Constructor for class org.apache.hadoop.examples.SleepJob.SleepInputFormat
-
- SleepJob - Class in org.apache.hadoop.examples
-
Dummy class for testing MR framefork.
- SleepJob() - Constructor for class org.apache.hadoop.examples.SleepJob
-
- SleepJob.EmptySplit - Class in org.apache.hadoop.examples
-
- SleepJob.SleepInputFormat - Class in org.apache.hadoop.examples
-
- SMALL_THRESH - Static variable in class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream
-
This constant is accessible by subclasses for historical purposes.
- SMARTParser - Class in org.apache.hadoop.contrib.failmon
-
Objects of this class parse the output of smartmontools to
gather information about the state of disks in the system.
- SMARTParser() - Constructor for class org.apache.hadoop.contrib.failmon.SMARTParser
-
Constructs a SMARTParser and reads the list of disk
devices to query
- SocketInputStream - Class in org.apache.hadoop.net
-
This implements an input stream that can have a timeout while reading.
- SocketInputStream(ReadableByteChannel, long) - Constructor for class org.apache.hadoop.net.SocketInputStream
-
Create a new input stream with the given timeout.
- SocketInputStream(Socket, long) - Constructor for class org.apache.hadoop.net.SocketInputStream
-
Same as SocketInputStream(socket.getChannel(), timeout):
Create a new input stream with the given timeout.
- SocketInputStream(Socket) - Constructor for class org.apache.hadoop.net.SocketInputStream
-
Same as SocketInputStream(socket.getChannel(), socket.getSoTimeout())
:
Create a new input stream with the given timeout.
- SocketOutputStream - Class in org.apache.hadoop.net
-
This implements an output stream that can have a timeout while writing.
- SocketOutputStream(WritableByteChannel, long) - Constructor for class org.apache.hadoop.net.SocketOutputStream
-
Create a new ouput stream with the given timeout.
- SocketOutputStream(Socket, long) - Constructor for class org.apache.hadoop.net.SocketOutputStream
-
Same as SocketOutputStream(socket.getChannel(), timeout):
Create a new ouput stream with the given timeout.
- SocksSocketFactory - Class in org.apache.hadoop.net
-
Specialized SocketFactory to create sockets with a SOCKS proxy
- SocksSocketFactory() - Constructor for class org.apache.hadoop.net.SocksSocketFactory
-
Default empty constructor (for use with the reflection API).
- SocksSocketFactory(Proxy) - Constructor for class org.apache.hadoop.net.SocksSocketFactory
-
Constructor with a supplied Proxy
- solution(List<List<ColumnName>>) - Method in interface org.apache.hadoop.examples.dancing.DancingLinks.SolutionAcceptor
-
A callback to return a solution to the application.
- solve(int[], DancingLinks.SolutionAcceptor<ColumnName>) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
-
Given a prefix, find solutions under it.
- solve(DancingLinks.SolutionAcceptor<ColumnName>) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
-
Solve a complete problem
- solve(int[]) - Method in class org.apache.hadoop.examples.dancing.Pentomino
-
Find all of the solutions that start with the given prefix.
- solve() - Method in class org.apache.hadoop.examples.dancing.Pentomino
-
Find all of the solutions to the puzzle.
- solve() - Method in class org.apache.hadoop.examples.dancing.Sudoku
-
- Sort<K,V> - Class in org.apache.hadoop.examples
-
This is the trivial map/reduce program that does absolutely nothing
other than use the framework to fragment and sort the input values.
- Sort() - Constructor for class org.apache.hadoop.examples.Sort
-
- sort(Path[], Path, boolean) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Perform a file sort from a set of input files into an output file.
- sort(Path, Path) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
The backwards compatible interface to sort.
- sort(IndexedSortable, int, int) - Method in class org.apache.hadoop.util.HeapSort
-
Sort the given range of items using heap sort.
- sort(IndexedSortable, int, int, Progressable) - Method in class org.apache.hadoop.util.HeapSort
-
- sort(IndexedSortable, int, int) - Method in interface org.apache.hadoop.util.IndexedSorter
-
Sort the items accessed through the given IndexedSortable over the given
range of logical indices.
- sort(IndexedSortable, int, int, Progressable) - Method in interface org.apache.hadoop.util.IndexedSorter
-
- sort(IndexedSortable, int, int) - Method in class org.apache.hadoop.util.QuickSort
-
Sort the given range of items using quick sort.
- sort(IndexedSortable, int, int, Progressable) - Method in class org.apache.hadoop.util.QuickSort
-
- sortAndIterate(Path[], Path, boolean) - Method in class org.apache.hadoop.io.SequenceFile.Sorter
-
Perform a file sort from a set of input files and return an iterator.
- SortedMapWritable - Class in org.apache.hadoop.io
-
A Writable SortedMap.
- SortedMapWritable() - Constructor for class org.apache.hadoop.io.SortedMapWritable
-
default constructor.
- SortedMapWritable(SortedMapWritable) - Constructor for class org.apache.hadoop.io.SortedMapWritable
-
Copy constructor.
- Sorter(FileSystem, Class<? extends WritableComparable>, Class, Configuration) - Constructor for class org.apache.hadoop.io.SequenceFile.Sorter
-
Sort and merge files containing the named classes.
- Sorter(FileSystem, RawComparator, Class, Class, Configuration) - Constructor for class org.apache.hadoop.io.SequenceFile.Sorter
-
- SortGenMapper() - Constructor for class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
-
- SOURCE_TAGS_FIELD - Static variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
-
- specialToken - Variable in class org.apache.hadoop.record.compiler.generated.Token
-
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
- split(int) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
-
Generate a list of row choices to cover the first moves.
- split - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
-
- split(String) - Static method in class org.apache.hadoop.util.StringUtils
-
Split a string using the default separator
- split(String, char, char) - Static method in class org.apache.hadoop.util.StringUtils
-
Split a string using the given separator
- splitKeyVal(byte[], int, int, Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
-
split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], int, int, Text, Text, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
-
split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
-
split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], Text, Text, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
-
split a UTF-8 byte array into key and value
assuming that the delimilator is at splitpos.
- splitKeyVal(byte[], int, int, Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
-
- splitKeyVal(byte[], int, int, Text, Text, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
-
- splitKeyVal(byte[], Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
-
- splitKeyVal(byte[], Text, Text, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
-
- SplitSampler(int) - Constructor for class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
-
Create a SplitSampler sampling all splits.
- SplitSampler(int, int) - Constructor for class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler
-
Create a new SplitSampler.
- StackServlet() - Constructor for class org.apache.hadoop.http.HttpServer.StackServlet
-
- StandardSocketFactory - Class in org.apache.hadoop.net
-
Specialized SocketFactory to create sockets with a SOCKS proxy
- StandardSocketFactory() - Constructor for class org.apache.hadoop.net.StandardSocketFactory
-
Default empty constructor (for use with the reflection API).
- start() - Method in class org.apache.hadoop.fs.DU
-
Start the disk usage checking thread.
- start() - Method in class org.apache.hadoop.http.HttpServer
-
Start the server.
- start() - Method in class org.apache.hadoop.ipc.Server
-
Starts the service.
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.ChecksumFileSystem
-
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.FileSystem
-
Returns a local File that the user can write output to.
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.FilterFileSystem
-
Returns a local File that the user can write output to.
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.HarFileSystem
-
not implemented.
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.kfs.KosmosFileSystem
-
- startLocalOutput(Path, Path) - Method in class org.apache.hadoop.fs.RawLocalFileSystem
-
- startMap(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startMap(TreeMap, String) - Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startMap(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- startMap(TreeMap, String) - Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startMap(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Check the mark for start of the serialized map.
- startMap(TreeMap, String) - Method in interface org.apache.hadoop.record.RecordOutput
-
Mark the start of a map to be serialized.
- startMap(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- startMap(TreeMap, String) - Method in class org.apache.hadoop.record.XmlRecordOutput
-
- startMonitoring() - Method in class org.apache.hadoop.metrics.file.FileContext
-
Starts or restarts monitoring, by opening in append-mode, the
file specified by the fileName
attribute,
if specified.
- startMonitoring() - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Starts or restarts monitoring, the emitting of metrics records as they are
updated.
- startMonitoring() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Starts or restarts monitoring, the emitting of metrics records.
- startMonitoring() - Method in class org.apache.hadoop.metrics.spi.CompositeContext
-
- startMonitoring() - Method in class org.apache.hadoop.metrics.spi.NullContext
-
Do-nothing version of startMonitoring
- startNextPhase() - Method in class org.apache.hadoop.util.Progress
-
Called during execution to move to the next phase at this level in the
tree.
- startNotifier() - Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- startRecord(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startRecord(Record, String) - Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startRecord(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- startRecord(Record, String) - Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startRecord(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Check the mark for start of the serialized record.
- startRecord(Record, String) - Method in interface org.apache.hadoop.record.RecordOutput
-
Mark the start of a record to be serialized.
- startRecord(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- startRecord(Record, String) - Method in class org.apache.hadoop.record.XmlRecordOutput
-
- startTracker(JobConf) - Static method in class org.apache.hadoop.mapred.JobTracker
-
Start the JobTracker with given configuration.
- startTracker(JobConf, String) - Static method in class org.apache.hadoop.mapred.JobTracker
-
- startupShutdownMessage(Class<?>, String[], Log) - Static method in class org.apache.hadoop.util.StringUtils
-
Print a log message for starting up and shutting down
- startVector(String) - Method in class org.apache.hadoop.record.BinaryRecordInput
-
- startVector(ArrayList, String) - Method in class org.apache.hadoop.record.BinaryRecordOutput
-
- startVector(String) - Method in class org.apache.hadoop.record.CsvRecordInput
-
- startVector(ArrayList, String) - Method in class org.apache.hadoop.record.CsvRecordOutput
-
- startVector(String) - Method in interface org.apache.hadoop.record.RecordInput
-
Check the mark for start of the serialized vector.
- startVector(ArrayList, String) - Method in interface org.apache.hadoop.record.RecordOutput
-
Mark the start of a vector to be serialized.
- startVector(String) - Method in class org.apache.hadoop.record.XmlRecordInput
-
- startVector(ArrayList, String) - Method in class org.apache.hadoop.record.XmlRecordOutput
-
- stat2Paths(FileStatus[]) - Static method in class org.apache.hadoop.fs.FileUtil
-
convert an array of FileStatus to an array of Path
- stat2Paths(FileStatus[], Path) - Static method in class org.apache.hadoop.fs.FileUtil
-
convert an array of FileStatus to an array of Path.
- staticFlag - Static variable in class org.apache.hadoop.record.compiler.generated.SimpleCharStream
-
Whether parser is static.
- statistics - Variable in class org.apache.hadoop.fs.FileSystem
-
The statistics for this file system.
- Statistics(String) - Constructor for class org.apache.hadoop.fs.FileSystem.Statistics
-
- StatusReporter - Class in org.apache.hadoop.mapreduce
-
- StatusReporter() - Constructor for class org.apache.hadoop.mapreduce.StatusReporter
-
- statusUpdate(TaskAttemptID, TaskStatus) - Method in class org.apache.hadoop.mapred.TaskTracker
-
Called periodically to report Task progress, from 0.0 to 1.0.
- stop() - Method in class org.apache.hadoop.http.HttpServer
-
stop the server
- stop() - Method in class org.apache.hadoop.ipc.Client
-
Stop all threads related to this client.
- stop() - Method in class org.apache.hadoop.ipc.Server
-
Stops the service.
- stop() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
set the thread state to STOPPING so that the
thread will stop when it wakes up.
- stopMonitoring() - Method in class org.apache.hadoop.metrics.file.FileContext
-
Stops monitoring, closing the file.
- stopMonitoring() - Method in interface org.apache.hadoop.metrics.MetricsContext
-
Stops monitoring.
- stopMonitoring() - Method in class org.apache.hadoop.metrics.spi.AbstractMetricsContext
-
Stops monitoring.
- stopMonitoring() - Method in class org.apache.hadoop.metrics.spi.CompositeContext
-
- stopNotifier() - Static method in class org.apache.hadoop.mapred.JobEndNotifier
-
- stopProxy(VersionedProtocol) - Static method in class org.apache.hadoop.ipc.RPC
-
Stop this proxy and release its invoker's resource
- stopTracker() - Method in class org.apache.hadoop.mapred.JobTracker
-
- store(Configuration, K, String) - Static method in class org.apache.hadoop.io.DefaultStringifier
-
Stores the item in the configuration with the given keyName.
- storeArray(Configuration, K[], String) - Static method in class org.apache.hadoop.io.DefaultStringifier
-
Stores the array of items in the configuration with the given keyName.
- storeBlock(Block, File) - Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- storeINode(Path, INode) - Method in interface org.apache.hadoop.fs.s3.FileSystemStore
-
- StreamBackedIterator<X extends Writable> - Class in org.apache.hadoop.mapred.join
-
This class provides an implementation of ResetableIterator.
- StreamBackedIterator() - Constructor for class org.apache.hadoop.mapred.join.StreamBackedIterator
-
- StreamBaseRecordReader - Class in org.apache.hadoop.streaming
-
Shared functionality for hadoopStreaming formats.
- StreamBaseRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) - Constructor for class org.apache.hadoop.streaming.StreamBaseRecordReader
-
- StreamInputFormat - Class in org.apache.hadoop.streaming
-
An input format that selects a RecordReader based on a JobConf property.
- StreamInputFormat() - Constructor for class org.apache.hadoop.streaming.StreamInputFormat
-
- StreamJob - Class in org.apache.hadoop.streaming
-
All the client-side work happens here.
- StreamJob(String[], boolean) - Constructor for class org.apache.hadoop.streaming.StreamJob
-
- StreamJob() - Constructor for class org.apache.hadoop.streaming.StreamJob
-
- StreamKeyValUtil - Class in org.apache.hadoop.streaming
-
- StreamKeyValUtil() - Constructor for class org.apache.hadoop.streaming.StreamKeyValUtil
-
- StreamUtil - Class in org.apache.hadoop.streaming
-
Utilities not available elsewhere in Hadoop.
- StreamUtil() - Constructor for class org.apache.hadoop.streaming.StreamUtil
-
- StreamXmlRecordReader - Class in org.apache.hadoop.streaming
-
A way to interpret XML fragments as Mapper input records.
- StreamXmlRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) - Constructor for class org.apache.hadoop.streaming.StreamXmlRecordReader
-
- STRING - Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- string2long(String) - Static method in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
-
Convert a string to long.
- STRING_VALUE_MAX - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- STRING_VALUE_MIN - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
-
- Stringifier<T> - Interface in org.apache.hadoop.io
-
Stringifier interface offers two methods to convert an object
to a string representation and restore the object given its
string representation.
- stringifyException(Throwable) - Static method in class org.apache.hadoop.util.StringUtils
-
Make a string representation of the exception.
- stringifySolution(int, int, List<List<Pentomino.ColumnName>>) - Static method in class org.apache.hadoop.examples.dancing.Pentomino
-
Convert a solution to the puzzle returned by the model into a string
that represents the placement of the pieces onto the board.
- stringToPath(String[]) - Static method in class org.apache.hadoop.util.StringUtils
-
- stringToURI(String[]) - Static method in class org.apache.hadoop.util.StringUtils
-
- StringTypeID - Static variable in class org.apache.hadoop.record.meta.TypeID
-
- StringUtils - Class in org.apache.hadoop.util
-
General string utils
- StringUtils() - Constructor for class org.apache.hadoop.util.StringUtils
-
- StringUtils.TraditionalBinaryPrefix - Enum in org.apache.hadoop.util
-
The traditional binary prefixes, kilo, mega, ..., exa,
which can be represented by a 64-bit integer.
- StringValueMax - Class in org.apache.hadoop.mapred.lib.aggregate
-
This class implements a value aggregator that maintain the biggest of
a sequence of strings.
- StringValueMax() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
-
the default constructor
- StringValueMin - Class in org.apache.hadoop.mapred.lib.aggregate
-
This class implements a value aggregator that maintain the smallest of
a sequence of strings.
- StringValueMin() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
-
the default constructor
- StrToken(Parser.TType, String) - Constructor for class org.apache.hadoop.mapred.join.Parser.StrToken
-
- STRUCT - Static variable in class org.apache.hadoop.record.meta.TypeID.RIOType
-
- StructTypeID - Class in org.apache.hadoop.record.meta
-
Represents typeID for a struct
- StructTypeID(RecordTypeInfo) - Constructor for class org.apache.hadoop.record.meta.StructTypeID
-
Create a StructTypeID based on the RecordTypeInfo of some record
- subMap(WritableComparable, WritableComparable) - Method in class org.apache.hadoop.io.SortedMapWritable
- submit() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
-
Submit this job to mapred.
- submit() - Method in class org.apache.hadoop.mapreduce.Job
-
Submit the job to the cluster and return immediately.
- submitAndMonitorJob() - Method in class org.apache.hadoop.streaming.StreamJob
-
- submitJob(String) - Method in class org.apache.hadoop.mapred.JobClient
-
Submit a job to the MR system.
- submitJob(JobConf) - Method in class org.apache.hadoop.mapred.JobClient
-
Submit a job to the MR system.
- submitJob(JobID) - Method in class org.apache.hadoop.mapred.JobTracker
-
JobTracker.submitJob() kicks off a new job.
- submitJob(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
-
- submitJobInternal(JobConf) - Method in class org.apache.hadoop.mapred.JobClient
-
Internal method for submitting jobs to the system.
- Submitter - Class in org.apache.hadoop.mapred.pipes
-
The main entry point and job submitter.
- Submitter() - Constructor for class org.apache.hadoop.mapred.pipes.Submitter
-
- Submitter(Configuration) - Constructor for class org.apache.hadoop.mapred.pipes.Submitter
-
- SUCCEEDED - Static variable in class org.apache.hadoop.mapred.JobStatus
-
- SUCCESS - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
-
- Sudoku - Class in org.apache.hadoop.examples.dancing
-
This class uses the dancing links algorithm from Knuth to solve sudoku
puzzles.
- Sudoku(InputStream) - Constructor for class org.apache.hadoop.examples.dancing.Sudoku
-
Set up a puzzle board to the given size.
- Sudoku.ColumnName - Interface in org.apache.hadoop.examples.dancing
-
This interface is a marker class for the columns created for the
Sudoku solver.
- suffix(String) - Method in class org.apache.hadoop.fs.Path
-
Adds a suffix to the final name in the path.
- sum(Counters, Counters) - Static method in class org.apache.hadoop.mapred.Counters
-
Deprecated.
Convenience method for computing the sum of two sets of counters.
- suspend() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
-
suspend the running thread
- swap(int, int) - Method in interface org.apache.hadoop.util.IndexedSortable
-
Swap items at the given addresses.
- SwitchTo(int) - Method in class org.apache.hadoop.record.compiler.generated.RccTokenManager
-
Switch to specified lex state.
- SYMBOL - Variable in enum org.apache.hadoop.fs.permission.FsAction
-
Symbolic representation
- symbol - Variable in enum org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix
-
- symLink(String, String) - Static method in class org.apache.hadoop.fs.FileUtil
-
Create a soft link between a src and destination
only on a local disk.
- sync() - Method in class org.apache.hadoop.fs.FSDataOutputStream
-
Synchronize all buffer with the underlying devices.
- sync() - Method in interface org.apache.hadoop.fs.Syncable
-
Synchronize all buffer with the underlying devices.
- sync(long) - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Seek to the next sync mark past a given position.
- sync() - Method in class org.apache.hadoop.io.SequenceFile.Writer
-
create a sync point
- SYNC_INTERVAL - Static variable in class org.apache.hadoop.io.SequenceFile
-
The number of bytes between sync points.
- Syncable - Interface in org.apache.hadoop.fs
-
This interface declare the sync() operation.
- syncLogs(TaskAttemptID, TaskAttemptID) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- syncLogs(TaskAttemptID, TaskAttemptID, boolean) - Static method in class org.apache.hadoop.mapred.TaskLog
-
- syncSeen() - Method in class org.apache.hadoop.io.SequenceFile.Reader
-
Returns true iff the previous call to next passed a sync mark.
- SystemLogParser - Class in org.apache.hadoop.contrib.failmon
-
An object of this class parses a Unix system log file to create
appropriate EventRecords.
- SystemLogParser(String) - Constructor for class org.apache.hadoop.contrib.failmon.SystemLogParser
-
Create a new parser object .