org.apache.hadoop.http.HttpServer.addInternalServlet(String, String, Class<? extends HttpServlet>)
|
org.apache.hadoop.http.HttpServer.addSslListener(InetSocketAddress, String, String, String)
|
org.apache.hadoop.fs.FsShell.byteDesc(long)
|
org.apache.hadoop.ipc.RPC.call(Method, Object[][], InetSocketAddress[], Configuration)
|
org.apache.hadoop.ipc.Client.call(Writable[], InetSocketAddress[])
|
org.apache.hadoop.ipc.Client.call(Writable, InetSocketAddress)
|
org.apache.hadoop.ipc.Client.call(Writable, InetSocketAddress, UserGroupInformation)
|
org.apache.hadoop.ipc.Server.call(Writable, long)
|
org.apache.hadoop.io.WritableUtils.cloneInto(Writable, Writable)
|
org.apache.hadoop.util.ReflectionUtils.cloneWritableInto(Writable, Writable) |
org.apache.hadoop.io.file.tfile.TFile.Reader.createScanner(byte[], byte[])
|
org.apache.hadoop.io.file.tfile.TFile.Reader.createScanner(RawComparable, RawComparable)
|
org.apache.hadoop.fs.FilterFileSystem.delete(Path) |
org.apache.hadoop.fs.RawLocalFileSystem.delete(Path) |
org.apache.hadoop.fs.FileSystem.delete(Path)
|
org.apache.hadoop.fs.ftp.FTPFileSystem.delete(Path)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.delete(Path) |
org.apache.hadoop.fs.s3.S3FileSystem.delete(Path) |
org.apache.hadoop.fs.s3native.NativeS3FileSystem.delete(Path) |
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findByte(byte[], int, int, byte)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findBytes(byte[], int, int, byte[])
|
org.apache.hadoop.mapred.Counters.findCounter(String, int, String) |
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findNthByte(byte[], byte, int)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findNthByte(byte[], int, int, byte, int)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findTab(byte[])
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findTab(byte[], int, int)
|
org.apache.hadoop.fs.FileUtil.fullyDelete(FileSystem, Path)
|
org.apache.hadoop.io.BytesWritable.get()
|
org.apache.hadoop.fs.FileSystem.getBlockSize(Path)
|
org.apache.hadoop.streaming.StreamJob.getClusterNick() |
org.apache.hadoop.mapred.JobTracker.getClusterStatus()
|
org.apache.hadoop.record.compiler.generated.SimpleCharStream.getColumn() |
org.apache.hadoop.io.SequenceFile.getCompressionType(Configuration)
|
org.apache.hadoop.mapred.Counters.Group.getCounter(int, String)
|
org.apache.hadoop.mapred.JobClient.getJob(String)
|
org.apache.hadoop.mapred.JobProfile.getJobId()
|
org.apache.hadoop.mapred.JobStatus.getJobId()
|
org.apache.hadoop.mapred.RunningJob.getJobID()
|
org.apache.hadoop.mapred.JobID.getJobIDsPattern(String, Integer) |
org.apache.hadoop.fs.FileSystem.getLength(Path)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.getLength(Path) |
org.apache.hadoop.record.compiler.generated.SimpleCharStream.getLine() |
org.apache.hadoop.mapred.jobcontrol.Job.getMapredJobID()
|
org.apache.hadoop.mapred.JobClient.getMapTaskReports(String)
|
org.apache.hadoop.mapred.JobConf.getMaxPhysicalMemoryForTask()
|
org.apache.hadoop.mapred.JobConf.getMaxVirtualMemoryForTask()
|
org.apache.hadoop.fs.FilterFileSystem.getName()
|
org.apache.hadoop.fs.FileSystem.getName()
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.getName() |
org.apache.hadoop.fs.FileSystem.getNamed(String, Configuration)
|
org.apache.hadoop.mapred.JobClient.getReduceTaskReports(String)
|
org.apache.hadoop.fs.FileSystem.getReplication(Path)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.getReplication(Path) |
org.apache.hadoop.net.NetUtils.getServerAddress(Configuration, String, String, String) |
org.apache.hadoop.io.BytesWritable.getSize()
|
org.apache.hadoop.fs.FileSystem.getStatistics()
|
org.apache.hadoop.mapred.TaskAttemptID.getTaskAttemptIDsPattern(String, Integer, Boolean, Integer, Integer) |
org.apache.hadoop.mapred.TaskCompletionEvent.getTaskId()
|
org.apache.hadoop.mapred.TaskReport.getTaskId()
|
org.apache.hadoop.mapred.TaskID.getTaskIDsPattern(String, Integer, Boolean, Integer) |
org.apache.hadoop.mapred.JobClient.getTaskOutputFilter() |
org.apache.hadoop.streaming.StreamJob.go()
|
org.apache.hadoop.fs.FileSystem.isDirectory(Path)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.isDirectory(Path) |
org.apache.hadoop.fs.kfs.KosmosFileSystem.isFile(Path) |
org.apache.hadoop.mapred.RunningJob.killTask(String, boolean)
|
org.apache.hadoop.fs.FsShell.limitDecimalTo2(double)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.lock(Path, boolean) |
org.apache.hadoop.mapred.JobHistory.MapAttempt.logFailed(TaskAttemptID, long, String, String)
|
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logFailed(TaskAttemptID, long, String, String)
|
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logFinished(TaskAttemptID, long, long, long, String)
|
org.apache.hadoop.mapred.JobHistory.MapAttempt.logFinished(TaskAttemptID, long, String)
|
org.apache.hadoop.mapred.JobHistory.JobInfo.logJobInfo(JobID, long, long, int)
|
org.apache.hadoop.mapred.JobHistory.MapAttempt.logKilled(TaskAttemptID, long, String, String)
|
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logKilled(TaskAttemptID, long, String, String)
|
org.apache.hadoop.mapred.JobHistory.JobInfo.logStarted(JobID, long, int, int)
|
org.apache.hadoop.mapred.JobHistory.MapAttempt.logStarted(TaskAttemptID, long, String)
|
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logStarted(TaskAttemptID, long, String)
|
org.apache.hadoop.mapred.JobHistory.JobInfo.logSubmitted(JobID, JobConf, String, long)
|
org.apache.hadoop.io.SequenceFile.Reader.next(DataOutputBuffer)
|
org.apache.hadoop.mapred.TaskAttemptID.read(DataInput) |
org.apache.hadoop.mapred.TaskID.read(DataInput) |
org.apache.hadoop.mapred.JobID.read(DataInput) |
org.apache.hadoop.streaming.UTF8ByteArrayUtils.readLine(LineReader, Text)
|
org.apache.hadoop.fs.kfs.KosmosFileSystem.release(Path) |
org.apache.hadoop.io.SequenceFile.setCompressionType(Configuration, SequenceFile.CompressionType)
|
org.apache.hadoop.security.UserGroupInformation.setCurrentUGI(UserGroupInformation)
|
org.apache.hadoop.mapreduce.Counter.setDisplayName(String) |
org.apache.hadoop.mapred.jobcontrol.Job.setMapredJobID(String)
|
org.apache.hadoop.mapred.JobConf.setMaxPhysicalMemoryForTask(long) |
org.apache.hadoop.mapred.JobConf.setMaxVirtualMemoryForTask(long)
|
org.apache.hadoop.mapred.TaskCompletionEvent.setTaskId(String)
|
org.apache.hadoop.mapred.JobClient.setTaskOutputFilter(JobClient.TaskStatusFilter) |
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], int, int, Text, Text, int)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], int, int, Text, Text, int, int)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], Text, Text, int)
|
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], Text, Text, int, int)
|
org.apache.hadoop.mapred.pipes.Submitter.submitJob(JobConf)
|