程序包 | 说明 |
---|---|
org.apache.hadoop.examples |
Hadoop example code.
|
org.apache.hadoop.examples.dancing |
This package is a distributed implementation of Knuth's dancing links
algorithm that can run under Hadoop.
|
org.apache.hadoop.examples.terasort |
This package consists of 3 map/reduce applications for Hadoop to
compete in the annual terabyte sort
competition.
|
org.apache.hadoop.fs |
An abstract file system API.
|
org.apache.hadoop.fs.s3 |
A distributed, block-based implementation of
FileSystem that uses Amazon S3
as a backing store. |
org.apache.hadoop.hdfs.server.balancer | |
org.apache.hadoop.hdfs.tools | |
org.apache.hadoop.mapred |
A software framework for easily writing applications which process vast
amounts of data (multi-terabyte data-sets) parallelly on large clusters
(thousands of nodes) built of commodity hardware in a reliable, fault-tolerant
manner.
|
org.apache.hadoop.mapred.lib |
Library of generally useful mappers, reducers, and partitioners.
|
org.apache.hadoop.mapred.pipes |
Hadoop Pipes allows C++ code to use Hadoop DFS and map/reduce.
|
org.apache.hadoop.mapred.tools | |
org.apache.hadoop.streaming |
Hadoop Streaming is a utility which allows users to create and run
Map-Reduce jobs with any executables (e.g.
|
org.apache.hadoop.tools | |
org.apache.hadoop.tools.rumen |
Rumen is a data extraction and analysis tool built for
Apache Hadoop.
|
org.apache.hadoop.util |
Common utilities.
|
限定符和类型 | 类和说明 |
---|---|
class |
DBCountPageView
This is a demonstrative program, which uses DBInputFormat for reading
the input data from a database, and DBOutputFormat for writing the data
to the database.
|
class |
Grep |
class |
Join
This is the trivial map/reduce program that does absolutely nothing
other than use the framework to fragment and sort the input values.
|
class |
MultiFileWordCount
MultiFileWordCount is an example to demonstrate the usage of
MultiFileInputFormat.
|
class |
PiEstimator
A Map-reduce program to estimate the value of Pi
using quasi-Monte Carlo method.
|
class |
RandomTextWriter
This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task writes a large unsorted
random sequence of words.
|
class |
RandomWriter
This program uses map/reduce to just run a distributed job where there is
no interaction between the tasks and each task write a large unsorted
random binary sequence file of BytesWritable.
|
class |
SleepJob
Dummy class for testing MR framefork.
|
class |
Sort<K,V>
This is the trivial map/reduce program that does absolutely nothing
other than use the framework to fragment and sort the input values.
|
限定符和类型 | 类和说明 |
---|---|
class |
DistributedPentomino
Launch a distributed pentomino solver.
|
限定符和类型 | 类和说明 |
---|---|
class |
TeraGen
Generate the official terasort input data set.
|
class |
TeraSort
Generates the sampled split points, launches the job, and waits for it to
finish.
|
class |
TeraValidate
Generate 1 mapper per a file that checks to make sure the keys
are sorted within each file.
|
限定符和类型 | 类和说明 |
---|---|
class |
FsShell
Provide command line access to a FileSystem.
|
限定符和类型 | 类和说明 |
---|---|
class |
MigrationTool
This class is a tool for migrating data from an older to a newer version
of an S3 filesystem.
|
限定符和类型 | 类和说明 |
---|---|
class |
Balancer
The balancer is a tool that balances disk space usage on an HDFS cluster
when some datanodes become full or when new empty nodes join the cluster.
|
限定符和类型 | 类和说明 |
---|---|
class |
DFSAdmin
This class provides some DFS administrative access.
|
class |
DFSck
This class provides rudimentary checking of DFS volumes for errors and
sub-optimal conditions.
|
限定符和类型 | 类和说明 |
---|---|
class |
JobClient
JobClient is the primary interface for the user-job to interact
with the JobTracker . |
限定符和类型 | 类和说明 |
---|---|
class |
InputSampler<K,V>
Utility for collecting samples and writing a partition file for
TotalOrderPartitioner . |
限定符和类型 | 类和说明 |
---|---|
class |
Submitter
The main entry point and job submitter.
|
限定符和类型 | 类和说明 |
---|---|
class |
MRAdmin
Administrative access to Hadoop Map-Reduce.
|
限定符和类型 | 类和说明 |
---|---|
class |
DumpTypedBytes
Utility program that fetches all files that match a given pattern and dumps
their content to stdout as typed bytes.
|
class |
LoadTypedBytes
Utility program that reads typed bytes from standard input and stores them in
a sequence file for which the path is given as an argument.
|
class |
StreamJob
All the client-side work happens here.
|
限定符和类型 | 类和说明 |
---|---|
class |
DistCh
A Map-reduce program to recursively change files properties
such as owner, group and permission.
|
class |
DistCp
A Map-reduce program to recursively copy directories between
different file-systems.
|
class |
HadoopArchives
a archive creation utility.
|
限定符和类型 | 类和说明 |
---|---|
class |
HadoopLogsAnalyzer
已过时。
|
class |
TraceBuilder
The main driver of the Rumen Parser.
|
限定符和类型 | 方法和说明 |
---|---|
static int |
ToolRunner.run(Configuration conf,
Tool tool,
String[] args)
Runs the given
Tool by run(String[]) , after
parsing with the given generic arguments. |
static int |
ToolRunner.run(Tool tool,
String[] args)
Runs the
Tool with its Configuration . |
Copyright © 2009 The Apache Software Foundation