Uses of Package
org.apache.hadoop.mapreduce

Packages that use org.apache.hadoop.mapreduce
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.output   
org.apache.hadoop.mapreduce.lib.partition   
org.apache.hadoop.mapreduce.lib.reduce   
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.examples
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
           
Partitioner
          Partitions the key space.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
           
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapred
Counter
          A named counter that tracks the progress of a map/reduce job.
ID
          A general identifier, which internally stores the id as an integer.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskID
          TaskID represents the immutable and unique identifier for a Map or Reduce Task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce
Counter
          A named counter that tracks the progress of a map/reduce job.
CounterGroup
          A group of Counters that logically belong together.
Counters
           
ID
          A general identifier, which internally stores the id as an integer.
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job.JobState
           
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
MapContext
          The context that is given to the Mapper.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
           
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
ReduceContext
          The context passed to the Reducer.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
           
StatusReporter
           
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskID
          TaskID represents the immutable and unique identifier for a Map or Reduce Task.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.input
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
TaskAttemptContext
          The context for task attempts.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.map
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
           
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.output
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
TaskAttemptContext
          The context for task attempts.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.partition
Partitioner
          Partitions the key space.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.reduce
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
           
 



Copyright © 2009 The Apache Software Foundation