Package | Description |
---|---|
org.apache.hadoop.examples |
Hadoop example code.
|
org.apache.hadoop.mapred |
A software framework for easily writing applications which process vast
amounts of data (multi-terabyte data-sets) parallelly on large clusters
(thousands of nodes) built of commodity hardware in a reliable, fault-tolerant
manner.
|
org.apache.hadoop.mapreduce | |
org.apache.hadoop.mapreduce.lib.db |
org.apache.hadoop.mapred.lib.db Package
|
org.apache.hadoop.mapreduce.lib.input | |
org.apache.hadoop.mapreduce.lib.map | |
org.apache.hadoop.mapreduce.lib.output | |
org.apache.hadoop.mapreduce.lib.reduce | |
org.apache.hadoop.mapreduce.task |
Modifier and Type | Method and Description |
---|---|
RecordReader<MultiFileWordCount.WordOffset,org.apache.hadoop.io.Text> |
MultiFileWordCount.MyInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
void |
MultiFileWordCount.CombineFileLineRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
Constructor and Description |
---|
MultiFileWordCount.CombineFileLineRecordReader(CombineFileSplit split,
TaskAttemptContext context,
java.lang.Integer index) |
Modifier and Type | Interface and Description |
---|---|
interface |
TaskAttemptContext
Deprecated.
Use
TaskAttemptContext
instead. |
Modifier and Type | Method and Description |
---|---|
void |
OutputCommitter.abortTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method.
|
void |
OutputCommitter.commitTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method.
|
boolean |
OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method.
|
void |
OutputCommitter.setupTask(TaskAttemptContext taskContext)
This method implements the new interface by calling the old method.
|
Modifier and Type | Interface and Description |
---|---|
interface |
MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context that is given to the
Mapper . |
interface |
ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context passed to the
Reducer . |
interface |
TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
A context object that allows input and output from the task.
|
Modifier and Type | Class and Description |
---|---|
class |
Mapper.Context
The
Context passed on to the Mapper implementations. |
class |
Reducer.Context
The
Context passed on to the Reducer implementations. |
Modifier and Type | Method and Description |
---|---|
abstract void |
OutputCommitter.abortTask(TaskAttemptContext taskContext)
Discard the task output
|
abstract void |
RecordWriter.close(TaskAttemptContext context)
Close this
RecordWriter to future operations. |
abstract void |
OutputCommitter.commitTask(TaskAttemptContext taskContext)
To promote the task's temporary output to final output location
The task's output is moved to the job's output directory.
|
abstract RecordReader<K,V> |
InputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for a given split.
|
abstract OutputCommitter |
OutputFormat.getOutputCommitter(TaskAttemptContext context)
Get the output committer for this output format.
|
abstract RecordWriter<K,V> |
OutputFormat.getRecordWriter(TaskAttemptContext context)
Get the
RecordWriter for the given task. |
abstract void |
RecordReader.initialize(InputSplit split,
TaskAttemptContext context)
Called once at initialization.
|
abstract boolean |
OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
Check whether task needs a commit
|
abstract void |
OutputCommitter.setupTask(TaskAttemptContext taskContext)
Sets up output for the task.
|
Modifier and Type | Method and Description |
---|---|
void |
DBOutputFormat.DBRecordWriter.close(TaskAttemptContext context)
Close this
RecordWriter to future operations. |
RecordReader<org.apache.hadoop.io.LongWritable,T> |
DBInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for a given split.
|
OutputCommitter |
DBOutputFormat.getOutputCommitter(TaskAttemptContext context) |
RecordWriter<K,V> |
DBOutputFormat.getRecordWriter(TaskAttemptContext context)
Get the
RecordWriter for the given task. |
void |
DBRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
Modifier and Type | Field and Description |
---|---|
protected TaskAttemptContext |
CombineFileRecordReader.context |
Modifier and Type | Method and Description |
---|---|
abstract RecordReader<K,V> |
CombineFileInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context)
This is not implemented yet.
|
RecordReader<K,V> |
DelegatingInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> |
KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context) |
RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> |
NLineInputFormat.createRecordReader(InputSplit genericSplit,
TaskAttemptContext context) |
RecordReader<org.apache.hadoop.io.BytesWritable,org.apache.hadoop.io.BytesWritable> |
SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
RecordReader<K,V> |
SequenceFileInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> |
SequenceFileAsTextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
RecordReader<K,V> |
SequenceFileInputFilter.createRecordReader(InputSplit split,
TaskAttemptContext context)
Create a record reader for the given split
|
RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> |
TextInputFormat.createRecordReader(InputSplit split,
TaskAttemptContext context) |
void |
CombineFileRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
void |
DelegatingRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
void |
KeyValueLineRecordReader.initialize(InputSplit genericSplit,
TaskAttemptContext context) |
void |
LineRecordReader.initialize(InputSplit genericSplit,
TaskAttemptContext context) |
void |
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
void |
SequenceFileAsTextRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
void |
SequenceFileRecordReader.initialize(InputSplit split,
TaskAttemptContext context) |
Constructor and Description |
---|
CombineFileRecordReader(CombineFileSplit split,
TaskAttemptContext context,
java.lang.Class<? extends RecordReader<K,V>> rrClass)
A generic RecordReader that can hand out different recordReaders
for each chunk in the CombineFileSplit.
|
DelegatingRecordReader(InputSplit split,
TaskAttemptContext context)
Constructs the DelegatingRecordReader.
|
Modifier and Type | Class and Description |
---|---|
class |
WrappedMapper.Context |
Modifier and Type | Method and Description |
---|---|
void |
FileOutputCommitter.abortTask(TaskAttemptContext context)
Delete the work directory
|
void |
FilterOutputFormat.FilterRecordWriter.close(TaskAttemptContext context) |
void |
TextOutputFormat.LineRecordWriter.close(TaskAttemptContext context) |
void |
FileOutputCommitter.commitTask(TaskAttemptContext context)
Move the files from the work directory to the job output directory
|
org.apache.hadoop.fs.Path |
FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context,
java.lang.String extension)
Get the default path and filename for the output format.
|
OutputCommitter |
FileOutputFormat.getOutputCommitter(TaskAttemptContext context) |
OutputCommitter |
FilterOutputFormat.getOutputCommitter(TaskAttemptContext context) |
OutputCommitter |
LazyOutputFormat.getOutputCommitter(TaskAttemptContext context) |
OutputCommitter |
NullOutputFormat.getOutputCommitter(TaskAttemptContext context) |
abstract RecordWriter<K,V> |
FileOutputFormat.getRecordWriter(TaskAttemptContext job) |
RecordWriter<K,V> |
FilterOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<K,V> |
LazyOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<org.apache.hadoop.io.WritableComparable<?>,org.apache.hadoop.io.Writable> |
MapFileOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<K,V> |
NullOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<org.apache.hadoop.io.BytesWritable,org.apache.hadoop.io.BytesWritable> |
SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<K,V> |
SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context) |
RecordWriter<K,V> |
TextOutputFormat.getRecordWriter(TaskAttemptContext job) |
protected org.apache.hadoop.io.SequenceFile.Writer |
SequenceFileAsBinaryOutputFormat.getSequenceWriter(TaskAttemptContext context,
java.lang.Class<?> keyClass,
java.lang.Class<?> valueClass) |
static java.lang.String |
FileOutputFormat.getUniqueFile(TaskAttemptContext context,
java.lang.String name,
java.lang.String extension)
Generate a unique filename, based on the task id, name, and extension
|
boolean |
FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)
Did this task write any files in the work directory?
|
void |
FileOutputCommitter.setupTask(TaskAttemptContext context)
No task setup required.
|
Constructor and Description |
---|
FileOutputCommitter(org.apache.hadoop.fs.Path outputPath,
TaskAttemptContext context)
Create a file output committer
|
Modifier and Type | Class and Description |
---|---|
class |
WrappedReducer.Context |
Modifier and Type | Class and Description |
---|---|
class |
MapContextImpl<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context that is given to the
Mapper . |
class |
ReduceContextImpl<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
The context passed to the
Reducer . |
class |
TaskAttemptContextImpl
The context for task attempts.
|
class |
TaskInputOutputContextImpl<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
A context object that allows input and output from the task.
|
Copyright © 2009 The Apache Software Foundation