public class WholeTextFileInputFormat
extends org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String>
implements org.apache.hadoop.conf.Configurable
CombineFileInputFormat for
reading whole text files. Each file is read as key-value pair, where the key is the file path and
the value is the entire content of file.| Constructor and Description |
|---|
WholeTextFileInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.mapreduce.RecordReader<String,String> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
org.apache.hadoop.conf.Configuration |
getConf() |
void |
setConf(org.apache.hadoop.conf.Configuration c) |
void |
setMinPartitions(org.apache.hadoop.mapreduce.JobContext context,
int minPartitions)
Allow minPartitions set by end-user in order to keep compatibility with old Hadoop API,
which is set through setMaxSplitSize
|
public void setConf(org.apache.hadoop.conf.Configuration c)
setConf in interface org.apache.hadoop.conf.Configurablepublic org.apache.hadoop.conf.Configuration getConf()
getConf in interface org.apache.hadoop.conf.Configurablepublic org.apache.hadoop.mapreduce.RecordReader<String,String> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
createRecordReader in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String>public void setMinPartitions(org.apache.hadoop.mapreduce.JobContext context,
int minPartitions)