public class TableInputFormat
extends org.apache.hadoop.mapreduce.InputFormat<org.ojai.Value,org.ojai.Document>
implements org.apache.hadoop.conf.Configurable
Modifier and Type | Field and Description |
---|---|
protected com.mapr.db.impl.ConditionImpl |
cond |
static java.lang.String |
COND_OBJ |
static java.lang.String |
EXCLUDE_EMBEDDEDFAMILY |
static java.lang.String |
FIELD_PATH |
static java.lang.String |
GET_DELETES |
static java.lang.String |
INPUT_TABLE |
protected com.mapr.db.impl.MapRDBTableImpl |
jTable |
static java.lang.String |
READ_ALL_CFS |
static java.lang.String |
START_ROW |
static java.lang.String |
STOP_ROW |
Constructor and Description |
---|
TableInputFormat() |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.RecordReader<org.ojai.Value,org.ojai.Document> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Creates a record reader for a given split.
|
org.apache.hadoop.conf.Configuration |
getConf()
Returns the current configuration.
|
java.util.List<org.apache.hadoop.mapreduce.InputSplit> |
getSplits(org.apache.hadoop.mapreduce.JobContext context)
Set up the splits which will be served as inputs to map tasks.
|
void |
setConf(org.apache.hadoop.conf.Configuration arg0)
This function is used to set parameters in the configuration.
|
protected com.mapr.db.impl.MapRDBTableImpl jTable
protected com.mapr.db.impl.ConditionImpl cond
public static final java.lang.String INPUT_TABLE
public static final java.lang.String FIELD_PATH
public static final java.lang.String COND_OBJ
public static final java.lang.String START_ROW
public static final java.lang.String STOP_ROW
public static final java.lang.String EXCLUDE_EMBEDDEDFAMILY
public static final java.lang.String GET_DELETES
public static final java.lang.String READ_ALL_CFS
public org.apache.hadoop.conf.Configuration getConf()
getConf
in interface org.apache.hadoop.conf.Configurable
Configurable.getConf()
public java.util.List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext context) throws java.io.IOException, java.lang.InterruptedException
getSplits
in class org.apache.hadoop.mapreduce.InputFormat<org.ojai.Value,org.ojai.Document>
context
- The current job context.java.io.IOException
java.lang.InterruptedException
InputFormat.getSplits(org.apache.hadoop.mapreduce.JobContext)
public org.apache.hadoop.mapreduce.RecordReader<org.ojai.Value,org.ojai.Document> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) throws java.io.IOException, java.lang.InterruptedException
createRecordReader
in class org.apache.hadoop.mapreduce.InputFormat<org.ojai.Value,org.ojai.Document>
split
- The input split.context
- Current job context.java.io.IOException
java.lang.InterruptedException
InputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit, org.apache.hadoop.mapreduce.TaskAttemptContext)
public void setConf(org.apache.hadoop.conf.Configuration arg0)
setConf
in interface org.apache.hadoop.conf.Configurable
arg0
- Configuration object with parameters for TableInputFormat.Configurable.setConf(
org.apache.hadoop.conf.Configuration)