Skip to content

Instantly share code, notes, and snippets.

@jeromatron
Created February 16, 2013 17:23
Show Gist options
  • Star 1 You must be signed in to star a gist
  • Fork 1 You must be signed in to fork a gist
  • Save jeromatron/4967799 to your computer and use it in GitHub Desktop.
Save jeromatron/4967799 to your computer and use it in GitHub Desktop.
//
// Old Hadoop API
//
public org.apache.hadoop.mapred.InputSplit[] getSplits(JobConf jobConf, int numSplits) throws IOException
{
//TaskAttemptContext tac = new TaskAttemptContext(jobConf, new TaskAttemptID());
TaskAttemptContext tac = new org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl(jobConf, new TaskAttemptID());
List<org.apache.hadoop.mapreduce.InputSplit> newInputSplits = this.getSplits(tac);
org.apache.hadoop.mapred.InputSplit[] oldInputSplits = new org.apache.hadoop.mapred.InputSplit[newInputSplits.size()];
for (int i = 0; i < newInputSplits.size(); i++)
oldInputSplits[i] = (ColumnFamilySplit) newInputSplits.get(i);
return oldInputSplits;
}
public org.apache.hadoop.mapred.RecordReader<ByteBuffer, SortedMap<ByteBuffer, IColumn>> getRecordReader(org.apache.hadoop.mapred.InputSplit split, JobConf jobConf, final Reporter reporter) throws IOException
{
TaskAttemptContext tac = new org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl(jobConf, TaskAttemptID.forName(jobConf.get(MAPRED_TASK_ID)))
{
@Override
public void progress ()
{
reporter.progress();
}
};
ColumnFamilyRecordReader recordReader = new ColumnFamilyRecordReader(jobConf.getInt(CASSANDRA_HADOOP_MAX_KEY_SIZE, CASSANDRA_HADOOP_MAX_KEY_SIZE_DEFAULT));
recordReader.initialize((org.apache.hadoop.mapreduce.InputSplit) split, tac);
return recordReader;
}
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment