Hello Jeff,
Thank you for your comments, bu the problem is not about the RangeBatchSize.
In the case of the configuration parameter,
mapred.tasktracker.map.tasks.maximum > 1
all the map task times out, they don't even run a single line of code in the
Mapper.map() function.
In the case of the con
The default batch size is 4096, which means that each call to
get_range_slices retrieves 4,096 rows. I have found that this causes
timeouts when cassandra is under load. Try reducing the batchsize
with a call to ConfigHelper.setRangeBatchSize(). This has eliminated
the TimedOutExceptions for us.
Hey All,
I'm trying to run some tests on cassandra an Hadoop integration. I'm
basically following the word count example at
https://svn.apache.org/repos/asf/cassandra/trunk/contrib/word_count/src/WordCount.javausing
the ColumnFamilyInputFormat.
Currently I have one-node cassandra and hadoop setup