Hi guys :
I have a map/r job that has always worked fine, but which fails due to a
heap space error on my local machine during unit tests.
It runs in hadoop's default mode, and just fails durring the constructor of
the MapOutputBuffer.... Any thoughts on why ?
I dont do any custom memory settings in by unit tests, because they aren't
really needed --- So I assume this is related to /tmp files
or something ... but cant track down the issue.
Any thoughts would be very much appreciated ..
12/05/01 19:15:53 WARN mapred.LocalJobRunner: job_local_0002
java.lang.OutOfMemoryError: Java heap space
at
org.apache.hadoop.mapred.MapTask$MapOutputBuffer.<init>(MapTask.java:807)
at
org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:557)
--
Jay Vyas
MMSB/UCHC