I use package multicore, and it works very well. There is, however, one thing I 
wonder if I don't do correctly, here is one example:  I read ca 5.5 mill 
records into a dataframe, using read.dta through a one-line function rdam1. It 
runs nicely in parallel with other activitites.

  p1 <-  parallel(rdam1()) ;

Then I recover the data frame with collect:

 am1 <- (collect(p1))[[1]]

All goes well, but the collecting step takes quite some time, and as the data 
frame is already in memory, should this be necessary? I experience the same 
with all parallelized steps resulting in large objects.  Is there a way to 
avoid this, for instance, accessing the object through a pointer?

Trond Ydersbond




        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to