i would suggest that if you want to use R that you get a 64-bit version with 24GB of memory to start. if your data is a numeric matrix, you will need 8GB for a single copy.
Do you really need it all in memory at once, or can you partition the problem? Can you use a database to access the portion you need at any time? If you only need one, or two, columns at a time, then the use of a database storing the columns might work. You probably need some more analysis on exactly how you want to solve your problem understanding the limitations of the system. Sent from my iPad On Sep 2, 2011, at 1:13, Worik R <wor...@gmail.com> wrote: > Friends > > I am starting on a (section of the) project where I need to build a matrix > with on the order of 5 million rows and 200 columns > > I am wondering if I can stay in R. > > I need to do rollapply type operations on the columns, including some that > will be functions of (windows of) two columns. > > I have been looking at the ff and bigmemory packages but am not sure that > they will do. > > Before I get too deep can some one offer some wisdom about what the best > direction to go would be? > > Switching to C/C++ is definitely an option if it is all too hard > > cheers > Worik > > [[alternative HTML version deleted]] > > ______________________________________________ > R-help@r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.