Hi, I have got a lot of SPSS data for years 1993-2010. I load all data into lists so I can easily index the values over the years. Unfortunately loaded data occupy quite a lot of memory (10Gb) - so my question is, what's the best approach to work with big data files? Can R get a value from the file data without full loading into memory? How can a slower computer with not enough memory work with such data?
I use the following commands: data1993 = vector("list", 4); data1993[[1]] = read.spss(...) # first trimester data1993[[2]] = read.spss(...) # second trimester ... data_all = vector("list", 17); data_all[[1993]] = data1993; ... and indexing, e.g.: data_all[[1993]][[1]]$DISTRICT, etc. Thanks, Petr Kurtin ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.