Yet another option is provided by the 'rhdf5' package on Bioconductor,
which interfaces to HDF5, a cross-platform infrastructure for exactly
this purpose: http://www.hdfgroup.org/HDF5/whatishdf5.html
Best wishes
Wolfgang
Aug/31/12 4:44 PM, Gabriel Becker scripsit:
I'm not an expert in its use, but I believe the bigmemory package offers
the functionality you are looking for (or at least similar functionality
that can be co-opted for your use-case). See the sub.big.matrix function.
Depending on what you mean by "huge" it may offer other benefits as well.
HTH,
~G
On Fri, Aug 31, 2012 at 6:47 AM, Damien Georges
<damien.georg...@gmail.com>wrote:
Hi all,
I'm working with some huge array in R and I need to load several ones to
apply some functions that requires to have all my arrays values for each
cell...
To make it possible, I would like to load only a part (for example 100
cells) of all my arrays, apply my function, delete all cells loaded, loaded
following cells and so on.
Is it possible to unserialize (or load) only a defined part of an R array ?
Do you know some tools that might help me?
Finally, I did lot of research to find the way array (and all other R
object) are serialized into binary object, but I found nothing explaining
really algorithms involved. If someone has some information on this topic,
I'm interesting in.
Hoping my request is understandable,
All the best,
Damien.G
______________________________**________________
R-devel@r-project.org mailing list
https://stat.ethz.ch/mailman/**listinfo/r-devel<https://stat.ethz.ch/mailman/listinfo/r-devel>
--
Best wishes
Wolfgang
Wolfgang Huber
EMBL
http://www.embl.de/research/units/genome_biology/huber
______________________________________________
R-devel@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-devel