Hi all,

I am trying to cluster data file with 50K rows and 10 columns with R. Since
data contains mixed data types (nominal, continuous etc.) I am using gower
distance to create a dissimilarity matrix. But while I was creating the
matrix, I got Error: cannot allocate vector of size 4.1 Gb error. I am
using 64 bit 16 GB RAM.

1-My question is how to use all my memory for R (memory.limit does not work
for linux)?
2-Can I chunk dissimilarty matrix while I am creting it with ;

gower_dist <- daisy(i_cluster_ready,metric = "gower")

3-Can daisy function directly write to ROM to avoid error of RAM usace?


I have tried ff and bigmemory packages. Bigmemory does not work with
mixed data. and I could not figure ff function out.


Thanks for your help


Kind regards

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list -- To UNSUBSCRIBE and more, see
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to