Good day,
I notice that summing rows of a large dgTMatrix fails.
library(Matrix)
aMatrix <- new("dgTMatrix",
i = as.integer(sample(200000, 10000)-1), j =
as.integer(sample(50000, 10000)-1), x = rnorm(10000),
Dim = c(200000L, 50000L)
)
totals <- rowSums(aMatrix == 0) # Segmentation fault.
The server has 768 GB of RAM and it was never close to being consumed by this.
Converting it to an ordinary matrix works fine.
big <- as.matrix(aMatrix)
totals <- rowSums(big == 0) # Uses more RAM but there is no segmentation
fault and result is returned.
May it be made more robust for dgTMatrix?
--------------------------------------
Dario Strbenac
University of Sydney
Camperdown NSW 2050
Australia
______________________________________________
[email protected] mailing list
https://stat.ethz.ch/mailman/listinfo/r-devel