One more note. In case it is helpful, I am including the code for my loop:

# data is read in

numSNPs <- ncol(myData);
pvalues <- rep(-1,numSNPs);
names(pvalues) <- colnames(myData);


for (SNPnum in 1:numSNPs)
{
   is.na(pvalues[SNPnum]) <- TRUE;
   try({
      fit.yags <- yags(log(myPhenos$PHENOTYPE) ~
myPhenos$AGE+myPhenos$SEX*myData[,SNPnum], id=myPhenos$id,
family=gaussian,corstr="exchangeable",alphainit=0.05)
      z.gee <- fit.y...@coefficients[5]/sqrt(fit.y...@robust.parmvar[5,5]);
      pval <- 2 * pnorm(abs(z.gee), lower.tail = FALSE);
      pvalues[SNPnum] <- pval;
})
}

pvalues <- format(pvalues,digits=3);

On Mon, Dec 29, 2008 at 11:59 AM, Juliet Hannah <juliet.han...@gmail.com> wrote:
> I monitored the usage of memory on a script that I ran. It ran 30K
> regressions and it stores p-values for one of the
> coefficients. It read in a file that has 3000 rows and about 30K
> columns. The size of the file is about 170 MB.
>
> My understanding is that memory usage started out at 2.2G and went up to 23G:
>
>
> cpu=00:03:08, mem=172.75822 GBs, io=0.00000, vmem=2.224G, maxvmem=2.224G
> cpu=00:42:35, mem=29517.64894 GBs, io=0.00000, vmem=23.612G, maxvmem=23.612G
>
> I know very little about how memory works, but I thought the hardest
> part would be reading the file in. Could
> someone explain why there is such a substantial increase over the
> course of the script.
>
> Thanks,
>
> Juliet
>

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to