Hi All,

This is not really an R question but a statistical one. If someone could
either give me the brief explanation or point me to a reference that might
help, I'd appreciate it.

I want to estimate the mean of a log-normal distribution, given the (log
scale normal) parameters mu and sigma squared (sigma2). I understood this
should simply be:

exp(mu + sigma2)

... but I the following code gives me something strange:

R <- 10000000
mu <- -400
sigma2 <- 200
tmp <- rlnorm(R, mu, sqrt(sigma2)) # a sample from the desired log-normal
distribution
muh <- mean(log(tmp))
sigma2h <- var(log(tmp))

#by my understanding, all of the the following vectors should then contain
very similar numbers
c(mu, muh)
c(sigma2, sigma2h)
c(exp(mu + sigma2/2), exp(muh + sigma2h/2), mean(tmp))


I get the following (for one sample):
> c(mu, muh)
[1] -400.0000 -400.0231
> c(sigma2, sigma2h)
[1] 200.0000 199.5895
> c(exp(mu + sigma2/2), exp(muh + sigma2h/2), mean(tmp))
[1] 5.148200e-131 4.097249e-131 5.095888e-150

so they do all contain similar numbers, with the exception of the last
vector, which is out by a factor of 10^19. Is this likely to be because one
needs **very** large samples to get a reasonable estimate of the mean... or
am I missing something?

Regards,
Simon

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to