Duncan Murdoch <murdoch.duncan <at> gmail.com> writes:

> 
> On 11-10-18 4:30 AM, Seref Arikan wrote:
> > Hi Dan,
> > I've tried the log likelihood, but it reaches zero again, if I work with say
> > 1000 samples.
> > I need an approach that would scale to quite large sample sizes. Surely I
> > can't be the first one to encounter this problem, and I'm sure I'm missing
> > an option that is embarrassingly obvious.
> 
> I think you are calculating the log likelihood incorrectly.  Don't 
> calculate the likelihood and take the log; work out the formula for the 
> log of the likelihood, and calculate that.  (If the likelihood contains 
> a sum of terms, as in a mixture model, this takes some thinking, but it 
> is still worthwhile.)
> 
> With most models, it is just about impossible to cause the log 
> likelihood to underflow if it is calculated carefully.
> 
> Duncan Murdoch
> 

  I haven't followed this carefully, but there is a special problem
in Bayesian situations where at some point you may have to *integrate*
the likelihoods, which implies dealing with them on the likelihood
(not log-likelihood) scale.  There are various ways of dealing with
this, but one is to factor a constant (which will be a very small value)
out of the elements in the integrand.

  Ben Bolker

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to