Dear R team,
 
I'm not sure if I use the right distribution list, but I hope in case if
not, you will forward it to the reference person.
 
Following problem occured:
I used R to calculate the p-value for the two sided binomial test (exact -
Pearson).
For a very little difference for my forecast I get a very big difference in
my p-value
 
 >  binom.test(1,101, 0.02402)

        Exact binomial test

data:  1 and 101
number of successes = 1, number of trials = 101, p-value = 0.7375
alternative hypothesis: true probability of success is not equal to 0.02402
95 percent confidence interval:
0.00025064 0.05393235
sample estimates:
probability of success 
            0.00990099

> binom.test(1,101, 0.02403)

        Exact binomial test

data:  1 and 101
number of successes = 1, number of trials = 101, p-value = 0.5243
alternative hypothesis: true probability of success is not equal to 0.02403
95 percent confidence interval:
0.00025064 0.05393235
sample estimates:
probability of success 
            0.00990099


Can you please explain where this huge difference come from?
Which mathematical explanation is given for this topic?

Please help.

I hope for your soon feedback.

Kind Regards,
Alwina
 

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to