Noah Silverman wrote:
Frank,
That makes sense.
I just had a look at the actual algorithm calculating the Briar score.
One thing that confuses me is how the score is calculated.
If I understand the code correctly, it is just: sum((p - y)^2)/n
If I have an example with a label of 1 and a probability prediction of
.4, it is (.4 - 1)^2
(I know it is the average of these value across all the examples)
Yes and I seem to remember the original score is 1 minus that.
Wouldn't it make more sense to stratify the probabilities and then check
the accuracy of each level.
The stratification will bring a great deal of noise into the problem.
Better: loess calibration curves or decomposition of the Brier score
into discrimination and calibration components (which is not in the
software).
Frank
i.e. For predicted probabilities of .10 to .20 the data was actually
labeled true .18 percent of the time. mean(label)
On 8/19/09 11:51 AM, Frank E Harrell Jr wrote:
Noah Silverman wrote:
Thanks for the suggestion.
You explained that Briar combines both accuracy and discrimination
ability. If I understand you right, that is in relation to binary
classification.
I'm not concerned with binary classification, but the accuracy of the
probability predictions.
Is there some kind of score that measures just the accuracy?
Thanks!
-N
The Brier score has nothing to do with classification. It is a
probability accuracy score.
Frank
On 8/19/09 10:42 AM, Frank E Harrell Jr wrote:
Noah Silverman wrote:
Hello,
I working on a model to predict probabilities.
I don't really care about binary prediction accuracy.
I do really care about the accuracy of my probability predictions.
Frank was nice enough to point me to the val.prob function from the
Design library. It looks very promising for my needs.
I've put together some tests and run the val.prob analysis. It
produces some very informative graphs along with a bunch of
performance measures.
Unfortunately, I'm not sure which measure, if any, is the "best"
one. I'm comparing hundreds of different models/parameter
combinations/etc. So Ideally I'd like a single value or two as the
"performance measure" for each one. That way I can pick the
"best" model from all my experiments.
As mentioned above, I'm mainly interested in the accuracy of my
probability predictions.
Does anyone have an opinion about which measure I should look at??
(I see Dxy, C, R2, D, U, Briar, Emax, Eavg, etc.)
Thanks!!
-N
It all depends on the goal, i.e., the relative value you place on
absolute accuracy vs. discrimination ability. The Brier score
combines both and other than interpretability has many advantages.
Frank
______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide
http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.
--
Frank E Harrell Jr Professor and Chair School of Medicine
Department of Biostatistics Vanderbilt University
______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.