On Jul 17, 2013, at 5:26 PM, Axel Urbiz wrote:

Dear List,

I'm running simulations using the glmnet package. I need to use an
'automated' method for model selection at each iteration of the simulation.
The cv.glmnet function in the same package is handy for that purpose.
However, in my simulation I have p >> N, and in some cases the selected model from cv.glmet is essentially shrinking all coefficients to zero. In
this case, the prediction for each instance equals the average of the
response variable. A reproducible example is shown below.

Is there a reasonable way to prevent this from happening in a simulation setting with glmnet? That is, I'd like the selected model to give me some
useful predictions.

I'd like to expose the premise of the request to criticism. Reporting the sample mean in cases where no preditctors meet the criteria for significance under penalsization IS an informative response under conditions of simulation. The simulated result is telling you that in some data situations of modest size assess under a penalized process will not deliver a "significant" result. Why does this bother yu\ou? The number of such messages would seem to be one measure of the power of the method, although other departures from the "true" result would also be substracted from teh count of runs.

If you choose to ignore the "evidence", then I "predict" that you are also one who chooses to throw out outliers. Both would have a similar effect of inflating measures of significance at the expense of fideltity to the data. If you want to vary the parameter, then vary the penalization and determine the effect of that hyper-parameter.

David Winsemius


I've tested using alternative loss measures (type.measure argument), but
none is satisfactory in all cases.

This question is not necessarily R related (so sorry for that): when
comparing glmnet with other models in terms of predictive accuracy, is it
fair to make the comparison including those cases in which the `best'
cv.glmnet can do in an automated setting is pred = avg(response)?

library(glmnet)
set.seed(1010)
n=100;p=3000
nzc=trunc(p/10)
x=matrix(rnorm(n*p),n,p)
beta=rnorm(nzc)
fx= x[,seq(nzc)] %*% beta
eps=rnorm(n)*5
y=drop(fx+eps)
px=exp(fx)
px=px/(1+px)
ly=rbinom(n=length(px),prob=px,size=1)

fit.net <- cv.glmnet(x,
                    ly,
                    family = "binomial",
                    alpha = 1, # lasso penalty
                    type.measure = "deviance",
                    standardize = FALSE,
                    intercept = FALSE,
                    nfolds = 10,
                    keep = FALSE)

plot(fit.net)
log(fit.net$lambda.1se)
pred <- predict(fit.net, x,
               type = "response", s = "lambda.1se")
all(coef(fit.net) == 0)
all(pred ==0.5)


Thanks in advance for your thoughts.

Regards,
Lars.

        [[alternative HTML version deleted]]

No problems with this posting for my mail client but you should learn to use the facilities in gmail to send palin text. Yhey are easy to fnd.

--
David Winsemius, MD
Alameda, CA, USA

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to