Displaying 20 results from an estimated 11000 matches similar to: "GLMNET warning msg"
2011 Nov 01
1
predict for a cv.glmnet returns an error
Hi there,
I am trying to use predict() with an object returned by cv.glmnet(), and get
the following error:
no applicable method for 'predict' applied to an object of class "cv.glmnet"
What's wrong?
my code:
x=matrix(rnorm(100*20),100,20)
y=rnorm(100)
cv.fit=cv.glmnet(x,y)
predict(cv.fit,newx=x[1:5,])
coef(cv.fit)
Thanks so much,
Asaf
--
View this message in context:
2010 Jun 02
2
glmnet strange error message
Hello fellow R users,
I have been getting a strange error message when using the cv.glmnet
function in the glmnet package. I am attempting to fit a multinomial
regression using the lasso. covars is a matrix with 80 rows and roughly 4000
columns, all the covariates are binary. resp is an eight level factor. I can
fit the model with no errors but when I try to cross-validate after about 30
seconds
2013 Jul 17
1
glmnet on Autopilot
Dear List,
I'm running simulations using the glmnet package. I need to use an
'automated' method for model selection at each iteration of the simulation.
The cv.glmnet function in the same package is handy for that purpose.
However, in my simulation I have p >> N, and in some cases the selected
model from cv.glmet is essentially shrinking all coefficients to zero. In
this case,
2011 Feb 17
1
cv.glmnet errors
Hi,
I am trying to do multinomial regression using the glmnet package, but the
following gives me an error (for no reason apparent to me):
library(glmnet)
cv.glmnet(x=matrix(c(1,2,3,4,5,6,1,2,3,4,5,6),
nrow=6),y=as.factor(c(1,2,1,2,3,3)),family='multinomial',alpha=0.5,
nfolds=2)
The error i get is:
Error in if (outlist$msg != "Unknown error") return(outlist) :
argument is of
2012 Jul 12
1
using glmnet for the dataset with numerical and categorical
Dear R users,
if all my numerical variables in my datasets having the same units, may I
leave them unnormalized, just do cv.glmnet
directly(cv.glmnet(data,standardize=FALSE))?
i know normally if there is a mixture of numerical and categorical , one has
to standardize the numerical part before applying cv.glmnet with
standardize=fase, but that's due to the different units in the numerical
2011 Jul 22
4
glmnet with binary logistic regression
Hi all,
I am using the glmnet R package to run LASSO with binary logistic
regression. I have over 290 samples with outcome data (0 for alive, 1 for
dead) and over 230 predictor variables. I currently using LASSO to reduce
the number of predictor variables.
I am using the cv.glmnet function to do 10-fold cross validation on a
sequence of lambda values which I let glmnet determine. I then take
2011 Mar 25
2
A question on glmnet analysis
Hi,
I am trying to do logistic regression for data of 104 patients, which
have one outcome (yes or no) and 15 variables (9 categorical factors
[yes or no] and 6 continuous variables). Number of yes outcome is 25.
Twenty-five events and 15 variables mean events per variable is much
less than 10. Therefore, I tried to analyze the data with penalized
regression method. I would like please some of the
2012 Mar 21
2
glmnet: obtain predictions using predict and also by extracting coefficients
All,
For my understanding, I wanted to see if I can get glmnet predictions
using both the predict function and also by multiplying coefficients
by the variable matrix. This is not worked out. Could anyone suggest
where I am going wrong?
I understand that I may not have the mean/intercept correct, but the
scaling is also off, which suggests a bigger mistake.
Thanks for your help.
Juliet Hannah
2012 Mar 20
2
cv.glmnet
Hi, all:
Does anybody know how to avoid the intercept term in cv.glmnet coefficient?
When I say "avoid", it does not mean using coef()[-1] to omit the printout
of intercept, it means no intercept at all when doing the analysis. Thanks.
[[alternative HTML version deleted]]
2011 May 01
1
Different results of coefficients by packages penalized and glmnet
Dear R users:
Recently, I learn to use penalized logistic regression. Two packages
(penalized and glmnet) have the function of lasso.
So I write these code. However, I got different results of coef. Can someone
kindly explain.
# lasso using penalized
library(penalized)
pena.fit2<-penalized(HRLNM,penalized=~CN+NoSus,lambda1=1,model="logistic",standardize=TRUE)
pena.fit2
2011 Feb 03
1
glmnet with binary predictors
Hi Everybody!
I must start with a declaration that I am a sparse user of R. I am
creating a credit scorecard using a dataset which has a variable
depicting actual credit history (good/bad) and 41 other variables of
yes/no type. The procedure I am asked to follow is to use a penalized
logistic procedure for variable selection. I have located the package
"glmnet" which gives the complete
2011 Aug 10
2
glmnet
Hi All,
I have been trying to use glmnet package to do LASSO linear regression. my x data is a matrix n_row by n_col and y is a vector of size n_row corresponding to the vector data. The number of n_col is much more larger than the number of n_row. I do the following:
fits = glmnet(x, y, family="multinomial")I have been following this
2023 Oct 23
2
running crossvalidation many times MSE for Lasso regression
For what it's worth it looks like spm2 is specifically for *spatial*
predictive modeling; presumably its version of CV is doing something
spatially aware.
I agree that glmnet is old and reliable. One might want to use a
tidymodels wrapper to create pipelines where you can more easily switch
among predictive algorithms (see the `parsnip` package), but otherwise
sticking to glmnet
2009 Jun 08
3
caret package
Hi all
I am using the caret package and having difficulty in obtaining the results
using regression, I used the glmnet to model and trying to get the
coefficients and the model parameters I am trying to use the
extractPrediction to obtain a confusion matrix and it seems to be giving me
errors.
x<-read.csv("x.csv", header=TRUE);
y<-read.csv("y.csv", header=TRUE);
2023 Oct 24
1
running crossvalidation many times MSE for Lasso regression
?s 20:12 de 23/10/2023, varin sacha via R-help escreveu:
> Dear R-experts,
>
> I really thank you all a lot for your responses. So, here is the error (and warning) messages at the end of my R code.
>
> Many thanks for your help.
>
>
> Error in UseMethod("predict") :
> ? no applicable method for 'predict' applied to an object of class
2011 Apr 14
2
How to see a R code from a package?
Dear R users,
Hi. I want know R code of a function: predict.cv.glmnet (which is
included in glmnet package).
Could you let me know how I can see the R code of the function?
Thank you,
Soyeon Kim
2011 May 28
1
Questions regrading the lasso and glmnet
Hi all. Sorry for the long email. I have been trying to find someone local to work on this with me, without much luck. I went in to our local stats consulting service here, and the guy there told me that I already know more about model selection than he does. :-< He pointed me towards another professor that can perhaps help, but that prof is busy until mid-June, so I want to get as much
2023 Oct 23
1
running crossvalidation many times MSE for Lasso regression
Dear R-experts,
I really thank you all a lot for your responses. So, here is the error (and warning) messages at the end of my R code.
Many thanks for your help.
Error in UseMethod("predict") :
? no applicable method for 'predict' applied to an object of class "c('matrix', 'array', 'double', 'numeric')"
> mean(unlist(lst))
[1] NA
2013 May 23
0
glmnet package: command meanings
Hi List,
I have a little confused when to glmnet() vs cv.glmnet().
I know that,
glmnet(): gives the fit
cv.glment(): does the cv after the fit
I just want to get the beta coefficients after the fit, that's it!
But of all the glmnet examples I've seen, the beta coefficient is
obtained ONLY AFTER cv.glmnet().
Why is that? Also, why is there so many more extra beta's after the fit?
2012 May 28
0
GLMNET AUC vs. MSE
Hello -
I am using glmnet to generate a model for multiple cohorts i. For each i, I
run 5 separate models, each with a different x variable. I want to compare
the fit statistic for each i and x combination.
When I use auc, the output is in some cases is < .5 (.49). In addition, if
I compare mean MSE (with upper and lower bounds) ... there is no difference
across my various x variables, but