similar to: glmnet 1.9-3 uploaded to CRAN (with intercept option)

Displaying 20 results from an estimated 6000 matches similar to: "glmnet 1.9-3 uploaded to CRAN (with intercept option)"

2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2010 Nov 04
0
glmnet_1.5 uploaded to CRAN
This is a new version of glmnet, that incorporates some bug fixes and speedups. * a new convergence criterion which which offers 10x or more speedups for saturated fits (mainly effects logistic, Poisson and Cox) * one can now predict directly from a cv.object - see the help files for cv.glmnet and predict.cv.glmnet * other new methods are deviance() for "glmnet" and coef() for
2013 Apr 25
0
glmnet webinar Friday May 3 at 10am PDT
I will be giving a webinar on glmnet on Friday May 3, 2013 at 10am PDT (pacific daylight time) The one-hour webinar will consist of: - Intro to lasso and elastic net regularization, and coefficient paths - Why is glmnet so efficient and flexible - New features of the latest version of glmnet - Live glmnet demonstration - Question and Answer period To sign up for the webinar, please go to
2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2012 Jul 03
0
Glmnet_1.8 uploaded to CRAN
This is a major revision, with two additional models included. 1) Multiresponse regression - family="mgaussian" Here we have a matrix of M responses, and we fit a series of linear models in parallel. We use a group-lasso penalty on the set of M coefficients for each variable. This means they are all in or out together 2) family="multinomial, type.multinomial="grouped"
2010 Jun 02
2
glmnet strange error message
Hello fellow R users, I have been getting a strange error message when using the cv.glmnet function in the glmnet package. I am attempting to fit a multinomial regression using the lasso. covars is a matrix with 80 rows and roughly 4000 columns, all the covariates are binary. resp is an eight level factor. I can fit the model with no errors but when I try to cross-validate after about 30 seconds
2011 May 28
1
Questions regrading the lasso and glmnet
Hi all. Sorry for the long email. I have been trying to find someone local to work on this with me, without much luck. I went in to our local stats consulting service here, and the guy there told me that I already know more about model selection than he does. :-< He pointed me towards another professor that can perhaps help, but that prof is busy until mid-June, so I want to get as much
2013 Feb 10
0
glmnet_1.9-1 submitted to CRAN
This new version of glmnet has some bug fixes, and some new features * new arguments lower.limits=-Inf and upper.limits=Inf (defaults shown) for all the coefficients in glmnet. Users can provide limits on coefficients. See the documentation for glmnet. Typical usage: glmnet(x,y,lower=0) Here the argument is abbreviated, and by giving a single value, this uses the same value for all parameters.
2013 Feb 10
0
glmnet_1.9-1 submitted to CRAN
This new version of glmnet has some bug fixes, and some new features * new arguments lower.limits=-Inf and upper.limits=Inf (defaults shown) for all the coefficients in glmnet. Users can provide limits on coefficients. See the documentation for glmnet. Typical usage: glmnet(x,y,lower=0) Here the argument is abbreviated, and by giving a single value, this uses the same value for all parameters.
2003 Apr 30
0
Least Angle Regression packages for R
Least Angle Regression software: LARS "Least Angle Regression" ("LAR") is a new model selection algorithm; a useful and less greedy version of traditional forward selection methods. LAR is described in detail in a paper by Brad Efron, Trevor Hastie, Iain Johnstone and Rob Tibshirani, soon to appear in the Annals of Statistics. The paper, as well as R and Splus packages, are
2003 Apr 30
0
Least Angle Regression packages for R
Least Angle Regression software: LARS "Least Angle Regression" ("LAR") is a new model selection algorithm; a useful and less greedy version of traditional forward selection methods. LAR is described in detail in a paper by Brad Efron, Trevor Hastie, Iain Johnstone and Rob Tibshirani, soon to appear in the Annals of Statistics. The paper, as well as R and Splus packages, are
2011 Aug 10
2
glmnet
Hi All,  I have been trying to use glmnet package to do LASSO linear regression. my x data is a matrix n_row by n_col and y is a vector of size n_row corresponding to the vector data. The number of n_col is much more larger than the number of n_row. I do the following: fits = glmnet(x, y, family="multinomial")I have been following this
2004 Jan 07
0
Statistical Learning and Datamining course based on R/Splus tools
Short course: Statistical Learning and Data Mining Trevor Hastie and Robert Tibshirani, Stanford University Sheraton Hotel Palo Alto, CA Feb 26-27, 2004 This two-day course gives a detailed overview of statistical models for data mining, inference and prediction. With the rapid developments in internet technology, genomics and other high-tech industries, we rely increasingly more on data
2004 Jul 12
0
Statistical Learning and Data Mining Course
Short course: Statistical Learning and Data Mining Trevor Hastie and Robert Tibshirani, Stanford University Georgetown University Conference Center Washington DC September 20-21, 2004 This two-day course gives a detailed overview of statistical models for data mining, inference and prediction. With the rapid developments in internet technology, genomics and other high-tech industries, we
2005 Jan 04
0
Statistical Learning and Data Mining Course
Short course: Statistical Learning and Data Mining Trevor Hastie and Robert Tibshirani, Stanford University Sheraton Hotel, Palo Alto, California February 24 & 25, 2005 This two-day course gives a detailed overview of statistical models for data mining, inference and prediction. With the rapid developments in internet technology, genomics and other high-tech industries, we rely
2010 Jul 31
1
Feature selection via glmnet package (LASSO)
Hello, I'm trying to select features of cetain numbers(like 100 out of 1000) via LASSO, based on multinomial model, however, it seems the glmnet package provides a very sparse estimation of coefficients(most of coefficients are 0), which selects very few number of variables, like only 10, based on my easy dataset. I try to connect the choice of lambda to the selecting