similar to: glmnet webinar Friday May 3 at 10am PDT

Displaying 20 results from an estimated 2000 matches similar to: "glmnet webinar Friday May 3 at 10am PDT"

2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2013 Mar 02
0
glmnet 1.9-3 uploaded to CRAN (with intercept option)
This update adds an intercept option (by popular request) - now one can fit a model without an intercept Glmnet is a package that fits the regularization path for a number of generalized linear models, with with "elastic net" regularization (tunable mixture of L1 and L2 penalties). Glmnet uses pathwise coordinate descent, and is very fast. The current list of models covered are:
2013 Mar 02
0
glmnet 1.9-3 uploaded to CRAN (with intercept option)
This update adds an intercept option (by popular request) - now one can fit a model without an intercept Glmnet is a package that fits the regularization path for a number of generalized linear models, with with "elastic net" regularization (tunable mixture of L1 and L2 penalties). Glmnet uses pathwise coordinate descent, and is very fast. The current list of models covered are:
2010 Nov 04
0
glmnet_1.5 uploaded to CRAN
This is a new version of glmnet, that incorporates some bug fixes and speedups. * a new convergence criterion which which offers 10x or more speedups for saturated fits (mainly effects logistic, Poisson and Cox) * one can now predict directly from a cv.object - see the help files for cv.glmnet and predict.cv.glmnet * other new methods are deviance() for "glmnet" and coef() for
2013 May 14
0
apcluster webinar: Thursday, June 13, 2013, 7:00pm CEST
Dear colleagues, This is to inform you that I will be giving a webinar on the apcluster package on Thursday, June 13, 2013, 7:00pm CEST (10:00am PDT). The outline of the one-hour webinar is as follows: - Introduction to affinity propagation (AP) clustering - The apcluster package, its algorithms, and visualization tools - Live apcluster demonstration - Question and Answer period To register
2013 Feb 10
0
glmnet_1.9-1 submitted to CRAN
This new version of glmnet has some bug fixes, and some new features * new arguments lower.limits=-Inf and upper.limits=Inf (defaults shown) for all the coefficients in glmnet. Users can provide limits on coefficients. See the documentation for glmnet. Typical usage: glmnet(x,y,lower=0) Here the argument is abbreviated, and by giving a single value, this uses the same value for all parameters.
2013 Feb 10
0
glmnet_1.9-1 submitted to CRAN
This new version of glmnet has some bug fixes, and some new features * new arguments lower.limits=-Inf and upper.limits=Inf (defaults shown) for all the coefficients in glmnet. Users can provide limits on coefficients. See the documentation for glmnet. Typical usage: glmnet(x,y,lower=0) Here the argument is abbreviated, and by giving a single value, this uses the same value for all parameters.
2010 Nov 19
0
glmnet_1.5.1 uploaded to CRAN
In glmnet_1.5 a poor default was set for the argument type which caused the program to be very slow or even crash when nvar (p) is very large. The argument type (now called type.gaussian) has two options, "covariance" or "naive", and is used for the default family="gaussion" model (squared error loss). When type.gaussian="covariance", all inner-products
2009 Oct 30
0
different L2 regularization behavior between lrm, glmnet, and penalized? (original question)
Dear Robert, The differences have to do with diffent scaling defaults. lrm by default standardizes the covariates to unit sd before applying penalization. penalized by default does not do any standardization, but if asked standardizes on unit second central moment. In your example: x = c(-2, -2, -2, -2, -1, -1, -1, 2, 2, 2, 3, 3, 3, 3) z = c(0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1) You
2011 Jul 18
0
Reminder: Monitoring GlusterFS Webinar is Tomorrow
Greetings - if you're curious about monitoring GlusterFS performance, be sure and sign up for tomorrow's webinar. We will also post the recording online should you not be able to make it. Introducing Gluster for Geeks Technical Webinar Series In this Gluster for Geeks technical webinar, Craig Carl, Senior Systems Engineer, will explain and demonstrate how to monitor your Gluster
2009 Oct 14
1
different L2 regularization behavior between lrm, glmnet, and penalized?
The following R code using different packages gives the same results for a simple logistic regression without regularization, but different results with regularization. This may just be a matter of different scaling of the regularization parameters, but if anyone familiar with these packages has insight into why the results differ, I'd appreciate hearing about it. I'm new to
2005 Nov 28
0
glmpath: L1 regularization path for glms
We have uploaded to CRAN the first version of glmpath, which fits the L1 regularization path for generalized linear models. The lars package fits the entire piecewise-linear L1 regularization path for the lasso. The coefficient paths for L1 regularized glms, however, are not piecewise linear. glmpath uses convex optimization - in particular predictor-corrector methods- to fit the
2005 Nov 28
0
glmpath: L1 regularization path for glms
We have uploaded to CRAN the first version of glmpath, which fits the L1 regularization path for generalized linear models. The lars package fits the entire piecewise-linear L1 regularization path for the lasso. The coefficient paths for L1 regularized glms, however, are not piecewise linear. glmpath uses convex optimization - in particular predictor-corrector methods- to fit the
2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2015 Mar 19
0
RFC: Matrix package: Matrix products (%*%, crossprod, tcrossprod) involving "nsparseMatrix" aka sparse pattern matrices
Hi Martin I got stung by this last week. glmnet produces a coefficient matrix of class ?dgCMatrix? If a predictor matrix was created using sparseMatrix as follows, one gets unexpected results, as this simple example shows. My fix was easy (I always convert the predictor matrix to class ?dgCMatrix? now) Trevor > y=Matrix(diag(4)) > y 4 x 4 diagonal matrix of class "ddiMatrix"
2013 Jul 06
1
problem with BootCV for coxph in pec after feature selection with glmnet (lasso)
Hi, I am attempting to evaluate the prediction error of a coxph model that was built after feature selection with glmnet. In the preprocessing stage I used na.omit (dataset) to remove NAs. I reconstructed all my factor variables into binary variables with dummies (using model.matrix) I then used glmnet lasso to fit a cox model and select the best performing features. Then I fit a coxph model