similar to: sparsenet: a new package for sparse model selection

Displaying 20 results from an estimated 1000 matches similar to: "sparsenet: a new package for sparse model selection"

2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2013 Apr 02
0
softImpute_1.0 uploaded to CRAN
SoftImpute is a new package for matrix completion - i.e. for imputing missing values in matrices. SoftImpute was written by myself and Rahul Mazumder. softImpute uses uses squared-error loss with nuclear norm regularization - one can think of it as the "lasso" for matrix approximation - to find a low-rank approximation to the observed entries in the matrix. This low-rank approximation
2013 Mar 02
0
glmnet 1.9-3 uploaded to CRAN (with intercept option)
This update adds an intercept option (by popular request) - now one can fit a model without an intercept Glmnet is a package that fits the regularization path for a number of generalized linear models, with with "elastic net" regularization (tunable mixture of L1 and L2 penalties). Glmnet uses pathwise coordinate descent, and is very fast. The current list of models covered are:
2013 Mar 02
0
glmnet 1.9-3 uploaded to CRAN (with intercept option)
This update adds an intercept option (by popular request) - now one can fit a model without an intercept Glmnet is a package that fits the regularization path for a number of generalized linear models, with with "elastic net" regularization (tunable mixture of L1 and L2 penalties). Glmnet uses pathwise coordinate descent, and is very fast. The current list of models covered are:
2010 Nov 04
0
glmnet_1.5 uploaded to CRAN
This is a new version of glmnet, that incorporates some bug fixes and speedups. * a new convergence criterion which which offers 10x or more speedups for saturated fits (mainly effects logistic, Poisson and Cox) * one can now predict directly from a cv.object - see the help files for cv.glmnet and predict.cv.glmnet * other new methods are deviance() for "glmnet" and coef() for
2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2008 Jun 02
0
New glmnet package on CRAN
glmnet is a package that fits the regularization path for linear, two- and multi-class logistic regression models with "elastic net" regularization (tunable mixture of L1 and L2 penalties). glmnet uses pathwise coordinate descent, and is very fast. Some of the features of glmnet: * by default it computes the path at 100 uniformly spaced (on the log scale) values of the
2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2010 Apr 04
0
Major glmnet upgrade on CRAN
glmnet_1.2 has been uploaded to CRAN. This is a major upgrade, with the following additional features: * poisson family, with dense or sparse x * Cox proportional hazards family, for dense x * wide range of cross-validation features. All models have several criteria for cross-validation. These include deviance, mean absolute error, misclassification error and "auc" for logistic or
2005 Nov 28
0
glmpath: L1 regularization path for glms
We have uploaded to CRAN the first version of glmpath, which fits the L1 regularization path for generalized linear models. The lars package fits the entire piecewise-linear L1 regularization path for the lasso. The coefficient paths for L1 regularized glms, however, are not piecewise linear. glmpath uses convex optimization - in particular predictor-corrector methods- to fit the
2005 Nov 28
0
glmpath: L1 regularization path for glms
We have uploaded to CRAN the first version of glmpath, which fits the L1 regularization path for generalized linear models. The lars package fits the entire piecewise-linear L1 regularization path for the lasso. The coefficient paths for L1 regularized glms, however, are not piecewise linear. glmpath uses convex optimization - in particular predictor-corrector methods- to fit the
2006 Mar 02
0
glmpath (new version 0.91)
We have uploaded to CRAN a new version of glmpath, a package which fits the L1 regularization path for generalized linear models. The revision includes: - coxpath, a function for fitting the L1-regularization path for the Cox ph model; - bootstrap functions for analyzing sparse solutions; - the ability to mix in L2 regularization along with L1 (elasticnet). We have also completed a report that
2006 Mar 02
0
glmpath (new version 0.91)
We have uploaded to CRAN a new version of glmpath, a package which fits the L1 regularization path for generalized linear models. The revision includes: - coxpath, a function for fitting the L1-regularization path for the Cox ph model; - bootstrap functions for analyzing sparse solutions; - the ability to mix in L2 regularization along with L1 (elasticnet). We have also completed a report that
2006 Dec 19
1
preserving sparse matrices (Matrix)
Hi, I have sparse (tridiagonal) matrices, and I use the Matrix package for handling them. For certain computations, I need to either set the first row to zero, or double the last row. I find that neither operation preserves sparsity, eg > m <- Diagonal(5) > m 5 x 5 diagonal matrix of class "ddiMatrix" [,1] [,2] [,3] [,4] [,5] [1,] 1 . . . . [2,] . 1
2011 Jan 12
0
Multivariate autoregressive models with lasso penalization
I wish to estimate sparse causal networks from simulated time series data. Although there's some discussion about this problem in the literature (at least a few authors have used lasso and l(1,2) regularization to enforce sparsity in multivariate autoregressive models, e.g., http://user.cs.tu-berlin.de/~nkraemer/papers/grplasso_causality.pdf), I can't find any R packages with these
2010 Dec 20
0
survexp - unable to reproduce example
Dear All, when I try to reproduce an example of survexp, taken from the help page of survdiff, I receive the error message "Error in floor(temp) : Non-numeric argument to mathematical function" . It seems to come from match.ratetable. I think, it has to do with character variables in a ratetable. I would be interested to know, if it works for others. With an older version of
2010 Dec 31
3
survexp - example produces error
Dear All, reposting, because I did not find a solution, maybe someone could check the example below. It's taken from the help page of survdiff. Executing it, gives the error "Error in floor(temp) : Non-numeric argument to mathematical function" best regards, Heinz library(survival) ## Example from help page of survdiff ## Expected survival for heart transplant patients based
2017 Oct 21
1
What exactly is an dgCMatrix-class. There are so many attributes.
> On Oct 21, 2017, at 7:50 AM, Martin Maechler <maechler at stat.math.ethz.ch> wrote: > >>>>>> C W <tmrsg11 at gmail.com> >>>>>> on Fri, 20 Oct 2017 15:51:16 -0400 writes: > >> Thank you for your responses. I guess I don't feel >> alone. I don't find the documentation go into any detail. > >> I also find
2009 Nov 17
0
Marginal Homogeneity tests for sparse matrixes ?
Hello people, I am in need for testing Marginal Homogeneity for sparse (more then 2X2) matrixes. After searching, what I found by now is that for more then 2 by 2 matrixes, one turns to "stuart maxwell tests" that are available in two packages: irr - see: stuart.maxwell.mh coin - see: mh_test But I couldn't find in the documentation how valid the results are for sparse matrixes,
2009 Feb 15
1
GLMM, ML, PQL, lmer
Dear R community, I have two questions regarding fitting GLMM using maximum likelihood method. The first one arises from trying repeat an analysis in the Breslow and Clayton 1993 JASA paper. Model 3 of the epileptic dataset has two random effects, one subject specific, and one observation specific. Thus if we count random effects, there are more parameters than observations. When I try to run the