similar to: step, leaps, lasso, LSE or what?

Displaying 20 results from an estimated 1000 matches similar to: "step, leaps, lasso, LSE or what?"

2012 Mar 27
2
lasso constraint
In the package lasso2, there is a Prostate Data. To find coefficients in the prostate cancer example we could impose L1 constraint on the parameters. code is: data(Prostate) p.mean <- apply(Prostate, 5,mean) pros <- sweep(Prostate, 5, p.mean, "-") p.std <- apply(pros, 5, var) pros <- sweep(pros, 5, sqrt(p.std),"/") pros[, "lpsa"] <-
2011 May 28
1
Questions regrading the lasso and glmnet
Hi all. Sorry for the long email. I have been trying to find someone local to work on this with me, without much luck. I went in to our local stats consulting service here, and the guy there told me that I already know more about model selection than he does. :-< He pointed me towards another professor that can perhaps help, but that prof is busy until mid-June, so I want to get as much
2006 Jan 16
2
New RPM packages for CentOS4.0
Greetings list, It's been a while since I've been able to focus on asterisk packaging but this weekend I took some time to audit and recompile packages for CentOS 4.2. You can find them here. ftp://ftp.linuxsys.com/ftp/pub/releases/CentOS-4.0 You have your choice of 1.2.1 or 1.0.10 releases. If you need zaptel modules then install this kernel as well:
2005 Mar 02
2
subset selection for logistic regression
R-packages leaps and subselect implement various methods of selecting best or good subsets of predictor variables for linear regression models, but they do not seem to be applicable to logistic regression models. Does anyone know of software for finding good subsets of predictor variables for linear regression models? Thanks. -Ben p.s., The leaps package references "Subset Selection
2012 May 26
2
avoid error within for loop, try, trycatch, while, move to next iteration, unlist
Hi there, I would like to ask something about how to avoid a possible error message within a for loop. I am running a simulation and in some repetitions there may be an error that will cause a crash and stop the whole procedure, what I want is to simply move on to the next iteration automatically and discard the "bad" repetitions from my results. I used the "try" function to
2006 Jan 26
5
Asterisk 1.2.3 CentOS 4.x RPMS
Available in the usual place. ftp://ftp.linuxsys.com/pub/releases/CentOS-4.0 This release includes minor spec changes, spandsp 0.0.2pre23, a new Sangoma wanpipe RPM for use with the LSE kernel rpm and an AMP installation document. Best Regards, -- Andrew McRory - President/CTO Linux Systems Engineers, Inc. - http://www.linuxsys.com Located in beautiful Tallahassee, Florida Office
2006 Apr 11
2
variable selection when categorical variables are available
Dear All, Probably it is not highly relevant question: Why do stepwise regression functions in R (step() or stepAIC()) add/delete categorical variables as a set? For example, I have a four-level factor variable d, so dummies are d1,d2,d3, as stepwise regression operates d, adding or removing, d1,d2,d3 are simultaneously added/removed. What's the concern here if operating dummies individually?
2006 Jan 16
1
Problem with installation of rpm's, Please help me.
Hi All, I am a newbie and trying to install Asterisk from instructions given in http://www.voip-info.org/tiki-index.php?page=Asterisk+RPM. We have Centos 3.3 so I downloaded rpm's from ftp://ftp.linuxsys.com/pub/LSE/packages/CentOS-3.4/asterisk-1.0.9/ and tried installing one by one but I get the following errors error: Failed dependencies: asterisk = v1.0.9 is needed by
2011 Aug 20
4
I have a problem with R!!
Dear all i?m working with a program i?ve made in R (using functions that others created) to run my program i need a sample. if i generate the sample using for example, rnorm(n, mu, sigma) i have no problem but if i obtain a sample from a column in excel and i copy it, the program says that there is a mistake: it says "Error en `[.data.frame`(data, indices) : undefined columns
2005 Mar 03
1
total variation penalty
Hi, I was recently plowing through the docs of the quantreg package by Roger Koenker and came across the total variation penalty approach to 1-dimensional spline fitting. I googled around a bit and have found some papers originated in the image processing community, but (apart from Roger's papers) no paper that would discuss its statistical aspects. I have a couple of questions in this
2005 Apr 19
2
cross validation and parameter determination
Hi all, In Tibshirani's PNAS paper about nearest shrunken centroid analysis of microarrays (PNAS vol 99:6567), they used cross validation to choose the amount of shrinkage used in the model, and then test the performance of the model with the cross-validated shrinkage in separate independent testing set. If I don't have the luxury of having independent testing set, can I just use the
2009 Aug 14
1
Permutation test and R2 problem
Hi, I have optimized the shrinkage parameter (GCV)for ridge and got my r2 value is 70% . to check the sensitivity of the result, I did permutation test. I permuted the response vector and run for 1000 times and draw a distribution. But now, I get r2 values highest 98% and some of them more than 70 %. Is it expected from such type of test? *I was under impression that, r2 with real data set
2005 Feb 15
1
shrinkage estimates in lme
Hello. Slope estimates in lme are shrinkage estimates which pull the OLS slope estimates towards the population estimates, the degree of which depends on the group sample size and the distance between the group-based estimate and the overall population estimate. Although these shrinkage estimates as said to be more precise with respect to the true values, they are also biased. So there is a
2011 Jun 06
1
Lasso for k-subset regression
Dear R-users I'm trying to use lasso in lars package for subset regression, I have a large matrix of size 1000x100 and my aim is to select a subset k of the 100 variables. Is there any way in lars to fix the number k (i.e. to select the best 10 variables) library(lars) aa=lars(X,Y,type="lasso",max.steps=200) plot(aa,plottype="Cp") aa$RSS which.min(aa$RSS)
2010 Dec 06
2
How to get lasso fit coefficient(given penalty tuning parameter \lambda) using lars package
Hi, all, I am using the lars package for lasso estimate. So I get a lasso fit first: lassofit = lars(x,y,type ="lasso",normalize=T, intercept=T) Then I want to get coefficient with respect to a certain value of \lambda (the tuning parameter), I know lars has three mode options c("step", "fraction", "norm"), but can I use the \lambda value instead
2007 Jun 12
1
LASSO coefficients for a specific s
Hello, I have a question about the lars package. I am using this package to get the coefficients at a specific LASSO parameter s. data(diabetes) attach(diabetes) object <- lars(x,y,type="lasso") cvres<-cv.lars(x,y,K=10,fraction = seq(from = 0, to = 1, length = 100)) fits <- predict.lars(object, type="coefficients", s=0.1, mode="fraction") Can I assign
2012 Jun 16
0
Selecting correlated predictors with LASSO
I'm using the package 'lars' in R with the following code: > library(lars) > set.seed(3) > n <- 1000 > x1 <- rnorm(n) > x2 <- x1+rnorm(n)*0.5 > x3 <- rnorm(n) > x4 <- rnorm(n) > x5 <- rexp(n) > y <- 5*x1 + 4*x2 + 2*x3 + 7*x4 + rnorm(n) > x <- cbind(x1,x2,x3,x4,x5) > cor(cbind(y,x)) y x1 x2
2010 Jan 06
0
parcor 0.2-2 - Regularized Partial Correlation Matrices with (adaptive) Lasso, PLS, and Ridge Regression
Dear R-users, we are happy to announce the release of our R package parcor. The package contains tools to estimate the matrix of partial correlations based on different regularized regression methods: Lasso, adaptive Lasso, PLS, and Ridge Regression. In addition, parcor provides cross-validation based model selection for Lasso, adaptive Lasso and Ridge Regression. More details can be found
2010 Jan 06
0
parcor 0.2-2 - Regularized Partial Correlation Matrices with (adaptive) Lasso, PLS, and Ridge Regression
Dear R-users, we are happy to announce the release of our R package parcor. The package contains tools to estimate the matrix of partial correlations based on different regularized regression methods: Lasso, adaptive Lasso, PLS, and Ridge Regression. In addition, parcor provides cross-validation based model selection for Lasso, adaptive Lasso and Ridge Regression. More details can be found
2012 Jun 05
1
Piecewise Lasso Regression
Hi All, I am trying to fit a piecewise lasso regression, but package Segmented does not work with Lars objects. Does any know of any package or implementation of piecewise lasso regression? Thanks, Lucas