search for: shrinkag

Displaying 20 results from an estimated 102 matches for "shrinkag".

Did you mean: shrinkage
2005 Feb 15
1
shrinkage estimates in lme
Hello. Slope estimates in lme are shrinkage estimates which pull the OLS slope estimates towards the population estimates, the degree of which depends on the group sample size and the distance between the group-based estimate and the overall population estimate. Although these shrinkage estimates as said to be more precise with respect to...
2008 May 06
1
mgcv::gam shrinkage of smooths
...zero. IE, my informal prior is to keep the contribution of a specific term small. 1) Is adding eps*I to the penalty matrix an effective way to achieve this goal? 2) How do I accomplish this in practice using mgcv::gam? Thanks. -- View this message in context: http://www.nabble.com/mgcv%3A%3Agam-shrinkage-of-smooths-tp17093645p17093645.html Sent from the R help mailing list archive at Nabble.com.
2005 Jun 01
0
determine the shrinkage threshold in PAMR?
1. According to the doc of PAMR, the shrinkage threshold is determined by cross-validation. Does this mean that user need not tune any parameter? 2. I tried two applications using PAMR, the results are very disappointing. The attached are the cross-validation results. You can see that the classification errors are relatively high (0.2 at the...
2008 Jan 02
0
How to select a reasonable shrinkage coefficient in stepplr?
Dear R-users, I am using stepplr for L2 regularized logistic regression. Since number of attribute is too large i discarded interaction terms. Everything is fine but only problem i have faced that i cannot choose a good shrinkage coefficient (lambda). If CV is the best way to estimate, can you please elaborately tell me how to select lambda in stepplr using CV? Except CV is there any other procedure available? Thanks. -- View this message in context: http://www.nabble.com/How-to-select-a-reasonable-shrinkage-coefficient...
2005 Apr 19
2
cross validation and parameter determination
Hi all, In Tibshirani's PNAS paper about nearest shrunken centroid analysis of microarrays (PNAS vol 99:6567), they used cross validation to choose the amount of shrinkage used in the model, and then test the performance of the model with the cross-validated shrinkage in separate independent testing set. If I don't have the luxury of having independent testing set, can I just use the cross validation performance as the performance estimate? In other words, can I...
2010 Apr 13
0
exract Shrinkage intensity lambda and lambda.var
does anyone know how to extract Shrinkage intensity lambda and lambda.var values after run cov.shrink(x)? thanks, KZ [[alternative HTML version deleted]]
2009 Aug 14
1
Permutation test and R2 problem
Hi, I have optimized the shrinkage parameter (GCV)for ridge and got my r2 value is 70% . to check the sensitivity of the result, I did permutation test. I permuted the response vector and run for 1000 times and draw a distribution. But now, I get r2 values highest 98% and some of them more than 70 %. Is it expected from such type...
2012 Dec 08
0
Oracle Approximating Shrinkage in R?
Hi, Can anyone point me to an implementation in R of the oracle approximating shrinkage technique for covariance matrices? Rseek, Google, etc. aren't turning anything up for me. Thanks in advance, Matt Considine
2007 Jun 21
1
mgcv: lowest estimated degrees of freedom
...02, Ecol. Model. 157, p. 157-177). One criterion to decide if a term should be dropped from a model is if the estimated degrees of freedom (EDF) for the term are close to their lower limit. What would be the minimum number of EDF's for a) Univariate thin plate regression splines(TPRS) with shrinkage, i.e. s(...,bs="ts") b) Bivariate tensor products of TPRS with shrinkage? Thanks for any help, Julian Burgos -- Julian M. Burgos Fisheries Acoustics Research Lab School of Aquatic and Fishery Science University of Washington 1122 NE Boat Street Seattle, WA 98105 Phone: 206-221...
2006 May 27
2
boosting - second posting
...ease, + # 0: no monotone restrictions + distribution="gaussian", # bernoulli, adaboost, gaussian, + # poisson, and coxph available + n.trees=3000, # number of trees + shrinkage=0.005, # shrinkage or learning rate, + # 0.001 to 0.1 usually work + interaction.depth=3, # 1: additive model, 2: two-way interactions, etc. + bag.fraction = 0.5, # subsampling fraction, 0.5 is probably best +...
2006 Apr 11
2
variable selection when categorical variables are available
...as a set? For example, I have a four-level factor variable d, so dummies are d1,d2,d3, as stepwise regression operates d, adding or removing, d1,d2,d3 are simultaneously added/removed. What's the concern here if operating dummies individually? Model interpretability or anything else? (it seems shrinkage methods can operate them one by one) Thanks mike [[alternative HTML version deleted]]
2014 Jul 02
0
How do I call a C++ function (for k-means) within R?
...ribution=as.character(distribution.call.name), n.trees=as.integer(n.trees), interaction.depth=as.integer(interaction.depth), n.minobsinnode=as.integer(n.minobsinnode), n.classes = as.integer(nClass), shrinkage=as.double(shrinkage), bag.fraction=as.double(bag.fraction), nTrain=as.integer(nTrain), fit.old=as.double(NA), n.cat.splits.old=as.integer(0), n.trees.old=as.integer(0), verbose=a...
2002 Mar 01
2
step, leaps, lasso, LSE or what?
...ng overfitting. In Hastie, Tibshirani and Friedman "The Elements of Statistical Learning" chapter 3, they describe a number of procedures that seem better. The use of cross-validation in the training stage presumably helps guard against overfitting. They seem particularly favorable to shrinkage through ridge regressions, and to the "lasso". This may not be too surprising, given the authorship. Is the lasso "generally accepted" as being a pretty good approach? Has it proved its worth on a variety of problems? Or is it at the "interesting idea" stage? What,...
2013 Jun 24
2
Nomogram (rms) for model with shrunk coefficients
...ach(d) ddist<-datadist(d) options(datadist='ddist') model<-lrm(y~x1+x2, x=TRUE, y=TRUE, data=d) plot(nomogram(model)) ##Nomogram is printed, as expected ##Now the model is internally validated, and regression coefficients are penalized bootstrap<-validate(model, bw=FALSE, B=100) shrinkage<-round(bootstrap[4,5],2) final<-round(model$coef*shrinkage, 3) final.lp<-cbind(model$x)%*%final[-1] final["Intercept"]<-round(lrm.fit(y=d$y, offset=final.lp)$coef,3) final.lp<-final[1]+model$x%*%final[-1] ##The object 'final' now contains all model parameters, ye...
2009 Jun 17
1
gbm for cost-sensitive binary classification?
...ne has similar experience and can advise me how to implement cost-sensitive classification with gbm. model.gbm <- gbm.fit(tr[,1:DIM],tr.y,offset = NULL,misc = NULL,distribution = "bernoulli",w = tr.w,var.monotone = NULL,n.trees = NTREE,interaction.depth = TREEDEPTH,n.minobsinnode = 10,shrinkage = 0.05,bag.fraction = BAGRATIO,train.fraction = 1.0,keep.data = TRUE,verbose = TRUE,var.names = NULL,response.name = NULL); or model.gbm <- gbm(tr.y ~ .,distribution = "bernoulli",data=data.frame(cbind(tr[,1:DIM],tr.y)),weights = tr.w,var.monotone=NULL,n.trees=NTREE,interaction.de...
2008 Sep 18
1
caret package: arguments passed to the classification or regression routine
...on. here is the code I used and the error: gbm.test <- train(x.enet, y.matrix[,7], method="gbm", distribution=list(name="quantile",alpha=0.5), verbose=FALSE, trControl=trainControl(method="cv",number=5), tuneGrid=gbmGrid ) Model 1: interaction.depth=1, shrinkage=0.1, n.trees=300 collapsing over other values of n.trees Error in gbm.fit(trainX, modY, interaction.depth = tuneValue$.interaction.depth, : formal argument "distribution" matched by multiple actual arguments The same error occured with distribution="laplace". I also tried...
2011 Dec 05
1
finding interpolated values along an empirical parametric curve
...1)) lambdaf <- c(expression(~widehat(beta)^OLS), ".005", ".01", ".02", ".04", ".08") op <- par(mar=c(4, 4, 1, 1) + 0.2, xpd=TRUE) with(pd, {plot(norm.beta, log.det, type="b", cex.lab=1.25, pch=16, cex=1.5, col=clr, xlab='shrinkage: ||b||', ylab='variance: log |(Var(b)|)') text(norm.beta, log.det, lambdaf, cex=1.25, pos=2) text(min(norm.beta), max(log.det), "Variance vs. Shrinkage", cex=1.5, pos=4) }) # How to find the (x,y) positions for these values of lambda along the curve of...
2003 Sep 23
1
Very small estimated random effect variance (lme)
...cts model (lme), of the type: lme1 <- lme(y ~ x, random=~x|group, ...) For some datasets, i obtain very small standard deviations of the random effects. I compared these to standard deviations of the slope and intercept using a lmList approach. Of course, the SD from the lme is always smaller (shrinkage estimator), but in some cases (the problem cases) the SD from the lme seems way too small. E.g.: SD of intercept = 0.14, SD of slope = 0.0004, SD residual=0.11. An lmList gives a slope SD of 0.07. I have about n=6 observations per group, and about 20-100 groups depending on the dataset. thank yo...
2013 Mar 24
3
Parallelizing GBM
...rallelization. I normally rely on gbm.fit for speed reasons, and I usually call it this way gbm_model <- gbm.fit(trainRF,prices_train, offset = NULL, misc = NULL, distribution = "multinomial", w = NULL, var.monotone = NULL, n.trees = 50, interaction.depth = 5, n.minobsinnode = 10, shrinkage = 0.001, bag.fraction = 0.5, nTrain = (n_train/2), keep.data = FALSE, verbose = TRUE, var.names = NULL, response.name = NULL) Does anybody know an easy way to parallelize the model (in this case it means simply having 4 cores on the same machine working on the problem)? Any suggestion is welco...
2010 Apr 26
3
R.GBM package
HI, Dear Greg, I AM A NEW to GBM package. Can boosting decision tree be implemented in 'gbm' package? Or 'gbm' can only be used for regression? IF can, DO I need to combine the rpart and gbm command? Thanks so much! -- Sincerely, Changbin -- [[alternative HTML version deleted]]