similar to: Bad points in regression [Broadcast]

Displaying 20 results from an estimated 4000 matches similar to: "Bad points in regression [Broadcast]"

2003 Nov 14
1
What goodness-of-fit measure for robust regression ?
Hi, i. After estimating some coefficients using robust regression with rlm() or lqs(), I wonder if there exist some measures of the goodness-of-fit as those for standard linear model(R2)... or evenly if it's a statistics non-sense to look for since I do not find any mention of that in differents chapters on robust and resistant regression or in severals R documentation (Fox, Ripley and
1998 Aug 31
0
Packages aov, modreg, lqs, psplines
I now have versions of code that is destined (I believe) for 0.63 which is in a suitable state for comment. The files are at ftp://ftp.stats.ox.ac.uk/pub/R (Our www server is being moved, so may be intermittently down, but this ftp server should be stable.) All are R packages, for the moment for personal use only (no re-distribution). Use with 0.62.3 or 0.63 (although I am aware of some
2003 Jul 30
2
robust regression
Hi, trying to do a robudt regression of a two-way linear model, I keep getting the following error: > lqs(obs ~ y + s -1,method="lms", contrasts=list(s=("contr.sum"))) Error: lqs failed: all the samples were singular Robust regression with M-estimators works (also regular least square fits, of course): rlm.formula(formula = obs ~ y + s - 1, method = "M",
2006 Feb 21
3
How to get around heteroscedasticity with non-linear leas t squares in R?
Your understanding isn't similar to mine. Mine says robust/resistant methods are for data with heavy tails, not heteroscedasticity. The common ways to approach heteroscedasticity are transformation and weighting. The first is easy and usually quite effective for dose-response data. The second is not much harder. Both can be done in R with nls(). Andy From: Quin Wills > > I am
2012 Nov 22
1
help in M-estimator by R
hi guys and gals ... How are you all ... i have to do something in robust regression by R programm , and i have some problems as following: *the first :* suppose w(r) =1/(1 r^2) and r <- c(7.01,2.07,7.061,5.607,8.502,54.909,12.222) and i want to exclude some values from r so that (abs(r)>4.9 )... after ,i want to used (w) to get on coefficients beta0 and beta1 (B1 <-
2006 Jun 22
1
High breakdown/efficiency statistics -- was RE: Rosner's test [Broadcast]
What would be nice is to have something like a "robust" task view... Andy From: Berton Gunter > > Many thanks for this Martin. There now are several packages > with what appear to be overlapping functions (or at least > algorithms). Besides those you mentioned, "robust" and > "roblm" are at least two others. Any recommendations about > how or
2005 Sep 06
0
MASS: rlm, MM and errors in observations AND regressors
Hello, I need to perform a robust regression on data which contains errors in BOTH observations and regressors. Right now I am using rlm from the MASS package with 'method="MM"' and get visually very nice results. MASS is quite clear, however, that the described methodologies are only applicable to observation-error only data (p. 157, 4th Ed.). So here's the questions now:
2012 Jan 23
1
R not giving significance tests for coefficients/estimates?
> 3x4 Error: unexpected symbol in "3x4" R has no idea that you equate "x" as multiplication.. use an astrix > 3*4 [1] 12 dominic wrote > > This is basically my code: > > library(MASS) > lmsreg(formula = b0 ~ b1 + b3 + b1xb2, data=mydata) > > b1xb2 is an interaction but it was the centered value for a continuous > variable times a
2003 Dec 12
2
Reorganization of packages in the R distribution
After long but intermittent discussion (it was mentioned at DSC99, for example), we have reorganized the standard packages, with base graphics stats utils methods normally loaded, mle splines stepfun tcltk tools available for loading, and ctest eda lqs modreg mva nls ts as stub packages which ensure back-compatibility. (These have all been merged into stats except lqs which rejoins
2006 Feb 21
2
How to get around heteroscedasticity with non-linear least squares in R?
I am using "nls" to fit dose-response curves but am not sure how to approach more robust regression in R to get around the problem of the my error showing increased variance with increasing dose. My understanding is that "rlm" or "lqs" would not be a good idea here. 'Fairly new to regression work, so apologies if I'm missing something obvious.
2006 Jul 31
3
Soapbox
Hi all, I thought y''all might be interested in seeing a newly released website named Soapbox which was written in Rails. Soapbox features reviews of products, businesses, services, and anything else you can think of written by the people *you* care about. http://soapboxit.com Thanks! Duff OMelia -- Posted via http://www.ruby-forum.com/.
2011 Mar 14
1
discrepancy between lm and MASS:rlm
Dear R-devel, There seems to be a discrepancy in the order in which lm and rlm evaluate their arguments. This causes rlm to sometimes produce an error where lm is just fine. Here is a little script that illustrate the issue: > library(MASS) > ## create data > n <- 100 > dat <- data.frame(x=rep(c(-1,0,1), n), y=rnorm(3*n)) > > ## call lm, works fine > summary(lm(y ~
2005 Mar 24
1
Robust multivariate regression with rlm
Dear Group, I am having trouble with using rlm on multivariate data sets. When I call rlm I get Error in lm.wfit(x, y, w, method = "qr") : incompatible dimensions lm on the same data sets seem to work well (see code example). Am I doing something wrong? I have already browsed through the forums and google but could not find any related discussions. I use Windows XP and R
2008 Jan 11
0
Behaviour of standard error estimates in lmRob and the like
I am looking at MM-estimates for some interlab comparison work. The usual situation in this particular context is a modest number of results from very expensive methods with abnormally well-characterised performance, so for once we have good "variance" estimates (which can differ substantially for good reason) from most labs. But there remains room for human error or unexpected chemistry
2012 Jul 06
1
How to do goodness-of-fit diagnosis and model checking for rlm in R?
Hi all, I am reading the MASS book but it doesn't give examples about the diagnosis and model checking for rlm... My data is highly non-Gaussian so I am using rlm instead of lm. My questions are: 0. Are goodness-of-fit and model-checking using rlm completely the same as usual regression? 1. Please give me some pointers about how to do goodness-of-fit and residual diagnosis for
2004 Jun 11
1
comparing regression slopes
Dear List, I used rlm to calculate two regression models for two data sets (rlm due to two outlying values in one of the data sets). Now I want to compare the two regression slopes. I came across some R-code of Spencer Graves in reply to a similar problem: http://www.mail-archive.com/r-help at stat.math.ethz.ch/msg06666.html The code was: > df1 <- data.frame(x=1:10, y=1:10+rnorm(10))
2010 Nov 08
1
Add values of rlm coefficients to xyplot
Hello, I have a simple xyplot with rlm lines. I would like to add the a and b coefficients (y=ax+b) of the rlm calculation in each panel. I know I can do it 'outside' the xyplot command but I would like to do all at the same time. I found some posts with the same question, but no answer. Is it impossible ? Thanks in advance for your help. Ptit Bleu. x11(15,12) xyplot(df1$col2 ~
2003 Oct 02
4
using a string as the formula in rlm
Hi, I am trying to build a series of rlm models. I have my data frame and the models will be built using various coulmns of the data frame. Thus a series of models would be m1 <- rlm(V1 ~ V2 + V3 + V4, data) m2 <- rlm(V1 ~ V2 + V5 + V7, data) m3 <- rlm(V1 ~ V2 + V8 + V9, data) I would like to automate this. Is it possible to use a string in place of the formula? I tried doing: fmla
2003 Aug 04
0
Specifying weird models
[re-sending this one since it apparently didn't get through yesterday] Hi Folks, I'm pondering the following type of question in the context of specifying a linear model formula. Basically, it's a matter of specifying a "non-homogeneous" model. The following example (not a real case, and over-simplified, but it illustrates the point cleanly) shows what I mean. There are 3
2010 Dec 13
1
Wrong contrast matrix for nested factors in lm(), rlm(), and lmRob()
This message also reports wrong estimates produced by lmRob.fit.compute() for nested factors when using the correct contrast matrix. And in these respects, I have found that S-Plus behaves the same way as R. Using the three available contrast types (sum, treatment, helmert) with lm() or lm.fit(), but just contr.sum with rlm() and lmRob(), and small examples, I generated contrast matrices for