Displaying 20 results from an estimated 4000 matches similar to: "Bad points in regression [Broadcast]"
2003 Nov 14
1
What goodness-of-fit measure for robust regression ?
Hi,
i. After estimating some coefficients using robust regression with rlm() or lqs(), I wonder if there exist some measures of the goodness-of-fit as those for standard linear model(R2)... or evenly if it's a statistics non-sense to look for since I do not find any mention of that in differents chapters on robust and resistant regression or in severals R documentation (Fox, Ripley and
1998 Aug 31
0
Packages aov, modreg, lqs, psplines
I now have versions of code that is destined (I believe) for 0.63 which
is in a suitable state for comment. The files are at
ftp://ftp.stats.ox.ac.uk/pub/R
(Our www server is being moved, so may be intermittently down, but this
ftp server should be stable.) All are R packages, for the moment for
personal use only (no re-distribution). Use with 0.62.3 or 0.63 (although
I am aware of some
2003 Jul 30
2
robust regression
Hi,
trying to do a robudt regression of a two-way linear model, I keep
getting the following error:
> lqs(obs ~ y + s -1,method="lms", contrasts=list(s=("contr.sum")))
Error: lqs failed: all the samples were singular
Robust regression with M-estimators works (also regular least square
fits, of course):
rlm.formula(formula = obs ~ y + s - 1, method = "M",
2006 Feb 21
3
How to get around heteroscedasticity with non-linear leas t squares in R?
Your understanding isn't similar to mine. Mine says robust/resistant
methods are for data with heavy tails, not heteroscedasticity. The common
ways to approach heteroscedasticity are transformation and weighting. The
first is easy and usually quite effective for dose-response data. The
second is not much harder. Both can be done in R with nls().
Andy
From: Quin Wills
>
> I am
2012 Nov 22
1
help in M-estimator by R
hi guys and gals ... How are you all ...
i have to do something in robust regression by R programm , and i have some
problems as following:
*the first :*
suppose
w(r) =1/(1 r^2) and r <- c(7.01,2.07,7.061,5.607,8.502,54.909,12.222)
and i want to exclude some values from r so that (abs(r)>4.9 )...
after ,i want to used (w) to get on coefficients beta0 and beta1 (B1 <-
2006 Jun 22
1
High breakdown/efficiency statistics -- was RE: Rosner's test [Broadcast]
What would be nice is to have something like a "robust" task view...
Andy
From: Berton Gunter
>
> Many thanks for this Martin. There now are several packages
> with what appear to be overlapping functions (or at least
> algorithms). Besides those you mentioned, "robust" and
> "roblm" are at least two others. Any recommendations about
> how or
2005 Sep 06
0
MASS: rlm, MM and errors in observations AND regressors
Hello,
I need to perform a robust regression on data which contains errors in BOTH
observations and regressors. Right now I am using rlm from the MASS package
with 'method="MM"' and get visually very nice results. MASS is quite clear,
however, that the described methodologies are only applicable to
observation-error only data (p. 157, 4th Ed.). So here's the questions now:
2012 Jan 23
1
R not giving significance tests for coefficients/estimates?
> 3x4
Error: unexpected symbol in "3x4"
R has no idea that you equate "x" as multiplication.. use an astrix
> 3*4
[1] 12
dominic wrote
>
> This is basically my code:
>
> library(MASS)
> lmsreg(formula = b0 ~ b1 + b3 + b1xb2, data=mydata)
>
> b1xb2 is an interaction but it was the centered value for a continuous
> variable times a
2003 Dec 12
2
Reorganization of packages in the R distribution
After long but intermittent discussion (it was mentioned at DSC99, for
example), we have reorganized the standard packages, with
base graphics stats utils methods
normally loaded,
mle splines stepfun tcltk tools
available for loading, and
ctest eda lqs modreg mva nls ts
as stub packages which ensure back-compatibility. (These have all been
merged into stats except lqs which rejoins
2006 Feb 21
2
How to get around heteroscedasticity with non-linear least squares in R?
I am using "nls" to fit dose-response curves but am not sure how to approach
more robust regression in R to get around the problem of the my error
showing increased variance with increasing dose.
My understanding is that "rlm" or "lqs" would not be a good idea here.
'Fairly new to regression work, so apologies if I'm missing something
obvious.
2006 Jul 31
3
Soapbox
Hi all,
I thought y''all might be interested in seeing a newly released website
named Soapbox which was written in Rails.
Soapbox features reviews of products, businesses, services, and anything
else you can think of written by the people *you* care about.
http://soapboxit.com
Thanks!
Duff OMelia
--
Posted via http://www.ruby-forum.com/.
2011 Mar 14
1
discrepancy between lm and MASS:rlm
Dear R-devel,
There seems to be a discrepancy in the order in which lm and rlm evaluate their arguments. This causes rlm to sometimes produce an error where lm is just fine.
Here is a little script that illustrate the issue:
> library(MASS)
> ## create data
> n <- 100
> dat <- data.frame(x=rep(c(-1,0,1), n), y=rnorm(3*n))
>
> ## call lm, works fine
> summary(lm(y ~
2005 Mar 24
1
Robust multivariate regression with rlm
Dear Group,
I am having trouble with using rlm on multivariate data sets. When I
call rlm I get
Error in lm.wfit(x, y, w, method = "qr") :
incompatible dimensions
lm on the same data sets seem to work well (see code example). Am I
doing something wrong?
I have already browsed through the forums and google but could not find
any related discussions.
I use Windows XP and R
2008 Jan 11
0
Behaviour of standard error estimates in lmRob and the like
I am looking at MM-estimates for some interlab comparison work. The
usual situation in this particular context is a modest number of results
from very expensive methods with abnormally well-characterised
performance, so for once we have good "variance" estimates (which can
differ substantially for good reason) from most labs. But there remains
room for human error or unexpected chemistry
2012 Jul 06
1
How to do goodness-of-fit diagnosis and model checking for rlm in R?
Hi all,
I am reading the MASS book but it doesn't give examples about the diagnosis
and model checking for rlm...
My data is highly non-Gaussian so I am using rlm instead of lm.
My questions are:
0. Are goodness-of-fit and model-checking using rlm completely the same as
usual regression?
1.
Please give me some pointers about how to do goodness-of-fit and
residual diagnosis for
2004 Jun 11
1
comparing regression slopes
Dear List,
I used rlm to calculate two regression models for two data sets (rlm
due to two outlying values in one of the data sets). Now I want to
compare the two regression slopes. I came across some R-code of Spencer
Graves in reply to a similar problem:
http://www.mail-archive.com/r-help at stat.math.ethz.ch/msg06666.html
The code was:
> df1 <- data.frame(x=1:10, y=1:10+rnorm(10))
2010 Nov 08
1
Add values of rlm coefficients to xyplot
Hello,
I have a simple xyplot with rlm lines.
I would like to add the a and b coefficients (y=ax+b) of the rlm calculation
in each panel.
I know I can do it 'outside' the xyplot command but I would like to do all
at the same time.
I found some posts with the same question, but no answer.
Is it impossible ?
Thanks in advance for your help.
Ptit Bleu.
x11(15,12)
xyplot(df1$col2 ~
2003 Oct 02
4
using a string as the formula in rlm
Hi,
I am trying to build a series of rlm models. I have my data frame and
the models will be built using various coulmns of the data frame.
Thus a series of models would be
m1 <- rlm(V1 ~ V2 + V3 + V4, data)
m2 <- rlm(V1 ~ V2 + V5 + V7, data)
m3 <- rlm(V1 ~ V2 + V8 + V9, data)
I would like to automate this. Is it possible to use a string in place
of the formula?
I tried doing:
fmla
2003 Aug 04
0
Specifying weird models
[re-sending this one since it apparently didn't get through yesterday]
Hi Folks,
I'm pondering the following type of question in the context
of specifying a linear model formula. Basically, it's a matter
of specifying a "non-homogeneous" model. The following example
(not a real case, and over-simplified, but it illustrates the
point cleanly) shows what I mean.
There are 3
2010 Dec 13
1
Wrong contrast matrix for nested factors in lm(), rlm(), and lmRob()
This message also reports wrong estimates produced by lmRob.fit.compute()
for nested factors when using the correct contrast matrix.
And in these respects, I have found that S-Plus behaves the same way as R.
Using the three available contrast types (sum, treatment, helmert)
with lm() or lm.fit(), but just contr.sum with rlm() and lmRob(),
and small examples, I generated contrast matrices for