Displaying 20 results from an estimated 600 matches similar to: "dfbetas vs dfbeta"
2011 Apr 29
1
logistic regression with glm: cooks distance and dfbetas are different compared to SPSS output
Hi there,
I have the problem, that I'm not able to reproduce the SPSS residual
statistics (dfbeta and cook's distance) with a simple binary logistic
regression model obtained in R via the glm-function.
I tried the following:
fit <- glm(y ~ x1 + x2 + x3, data, family=binomial)
cooks.distance(fit)
dfbetas(fit)
When i compare the returned values with the values that I get in SPSS,
2003 Jun 12
1
What PRECISELY is the dfbetas() or lm.influence()$coef ?
Hello. I want to get the proper influence function for the glm
coefficients in R. This is supposed to be inv(information)*(y-yhat)*x. So
I am wondering what is the exact mathematical formula for the output that
the functions:
dfbeta() OR lm.influence()$coefficients
return for a glm model. I am confused because:
1. Their columns don't sum to zero as influences should.
2. They
2003 Jul 12
1
Problem with library "car"
I am using the Unix version of R (version 1.7.0), installed via fink on a G4
Macintosh. I recently upgraded from version 1.6.0 and found that the "car"
library now has a problem:
---Begin transcript---
>library(car)
Attaching package 'car':
The following object(s) are masked from package:base :
dfbeta dfbeta.lm dfbetas dfbetas.lm hatvalues hatvalues.lm
2009 Jan 14
1
dfbetas without intercept
Hello
I am running a regression without the intercept, and want to compute dfbetas. How do I do this? The dfbetas function only works when the intercept is included in the model.
Regards
K
[[alternative HTML version deleted]]
2001 Apr 28
9
two new packages
I've prepared preliminary versions of two packages that I plan eventually
to contribute to CRAN:
car (for "Companion to Applied Regression") is a package that provides a
variety of functions in support of linear and generalized linear models,
including regression diagnostics (e.g., studentized residuals, hat-values,
Cook's distances, dfbeta, dfbetas, added-variable plots,
2001 Apr 28
9
two new packages
I've prepared preliminary versions of two packages that I plan eventually
to contribute to CRAN:
car (for "Companion to Applied Regression") is a package that provides a
variety of functions in support of linear and generalized linear models,
including regression diagnostics (e.g., studentized residuals, hat-values,
Cook's distances, dfbeta, dfbetas, added-variable plots,
2001 Apr 28
9
two new packages
I've prepared preliminary versions of two packages that I plan eventually
to contribute to CRAN:
car (for "Companion to Applied Regression") is a package that provides a
variety of functions in support of linear and generalized linear models,
including regression diagnostics (e.g., studentized residuals, hat-values,
Cook's distances, dfbeta, dfbetas, added-variable plots,
2013 May 01
1
Trouble with methods() after loading gdata package.
Greetings to r-help land.
I've run into some program crashes and I've traced them back to methods()
behavior
after the package gdata is loaded. I provide now a minimal re-producible
example. This seems bugish to me. How about you?
dat <- data.frame(x = rnorm(100), y = rnorm(100))
lm1 <- lm(y ~ x, data = dat)
methods(class = "lm")
## OK so far
library(gdata)
2010 Aug 02
7
Linux Editor
Hi Everyone,
I recently have started using R again on a Linux box after spending several
years on a Mac. Last I checked, the best way to use R was through EMACS
using something like ESS. I remember that being serviceable but not always
the most convenient.
Is there anything comparable to the mac version of R with its built in
console, editor, etc??
thanks!
--
View this message in context:
2010 Feb 21
1
tests for measures of influence in regression
influence.measures gives several measures of influence for each
observation (Cook's Distance, etc) and actually flags observations
that it determines are influential by any of the measures. Looks
good! But how does it discriminate between the influential and non-
influential observations by each of the measures? Like does it do a
Bonferroni-corrected t on the residuals identified by
2008 May 07
1
coxph - weights- robust SE
Hi,
I am using coxph with weights to represent sampling fraction of subjects.
Our simulation results show that the robust SE of beta systematically
under-estimate the empirical SD of beta.
Does anyone know how the robust SE are estimated in coxph using weights?
Is there any analytical formula for the “weighted” robust SE?
Any help is appreciated!
Thanks so much in advance
Willy
1999 Oct 21
1
left.solve
I have sort of an emergency question for the list. One of my professors
for an S-Plus intensive class distributed a function to produce partial
regression plots. I need to run it under R, because I'm doing the
homework on my home computer with a modem; hence I don't have the speed
required to emulate X-Windows and run S Plus off one of the campus
servers. Bottom line: I'm using R.
2017 Apr 04
0
Some "lm" methods give wrong results when applied to "mlm" objects
I had a look at some influence measures, and it seems to me that currently several methods handle multiple lm (mlm) objects wrongly in R. In some cases there are separate "mlm" methods, but usually "mlm" objects are handled by the same methods as univariate "lm" methods, and in some cases this fails.
There are two general patterns of problems in influence measures:
2005 Jun 27
1
delta-beta's
Hi there
I have created a multivariate logistic regression model looking at the
presence/absence of disease on farms. I would like to plot the diagnostic
plots recommended by Hosmer & Lemeshow to look particularly for any points of
high influence. In order to do this I need to extract values for delta-beta.
The function dfbeta gives a value for change in each coefficient but I am
looking
2009 Oct 30
1
Package zelig
hello all
I am using the R package Zelig for some tobit regression with robust
standard errors.
I have got R version 2.9.2 (2009-08-24)
and Zelig Version: 3.4-5
when i do demo(robust)
It ends like this way
data(coalition)
> # Fit the model with robust standard error
> user.prompt()
Press <return> to continue:
> z.out3 <- zelig(Surv(duration, ciep12) ~ polar + numst2 +
2012 Feb 09
1
passing an extra argument to an S3 generic
I'm trying to write some functions extending influence measures to
multivariate linear models and also
allow subsets of size m>=1 to be considered for deletion diagnostics.
I'd like these to work roughly parallel
to those functions for the univariate lm where only single case deletion
(m=1) diagnostics are considered.
Corresponding to stats::hatvalues.lm, the S3 method for class
2009 Mar 25
3
very fast OLS regression?
Dear R experts:
I just tried some simple test that told me that hand computing the OLS
coefficients is about 3-10 times as fast as using the built-in lm()
function. (code included below.) Most of the time, I do not care,
because I like the convenience, and I presume some of the time goes
into saving a lot of stuff that I may or may not need. But when I do
want to learn the properties of an
2016 Apr 26
0
survival::clogit, how to extract residuals for GOF assessment
Hi Folks,
Hopefully this question has enough R and not too much stats to be
appropriate for this list. Based on,* Hosmer et al. 2013. Logistic
regression for matched case-control studies. Applied Logistic
Regression *(eqtn.
7.8)*, *I am assessing GOF of conditional (or matched) logistic regression
models with the *standardized Pearson residuals*. The authors define
?large? as delta chi-squared
2008 Sep 30
2
weird behavior of drop1() for polr models (MASS)
I would like to do a SS type III analysis on a proportional odds logistic
regression model. I use drop1(), but dropterm() shows the same behaviour. It
works as expected for regular main effects models, however when the model
includes an interaction effect it seems to have problems with matching the
parameters to the predictor terms. An example:
library("MASS");
options(contrasts =
2005 Apr 13
2
multinom and contrasts
Hi,
I found that using different contrasts (e.g.
contr.helmert vs. contr.treatment) will generate
different fitted probabilities from multinomial
logistic regression using multinom(); while the fitted
probabilities from binary logistic regression seem to
be the same. Why is that? and for multinomial logisitc
regression, what contrast should be used? I guess it's
helmert?
here is an example