similar to: colinearity?

Displaying 20 results from an estimated 11000 matches similar to: "colinearity?"

2006 Jul 05
2
Colinearity Function in R
Is there a colinearty function implemented in R? I have tried help.search("colinearity") and help.search("collinearity") and have searched for "colinearity" and "collinearity" on http://www.rpad.org/Rpad/Rpad-refcard.pdf but with no success. Many thanks in advance, Peter Lauren.
2005 Apr 11
2
dealing with multicollinearity
I have a linear model y~x1+x2 of some data where the coefficient for x1 is higher than I would have expected from theory (0.7 vs 0.88) I wondered whether this would be an artifact due to x1 and x2 being correlated despite that the variance inflation factor is not too high (1.065): I used perturbation analysis to evaluate collinearity library(perturb)
2012 Nov 06
2
R and SPSS
Hi group: I have a data set, which has severe colinearity problem. While running linear regression in R and SPSS, I got different models. I am wondering if somebody knows how to make the two software output the same results. (I guess the way R and SPSS handling singularity is different, which leads to different models.) Thanks. [[alternative HTML version deleted]]
2012 Nov 10
1
colineraity among categorical variables (multinom)
Dear all users, I"d like to ask you how to make decision about colinearity among categorical independent variables when the model is multinomial logistic regression. Any help is appreciated, Niklas [[alternative HTML version deleted]]
2005 Sep 13
3
Collineariy Diagnostics
Hi, and thanks for your help in order to do collinearity analysis I downloaded the perturb package. I run a lm (regression) and on that the ??calldiag?? commad to get condition numbers but i get the following message: the variable XY with modus ??numeric?? was not found (it does the same with all predictors despite all variables are numeric and exists). Can anyone tell me how can I go arround
2004 Jun 30
1
linear models and colinear variables...
Hi! I'm having some issues on both conceptual and technical levels for selecting the right combination of variables for this model I'm working on. The basic, all inclusive form looks like lm(mic ~ B * D * S * U * V * ICU) Where mic, U, V, and ICU are numeric values and B D and S are factors with about 16, 16 and 2 levels respectively. In short, there's a ton of actual explanatory
2008 May 28
1
Fixing the coefficient of a regressor in formula
Dear R users, I want to estimate a Cox PH model with time-dependent covariates so I am using a counting process format with the following formula: Surv(data$start, data$stop, data$event.time) ~ cluster(data$id) + G1 + G2 + G3 + G4 + G5 +G6 Gs represent a B-spline basis functions so they sum to 1 and I can't estimate the model as is without getting the last coefficient to be NA, which
2009 Mar 12
2
MANOVA
Hi All, I have questions about MANOVA which I am still not sure if appropriately I should use it. For example I have a data set like this: BloodPressure (BP) Weight Height 120 115 165 125 145 198 156 99 176 I know that BloodPressure is correlated with both Weight and Height, however colinearity exists between Weight and Height. When I use BP = Weight + Height
2004 Aug 16
2
mutlicollinearity and MM-regression
Dear R users, Usually the variance-inflation factor, which is based on R^2, is used as a measure for multicollinearity. But, in contrast to OLS regression there is no robust R^2 available for MM-regressions in R. Do you know if an equivalent or an alternative nmeasure of multicollinearity is available for MM-regression in R? With best regards, Carsten Colombier Dr. Carsten Colombier Economist
2010 Jan 07
1
logistic regression based on principle component analysis
Dear all: I try to analyse a dataset which contain one binary response variable and serveral predict variables, but multiple colinear problem exists in my dataset, some paper suggest that logistic regression for principle components is suit for these noise data, but i only find R can done principle component regression using "pls" package, is there any package that can do the task i
2008 Mar 06
0
Help with colinearity problem in multiple linear regression
Hello, For basic linear regression lm() does the job well, for datasets that are larger than memory biglm() seems to work. I'm working on a parallel implementation of multiple linear regression for datasets that are too large for memory. Currently I am working over least squares: calculating: t(X) %*% X and t(X) %*% y separately in parallel on each node This generates a
2008 Nov 20
1
Checking collinearity using lmer
I am running a logistic regression model with a random effect using lmer. I am uncertain how to check for collinearity between my parameters. I have already run cor() and linear regression for each combination of parameters, and all Rsqr values were <0.8….but I am analyzing ecological data so a 0.8 cutoff may be unrealistic. -is there a way to check variance inflation factors or tolerance
2005 Jul 25
5
passing formula arguments cv.glm
I am trying to write a wrapper for the last example in help(cv.glm) that deals with leave-one-out-cross-validation (LOOCV) for a logistic model. This wrapper will be used as part of a bigger program. Here is my wrapper funtion : logistic.LOOCV.err <- function( formu=NULL, data=NULL ){ cost.fn <- function(cl, pred) mean( abs(cl-pred) > 0.5 ) glmfit <- glm(
2012 Apr 03
1
how to use condition indexes to test multi-collinearity
Dear Users, I try to calculate condition indexes and variance decomposition proportions in order to test for collinearity using colldiag() in perturb package, I got a large index and two variables with large variance decomposition proportions,but one of them is constant item.I also checked the VIF for that variable, the value is small.The result is as follows: Index intercept V1
2006 Nov 28
3
Predicted values in lmer modeling
Dear All, I am working with linear mixed-effects models using the lme4 package in R. I created a model with the lmer function including some main effects, a two-way interaction and a random effect. Now I am searching for a way to save the predicted values for this model. As far as I can see, there is no command in lme4 to save the predicted values (like the predict(model) function in e.g.
2006 Sep 15
2
prediction interval for new value
Hi, 1. How do I construct 95% prediction interval for new x values, for example - x = 30000? 2. How do I construct 95% confidence interval? my dataframe is as follows : >dt structure(list(y = c(26100000, 60500000, 16200000, 30700000, 70100000, 57700000, 46700000, 8600000, 10000000, 61800000, 30200000, 52200000, 71900000, 55000000, 12700000 ), x = c(108000, 136000,
2006 Nov 26
1
plot p(Y=1) vs as
I am trying to fit a logistic regression model for this data set. Firstly, I want to plot P(Y=1) vs As and P(Y=1) vs Aa. Does any body know how to do these in R. Thanks, Aimin > p5 <- read.csv("http://www.public.iastate.edu/~aiminy/data/p_5_2.csv") > str(p5) 'data.frame': 1030 obs. of 6 variables: $ P : Factor w/ 5 levels "821p","8ABP",..: 1
2007 Dec 18
1
How can I extract the AIC score from a mixed model object produced using lmer?
I am running a series of candidate mixed models using lmer (package lme4) and I'd like to be able to compile a list of the AIC scores for those models so that I can quickly summarize and rank the models by AIC. When I do logistic regression, I can easily generate this kind of list by creating the model objects using glm, and doing: > md <- c("md1.lr", "md2.lr",
2006 Oct 19
2
How to get multiple Correlation Coefficients
Hi I have used a polycor package for categorical correlation coefficients. I run the following script. But there were no results. Could you tell me how to correct the script? Thanks in advance, vars <- names(sdi) for (i in 1:length(vars)) { for (j in 1:length(vars)) { paste(vars[i]," and ", vars[j]) polychor(vars[i], vars[j]) # corr } } -- Kum-Hoe Hwang, Ph.D.Phone :
2006 Oct 22
2
"glm" function question
I am creating a model attempting to predict the probability someone will reoffend after being caught for a crime. There are seven total inputs and I planned on using a logistic regression. I started with a null deviance of 182.91 and ended up with a residual deviance of 83.40 after accounting for different interactions and such. However, I realized after that my code is different from that in