Displaying 20 results from an estimated 1000 matches similar to: "Rms package - problems with fit.mult.impute"
2008 Nov 26
1
multiple imputation with fit.mult.impute in Hmisc - how to replace NA with imputed value?
I am doing multiple imputation with Hmisc, and
can't figure out how to replace the NA values with
the imputed values.
Here's a general ourline of the process:
> set.seed(23)
> library("mice")
> library("Hmisc")
> library("Design")
> d <- read.table("DailyDataRaw_01.txt",header=T)
> length(d);length(d[,1])
[1] 43
[1] 2666
2011 Mar 31
2
fit.mult.impute() in Hmisc
I tried multiple imputation with aregImpute() and
fit.mult.impute() in Hmisc 3.8-3 (June 2010) and R-2.12.1.
The warning message below suggests that summary(f) of
fit.mult.impute() would only use the last imputed data set.
Thus, the whole imputation process is ignored.
"Not using a Design fitting function; summary(fit)
will use standard errors, t, P from last imputation only.
Use
2003 Jul 27
1
multiple imputation with fit.mult.impute in Hmisc
I have always avoided missing data by keeping my distance from
the real world. But I have a student who is doing a study of
real patients. We're trying to test regression models using
multiple imputation. We did the following (roughly):
f <- aregImpute(~ [list of 32 variables, separated by + signs],
n.impute=20, defaultLinear=T, data=t1)
# I read that 20 is better than the default of
2008 May 05
3
troubles with R CMD check and examples under Ubuntu gutsy
Dear listers,
I was used to package pgirmess under Windows with everything OK, but,
for the first time, I had a trial this afternoon on Ubuntu 7.10 gutsy (I
have a double boot computer and work more and more under unix) and R
2.7.0. Everything went OK except this:
sudo R CMD check pgirmess
.....
* checking examples ... ERROR
Running examples in 'pgirmess-Ex.R' failed.
The error most
2005 Jul 09
1
aregImpute: beginner's question
Hello R-help,
Thanks for everyone's very helpful suggestions so far. I am now trying to
use aregImpute for my missing data imputation. Here are the code and error
messages. Any suggestions would be very much appreciated.
Sincerely,
Anders Corr
########################################
#Question for R-Help on aregImpute
########################################
#DOWNLOAD DATA (61Kb)
2010 Nov 01
1
Error message in fit.mult.impute (Hmisc package)
Hello,
I would like to use the aregImpute and fit.mult.impute to impute missing
values for my dataset and then conduct logistic regression analyses on the
data, taking into account that we imputed values. I have no problems
imputing the values using aregImpute, but I am getting an error at the
fit.mult.impute stage.
Here is some sample code (I actually have more observations and variables to
2010 Aug 10
1
Multiple imputation, especially in rms/Hmisc packages
Hello, I have a general question about combining imputations as well as a
question specific to the rms and Hmisc packages.
The situation is multiple regression on a data set where multiple
imputation has been used to give M imputed data sets. I know how to get
the combined estimate of the covariance matrix of the estimated
coefficients (average the M covariance matrices from the individual
2007 May 21
2
Questions about bwplot
Dear R-experts,
I have some questions about boxplots with lattice.
My data is similar as in the example below, I have two factors
(Goodness of Fit and Algorithms) and data values but in each panels the scales are quite different, therefore the normal boxplots produced by
set.seed(1)
GOF <- factor(rep(c("GOF1","GOF2","GOF3"),each=40))
Alg <-
2004 Jun 15
1
fit.mult.impute and quantile regression
I have a largish dataset (1025) with around .15 of the data missing at random overall, but more like .25 in the dependent variable. I am interested in modelling the data using quantile regression, but do not know how to do this with multiply imputed data (which is what the dataset seems to need). The original plan was to use qr (or whatever) from the quantreg package as the 'fitter'
2012 Mar 08
2
xyplot without external box
Dear list members,
Within a loop, I need to create an xyplot with only a legend, not even
with the default external box drawn by lattice.
I already managed to remove the axis labels and tick marks, but I
couldn't find in the documentation of xyplot how to remove the
external box.
I would really appreciate any help with this
------------- START -----------
library(lattice)
x<-1:100
2010 Dec 02
1
problem with package rsm: running fit.mult.impute with cph
Hi all (and especially Frank),
I'm trying to use x=T, y=T in order to run a validated stepwise cox
regression in rsm, having multiply imputed using mice. I'm coding
model.max<-fit.mult.impute(baseform,cph,miced2,dated.sexrisk2,x=T,y=T)
baseform is
baseform<-Surv(si.age,si=="Yes")~ peer.press + copy.press + excited +
worried + intimate.friend + am.pill.times +
2010 Sep 01
1
[Q] Goodness-of-fit test of a logistic regression model using rms package
Hello,
I was looking for a way to evaluate the goodness-of-fit of a logistic regression model. After googling, I found that I could use "resid(fit, 'gof')" method implemented in the rms package. However, since I am not used to the "le Cessie-van Houwelingen normal test statistic," I do not know which statistic from the returned from the "resid(fit,
2008 Jun 05
1
(baseline) logistic regression + gof functions?
?
Hallo,
which function can i use to do (baseline) logistic regression +
goodness
of fit tests?
so far i found:
# logistic on binary data
lrm combined with resid(model,'gof')
# logistic on binary data
glm with no gof-test
# baseline logit on binary data
2010 Jul 07
1
Different goodness of fit tests leads to contradictory conclusions
I am trying to test goodness of fit for my legalistic regression using several options as shown below. Hosmer-Lemeshow test (whose function I borrowed from a previous post), Hosmer–le Cessie omnibus lack of fit test (also borrowed from a previous post), Pearson chi-square test, and deviance test. All the tests, except the deviance tests, produced p-values well above 0.05. Would anyone please
2011 Nov 20
2
ltm: Simplified approach to bootstrapping 2PL-Models?
Dear R-List,
to assess the model fit for 2PL-models, I tried to mimic the
bootstrap-approach chosen in the GoF.rasch()-function. Not being a
statistician, I was wondering whether the following simplification
(omit the "chi-squared-expressed model fit-step") would be appropriate:
GoF.ltm <- function(object, B = 50, ...){
liFits <- list()
for(i in 1:B){
rndDat <-
2008 Oct 15
1
Parameter estimates from an ANCOVA
Hi all,
This is probably going to come off as unnecessary (and show my ignorance)
but I am trying to understand the parameter estimates I am getting from R
when doing an ANCOVA. Basically, I am accustomed to the estimate for the
categorical variable being equivalent to the respective cell means minus the
grand mean. I know is the case in JMP - all other estimates from these data
match the
2013 Jan 06
4
random effects model
Hi A.K
Regarding my question on comparing normal/ obese/overweight with blood
pressure change, I did finally as per the first suggestion of stacking the
data and creating a normal category . This only gives me a obese not obese
14, but when I did with the wide format hoping to get a
obese14,normal14,overweight 14 Vs hibp 21, i could not complete any of the
models.
This time I classified obese=1
2009 Jun 11
2
How to order an data.table by values of an column?
Hello!
Can you help me? How to order an data.table by values of an column?
Per example:
Table no initial
Categ Perc
468 31.52
351 27.52
0 0.77
234 22.55
117 15.99
table final
Categ Perc
0 0.77
117 15.99
234 22.55
351 27.52
468 31.52
Lesandro
Veja quais são os assuntos do momento no Yahoo! +Buscados
[[alternative HTML version deleted]]
2007 May 18
4
Simple programming question
Hi R-users,
I have a simple question for R heavy users. If I have a data frame like this
dfr <- data.frame(id=1:16, categ=rep(LETTERS[1:4], 4),
var3=c(8,7,6,6,5,4,5,4,3,4,3,2,3,2,1,1))
dfr <- dfr[order(dfr$categ),]
and I want to score values or points in variable named "var3" following this
kind of logic:
1. the highest value of var3 within category (variable named
2010 Nov 09
1
Bootstrap confidence intervals using bootcov from the rms package
Hello,
I am using R.12.2.0. I am trying to generate bootstrap confidence intervals
using bootcov from the rms package. I am able to impute the missing data
using aregImpute and to perform a linear regression on the imputed datasets
using fit.mult.impute, but I am unable to use bootcov to generate the
confidence intervals for the R-squared. Here is a small example that should
duplicate the