similar to: random effect nested within fixed effects (binomial lmer)

Displaying 20 results from an estimated 4000 matches similar to: "random effect nested within fixed effects (binomial lmer)"

2007 Mar 23
1
lmer estimated scale
I have data consisting of binary responses from a large number of subjects on seven similar items. I have been using lmer with (crossed) random effects for subject and item. These effects are almost always (in the case of subject, always) significant additions to the model, testing this with anova. Including them also increases the Somers' Dxy value substantially. Even without those
2008 Mar 13
1
strange results from binomial lmer?
I'm running lmer repeatedly on artificial data with two fixed factors (called 'gender' and 'stress') and one random factor ('speaker'). Gender is a between-speaker variable, stress is a within-speaker variable, if that matters. Each dataset has 100 rows from each of 20 speakers, 2000 rows in all. About 5% of the time I get a strange result, where the lmer() model with
2011 Jun 13
1
Somers Dyx
Hello R Community, I'm continuing to work through logistic regression (thanks for all the help on score test) and have come up against a new opposition. I'm trying to compute Somers Dyx as some suggest this is the preferred method to Somers Dxy (Demaris, 1992). I have searchered the [R] archieves to no avail for a function or code to compute Dyx (not Dxy). The overview of Hmisc has
2009 Jul 15
0
Nagelkerkes R2N
I am interested Andrea is whether you ever established why your R2 was 1. I have had a similar situation previously. My main issue though, which I'd be v grateful for advice on, is why I am obtaining such negative values -0.3 for Somers Dxy using validate.cph from the Design package given my value of Nagelkerke R2 is not so low 13.2%. I have this output when fitting 6 variables all with
2011 May 05
1
Confidence interval for difference in Harrell's c statistics (or equivalently Somers' D statistics)
Dear All, I am trying to calculate a 95% confidence interval for the difference in two c statistics (or equivalently D statistics). In Stata I gather that this can be done using the lincom command. Is there anything similar in R? As you can see below I have two datasets (that are actually two independent subsets of the same data) and the respective c statistics for the variables in both cases.
2011 Sep 14
0
Confidence interval or p-value for difference in two c-statistics
Dear All, Apologies if you have a seen a question like this from me before. I am hoping that if I re-word my question more carefully someone may be able to offer more help than the last time I asked something similar. I am using R 2.9.2 and Windows XP. I am trying to determine if there is a statistically significant difference between two c-statistics (or equivalently D statistics). In Stata
2009 Jul 15
1
negative Somers D from Design package
Dear R help My problem is very similar to the analysis detailed here. If we use the mayo dataset provided with the survivalROC package the estimate for Somer's Dxy is very negative -0.56. The Nagelkerke R2 is positive though 0.32. I know there is a difference between explained variation and predictive ability but I am surprised there is usch a difference given that even a non predictive model
2003 Dec 04
2
Comparing Negative Binomial Regression in Stata and R. Constants differ?
I looked for examples of count data that might interest the students and found this project about dropout rates in Los Angeles High Schools. It is discussed in the UCLA stats help pages for the Stata users: http://www.ats.ucla.edu/stat/stata/library/count.htm and See: http://www.ats.ucla.edu/stat/stata/library/longutil.htm To replicate those results, I used R's excellent foreign package to
2011 Feb 19
0
contrasting Somer's D from Design package
Dear R help, I am having a problem with the Design package and my problem is detailed here. I fit a cox model to my data and validate the Somer's Dxy using the Design package. (Because of computation time problem, i only try 10 bootstrap samples for the time being) This is the model without stratification: > library(Design) >
2008 Jul 19
0
fixed effect significance with lmer() vs. t-test
I am looking at data of the following structure: n <- 100 dataset <- data.frame(gender=NULL,subject=NULL,outcome=NULL) for (i in 1:n){ gender <- c(rep("m",5),rep("f",5)) subject <- letters[1:10] outcome <- c(rbinom(5,1,0.6),rbinom(5,1,0.4)) dataset <- rbind(dataset,cbind(gender,subject,outcome))} I am interested in the significance of
2011 Feb 21
2
Interpreting the example given by Prof Frank Harrell in {Design} validate.cph
Dear R-help, I am having a problem with the interpretation of result from validate.cph in the Design package. My purpose is to fit a cox model and validate the Somer's Dxy. I used the hypothetical data given in the help manual with modification to the cox model fit. My research problem is very similar to this example. This is the model without stratification: > library(Design) > f1
2012 Dec 10
3
Warning message: In eval(expr, envir, enclos) : non-integer #successes in a binomial glm!
Hi there I'm trying to fit a logistic regression model to data that looks very similar to the data in the sample below. I don't understand why I'm getting this error; none of the data are proportional and the weights are numeric values. Should I be concerned about the warning about non-integer successes in my binomial glm? If I should be, how do I go about addressing it? I'm
2008 Apr 10
1
Degrees of freedom in binomial glm
Hello, I am looking at the job satisfaction data below, from a problem in Agresti's book, and I am not sure where the degrees of freedom come from. The way I am fitting a binomial model, I have 168 observations, so in my understanding that should also be the number of fitted parameters in the saturated model. Since I have one intercept parameter, I was thinking to get 167 df for the Null
2007 May 14
2
lmer function
Does anyone know if the lmer function of lme4 works fine for unbalanced designs? I have the examination results of 1000 pupils on three subjects, one score every term. So, I have three scores for English (one for every term), three scores for maths etc. However, not everybody was examined in maths, not everybody was examined in English etc, but everybody was in effect examined on four subjects. I
2011 Dec 19
2
summary vs anova
Hi, I'm sure this is simple, but I haven't been able to find this in TFM, say I have some data in R like this (pasted here: http://pastebin.com/raw.php?i=sjS9Zkup): > head(df) gender age smokes disease Y 1 female 65 ever control 0.18 2 female 77 never control 0.12 3 male 40 state1 0.11 4 female 67 ever control 0.20 5 male 63 ever state1 0.16
2011 Feb 16
1
Saturated model in binomial glm
Hi all, Could somebody be so kind to explain to me what is the saturated model on which deviance and degrees of freedom are calculated when fitting a binomial glm? Everything makes sense if I fit the model using as response a vector of proportions or a two-column matrix. But when the response is a factor and counts are specified via the "weights" argument, I am kind of lost as far as
2008 Jul 28
1
Negative Binomial Regression
Hello. I am attempting to duplicate a negative binomial regression in R. SAS uses generalized estimating equations for model fitting in the GENMOD procedure. proc genmod data=mydata (where=(gender='F')); by agegroup; class id gender type; model count = var1 var2 var3 /dist=NB link=log offset=lregtm; repeated subject=id /type=exch; run; Since my dataset has several observations for
2003 Apr 25
1
validate function in Design library does not work with small samples
Hi, I am using the validate function in the design library to get corrected Somer's Dxy for cox ph models. When my sample size is reduced from 300 to 150, the function complains (length of dimnames[1] not equal to array) and does not produce any results. There are no missing values in the data. Any suggestions for a work-around? Thank you in Advance. >
2011 Mar 01
1
which does the "S.D." returned by {Hmisc} rcorr.cens measure?
Dear R-help, This is an example in the {Hmisc} manual under rcorr.cens function: > set.seed(1) > x <- round(rnorm(200)) > y <- rnorm(200) > round(rcorr.cens(x, y, outx=F),4) C Index Dxy S.D. n missing uncensored Relevant Pairs Concordant Uncertain 0.4831 -0.0338 0.0462 200.0000
2012 May 04
2
Binomial GLM, chisq.test, or?
Hi, I have a data set with 999 observations, for each of them I have data on four variables: site, colony, gender (quite a few NA values), and cohort. This is how the data set looks like: > str(dispersal) 'data.frame': 999 obs. of 4 variables: $ site : Factor w/ 2 levels "1","2": 1 1 1 1 1 1 1 1 2 2 ... $ gender: Factor w/ 2 levels "0","1":