similar to: lmer() with no intercept

Displaying 20 results from an estimated 500 matches similar to: "lmer() with no intercept"

2002 Jun 21
1
lme: anova vs. intervals
Windows 2000 (v.5.00.2195), R 1.5.1 I have an lme object, fm0, which I examine with anova() and intervals(). The anova output indicates one of the interaction terms is significant, but the intervals output shows that the single parameter for that term includes 0.0 in its 95% CI. I believe that the anova is a conditional (sequential) test; is intervals marginal or approximate? Which should I
2002 Jun 08
3
contour plot for non-linear models
Hello all, I've tried to reproduce the contour plot that appears in the book of Venables and Ripley, at page 255. Is a F-statistic surface and a confidence region for the regression parameters of a non-linear model. It uses the stormer data that are in the MASS package. I haven't been able to reproduce the plot either in R ( version 1.5 ) and S. It makes the axes and it puts the
2004 Jun 01
2
GLMM(..., family=binomial(link="cloglog"))?
I'm having trouble using binomial(link="cloglog") with GLMM in lme4, Version: 0.5-2, Date: 2004/03/11. The example in the Help file works fine, even simplified as follows: fm0 <- GLMM(immun~1, data=guImmun, family=binomial, random=~1|comm) However, for another application, I need binomial(link="cloglog"), and this generates an error for me: >
2004 Dec 31
4
R-intro
Hello! I was reading R-intro and I have some suggestions: R-intro.html#A-sample-session rm(fm, fm1, lrf, x, dummy) suggestion rm(fm, fm1, lrf, x, y, w, dummy) The next section will look at data from the classical experiment of Michaelson and Morley to measure the speed of light. file.show("morley.tab") mm <- read.table("morley.tab") suggestion mm <- data(morley)
2010 Dec 06
1
waldtest and nested models - poolability (parameter stability)
Dear All, I'm trying to use waldtest to test poolability (parameter stability) between two logistic regressions. Because I need to use robust standard errors (using sandwich), I cannot use anova. anova has no problems running the test, but waldtest does, indipendently of specifying vcov or not. waldtest does not appear to see that my models are nested. H0 in my case is the the vector of
2013 Jun 07
1
Function nlme::lme in Ubuntu (but not Win or OS X): "Non-positive definite approximate variance-covariance"
Dear all, I am estimating a mixed-model in Ubuntu Raring (13.04ΒΈ amd64), with the code: fm0 <- lme(rt ~ run + group * stim * cond, random=list( subj=pdSymm(~ 1 + run), subj=pdSymm(~ 0 + stim)), data=mydat1) When I check the approximate variance-covariance matrix, I get: > fm0$apVar [1] "Non-positive definite
2006 Dec 04
1
stepAIC for lmer
Dear All, I am trying to use stepAIC for an lmer object but it doesn't work. Here is an example: x1 <- gl(4,100) x2 <- gl(2,200) time <- rep(1:4,100) ID <- rep(1:100, each=4) Y <- runif(400) <=.5 levels(Y) <- c(1,0) dfr <- as.data.frame(cbind(ID,Y,time,x1,x2)) fm0.lmer <- lmer(Y ~ time+x1+x2 + (1|ID), data = dfr, family = binomial)
2012 Jan 17
1
MuMIn package, problem using model selection table from manually created list of models
The subject says it all really. Question 1. Here is some code created to illustrate my problem, can anyone spot where I'm going wrong? Question 2. The reason I'm following a manual specification of models relates to the fact that in reality I am using mgcv::gam, and I'm not aware that dredge is able to separate individual smooth terms out of say s(a,b). Hence an additional request,
2008 Aug 16
1
Pseudo R2 for Tobit Regression
Dear All: I need some guidance in calculating a goodness-of-fit statistic for a Tobit Regression model. To develop the Tobit regression, I used the tobit() method from the AER package, which is basically a simpler interface to the survreg() method. I've read about pseudo R2 and C-index and was wondering if there is a package that calculates this for me. Also, is there a reason to select
2007 Jun 06
1
Chow Test
Hello R-users! I tried to find a package to run a CHOW TEST. As a reference package I found the STRUCCHANGE package. Do you know if it works well otherwise can you recommend a different one? Thanks, Bernd -- View this message in context: http://www.nabble.com/Chow-Test-tf3878416.html#a10990270 Sent from the R help mailing list archive at Nabble.com.
2010 Jul 08
1
xyplot help
Hi, I am learning xyplot. I have an example dataset attached. plotdata<-read.table("plotdata.txt",sep='\t',header=T,row.names=1) head(plotdata,n=4) y x type 1 -4.309601 -0.7448405 A 2 -4.715421 0.7875994 A 3 -2.310638 0.5455310 A 4 -2.685803 10.4116868 A xyplot(y~x,groups=type,plotdata, auto.key=T) This shows different colors for different
2010 Jun 09
3
comparing two regression models with different dependent variable
Hi, I would like to compare to regression models - each model has a different dependent variable. The first model uses a number that represents the learning curve for reward. The second model uses a number that represents the learning curve from punishment stimuli. The first model is significant and the second isn't. I want to compare those two models and show that they are significantly
2011 Mar 16
5
Strange R squared, possible error
k=lm(y~x) summary(k) returns R^2=0.9994 lm(y~x) is supposed to find coef. a anb b in y=a*x+b l=lm(y~x+0) summary(l) returns R^2=0.9998 lm(y~x+0) is supposed to find coef. a in y=a*x+b while setting b=0 The question is why do I get better R^2, when it should be otherwise? Im sorry to use the word "MS exel" here, but I verified it in exel and it gives: R^2=0.9994 when y=a*x+b is used
2006 Nov 22
2
help
consider p as random effect with 5 levels, what is difference between these two models? > p5.random.p <- lmer(Y ~p+(1|p),data=p5,family=binomial,control=list(usePQL=FALSE,msV=1)) > p5.random.p1 <- lmer(Y ~1+(1|p),data=p5,family=binomial,control=list(usePQL=FALSE,msV=1)) thanks, Aimin Yan
2008 Mar 08
1
ask for help on nonlinear fitting
I have a table like the following. I want to fit Cm to Vm like this: Cm ~ Cl+Q1*b1*38.67*exp(-b1*(Vm-Vp1)*0.03867)/(1+exp(-b1*(Vm-Vp1)*0.03867))^2+Q2*b2*38.67*exp(-b2*(Vm-Vp2)*0.03867)/(1+exp(-b2*(Vm-Vp2)*0.03867))^2 I use nls, with start=list(Q1=2e-3, b1=1, Vp1=-25, Q2=3e-3, b2=1, Vp2=200). But I always get 'singlular gradient' error like this. But in SigmaPlot I can get the result. How
2010 May 18
1
BIC() in "stats" {was [R-sig-ME] how to extract the BIC value}
>>>>> "MM" == Martin Maechler <maechler at stat.math.ethz.ch> >>>>> on Tue, 18 May 2010 12:37:21 +0200 writes: >>>>> "GaGr" == Gabor Grothendieck <ggrothendieck at gmail.com> >>>>> on Mon, 17 May 2010 09:45:00 -0400 writes: GaGr> BIC seems like something that would logically go into stats
2009 Apr 15
2
AICs from lmer different with summary and anova
Dear R Helpers, I have noticed that when I use lmer to analyse data, the summary function gives different values for the AIC, BIC and log-likelihood compared with the anova function. Here is a sample program #make some data set.seed(1); datx=data.frame(array(runif(720),c(240,3),dimnames=list(NULL,c('x1','x2','y' )))) id=rep(1:120,2); datx=cbind(id,datx) #give x1 a
2017 Jun 17
3
Prediction with two fixed-effects - large number of IDs
Dear all, I am running a panel regression with time and location fixed effects: ### reg1 <- lm(lny ~ factor(id) + factor(year) + x1+ I(x1)^2 + x2+ I(x2)^2 , data=mydata, na.action="na.omit") ### My goal is to use the estimation for prediction. However, I have 8,500 IDs, which is resulting in very slow computation. Ideally, I would like to do the following: ### reg2 <-
2017 Jun 17
0
Prediction with two fixed-effects - large number of IDs
I have no direct experience with such horrific models, but your formula is a mess and Google suggests the biglm package with ffdf. Specifically, you should convert your discrete variables to factors before you build the model, particularly since you want to use predict after the fact, for which you will need a new data set with the exact same levels in the factors. Also, your use of I() is
2012 Nov 15
1
Stepwise regression scope: all interacting terms (.^2)
Dear Gurus, Thank you in advance for your assistance. I'm trying to understand scope better when performing stepwise regression using "step." I have a model with a binary response variable and 10 predictor variables. When I perform stepwise regression I define scope=.^2 to allow interactions between all terms. But I am missing something. When I perform stepwise regression (both