I'm trying to run bGLMM on a large dataset and it is failing after trying to create huge vectors with the error: Error: cannot allocate vector of size 553.7 Gb The dataset that I'm using is fairly large, around 1.5 million observations. The command is as follows: boost1 <- bGLMM(A~ B + C, rnd = list(D=~1), data = train, family =binomial) A is binary, B and C are continuous and D is a factor variable with 250,000 levels that I'm trying to model as a random effect. A similar model runs fine on lme4. Is my dataset simply much too large for this package or am I doing something obviously wrong? [[alternative HTML version deleted]]