similar to: AIC model selection

Displaying 20 results from an estimated 4000 matches similar to: "AIC model selection"

2007 Jan 12
1
R2WinBugs and Compare DIC versus BIC or AIC
Dear All 1) I'm fitting spatial CAR models using R2Winbugs and although everything seems to go reasonably well (or I think so) the next message appears from WINBUGS 1.4 window: gen.inits() Command #Bugs: gen.inits cannot be executed (is greyed out) The question is if this message means that something is wrong and the results are consequently wrong, or Can I assume it as a simple warning
2006 Feb 20
1
Nested AIC
Greetings, I have recently come into some confusion over weather or not AIC results for comparing among models requires that they be nested. Reading Burnham & Anderson (2002) they are explicit that nested models are not required, but other respected statisticians have suggested that nesting is a pre-requisite for comparison. Could anyone who feels strongly regarding either position
2007 Feb 13
1
lag orders with ADF.test
Hello! I do not understand what is meant by: "aic" and "bic" follow a top-down strategy based on the Akaike's and Schwarz's information criteria in the datails to the ADF.test function. What does a "top-down strategy" mean? Probably the respective criterion is minimized and the mode vector contains the lag orders at which the criterion attains it
2012 Jul 06
1
Definition of AIC (Akaike information criterion) for normal error models
Dear R users (r-help@r-project.org), The definition of AIC (Akaike information criterion) for normal error models has just been changed. Please refer to the paper below on this matter. Eq.(22) is the new definition. The essential part is RSS(n+q+1)/(n-q-3); it is close to GCV. The paper is temporarily available at the "Papers In Press" place. Kunio Takezawa(2012): A Revision of
2008 Mar 11
1
Problem comparing Akaike's AIC - nlme package
Hello, I am comparing models made with nlme functions and non-nlme functions, based on Akaike's AIC. The AIC values I get for exactly the same model formulation --for example a linear model with no random effects fit with gls and lm, respectively-- do not fit, although the values of the four model parameters are exactly the same. For example: m1 <- gls(height ~ age, data = Loblolly) m2
2006 Jun 23
1
Bug in R-intro.html ? (PR#9028)
Full_Name: Ommo H?ppop Version: 2.0.1 OS: XP Submission from: (NULL) (84.143.196.187) Hi, Presumably, I've found an error in http://finzi.psych.upenn.edu/R/doc/manual/R-intro.html In Chapter 11.3 Generic functions for extracting model information it says "Select a suitable model by adding or dropping terms and preserving hierarchies. The model with the LARGEST value of AIC
2013 May 21
1
Calculating AIC for the whole model in VAR
Hello! I am using package "VAR". I've fitted my model: mymodel<-VAR(mydata,myp,type="const") I can extract the Log Liklihood for THE WHOLE MODEL: logLik(mymodel) How could I calculate (other than manually) the corresponding Akaike Information Criterion (AIC)? I tried AIC - but it does not take mymodel: AIC(mymodel) # numeric(0) Thank you! -- Dimitri Liakhovitski
2001 Sep 13
2
akaike's information criterion
Hello all, i hope you don't mind my off topic question. i want to use the Akaike criterion for variable selection in a regression model. Does anyone know some basic literature about that topic? Especially I'm interested in answers to the following questions: 1. Has (and if so how has) the criterion to be modified, if i estimate the transformations of the variables too? 2. How is the
2010 Apr 02
2
Cross-validation for parameter selection (glm/logit)
If my aim is to select a good subset of parameters for my final logit model built using glm(). What is the best way to cross-validate the results so that they are reliable? Let's say that I have a large dataset of 1000's of observations. I split this data into two groups, one that I use for training and another for validation. First I use the training set to build a model, and the the
2008 Nov 28
2
AIC function and Step function
I would like to figure out the equations for calculating "AIC" in both "step() function" and "AIC () function". They are different. Then I just type "step" in the R console, and found the "AIC" used in "step() function" is "extractAIC". I went to the R help, and found: "The criterion used is AIC = - 2*log L + k *
2007 Aug 03
3
question about logistic models (AIC)
Een ingesloten tekst met niet-gespecificeerde tekenset is van het bericht gescrubt ... Naam: niet beschikbaar Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070803/79b6292b/attachment.pl
2008 Feb 26
2
AIC and anova, lme
Dear listers, Here we have a strange result we can hardly cope with. We want to compare a null mixed model with a mixed model with one independent variable. > lmmedt1<-lme(mediane~1, random=~1|site, na.action=na.omit, data=bdd2) > lmmedt9<-lme(mediane~log(0.0001+transat), random=~1|site, na.action=na.omit, data=bdd2) Using the Akaike Criterion and selMod of the package pgirmess
2007 Sep 07
1
negative value for AIC and BIC
Hi all, I obtained negative values for AIC and BIC criteria for a particular model that I have developped... I don't remember to have negative values for these crietria for others applications, so I am a little suprised... Could anyone tell me if something is wrong or his conclusion concerning my model? Best regards, Olivier.
2012 Mar 30
1
Akaike's Final Prediction Error (FPE)
Hello, first of all I have found lots of different versions of the FPE which have given me different results. I was wondering if there was an explicit command in R to compute the FPE of a model. Thank you in advance, Jonny -- View this message in context: http://r.789695.n4.nabble.com/Akaike-s-Final-Prediction-Error-FPE-tp4519011p4519011.html Sent from the R help mailing list archive at
2008 Dec 11
2
negative binomial lmer
Hi; I am running generalized linear mixed models (GLMMs) with the lmer function from the lme4 package in R 2.6.2. My response variable is overdispersed, and I would like (if possible) to run a negative binomial GLMM with lmer if possible. I saw a posting from November 15, 2007 which indicated that there was a way to get lmer to work with negative binomial by assigning: family =
2005 Oct 29
2
LaTex error when creating DVI version when compiling package
Dear Listers, I got this message when compiling a package: * creating pgirmess-manual.tex ... OK * checking pgirmess-manual.text ... ERROR LaTex errors when creating DVI version. This typically indicates Rd problems. The message is quite explicit but I struggled a lot before understanding that the trouble comes from a single file "selMod.rd" among 44 topics. Even though I have
2004 Mar 09
4
aic calculation
hello, could somebody refer me to the reason R uses -2*loglik + 2*(#param)+2 to calculate AIC? thank you -- Stoyan Iliev
2005 Feb 24
2
Forward Stepwise regression based on partial F test
I am hoping to get some advise on the following: I am looking for an automatic variable selection procedure to reduce the number of potential predictor variables (~ 50) in a multiple regression model. I would be interested to use the forward stepwise regression using the partial F test. I have looked into possible R-functions but could not find this particular approach. There is a function
2005 May 23
1
comparing glm models - lower AIC but insignificant coefficients
Hello, I am a new R user and I am trying to estimate some generalized linear models (glm). I am trying to compare a model with a gaussian distribution and an identity link function, and a poisson model with a log link function. My problem is that while the gaussian model has significantly lower (i.e. "better") AIC (Akaike Information Criterion) most of the coefficients are not
2004 Jan 16
2
individual likelihoods
Dear all, is there a way to extract individual likelihoods from a glm/lrm object? By individual likelihoods, I mean the likelihoods whose product give the overall likelihood of the model. I guess the code in the base package, used to compute the Akaike Information Criterion may help me. However, I couldn't figure it out, probably because I'm rather new to likelihood theory and ML