"Rich Ulrich" wrote in message news:***@4ax.com...
On Mon, 8 Sep 2014 15:30:01 -0700 (PDT), Fijoy Vadakkumpadan
Post by Fijoy VadakkumpadanHi Rich,
1.http://scitation.aip.org/content/aapm/journal/medphys/34/11/10.1118/1.2794176
2. http://www.ncbi.nlm.nih.gov/pubmed/19761098
They are not specifically comparing AIC to LRT, but since LRT uses
F-distribution, they are relevant to the question. Have you seen these
studies? If yes, do you think that they provide evidence for the
superiority of AIC?
First, I want to say that my recent knowledge of AIC has been
doubled, at least, by reading the Wikip article and David's post.
So I am not expert. If you want to use AIC, I suggest that you
find a good model in your own area. I don't remember ever seeing
an article with AIC in my own part of biostatistics, so I can't say
how it is used, or when.
And I think that I would distrust those two articles on the same
simulations -- though I judge them only by the abstracts, so I
could be wrong.
From David and from Wikip, I gather that the proper use of AIC
is to compare to choices that are of equal a-priori relevance and
desirability; it is not a test.
The authors of your articles also throw a bone in the direction of
"choice of models"; but it seems to me that they say "AIC is better"
because AICc favored models with fewer parameters. That ...
(a) would seem to be a bias, rather than an advantage; and
(b) would seem to be a matter of "tuning" what their approach
based on an F-test.
Without seeing the articles, I don't *know* what they are comparing.
As Bruce mentions, the LRT is ordinarily tested by chisquared.
However, I do get this clue from the Wikip article, followed by my
own inspired guessing.
- Near the end of the Wikip article on AIC -
"Further comparison of AIC and BIC, in the context of regression, is
given by Yang (2005). In particular, AIC is asymptotically optimal
in selecting the model with the least mean squared error, ..."
Now, in order to select the model with the least mean squared error,
you use the regression stepwise criterion which is equivalent to:
"Enter the new variable if the F-to-enter is greater than 1.0."
Keep in mind that the 1.0 is the average value for a random
predictor.
Further: The equations, if you can figure them out, show that the
AIC criterion (not AICc, the corrected AIC) is closely analogous
to using F-to-enter of 1.0. Thus, I figure when the authors say
that they are using F, they must be using that criterion of 1.0.
However, if you *prefer* fewer variables, apparently you should
not use a symmetric procedure like AIC. In the (explicit) interest of
parsimony, a default for "stepwise regression" was often taken as
a test of F at 5%, not at 50%.
Getting back to the article:
The authors, however, are testing AICc, which builds in a clear
bias (compared to AIC) against models with more parameters,
especially with small sample Ns (which is what they test).
So: I would not be surprised if there modeling would have had
*exactly* the same choices between AIC and their F-test method;
in that case, what they demonstrate is the tuning difference between
AIC and AICc.
I suppose that what AICc gives you is the justification hidden
inside the "black box" of obscurity, for using a criterion that is a
little stiffer than F=1.0 in deciding to accept a larger model.
... I end with apologies to the authors, for however much I
have mis-called what they were doing.
(David - Does this make sense?)
Rich Ulrich
==========================
It looks OK to me. In fact something similar is covered in Lecture 15 in the
set of lecture notes you linked to in your subsequent post:
http://myweb.uiowa.edu/cavaaugh/ms_seminar.html
Overall these seem very good in covering the background of the main
variants.
David Jones