Statalist


[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

st: RE: How to compare performance (goodness-of-fit) of very different modelling approaches?


From   "Nick Cox" <[email protected]>
To   <[email protected]>
Subject   st: RE: How to compare performance (goodness-of-fit) of very different modelling approaches?
Date   Thu, 15 Jan 2009 19:17:26 -0000

An objection in principle to correlation, meaning correlation between
observed and fitted response, is that it ignores bias. Thus a model that
always predicted 2y for response y would yield a correlation of 1,
ignoring the bias. That is one reason for preferring to use concordance
correlation, as implemented e.g. in -concord- from the SJ. 

In practice in problems like these gross bias appears rare and when it
occurs attributable to some major programming error. So correlation (and
indeed its square) remains fairly attractive. There is a nice paper by
Zheng and Agresti pointing out its simple virtues. The reference, and
some other comments, are in 

http://www.stata.com/support/faqs/stat/rsquared.html

Another objection to correlation is naturally that it takes no account
of model complexity and so we have various criteria that penalise for
the number of parameters (even though complexity has more dimensions
than that). You don't state your precise objections, but I often
encounter statements of the form "?IC is known to favour models that are
too complicated" (or "too simple"), but less often see explanations of
what equally objective criteria tell the researcher that is so, or
explanations of why researchers use criteria they believe to be
systematically flawed. 

My main positive suggestion is to suggest adding a graphical dimension
to model assessment. I take it that your response is essentially
continuous, in which case the best single kind of graph I suggest to be
a graph of residual versus fitted. Smoothing that in some way can spot
systematic structure missed by the model. Even if the models look
equally good (or bad) you will then have the interesting task of
discussing which model makes most scientific (in your case economic)
sense. 

Nick 
[email protected] 

Eva Poen

currently I am working on slightly complicated mixture models for my
data. My outcome variable is bounded between 0 and 20, and has mass at
either end of the interval. Whether or not I analyse the data on the
original [0,20] scale or a transformation to [0,1] (fractions) does
not make any difference to me.

My question concerns the goodness of fit. I would like to compare the
goodness fit of the complicated finite mixture model to much simpler
models, e.g. the tobit model, the glm model. and a hurdle
specification. Since the likelihood values of these models differ
substantially, likelihood based measures such as BIC appear to be
inadequate for the purpose. Also, measures that compare the model
likelihood of the fitted model to the null likelihood ("pseudo r2")
are difficult sine I can calculate them for the tobit and glm models,
but not for the mixture model, as it is unclear what the null model
would be.

So far I have been looking at crude measures like correlation between
predicted outcome and actual outcome, but I feel that this is
inadequate, especially since the outcome variable is bounded. I'd be
grateful for hints and comments. I am working with Stata 9.2.

*
*   For searches and help try:
*   http://www.stata.com/help.cgi?search
*   http://www.stata.com/support/statalist/faq
*   http://www.ats.ucla.edu/stat/stata/



© Copyright 1996–2024 StataCorp LLC   |   Terms of use   |   Privacy   |   Contact us   |   What's new   |   Site index