Hannan-Quinn Information Criterion
Jump to navigation
Jump to search
A Hannan-Quinn Information Criterion is a criterion for model selection .
- AKA: HQC.
- See: Deviance Information Criterion, Focused Information Criterion, Akaike Information Criterion, Bayesian Information Criterion, Kullback–Leibler Divergence.
References
2016
- (Wikipedia, 2016) ⇒ http://en.wikipedia.org/wiki/Hannan–Quinn_information_criterion Retrieved 2016-08-13
- In statistics, the Hannan–Quinn information criterion (HQC) is a criterion for model selection. It is an alternative to Akaike information criterion (AIC) and Bayesian information criterion (BIC). It is given as
- [math]\displaystyle{ \mathrm{HQC} = -2 L_{max} + 2 k \ln(\ln(n)), \ }[/math]
- where [math]\displaystyle{ L_{max} }[/math] is the log-likelihood, k is the number of parameters, and n is the number of observations.
- Burnham & Anderson (2002, p. 287) say that HQC, "while often cited, seems to have seen little use in practice". They also note that HQC, like BIC, but unlike AIC, is not an estimator of Kullback–Leibler divergence. Claeskens & Hjort (2008, ch. 4) note that HQC, like BIC, but unlike AIC, is not asymptotically efficient, and further point out that whatever method is being used for fine-tuning the criterion will be more important in practice than the term ln ln n, since this latter number is small even for very large n.