For a parametric model of distributions, the closest distribution in the model to the true distribution located outside the model is considered. Measuring the closeness between two distributions with the Kullback-Leibler (K-L) divergence, the closest distribution is called the "information projection." The estimation risk of the maximum likelihood estimator (MLE) is defined as the expectation of K-L divergence between the information projection and the predictive distribution with plugged-in MLE. Here, the asymptotic expansion of the risk is derived up to $n^{-2}$-order, and the sufficient condition on the risk for the Bayes error rate between the true distribution and the information projection to be lower than a specified value is investigated. Combining these results, the "$p-n$ criterion" is proposed, which determines whether the MLE is sufficiently close to the information projection for the given model and sample. In particular, the criterion for an exponential family model is relatively simple and can be used for a complex model with no explicit form of normalizing constant. This criterion can constitute a solution to the sample size or model acceptance problem. Use of the $p-n$ criteria is demonstrated for two practical datasets. The relationship between the results and information criteria is also studied.
翻译:对于参数化的分布模型,我们考虑模型外真实分布的最近似分布。使用Kullback-Leibler(K-L)散度度量两个分布间的接近程度时,最近似分布被称为“信息投影”。最大似然估计(MLE)的估计风险定义为信息投影与代入MLE的预测分布之间K-L散度的期望。本文推导了风险直至$n^{-2}$阶的渐近展开式,并研究了使真实分布与信息投影之间的贝叶斯错误率低于指定值时所需的风险充分条件。结合这些结果,提出了“$p-n$判定准则”,该准则可判断对于给定模型与样本,MLE是否足够接近信息投影。特别地,指数族模型的判定准则相对简洁,适用于归一化常数无显式表达的复杂模型。此准则可为样本量确定或模型接受问题提供解决方案。文中通过两个实际数据集展示了$p-n$准则的应用,并探讨了该结果与信息准则之间的关联。